Вы находитесь на странице: 1из 364

CMA Part 1

Volume 2: Sections D – F

Financial Planning,
Performance and, Analytics

Version 20.03
HOCK international books are licensed only for individual use and may not be
lent, copied, sold or otherwise distributed without permission directly from
HOCK international.

If you did not download this book directly from HOCK international, it is not a
genuine HOCK book. Using genuine HOCK books assures that you have complete,
accurate and up-to-date materials. Books from unauthorized sources are likely outdated
and will not include access to our online study materials or access to HOCK teachers.

Hard copy books purchased from HOCK international or from an authorized


training center should have an individually numbered orange hologram with the
HOCK globe logo on a color cover. If your book does not have a color cover or does
not have this hologram, it is not a genuine HOCK book.
2020 Edition

CMA
Preparatory Program

Part 1
Volume 2: Sections D – F

Financial Planning,
Performance and Analytics

Brian Hock, CMA, CIA


and
Lynn Roden, CMA
with
Kevin Hock
HOCK international, LLC
P.O. Box 6553
Columbus, Ohio 43206

(866) 807-HOCK or (866) 807-4625


(281) 652-5768

www.hockinternational.com
cma@hockinternational.com

Published August 2019

Acknowledgements

Acknowledgement is due to the Institute of Certified Management Accountants for


permission to use questions and problems from past CMA Exams. The questions and
unofficial answers are copyrighted by the Certified Institute of Management Accountants
and have been used here with their permission.

The authors would also like to thank the Institute of Internal Auditors for permission to
use copyrighted questions and problems from the Certified Internal Auditor Examinations
by The Institute of Internal Auditors, Inc., 247 Maitland Avenue, Altamonte Springs,
Florida 32701 USA. Reprinted with permission.

The authors also wish to thank the IT Governance Institute for permission to make use
of concepts from the publication Control Objectives for Information and related
Technology (COBIT) 3rd Edition, © 2000, IT Governance Institute, www.itgi.org.
Reproduction without permission is not permitted.

© 2019 HOCK international, LLC

No part of this work may be used, transmitted, reproduced or sold in any form or by any
means without prior written permission from HOCK international, LLC.

ISBN: 978-1-934494-71-4
Thanks

The authors would like to thank the following people for their assistance in the
production of this material:

§ All of the staff of HOCK Training and HOCK international for their patience in the
multiple revisions of the material,
§ The students of HOCK Training in all of our classrooms and the students of HOCK
international in our Distance Learning Program who have made suggestions,
comments and recommendations for the material,
§ Most importantly, to our families and spouses, for their patience in the long hours
and travel that have gone into these materials.

Editorial Notes

Throughout these materials, we have chosen particular language, spellings, structures


and grammar in order to be consistent and comprehensible for all readers. HOCK study
materials are used by candidates from countries throughout the world, and for many,
English is a second language. We are aware that our choices may not always adhere to
“formal” standards, but our efforts are focused on making the study process easy for all
of our candidates. Nonetheless, we continue to welcome your meaningful corrections and
ideas for creating better materials.

This material is designed exclusively to assist people in their exam preparation. No


information in the material should be construed as authoritative business, accounting or
consulting advice. Appropriate professionals should be consulted for such advice and
consulting.
Dear Future CMA:
Welcome to HOCK international! You have made a wonderful commitment to yourself
and your profession by choosing to pursue this prestigious credential. The process of
certification is an important one that demonstrates your skills, knowledge, and
commitment to your work.
We are honored that you have chosen HOCK as your partner in this process. We know
that this is a great responsibility, and it is our goal to make this process as efficient as
possible for you. To do so, HOCK has developed the following tools for your use:
Ÿ A Study Plan that guides you, week by week, through the study process. You
can also create a personalized study plan online to adapt the plan to fit your
schedule. Your personalized plan can also be emailed to you at the beginning of
each week.
Ÿ The Textbook that you are currently reading. This is your main study source and
contains all of the information necessary to pass the exam. This textbook follows
the exam contents and provides all necessary background information so that you
don’t need to purchase or read other books.
Ÿ The Flash Cards include short summaries of main topics, key formulas and
concepts. You can use them to review whenever you have a few minutes, but
don’t want to take your textbook along.
Ÿ ExamSuccess contains original questions and questions from past exams that
are relevant to the current syllabus. Answer explanations for the correct and in-
correct answers are also included for each question.
Ÿ Practice Essays taken from past CMA Exams that provide the opportunity to
practice the essay-style questions on the Exam.
Ÿ A Mock Exam enables you to make final preparations using questions that you
have not seen before.
Ÿ Teacher Support via our online student forum, e-mail, and telephone through-
out your studies to answer any questions that may arise.
Ÿ Videos using a multimedia learning platform that provide the same coverage as
a live-taught course, teaching all of the main topics on the exam syllabus.
We understand the commitment that you have made to the exams, and we will match
that commitment in our efforts to help you. Furthermore, we understand that your time
is too valuable to study for an exam twice, so we will do everything possible to make
sure that you pass the first time.
I wish you success in your studies, and if there is anything I can do to assist you, please
contact me directly at brian.hock@hockinternational.com.
Sincerely,

Brian Hock, CMA, CIA


President and CEO
CMA Part 1 Table of Contents

Table of Contents

Section D – Cost Management ......................................................................................... 1


D.1. Measurement Concepts ............................................................................................. 2
Why Cost Management? 2
Classifications of Costs 3
Product Costs vs. Period Costs 3
Costs Based on Level of Activity (Fixed, Variable and Mixed Costs) 5
Introduction to Costing Systems 11
Introduction to Cost Accumulation Methods 12
Introduction to Cost Measurement Systems 12
Cost of Goods Sold (COGS) and Cost of Goods Manufactured (COGM) 20
Variable and Absorption Costing ................................................................................... 22
Fixed Factory Overheads Under Absorption Costing 22
Fixed Factory Overheads Under Variable Costing 23
Effects of Changing Inventory Levels 23
Income Statement Presentation 25
Joint Products and Byproducts ..................................................................................... 33
Methods of Allocating Costs to Joint Products 33
Accounting for Byproducts 41
D.2. Costing Systems ...................................................................................................... 44
Review of Introduction to Costing Systems 44
Process Costing .............................................................................................................. 44
Steps in Process Costing 46
Process Costing Diagram – FIFO 59
Process Costing Diagram – Weighted Average 60
Process Costing Summary 61
Process Costing Examples 62
Spoilage in Process Costing 67
Job-Order Costing ........................................................................................................... 72
Operation Costing ........................................................................................................... 74
Life-Cycle Costing ........................................................................................................... 74
Customer Life-Cycle Costing.......................................................................................... 76
D.3. Overhead Costs........................................................................................................ 77
Introduction to Accounting for Overhead Costs 77
Traditional (Standard) Allocation Method 79
Activity-Based Costing 91
Shared Services Cost Allocation 98
Allocating Costs of a Single (One) Service or Support Department to Multiple Users 99
Allocating Costs of Multiple Shared Service Departments 102

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. i
Table of Contents CMA Part 1

Estimating Fixed and Variable Costs........................................................................... 109


High-Low Points Method 109
D.4. Supply Chain Management ................................................................................... 112
What is Supply Chain Management? 112
Lean Resource Management 113
Just-in-Time (JIT) Inventory Management Systems 117
Kanban 119
Introduction to MRP, MRPII, and ERP 120
Outsourcing 123
Theory of Constraints (TOC) 123
Capacity Level and Management Decisions 135
D.5. Business Process Improvement........................................................................... 138
The Value Chain and Competitive Advantage 138
Process Analysis 144
Business Process Reengineering 144
Benchmarking Process Performance 146
Activity-Based Management (ABM) 146
The Concept of Kaizen 147
The Costs of Quality 148
ISO 9000 158
Quality Management and Productivity 158
Other Quality Related Issues 158
Accounting Process Redesign 161

Section E – Internal Controls ....................................................................................... 166


E.1. Governance, Risk, and Compliance ..................................................................... 167
Corporate Governance 167
Responsibilities of the Board of Directors 173
Audit Committee Requirements, Responsibilities and Authority 173
Responsibilities of the Chief Executive Officer (CEO) 176
Election of Directors 176
Internal Control .............................................................................................................. 177
Internal Control Definition 177
The Importance of Objectives 179
Who Is Responsible for Internal Control? 179
Components of Internal Control 180
What is Effective Internal Control? 190
Transaction Control Objectives 190
Types of Transaction Control Activities 190
Safeguarding Controls 191
Legislative Initiatives About Internal Control ............................................................. 195

ii © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
CMA Part 1 Table of Contents

Foreign Corrupt Practices Act (FCPA) 195


Sarbanes-Oxley Act and the PCAOB 196
External Auditors’ Responsibilities and Reports ........................................................ 202
Financial Statement Opinion 202
Internal Control Opinion 204
Reviews and Compilations 204
Reports to the Audit Committee of the Board of Directors 205
E.2. Systems Controls and Security Measures ........................................................... 206
Introduction to Systems Controls 206
Threats to Information Systems 207
The Classification of Controls 208
General Controls 209
Application Controls 217
Controls Classified as Preventive, Detective and Corrective 225
Controls Classified as Feedback, Feedforward and Preventive 226
System and Program Development and Change Controls ........................................ 227
Internet Security ............................................................................................................ 232
Viruses, Trojan Horses, and Worms 233
Cybercrime 234
Business Continuity Planning ...................................................................................... 238
System Auditing ............................................................................................................ 240
Assessing Controls by Means of Flowcharts 240
Computerized Audit Techniques 241

Section F – Technology and Analytics ........................................................................ 244


Introduction to Technology and Analytics 244
F.1. – Information Systems ........................................................................................... 244
The Value Chain and the Accounting Information System 244
The Supply Chain and the Accounting Information System 245
Automated Accounting Information Systems (AIS) 245
Databases 258
Enterprise Resource Planning Systems 261
Data Warehouse, Data Mart, and Data Lake 264
Enterprise Performance Management 266
F.2. Data Governance .................................................................................................... 267
Definition of Data Governance 267
IT Governance and Control Frameworks 267
Data Life Cycle 278
Records Management 279
Cyberattacks 280

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. iii
Table of Contents CMA Part 1

Defenses Against Cyberattack 281


F.3. Technology-enabled Finance Transformation ..................................................... 285
Systems Development Life Cycle (SDLC) 285
Business Process Analysis 285
Robotic Process Automation (RPA) 286
Artificial Intelligence (AI) 288
Cloud Computing 290
Blockchains, Distributed Ledgers, and Smart Contracts 293
F.4. Data Analytics......................................................................................................... 301
Business Intelligence (BI) 302
Data Mining 306
Analytic Tools 313
Visualization 336

Answers to Questions ................................................................................................... 346

iv © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Section D – Cost Management

Section D – Cost Management


Section D represents 15% of the Part 1 Exam. Section D focuses on the process of determining how much
it costs to produce a product. Topics covered include several types of cost accumulation, cost measurement
and cost allocation systems as well as sources of operational efficiency and business process performance
for a firm. Important concepts in the business process performance topic are the concepts of competitive
advantage and how a firm can attain competitive advantage.

Major topics include:

• Variable and Absorption Costing

• Joint Product and Byproduct Costing

• Process Costing

• Job Order Costing

• Operation Costing

• Life-cycle Costing

• Overhead Cost Allocation

• Activity-based Costing

• Shared Services Cost Allocation

• Estimating Fixed and Variable Costs

• Supply Chain Management

• Business Process Improvement

The three topics that will be the most challenging are:

1) Variable and Absorption Costing

2) Process Costing

3) Activity-based Costing

The three topics above are not the only important topics, of course. The other topics are also important
and will be tested. However, variable and absorption costing, process costing, and activity-based costing
are areas where candidates will need to spend more time to ensure full understanding for the exam.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 1
D.1. Measurement Concepts CMA Part 1

D.1. Measurement Concepts


Why Cost Management?
Cost management systems are used as basic transaction reporting systems and for external financial re-
porting. Cost management systems not only provide reliable financial reporting, but they also track costs
in order to provide information for management decision-making.

The most important function of cost management is to help management focus on factors that make the
firm successful. The management accountant is an integral part of management, identifying, summarizing,
and reporting on the critical success factors necessary for the firm’s success. Critical success factors are
a limited number of characteristics, conditions, or variables that have a direct and important impact on the
efficiency, effectiveness and viability of an organization. They are the aspects of the company’s performance
that are essential to its competitive advantage1 and therefore to its success. Activities related to the critical
success factors must be performed at the highest possible level of excellence.

For example, the management accountant can provide information about the sources of the firm’s compet-
itive advantage, such as the cost, productivity or efficiency advantage the firm has relative to competitors
or about the additional prices the company can charge for additional features that make its offering distinc-
tive relative to the costs of adding those features. Strategic cost management is cost management that
specifically focuses on strategic issues such as these. Thus, cost management contributes to the company’s
achieving its strategic goals and objectives.

Evaluating Operating Performance


In determining whether a firm is achieving its goals and objectives, two aspects of operations are important:
effectiveness and efficiency.

As was emphasized in the section on Strategic Planning in Section B, “a publicly-owned for-profit company
must have maximizing shareholder value as its ultimate goal. The shareholders are the owners. They have
provided risk capital with the expectation that the managers will pursue strategies that will give them good
returns on their investments. Thus, managers have an obligation to invest company profits in such a way
as to maximize shareholder value.” Effectiveness and efficiency are important aspects of fulfilling that ob-
ligation.

An effective operation is one that achieves or exceeds the goals set for the operation. Maximizing share-
holder value is the ultimate goal. Effectiveness in reaching its goals can be measured by analyzing the
firm’s effectiveness in attaining its critical success factors. Critical success factors may be a desired level of
operating income, an increase in market share, new products introduced, or a specified return on invest-
ment. The master budget states the desired operating income for the period, and comparing actual results
with the planned results is a basic starting point for evaluating the effectiveness of the firm in attaining its
profitability goals.

An efficient operation is one that makes effective use of its resources in carrying out the operation. If a
firm attains its goal of increasing sales but it spends more of its resources than necessary to attain that
goal, the firm may be effective but it is not efficient. Alternatively, a firm may be efficient in its use of
resources, spending less than planned per unit sold, but if the firm’s goals for profitability and growth are
not achieved because sales are too low, the firm’s operation was not effective.

Therefore, assessments of efficiency are independent from assessments of effectiveness. Cost management
aids in assessing both effectiveness and efficiency.

1
Competitive advantage is an advantage that a company has over its competitors that it gains by offering consumers
greater value than they can get from its competitors.

2 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.1. Measurement Concepts

Classifications of Costs
It is important to understand the different types, classifications and treatments of costs.

The Difference Between Costs and Expenses


Costs and expenses are two different things.

1) Costs are resources given up to achieve an objective.

2) Expenses are costs that have been charged against revenue in a specific accounting period.

“Cost” is an economic concept, while “expense” is an accounting concept. A cost need not be an expense,
but every expense was a cost before it became an expense. Most costs eventually do become expenses,
such as manufacturing costs that reach the income statement as cost of goods sold when the units they
are attached to are sold, or the cost of administrative fixed assets that have been capitalized on the balance
sheet and subsequently expensed over a period of years as depreciation.

However, some costs do not reach the income statement. Implicit costs2 such as opportunity costs3 never
become expenses in the accounting records, but they are costs nonetheless because they represent re-
sources given up to achieve an objective.

Product Costs vs. Period Costs


Costs are classified according to their purpose. The two main classifications of costs based on purpose are
product (or production) costs and period costs.

Product Costs (also called Inventoriable Costs)


Product costs, or inventoriable costs, are costs for the production process without which the product could
not be made. Product costs are “attached” to each unit and are carried on the balance sheet as inventory
during production (as work-in-process inventory) and when production is completed (as finished goods
inventory) until the unit is sold. When a unit is sold, the item’s cost is transferred from the balance sheet
to the income statement where it is classified as cost of goods sold, which is an expense.

The main types of product costs are: 1) direct materials, 2) direct labor, and 3) manufacturing over-
head (both fixed and variable). These different product costs can be combined and given different names
as outlined in the tables below. Candidates need to know what types of costs are included in the different
classifications.

Direct labor and manufacturing overhead are called conversion costs. They are the costs of converting
the direct materials to finished goods in the production process.

Note: The definition above of product cost is the definition used for financial reporting purposes. How-
ever, other types of costs are considered “product costs” for pricing and other purposes, and those will
be discussed later.

2
Implicit costs are costs that do not involve any specific cash payment and are not recorded in the accounting records.
3
An opportunity cost is the contribution to income that is lost by not using a limited resource in its best alternative use.
An opportunity cost is a type of implicit cost. Both implicit costs and opportunity costs are discussed in more detail later.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 3
D.1. Measurement Concepts CMA Part 1

Types of Product Costs


The costs that follow are the main costs incurred in the production process.

Direct labor Direct labor costs are the costs of labor that can be directly traced to the production
of a product. The cost of wages for production workers is a direct labor cost for a
manufacturing company.

Direct material Direct materials are the raw materials that are directly used in producing the fin-
ished product. The costs included in the direct material cost are all of the costs
associated with acquiring the raw material: the raw material itself, shipping-in
cost, and insurance, among others. Common examples of direct materials are
plastic and components.

Manufacturing Manufacturing overhead costs are the company’s costs related to the production
overhead process that are not direct material or direct labor but are necessary costs of pro-
duction. Examples are indirect labor, indirect materials, rework costs, electricity
and other utilities, depreciation of plant equipment, and factory rent.

Indirect labor Indirect labor is labor that is part of the overall production process but does not
come into direct contact with the product. A common example is labor cost for
employees of the manufacturing equipment maintenance department. Indirect la-
bor is a manufacturing overhead cost.

Indirect material Similar to indirect labor, indirect materials are materials that are not the main
components of the finished goods. Examples are glue, screws, nails, and other
materials such as machine oils, lubricants, and miscellaneous supplies that may
not even be physically incorporated into the finished good. Indirect materials are
a manufacturing overhead cost.

Groupings of Product Costs


The five main types of product costs in the previous table can be further combined to create different cost
classifications. The three classifications that candidates need to be aware of for the CMA exam are in the
following table.

Prime costs Prime costs are the costs of direct material and direct labor. Direct material
and direct labor are the direct inputs, or the direct costs of manufacturing.

Manufacturing Manufacturing costs include the prime costs and manufacturing overhead ap-
costs plied. These are all of the costs that need to be incurred in order to produce the
product. Manufacturing costs do not include selling or administrative costs, which
are period costs.

Conversion costs Conversion costs include manufacturing overhead (both fixed and variable) and
direct labor. Conversion costs are the costs required to convert the direct ma-
terials into the final product.

Note: Direct labor is both a prime cost and a conversion cost.

4 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.1. Measurement Concepts

Period Costs, or Nonmanufacturing Overheads


Period costs are costs for activities other than the actual production of the product. Period costs are ex-
pensed as they are incurred.

The number of period costs is almost unlimited because period costs include essentially everything other
than the product costs, since all costs must be either product costs or period costs. The more commonly-
used examples of period costs include selling, administration, and accounting, but period costs are all the
costs of any department that is not involved in production.

Period costs can be variable, fixed or mixed, but they are not included in the calculation of cost of goods
sold or cost of goods manufactured (both of which are covered later). For financial reporting purposes,
period costs are expensed to the income statement as they are incurred.

However, for internal decision-making, some period costs may be allocated to the production
departments and then to the individual units. This allocation may be done internally so that the com-
pany can set a price for each product that covers all of the costs the company incurs. This type of allocation
will be covered in the topic of Shared Services Cost Allocation.

Note: Overhead allocation of period costs to production is not reported as such in the external financial
statements issued by the company, because it is not proper to do so according to U.S. GAAP or IFRS.
According to both U.S. GAAP and IFRS, period costs should be expensed in the period in which they are
incurred. Allocation of period costs to production would be used for internal decision-making purposes
only.

The number of classifications of period expenses that a company uses on its income statement will depend
on the company. Examples include general and administrative expense, selling expense, accounting, de-
preciation of administrative facilities, and so on.

Costs Based on Level of Activity (Fixed, Variable and Mixed Costs)


In the following table are the main groups of costs based on their behavior as the level of activity
changes. An activity is an event, task, or unit of work with a specified purpose. “Activity” in production
can refer to the number of units of a resource used such as hours of direct labor, or it can refer to the
number of units of product produced. Both production costs and period costs can be classified based on
activity level, though the type of activity used in the classification of period costs is different from that used
for production costs. For period costs, “activity” frequently refers to number of units sold, though it can be
used for any type of activity that incurs costs.

For the following three types of costs, candidates need to know both how the cost per unit of activity
changes and how the total cost changes as the level of activity changes.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 5
D.1. Measurement Concepts CMA Part 1

Fixed, Variable, and Mixed Costs


Fixed costs Fixed costs do not change within the relevant range of activity. As long as the
activity level remains within the relevant range, the total amount of fixed
costs does not change with a change in activity level such as production vol-
ume. However, the cost per unit decreases as the activity level increases and
increases as the activity level decreases.

Variable costs Variable costs are costs such as material and labor (among production costs)
or shipping-out costs (among period costs) that are incurred only when the
activity takes place. The per unit variable cost remains unchanged as the
activity increases or decreases while total variable cost increases as the ac-
tivity level increases and decreases as the activity level decreases.
Note: Because discounts are often received when more units are purchased, it
may appear that variable costs per unit decrease as activity increases. How-
ever, companies do not order units of production inputs one at a time. As part
of the budgeting process a company determines how many of a particular input
it will need to purchase during the year, and the cost per unit for that particular
quantity of inputs is used in the budget. Therefore, budgeted variable costs per
unit do not change as the production levels change for the company.
Mixed costs Mixed costs have both a fixed and a variable component. An example of a mixed
cost is a contract for electricity that includes a basic fixed fee that covers a
certain number of kilowatts of usage per month, and usage over that allowance
is billed at a specified amount per kilowatt used. The electricity plan has a fixed
component and a variable component. A mixed cost could also be an allocation
of overhead cost that contains both fixed and variable overheads.

Cost Behavior in the Production Process


Fixed costs, variable costs, and mixed costs behave in fundamentally different ways in the production pro-
cess as the production level changes. It is important for candidates to understand how total costs and costs
per unit change as production changes. This knowledge of the fundamental behavior of fixed and variable
costs will be required in other sections of the CMA Part 1 exam and also in the CMA Part 2 exam. Although
cost behavior is not inherently difficult, it is such an important underlying element of the production process
that it will be discussed in detail.

Variable Costs
Variable costs are incurred only when the company actually produces something. If a company produces
no units (sits idle for the entire period), the company will incur no variable costs. Direct material and direct
labor are usually variable costs.4

• As the production level increases, total variable costs will increase, but the variable cost per
unit will remain unchanged.
• As the production level decreases, total variable costs will decrease, but the variable cost
per unit will remain unchanged.

Note: The selling price per unit minus all unit variable costs is equal to the unit contribution. The unit
contribution is the amount from each sale available to cover fixed costs or to generate operating income
after the fixed costs have been covered. Contribution margin is a measure of contribution as a per-
centage of the sales price.

4
In some situations, direct labor may be considered a fixed cost, for example in the calculation of throughput contribution
margin, covered later under Theory of Constraints in Supply Chain Management, but those situations are not relevant
for this discussion.

6 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.1. Measurement Concepts

Fixed Costs
Fixed costs are costs that do not change in total as the level of production changes, as long as production
remains within the relevant range. The relevant range is the range of production within which the fixed
cost remains unchanged. As long as the production activity remains within the relevant range, an increase
in the number of units produced will not cause an increase in total fixed costs and a decrease in the number
of units produced will not cause a decrease in total fixed costs.

Fixed costs are best explained by using production in a factory as an example. A factory has the capacity
to produce a certain maximum number of units. As long as production is between zero and that maximum
number of units, the fixed cost for the factory will remain unchanged. However, once the level of production
exceeds the capacity of the factory, the company will need to build (or otherwise acquire) a second factory.
Building the second factory will increase the fixed costs as the company moves to another relevant range.

Within the relevant range of production, the total fixed costs will remain unchanged, but the fixed
costs per unit will decrease as the level of production increases.

Note: In an exam question, if a change in volume results in a volume that is outside the relevant range
given in the question, the question will provide information that would enable recalculation of the total
fixed costs at the changed volume. If a question does not mention anything about the relevant range,
assume that any changes in volume are within the relevant range and that fixed costs do not change in
total because of the change in volume.

Note: Over a large enough time period, all costs will behave like variable costs. In the short
term, some costs are fixed (such as a factory), but over a longer period of time, the company may be
able to change its factory situation by expanding or moving to another facility so that the factory cost
also becomes variable.

Note: Period costs can be fixed or variable, and production costs can also be fixed or variable.

Mixed Costs
In reality, many costs are mixed costs. Mixed costs have a combination of fixed and variable elements.
Mixed costs may be semi-variable costs or semi-fixed costs, which are also called step costs or step variable
costs.

A semi-variable cost has both a fixed component and a variable component. It has a basic fixed amount
that must be paid regardless of the amount of activity or even in the event of no activity, and added to that
fixed amount is an amount that varies with activity. Utilities provide an example. Some basic utility ex-
penses are required just to maintain a factory building, even if no production is taking place. Electric service,
water service, and other utilities usually must be continued, so that basic amount is the fixed component
of utilities. If production begins (or resumes), the cost for utilities increases by a variable amount, depend-
ing on the production level, because machines are running and using electricity and people are using the
water. The fixed component does not change, but the total cost increases incrementally by the amount of
the variable component when production activity increases. Another example of a semi-variable cost is the
cost of a salesperson who receives a base salary plus a commission for each sale made. The base salary is
the fixed component of the salesperson’s salary, and the commission is the variable component.

A semi-fixed cost, also called a step cost or a step variable cost, is fixed over a given, small range of
activity, and above that level of activity, the cost suddenly jumps. It stays fixed again for a while at the
higher range of activity, and when the activity moves out of that range, it jumps again. A semi-fixed cost
or step variable cost moves upward in a step fashion, staying at a certain level over a small range and then
moving to the next level quickly. All fixed costs behave this way, and a wholly fixed cost is also fixed only
as long as activity remains within the relevant range. However, a semi-fixed cost is fixed over a smaller
range than the relevant range of a wholly fixed cost.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 7
D.1. Measurement Concepts CMA Part 1

Example: The nursing staff in a hospital is an example of a semi-fixed cost. The hospital needs one
nurse for every 25 patients, so each time the patient load increases by 25 patients an additional nurse
will be hired. When each additional nurse is hired, the total cost of nursing salaries jumps by the amount
of the additional nurse’s salary.

In contrast, hospital administrative staff salaries remain fixed until the patient load increases by 250
patients, at which point an additional admitting clerk is needed. The administrative staff salaries are
wholly fixed costs over the relevant range, whereas the nursing staff salaries are semi-fixed costs be-
cause the relevant range for the administrative staff (250 patients) is greater than the relevant range
for the nursing staff (25 patients).

Note: The difference between a semi-variable and a semi-fixed cost (also called a step cost or a step
variable cost) is that the semi-variable cost starts out at a given base level and moves upward smoothly
from there as activity increases. A semi-fixed cost moves upward in steps.

Total Costs
Total costs consist of total fixed costs plus total variable costs. The lowest possible total cost occurs when
nothing is produced or sold, because at an activity level5 of zero, the only cost will be fixed costs. Total
costs begin at the fixed cost level and rise by the amount of variable cost per unit for each unit of increase
in activity. In theory at least, total costs graph as a straight line that begins at the fixed cost level on
the Y intercept and rises at the rate of the variable cost per unit for each unit of increase in activity.

The cost function for total manufacturing costs is

y = F + Vx

Where: y = Total Costs


F = Fixed Costs
V = Variable Cost Per Unit
x = Total Production

Note: The cost function can also be written as y = Vx + F. The order of the two terms on the right side
of the equals sign is not important.

To illustrate, following is a graph of total manufacturing costs for a company with fixed manufacturing costs
of $700,000 and variable manufacturing costs of $20 per unit produced. Total cost is on the y-axis, while
total production is on the x-axis. The cost function for this company’s total manufacturing costs is

y = 700,000 + 20x

The total cost line on the graph is a straight line beginning at $700,000 on the y-axis where x is zero and
increasing by $200,000 for each production increase of 10,000 units (because 10,000 units multiplied by
$20 equals $200,000). The graph of the above cost function follows. The graph should look familiar, because
the cost function and the line on the graph are the same as those for the equation of a linear regression
line. Linear regression was covered in Section B in Vol. 1 of this textbook in the Forecasting topic under
Trend Projection and Regression Analysis in relation to using simple regression analysis to make forecasts.
This same concept will be discussed again later in this section, in the topic of Estimating Fixed and Variable
Costs.

5
“Activity level” or “level of activity” can be used to refer to various types of activity. It can refer to production volume
in number of units of output, the number of units of inputs to the production process, sales volume, or to any other
activity being performed.

8 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.1. Measurement Concepts

Total Manufacturing Costs


$2,100,000

$1,900,000
Total Manufacturing Costs

y = 700,000 + 20x
$1,700,000

$1,500,000

$1,300,000

$1,100,000

Total Manufacturing Costs


$900,000

$700,000

$500,000

Production Volume

Direct Versus Indirect Costs


Direct costs are costs that can be traced directly to a specific cost object. A cost object is anything
for which a separate cost measurement is recorded. It can be a function, an organizational subdivision, a
contract, or some other work unit for which cost data are desired and for which provision is made to
accumulate and measure the cost of processes, products, jobs, capitalized projects, and so forth. Examples
of direct costs that will be covered in detail are direct materials and direct labor used in the production of
products.

Indirect costs are costs that cannot be identified with a specific cost object. In manufacturing, overhead
is an indirect cost. Manufacturing indirect costs are grouped into cost pools for allocation to units of product
manufactured. A cost pool is a group of indirect costs that are grouped together for allocation on the basis
of the same cost allocation base. Cost pools can range from very broad, such as all plant overhead costs,
to very narrow, such as the cost of operating a specific machine.

Other indirect costs are nonmanufacturing, or period, costs. Examples are support functions such as IT,
maintenance, security, and managerial functions such as executive management and other supervisory
functions.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 9
D.1. Measurement Concepts CMA Part 1

Other Costs and Cost Classifications


In addition to all of the costs and classifications listed above, candidates need to be familiar with several
other types of costs.

Explicit costs Explicit costs are also called out-of-pocket costs. Explicit costs involve payment of
cash and include wages and salaries, office supplies, interest paid on loans, payments
to vendors for raw materials, and so forth. Explicit costs are the opposite of implicit
costs. Most explicit costs eventually become expenses, though the timing of their recog-
nition as expenses may be delayed, for example when inventory is purchased and its
cost does not become an expense until it is sold.
Implicit costs An implicit cost, also called an imputed cost, is a cost that does not involve any specific
cash payment and is not recorded in the accounting records. Implicit costs are also
called economic costs. Implicit costs do not become expenses. They cannot be specif-
ically included in financial reports, but they are needed for use in a decision-making
process. For example, interest not earned on money that could have been invested in
an interest-paying security but instead was invested in manufacturing equipment is a
frequent implicit or imputed cost. The “lost” interest income is an opportunity cost of
investing in the machine. The lost interest is not reported as an expense on the income
statement, but it is a necessary consideration in making the decision to invest in the
machine, because it will be different if the machine is not purchased.
Opportunity An opportunity cost is a type of implicit cost. “Opportunity cost” is an economics term,
costs and opportunity cost is considered an economic cost. It is the contribution to income
that is lost by not using a limited resource in its best alternative use. Opportunity cost
includes the contribution that would have been earned if the alternative decision had
been made less any administrative expenditures that would have been made for the
other available alternative but that will not be made.
Anytime money is invested or used to purchase something, the potential return from
the next best use of that money is lost. Often times, the lost return is interest income.
If money were not used to purchase inventory, for example, it could be deposited in a
bank and could earn interest. The lost interest can be calculated only for the time period
during which the cash flows are different between the two options.
Carrying Carrying costs are the costs the company incurs when it holds inventory. Carrying costs
costs include: rent and utilities related to storage; insurance and taxes on the inventory; costs
of employees who manage and protect the inventory; damaged or stolen inventory; the
lost opportunity cost of having money invested in inventory (called cost of capital); and
other storage costs.
Because storage of inventory does not add value to the inventory items, carrying costs
are expensed on the income statement as incurred. They are not added to the cost of
the inventory and thus are not included on the balance sheet.
Sunk costs Sunk costs are costs that have already been incurred and cannot be recovered. Sunk
costs are irrelevant in any decision-making process because they have already been
incurred and no present or future decision can change that fact.
Committed Committed costs are costs for the company’s infrastructure. They are costs required to
costs establish and maintain the readiness to do business. Examples of committed costs are
intangible assets such as the purchase of a franchise and the purchase of fixed assets
such as property, plant, and equipment. They are fixed costs that are usually on the
balance sheet as assets and become expenses in the form of depreciation and amorti-
zation.

10 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.1. Measurement Concepts

Discretionary Discretionary costs, sometimes called flexible costs, are costs that may or may not be
costs, also incurred by either engaging in an activity or not engaging in it, at the discretion of the
sometimes manager bearing the cost. In the short term, not engaging in a discretionary activity
called flexible will not cause an adverse effect on the business. However, in the long run the activities
costs are necessary and the money does need to be spent. Discretionary cost decisions are
made periodically and are not closely related to input or output decisions. Furthermore,
the value added and the benefits obtained from spending the money cannot be precisely
defined. Advertising, research and development (R&D), and employee training are usu-
ally given as examples of discretionary costs. Discretionary costs, or flexible costs, may
be fixed costs, variable costs, or mixed costs.
Marginal Marginal costs are the additional costs necessary to produce one more unit.
costs

Note: When overtime is worked by production employees, the overtime premium paid to the workers
is considered to be factory overhead. The overtime premium is the amount of increase in the wages
paid per hour for the overtime work.

For example, direct labor is paid $20 per hour for regular hours and is paid and time and-a-half, or $30
per hour, for overtime hours worked in excess of 40 hours per week. Ten hours of overtime are needed
in a given week to complete the week’s required production. The regular rate of $20 per hour multiplied
by 10 hours, or $200, is classified as a direct labor cost, even though it is worked in excess of regular
hours. The half-time premium of $10 additional paid per hour multiplied by 10 hours, or $100, is
classified as factory overhead.

The half-time premium of $100 is not charged to the particular units worked on during the overtime
hours, because the units worked on during the overtime hours could just as easily have been different
units if the jobs to be done had simply been scheduled differently. As overhead, the overtime premium
paid is allocated equally among all units produced during the period.

However, if the need to work overtime results from a specific job or customer request, the overtime
premium should be charged to that specific job as part of the direct labor cost of that job and not
included in the overall overhead amount to be allocated.

Introduction to Costing Systems


Product costing involves accumulating, classifying and assigning direct materials, direct labor, and factory
overhead costs to products, jobs, or services.

In developing a costing system, management accountants need to make choices in three categories of
costing methods:

1) The cost measurement method to use in allocating costs to units manufactured (standard, nor-
mal, or actual costing).

2) The cost accumulation method to use (job costing, process costing, or operation costing).

3) The method to be used to allocate overhead (volume-based or activity-based).

Cost measurement systems are discussed here. Cost accumulation methods and allocation of overhead
will be discussed later. However, a few words about cost accumulation systems are needed before explain-
ing cost measurement systems.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 11
D.1. Measurement Concepts CMA Part 1

Introduction to Cost Accumulation Methods


Cost accumulation systems are used to assign costs to products or services. Job order costing (also called
job costing), process costing, and operation costing are different types of cost accumulation systems used
in manufacturing.

• Process costing is used when many identical or similar units of a product or service are being
manufactured, such as on an assembly line. Costs are accumulated by department or by process.

• Job order costing (also called job costing) is used when units of a product or service are distinct
and separately identifiable. Costs are accumulated by job.

• Operation costing is a hybrid system in which job costing is used for direct materials costs while
a departmental (process costing) approach is used to allocate conversion costs (direct labor and
overhead) to products or services.

Process costing, job order costing, and operation costing will be discussed in more detail later.

Introduction to Cost Measurement Systems


Costs are allocated to units manufactured in three main ways:

1) Standard costing systems

2) Normal costing systems

3) Actual costing systems

These three cost measurement systems are used for allocating both direct manufacturing costs (direct labor
and direct materials) and indirect manufacturing costs (overhead) in order to value the products manufac-
tured.

1) Standard Costing
In a standard cost system, standard, or planned, costs are assigned to units produced. The standard cost
of producing one unit of output is based on the standard cost for one unit of each of the inputs required to
produce that output unit, with each input multiplied by the number of units of that input allowed for one
unit of output. The inputs include direct materials, direct labor and allocated overhead. The standard cost
is what the cost should be for that unit of output.

Direct materials and direct labor are applied to production by multiplying the standard price or rate per
unit of direct materials or direct labor by the standard amount of direct materials or direct labor allowed
for the actual output. For example, if three direct labor hours are allowed to produce one unit and 100
units are actually produced, the standard number of direct labor hours for those 100 units is 300 hours (3
hours per unit × 100 units). The standard cost for direct labor for the 100 units is the standard hourly wage
rate multiplied by the 300 hours allowed for the actual output, regardless of how many direct labor
hours were actually worked and regardless of what actual wage rate was paid. The cost applied
to the actual output is the standard cost allowed for the actual output.

Note: In a standard cost system, the standard quantity of an input allowed for the actual output,
not the actual quantity of the input used for the actual output and the standard price allowed per
unit of the input, not the actual price paid per unit of the input, are used to calculate the amount of the
input’s cost applied to production. Some candidates find that a difficult concept to grasp, because it
requires using the standard price and quantity allowed per unit with the actual number of units pro-
duced.

12 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.1. Measurement Concepts

In a standard cost system, overhead is generally allocated to units produced by calculating a predeter-
mined, or standard, manufacturing overhead rate (a volume-based method) that is applied to the units
produced on the basis of the standard amount of the allocation base allowed for the actual output. When a
traditional method of overhead allocation is used, the standard manufacturing overhead application rate is
calculated as the budgeted overhead cost divided by the budgeted activity level of the allocation base.

The predetermined overhead rate is calculated as follows:

Budgeted Monetary Amount of Manufacturing Overhead


Budgeted Activity Level of Allocation Base

The best cost driver to use as the allocation base is the measure that best represents what causes overhead
cost to be incurred. The most frequently-used allocation bases are direct labor hours, direct labor costs, or
machine hours. For a labor-intensive manufacturing process, the proper base is probably direct labor hours
or direct labor costs. For an equipment-oriented manufacturing process, number of machine hours is the
better allocation base.

To apply overhead cost to production, the predetermined overhead rate is multiplied by the standard
amount of the allocation base allowed for producing one unit of product, and then that standard overhead
amount for one unit is multiplied by the number of units actually produced to calculate the standard over-
head cost to be applied to all the units produced.

Of course, the actual costs incurred will probably be different from the standard costs. The difference is a
variance. The difference is also called an under-applied or over-applied cost. At the end of each ac-
counting period, variances are accounted for in one of two basic ways.

• If the variances are immaterial, they may be closed out 100% to cost of goods sold expense on
the income statement.

• If the variances are material, they should be prorated among cost of goods sold and the relevant
Inventory accounts on the balance sheet according to the amount of overhead included in each
that was allocated to the current period’s production.

If the variances are closed out 100% to cost of goods sold, the cost of the goods in inventories will be equal
to their standard cost only.

Standard costing enables management to compare actual costs with what the costs should have been for
the actual amount produced. Moreover, it permits production to be accounted for as it occurs. Using actual
costs incurred for manufacturing inputs would cause an unacceptable delay in reporting, because those
costs may not be known until well after the end of each reporting period, when all the invoices have been
received.

The emphasis in standard costing is on flexible budgeting, where the flexible budget for the actual produc-
tion is equal to the standard cost per unit of output multiplied by the actual production volume.

Standard costing can be used in either a process costing or a job-order costing environment.

Note: The standard cost for each input per completed unit is the standard rate per unit of input multiplied
by the amount of inputs allowed per completed unit, not multiplied by the actual amount of inputs
used per completed unit.

Standard costing is applicable to a wide variety of companies. Manufacturing companies use standard cost-
ing with flexible budgeting to control direct materials and direct labor costs. Service companies such as
fast-food restaurants use standard costs, too, mainly to control their labor costs since they are labor-
intensive. Increasingly, firms are using Enterprise Resource Planning (ERP) systems to help them track
standard and actual costs and to assess variances in real time. Enterprise Resource Planning is covered
later in this section in Supply Chain Management.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 13
D.1. Measurement Concepts CMA Part 1

2) Normal Costing
In a normal cost system, direct materials and direct labor costs are applied to production differently
from the way they are applied in standard costing. In normal costing, direct materials and direct labor costs
are applied at their actual rates multiplied by the actual amount of the direct inputs used for produc-
tion.

To allocate overhead, a normal cost system uses a predetermined annual manufacturing overhead
rate, called a normal or normalized rate. The predetermined rate is calculated the same way the predeter-
mined rate is calculated under standard costing. However, under normal costing, that predetermined rate
is multiplied by the actual amount of the allocation base that was used in producing the product,
whereas under standard costing, the predetermined rate is multiplied by the amount of the allocation base
allowed for producing the product.

Normal costing is not appropriate in a process costing environment because it is too difficult to
determine the actual costs of the specific direct materials and direct labor used for a specific production
run. Process costing is used when many identical or similar units of a product or service are being manu-
factured, such as on an assembly line. Costs are accumulated by department or by process. In contrast,
job costing accumulates costs and assigns them to specific jobs, customers, projects, or contracts. Job
costing is used when units of a product or service are distinct and separately identifiable. Normal costing
is used mainly in job costing.

The purpose of using a predetermined annual manufacturing overhead rate in normal costing is to normalize
factory overhead costs and avoid month-to-month fluctuations in cost per unit that would be caused by
variations in actual overhead costs and actual production volume. It also makes current costs available. If
actual manufacturing overhead costs were used, those costs might not be known until well after the end of
each reporting period, when all the invoices had been received.

3) Actual Costing
In an actual costing system, no predetermined or estimated or standard costs are used. Instead, the actual
direct labor and materials costs and the actual manufacturing overhead costs are allocated to the units
produced. The cost of a unit is the actual direct cost rates multiplied by the actual quantities of the direct
cost inputs used and the actual indirect (overhead) cost rates multiplied by the actual quantities used of
the cost allocation bases.

Actual costing is practical only for job order costing for the same reasons that normal costing is practical
only for job order costing. In addition, actual costing is seldom used because it can produce costs per unit
that fluctuate significantly. This fluctuation in costs can lead to errors in management decisions such as
pricing of the product, decisions about adding or dropping product lines, and performance evaluations.

14 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.1. Measurement Concepts
efham CMA
Below is a summary of the three cost measurement methods:
Cost Accumu- Dir. Materials/ Dir. Materials/
lation Method Direct Labor Direct Labor Overhead Overhead
Cost Measure- Usually Application Application Application Allocation
ment Method Used With Rate Base Rate Base
Standard
Standard
Process Amount of Allo-
Amount Predetermined
Standard Costing or Standard cation Base
Allowed Standard
Costing Job Order Rate Allowed
for Actual Rate
Costing for Actual
Production
Production

Actual Amount
Actual Amount Estimated of Allocation
Normal Job Order Actual
Used for Actual Normalized Base Used
Costing Costing Rate
Production Rate for Actual
Production

Actual Amount
Actual Amount of Allocation
Actual Job Order Actual Actual
Used for Actual Base Used
Costing Costing Rate Rate
Production for Actual
Production

It is important in answering a question to identify what type of costing the company uses.

• If the company uses standard costing, the costs applied to each unit will be the standard costs
for the standard amount of inputs allowed for production of the actual number of units pro-
duced.

• Actual amount of inputs used for the actual production are used in calculating the costs applied
to each unit only when the company uses normal or actual costing.

Example of standard costing, normal costing, and actual costing used for the same product under the
same set of assumptions:

Log Homes for Dogs, Inc. (LHD) manufactures doghouses made from logs. It offers only one size and
style of doghouse. For the year 20X4, the company planned to manufacture 20,000 doghouses. Over-
head is applied on the basis of direct labor hours, and the direct labor standard is 2 DLH per doghouse.
The company’s planned costs were as follows:
Direct materials $45 per doghouse (5 units of DM/doghouse @ $9/ unit)
Direct labor $30 per doghouse (2 DLH/doghouse @ $15/ DLH)
Variable overhead $10 per doghouse (2 DLH/doghouse @ $5/DLH
Fixed overhead $260,000, or $13 per doghouse (2 DLH/doghouse @ $6.50 per DLH)

LHD actually produced and sold 21,000 doghouses during 20X4.

LHD’s actual costs incurred were:


Direct materials $882,000: 5.25 units of DM used per doghouse @ $8/unit of DM
Direct labor $617,400: 2.1 DLH used per doghouse @ $14/DLH
Variable overhead $224,910
Fixed overhead $264,600
(Continued)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 15
D.1. Measurement Concepts CMA Part 1

Total Costs Applied Under Standard Costing:


Direct materials cost applied: $9 std. cost/unit of DM × 5 units allowed/house × 21,000 = $945,000
Direct labor applied: $15 std. rate/DLH × 2 DLH allowed/house × 21,000 = $630,000
Variable overhead applied: $5 std. rate/DLH × 2 DLH. allowed/house × 21,000 = $210,000
Fixed overhead applied: $6.50 std. rate/DLH × 2 DLH. allowed/house × 21,000 = $273,000

Total Costs Applied Under Normal Costing:


Direct materials cost applied: $8 actual rate/DM unit × 5.25 units used/house × 21,000 = $882,000
Direct labor cost applied: $14 actual rate/DLH. × 2.1 DLH. used/house × 21,000 = $617,400
Variable overhead applied: $5 est. rate/DLH × 2.1 DLH used/house × 21,000 = $220,500
Fixed overhead applied: $6.50 est. rate/DLH × 2.1 DLH used/house × 21,000 = $286,650

Total Costs Applied Under Actual Costing:


Direct materials cost applied: $8 actual rate/DM unit × 5.25 units used/house × 21,000 = $882,000
Direct labor cost applied: $14 actual rate/DLH × 2.1 DLH used/house × 21,000 = $617,400
Variable overhead applied: $5.101 actual rate/DLH × 2.1 DLH used/house × 21,000 = $224,910
Fixed overhead applied: $6.002 actual rate/DLH × 2.1 DLH used/house × 21,000 = $264,600
1
Actual rate calculated as $224,910 ÷ 21,000 = $10.71/house. $10.71 ÷ 2.1 DLH/house = $5.10/DLH
2
Actual rate calculated as $264,600 ÷ 21,000 = $12.60/house. $12.60 ÷ 2.1 DLH/house = $6.00/DLH

The costs applied per unit under each of the cost measurement methods were:
Cost Applied per Unit Under Standard Costing:
Direct materials ($9 std. cost/unit of DM × 5 units of DM allowed) $45.00
Direct labor ($15 std. rate/DLH × 2 DLH allowed) 30.00
Variable overhead ($5/DLH allowed × 2 DLH allowed) 10.00
Fixed overhead ($6.50/DLH allowed × 2 DLH allowed) 13.00
Total cost per unit $98.00

Cost Applied per Unit Under Normal Costing:


Direct materials ($8 actual cost/DM unit × 5.25 units used) $42.00
Direct labor ($14 actual rate/DLH × 2.1 DLH used) 29.40
Variable overhead ($5 est. rate/DLH × 2.1 DLH used) 10.50
Fixed overhead ($6.50 est. rate/DLH × 2.1 DLH used) 13.65
Total cost per unit $95.55

Cost Applied per Unit Under Actual Costing:


Direct materials ($8 actual cost/DM unit × 5.25 units) $42.00
Direct labor ($14 actual rate/DLH × 2.1 DLH used) 29.40
Variable overhead ($5.10 actual rate/DLH × 2.1 DLH used) 10.71
Fixed overhead ($6.00 actual rate/DLH × 2.1 DLH used) 12.60
Total cost per unit $94.71

16 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.1. Measurement Concepts

Effect of Each Cost Measurement System on Variance Reporting


The manufacturing variances calculated will be affected by which cost measurement system is being used.

Effect of Standard Costing on Variance Reporting


When standard costing is being used, all of the manufacturing overhead variances as covered in Section C,
Performance Management, will be calculated, and their interpretation will be as presented in Section C.

Effect of Normal Costing on Variance Reporting


When normal costing is being used, direct materials and direct labor costs are applied to production at
their actual rates multiplied by the actual amount of the direct inputs used in the production after
the end of the period when the actual rates and the actual quantities of direct inputs used in production are
known. Therefore, under normal costing no direct material or direct labor variances can exist be-
cause the actual costs incurred are equal to the costs applied to production.

Under normal costing, overhead is applied at a predetermined overhead rate calculated the same way the
standard overhead rate is calculated under standard costing: budgeted total overhead divided by the num-
ber of hours of the allocation base (direct labor hours or machine hours) allowed for the budgeted output.
However, under normal costing, overhead is applied to production by multiplying that predetermined over-
head rate by the actual amount of the allocation base used in producing the product, instead of multiplying
it by the amount of the allocation base allowed for producing the product, as in standard costing.

Therefore, under normal costing, the difference between the actual variable overhead incurred and the
variable overhead applied consists only of the variable overhead spending variance. No variable overhead
efficiency variance arises under normal costing.

Under normal costing, the fixed overhead spending (flexible budget) variance is the same as it is under
standard costing, because that variance is actual fixed overhead incurred minus budgeted fixed overhead,
and neither of those is affected by the amount of fixed overhead applied.

However, the fixed overhead production-volume variance is different from what it would be under standard
costing because the amount of fixed overhead applied is different. The fixed overhead production-volume
variance is budgeted fixed overhead minus fixed overhead applied, and the amount of fixed overhead
applied under normal costing is based on the amount of the allocation base actually used for the actual
units of output rather than based on the standard amount of the allocation base allowed for the actual
output.

Effect of Actual Costing on Variance Reporting


In an actual costing system, no predetermined or estimated or standard costs are used. Instead, the actual
direct labor and direct materials costs and the actual manufacturing overhead costs are allocated to the
units produced after the end of the period when the actual costs and the actual quantities used of the cost
allocation bases are known. No direct input or variable overhead variances are calculated at all because the
costs applied are the same as the costs incurred.

Under actual costing, the fixed overhead spending (flexible budget) variance is the same as it is under
standard costing because it is calculated as actual fixed overhead incurred minus budgeted fixed overhead.

The fixed overhead production-volume variance is basically meaningless, however, because the fixed over-
head applied would be the same as the fixed overhead incurred. Therefore, the fixed overhead production-
volume variance, which is budgeted fixed overhead minus fixed overhead applied, would be the same as
the fixed overhead spending (flexible budget) variance, except it would have the opposite sign. For exam-
ple, if the fixed overhead spending variance were favorable, the fixed overhead production-volume variance
would be unfavorable by the same amount because the amount of fixed overhead applied would be the
same as the amount of actual fixed overhead incurred.

The two fixed overhead variances combined would net to zero.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 17
D.1. Measurement Concepts CMA Part 1

Question 1: Which one of the following is a variance that could appear if a company uses a normal
costing system?

a) Direct material price variance.

b) Direct labor efficiency variance.

c) Variable overhead spending variance.

d) Variable overhead efficiency variance.

(ICMA 2013)

Benefits and Limitations of Each Cost Measurement System

Benefits of Standard Costing

• Standard costing enables management to compare actual costs with what the costs should have been
for the actual production.

• It permits production to be accounted for as it occurs, since standard costs are used to apply costs
to units produced.

• Standard costing prescribes expected performance and provides control. Standard costs establish
what costs should be, who should be responsible, and what costs are under the control of specific
managers. Therefore, standard costs contribute to management of an integrated responsibility ac-
counting system.

• Standards can provide benchmarks for employees to use to judge their own performance, as long as
the employees view the standards as reasonable.

• Standard costing facilitates management by exception, because as long as the costs remain within
standards, managers can focus on other issues. Variances from standard costs alert managers when
problems require attention, which enables management to focus on those problems.

Limitations of Standard Costing

• Using a predetermined factory overhead rate to apply overhead cost to products can cause total
overhead applied to the units produced to be greater than the actual overhead incurred when pro-
duction is higher than expected; and overhead applied may be less than the amount incurred if actual
production is lower than expected.

• If the variances from the standards are used in a negative manner, for instance to assign blame,
employee morale suffers and employees are tempted to cover up unfavorable variances and to do
things they should not do in order to make sure the variances will be favorable.

• Output in many companies is not determined by how fast the employees work but rather by how fast
the machines work. Therefore, direct labor quantity standards may not be meaningful.

• The use of standard costing could lead to overemphasis on quantitative measures. Whether a vari-
ance is “favorable” or “unfavorable” and the amount of the variance is not the full story.

• There may be a temptation on the part of management to emphasize meeting the standards without
considering other non-financial objectives such as maintaining and improving quality, on-time deliv-
ery, and customer satisfaction. A balanced scorecard can be used to address the non-financial
objectives as well as the financial objectives.

• In environments that are constantly changing, it may be difficult to determine an accurate standard
cost.

• The usefulness of standard costs is limited to standardized processes. The less standardized the
process is, the less useful standard costs are.

18 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.1. Measurement Concepts

Benefits of Normal Costing

• The use of normal costing avoids the fluctuations in cost per unit that occur under actual costing
because of changes in the month-to-month volume of units produced and in month-to-month fluc-
tuations in overhead costs.

• Manufacturing costs of a job are available earlier under a normal costing system than under an actual
costing system.

• Normal costing allows management to keep direct product costs current because actual materials
and labor costs incurred are used in costing the production, while the actual incurred overhead costs
that would not be available until much later are applied based on a predetermined rate.

Limitations of Normal Costing

• Using a predetermined factory overhead rate to apply overhead cost to products can cause total
overhead applied to the units produced to be greater than the actual overhead incurred when pro-
duction is higher than expected; and overhead applied may be less than the amount incurred if actual
production is lower than expected.

• Normal costing is not appropriate for process costing because the actual costs would be too difficult
to trace to individual units produced, so it is used primarily for job costing.

Benefits of Actual Costing

• The primary benefit of using actual costing is that the costs used are actual costs, not estimated
costs.

Limitations of Actual Costing

• Because actual costs must be computed and applied, information is not available as quickly after the
end of a period as it is with standard costing.

• Actual costing leads to fluctuations in job costs because the amount of actual overhead incurred
fluctuates throughout the year.

• Like normal costing, actual costing is not appropriate for process costing because the actual costs
would be too difficult to trace to individual units produced. Therefore, it is used primarily in a job
costing environment.

Note: The focus of the remainder of this section will be on standard costing, because that is the
most commonly used system in manufacturing.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 19
D.1. Measurement Concepts CMA Part 1

Cost of Goods Sold (COGS) and Cost of Goods Manufactured (COGM)


The various classifications of costs are used in the calculation of cost of goods sold (COGS) and cost of
goods manufactured (COGM).

• Cost of goods sold (COGS) is the total of the costs directly attributable to producing items that
were sold during the period. For each unit sold, cost of goods sold includes the direct material
and direct labor costs of the unit and an allocation of a portion of manufacturing overhead costs.

• Cost of goods manufactured (COGM) is the total of the costs directly attributable to producing
items that were brought to completion in the manufacturing process during the period,
whether manufacturing on them was begun before the current period began or during the current
period. For each unit brought to completion, cost of goods manufactured includes the direct ma-
terial and direct labor costs of the unit and an allocation of a portion of manufacturing overhead
costs.

Neither cost of goods sold nor cost of goods manufactured includes indirect selling and administrative costs
such as sales, marketing and distribution, as those are period costs.

Though COGS and COGM are somewhat similar, they are very different in one key aspect. COGS is an
externally reported figure and it is reported on the income statement. It is the cost of producing the units
that were actually sold during the period. COGM, on the other hand, is an internal number and is not
reported on either the balance sheet or the income statement. COGM represents the cost of producing the
units that were completed during the period. COGM is, however, used in the calculation of cost of goods
sold for a company that produces its own inventory. The calculation of both numbers is covered in more
detail below.

The process of calculating the cost of producing an item is a very important one for any company. It is
critical for the calculated cost to represent the complete cost of production. If the company does not calcu-
late the cost of production correctly, it may charge a price for the product that will be incorrect. The result
will be either low sales volume if the price is too high or low profits if the price is too low.

The calculated production cost per unit is reported on the balance sheet as the value of each unit of finished
goods inventory when the items are completed. As the items in inventory are sold, the cost of each item
sold is transferred to the income statement as cost of goods sold.

Due to the need to determine the cost of production accurately, the information that accountants provide
to management regarding the company’s production costs is crucial. Furthermore, it is beneficial to provide
the information quickly and often, so that management can make any necessary adjustments such as
changes in pricing as soon as possible.

Note: Costs that are not production costs are period costs, and period costs are generally expensed as
incurred (for example, inventory carrying costs, general and administrative costs, and so on).

Calculating Cost of Goods Sold


Cost of goods sold represents the cost to produce or purchase the units that were sold during the period.
It is perhaps the largest individual expense item on the income statement, so it is important for cost of
goods sold to be calculated accurately.

Cost of goods sold is calculated using the following formula:

Beginning finished goods inventory


+ Purchases for a reseller or cost of goods manufactured for a manufacturer
− Ending finished goods inventory
= Cost of goods sold

20 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.1. Measurement Concepts

The formula above is a simplification of what is actually occurring because it assumes all of the units in
finished goods inventory at the beginning of the period were either sold during the period or were still in
finished goods inventory at the end of the period, which does not always happen. In reality, some units
may be damaged, stolen or lost. However, for the CMA Part 1 Exam, the above formula is sufficient.

Calculating Cost of Goods Manufactured


The cost of goods manufactured represents the cost of the units completed and transferred out of
work-in-process inventory during the period. COGM does not include the cost of work that was done on
units that were not finished during the period.

Cost of goods manufactured is calculated using the following formula:

Direct Materials Used*


+ Direct Labor Used
+ Manufacturing Overhead Applied
= Total Manufacturing Costs
+ Beginning Work-in-Process Inventory
− Ending Work-in-Process Inventory
= Cost of Goods Manufactured

* Direct Materials Used = Beginning Direct Materials Inventory + Purchases + Transportation-In − Net Returns −
Ending Direct Materials Inventory

The COGM of a manufacturing company will be part of its calculation of cost of goods sold (COGS).

As with the calculation of COGS, the calculation of COGM simplifies reality because it assumes that all items
of work-in-process inventory were either completed during the period or are in ending work-in-process
inventory. In reality, some of the work-in-process inventory may have been lost, damaged or otherwise
not used during the period, and therefore some of the raw materials entered into production or items
worked on during the period will not be in ending inventory. However, for the CMA Part 1 exam, the formula
above is sufficient.

Question 2: The Profit and Loss Statement of Madengrad Mining Inc. includes the following information
for the current fiscal year.

Sales $160,000
Gross profit 48,000
Year-end finished goods inventory 58,300
Opening finished goods inventory 60,190

The cost of goods manufactured by Madengrad for the current fiscal year is

a) $46,110

b) $49,890

c) $110,110

d) $113,890

(ICMA 2010)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 21
Variable and Absorption Costing CMA Part 1

Variable and Absorption Costing


Variable and absorption costing are two different methods of inventory costing. Under both variable and
absorption costing, all variable manufacturing costs (both direct and indirect) are inventoriable costs. The
only two differences between the two methods are in:

1) Their treatment of fixed manufacturing overhead

2) The income statement presentation of the different costs

Note: All other costs except for fixed factory overheads are treated in the same manner under both
variable and absorption costing, although they may be reported in a slightly different manner on the
income statement.

Note: Variable costing can be used only internally for decision-making. Variable costing is not acceptable
under generally accepted accounting principles for external financial reporting. It is also not acceptable
under U.S. tax regulations for income tax reporting.

Fixed Factory Overheads Under Absorption Costing


For a manufacturing company, absorption costing is required for external financial reporting by generally
accepted accounting principles. Under absorption costing, fixed factory overhead costs are allocated to
the units produced during the period according to a predetermined rate. Fixed manufacturing overhead
is therefore a product cost under absorption costing. Product costs are inventoried and they are expensed
as cost of goods sold only when the units they are attached to are sold.

The predetermined fixed overhead rate is calculated as follows:

Budgeted Monetary Amount of Fixed Manufacturing Overhead


Budgeted Activity Level of Allocation Base

The budgeted activity level of the allocation base is the number of budgeted direct labor hours, direct
labor cost, material cost, or machine hours—whatever is being used as the allocation base.

When standard costing is being used, the fixed overhead is applied to the units produced on the basis of
the number of units of the allocation base allowed for the actual output.

Example: Fixed overhead is applied to units produced on the basis of direct labor hours. The standard
number of direct labor hours allowed per unit produced is 0.5 hours. The company budgets $1,500,000
in fixed overhead costs for the year and budgets to produce 750,000 units. Thus, the standard number
of direct labor hours allowed for the budgeted number of units is 750,000 × 0.5, or 375,000 DLH. The
fixed overhead application rate is therefore $1,500,000 ÷ 375,000, or $4.00 per DLH.

The company actually produces 800,000 units, incurring $1,490,000 in actual fixed factory overhead.
The amount of fixed overhead applied to the units produced is $4.00 × (800,000 × 0.5), which equals
$1,600,000. Fixed factory overhead is over-applied by $110,000 ($1,600,000 applied − $1,490,000
actual cost).

Note that the fixed factory overhead incurred did not increase because a greater number of
units was produced than had been planned. In fact, fixed overhead incurred was actually lower
than the budgeted amount. That can occur because fixed factory overhead does not change in total
because of changes in the activity level, as long as the activity level remains within the relevant
range.

In contrast, variable factory overhead, which is applied to production in the same way as fixed factory
overhead, does change in total because of changes in the activity level.

22 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Variable and Absorption Costing

Note: Fixed factory overheads are allocated to the units produced as if they were variable costs,
even though fixed factory overheads are not variable costs.

Absorption costing is required not only by U.S. GAAP for external financial reporting but also by the U.S.
taxing authorities for tax reporting.

When absorption costing is being used, the operating income reported by a company is influenced by the
difference between the level of production and the level of sales. For example, when the level of production
is higher than the level of sales, some of the fixed manufacturing overhead costs incurred during the current
period are included on the balance sheet as inventory at year-end. As a result, the fixed costs that are in
inventory are not included on the income statement as an expense. This topic will be discussed in more
detail later.

Fixed Factory Overheads Under Variable Costing


Under variable costing (also called direct costing), fixed factory overheads are reported as period costs
and are expensed in the period in which they are incurred. Thus, no matter what the level of sales, all of
the fixed factory overheads will be expensed in the period when incurred.

Variable costing does not conform to GAAP. For external reporting purposes, GAAP requires the use
of absorption costing, and therefore variable costing cannot be used for external financial reporting.
However, many accountants feel that variable costing is a better tool to use for internal analysis, and
therefore variable costing is often used internally.

Note: It is important to remember that the only difference in operating income between absorption
costing and variable costing relates to the treatment of fixed factory overheads. Under absorption
costing, fixed factory overhead costs are included, or absorbed, into the product cost and reach the
income statement as part of cost of goods sold when the units the fixed costs are attached to are sold.
Under variable costing, only variable direct costs (direct materials and direct labor) and variable indirect
costs (variable overhead) are included as product costs. Fixed factory overhead costs are excluded from
the product cost and treated as a period cost.

Question 3: Which of the following statements is true for a firm that uses variable costing?

a) The cost of a unit of product changes because of changes in the number of manufactured units.

b) Profits fluctuate with sales.

c) An idle facility variation is calculated.

d) Product costs include "direct" (variable) administrative costs.

(CMA Adapted)

Effects of Changing Inventory Levels


Because fixed factory overheads are treated differently under absorption and variable costing, it is virtually
certain that variable and absorption costing will result in different amounts of operating income (or operat-
ing loss) for the same period of time.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 23
Variable and Absorption Costing CMA Part 1

Note: In addition to producing different amounts of operating income, variable costing and absorption
costing will always produce different values for ending inventory because different costs are included in
each unit of inventory. Ending inventory under absorption costing will always be higher than it would be
if variable costing were used because under absorption costing, each unit of inventory will include some
fixed factory overhead costs. In contrast, under variable costing fixed overhead costs are not allocated
to production and so are not included in inventory. Instead, they are expensed in the period incurred.

Only when production and sales are equal in a period (meaning there is no change in inventory levels
and everything that was produced was sold) will there not be a difference between the operating incomes
reported under variable costing and absorption costing. If sales and production are equal, the fixed factory
overheads will have been expensed as period costs under the variable costing method, and the fixed factory
overheads will have been “sold” and included in cost of goods sold under the absorption costing method.

Whenever inventory changes over a period of time, the two methods will produce different levels of oper-
ating income.

Production Greater than Sales (Inventory Increases)


Whenever production is greater than sales, the operating income calculated under the absorption
costing method is greater than operating income under variable costing because some of the fixed
factory overheads incurred were inventoried under absorption costing. Under absorption costing, fixed fac-
tory overheads are allocated to each unit. When a unit that was produced but not sold during the period
remains on the balance sheet as inventory, its inventory cost includes the fixed factory overhead applied
to that unit. As a result, that amount of fixed factory overhead is temporarily reported on the balance sheet
instead of the income statement. When the unit is sold during the following period, that amount of fixed
factory overhead moves to the income statement as cost of goods sold.

Under variable costing all of the fixed factory overheads incurred are expensed on the income statement in
the period incurred.

Sales Greater than Production (Inventory Decreases)


If production is lower than sales, the variable costing method will result in a greater operating in-
come than absorption costing will because under the variable method, the only fixed factory overheads
included as expenses in the current period were those that were incurred during the current period. Because
sales were greater than production, some of the products that were produced in previous years were sold
during the current period. Thus, under the absorption method, some of the fixed factory overhead costs
that had been inventoried in previous years will be expensed in the current period—in addition to all the
costs incurred during the current period—as cost of goods sold.

Note: Over a long period of time, the total operating income that will be presented under both
methods will be essentially the same. In the long term, operating income will be the same under
variable and absorption costing, because in the long term the company will not produce more than it
can sell and therefore sales will equal production. Rather, the difference between the two methods will
appear in the allocation of operating income to the different periods within that longer time period.

24 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Variable and Absorption Costing

The following table summarizes the effect on operating income of changing inventory levels (production
compared to sales) under the two methods:

Production & Sales Operating Income

Production = Sales Absorption = Variable

Production > Sales Absorption > Variable

Production < Sales Absorption < Variable

Note: Ending inventory under absorption costing will always be higher than ending inventory under
variable costing, because more costs are applied to each unit under absorption costing than under vari-
able costing, and the number of units in ending inventory are the same under both methods. The units
in ending inventory under absorption costing will always contain some fixed costs, whereas the units in
ending inventory under variable costing will always contain no fixed costs.

Income Statement Presentation


The presentation of the income statement will also be different under absorption costing and variable cost-
ing.

The Income Statement under Absorption Costing


Under absorption costing gross profit is calculated by subtracting from revenue the cost of goods sold,
which includes all variable and fixed manufacturing costs for goods sold. All variable and fixed non-
manufacturing costs (period costs) are then subtracted from gross profit to calculate operating income.

The income statement (through operating income) under absorption costing is as follows:

Sales revenue
− Cost of goods sold – variable and fixed manufacturing costs of items sold
= Gross profit
− Variable nonmanufacturing costs (expensed)
− Fixed nonmanufacturing costs (expensed)
= Operating Income

The Income Statement under Variable (Direct) Costing


Under variable costing a manufacturing contribution margin is calculated by subtracting all variable
manufacturing costs for goods that were sold from revenue. From this manufacturing contribution
margin, nonmanufacturing variable costs are subtracted to arrive at the contribution margin. All fixed
costs (manufacturing and non-manufacturing) are then subtracted from the contribution margin to calculate
operating income.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 25
Variable and Absorption Costing CMA Part 1

The income statement under variable costing is as follows:

Sales revenue
− Variable manufacturing costs of items sold
= Manufacturing contribution margin
− Variable nonmanufacturing costs (expensed)
= Contribution Margin
− All fixed manufacturing costs (expensed)
− All fixed nonmanufacturing costs (expensed)
= Operating Income

Note: The difference in presentation between the two methods does not affect the difference in the
treatment of fixed manufacturing overheads under the two different methods. Candidates need to know
that under the absorption method a gross profit is reported, while under the variable method a con-
tribution margin is reported; and the two are different.

The difference is demonstrated in the example (and the answer to the example) that follows this expla-
nation.

Absorption Costing versus Variable Costing: Benefits and Limitations


While absorption costing is required for external reporting (the GAAP financial statements) and in-
come tax reporting in the U.S., it is generally thought that variable costing is better for internal uses.

Benefits of Absorption Costing

• Absorption costing provides matching of costs and benefits.

• Absorption costing is consistent not only with U.S. Generally Accepted Accounting Principles
but also with U.S. Internal Revenue Service requirements for the reporting of income on income
tax returns.

Limitations of Absorption Costing

• Because of the way that fixed costs are allocated to units under absorption costing, managers have
an opportunity to manipulate reported operating income by overproducing in order to keep some of
the fixed costs on the balance sheet in inventory. Thus, the effect of this manipulation is to move
operating income from a future period to the current period. It also creates a buildup of inventories
that is not consistent with a profitable operation.

• When the number of units sold is greater than the number of units produced, inventory decreases.
Operating income under absorption costing will be lower than it would be under variable costing,
because some prior period fixed manufacturing costs will be expensed under absorption costing along
with the current period’s fixed manufacturing costs.

26 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Variable and Absorption Costing

Benefits of Variable Costing

• The impact on operating income of changes in sales volume is more obvious with variable costing
than with absorption costing.

• By not including fixed costs in the calculation of cost to produce, companies are able to make better
and more informed decisions about profitability and product mix.

• Operating income is directly related to sales levels and is not influenced by changes in inventory
levels due to production or sales variances.

• Variance analysis of fixed overhead costs is less confusing than it is with absorption costing.

• The impact of fixed costs on operating income is obvious and visible under variable costing because
total fixed costs are shown as expenses on the income statement.

• It is easier to determine the “contribution” to fixed costs made by a division or product—and thereby
helps determine whether the product or division should be discontinued.

• Variable costing tends to be less confusing than absorption costing because it presents costs in the
same way as they are incurred: variable costs are presented on a per-unit basis and fixed costs are
presented in total.

• Advocates argue that variable costing is more consistent with economic reality, because fixed costs
do not vary with production in the short run.

Limitations of Variable Costing

• Variable costing does not provide proper matching of costs and benefits and so is not acceptable for
external financial reporting under generally accepted accounting principles. Variable costing is also
not acceptable for income tax reporting.

• Because only variable manufacturing costs are charged to inventory, variable costing requires sepa-
rating all manufacturing costs into their fixed and variable components.

• To prepare an income statement based on variable costing, it is also necessary to separate the selling
and administrative costs into their fixed and variable components.

Note: The issue of absorption costing versus variable costing is relevant only for a manufacturing com-
pany. A company that does not do any manufacturing, for example a reseller or a service company,
would have only non-manufacturing fixed overhead. Non-manufacturing fixed overhead for such a com-
pany would simply be expensed as incurred.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 27
Variable and Absorption Costing CMA Part 1

Variable/Absorption Costing Example


Hardy Corp. uses the FIFO method to track inventory. The records for Hardy include the following infor-
mation. For simplicity, assume that the amounts given are both the budgeted amounts and the actual
amounts and thus, there were no variances.
Inventory (in units) Year 1 Year 2
Beginning balance, in units -0- 4,200
Production 12,000 10,000
Available for sale 12,000 14,200
Less units sold ( 7,800) (12,000)
Ending balance, in units 4,200 2,200
Other information
Sales ($2.10 per unit) $16,380 $25,200
Variable mfg. costs ($0.90/unit) 10,800 9,000
Fixed mfg. costs 5,000 5,400
Variable selling and admin. costs 2,250 3,750
Fixed selling and admin. costs 2,250 3,750
Required: Prepare the income statements for Year 1 and Year 2 using both the absorption and the
variable methods of costing. (The solution follows.)

Year 1 Absorption Costing: Year 1 Variable Costing:


Sales revenue $ Sales revenue $

+_________
Manuf. Cont. Margin $

________ ________

Gross Profit $ Contribution Margin $

________ ________

Operating Income $ Operating Income $

Year 2 Absorption Costing: Year 2 Variable Costing:


Sales revenue $ Sales revenue $

________

Manuf. Cont. Margin $

________ ________

Gross Profit $ Contribution Margin $

________ ________

Operating Income $ Operating Income $

28 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Variable and Absorption Costing

Answer to the Variable/Absorption Costing Example

Year 1 Income Statements

Absorption Costing Variable (Direct) Costing

Sales (7,800 × $2.10) $ 16,380 Sales (7,800 × $2.10) $ 16,380

Variable Mfg. COGS (7,800 × $0.90) 7,020 Variable Mfg. COGS (7,800 × $0.90) (7,020)

Fixed Mfg. COGS ($5,000÷12,000×7,800) 3,250 Manufacturing Contribution Margin $ 9,360

COGS (10,270) Less: Variable S&A (2,250)

Contribution margin $ 7,110

Gross profit $ 6,110

S&A exp. ($2,250 var. + $2,250 fixed) (4,500) Less: Fixed mfg. costs (5,000)

Fixed S&A (2,250)

Operating income $ 1,610 Operating income $ (140)

Year 2 Income Statements

Absorption Costing Variable (Direct) Costing

Sales (12,000 × $2.10) $ 25,200 Sales (12,000 × $2.10) $ 25,200

Variable Mfg. Cost (12,000 × $0.90) 10,800 Variable Mfg.COGS (12,000×$0.90) (10,800)
6
Fixed Mfg. COGS-Year 1 production 1,750 Manufacturing Contribution margin 14,400
7
Fixed Mfg. COGS-Year 2 production 4,212 Variable S&A (3,750)

COGS (16,762) Contribution margin 10,650

Gross profit 8,438 Less: Fixed mfg. costs (5,400)

S&A expenses (7,500) Fixed S&A (3,750)

Operating income $ 938 Operating income $ 1,500

6
Beginning inventory for Year 2 was 4,200 units, and those units were produced during Year 1 at Year 1 costs. Each
unit had $0.4167 of fixed manufacturing cost attached to it ($5,000 fixed manufacturing costs in Year 1 ÷ 12,000 units
produced in Year 1). Since the company uses FIFO, those units were the first units sold in Year 2. 4,200 units × $0.4167
of fixed manufacturing cost per unit = $1,750 of fixed manufacturing cost from Year 1 production that was sold in Year
2.
7
A total of 12,000 units were sold in Year 2. As noted above, 4,200 of those units came from Year 1’s production. That
means the remainder, or 12,000 – 4,200 = 7,800 units, came from Year 2’s production. Fixed manufacturing cost per
unit during Year 2 was $5,400 fixed manufacturing costs ÷ 10,000 units produced, or $0.54 per unit. So the fixed
manufacturing cost included in the units that were sold from Year 2’s production is $0.54 × 7,800 = $4,212.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 29
Variable and Absorption Costing CMA Part 1

Note: In certain situations, it is very easy to calculate the difference between the variable and absorption
method operating income. Given that the only difference between the two costing methods is the treat-
ment of fixed factory overheads, if one of the three situations below applies and the question asks for
the difference in operating income between the two methods, the only calculation needed is:

Fixed overhead cost per unit applied to production8


× Number of units of change in inventory
= Difference in operating income between the two methods

The three situations in which the formula above can be used to calculate the difference in operating
income between the two methods are:

(1) Beginning inventory is zero. In many questions there is a statement either that there is no beginning
inventory, or that it is the company’s first year of operations (in which case there is no beginning
inventory).

(2) The LIFO inventory cost flow assumption is being used and ending inventory is higher than beginning
inventory (in other words, none of the beginning inventory was sold during the period).

(3) If an inventory cost flow assumption other than LIFO is being used, and (a) the beginning inventories
are valued at the same per-unit fixed manufacturing cost as the current year’s planned per-unit fixed
manufacturing cost and (b) under- or over-applied fixed manufacturing overhead is closed out to
cost of goods sold only.

The following information is for the next four questions: The estimated unit costs for a company
using absorption (full) costing and planning to produce and sell at a level of 12,000 units per month are
as follows.

Cost Item Estimated Unit Cost


Direct materials $32
Direct labor 20
Variable manufacturing overhead 15
Fixed manufacturing overhead 6
Variable selling 3
Fixed selling 4

Question 4: Estimated conversion costs per unit are:

a) $35

b) $41

c) $48

d) $67

8
If the inventory level has fallen, the previous year’s fixed overhead cost per unit will need to be used as the “extra
inventory” sold that was produced during the previous year. If inventory has risen, the current period’s fixed overhead
cost per unit should be used.

30 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Variable and Absorption Costing

Question 5: Estimated prime costs per unit are:

a) $73

b) $32

c) $67

d) $52

Question 6: Estimated total variable costs per unit are:

a) $38

b) $70

c) $52

d) $18

Question 7: Estimated total costs that would be incurred during a month with a production level of
12,000 units and a sales level of 8,000 units are:

a) $692,000

b) $960,000

c) $948,000

d) $932,000

(CMA Adapted)

Question 8: When a firm prepares financial reports using absorption costing:

a) Profits will always increase with increases in sales.

b) Profits will always decrease with decreases in sales.

c) Profits may decrease with increased sales even if there is no change in selling prices and costs.

d) Decreased output and constant sales result in increased profits.

(CMA Adapted)

Question 9: Jansen, Inc. pays bonuses to its managers based on operating income. The company uses
absorption costing, and overhead is applied on the basis of direct labor hours. To increase bonuses,
Jansen's managers may do all of the following except:

a) Produce those products requiring the most direct labor.

b) Defer expenses such as maintenance to a future period.

c) Increase production schedules independent of customer demands.

d) Decrease production of those items requiring the most direct labor.

(CMA Adapted)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 31
Variable and Absorption Costing CMA Part 1

Question 10: Nance Corp began operations in January. The company produced 50,000 units and sold
45,000 units in its first year of operations. Costs for the year were as follows:

Fixed Manufacturing Costs $250,000


Variable Manufacturing Costs 180,000
Fixed General and Selling Costs 75,000
Variable General and Selling Costs 80,000

How would the operating income of Nance compare between the variable method and full absorption
costing methods?

a) Variable would be $25,000 higher.

b) Absorption would be $25,000 higher.

c) Variable would be $32,500 higher.

d) Absorption would be $32,500 higher.

(HOCK)

The following information is for the next two questions: Osawa planned to produce and actually
manufactured 200,000 units of its single product in its first year of operations. Variable manufacturing
costs were $30 per unit of product. Planned and actual fixed manufacturing costs were $600,000, and
the selling and administrative costs totaled $400,000. Osawa sold 120,000 units of product at a selling
price of $40 per unit.

Question 11: Osawa’s operating income using absorption costing is:

a) $200,000

b) $440,000

c) $600,000

d) $840,000

Question 12: Osawa’s operating income using variable costing is:

a) $200,000

b) $440,000

c) $800,000

d) $600,000

(CMA Adapted)

32 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Joint Products and Byproducts

Joint Products and Byproducts


When one production process leads to the production of two or more finished products, joint products
result. The products are not identical, but they share the same production process up to what is called the
splitoff point. The splitoff point is the point at which the two products stop sharing the same process and
become different, identifiable products.

The main issue with joint products is how to account for the joint costs (those costs incurred prior to the
splitoff point) and how to allocate the joint costs to the separate products. Accurate allocation is needed
primarily for financial reporting purposes and pricing decisions. The inventory cost of each unit of each joint
product needs to be determined accurately so that the balance sheet will be accurate. Since the inventory
cost of each unit becomes its cost of goods sold when it is sold, the amount of cost to be expensed to COGS
for each unit sold is needed.

An example of joint products would be the processing of pineapple. As a pineapple goes through processing
at the factory the rind is removed. The fruit is used to manufacture two products: bottled pineapple juice
and canned pineapple slices, two separate products that arise from the same initial production process
(removal of the rind). Therefore, the joint costs of the unprocessed pineapples (the direct materials) and
of processing the pineapples to remove the rinds need to be allocated between the juice and the slices.

Joint costs may include direct materials, direct labor, and overhead. Costs incurred after the splitoff point
may also include direct materials, direct labor, and overhead. The costs incurred after the splitoff point are
separable costs and they are allocated to each product as they are incurred by that product.

Byproducts are the low-value products that occur naturally in the process of producing higher value prod-
ucts. They are, in a sense, accidental results of the production process. The main differentiator between
main products and joint products is the market value. If the product has a comparatively low market value
when compared to the other products produced, it is a byproduct.

Methods of Allocating Costs to Joint Products


Several different measures may be used to allocate joint costs, but all of the different methods use some
sort of ratio between the two or more products. The cost allocation is largely a mathematical exercise, but
candidates need to learn the different allocation bases are calculated.

The primary methods are:

1) Relative Sales Value at Splitoff method

2) Estimated Net Realizable Value (NRV) method

3) Physical Measure and Average Cost methods

4) Constant Gross Profit (Gross Margin) Percentage method

1. Relative Sales Value at Splitoff Method (or Gross Market Value Method)
The Relative Sales Value at Splitoff method is also called the Gross Market Value method, the Sales
Value at Splitoff method, or, more simply, just the Sales Value method.

Under the Relative Sales Value at Splitoff method, joint costs are allocated on the basis of the sales values
of each product at the splitoff point, relative to the total sales value of all the joint products.

The formula to allocate the costs between or among the products is as follows, for each of the joint
products:

Sales Value of Product X Joint Cost Allocated to


× Joint Cost =
Total Sales Value of all Joint Products Product X

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 33
Joint Products and Byproducts CMA Part 1

The Relative Sales Value at Splitoff method can be used only if all of the joint products can be sold at
the splitoff point (in other words, with no further processing). Management may decide it would be more
profitable to the company to process some of the joint products further; but the Relative Sales Value at
Splitoff method can still be used to allocate joint costs up to the splitoff point, as long as sales prices at the
splitoff point do exist for all of the joint products.

Example of the Relative Sales Value at Splitoff method:

Cafe Industries manufactures two kinds of coffee percolators: an electric model and a stovetop model.
Part of the manufacturing process is the production of the coffee basket assembly, which includes a
basket and a spreader. The 8-cup electric model and the 8-cup stovetop model use the same basket
assembly, though the pump stems are different. The basket assembly is also sold separately as a re-
placement part for both percolators at a price of $10.00. Separate prices for the basket assemblies that
go on to be incorporated into the two percolators do not exist, since they become integral parts of the
percolators. However, the whole production run could be sold as replacement baskets, since a market
does exist for them at that stage of production.

One batch consists of 500 basket assemblies, of which 300 are destined to become part of electric
percolators, 150 are destined to become part of stovetop percolators, and 50 are sold separately as
replacement parts. The joint costs of one batch total $2,500, or $5.00 per unit.

Using the Sales Value at Splitoff method, how much of the joint cost is allocated to electric percolators,
how much to stovetop percolators, and how much to the replacement parts?

The sales value of the electric percolator basket assemblies is 300 × $10.00, or $3,000. The sales value
of the stovetop percolator basket assemblies is 150 × $10.00, or $1,500. The sales value of the replace-
ment baskets is 50 × $10.00, or $500. The total sales value of all 500 basket assemblies is therefore
$3,000 + $1,500 + $500, or $5,000.

The portion of the joint cost allocated to the electric percolators is ($3,000 ÷ $5,000) × $2,500, or
$1,500.
The portion of the joint cost allocated to the stovetop percolators is ($1,500 ÷ $5,000) × $2,500, or
$750.
The portion of the joint cost allocated to the replacement basket assemblies is ($500 ÷ $5,000) ×
$2,500, or $250.

To confirm that the full $2,500 of joint cost has been allocated, sum the allocated amounts: $1,500 +
$750 + $250 = $2,500.

Benefits of the Relative Sales Value at Splitoff (Gross Market Value) Method

• Costs are allocated to products in proportion to their expected revenues, in proportion to the individual
products’ ability to absorb costs.

• The method is easy to calculate and is simple, straightforward, and intuitive.

• The cost allocation base is expressed in terms of a common basis—amount of revenue—that is rec-
orded in the accounting system.

• It is the best measure of the benefits received from the joint processing. It is meaningful because
generating revenues is the reason for the company to incur the joint costs.

• It can be used when further processing is done, as long as selling prices exist for all the joint products.

34 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Joint Products and Byproducts

Limitations of the Relative Sales Value at Splitoff (Gross Market Value) Method

• Selling prices at the splitoff point must exist for all of the products in order to use this method.

• Market prices of joint products may vary frequently, but this method uses a single set of selling prices
throughout an accounting period, which can introduce inaccuracies into the allocations.

2. Estimated Net Realizable Value (NRV) Method


The Estimated Net Realizable Value (NRV) method can be used if one or more of the joint products must
be processed beyond the splitoff point in order to be sold. It may also be used under certain circumstances
if one or more of the joint products may be processed beyond the splitoff point in order to increase its
value above the selling price at the splitoff point.

This method is essentially the same as the Relative Sales Value method, and the allocation is done in the
same way, except an estimated Net Realizable Value (NRV) is used for the product or products that
must be or will be processed further.

The estimated NRV for a product to be processed further is calculated as:

Future sales price of items produced that will be sold in the future
− Separable costs incurred after the splitoff point
= Estimated net realizable value

Note: If one (or more) of the joint products is not processed further but is sold at the splitoff
point, instead of using NRV for those products, the company will simply use the Sales Values at the
splitoff point for the products that can be sold at the splitoff point, while using the NRVs for the products
that must be processed further to be marketable.

The Estimated NRV method would generally be used instead of the Relative Sales Value at Splitoff Point
method only when a market price at the splitoff point is not available for one or more of the joint
products, for example because a product is not marketable at the splitoff point. If a market price at the
splitoff point is available for one or more of the joint products because the products can be sold at that
point, that price is used instead of the products’ NRVs, even if the products will be processed
further.

If a sales value at splitoff is not available for one or more of the joint products, it is acceptable to use within
the same allocation the net realizable values of the products that must be processed further in order to be
sellable while using the sales values at splitoff for the products that can be sold at the splitoff point. For
instance, a company may have one product that cannot be sold at the splitoff point and must be processed
further (thus no sales value at splitoff is available for it), while the other product can be sold at the splitoff
point. In such a case, the NRV of the product that must be processed further is its estimated NRV (sales
price after further processing less cost to process further), while the NRV of the product that can be sold at
the splitoff point is its sales value at the splitoff point.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 35
Joint Products and Byproducts CMA Part 1

Note: The Net Realizable Value method is generally used in preference to the Relative Sales Value at
Splitoff method only when selling prices for one or more products at splitoff do not exist. However,
sometimes when sales prices at the splitoff do exist for all of the joint products but one or more products
can be processed further, an exam problem will say to use the Net Realizable Value method to allocate
the joint costs. If a problem says to use the Net Realizable Value method, use the net realizable
values for the products that can be processed further even though sales prices at splitoff do
exist, but only if the cost to process further is less than the additional revenue to be gained
from the further processing. (if the cost to process a product further is greater than the additional
revenue to be gained from the further processing, the product will not be processed further.)

If the problem does not say to use the Net Realizable Value method and sales values at the splitoff exist
for all products, then use the sales values of all of the joint products for the allocation, even if one or
more of the products can be or will be processed further.

Note: The joint costs of production are not relevant costs in the decision to process further or sell
immediately because they are sunk costs. In order to determine whether or not a product should be
processed further, the company should compare the incremental revenues (the increase in the sales
price that results from further processing) with the incremental cost (the increase in costs related to
the additional processing). If the incremental revenue is greater than the incremental cost, the product
should be processed further.

Costs incurred by each of the separate products after the splitoff point are simply allocated directly to
those products.

Example of the Net Realizable Value method:

Simpli Chili Company produces three flavors of its chili in a joint process. The flavors are mild, original
and extra spicy. 500,000 gallons of unspiced chili are produced per batch, and then varying amounts
and types of spices are added to produce the mild, original and extra spicy flavors. The three types of
chili are packaged in 16-ounce cans. The total joint cost of the unspiced chili is $1,850,000.

One batch results in 500,000 gallons of unspiced chili. The unspiced chili is, of course, not marketable
at that point. It needs spices.

After the spices have been added, Simpli has 800,000 cans of mild chili, 2,000,000 cans of original chili,
and 1,200,000 cans of extra spicy chili. The cost per can of adding the spices and blending them into
the unspiced chili are as follows:
Mild 0.065
Original 0.075
Extra spicy 0.080

The mild chili sells for $0.98 per can. The original chili sells for $1.05 per can. The extra spicy chili sells
for $1.09 per can.

Using the Net Realizable Value method of allocating the joint costs, how much of the joint costs will be
allocated to each type of chili?
Price/ Extended Cost to Pro- Percentage
Product # Cans Can Sales Value cess Further NRV of Total NRV
Mild 800,000 0.98 $ 784,000 $ 52,000 $ 732,000 18.8%
Original 2,000,000 1.05 2,100,000 150,000 1,950,000 50.1%
Extra spicy 1,200,000 1.09 1,308,000 96,000 1,212,000 31.1%

Total 4,000,000 $3,894,000

(Continued)

36 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Joint Products and Byproducts

The joint cost of $1,850,000 will be allocated as follows:

Mild $1,850,000 × 0.188 = $ 347,800


Original $1,850,000 × 0.501 = 926,850
Extra Spicy $1,850,000 × 0.311 = 575,350

Total allocated $1,850,000

Benefits of the Net Realizable Value Method

• The Net Realizable Value method can be used instead of the Sales Value at Splitoff method when
selling prices for one or more products at the splitoff do not exist, because it provides a better
measure of the benefits received than the other methods that could be used in this situation.

• The allocation results in comparable profitability among the joint products.

Limitations of the Net Realizable Value Method

• The NRV method is complex. It requires information on the specific sequence of further processing
and the separable costs of further processing, as well as the point at which individual products will
be sold.

• The NRV method is often implemented with simplified assumptions. Companies assume a specific set
of processing steps beyond the splitoff point, but they may actually do something else and in fact
may change the steps frequently.

• Selling prices of joint products may vary frequently, but the NRV method uses a single set of selling
prices throughout an accounting period, which can introduce inaccuracies into the allocations.

3. Physical Measure and Average Cost Methods


The Physical Measure and Average Cost methods are essentially the same. In the Physical Measure method,
the joint cost allocation is done based on the weight, volume, or other physical measure of the joint prod-
ucts, such as pounds, tons, or gallons. In the Average Cost method, the joint cost allocation is done based
on the physical units of output. In both methods, joint costs are allocated proportionately among the joint
products, so that each product is allocated the same amount of joint cost per unit of measure, whether that
unit is a unit of physical measure or a unit of output.

Physical Measure Method


Joint cost allocation may be done based on the weight, volume, or other physical measure of the joint
products. Joint costs are allocated based on some common unit of measurement of output at the splitoff
point, such as pounds, tons, gallons, or board feet (for lumber). The Physical Measure method may also be
called the Quantitative Unit method.

The total joint cost up to the splitoff point is prorated between or among the joint products based on the
physical measure being used. It stands to reason that it must be possible to measure all of the joint products
in the same unit of measurement. If all of the output that results from a joint process cannot be meas-
ured in the same terms—for instance, if some of the output is liquid and some of the output is dry—the
Physical Measure method cannot be used.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 37
Joint Products and Byproducts CMA Part 1

Example of the Physical Measure method:

The Simpli Chili Company produces three flavors of its chili in a joint process. The three flavors are mild,
original, and extra spicy. 500,000 gallons of unspiced chili are produced per batch, and then varying
amounts and types of spices are added to produce the mild, original and extra spicy flavors.

In this example, each of the three types of chili is packaged in a choice of can sizes: 12-ounce, 16-ounce
and 20-ounce cans. 100,000 gallons of the unspiced chili are used to produce the mild chili, 250,000
gallons are used for the original chili, and 150,000 gallons are used for the extra spicy chili. The total
joint cost of the unspiced chili (including direct materials, direct labor, and overhead) is $1,850,000. The
joint cost is allocated as follows:

Cost Per
Product Physical Measure Proportion Allocation of Joint Cost Gallon

Mild 100,000 gal. 0.20 $1,850,000 × 0.20 = $ 370,000 $3.70


Original 250,000 gal. 0.50 $1,850,000 × 0.50 = 925,000 $3.70
Extra spicy 150,000 gal. 0.30 $1,850,000 × 0.30 = 555,000 $3.70

Total 500,000 gal. $1,850,000

Average Cost Method


The Average Cost method may also be called the Physical Unit method. It is used when the joint costs
are to be allocated on the basis of physical units of output in completed form. It is basically the same as
the Physical Measure method, but because physical units of completed product are used, it is called by the
name Average Cost, or sometimes Average Unit Cost method.

The total joint cost is divided by the total number of units of all of the joint products produced to calculate
the average cost per unit. Then that average cost per unit is multiplied by the number of units of each
product produced to find the amount of cost to be allocated to each product.

Example of the Average Cost/Physical Unit Method:

In the Simpli Chili example, now all three flavors of chili are packaged in 16-ounce cans only. Since the
size of the cans is the same for all the different versions, units of output can be used for the allocation,
with the can as the unit of output. The 500,000 gallons of output are in 4,000,000 16-ounce cans. The
average cost per can is the total cost of $1,850,000 divided by 4,000,000 cans, or $0.4625 per can.

Now, the output and the allocations from the joint process are as follows:

Product Units of Output × Avg. Cost/Unit Allocation of Joint Cost

Mild 800,000 cans $0.4625 $ 370,000


Original 2,000,000 cans $0.4625 925,000
Extra spicy 1,200,000 cans $0.4625 555,000

Total 4,000,000 cans $1,850,000

Note that the allocation of the cost by product using the Average Cost/Physical Unit method is exactly
the same as it was when the Physical Measure method was used.

38 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Joint Products and Byproducts

Benefits of the Physical Measure and Average Cost Methods

• The Physical Measure and Average Cost methods are easy to use.

• The allocation is objective.

• The methods are useful when rates or prices are regulated.

Limitations of the Physical Measure and Average Cost Methods

• The Physical Measure and Average Cost methods can result in product costs that are greater than
the market values for some of the joint products. The physical measures of the individual products
may have no relationship to their respective abilities to generate revenue. If weight or size is used,
the heaviest or largest product will be allocated the greatest amount of the joint cost; but that
product may have the lowest sales value. Products with a high sales value per weight or size would
show large profits, while products with a low sales value per weight or size would show large losses.

• Physical measures are not always comparable for products. For example, some products might be in
liquid form (for example, petroleum), whereas some might be in gaseous form (for example, natural
gas). When the physical measures used for the joint products differ, this method cannot be used.

4. Constant Gross Profit (Gross Margin) Percentage Method


The Constant Gross Profit Percentage method allocates the joint costs in such a way that all of the joint
products will have the same gross profit margin percentage. It is done by “backing into” the amount of
joint cost to be allocated to each of the joint products.

Step 1: Calculate the gross profit margin percentage for the total of both (or all, if more than two) of the
joint products to be included in the allocation by subtracting the total joint and total separable costs from
the total final sales value and dividing the remainder by the total final sales value. This calculation should
be done for all of the joint products produced during the period, not for all of the joint products sold during
the period. The result is the total gross profit margin percentage.

Step 2: Calculate the gross profit for each of the individual products by multiplying the total gross profit
margin percentage calculated in Step 1 by each individual product’s final sales value.

Step 3: Subtract the gross profit calculated in Step 2 and any separable costs from each individual product’s
final sales value. The result of this subtraction process will be the amount of joint costs to allocate to each
product.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 39
Joint Products and Byproducts CMA Part 1
efham CMA
Example of the Constant Gross Profit (Gross Margin) Percentage method:
Pineapple Co. produces pineapple juice and canned slices at its Pineapple Processing Plant in Hawaii. The
information about the process and the two joint products is as follows:
• 10,000 pineapples are processed.
• The process results in 2,500 kg of juice and 7,500 kg of slices.
• The juice can be sold for $10 per kg and the slices can be sold for $15 per kg.
• The joint costs of production are $120,000.
• The juice can be processed further into a premium juice. The additional processing costs an additional
$8,000 ($3.20 per kg), but the sales price per kg will increase by $5 per kg to $15.
• The slices can be further processed into chunks. The additional processing costs $4,000 ($0.533 per
kg) and the chunks can be sold for $2 more per kg than the slices, or $17 per kg.
• Since the additional revenue to be earned from processing each product further exceeds the addi-
tional cost to process further, both products are processed further.
Step 1: Calculate the overall gross profit margin percentage for both products:

Gross Profit
Premium Juice Chunks Total Margin %
Margin
1 2
Final sales value $37,500 $127,500 $ 165,000
Separable costs 12,000
Joint costs 120,000
Overall gross profit $ 33,000 20%3

Step 2: Fill in the separable costs for each product and calculate what a 20% gross profit margin
percentage for each product would be:

Gross Profit
Premium Juice Chunks Total Margin %
Final sales value $37,500 $127,500 $165,000
Separable costs (given) 8,000 4,000 12,000
Joint costs J C 120,000
Required gross profit $ 7,500 4 $ 25,500 5
$ 33,000 20%

Step 3: The amount of joint cost allocated to each product is whatever amount would create the re-
quired gross profit for the product when the separable costs and the allocated joint cost are subtracted
from the final sales value.

Juice: $37,500 − $8,000 − J = $7,500 J = $22,000

Chunks: $127,500 − $4,000 − C = $25,500 C = $98,000

The $120,000 of joint costs is allocated $22,000 to juice and $98,000 to chunks.

Calculations:
1
2,500 × $15
2
7,500 × ($15 + $2)
3
$33,000 ÷ $165,000
4
$37,500 × 0.20
5
$127,500 × 0.20

40 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Joint Products and Byproducts

Benefits of the Constant Gross Profit (Gross Margin) Percentage Method

• The Constant Gross Profit method is the only method for allocating joint costs under which products
may receive negative allocations.

• This method allocates both joint costs and gross profits. Gross profit margin is allocated to the joint
products in order to determine the joint cost allocations so that the resulting gross profit margin per-
centage for each product is the same.

• The method is relatively easy to implement, so it avoids the complexities of the NRV method.

Limitations of the Constant Gross Profit (Gross Margin) Percentage Method

• It assumes that all products have the same ratio of cost to sales value, which is probably not the case.

Accounting for Byproducts


Byproducts are the low-value products that occur naturally in the process of producing higher value prod-
ucts. They are, in a sense, accidental results of the production process. Because they are so low in value,
the accounting method used to account for byproducts does not need to be very detailed or sophisticated.
The main issue in accounting for byproducts relates to the treatment of the associated costs and revenues.

Even though the byproducts are small and immaterial to the larger production cost, there are still costs
that will be allocated to the byproduct. At a minimum, there are some direct materials in the byproduct, so
some percentage of the direct materials used did not make it into the final main products but became
byproducts. The byproducts might be material remnants from cutting fabric, sawdust from cutting lumber,
or any other small amounts of materials.

The company may or may not have a policy of assigning costs to the byproducts, and that policy is reflected
in its accounting treatment of byproducts. The costs and revenues of byproducts can be accounted for in
either of two ways.

Note: The specific accounting for these two methods is outside the scope of the exam. Candidates need
to know only how to treat the associated costs and revenues of byproducts.

The two ways to treat the associated costs and revenues of the byproduct are:

1) Use the estimated net realizable value of the byproduct9 as a reduction of the costs of production
of the main product. This method is called the production method and in it the NRV of the
byproduct is inventoried as the cost of the byproduct.

2) All the costs of production are allocated to the main product or joint products. When the byproduct
is sold, the revenue is recorded as either revenue without any associated COGS or as a reduction
to the cost of production of the main products. This method is called the sales method and only
revenue is recognized from the byproduct.

Which Method is Better?


Both methods are acceptable. The Production Method, where the byproduct is inventoried at the time of
production, is conceptually correct because it is consistent with the matching principle. Byproduct inventory
is recognized as an asset in the accounting period in which it is produced, and it reduces the inventory cost
assigned to the main product or joint products.

9
The estimated net realizable value of a byproduct is the revenue from its sale less marketing and administrative costs
of selling the byproduct and less any subsequent costs of processing the byproduct after the splitoff point.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 41
Joint Products and Byproducts CMA Part 1

The Sales Method, in which the byproduct revenue is recognized at the time of sale, is simpler and is used
more frequently in practice if the monetary amounts of the byproduct or byproducts are immaterial.

The Sales Method has a disadvantage, though. The Sales Method makes it possible for managers to time
the sale of the byproducts and thus enables them to manage their earnings. A manager could store the
byproducts for a period of time and sell them to increase operating income during a period when sales or
gross profits from the main product or joint products are low.

Benefits of the Production Method of Accounting for Byproduct Costs

• The Production Method is conceptually correct because byproduct inventory is recognized as an


asset in the accounting period in which it is produced.

• Inventory cost assigned to the main product or joint products is reduced.

Limitations of the Production Method of Accounting for Byproduct Costs

• The Production Method is more complex than the Sales Method.

Benefits of the Sales Method of Accounting for Byproduct Costs

• The Sales Method is simpler to implement than the Production Method.

• It is more practical when the monetary amounts of the byproducts are immaterial.

Limitations of the Sales Method of Accounting for Byproduct Costs

• It is possible for managers to manage their earnings by timing the sale of the byproducts.

Note: An exam question will outline the treatment of the costs or revenue associated with the byproduct
and candidates just need to follow the mathematical steps required. If a question states that the
company inventories the byproduct, it means that it treats the byproduct’s net realizable
value as a reduction of the costs of production and uses the Production Method.

HOCK international books are licensed only for individual use and may not be lent,
copied, sold, or otherwise distributed without permission directly from HOCK
international.

If you did not download this book directly from HOCK international, it is not a
genuine HOCK book. Using genuine HOCK books assures that you have complete, accurate,
and up-to-date materials. Books from unauthorized sources are likely outdated and will not
include access to our online study materials or access to HOCK teachers.

Hard copy books purchased from HOCK international or from an authorized training
center should have an individually numbered orange hologram with the HOCK globe
logo on a color cover. If your book does not have a color cover or does not have this
hologram, it is not a genuine HOCK book.

42 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Joint Products and Byproducts

Question 13: Lankin Corp. produces two main products and a byproduct out of a joint process. The ratio
of output quantities to input quantities of direct materials used in the joint process remains consistent
from month-to-month. Lankin has employed the physical measure method to allocate joint production
costs to the two main products. The net realizable value of the byproduct is used to reduce the joint
production costs before the joint costs are allocated to the two main products. Data regarding Lankin’s
operations for the current month are presented in the chart below. During the month, Lankin incurred
joint production costs of $2,520,000. The main products are not marketable at the splitoff point and,
thus, have to be processed further.

1st Main Product 2nd Main Product Byproduct


Monthly input in pounds 90,000 150,000 60,000
Selling price per pound $30 $14 $2
Separable process costs $540,000 $660,000

The amount of joint production cost that Lankin would allocate to the Second Main Product by using the
physical measure method to allocate joint production costs would be:

a) $1,200,000

b) $1,260,000

c) $1,500,000

d) $1,575,000.

(CMA Adapted)

Question 14: Sonimad Sawmill manufactures two lumber products from a joint milling process. The two
products developed are mine support braces (MSB) and unseasoned commercial building lumber (CBL).
A standard production run incurs joint costs of $300,000 and results in 60,000 units of MSB and 90,000
units of CBL. Each MSB sells for $2 per unit, and each CBL sells for $4 per unit.

If there are no further processing costs incurred after the splitoff point, the amount of joint cost allo-
cated to the mine support braces (MSB) on a relative sales value basis would be:

a) $75,000

b) $180,000

c) $225,000

d) $120,000

(CMA Adapted)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 43
D.2. Costing Systems CMA Part 1

D.2. Costing Systems


Review of Introduction to Costing Systems
Product costing involves accumulating, classifying and assigning direct materials, direct labor, and factory
overhead costs to products, jobs, or services.

In developing a costing system, management accountants make choices in three categories of costing
methods:

1) The cost measurement method to use in allocating costs to units manufactured (standard, nor-
mal, or actual costing), as covered in D.1. Measurement Concepts.

2) The cost accumulation method to use (job costing, process costing, or operation costing).

3) The method to be used to allocate overhead, as will be covered in D.3. Overhead Costs.

Cost accumulation methods and the allocation of overhead will be covered in this topic.

Cost Accumulation Systems


Cost accumulation systems are used to assign costs to products or services. Job order costing (also called
job costing), process costing, and operation costing are different types of cost accumulation systems used
in manufacturing.

• Process costing is used when many identical or similar units of a product or service are being
manufactured, such as on an assembly line. Costs are accumulated by department or by process.

• Job order costing (also called job costing) is used when units of a product or service are distinct
and separately identifiable. Costs are accumulated by job.

• Operation costing is a hybrid system in which job costing is used for direct materials costs while
a departmental (process costing) approach is used to allocate conversion costs (direct labor and
overhead) to products or services.

Process costing, job order costing, and operation costing will be discussed in more detail later.

Process Costing
Process costing is used to allocate costs to individual products when the products are all relatively similar
and are mass-produced (the term “homogeneous” is used to describe the products, and it means “identical
to each other”). Process costing is basically applicable to assembly lines and products that share a similar
production process. In process costing, all of the costs incurred by a process (a process is often referred to
as a department) are collected and then allocated to the individual goods that were produced, or worked
on, during that period within that process (or department).

The basic exercise is to allocate all of the incurred costs to either the completed units that left the
department or to the units in ending work-in-process (EWIP) that are still in the department. It is
largely a mathematical operation and some basic formulas are used to make certain that everything is
accounted for.

Some general concepts and ideas of process costing will be presented first, and then the steps of process
costing will be discussed one-by-one. The topic will conclude with a review of the steps and a comprehensive
example.

All of the costs incurred during the current period and during all previous periods for the units worked on
during the period must be allocated to either finished goods (or to the next department for more work to
be done) or EWIP at the end of the current period.

44 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

The costs in the department, usually materials and conversion costs10 and sometimes transferred-in costs,
that require allocation can come from one of three places:

1) The costs are incurred by the department during the period. Materials and conversion costs are
accounted for separately.

2) The costs are transferred in from the previous department. Transferred-in costs include total
materials and conversion costs from previous departments that have worked on the units. Trans-
ferred-in costs are transferred in as total costs. They are not segregated according to direct
materials and conversion costs.

3) The costs were in the department on the first day of the period as costs for the beginning work-
in-process (BWIP). They were incurred by the department during the previous period to begin
the work on the units in the current period’s BWIP.

In reality, the categories of costs can be numerous. They may include more than one type of direct mate-
rials, more than one class of direct labor, indirect materials, indirect labor or other overheads. However, on
the CMA exam, generally only two classifications of costs are tested: direct materials and conversion
costs. Conversion costs include everything other than direct materials—specifically direct labor and over-
head—and are the costs necessary for converting the raw materials into the finished product.

Note: Conversion costs is a term used in process costing questions to refer to direct labor and
factory overhead. It encompasses everything except raw materials. Conversion costs are the costs
necessary to convert the raw materials into the finished product. Placing direct labor and factory over-
head into a single category reduces the number of individual allocations needed.

Note: Transferred-in costs are the total costs that come with the in-process product from the previous
department. They are similar to raw materials but they include all of the costs (direct materials and
conversion costs) from the previous department that worked on the units. The costs of the previous
department’s “completed units” are the current department’s transferred-in costs, and the transferred-
in costs and work are 100% complete (even though the units themselves are not complete when re-
ceived) because the work done in the previous department is 100% complete.

At the end of the period all of the costs within the department—including direct materials, conversion costs,
and transferred-in costs, if applicable—must either be moved to finished goods inventory (or to the next
department if further work is required) if the work on them in the current department was completed, or
they will remain in ending WIP if they are not complete (the allocation process will be explained later). The
Ending WIP inventory for the current period will be the beginning WIP inventory for the next period.

When the goods that have been completed and transferred to finished goods inventory are sold, the costs
associated with the units that were sold will end up in COGS. The costs of the units that have been com-
pleted but have not been sold will remain in finished goods inventory until they are sold.

Thus, the cost of every unit that goes through a particular process in a given period must be recorded in
one of the four following places at the end of the period:

1) Ending WIP inventory in the department or process

2) The next department in the assembly process

3) Ending finished goods inventory

4) Cost of goods sold

10
Conversion costs include direct labor and overhead costs.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 45
Process Costing CMA Part 1

Items 2 and 3 in the preceding list are classified together as completed units transferred out of the
department. The costs for all units on which the current process’s work has been completed are transferred
either to finished goods inventory or to the next department or process for further work.

Whether the units have been sold (and the costs are in COGS), are still being worked on (are in ending WIP
for the company), or finished but not sold (in ending finished goods inventory for the company) is irrelevant
to the process in a given department. The objective of process costing is to allocate costs incurred to date
on products worked on in one department during one period between completed units and ending work-
in-process inventory for that department.

Note: The basic accounting for a process costing system is as follows:

All of the manufacturing costs incurred are debited to a WIP inventory account. Manufacturing costs
include direct material, direct labor, and factory overhead consisting of indirect materials, in-
direct labor and other factory overhead such as facility costs. Direct materials are usually added
to units in process in a different pattern from the way conversion costs (direct labor and overhead) are
added to units in process, so they are usually accounted for separately.

The costs in WIP then need to be allocated between units completed during the period and units remain-
ing in ending WIP at the end of the period. Costs for units completed during the period are transferred
to either finished goods inventory or, if more work is needed on them, to the next department’s WIP
inventory. This cost allocation is done based on a per unit allocation basis. Candidates do not need to be
familiar with the accounting steps in the process, just the process of allocating the costs, but the infor-
mation is presented because it may help candidates to see what is happening in the process.

Steps in Process Costing


The following will examine the steps in process costing in more detail. It is important for candidates to be
very comfortable with equivalent units of production (covered in much more detail later) and how they
are calculated. Equivalent units of production are used to allocate costs between completed units trans-
ferred out during the period and the incomplete units remaining in ending work-in-process inventory at the
end of the period. Equivalent units of production, or EUP, are an important concept in process costing and
one that is likely to be tested.

The steps in process costing are:

1) Determine the physical flow of goods

2) Calculate how many units were started and completed during the period

3) Determine when materials are added to the process

4) Calculate the equivalent units of production for materials and conversion costs

5) Calculate the costs incurred during the period for materials and conversion costs

6) Calculate the cost per equivalent unit for materials and conversion costs

7) Allocate the costs for materials and conversion costs separately between units in EWIP and units
transferred out according to the equivalent units in each

Each step is explained in detail in the following pages.

46 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

1) Determine the Physical Flow of Goods

Note: The formula to determine the physical flow of goods is used as the first step in process costing in
order to understand how many units to account for.

One of the most important things needed to solve a process costing problem is tracking each of the physical
units that go through the department—knowing where they came from and where they were at the end of
the period. A problem will give enough information to solve this type of situation, but understanding the
way the goods move is essential.

A process costing problem will usually involve units that are partially completed at the beginning of the
period (beginning WIP inventory) and units that are partially completed at the end of the period (ending
WIP inventory). An example of ending WIP is water bottles in the filling department at the end of the period
that are partially full.

The physical flow of goods formula utilizes full units, no matter how complete they are (for example, the
number of actual water bottles in the filling department at the end of the period, even if some of the bottles
are not yet filled). The physical flow of goods formula enables tracking of how many physical units went
into production and where they are at the end of the period. The formula is:

Units in Beginning WIP + Units Started/Transferred In

Units in Ending WIP + Units Completed/Transferred Out

Candidates need to be able to use the above formula to solve for any of the items within it.

Note: When solving a problem about physical flow, the completion percentages for any of the beginning
or ending WIP units are not important. The percentage of completion of the units in beginning and ending
WIP inventory is used in calculating equivalent units of production, which will be covered later. In the
physical flow of goods formula, though, only the number of physical units is used.

Question 15: Ben Company had 4,000 units in its work-in-process (WIP) on January 1. Each unit was
50% complete in respect to conversion costs. During the first quarter, 15,000 units were completed.
On March 31, 5,000 units were in ending WIP that were 70% complete in respect to conversion cost.
For this product, all of the direct materials are added when the unit enters the facility. How many units
did Ben start during the first quarter?

a) 13,000

b) 15,000

c) 16,000

d) 16,500

(HOCK)

2) Calculate the Number of Units Started and Completed

Note: The formula for calculating the number of units started and completed by itself is not going to be
tested on the exam, but it is important. The number of units started and completed is used later in
calculating the equivalent units of production.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 47
Process Costing CMA Part 1

The next step is to calculate how many units were both started and completed during the period. In
other words, how many units had all of their work done during the period. (For example, if the product is
bottled water, of the water bottles that were transferred into the filling department during the period, how
many were completely filled during the period?)

All of the units that were completed during the period were either in beginning inventory at the beginning
of the period or were started on (or transferred in) during the period. Subtracting the number of physical
units in beginning WIP from the total number of units completed results in the number of units both started
and completed during the period.

The formula below is used to calculate the number of units that were both started and completed during
the period (meaning that 100% of the work was performed, start to finish, during the current period).

# Units Completed − # Units in Beginning WIP = Units Started and Completed This Period

3) Determine When the Materials Are Added to the Process


The point at which the materials are added is used to determine the amount of materials that were added
to beginning work-in-process inventory during the previous period and to ending work-in-process inventory
during the current period. The information will be provided in the question. It is very important to identify
when the materials are added, so always look for that information in the question and be able to identify
whether the materials are added:

• at the beginning of the process,

• at the end of the process,

• at some point in the middle of the process, or

• evenly throughout the process (in which case they will behave like conversion costs).

Usually, conversion costs are added evenly throughout the process, but if conversion costs are added at a
specific point in the process, then those conversion costs would need to be treated like materials.

4) Calculate the Equivalent Units of Production (EUP)

Note: The concept of “equivalent units of production” is similar to the concept of “full-time equivalent
employees.” When an employer has half-time employees, the employer expresses the number of em-
ployees as “full-time equivalents.”

For example, if the employer has only one half-time employee, the employer has 0.5 of a full-time
equivalent employee. If the employer has two half-time employees, the employer has one full-time
equivalent employee. If the employer has 10 half-time employees, those 10 half-time employees are
equal to 5 full-time equivalent employees (10 × 0.5).

Similarly, if a unit in ending work-in-process inventory is 30% complete as to conversion activities, for
example, it represents 0.30 of an equivalent unit of conversion costs added. Therefore, if 100 units in
ending work-in-process inventory are 30% complete as to conversion costs, they represent 30 equivalent
units (100 × 0.30) of conversion costs added.

Keep that in mind while going through the following explanation of equivalent units of production.

To allocate costs to the individual units produced, it is necessary to know how many “units” the costs should
be allocated to. In a system where beginning and ending WIP in the department are both zero, allocation
of costs to units would be a simple matter of dividing the costs by the number of units both started and
completed during the period, since all the units started during the period were also completed during the
period and all the units completed during the period were also started during the period.

48 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

However, since usually a number of incomplete units are in both beginning and ending WIP, the process is
more complicated. The allocation of costs is accomplished by calculating the number of units that would
have been “completely produced” (100% completed) during the period if the company had simply produced
one unit from start to finish, and then started another unit. “Completely produced” means completed from
start to finish with all direct materials added and all conversion activities completed.

Equivalent units are used to mathematically convert partially completed units of product into an equiva-
lent number of fully completed units. The number of complete units that could have been produced during
the period is called the number of equivalent units of production.

By way of illustration, because the units in beginning WIP are partially completed, only the remaining
portion of the work (a partial unit of work) needs to be done in the current period to finish each unit in
BWIP. (For example, if a 1-liter water bottle is 60% full on January 1, only 40% of a liter needs to be added
to fill that bottle.) Similarly, because each unit in ending WIP is partially completed, less than a full unit of
work has been done on each of those units.

The number of equivalent units of production (EUP) serves to quantify the partially completed units in
BWIP and EWIP.

Essentially, three different amounts of work may apply to an individual unit during the period:

1) Completed (beginning work-in-process inventory that has been completed) meaning that some
of the work was done in the previous period.

2) Started and Completed (see Step 2 above for the calculation) meaning that the unit was started
on or transferred in during the current period and was completely finished during the same period.

3) Started (but not completed), or units that were started on or transferred in during the period
but the work on them was not finished at the end of the period and thus they have not yet been
transferred out of the process.

The number of equivalent units of production work done during the period on each of these three categories
of work are used to allocate the costs.

The idea of EUP is probably best explained with some examples. Example #2 contains a formula that is
used for the calculation of number of equivalent units of production.

Example #1: A company has 100 units in beginning WIP and each unit is 25% complete. That 25% of
the work was done during the previous period. If no other units were added to the system this period,
and if at the end of the period 100 units are complete, the number of equivalent units produced this
period is 75. The 100 units needed 75% of the work to be done on them during this period in order to
be complete (100% minus the 25% completed during the previous period). Therefore, the number of
equivalent units of work done during the current period is calculated as 100 × 0.75.

For example, 100 1-liter bottles in beginning WIP inventory are each 25% full. Twenty-five liters of water
were added to all of the bottles during the previous period. Therefore, 75 liters of water will need to be
added to all of the bottles to make them all complete. In other words, 75 “equivalent units” of water (or
75% of 100 liters) need to be added to the units in beginning WIP during the period to complete the
units.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 49
Process Costing CMA Part 1

Example #2: Assume that in addition to the 100 units in beginning WIP that are 25% complete at the
beginning of the period, 100 more units (empty bottles to be filled) were transferred in during the period,
and at the end of the period 10 units are in ending WIP that are 40% complete. Calculate

1) The number of units completed,


2) The number of units started and completed during the period,
3) The amount of work performed in terms of equivalent units to start the incomplete units that were
in ending WIP inventory, and
4) The total number of EUP during the period.

Solutions:

1) The number of units completed is 190, calculated using the physical flow of goods formula: Units in
BWIP (100) + Units transferred in (100) = Units in EWIP (10) + Units Completed (must be 190).

2) Having calculated the number of units completed, the number of units started and completed can
be determined (190 units completed − 100 physical units in BWIP = 90 units started and completed).
The number of units started and completed is used to calculate the number of equivalent units
produced.

3) The total number of EUP is calculated using the three different amounts of work that may apply to
an individual unit during the period.

The calculation of equivalent units of production is:

To Complete BWIP (100 × 0.75) = 75


Started and Completed (90 × 1.00) = 90
To Start EWIP (10 × 0.40) = 4
Total EUP 169

Though the concept of equivalent units of production is mathematically simple, candidates sometimes ini-
tially have trouble with it. So please make certain to spend the necessary time to become comfortable with
the way it works. The process will be explained further in the following pages.

Question 16: Ben Company operates a production facility that has three departments. The following
information is in respect to the second production department for the month of May:

Number of units in BWIP 200


% complete for BWIP 20%
Number of units started 1,300
Number of units completed 1,100
Number of units in EWIP 400
% complete for EWIP 80%

The equivalent units of production for Ben Company for May was:

a) 900

b) 1,260

c) 1,380

d) 1,620

(HOCK)

50 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

EUP for Materials and Conversion Costs Separately


In most problems, candidates will need to allocate direct materials costs and conversion costs separately
and so will need to calculate EUP for materials and conversion costs separately because materials are not
added to production in the same way as conversion costs are added to production. Usually, conversion
costs are added proportionately throughout the process, whereas direct materials are added at a specific
point in the process. This does not make the problem any more difficult; it simply means that twice as
many calculations are needed.

When the Materials Are Added


• In process costing questions, candidates must pay attention to when materials are added to the
process in order to calculate the EUP for materials in both beginning WIP inventory and ending
WIP inventory. If the materials are added at the beginning of the process, beginning WIP in-
ventory will have had 100% of the needed materials added, and no materials will need to be
added to complete beginning WIP inventory during the current period. Ending WIP inventory will
also have had 100% of the materials added.

• If the materials are added at the end of the process, no materials will have been added to
beginning WIP inventory, and all of the direct materials (100%) will need to be added to com-
plete the beginning WIP inventory during the current period. Ending WIP inventory will also have
had no materials added.

• If the materials are added at some point in the middle of the process (such as when the
process is 40% complete), compare that percentage with the percentage complete as to conversion
costs for the units in beginning WIP inventory and compare that percentage with the percentage
complete as to conversion costs for the units in ending WIP inventory.
o If the conversion activities have been completed beyond the point at which materials are added
(for example, materials are added when the process is 40% complete and BWIP is 60% com-
plete as to conversion costs), the units in BWIP will have had 100% of the needed materials
added during the previous period, and no materials will need to be added to complete the
beginning WIP inventory during the current period.
o If the conversion activities are not completed beyond that point (for example, materials are
added when the process is 40% complete and BWIP is 30% complete as to conversion costs),
the units in BWIP will have had no materials added during the previous period, and beginning
WIP inventory will need to have 100% of its materials added during the current period to
complete production.

• If the materials are added evenly throughout the process, the percentage of materials already
added to beginning WIP will be equal to the percentage of conversion costs added to beginning
WIP inventory. The remaining amount (100% minus the percentage already added) will need to
be added to complete beginning WIP inventory during the current period. A similar process applies
to ending WIP inventory. The percentage of materials added to ending WIP inventory will be equal
to the percentage of conversion costs added to ending WIP inventory.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 51
Process Costing CMA Part 1

Example – ending WIP inventory: The materials are added when the unit is 50% complete as to
conversion costs, and the ending WIP was 40% complete as to conversion costs. Therefore, the materials
had not yet been added to the units in ending WIP inventory, and the units in ending WIP inventory were
0% complete as to direct materials although the units were 40% complete as to conversion costs. There
were zero equivalent units of materials in ending WIP inventory, but there were equivalent units of
conversion costs equal to 40% of the number of physical units in ending WIP inventory.

However, if the units in ending WIP inventory were 60% complete for conversion costs, the full amount
of material needed for each unit would have been added to the units in ending WIP inventory and the
units in ending WIP inventory would have been 100% complete as to direct materials though they were
only 60% complete for conversion costs. In this case, the number of equivalent units of materials in
ending WIP inventory would be 100% of the number of physical units in ending WIP inventory, while the
number of equivalent units of conversion costs in ending WIP inventory would be 60% of the number of
physical units in ending WIP inventory.

Similar calculations are necessary for beginning WIP in order to calculate the work that was done in the
current period to complete those units in beginning WIP, as follows:

Example – beginning WIP inventory: Materials are added when the units are 50% complete as to
conversion costs, and the beginning WIP inventory was 35% complete as to conversion costs. Therefore,
the materials had not yet been added to the units in beginning WIP inventory although the units were
35% complete as to conversion costs. Thus, 100% of the materials, or equivalent units of materials
equal to 100% of the number of physical units, needed to be added during the current period to complete
the units in beginning WIP inventory.

The equivalent units of conversion costs that needed to be added to the units in beginning WIP inventory
during the current period were 65% of the number of physical units in beginning WIP inventory (100%
− 35% added during the previous period).

However, if the units in beginning WIP inventory were 70% complete as to conversion costs, 100% of
the direct materials would have been added to each unit in beginning WIP inventory during the previous
period although the units were 70% complete as to conversion costs. In this case no materials, or zero
equivalent units of materials, would need to be added during the current period to complete the units in
beginning WIP inventory.

The equivalent units of conversion costs that would have needed to be added to the units in beginning
WIP inventory during the current period would have been 30% of the number of physical units in begin-
ning WIP inventory (100% − 70% added during the previous period).

Note: How and when in the process materials are added is very important. It is used to deter-
mine how many equivalent units of direct materials, if any, were added in order to complete beginning
WIP inventory during the period and how many equivalent units of direct materials, if any, were added
to the incomplete units in ending WIP inventory before the current period ended.

When answering a process costing question, always look for the information on how and when the ma-
terials are added.

52 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

Question 17: Hoeppner Corp. uses process costing to allocate costs. In the Pressing Department, all of
the materials are added at the very beginning of the process. After this first addition, no additional
materials are added during the process. At the end of January, Hoeppner was presented with the fol-
lowing information:

Beginning work-in-process (60% complete for conversion costs) 2,000


Units started in January 5,000
Transferred out from Pressing during January 6,000
Ending work-in-process, (40% complete for conversion costs) 1,000

Under the EUP method already discussed, what are the equivalent units of production for the month of
January?

Materials Conversion Costs

a) 5,000 5,200

b) 5,000 5,600

c) 5,200 5,400

d) 5,200 5,200

(HOCK)

EUP and Inventory Tracking Methods


Another complication in the EUP calculation and the allocation of costs (and a much more significant com-
plication) is that the calculation of EUP is influenced by the inventory cost flow assumption the company
uses in its process costing. As with any other kind of inventory, a system is needed for determining whether
or not each unit that was completed was an old unit taken from the units that were in BWIP or not. In
process costing, two inventory cost flow assumptions are used:

1) First-in-first-out (FIFO) – In FIFO, the assumption is made that in each period the company will
finish what is in BWIP inventory before starting any new units. The FIFO method is what has been
used in the previous explanations and examples where total equivalent units for a period were
calculated.

2) Weighted Average (WAVG) – In WAVG the assumption is not made that the units in BWIP are
finished first. As a result, all of the units (both those from BWIP and those transferred in or started
in the current period) will be treated the same way. The costs of the work done during the previous
period on beginning WIP and the work done during the current period are combined and averaged,
creating the weighted averaging.

The main difference between FIFO and WAVG is the treatment of the costs that were assigned to the
units already in BWIP at the start of the period. The treatment of costs will be covered in much more
detail later, but very briefly:

• Under FIFO all of the costs in BWIP that were incurred during the previous period are automatically
allocated 100% to units completed and transferred to finished goods or to the next department,
while costs incurred during the period are allocated between units completed and units in ending
WIP inventory on the basis of equivalent units of production.

• Under WAVG the costs in BWIP incurred during the previous period are combined with the costs
incurred during the current period, and the total costs are divided by all the equivalent units worked
on during the period (whether the units were in BWIP or were started during the period) to calcu-
late a weighted average to use in allocating the total costs between completed goods and ending
WIP inventory on the basis of equivalent units.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 53
Process Costing CMA Part 1

Note: Candidates need to be able to make the calculations for EUP and to allocate costs under both the
FIFO and the WAVG method. The two methods are discussed in more detail individually below and a
simple formula for calculating EUP under each method is provided. To be able to answer process costing
questions, be sure to memorize the formulas.

EUP Under the FIFO Method


Under the FIFO method it is assumed that all of the units in BWIP at the beginning of the period were
completed during the current period before any other new units were started. Thus, under the FIFO
method, all of the units that were in BWIP were completed and were transferred out at the end of the
period.

Because the units in BWIP have been transferred out of the department, the costs that were already asso-
ciated with those units in BWIP (costs incurred in the current department during the previous period) are
kept separate from the costs incurred during the current period. The costs that were in BWIP at the begin-
ning of the period are 100% allocated to completed units transferred out during the current period. They
are not allocated between units transferred out and units remaining in ending WIP.

Therefore, only the costs incurred during the current period are allocated between completed units trans-
ferred out and units in EWIP. The costs incurred during the current period (only) are allocated
according to the work actually done during the period (only) as determined by the equivalent units of
production.

Since the costs are allocated on the basis of equivalent units of production, the equivalent units of produc-
tion used in the allocation include only the work done during the current period on units in beginning WIP
inventory and units started during the period. The work done during the previous period on units in begin-
ning WIP inventory is excluded from the calculation of EUP and from the allocation of current period costs
between units completed and units in EWIP.

In other words, under FIFO the cost per EUP calculation will include only the costs incurred during the
current period and three elements of work:

1) The work required to complete the units in BWIP

2) The work required for all of the units started and completed

3) The work done to start the EWIP

The costs incurred during the current period are allocated between units completed and transferred out
(items numbered 1 and 2 above) and units in EWIP (item 3 above). The costs and the EUP that were in
BWIP at the beginning of the period (for work done during the previous period) are not included in the
allocation between completed units and ending WIP inventory. Instead, the units in BWIP are considered
completed units for the period, and the costs incurred during the previous period for the units in BWIP are
transferred out in full either to finished goods inventory or to the next process because under the FIFO
assumption, first the units in BWIP are finished, and then other new units are started.

Thus, the costs transferred out for completed units equal (1) 100% of the costs in BWIP plus (2) an allo-
cation of the costs incurred during the period. The costs remaining in ending WIP are an allocation of the
costs incurred during the period.

Note: When the units are transferred out, the costs for BWIP, the costs incurred during the current
period, and any transferred-in costs are combined and become the transferred-in costs for the next
department. The costs are not kept separate in future departments. This means that the next department
will use a weighted average for the cost of the units transferred in and will not have similar units that
have different transferred-in costs.

54 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

The FIFO calculation of EUP during the period consists of three steps as follows (this is very important):

How Calculated
1) Completion of BWIP Units in BWIP × % of work done this period
2) + Started & Completed # of Started & Completed Units × 1
3) + Starting of EWIP # of units in EWIP × % of work done this period
= EUP this period TOTAL

EUP Under the Weighted Average Method


Under the Weighted Average cost flow assumption, the assumption is made that all of the costs and the
equivalent units of production in BWIP at the beginning of the period are simply added together with all of
the work done and costs incurred during the current period and averaged together for allocation between
units completed and units in EWIP. No distinction is made between the costs of the units that were
in BWIP and the units that were transferred in or started on during the period. Therefore, no dis-
tinction is made between the work done and costs incurred on the BWIP last period and the work done and
costs incurred this period to finish the BWIP.

The costs associated with BWIP that were incurred during the previous period are combined with the costs
that were incurred during the current period. Thus, the costs incurred during the previous period to start
the beginning WIP are treated as though they were incurred during the current period, and all of the
units that were in BWIP are considered to have been 100% worked on and completed during the current
period. The total combined cost is therefore allocated to all units worked on to determine their average
cost.

Again, in WAVG, the costs and EUP of work that already were held in BWIP at the start of the period are
considered to have been incurred and worked during the current period. Therefore, an average cost is
calculated for the work done last period that was still in the department at the beginning of this period as
well as for the work done this period.

Because the assumption is made that all of the work done on BWIP was done during the current period,
the calculation of EUP for the weighted average method is simpler than for FIFO. For WAVG, equivalent
units of production need to be calculated for only two categories of units: Units Completed and the Starting
of EWIP.

The formula for EUP under the WAVG Method has two steps, as follows (candidates must know this):

How Calculated
1) Units Completed # of Units completed during the period × 1
2) + Starting of EWIP # of units in EWIP × % of work done this period
= EUP this period TOTAL

Units completed is calculated as:

Physical Units in BWIP + Units Started/Transferred In – Physical Units in EWIP = Units Completed

The weighted average method essentially creates a weighted average of the production costs of the two
periods by putting them together (and hence, the name).

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 55
Process Costing CMA Part 1

Note: The total equivalent units of production under the weighted average method can never
be lower than the total equivalent units of production under FIFO. The two calculations of EUP
will be equal if no units were in beginning work-in-process inventory, but total EUP under FIFO can never
be higher than total EUP under the weighted average method. The difference in the EUP between the
two methods will be equal to the EUP that were in BWIP at the start of the period.

The following information is for the next four questions: Levittown Company employs a process
cost system for its manufacturing operations. All direct materials are added at the beginning of the
process, and conversion costs are added proportionately. The production schedule for November is:

Units
Work-in-process on November 1 (60% complete as to conversion costs) 1,000
Units started during November 5,000
Total units to account for 6,000

Units completed and transferred out from beginning WIP inventory 1,000
Units started and completed during November 3,000
Work-in-process on November 30 (20% complete as to conversion costs) 2,000
Total units accounted for 6,000

Question 18: Using FIFO, the EUP for direct materials for November are:

a) 5,000 units

b) 6,000 units

c) 4,400 units

d) 3,800 units

Question 19: Using FIFO, the EUP of conversion costs for November are:

a) 3,400 units

b) 3,800 units

c) 4,000 units

d) 4,400 units

Question 20: Using weighted-average, the EUP for materials for November are:

a) 3,400 units

b) 4,400 units

c) 5,000 units

d) 6,000 units

Question 21: Using weighted-average, the EUP for conversion costs for November are:

a) 3,400 units

b) 3,800 units

c) 4,000 units

d) 4,400 units

(CMA Adapted)

56 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

5) Calculation of Costs Incurred During the Period


After calculating the EUP during the period the next step is to determine the costs that will be considered
to have been incurred during the period under each inventory method.

Costs Incurred Under the FIFO Method


Under the FIFO method all of the costs associated with the units in BWIP at the start of the period are
transferred out. The costs in BWIP do not go into the “process” during the period because under FIFO, the
units in BWIP are all assumed to have been finished before any other units are started. Therefore, all of
the costs in BWIP will end up as costs for units completed and transferred out during the current period.
Only the costs that were actually incurred during the current period will be allocated between units com-
pleted and transferred out and ending WIP according to the EUP in each during the period. Thus, the costs
allocated to units completed and transferred out will consist of two components:

1) All of the costs incurred by this department during the previous period for the EUP in BWIP.

2) An allocated portion of costs incurred by this department during the current period for the units
that were started and completed during the period.

Costs Incurred Under the Weighted Average Method


Under the weighted average method all of the costs in BWIP are added to the costs that were incurred
during the current period. This treatment is the same thing that was done with the calculation of EUP under
the weighted average method. The costs in BWIP and the costs incurred during the period are combined
and allocated to either completed units or to ending WIP inventory according to the EUP in each during the
period as calculated under the WAVG method (which included the EUP that were in BWIP at the beginning
of the period).

Because the costs of the current and previous period are mixed, the result is a weighted average.

Selecting an Inventory Cost Flow Method


The weighted average method is simpler to use than the FIFO method since only one calculation is needed
to allocate the costs. However, the weighted average method mixes together costs, so any change in the
cost of an input is in a sense covered up by the weighted average. The weighted average method is
best used when prices are stable.

Because the FIFO method keeps the costs of the two periods separate, FIFO is preferable when prices
of inputs are changing frequently.

6) Calculation of the Cost per EUP


Once the EUP have been determined (under either FIFO or WAVG), and the costs to be allocated have been
identified, the next step is to determine a rate (or unit cost) per EUP for each cost element, usually raw
materials and conversion costs. The cost per equivalent unit is calculated by dividing the total costs to be
allocated for each cost element (Step #5) by the total equivalent units for that cost element. The calculation
is done separately for materials and conversion costs (Step #4). The cost per EUP for materials and
conversion costs must be calculated separately because the total equivalent units for materials
and conversion costs may be different.

Note: Remember that if using the FIFO cost flow assumption, all of the costs associated with the BWIP
are transferred 100% to FG or to the next department. They do not need to be allocated between
completed units and ending WIP inventory and thus they are excluded from the calculation of equivalent
units of production and from the allocation based on equivalent units of production.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 57
Process Costing CMA Part 1

7) Allocation of the Costs to the Products


After the rates per EUP for materials and conversion costs have been determined, the last step is to allocate
the costs to completed units transferred out and to EWIP based on the number of equivalent units that
were completed and the number of equivalent units in EWIP. This allocation is simply a mathematical
exercise that requires multiplying the EUP in each place by the rate (or cost) per EUP.

Note: It is usually easiest to allocate the costs to EWIP first and then all remaining costs end up in units
transferred out. When FIFO is being used, however, it is critical not to forget to allocate the costs in
BWIP to units transferred out, as well.

On the following two pages are diagrams illustrating the FIFO and Weighted Average process costing sys-
tems. Note that the only difference between them is the treatment of BWIP.

• Under FIFO, the costs of BWIP go directly to completed units transferred out.

• Under the weighted average method, the costs in BWIP are included with the costs added into the
“process” for the current period and are allocated. Similarly, under the weighted average method
the work that had been done in previous periods on the BWIP is also added to the “process” and
is considered to have been done this period.

The diagrams will again go through the steps of process costing, and the diagrams will be followed by an
example that goes through all of the steps for both FIFO and WAVG.

58 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

Process Costing Diagram – FIFO


1
Beginning WIP Conversion Costs Added Materials Added During
During Period Period

Beginning WIP contains a These are the costs paid for These are the costs paid for
number of units that have labor and overhead during materials during the period
had some work (EUP) done the period to convert the to produce the product.
on them and have been allo- materials into the finished
cated costs incurred during product.
the previous period.

Costs associated with Cost of Conversion Costs Costs of Materials


BWIP

The Process

In the process, the EUP for ma-


terials and conversion costs
are calculated.

A cost per EUP is calculated


and used to allocate the costs
incurred between units Com-
pleted and Transferred Out and
EUP in Ending WIP.

Transferred Out Ending WIP

# of units completed that EUP of Materials ×


were NOT in BWIP × Cost/EUP of Materials
cost/EUP for whole unit +
(Materials and Conv. Costs) EUP of Conv. Costs ×
Cost/EUP of Conv. Costs

1
Under FIFO the units in BWIP are considered to have been completed during the period, so the costs
associated with BWIP are all transferred out with the completed units. The EUP in BWIP are not used in
the allocation of costs incurred during the current period because the costs in BWIP are allocated 100%
to units completed and transferred out.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 59
Process Costing CMA Part 1

Process Costing Diagram – Weighted Average


2
Beginning WIP Conversion Costs Added Materials Added During
During Period Period
Beginning WIP contains a
number of units that have These are the costs paid for These are the costs paid for
had some work (EUP) done labor and overhead during materials during the period
on them and have been allo- the period to convert the to produce the product.
cated costs incurred during materials into the finished
the previous period. product.

Costs associated with


BWIP AND EUP associated Cost of Conversion Costs Cost of Materials
with BWIP

The Process

In the process, the EUP of


materials and conversion
costs are calculated. The
units that were in BWIP are
included and it is assumed
that that work was done
during this period

A cost per EUP is calculated


and used to allocate the
costs incurred between units
Transferred Out and EUP in
Ending WIP

Transferred Out Ending WIP

# of units completed × EUP of Materials ×


cost/EUP for whole unit Cost/EUP of materials
(Materials and Conv. Costs) +
EUP of Conv. Costs ×
Cost/EUP of Conv. Costs

2
Under WAVG the units in BWIP are considered to have been done completely during the current period.
The costs in BWIP are combined with the costs incurred during the current period and the total cost is
divided by the work done during the previous period on the units in BWIP and the work done during the
current period (total EUP) to calculate the cost per EUP.

60 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

Process Costing Summary


The seven steps in a process costing question are shown again below in an abbreviated form.

1) Determine the physical flow of units of goods. The formula is:

Units in BWIP + Started or Transferred In = Units in EWIP + Completed/Transferred


Out

2) Calculate how many units were started and completed during the period. The formula is:

Started and Completed = Units Completed or Transferred Out – Units in BWIP

3) Determine when the materials are added to the process.

4) Calculate the equivalent units of production during the period. The calculation of equivalent
units will most likely need to be done twice—once for materials and once for conversion costs. If
materials are added at the beginning of the process (or at any other point in the process) the EUP
calculation for materials will need to be done separately. However, if the materials and the con-
version costs are both added evenly (or proportionately) throughout the process, only one
calculation of equivalent units is needed, as the total equivalent units will be the same for materials
and conversion costs.

For FIFO:

1) Completion of BWIP Units in BWIP × % work done this period


2) + Started and Completed Number of Started & Completed Units × 1
3) + Starting of EWIP Units in EWIP × % work done this period
= EUP this period TOTAL

For Weighted Average:

1) Units Completed* Units completed during the period × 1


2) + Starting of EWIP Units in EWIP × % work done this period
= EUP this period TOTAL
* Units Completed is calculated as:
Physical Units in BWIP + Units Started or Transferred In – Physical Units in EWIP = Units Completed

5) Calculate the costs incurred during the period.

For FIFO, only costs that were actually incurred during the period are included in the amount to
be allocated between units completed and units in Ending WIP. Costs in BWIP are allocated 100%
to units completed.

For WAVG, the costs allocated between units completed and units in Ending WIP will include those
actually incurred during the period, plus the costs that were incurred during the previous period in
this department to start the units that were in BWIP at the start of the period.

The costs for materials and conversion costs need to be calculated separately.

6) Calculate the cost per equivalent unit for materials and conversion costs by dividing the
costs from Step 5 by the EUP calculated in Step 4. The cost per equivalent unit is done separately
for materials and conversion costs.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 61
Process Costing CMA Part 1

7) Allocate the costs between completed units and EWIP. The allocation of costs between com-
pleted units and units in EWIP is done by multiplying separately for materials and conversion costs
the equivalent units of production for each that were calculated in Step 4 by the cost per EUP for
each that were calculated in Step 6. If FIFO is being used, also allocate 100% of the costs in
beginning WIP inventory for materials and conversion costs to completed units.

Process Costing Examples


On the following two pages are examples showing how the entire process works for both the FIFO and the
Weighted Average methods. The same facts will be used for both methods and the monetary value of EWIP
and FG will be calculated using both the FIFO and Weighted Average methods. Exam questions will be
easier when candidates understand the entire process, as the exam questions generally ask about only one
part of the process.

62 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

FIFO Example: Beginning WIP inventory is 150 units (60% complete as to conversion costs). Beginning
WIP inventory material costs are $250 and conversion costs are $500. All material for the product is
added at the beginning of the production process. Conversion takes place continuously throughout the
process. During the period 550 units were started. Material costs for the period are $800 and conversion
costs for the period are $6,000. Ending WIP is 80 units at 20% completion as to conversion costs.

Calculate the amounts to be in EWIP and FG using the FIFO method.

1) Determine the physical flow.


BWIP + Started = EWIP + Units Completed
150 + 550 = 80 + 620

2) Calculate the Units Started and Completed.


S&C = Units Completed – BWIP
470 = 620 – 150

3) Materials are added at the beginning of the process. Therefore, 100% of the required materials
were added to the units in BWIP during the previous period and 100% of the required materials have
been added to the units in EWIP during the current period. Thus, the number of equivalent units of
direct material in each is equal to the number of physical units in each.

4) Calculate equivalent units of production. Under FIFO, the costs in BWIP are allocated 100% to
units completed and transferred out. Therefore, only costs incurred during the current period are
allocated between completed units and units in EWIP on the basis of equivalent units.
Materials Conversion Costs
Complete BWIP 0 60 = 150 × 40%
Started & Completed 470 470
Start EWIP 80 16 = 80 × 20%
TOTAL EUP 550 546

5) Calculate costs incurred. Under FIFO, only costs actually incurred during the period are allocated
between units completed and units in EWIP.
Materials $ 800
Conversion Costs $6,000

6) Incurred costs per equivalent unit.


Materials $ 800 ÷ 550 = $ 1.46
Conversion Costs $ 6,000 ÷ 546 = $10.99

7) Allocate incurred costs to EWIP.


Materials $ 1.46 × 80 = $ 116.80
Conversion Costs $10.99 × 16 = 175.84
Total $ 292.64

Allocate BWIP costs and incurred costs to finished goods.


Costs of BWIP (100% to finished goods) $250 + $500 = $ 750.00
Materials $1.46 × 470 = 686.20
Conversion Costs $10.99 × 530 = 5,824.70
Total $ 7,260.90

Total allocated costs $7,553.54 *

* The total cost to allocate is $750 in BWIP, $800 materials added, and $6,000 conversion costs added,
for a total of $7,550. The $3.54 difference between the total cost to allocate and the total of the
allocated costs is a rounding difference that can be dealt with by adjusting the individual components
by small amounts so that the allocations total to $7,550.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 63
Process Costing CMA Part 1

WAVG Example: Beginning WIP inventory is 150 units (60% complete as to conversion costs). Begin-
ning WIP inventory material costs are $250 and conversion costs are $500. All material for the product
is added at the beginning of the production process. Conversion takes place continuously throughout the
process. During the period 550 units were started. Material costs for the period are $800 and conversion
costs for the period are $6,000. Ending WIP is 80 units at 20% completion as to conversion costs.

Calculate the amounts to be in EWIP and FG using the Weighted Average method.

1) Determine the physical flow.


BWIP + Started = EWIP + Units Completed
150 + 550 = 80 + 620

2) Calculate the Units Started and Completed.


S&C = Units Completed – BWIP
470 = 620 – 150

3) Materials are added at the beginning of the process. Therefore, 100% of the required materials
were added to the units in BWIP during the previous period and 100% of the required materials have
been added to the units in EWIP during the current period. Thus, the number of equivalent units of
direct material in each is equal to the number of physical units in each.

4) Calculate equivalent units of production.


Materials Conversion Costs
Units COMPLETED 620 620
EWIP 80 16 = 80 × 20%
TOTAL 700 636

5) Calculate costs to allocate using EUP. Under WAVG the costs associated with BWIP are added to
the costs incurred during the period.
Materials $250 in BWIP + $ 800 incurred = $1,050
Conversion Costs $ 500 in BWIP + $ 6,000 incurred = 6,500
Total costs to allocate $7,550

6) Costs per equivalent unit (including costs in BWIP).


Materials ($250 + $800) ÷ 700 = $ 1.50
Conversion Costs ($ 500 + $ 6,000) ÷ 636 = $10.22

7) Allocate costs to EWIP.


Materials $ 1.50 × 80 = $ 120.00
Conversion Costs $10.22 × 16 = 163.52
Total $ 283.52

Allocate costs to finished goods.


Materials $1.50 × 620 = $ 930.00
Conversion Costs $10.22 × 620 = 6,336.40
Total $ 7,266.40

Total allocated costs $7,549.92 *

* The total cost to allocate is $7,550. The $0.08 difference between the total cost to allocate and the
total of the allocated costs is a rounding difference that can be dealt with by adjusting the individual
components by small amounts so that the allocations total to $7,550.

Note: If the company is in its first month of operation, it will have no BWIP. In that case, the cost
allocations under FIFO and WAVG will be the same because the treatment of BWIP is the only difference
between the two methods.

64 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing
efham CMA
The following information is for the next two questions: A sporting goods manufacturer buys
wood as a direct material for baseball bats. The Forming Department processes the baseball bats, which
are then transferred to the Finishing Department where a sealant is applied. The Forming Department
began manufacturing 10,000 "Casey Sluggers" during the month of May. There was no beginning in-
ventory.

Costs for the Forming Department for the month of May were as follows:

Direct materials $33,000


Conversion costs 17,000
Total $50,000

A total of 8,000 bats were completed and transferred to the Finishing Department; the remaining 2,000
bats were still in the forming process at the end of the month. All of the Forming Department's direct
materials were placed in process, but, on average, only 25% of the conversion cost was applied to the
ending work-in-process inventory.

Question 22: The cost of the units transferred to the Finishing Department is:

a) $50,000

b) $40,000

c) $53,000

d) $42,400

Question 23: The cost of the work-in-process inventory in the Forming Department at the end of May
is:

a) $10,000

b) $2,500

c) $20,000

d) $7,600

(CMA Adapted)

Question 24: Nance Co. began operations in January 20X5 and uses the FIFO Process Costing method
in its accounting system. After reviewing the first quarter results, Nance is interested in how the equiv-
alent units of production would have been different if the weighted average method had been used
instead. Assume that the number of units in ending work-in-process was the same at the end of January,
February and March and that at the end of each month they were the same percentage complete. If
Nance had used the weighted average method, the EUP for conversion costs for each of the first two
months would have compared to the FIFO method in what way?

January February

a) Same Same

b) Greater number Greater number

c) Greater number Same

d) Same Greater number

(HOCK)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 65
Process Costing CMA Part 1

The following information is for the next two questions: Kimbeth Manufacturing uses a process
cost system to manufacture Dust Density Sensors for the mining industry. The following information
pertains to operations for the month of May.

Units
Beginning work-in-process inventory, May 1 16,000
Started in production during May 100,000
Completed production during May 92,000
Ending work-in-process inventory, May 31 24,000

The beginning inventory was 60% complete for materials and 20% complete for conversion costs. The
ending inventory was 90% complete for materials and 40% complete for conversion costs.

Costs pertaining to the month of May are as follows:

• Beginning inventory costs are materials, $54,560; direct labor, $20,320; and factory overhead,
$15,240.

• Costs incurred during May are materials used, $468,000; direct labor, $182,880; and factory over-
head, $391,160.

Question 25: Using the weighted-average method, the cost per equivalent unit for materials during May
is

a) $4.12

b) $4.50

c) $4.60

d) $5.02

Question 26: Using the weighted-average method, the cost per equivalent unit for conversion costs
during May is

a) $5.65

b) $5.83

c) $6.00

d) $6.20

(CMA Adapted)

Question 27: At the end of the first month of operations, Larkin had 2,000 units in ending WIP that
were 25% complete as to conversion costs. All materials are added at the end of the process. The
number of equivalent units of conversion costs would be:

a) Equal to the number of units completed during the period.

b) Less than the number of equivalent units of materials.

c) Less than the number of units completed during the period.

d) Less than the number of units placed into the process during the period.

(HOCK)

66 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

Spoilage in Process Costing


Some units worked on during a period are never completed and also are no longer in EWIP at the end of
the period because they are spoiled and are identified as defective during the process. When spoiled units
are detected, the spoiled units are removed from processing immediately. No further direct materials are
added to them and no further conversion work is done on them. The term used for defective units that are
removed from processing is spoilage.

Up to this point, all of the costs that are in the department (costs in beginning WIP, costs incurred during
the period, or costs transferred in) have been allocated either to completed units—units transferred to
finished goods inventory (FG) or to the next process—or to ending WIP inventory. Now that spoilage has
been added to the equation, spoiled units must be included as another possible destination for a portion of
the costs that are in the department.

The formula to calculate the physical flow of units of goods when spoilage is included needs to be expanded
to the following:

Units in BWIP + Started/Transferred In = Units in EWIP + Completed/Transferred Out + Spoiled Units

When units are spoiled, material will have been added to the spoiled units and some of the needed conver-
sion work will have been done on them before they were identified as spoiled. Therefore, a portion of the
material costs, conversion costs, and transferred-in costs, if any, must be allocated to the spoiled units.
The treatment of spoilage is similar to the treatment of ending WIP inventory—the units were started but
not finished. The spoiled units have costs associated with them based on what has been done to them,
similar to EWIP.

The allocation of costs to the spoiled units is based on equivalent units, and the number of equivalent units
worth of each cost element used in the allocation depends on when the spoiled units are identified as
spoiled.

Whenever spoilage exists, the following issues need to be addressed:

1) How many units were spoiled?

2) How are the spoiled units classified—as normal spoilage or abnormal spoilage?

3) What costs are allocated to each spoiled unit?

4) What is done with the costs allocated to the spoiled units?

1) How many units were spoiled?


The number of units (physical units) spoiled are the units that did not pass inspection and that were re-
moved from the production process at the inspection point.

Though the number of spoiled units is usually given in the question, it is important to identify the point at
which the spoiled units were removed from the production process, because that information will
be needed to calculate the equivalent units associated with the spoiled units.

2) How are the spoiled units classified – as normal spoilage or abnormal spoilage?
For each process, the company will have established standards regarding the number of defective (spoiled)
units that are expected to occur. The expected spoilage is normal spoilage and spoilage above that is
abnormal spoilage. All spoiled units are classified as either normal or abnormal.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 67
Process Costing CMA Part 1

The way the standard for normal spoilage is presented in the question is used to determine the number of
spoiled units to be classified as normal spoilage. The expected amount of spoilage (normal spoilage) is
usually presented in a question in one of four ways, as follows:

1) As a percentage of the number of units started into production

2) As a percentage of the units that completed production

3) As a percentage of the good units that passed inspection

4) As a percentage of the total number of units that were inspected

The number of spoiled units up to and including, but not exceeding, the number of spoiled units expected
to occur are classified as normal spoilage. Any spoiled units in excess of the expected spoilage are classified
as abnormal spoilage. The distinction between normal and abnormal spoilage is important because the
costs associated with the two different classifications of spoilage are treated differently.

3) What costs are allocated to each spoiled unit?


If the spoiled units are identified as spoiled after processing has begun in the current department, they will
have had some conversion costs allocated to them and possibly materials as well, depending on when in
the process the materials are added. If costs incurred in a previous department or departments were trans-
ferred in along with the units, some of the transferred-in costs will also need to be allocated to the spoiled
units. The different costs are allocated to the spoiled units based on their EUP. The equivalent units of
spoiled units are calculated the same way as equivalent units are calculated for ending WIP inventory, so
another line is added to the EUP worksheet calculation.

In a problem with spoilage, the formula for FIFO now looks like the following:

1) Completion of BWIP Units in BWIP × % work done this period


2) + Started and Completed Number of Started & Completed Units × 1
3) + Starting of EWIP Units in EWIP × % work done this period
4) + Starting of Spoiled Units Units Spoiled × % work done this period
= EUP this period TOTAL

In a problem with spoilage, the formula for WAVG now looks like the following:

1) Units Completed Units completed during the period × 1


2) + Starting of EWIP Units in EWIP × % work done this period
3) + Starting of Spoiled Units Units Spoiled × % work done this period
= EUP this period TOTAL

Allocation of Costs to the Spoiled Units


After the cost per EUP for materials and conversion costs and transferred-in costs (if applicable) are calcu-
lated as discussed above, the company will allocate costs to each unit, including the spoiled units. The
allocation of costs is made in the same manner as the allocation of costs are made to EWIP: the number
of EUP for spoiled units is multiplied by the cost per EUP. Again, this allocation will probably be done
separately for materials and conversion costs and will definitely be done separately for transferred-in costs,
if any.

68 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

4) What is done with the costs allocated to the spoiled units?


After determining the direct materials costs, conversion costs, and transferred-in costs to be allocated to
the normally and abnormally spoiled units, those costs must be transferred somewhere at the end of the
period. The spoiled units are neither finished goods nor ending work-in-process, but the costs have been
incurred in the department and must be moved out of the department.

The treatment of spoilage costs will depend on the type of spoilage.

Normal Spoilage
If the spoilage is normal spoilage, the costs that have been allocated to the normally spoiled units are
added to the costs of the good units that are transferred out to finished goods (or to the next
department). As a result, the cost per unit transferred out will be higher than the actual cost of producing
a good unit in the current department.

Abnormal Spoilage
It is generally considered that production management can control abnormal spoilage. Normal spoilage is
expected to occur and generally cannot be prevented. However, abnormal spoilage should not occur and
should therefore be preventable.

Therefore, the costs that have been allocated to the abnormally spoiled units will be expensed on the
income statement in that period as a loss from abnormal spoilage.

Note for FIFO only: To simplify the allocation of costs to spoiled units under FIFO, the spoiled units
are accounted for as though they were all started in the current period. Although the spoiled units
probably included some units from beginning work-in-process inventory, all spoilage is treated as if it
came from current production and thus a portion of only the current production costs incurred, along
with a portion of any costs transferred in during the period, are allocated to the spoiled units.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 69
Process Costing CMA Part 1

Question 28: A company employs a process cost system using the FIFO method. The product passes
through both Department 1 and Department 2 in order to be completed. Units enter Department 2 upon
completion in Department 1. Additional direct materials are added in Department 2 when the units have
reached the 25% stage of completion with respect to conversion costs. Conversion costs are added
proportionally in Department 2. The production activity in Department 2 for the current month was:

Beginning work-in-process inventory (40%


complete with respect to conversion costs) 15,000
Units transferred in from Department 1 80,000
Units to account for 95,000

Units completed and transferred to finished goods 85,000


Ending work-in-process inventory (20%
complete with respect to conversion costs) 10,000
Units accounted for 95,000

How many equivalent units for direct materials were added in Department 2 for the current month?

a) 70,000 units.

b) 80,000 units.

c) 85,000 units.

d) 90,000 units.

(CMA Adapted)

Question 29: During May 20X5, Mercer Company completed 50,000 units costing $600,000, exclusive
of spoilage allocation. Of these completed units, 25,000 were sold during the month. An additional
10,000 units, costing $80,000, were 50% complete at May 31. All units are inspected between the
completion of manufacturing and transfer to finished goods inventory. Normal spoilage for the month
was $20,000, and abnormal spoilage of $50,000 was also incurred during the month. The portion of
total spoilage that should be charged against revenue in May is

a) $50,000

b) $20,000

c) $70,000

d) $60,000

(CMA Adapted)

70 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Process Costing

Benefits of Process Costing

• Process costing is the easiest, most practical costing system to use to allocate costs for homogeneous
items. It is a simple and direct method that reduces the volume of data that must be collected.

• Process costing can aid in establishing effective control over the production process. Process costing
can be used with standard costing by using standard costs as the costs to be allocated. Use of standard
costing with process costing makes it possible to track variances and to recognize inefficiency in a
specific process.

• It is flexible. If the company adds or removes a process, it can adapt its process costing system easily.

• Management accountants can review the amounts of materials and labor used in each process to look
for possible cost savings.

• Process costing enables obtaining and predicting the average cost of a product, an aid in providing
pricing estimates to customers.

Limitations of Process Costing

• Process costing can introduce large variances into the costing system if standard costs allocated to the
units are not up to date. Depending on how the variances are resolved, the variances could cause the
product to be over- or under-costed, which could lead to pricing errors.

• Process costing can be time-consuming for management accountants.

• Calculating the equivalent units for beginning and ending work-in-process inventory can lead to inac-
curacies, because the percentage of completion of those inventories may be subjective (an estimate
or even a guess).

• Process costing cannot provide an accurate cost estimate when a single process is used to produce
different (joint) products.

• Process costing is not suitable for custom orders or diverse products.

• Because the work of an entire department is combined in the process costing system, process costing
makes it difficult to evaluate the productivity of an individual worker.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 71
Job-Order Costing CMA Part 1

Job-Order Costing
Job-order costing is a cost system in which all of the costs associated with a specific job or client are
accumulated and charged to that job or client. The costs are accumulated on what is called a job-cost sheet.
All of the job sheets that are still being worked on equal the work-in-process at that time. In a job-order
costing system, costs are recorded on the job-cost sheets and not necessarily in an inventory account.

Job-order costing can be used when all of the products or production runs are unique and identifiable from
each other. Audit firms and legal firms are good examples of job-order costing environments. As employees
work on a particular client or case, they charge their time and any other costs to that specific job. At the
end of the project, the company simply needs to add up all of the costs assigned to it to determine the
project’s cost. Performance measurement can be done by comparing each individual job to its budgeted
amounts or by using a standard cost system.

While direct materials and direct labor are accumulated on an actual basis, manufacturing overhead must
be allocated to each individual job. Overhead allocation is done in much the same manner as has already
been explained. A predetermined overhead rate is calculated and applied to each product based either on:

• Actual usage of the allocation base (as in normal costing)

• Standard usage of the allocation base (as in standard costing)

Multiple cost allocation bases may be used if different overheads have different cost drivers. For example,
in a manufacturing environment, machine hours for each job may be used to allocate overhead costs such
as depreciation and machine maintenance, whereas direct labor hours for each job may be used to allocate
plant supervision and production support costs to jobs. If normal costing is being used, actual machine
hours and actual direct labor hours will be used. If standard costing is being used, the standard machine
hours allowed and the standard direct labor hours allowed for the actual output on each job will be used.

Note: Under job-order costing, selling and administrative costs are not allocated to the products
in order to determine the COGS per unit. Selling and administrative costs are expensed as period costs.

Question 30: Lucy Sportswear manufactures a line of T-shirts using a job-order cost system. During
March, the following costs were incurred completing Job ICU2: direct materials, $13,700; direct labor,
$4,800; administrative, $1,400; and selling, $5,600. Factory overhead was applied at the rate of $25
per machine hour, and Job ICU2 required 800 machine hours. If Job ICU2 resulted in 7,000 good shirts,
the cost of goods sold per unit would be:

a) $6.50

b) $6.30

c) $5.70

d) $5.50

(CMA Adapted)

Note: On the Exam, a numerical question about job-order costing is most likely to be nothing more than
a question where a standard overhead application rate needs to be used to apply overhead to specific
jobs.

72 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Job-Order Costing

Benefits of Job Order Costing

• Job order costing is best for businesses that do custom work or service work.

• Job order costing enables the calculation of gross profit on individual jobs, which can help management
determine in the future which kinds of jobs are desirable.

• Managers are able to keep track of the performance of individuals for cost control, efficiency, and
productivity.

• The records kept result in accurate costs for items produced.

• Management can see and analyze each cost incurred on a job in order to determine how it can be
better controlled in the future.

• Costs can be seen as they are added, rather than waiting until the job is complete.

Limitations of Job Order Costing

• Employees are required to keep track of all the direct labor hours used and all the materials used.

• The focus is on direct costs of products produced. The focus on direct costs can allow for inefficiencies
and increasing overhead costs.

• Depending on the type of costing being used, overhead may be applied to jobs on the basis of prede-
termined rates. If the overhead application rates are out of date, the costing can be inaccurate.

• If overhead is applied on the basis of predetermined rates and the rates are not calculated on any
meaningful basis, the cost of each job will not be meaningful.

• To produce meaningful results, job order costing requires a lot of accurate data entry. Because of the
massive amount of data recording required, input errors can occur and if not corrected, the errors can
lead to poor management decisions.

• The use of job order costing is limited to businesses that do custom manufacturing or service work.
It is not appropriate for high volume manufacturing or for retailing.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 73
Operation Costing CMA Part 1

Operation Costing
Operation costing is a hybrid, or combination, of job-order costing and process costing. In operating
costing, a company applies the basic operation of process costing to a production process that produces
batches of items. The different batches all follow a similar process, but the direct materials input to each
batch are different.

Examples of manufacturing processes where operation costing would be appropriate are clothing, furniture,
shoes and similar items. For each of these items, the general product is the same (for example, a shirt),
but the materials used in each shirt may be different.

In operation costing the direct materials are charged to the specific batch where they are used, but con-
version costs are accumulated and distributed using a predetermined conversion cost per unit. Conversion
costs are allocated by batch.

An operation costing worksheet would look very much like a process costing worksheet, except it would
require a separate column for each product’s direct materials, while the worksheet would have one con-
version costs column that would pertain to the conversion of all the products.

Life-Cycle Costing
Life-cycle costing is another type of costing that is useful only for internal decision-making.

In life-cycle costing a company does not determine the production cost in the short-term sense of the
production of one unit. Rather, the company takes a much longer view to the cost of production and at-
tempts to allocate to each product all of the research and development, marketing, after-sale service and
support costs, and any other cost associated with the product during its life cycle. The life cycle of the
product may be called its value chain.11

The longer-term view used in life-cycle costing is of particular importance when the product has significant
research and development (R&D) costs or other non-production costs such as after sale service and support
costs associated with it. For the product to be profitable over its life, these non-production costs also need
to be covered by the sales price. If the company fails to take into account the large costs of R&D, it runs
the risk of the sales price covering the costs of the actual production of each unit sold but not covering the
costs of R&D, marketing, after-sale service and support, and other costs.

In the process of looking at all of the costs, the company should be able to determine the ultimate value of
developing a better product. In addition to R&D, costs include after-sale costs such as warranties and repair
work and product support expense. It may be that a larger investment in the design or development of the
product will be recovered through smaller after-sale costs. Alternatively, the company may realize that
additional design costs will not provide sufficient benefit later to make the additional investment in design
and development feasible.

Note: Life-cycle costing is different from other costing methods because it treats pre-production and
after-sale costs as part of the product costs, whereas other methods treat those costs as period expenses
that are expensed as incurred. Therefore, under other methods, pre-production and after-sale costs are
not directly taken into account when determining the profitability of a product or product line.

11
The term “value chain” refers to the steps a business goes through to transform inputs such as raw materials into
finished products by adding value to the inputs by means of various processes, and finally to sell the finished products
to customers. The goal of value chain analysis is to provide maximum value to the customer for the minimum possible
cost.

74 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Life-Cycle Costing

All of the costs in the life cycle of a product can be broken down into three categories. The three categories
and the types of costs included in each are:

Upstream Costs (before production)

• Research and Development

• Design – prototyping (the first model), testing, engineering, quality development

Manufacturing Costs

• Purchasing

• Direct and indirect manufacturing costs (labor, materials and overhead)

Downstream Costs (after production)

• Marketing and distribution

• Services and warranties

For external financial reporting under GAAP, R&D and design costs and the other life-cycle costs other than
manufacturing costs are expensed as incurred. However, for internal decision-making purposes, it is im-
portant for the company to treat all of these costs as product costs that need to be recovered from the sale
of the product.

Life-cycle costing plays a role in strategic planning and decision-making about products. After making the
life-cycle cost calculations, the company can make an assessment as to whether or not the product should
be manufactured. If management believes the company will not be able to charge a price high enough to
recover all the life-cycle product costs for the product, then it should not be produced.

Furthermore, by looking at all of the costs expected to be incurred in the process of developing, producing
and selling the product, the company can identify any non-value-adding costs, which can then be reduced
or eliminated without reducing the value of the product to the customer.

Benefits of Life-Cycle Costing

• Life-cycle costing provides a long-term, more complete perspective on the costs and profitability of
a product or service when compared to other costing methods, which typically report costs for a short
period such as a month or a year.

• When long-term costs are recognized in advance, life-cycle costing can be used to lower those long-
term costs.

• Life-cycle costing includes research and development costs as well as future costs such as warranty
work, enabling better pricing for profitability over a product’s lifetime.

• Life-cycle costing can be used to assess future resource requirements such as needed operational
support for the product during its life.

• Life-cycle costing can help in determining when a product will reach the end of its economic life.

Limitations of Life-Cycle Costing

• When life-cycle costing is used to spread the cost of fixed assets over the life of a product, the
assumption may be made that the fixed assets will be as productive in later years as they were when
they were new.

• Accurate estimation of the operational and maintenance costs for a product during its whole lifetime
can be difficult.

• Cost increases over the life of the product need to be considered.

• Life-cycle costing can require considerable time and resources, and the costs may outweigh the ben-
efits.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 75
Customer Life-Cycle Costing CMA Part 1

Customer Life-Cycle Costing


Customer life-cycle costing looks at the cost of the product from the customer’s (the buyer’s) standpoint.
It focuses on the total costs that will be paid by the customer during the whole time the customer owns the
product: the customer’s purchase costs plus costs to use, maintain, and dispose of the product or service.
For example, the life-cycle cost of laundry equipment includes the purchase cost plus the cost for energy
to operate it over its lifetime, the cost of repairs, and the cost to dispose of it at the end of its life, if any.

Customer life-cycle costing is important to a company because it is part of the pricing decision. If a product
is expected to require minimal maintenance when compared with its competition, the company can charge
a higher price for the product than the price the competition is charging for their products, and the total
cost to the customer may still be lower than the cost for the competitor’s product.

Example: BusinessSoft Co. is about to launch a new product. The company expects a 6-year life cycle
from the moment it starts developing its product through its last sale and installation of the product.
However, it also expects to provide after-sale services as part of the contract during and beyond the 6-
year life cycle period.

The company’s cost estimates are:

R&D $750,000
Design 500,000
Manufacturing costs 300,000
Marketing 200,000
Distribution 100,000
Customer service 250,000
After-sale support 60,000

The company plans to produce and sell 1,500 installations of the product and would like to earn a 40%
mark-up over its life-cycle costs relating to the new product. The company envisions that an average
client would incur around $500 of installation, training, operating, maintenance, and disposal costs
relating to usage of the product per installation. What is the expected total life-cycle cost per
installation for BusinessSoft, and how much does the company need to charge per
installation?

Solution: The life-cycle costs to BusinessSoft include all of the upstream, manufacturing, and down-
stream costs related to the new product. The upstream costs are the research and development
($750,000) and the design ($500,000) costs. Manufacturing costs are $300,000. Downstream costs
include all of the other costs listed above, from marketing to after-sale support, totaling $610,000. In
total, the life-cycle costs are $2,160,000.

With a required 40% mark-up over the life-cycle cost, a mark-up of $864,000 ($2,160,000 × 0.40) will
be required. Therefore, the price that BusinessSoft needs to charge for each of the 1,500 installations is
calculated as ($2,160,000 + $864,000) ÷ 1,500, or $2,016 per installation.

If the company charges $2,016 per installation, it will be able to recover all of the life-cycle costs
associated with the new product and earn a 40% mark-up on those costs.

The customer’s life-cycle cost will be the customer’s purchase price of $2,016 per installation plus the
estimated $500 per installation for the installation, training, operating, maintenance, and disposal costs,
for a total of $2,516. The additional $500 cost to the customer is relevant information to the company
in making its final pricing decision, but the customer’s additional $500 costs are not included with the
company’s costs when calculating the company’s total life-cycle cost for the product.

Note: The costs incurred by the customer are relevant only for customer life-cycle costing and as input
into pricing decisions made by the company.

76 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

D.3. Overhead Costs


Introduction to Accounting for Overhead Costs
In general, overheads are costs that cannot be traced directly to a specific product or unit. Overheads are
actually of two main types: manufacturing (or factory) overheads and nonmanufacturing overheads. Man-
ufacturing overheads are overheads related to the production process (factory rent and electricity, for
example), whereas nonmanufacturing overheads are not related to the production process. Examples of
nonmanufacturing overheads are accounting, advertising, sales, legal counsel and general corporate ad-
ministration costs.

The allocation of manufacturing overheads is covered first in the topic that follows. Allocation of manufac-
turing overheads can be accomplished by a variety of methods that include traditional allocation, process
costing, job order costing, operation costing, activity-based costing, and life-cycle costing. All of those
methods, with the exception of life-cycle costing, can be used for external financial reporting, although
some of the principles of activity-based costing must be adapted in order for it to be used for external
reporting because in principle, activity-based costing does not conform to generally accepted accounting
principles. Methods that cannot be used for external financial reporting can be used internally for decision-
making.

The allocation of nonmanufacturing overheads is covered next in Shared Services Cost Allocation. However,
some of the concepts and ideas covered in manufacturing overhead are also applicable in the allocation of
nonmanufacturing overheads.

Manufacturing Overhead Allocation

Note: In order to help the following explanations flow more easily, the term “overhead” will be used in
the majority of situations, even though the term “manufacturing overhead” would be more technically
accurate. If “manufacturing overhead” were used in every situation, the language would become cum-
bersome and be more difficult to read. Also, the term “factory overhead” can be used in place of
“manufacturing overhead” because the two are interchangeable terms.

The three main classifications of production costs are:

1) Direct materials

2) Direct labor

3) Manufacturing (or factory) overhead

Direct materials (DM) and direct labor (DL) are usually simple to trace to individual units or products be-
cause direct materials and direct labor costs are directly and obviously part of the production process.
Therefore, the CMA exam does not put much emphasis on the determination of DM and DL. Rather, the
emphasis is on the allocation of overhead.

Overhead costs are production and operation costs that a company cannot trace to any specific product or
unit of a product. Because overhead costs are incurred and paid for by the company and are necessary for
the production process, it is essential that the company know what these costs are and allocate them to
the various products being manufactured. Allocation to products manufactured must occur so that the full
costs of production and operation are known in order to set the selling prices for the different products. If
a company does not take overhead costs into account when it determines its selling price for a product, it
runs a significant risk of pricing the product at a loss because the price the company charges may cover
the direct costs of production but not the indirect costs of production.

Furthermore, generally accepted accounting principles require the use of absorption costing for external
financial reporting. In absorption costing, all overhead costs associated with manufacturing a product be-
come a part of the product’s inventoriable cost along with the direct costs. Therefore, all manufacturing
overhead costs must be allocated to the units produced.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 77
D.3. Overhead Costs CMA Part 1

The categories of costs included in factory overhead (OH) are:

1) Indirect materials – materials not identifiable with a specific product or job, such as cleaning
supplies, small or disposable tools, machine lubricant, and other supplies.

2) Indirect labor – salaries and wages not directly attributable to a specific product or job, such as
those of the plant superintendent, janitorial services, and quality control.

3) General manufacturing overheads, such as facilities costs (factory rent, electricity, and utili-
ties) and equipment costs, including depreciation and amortization on plant facilities and
equipment.

Note: Remember that factory overhead and manufacturing overhead are interchangeable terms
that mean the same thing. Either may be used in a question.

Overheads may be fixed or variable or mixed.

• Fixed overhead, like any fixed cost, does not change with changes in activity as long as the
activity remains within the relevant range. Examples of fixed manufacturing overhead are factory
rent, depreciation on production equipment, and the plant superintendent’s salary.

• Variable overheads are costs that change as the level of production changes. Examples of vari-
able manufacturing overheads are indirect materials and equipment maintenance.

• Mixed overheads contain elements of both fixed and variable costs. Electricity is an example of
a mixed overhead cost because electricity may be billed as a basic fixed fee that covers a certain
number of kilowatts of usage per month and usage over that allowance is billed at a specified
amount per kilowatt used. A mixed overhead cost could also be an allocation of overhead cost from
a cost pool containing both fixed and variable overhead costs.

The number of ways in which a company can allocate overhead are numerous and limited only by the
imagination of the accountant, and now, the computer programmer. However, for the CMA Exam, candi-
dates need to be familiar with only the primary methods of overhead allocation. The primary methods are
the traditional overhead allocation method and activity-based costing.

• The traditional method of overhead allocation involves grouping manufacturing overhead costs
into a cost pool12 and allocating them to individual products based on a single cost driver, 13 such
as direct labor hours or machine hours. Traditional overhead allocation may involve the use of
separate cost pools for fixed overhead and variable overhead, though fixed and variable overhead
may also be combined into a single cost pool.

• Activity-based costing (ABC) involves using multiple cost pools and multiple cost drivers to
allocate overhead on the basis of cost drivers specific to each cost pool. Activity-based costing in
its pure form does not conform to GAAP and thus cannot be used for external financial reporting.
Product costs used in external financial reports must include all manufacturing costs and only
manufacturing costs, whereas under pure activity-based costing, some manufacturing costs are
excluded and some non-manufacturing costs are included in the cost pools. However, if activity-
based costing is adapted so that all manufacturing overhead costs and only manufacturing over-
head costs are allocated to products, ABC can be used for external financial reporting.

Regardless of the manner of allocation used, overhead allocation is simply a mathematical exercise of
distributing the overhead costs to the products that were produced using some sort of basis and
formula.

12
A cost pool is a group of indirect costs that are being grouped together for allocation on the basis of some cost
allocation base.
13
A cost driver is anything (it can be an activity, an event or a volume of something) that causes costs to be incurred
each time the driver occurs.

78 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

Traditional (Standard) Allocation Method


Traditionally, manufacturing overhead costs have been allocated to the individual products based on
either the direct labor hours, machine hours, materials cost, units of production, weight of production or
some similar measure that is easy to measure and calculate. The measure used is called the activity base.

For example, if a company allocates factory overhead based on direct labor hours, for every hour of direct
labor allowed per unit of output (under standard costing) or used per unit of output (under any other type
of cost measurement system), a certain amount of factory overhead is allocated to, or applied to, each
unit actually produced. The determination of how much overhead is allocated per unit is covered below.

By summing the direct materials, direct labor, and allocated manufacturing overhead costs, a company
determines the total cost of producing each specific unit of product.

Determining the Basis of Allocation


When choosing the basis of allocation (for example, direct labor hours or machine hours), the basis used
should closely reflect the reality of the way in which the costs are actually incurred. For example, in a
highly-automated company, direct labor would most likely not be a good allocation basis for factory over-
head because labor would not be a large part of the production process. The allocation basis does not need
to be direct labor hours or machine hours, though those are the most common bases used. For example,
in a company that produces very large, heavy items (such as an appliance manufacturer), the best basis
on which to allocate overhead may be the weight of each product.

Plant-Wide versus Departmental Overhead Allocation


A company can choose to use plant-wide overhead allocation or departmental overhead allocation.

Plant-wide overhead allocation involves putting all of the plant-wide overhead costs into one cost pool
and then allocating the costs in that cost pool to products using one allocation basis, usually machine hours
or labor hours.

Alternatively, a company can choose to have a separate cost pool for each department that the products
pass through in production. This second method is called departmental overhead allocation. Each de-
partment’s overhead costs are put into a separate cost pool, and then the overhead is allocated according
to the cost basis that managers believe is best for that department.

Note: When process costing is being used and products pass through several different processes, or
departments, before becoming finished goods, departmental overhead allocation is essential so that
overhead costs can be applied to the products as production activities take place in each department.

In both plant-wide and department overhead allocation, fixed overhead costs can be segregated in a sep-
arate cost pool from variable overhead costs and the fixed and variable overheads allocated separately.
The fixed and variable overheads can be allocated using the same cost basis, or they can be allocated using
different cost bases. For planning purposes and in order to calculate fixed and variable overhead variances,
it is virtually essential to segregate fixed and variable overhead costs.

Note: A cost basis, or basis of cost allocation, is a measure of activity such as direct labor hours or
machine hours that is used to assign costs to cost objects. A cost object is a function, organizational
subdivision, contract, or other work unit for which cost information is desired and for which provision is
made to accumulate and measure the cost of processes, products, jobs, capital projects, and so forth.

For example, if Department A uses very little direct labor but a lot of machine time, Department A’s over-
head costs would probably be allocated to products on the basis of machine hours. If Department B uses a
lot of direct labor and very little machine time, Department B’s overhead costs would probably be allocated
to products on the basis of direct labor hours. If Department C is responsible for painting the products,
Department C’s overhead costs might be allocated based on square footage or square meterage of the

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 79
D.3. Overhead Costs CMA Part 1

painted products. A department that assembles products may allocate overhead costs based on the number
of parts in each product.

To best reflect the way that manufacturing overhead is incurred, departmental overhead allocation is
preferable to plant-wide overhead allocation. The greater the number of manufacturing overhead allocation
rates used, the more accurate or more meaningful the overhead allocation will be. However, departmental
overhead allocation requires a lot more administrative and accounting time than plant-wide allocation and
thus is more costly. The more bases used to allocate overhead, the more costs will be incurred to obtain
the needed information for the allocation. Therefore, a company needs to find a balance between the use-
fulness of having more than one overhead allocation basis against the cost of making the needed
calculations for the additional bases.

Departmental overhead allocation would be chosen by a company’s management if it felt the benefit of the
additional information produced would be greater than the cost to produce the information. For example,
the additional information could be used to develop more accurate product costs for use in setting prices
and making other decisions.

Note: The only time that it may be applicable to use only one overhead application rate for the
factory is when production is limited to a single product or to different but very similar products.

Calculating the Manufacturing Overhead Allocation Rate


Once the method, or basis, of manufacturing overhead allocation is determined, the predetermined man-
ufacturing overhead allocation rate is calculated. The predetermined rate is the amount of
manufacturing overhead that will be charged (allocated) to each unit of a product for each unit of the
allocation basis (direct labor hours, machine hours, and so on) used by that product during production.

The predetermined overhead rate may be a combined rate including both variable and fixed overheads; or
it may be calculated separately for variable overhead and fixed overhead and applied separately. Whichever
way it is done, the total overhead allocated to production will be the same if the same allocation base is
used for both fixed and variable overhead.

It is important to note that fixed overhead is applied to units produced as if it were variable over-
head, even though fixed costs do not behave the same way variable costs behave. Actual variable costs
increase in total as production increases and decrease in total as production decreases. However, as long
as the production level remains within the relevant range, actual fixed manufacturing costs do not change
in total as production increases and decreases. Therefore, actual fixed manufacturing cost per unit (total
fixed manufacturing cost divided by the actual number of units produced) must change as the production
level increases and decreases.

Although actual fixed manufacturing overhead may not vary much in total from budgeted fixed manufac-
turing overhead, the variance between the amount of actual fixed manufacturing overhead incurred and
the amount of fixed manufacturing overhead applied to production can be significant because of the fact
that fixed overhead is applied to production the same way variable overhead is, on a per-unit basis, but it
is not incurred that way. Therefore, the fixed manufacturing overhead component of total overhead costs
creates a large part of the reported variances.

The rate used to allocate overhead is usually calculated at the beginning of the year, based on budgeted
overhead for the coming year and the budgeted level of activity for the coming year.

Unless material changes in actual overhead costs incurred during the year necessitate a change to the
predetermined rate, that rate (or those rates, if fixed and variable overheads are allocated separately) will
be used to allocate manufacturing overheads throughout the year. Because the manufacturing overhead
allocation rate is set before the production takes place, it must use budgeted, or expected, amounts. The
manufacturing overhead allocation rate is called the predetermined rate because it is calculated at the
beginning of the period.

80 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

Note: It is important to remember that the manufacturing overhead allocation rate is calculated at the
beginning of the year and then used throughout the year unless it becomes necessary to change it during
the year.

The predetermined overhead rate is calculated as follows:

Budgeted Monetary Amount of Manufacturing Overhead


Budgeted Activity Level of the Allocation Base

The budgeted activity level of the allocation base is the number of budgeted direct labor hours, direct
labor cost, material cost, or machine hours—whatever is being used as the allocation basis—allowed (under
standard costing) for the budgeted output. The budgeted activity level will be discussed in greater detail
later in this explanation.

However, as noted above, the application rate should be reviewed periodically and adjusted if necessary so
that the amount applied is a reasonable approximation of the current overhead costs per unit.

Note: Since the predetermined rate is a calculated rate using Budgeted Manufacturing Overhead divided
by Budgeted Activity Level of the Allocation Base, the predetermined rate and the budgeted activity level
can easily be used to reverse the process and calculate the budgeted overhead amount that was used
to calculate the predetermined rate. Such an exercise will frequently be required, particularly for fixed
overhead.

Predetermined rate × Budgeted Activity (Production) Level = Budgeted Manufacturing Overhead

The above formula is particularly important to remember when working with fixed manufacturing over-
head, since neither actual nor budgeted fixed overhead is affected in total by the number of
units actually produced or the amount of the allocation base actually used as long as production
remains within the relevant range.

Only the amount of fixed overhead applied is affected by the actual production activity level.

It is critical to use the budgeted amount of manufacturing overhead for the numerator and the budg-
eted activity level of the allocation base for the denominator in calculating the predetermined overhead
allocation rate. Do not use budgeted for one and actual for the other. The budgeted amount of manufac-
turing overhead is the cost the company expects to incur and the budgeted activity level is the number of
units of the allocation base allowed for the expected production during the upcoming year. The company
determines the allocation rate at the beginning of the year and uses it for the entire year for the application
of the manufacturing overhead costs unless something changes that requires the allocation rate to be
adjusted.

Note: The calculation of the allocation rate can also be done on a weekly or a monthly basis. In such a
situation, the process would be exactly the same except that the budgeted overhead cost and activity
level used would be for the upcoming week or month (or whatever time period is used).

Clearly, the budgeted rate is not going to be the actual rate that occurs during the year. However, in order
to determine the cost of goods produced throughout the year so their cost can flow to inventory when
produced and then to cost of goods sold when they are sold, an estimated rate must be used. A company
cannot wait until the end of the year to determine what its cost of production was. As long as the rate is
reviewed periodically and adjusted when necessary, however, variances can be minimized.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 81
D.3. Overhead Costs CMA Part 1

Determining the Level of Activity


In relation to the allocation rate, the company must decide what level of activity to use for its budgeted
activity level of the allocation base in the denominator of the overhead predetermined rate calculation. The
level of activity used is the number of machine hours, direct labor hours, or whatever other activity base
the company plans to use during the year. The activity level is estimated in advance. As the activity level
is one of the two figures used in the determination of the manufacturing overhead rate, it will greatly impact
the allocation rate.

A company has several choices for the activity level to use in calculating a predetermined allocation rate
for overhead. U.S. GAAP, in ASC 330, specifically prescribes that normal capacity should be used for
external financial reporting. The other activity levels can be used for internal reporting and decision-making,
however.

The choices, the impact of using each option, and what each measure is best used for by a company are in
the table below.

Activity Level What it is Impact of Using It Best Used For

Normal capacity The level of activity that Normal capacity utilization Long-term planning
utilization will be achieved in the long is the level of activity that
run, taking into account will satisfy average cus-
seasonal changes in the tomer demand over a long-
business as well as cyclical term period such as 2 to 3
changes. years.

This is required by U.S.


GAAP.

Practical (or The theoretical activity The practical capacity basis Pricing decisions
currently attain- level reduced by allowances is greater than the level
able) capacity for unavoidable interrup- that will be achieved and
tions such as shutdowns for will result in an under-
holidays or scheduled application of manufactur-
maintenance but not de- ing overhead.
creased for any expected
decrease in sales demand.

Master budget The amount of output actu- Results in a different over- Developing the
capacity utiliza- ally expected during the head rate for each budget master budget
tion (expected next budget period based period because of increases
Current performance
actual capacity on expected demand. or decreases in planned
measurement
utilization) production due to expected
increases or decreases in
demand.

Theoretical, or The level of activity that A company will not be able Has very little use in
ideal, capacity will occur if the company to achieve the theoretical practical situations,
produces at its absolute level of activity, and if that because it will not be at-
most efficient level at all level is used to calculate tained.
times. the overhead application
rate, manufacturing over-
head will be under-
applied because the re-
sulting application rate will
be too low.

82 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

Example: RedHawks Co. produces bookshelves for shipment to distributors. RedHawks’ fixed manufac-
turing overheads for the coming year are expected to be $250,000. In a perfect situation, RedHawks
has the capacity to produce 40,000 bookshelves. During the current year, RedHawks produced 38,000
bookshelves, the most in company history. Management thinks this high production level was attributa-
ble to a performance bonus that was in place for management during the current year, but that bonus
will not be repeated next year. For the seven years prior to the current year, RedHawks produced an
average of 34,000 bookshelves per year, with production always between 32,500 and 35,500. Due to
an expected decrease in demand next year, the company expects to produce 36,000 bookshelves next
year.

The CFO of RedHawks is trying to determine the cost to produce each bookshelf in order to determine
the price to charge. The CFO knows that a portion of the fixed manufacturing overhead must be allocated
to each of the units. Below are the calculations for the fixed overhead allocation rate using the four
different levels of production.

Calculating the Predetermined Rate


1) If RedHawks uses theoretical capacity, the fixed overhead allocation rate for next year will be
$6.25 per unit ($250,000 ÷ 40,000 units).

2) If the fixed manufacturing overhead allocation rate is determined using the current year’s perfor-
mance as the practical capacity, the allocation rate will be $6.58 per unit ($250,000 ÷ 38,000
units).

3) If the master budget capacity is used in calculating the predetermined fixed overhead allocation
rate, the rate will be $6.94 per unit ($250,000 ÷ 36,000 units).

4) Finally, if Redhawks uses what had been the normal capacity prior to the current year, the alloca-
tion rate per unit would be $7.35 ($250,000 ÷ 34,000 units).

Allocation of Fixed Manufacturing Overhead Under the Traditional Method


The year is now complete. RedHawks actually produced 35,750 units during the year. Fixed overhead
was allocated per unit. The allocation was made by multiplying the actual number of units produced by
the predetermined rate per unit. Under each of the four levels of capacity, a different amount of fixed
manufacturing overhead would have been allocated to the products.

1) Using theoretical capacity, RedHawks would have allocated $223,438 ($6.25 × 35,750 units) to the
units produced.

2) Using practical capacity, it would have allocated $235,235 ($6.58 × 35,750 units).

3) Using master budget capacity, it would have allocated $248,105 ($6.94 × 35,750 units).

4) Using normal capacity, it would have allocated $262,763 ($7.35 × 35,750 units).

Summary Analysis
Because the actual production was different from any of the bases that could be used, the amount of
fixed overhead allocated would not have been exactly correct no matter which capacity level was chosen.
However, the amount of overhead allocated under master budget capacity would be the closest to the
expected amount of $250,000.

If RedHawks had used the theoretical base at the beginning of the year (40,000) in its calculation of cost
per unit, the company would have run the risk of underpricing the bookshelves. The fixed overhead rate
per unit would have been too low and not enough overhead would have been allocated to the bookshelves
that were made. If management had used normal capacity, too much overhead would have been allo-
cated to each bookshelf, and management might have overpriced the bookshelves.

If the company used master budget capacity (36,000), the price charged would fluctuate from one year
to the next, as the expected production level fluctuated from budget period to budget period based on
anticipated demand and in turn cost per unit would have fluctuated.

(Continued)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 83
D.3. Overhead Costs CMA Part 1

Using practical capacity (38,000) would have resulted in an allocation of fixed overhead that would be
lower than that under normal or master budget capacity. The total cost per unit would not have increased
because of a lower expected demand during the coming year, and the company would not have increased
the price next year to reflect the higher cost, possibly avoiding causing further decreases in demand.

For pricing decisions, practical capacity is the best capacity level to use for calculating the
fixed overhead allocation rate because it can avoid the downward demand spiral caused by cost
increases when production falls, which leads to price increases and further reductions in demand and
further production decreases, and so on.

The result of using different capacity levels for overhead allocation will be discussed in more detail later
under the topic Capacity Level and Management Decisions.

Allocating Manufacturing Overhead to the Units


Once the overhead allocation rate has been determined, the company can allocate overhead to the individ-
ual units produced. Overhead is allocated by multiplying the predetermined rate by the number of units
of the allocation basis that were either allowed to be used (under standard costing) or were actually used
(under other costing methods) in the production of each unit.

Overhead allocation is a very simple mathematical operation, but it becomes a little more involved because
management must decide which cost allocation method to use: standard, normal, or actual. The three
methods were discussed in the topic of Costing Systems, so they will just be reviewed quickly here as they
apply to overhead application.

Overhead Overhead
Application Rate Allocation Base

Predetermined Stand- Standard Amount of Alloca-


Standard Costing ard tion Base Allowed
Rate for Actual Production

Estimated Amount of Allocation Base


Normal Costing Normalized Actually Used for
Rate Actual Production

Amount of Allocation Base


Actual
Actual Costing Actually Used for
Rate
Actual Production

84 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

Example: A company that allocates overhead on the basis of machine hours has the following budgeted
and actual results for 20XX:

Budgeted Actual
Overhead cost $250,000 $288,000
Production volume (units) 100,000 125,000
Total machine hours 200,000 240,000

How much overhead would be allocated to the production under standard, normal and actual costing?

Standard Normal Actual


Predetermined OH Rate: $250,000 ÷ 200,000 $1.25 $1.25
Actual OH Rate: $288,000 ÷ 240,000 $1.20
Allocation Base:
Standard number of machine hours 250,0001
Actual number of Machine hours 240,000 240,000
OH applied to production $312,500 $300,000 $288,000
1
Two hours are allowed for each unit, calculated as 200,000 total machine hours budgeted divided by
budgeted production volume of 100,000 units. The standard number of machine hours is the actual
production volume of 125,000 multiplied by 2 machine hours allowed per unit.

Why would the company not want to use actual costing, since the amount of overhead applied under
that system is correct whereas the amounts applied under the other systems are not accurate?

The total actual costs for the year are not known until sometime after the end of the reporting period
because of the delay in receiving and recording invoices. In most cases, overhead needs to be applied
to production as the production takes place. It cannot wait until the year has been closed out and the
total actual costs are known. Therefore, an estimated rate is used throughout the year. The difference
between the actual amount of overhead incurred and the amount applied (called over-applied or under-
applied overhead) is closed out after the total costs for the year are known by either debiting or crediting
cost of goods sold for the total amount of the variance or by distributing the variance between cost of
goods sold and inventories. See the following topic for further information.

Fixed Overhead Application to Production


For direct materials, direct labor, and variable overhead, the amount of standard cost applied to production
is the same as the flexible budget amount of cost allowed for the actual production. For example, the
amount of variable overhead applied to production is the same as the flexible budget amount of variable
overhead. If variable overhead is under-applied, it means that the actual cost incurred was greater than
the amount applied to production and also greater than the flexible budget by the same amount.

However, fixed overhead is different. Fixed overhead does not change in total as the level of production
changes; however, fixed overhead is applied to production as if it were a variable cost that does
change in total as the level of production changes. Therefore, the amount of fixed overhead applied
to production is not the same as the amount of fixed cost in the flexible budget, as is the case with variable
overhead and other variable manufacturing costs.

The amount of fixed cost in the flexible budget is the same as the amount of fixed cost in the static budget
because the amount of budgeted fixed cost does not change with changes in production level as long as
the activity remains within the relevant range.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 85
D.3. Overhead Costs CMA Part 1

Example: Total manufacturing overhead is budgeted to be $960,000 for the budget year: $360,000 for
variable manufacturing overhead and $600,000 for fixed manufacturing overhead. Output, based on
normal capacity, is budgeted at 600,000 units for the year. Overhead is allocated based on machine
hours. Two machine hours are allowed per unit, so the budgeted activity level of the allocation base is
1,200,000 machine hours per year (600,000 budgeted units × 2 hours per unit).

The standard variable overhead allocation rate is $360,000 ÷ 1,200,000 machine hours, or $0.30 per
MH. Since two machine hours are allowed per unit, the standard variable overhead allowed per unit is
$0.30 × 2, or $0.60 per unit.

The standard fixed overhead allocation rate is $600,000 ÷ 1,200,000 machine hours, or $0.50 per MH.
Since two machine hours are allowed per unit, the standard fixed overhead allowed per unit is $0.50 ×
2, or $1.00 per unit.

The actual number of units produced during the first quarter were 48,000 units in January, 52,000 in
February, and 58,000 in March, or 158,000 units for the quarter. Actual fixed overhead incurred was the
same as budgeted fixed overhead, or $50,000 per month ($600,000 ÷ 12). The total variable and fixed
overhead applied to production during each month of the first quarter, compared with the actual costs
incurred, are:
Actual Variable Variable Overhead Actual Fixed Fixed Overhead Ap-
Overhead Incurred Applied @ $0.60/Unit Overhead Incurred plied @ $1.00/Unit
January $29,500 $28,800 $50,000 $48,000
February $31,000 $31,200 $50,000 $52,000
March $35,000 $34,800 $50,000 $58,000
The amounts applied can be calculated either by multiplying the actual number of units produced by the
allocation rate per unit ($0.60 per unit for variable overhead and $1.00 per unit for fixed overhead) or
by multiplying the number of machine hours allowed for the actual number of units produced (2 hours
per unit × the actual production) by the allocation rate per machine hour ($0.30 per MH for variable
overhead and $0.50 per MH for fixed overhead).

The allocation of variable overhead for January has been calculated below using both methods to demon-
strate that the result is the same:

48,000 units produced × $0.60 per unit = $28,800.

48,000 units produced × 2 machine hours allowed per unit × $0.30 per machine hour = $28,800.

Variable overhead is under-applied by $700 for January ($29,500 − $28,800) and fixed overhead is
under-applied by $2,000 for January ($50,000 − $48,000). When production increases to 52,000 in
February, variable overhead is over-applied by $200 while fixed overhead is over-applied by $2,000.

The larger variances in the application of fixed overhead versus those of variable overhead are due to
the fact that fixed overhead is applied to production as though it were a variable cost even
though it is not a variable cost. Because of that, variances between actual production volume and
planned production volume will have a larger effect on fixed overhead manufacturing cost variances than
they will have on variable overhead manufacturing cost variances.

If actual production continues to exceed budgeted production, the standard application rate for fixed
overhead should be decreased so that fixed overhead is not over-applied.

Note: The difference in cost behavior between fixed and variable manufacturing overhead costs is very
important to understand.

86 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

The Process of Accounting for Factory Overhead


Factory overhead includes all costs, both fixed and variable, that cannot be traced to the production of a
specific unit or group of units. Factory overhead includes supervisory and other indirect salaries and wages,
depreciation on production facilities, indirect materials used such as cleaning rags, indirect facility costs
such as utilities, and an almost limitless list of other costs.

As factory overhead costs are actually incurred, the actual incurred costs are debited to a factory over-
head (OH) control account with the following journal entry:
Dr Factory overhead control ................................................ X
Cr Cash (or accounts payable) ................................................... X

The factory overhead control account is used to accumulate (collect) all the actual factory overhead costs
that the company incurs during the year. If the company has several cost pools, for instance for variable
overhead and fixed overhead in each of its various departments, it will have a separate factory overhead
control account for each cost pool. It is from this account that overhead will be allocated to work-in-process
and then to finished goods.

As each unit is produced, some of the overhead cost that has accumulated in the factory overhead control
account is transferred to the work-in-process (WIP) account with the following journal entry, using the
calculated, predetermined overhead rate and the amount of the allocation base allowed for the actual
output (assuming standard costing is being used):
Dr WIP inventory ................................................................... X
Cr Factory overhead control .................................................. X

The credit may instead be to an account called factory overhead applied, simply to keep the debits in
one account and the credits in a different account. The factory overhead applied account follows the factory
overhead control account in the chart of accounts and carries a credit balance. The net of the balances in
the two accounts at any point in time represents under-applied or over-applied overhead.

The work-in-process inventory account is one of the inventory accounts for the company (the others are
finished goods and raw materials). The process of debiting WIP inventory and crediting either factory OH
control or factory OH applied is called applying the overhead to production, and when the overhead is
transferred into the WIP account, it becomes applied.

Thus, the factory OH control account receives the debits for the actual costs the company incurs, and then
the costs are allocated out to the WIP account to apply them to the work-in-process inventory using the
predetermined overhead application rate. If the credits used to apply the overhead are made directly to the
factory OH control account, graphically it looks like the following:

Work-in-Process Inventory
Factory Over- Account: Costs Applied to
Actual Costs head Control Production @ Predetermined
Account Manufacturing OH Rate ×
Quantity of Allocation Base

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 87
D.3. Overhead Costs CMA Part 1

If a separate account, factory overhead applied, is used to apply the costs to production, it would look like
this:

Factory Over-
head Control
Account
Actual Costs
Debit Balance
Work-in-Process Inventory
Factory Over- Account: Costs Applied to
head Applied Production @ Predetermined
Account Manufacturing OH Rate ×
(Credit Balance) Quantity of Allocation Base

The amount of overhead applied to production from the factory OH control account will not be exactly the
same amount as the actual costs that were debited to the factory OH control account when the costs were
incurred.

• If the amount of overhead applied to units produced during a period is less than the actual incurred
amounts that were debited to the factory OH control account, the overhead is under-applied.
When overhead is under-applied, each unit received a lower amount of overhead than it should
have.

• If the amount of overhead applied to units produced during a period is greater than the actual
incurred amounts debited to the factory OH control account, overhead is over-applied. When
overhead is over-applied, each unit received a greater amount of overhead than it should have.

The amount of the difference between the amount incurred and the amount applied is a variance, and it
remains in the factory OH control account until the period is closed. The process of closing out variances
will be explained in more detail in the pages that follow.

After the overhead costs are moved into the WIP account, they will be in one of three places in the financial
statements at the end of the year. Where the cost of a particular unit is reported in the financial statements
depends on what happened to that unit by the end of the year. The three possible locations in the financial
statements are:

1) In ending WIP inventory on the balance sheet if the item has not yet been completed at the end
of the period.

2) In finished goods inventory on the balance sheet if the item has been completed but not sold.

3) As an expense in cost of goods sold on the income statement if the item has been sold.

88 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

Over-Applied and Under-Applied Manufacturing Overhead


Chances are very good that the actual costs incurred and the actual level of activity during the period
will be different from the estimates made at the beginning of the year. Since the costs are being allocated
using the budgeted costs and budgeted usage of the allocation base, if the actual costs or usage are differ-
ent from the budgeted, the allocation will be, in essence, “incorrect.” As a result, the factory overhead
control account (or the net of the two accounts, if two separate accounts are used) will have a balance in
it at the end of the year because the credits for the allocated costs will not be equal to the debits repre-
senting the incurred costs. The remaining balance may be a positive (debit) or negative (credit) balance,
but whichever it is, it needs to be eliminated from the account (or accounts).

• If the balance is a debit balance, it means that factory overhead was under-applied to the
production. In other words, the amount of factory overhead that was applied (credited to factory
OH control or to factory OH applied) during the period was less than the actual factory overhead
incurred (debited to factory OH control) during the period. If two accounts are being used, the net
of the two accounts will be a debit balance.

• If the balance is a credit balance (negative balance), it means factory overhead was over-
applied to the production. In other words, the amount of factory overhead that was applied (cred-
ited to factory OH control or to factory OH applied) during the period was greater than the actual
factory overhead costs incurred (debited to the account) during the period. If two accounts are
being used, the net of the two accounts will be a credit balance.

The amount of over- or under-applied factory overhead is calculated as follows:

Actual costs incurred


− Factory overhead applied during the period
= Under- (over-) applied factory overhead

Whether factory overhead has been under-applied or over-applied, the company needs to correct the im-
balance, because it is not reasonable to have any balance in the factory OH control account (or the net of
the two accounts) at the end of the period. Therefore, the remaining balance (the amount over- or under-
applied) in the factory overhead control account must be removed from the account(s) as part of the period-
end closing entries.

The way this remaining over- or under-applied overhead is closed out depends on whether the balance is
immaterial or material.

Immaterial Amount
If an under-applied amount is immaterial it can simply be debited to COGS in that period. Cost of goods
sold is debited and the factory overhead control account is credited for the amount of the debit balance in
the factory overhead control account to reduce the balance in the factory overhead control account to zero.
COGS for the period is increased, and operating income for the period is decreased.

If an over-applied amount is immaterial, it can be credited to COGS, reducing COGS and increasing the
operating income for the period. The factory overhead control account is debited for the amount of the
credit balance in the account to reduce it to zero and cost of goods sold is credited for the same amount.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 89
D.3. Overhead Costs CMA Part 1
efham CMA
Material Amount
If the amount of overhead that was over- or under-applied is material, it must be distributed among the
WIP inventory, finished goods inventory, and cost of goods sold accounts. The variances are pro-
rated according to the amount of overhead included in each that was allocated to the current period’s
production, not according to the ending balances in each.

Note: The pro-ration of under- or over-applied overhead should be done on the basis of the
overhead allocated to production during the current period only, not on the basis of the balances
in the inventories and cost of goods sold accounts at year end. Information on the amount of overhead
allocated to production during the current year should be available in the accounting system.

An under-applied amount will be debited to the inventories and cost of goods sold accounts while an over-
applied amount will be credited to those accounts. The opposing credit or debit to the factory overhead
control account will bring the balance in that account to zero as of the end of the period.

Whenever the variances (under-applied or over-applied amounts) are allocated proportionately among in-
ventories and cost of goods sold on the basis of the costs applied during the period, the cost per unit for
those costs will be the same as if the actual costs per unit instead of the budgeted costs per unit had been
allocated to production during the year.

Accumulating Overhead Variances


Instead of immediately charging over- or under-applied overhead amounts to COGS or to inventories and
COGS each month, companies may accumulate the over- or under-applied overhead (also called the vari-
ance) in variance accounts throughout the year by closing out the over- or under-applied monthly amount
to the variance accounts.

An advantage of waiting until year-end to resolve the variances is that some of the variances may reverse
by the end of the year. That is particularly true if, based on year-to-date variances, the standard overhead
application rates are adjusted mid-year.

Note: In questions about overhead allocation on the exam, candidates should make certain to identify
the allocation base. It may be machine hours, direct labor hours, direct labor costs, weight, size, or
something similar. It is also possible that the company in a question will use more than one overhead
allocation base for the same cost pool. If the company uses more than one overhead allocation base,
the mathematics of overhead allocation will need to be performed several times and then summed to
find the total to be allocated to production. Doing this does not make the question any more difficult; it
just takes longer to do since the same operation needs to be performed more than once.

90 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

Question 31: On January 1, 20X5, the first year of operations, Medina Co. had the following annual
budget.

Units produced 20,000

Sales $120,000
Minus:
Total variable expenses 70,000
Total fixed expenses 25,000
Operating income $ 25,000
Factory overhead:
Variable $40,000
Fixed 20,000

At the end of the first year, no units were in progress and the actual total factory overhead incurred
was $45,000. There was also $3,000 of over-applied factory overhead. Factory overhead was allocated
on a basis of budgeted units of production. How many units did Medina produce this year?

a) 14,000

b) 16,000

c) 20,000

d) 23,333

(HOCK)

Activity-Based Costing
Activity-based costing (ABC) is another way of allocating overhead costs to products, and in ABC the method
of allocation is based on cost drivers.14 As with the other overhead allocation methods, ABC is a mathe-
matical process. It requires identification of the costs to be allocated, followed by some manner of allocating
them to departments, processes, products, or other cost objects. ABC can be used in a variety of situations
and can be applied to both manufacturing and nonmanufacturing overheads. It can also be used in service
businesses.

The Institute of Management Accountants defines activity-based costing as

“a methodology that measures the cost and performance of activities, resources, and cost
objects based on their use. ABC recognizes the causal relationships of cost drivers to
activities.”15

Activity-based costing is a costing system that focuses on individual activities as the fundamental cost
objects to determine what the activities cost.

• An activity is an event, task or unit of work with a specified purpose. Examples of activities are
designing products, setting up machines, operating machines, making orders or distributing prod-
ucts.

• A cost object is anything for which costs are accumulated for managerial purposes. Examples of
cost objects are a specific job, a product line, a market, or certain customers.

14
Direct costs such as direct materials and direct labor can easily be traced to products, so activity-based costing focuses
on allocating indirect, or overhead, costs to departments, processes, products, or other cost objects.
15
Institute of Management Accountants, Implementing Activity-Based Management: Avoiding the Pitfalls (Montvale, NJ:
Statements on Management Accounting, Strategic Cost Management, 1998), 2.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 91
D.3. Overhead Costs CMA Part 1

• A cost driver is anything (it can be an activity, an event, or a volume of some substance) that
causes costs to be incurred each time the driver occurs. Cost drivers can be structural or execu-
tional.

o Structural cost drivers relate to the actual structure of the company’s operations and the
complexity of the technologies the company uses. A more complex working environment leads
to higher structural costs. Decisions to offer a new service or to open a new retail location and
if so, what size of store to open are examples of structural cost drivers.

o Organizational cost drivers relate to the organization of activities and the delegation of
authorities. Authorizing employees to make decisions is an organizational cost driver, for ex-
ample granting purchasing authority up to a specific amount to a purchasing agent. Any
purchases above that amount require a manager’s approval.

o Executional, or activity, cost drivers relate to the actual processes performed. The cost of
executing activities is determined by the company’s effective use of staff and processes used.
Examples of executional cost drivers are set-ups, moving, number of parts used in assembling
a product, casting, packaging, or handling. In a service business, a service call is an activity
cost driver.

Traditional Costing versus ABC


Differences between traditional costing and activity-based costing include:

Allocations Based On Different Things


Traditional costing systems allocate costs according to general usage of resources, such as machine hours
or direct labor hours. The resources on which the allocations are based may or may not have a connection
with the costs being allocated.

With ABC, the cost allocations are instead based on activities performed and the costs of those activities.
ABC is much more detailed than traditional costing, because it uses many more cost pools and each cost
pool has its own cost driver.

Costs Attached to Low-Volume Products


The use of activity-based costing can result in greater per-unit costs for products produced in low volume
relative to other products than would be reported under traditional costing.

If a product is produced in low volume, it will require fewer resources such as machine hours or direct labor
hours used as cost drivers in traditional overhead allocation than a product that is produced in high volume.
Therefore, under traditional overhead costing, that low-volume product would be allocated a small amount
of total overhead costs. However, a low-volume product may require just as much time and cost per
production run as a high-volume product.

One example is product setups. Production setup time is the same for a low-volume production run as it is
for a high-volume production run. If the cost of product setups is included in total overhead and is allocated
according to machine hours or direct labor hours as in traditional costing, not much product setup cost will
be allocated to the low-volume product because the volume of products produced and the cost drivers used
to produce them are low relative to other products. However, if the cost of product setups is segregated
from other overhead costs, as in ABC, and is allocated according to how many product setups are done for
each product instead of how many units of each product are produced and the machine hours or direct
labor hours per unit, then a more realistic amount of cost for product setups will be allocated to the low-
volume product. The amount of product setup cost allocated to the low-volume product will probably be
higher than it would be under traditional costing, so cost per unit allocated to the low-volume product will
probably be higher under ABC than under traditional costing.

92 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

Note: ABC is becoming more of a necessity for many companies, since traditional systems may use
direct labor to allocate overheads and direct labor is becoming a smaller part of the overall production
process. Essentially, activity-based costing is very similar to the standard method of overhead
allocation, except for the fact that there are many cost pools, each with a different cost driver, and the
cost drivers should have a direct relationship to the incurrence of costs by an individual product. There-
fore, in the application of ABC, many different overhead allocations are performed. The main difference
is the determination of the allocation bases. In a numerical ABC exam question, candidates should be
prepared to make three or four allocations of different cost pools to the final product. Candidates may
also need to determine the appropriate allocation base for each of the cost pools.

The ABC Process


Setting up an ABC system is more difficult, time-consuming, and costly than setting up a traditional system.
It involves the

1) Identification of activities

2) Identification of cost drivers

3) Identification of cost pools

1) Identification of Activities
The company must first analyze the production process and identify the various activities that help to
explain why the company incurs the costs it considers to be indirect costs.

As part of the process of analyzing the production process, the company may identify some non-value-
adding activities. Non-value-adding activities are activities that do not add any value for the end con-
sumer, and the company should try to reduce or eliminate them. Reducing non-value-adding costs provides
an additional benefit to the company (in addition to the more accurate costing of products that ABC pro-
vides) because it can lead to a reduction in the cost of production. In turn, the cost reduction can enable
the company to either reduce the sales price or to recognize more gross profit from the sale of each unit.

Value-adding activities are the opposite of non-value-adding activities. As the name suggests, value-
adding activities are activities (and their costs) that add value to the customer. Value-adding activities add
something to the product that customers are willing to pay for. Even though these activities are value-
adding activities, they must be monitored to make certain that their costs are not excessive.

Note: This identification of non-value adding activities is potentially one of the greatest benefits a com-
pany can receive from implementing an ABC system. Some companies decide to analyze their costs as
value adding o

r non-value adding activities even if they choose to not fully implement ABC.

2) Identification of Cost Drivers and Cost Pools


The company must evaluate all of the tasks performed and select the activities that will form the basis of
its ABC system. Each activity or cost driver identified requires the company to keep records related to that
activity and increases the complexity of the ABC system. Therefore, an activity-based costing system that
includes many different activities can be overly detailed and difficult to use. On the other hand, an activity-
based costing system with too few activities may not be able to measure cause-and-effect relationships
adequately between the cost drivers and the indirect costs.

Activities should be selected that account for a majority of indirect costs, and management should combine
activities that have the same cause-and-effect relationship with the same cost driver or drivers into a single
cost pool. Cost pools should contain costs that are homogeneous. In other words, all of the costs in a cost
pool should have the same or a similar cause-and-effect or benefits-received relationship to a cost driver
or drivers. Identification of the cost pools and identification of the cost driver or drivers for each cost pool
should therefore occur simultaneously.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 93
D.3. Overhead Costs CMA Part 1

A cost pool is similar to the traditional factory overhead control account, except the company maintains
many of them, each with its own cost driver or drivers. The cost pools are used to collect the costs associ-
ated with the various activities (drivers) that incur the costs, and then the costs are allocated as the drivers
are used or consumed in the production of the product.

For example, management may decide to combine equipment maintenance, machine operations, and pro-
cess control into a single cost pool because it determines that machine hours are the most appropriate cost
driver for all of those costs and have the same cause-and-effect relationship with them.

Note: On the CMA Exam, candidates will probably not need to identify the cost drivers or cost pools in
a large question. The cost drivers and cost pools will likely be provided in the question and candidates
will simply need to use the provided information to determine the amount of overhead to be allocated
to any product. In a smaller question, however, candidates may need to determine the appropriate driver
for each cost pool from among those given. Usually, though, the cost drivers and cost pools are fairly
obvious, and candidates can determine what cost driver should be used for what cost pool by simply
thinking about what would cause costs to be incurred.

3) Calculation of the Allocation Rate and Allocation


After the activities have been identified and the cost pools and cost drivers determined, an allocation rate
is calculated for each of the cost allocation bases. The identified cost drivers should be used as the
allocation bases for each cost pool.

The calculation of the allocation rate is done the same way as it is done under the traditional method:
expected costs in each cost pool are divided by the expected usage of the cost allocation base (cost driver)
for the cost pool. The costs in the cost pools are then allocated to the products based upon the usage of
the cost allocation base/driver for each product. This process of cost allocation under ABC is similar to the
accumulation of actual manufacturing overhead costs and the allocation of the costs done under the tradi-
tional method of overhead allocation. The difference is that ABC involves the use of many cost pools and
cost drivers, so the allocation calculations need to be performed many times.

Because of the increased number of allocation bases, ABC provides a more accurate costing of products
when the company produces multiple products. If a company produces only one product, it does not need
to use ABC because all of the costs of production are costs of that one product. ABC is meaningful only if a
company produces more than one product, because ABC affects how much overhead is allocated to each
product.

Note: ABC uses multiple cost pools and multiple allocation bases to more accurately reflect the different
consumptions of overhead activities between and among products.

Categories of Activities
Four different types of activities are used in ABC, based upon where the activity occurs in relation to the
final product and the facility as a whole. The four categories are:

1) Unit-level activities are performed for each unit produced. Examples of unit-level activities are
hours of work, inspecting each item, operating a machine and performing a specific assembly task.

2) Batch-level activities occur each time a batch is produced. Examples of batch-level activities
are machine setups, purchasing, scheduling, materials handling and batch inspection.

3) Product-sustaining activities are activities performed in order to support the production of


specific products. Examples of product-sustaining activities include product design, engineering
changes, maintaining product specifications, and developing a special testing routine for a partic-
ular product.

4) Facility-sustaining activities are activities undertaken to support production in general, such


as security, maintenance, and plant management. Facility-sustaining activities and their associated

94 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

costs present an issue that requires a company decision. Given the broad nature of facility-sus-
taining costs, it is difficult to allocate them in any reasonable manner to the final goods. For internal
decision-making purposes, a company may either try to allocate these costs or may simply expense
them in the period incurred. However, for external financial reporting and tax reporting purposes,
the costs are product costs and must be allocated to products in order to maintain absorption
costing in accordance with U.S. GAAP and tax regulations.

Note: The four different categories of activities may be needed more for answering word questions than
for numerical questions. The cost allocations are done the same way for all the different types of activi-
ties.

Benefits of Activity-Based Costing

• ABC provides a more accurate product cost for use in pricing and strategic decisions because overhead
rates can be determined more precisely and overhead application occurs based on specific actions.

• By identifying the activities that cause costs to be incurred, ABC enables management to identify
activities that do not add value to the final product.

Limitations of Activity-Based Costing

• Not everything can be allocated strictly on a cost driver basis. In particular, facility-sustaining costs
cannot be allocated strictly on a cost driver basis.

• ABC is expensive and time consuming to implement and maintain.

• Activity-based costing systems cannot take the place of traditional costing systems without significant
adaptation because standard ABC is not acceptable for U.S. GAAP or U.S. tax reporting.

ABC will provide the most benefits to companies that produce very diverse products or have com-
plex activities. For companies that produce relatively similar products or have fairly straightforward
processes that are consumed equally by all products, the costs of ABC would probably outweigh the bene-
fits.

Question 32: Cost drivers are:

a) Accounting techniques used to control costs.

b) Accounting measurements used to evaluate whether performance is proceeding according to plan.

c) A mechanical basis, such as machine hours, computer time, size of equipment or square footage
of factory used to assign costs to activities.

d) Activities that cause costs to increase as the activity increases.

(CMA Adapted)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 95
D.3. Overhead Costs CMA Part 1

Below are two examples of how activity-based costing might be used in a real situation.

Example #1: A manufacturing firm. In a manufacturing environment, machine setup is required


every time the production line changes from producing one product to producing another product.

The cost driver is machine setups. An engineer might be required to supervise the setup of the machine
for the product change. The engineer spends 20% of his or her time supervising setups (the resource
driver). Therefore, 20% of the engineer's salary and other expenses will be costs to be allocated ac-
cording to the amount of time the engineer spends supervising each product's machine setup as a
percentage of the amount of time spent supervising all product setups (the resource driver).

In addition to the engineer, other personnel are required for machine setups. Production supervisors are
also needed to supervise machine setups, and they spend 40% of their time doing that (the resource
driver again).

All of the costs of machine setups are collected in a "machine setups" cost pool. Setup time spent on
each product as a percentage of setup time spent on all products is the activity driver. The total costs
in the pool are allocated to the different products being produced based on the percentage of total setup
time used for each product.

Example #2: A service firm. Bank tellers process all kinds of transactions. The transactions relate to
the many different banking services the bank offers. Transactions processed are the cost drivers. How
should the tellers' time, the teller machines used by the tellers, the supplies used by the tellers and the
space occupied by the tellers be allocated among the various services offered by the bank (checking
accounts, savings accounts, and so forth) in order to determine which services are most profitable?

The bank does time and motion studies to determine the average time it takes tellers to process each
type of transaction (checking account transactions, savings account transactions, and so forth). Then,
information on how many of each type of transactions are processed by tellers is captured. The average
time for each type of transaction is multiplied by the number of transactions processed. The percentage
of teller time spent on each type of transaction as a percentage of the amount of teller time spent on all
types of transactions is the activity driver.

The tellers spend 90% of their time performing teller transactions and 10% of their time doing something
else like answering telephones, so that 90% is the resource driver for the tellers' salaries. 90% of the
tellers' salaries will be put into the "tellers" cost pool along with 100% of the costs of their teller ma-
chines, their supplies, and the square footage occupied by their teller stations. Then the percentage of
tellers' time spent on each type of transaction in relation to their total time spent on all teller transactions
(the activity driver) is used to allocate the teller costs in the cost pool proportionately among the bank's
various services.

96 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

The following information is for the next two questions: Zeta Company is preparing its annual
profit plan. As part of its analysis of the profitability of individual products, the controller estimates the
amount of manufacturing overhead that should be allocated to the individual product lines from the
information given as follows:

Wall Mirrors Specialty Windows


Units produced 25 25
Material moves per product line 5 15
Direct labor hours per unit 200 200

Budgeted materials handling costs - $50,000 in total

Question 33: Under a costing system that allocates manufacturing overhead on the basis of direct labor
hours, the materials handling costs allocated to one unit of wall mirrors would be:

a) $1,000

b) $500

c) $2,000

d) $5,000

Question 34: Under activity-based costing (ABC), the materials handling costs allocated to one unit of
wall mirrors would be:

a) $1,000

b) $500

c) $1,500

d) $2,500

(CMA Adapted)

Note: The above question is a very good example of a question that requires determining a cost under
both ABC and the traditional method. Candidates should make sure to be able to do that.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 97
D.3. Overhead Costs CMA Part 1

Shared Services Cost Allocation


Shared services are administrative services provided by a central department to the company’s operating
units. Shared services are usually services such as human resources, information technology, maintenance,
legal, and many accounting services such as payroll processing, invoicing and accounts payable. Usage of
the services by the individual departments (cost objects) can be traced in a meaningful way based on a
cost driver that fairly represents each one’s usage of the service.

These shared service, or support, departments incur costs (salaries, rent, utilities, and so on). For internal
decision-making, the costs of shared service departments need to be allocated to the operating departments
that use their services in order to calculate the full cost of operations or production. The methods of allo-
cating shared services costs are different from the methods of allocating manufacturing overhead costs
covered previously. Shared services costs are allocated for each shared service department as a single cost
pool containing all of the service department’s costs, and the costs are allocated to user departments on
the basis of a single cost driver, such as hours of service used.

Note: The cost allocation methods that follow are all stand-alone cost allocation methods.

Reasons for allocating shared services costs include:

• Cost allocation provides accurate departmental and product costs for use in making decisions,
valuing inventory, and evaluating the efficiency of departments and the profitability of individual
products.

• It motivates managers and other employees to make their best efforts in their own areas of re-
sponsibility to achieve the company’s strategic objectives by reminding them that earnings must
be adequate to cover indirect costs.

• It provides an incentive for managers to make decisions that are consistent with the goals of top
management.

• It fixes accountability and provides a fair evaluation of the performance of segments and segment
managers.

• It can be used to create competition among segments of the organization and their managers to
stimulate improved performance.

• It justifies costs, such as transfer prices.

• It can be used to compute reimbursement when a contract provides for cost reimbursement.

Cost allocations may be performed for just one shared service or support department whose costs are
allocated to multiple user departments, or they may be performed for multiple shared service or support
departments whose costs are being allocated both to other service and support departments and to oper-
ating departments. But even when costs of shared service departments are allocated to other service
departments, ultimately all the shared service costs will be allocated fully to operating departments.

98 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

Allocating Costs of a Single (One) Service or Support Department to Multiple Users


Before allocating the costs of a shared service department to operating departments, management must
decide whether the shared service department’s fixed costs and its variable costs should be allocated as
one amount or whether the fixed and variable costs should be allocated separately, which would enable the
fixed costs to be allocated in a different manner from the variable costs. Allocating all costs as one amount
is called the single-rate method. Allocating the fixed and variable costs as two separate cost pools is called
the dual-rate method.

• Single-Rate Method – The single-rate method does not separate fixed costs of service depart-
ments from their variable costs. All of the service department costs are put into one cost pool and
the costs are allocated using one allocation base.

• Dual-Rate Method – The dual-rate method breaks the cost of each service department into two
pools, a variable-cost pool and a fixed-cost pool. Each cost pool is allocated using a different cost-
allocation base.

Allocation bases for either the single-rate method or the dual-rate method can be:

• Budgeted rate and budgeted hours (or other cost driver) to be used by the operating divisions.

• Budgeted rate and actual hours (or other cost driver) used by the operating divisions.

Following are examples of the single-rate method and the dual-rate method, using the same data for both.

Example #1: Single-rate method of allocation:

The following information is from the 20X0 budget for EZYBREEZY Co. EZYBREEZY has one service de-
partment, its Maintenance Department, that serves its Manufacturing and Sales departments. The
Maintenance Department has a practical capacity of 5,000 hours of maintenance service available each
year. The fixed costs of operating the Maintenance Department (physical facilities, tools, and so forth)
are budgeted at $247,500 per year. The wages for the maintenance employees and the supplies they
require are the variable costs, and those are budgeted at $25 per hour.

Budgeted Maintenance usage by the Manufacturing and Sales departments is as follows:


Manufacturing 3,500 hours
Sales 1,000 hours
Total budgeted usage 4,500 hours

Using the single-rate method, the budgeted total cost pool will be:

$247,500 fixed cost + (4,500 hours × $25) variable cost = $360,000

The allocation rate for the total maintenance cost is $360,000 ÷ 4,500 hours, which is $80 per hour.

The single-rate method is usually used with the second allocation base option: budgeted rate and actual
hours used. The actual maintenance hours used by the Manufacturing and Sales departments are as
follows:
Manufacturing 3,600 hours
Sales 1,100 hours
Total actual usage 4,700 hours

(Continued)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 99
D.3. Overhead Costs CMA Part 1

Therefore, the amounts of Maintenance department costs that will be allocated to the Manufacturing and
Sales departments are:
Manufacturing 3,600 × $80 $288,000
Sales 1,100 × $80 88,000
Total cost allocated $376,000

The problem with the single-rate method is that, from the perspective of the user departments, the
amount they are charged is a variable cost, even though it includes costs that are fixed. The manager
of the Manufacturing department could be tempted to cut his department’s costs and outsource the
maintenance function if he could find a company to supply maintenance for less than $80 per hour.

Assume that an outside maintenance company offers to do the maintenance for the Manufacturing de-
partment for $60 per hour. The amount paid to the outside company will be 3,600 hours @ $60 per
hour, or $216,000. The Manufacturing department manager is happy, because he has saved $72,000
($288,000 − $216,000) for his department.

However, the fixed costs of the in-house Maintenance department do not go away just because Manu-
facturing is not using its services any longer. The total cost of the Maintenance department (now being
used only by the Sales department) will be $247,500 fixed cost + (1,100 hours used by the Sales de-
partment × $25) variable cost = $275,000, assuming the actual costs incurred are equal to the budgeted
amounts. This is $85,000 less than was budgeted ($360,000 − $275,000). However, the company as a
whole is paying an additional cost of $216,000 to the outside maintenance company. Thus, for the
company as a whole, total maintenance cost is now $491,000 ($275,000 for the internal maintenance
department + $216,000 for the outside company), which is $131,000 greater than the budgeted amount
of $360,000 for maintenance. The Manufacturing department’s actions have caused a variance in costs
for the company as a whole of $131,000 over the budgeted amount.

Example #2: Dual-rate method of allocation:

The following information is from the 20X0 budget for EZYBREEZY Co. EZYBREEZY has one service de-
partment, its Maintenance Department, that serves its Manufacturing and Sales departments. The
Maintenance Department has a practical capacity of 5,000 hours of maintenance service available each
year. The fixed costs of operating the Maintenance Department (physical facilities, tools, and so forth)
are budgeted at $247,500 per year. The wages for the maintenance employees and the supplies they
require are the variable costs, and those are budgeted at $25 per hour.

Budgeted Maintenance usage by the Manufacturing and Sales departments is as follows:


Manufacturing 3,500 hours
Sales 1,000 hours
Total budgeted usage 4,500 hours

Because a dual-rate method is being used, EZYBREEZY selects an allocation base for the variable costs
and an allocation base for the fixed costs. The company allocates the variable costs based on the budg-
eted variable cost per hour ($25) and the actual hours used. It allocates fixed costs based on budgeted
fixed costs per hour and the budgeted number of hours for each department (option #1).

(Continued)

100 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

The actual maintenance hours used by the Manufacturing and Sales departments are as follows:
Manufacturing 3,600 hours
Sales 1,100 hours
Total actual usage 4,700 hours

The allocation rate for the fixed cost is $247,500 ÷ 4,500 hours, or $55 per hour. The allocation rate
for the variable cost is $25 per hour.

The amounts allocated to each user department are now:

Manufacturing:
Fixed Costs: $55 × 3,500 hours $ 192,500
Variable Costs: $25 × 3,600 hours 90,000
Total allocated to Manufacturing $ 282,500

Sales:
Fixed Costs $55 × 1,000 hours $ 55,000
Variable Costs: $25 × 1,100 hours 27,500
Total allocated to Sales $ 82,500

Total cost allocated $ 365,000

Under the dual-rate method, the total costs allocated to each user department are different from the
costs allocated under the single-rate method because the fixed costs are allocated based on the budg-
eted usage under the dual-rate method and based on the actual usage under the single-rate method.

Under the dual-rate method, the Manufacturing and Sales departments would each be charged for its
budgeted fixed allocation of costs even if it does not use the internal Maintenance department.
That should discourage the manager of the Manufacturing department from contracting with an outside
maintenance service.

Benefits of the Single-Rate Method

• The cost to implement the single-rate method is low because it avoids the analysis needed to classify
all of the service department’s cost into fixed and variable costs.

Limitations of the Single-Rate Method

• The single-rate method makes fixed costs of the service department appear to be variable costs to
the user departments, possibly leading to outsourcing that hurts the organization as a whole.

Benefits of the Dual-Rate Method

• The dual-rate method helps user department managers see the difference in the ways that fixed costs
and variable costs behave.

• The dual-rate method encourages user department managers to make decisions that are in the best
interest of the organization as a whole, as well as in the best interest of each individual department.

Limitations of the Dual-Rate Method

• The cost is higher than the cost of the single-rate method because of the need to classify all of the
costs of the service department into fixed and variable costs.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 101
D.3. Overhead Costs CMA Part 1

Allocating Costs of Multiple Shared Service Departments


Special cost allocation problems arise when an organization has several shared service departments and
the shared service departments provide support not only to the operating departments but to the other
shared service departments as well. The factor that complicates the process is service departments using
the services of other service departments. For example, people in the maintenance department use the
services of the IT department and eat in the cafeteria.

Three different methods of allocating costs of multiple shared service departments are used when service
departments use the services of other service departments. All three methods are simply mathematical
allocations, much like the way manufacturing overhead is allocated to the products. The different methods
of multiple shared service cost allocation treat these reciprocal services differently.

The three methods of allocation are:

1) The direct method


2) The step (or step-down) method
3) The reciprocal method

Note: On an exam question, treat the service departments as shared service departments even if they
could be treated in a different manner. Accounting is an example of a shared service that may be treated
differently, for example. In some companies, the costs of accounting may simply be expensed and not
allocated. But if accounting is given as a service department in a question, its costs should be allocated.

Note: The following discussion of multiple service departments uses only the single rate method (fixed
and variable costs allocated together). However, the dual-rate method can also be used to allocate costs
of multiple service departments. To do so, a company would need to do two separate allocations, one of
fixed costs and one of variable costs. There is no reason to expect that an exam question will require a
dual-rate allocation of multiple service departments’ costs, though, so the following discussion is limited
to use of the single rate method for multiple service departments.

1) The Direct Method – IGNORING Services Provided Between Service Departments


Under the direct method the reciprocal services provided by the different shared service departments to
each other are ignored. The company simply allocates all of the shared service departments’ costs directly
to the operating departments. The allocation is made on a basis that is reasonable and equitable to the
operating departments for each service department. For example, the costs of a subsidized employee caf-
eteria should be allocated to the operating departments based on the number of employees, while the
maintenance department’s costs may be allocated based on the number of maintenance hours used by
each operating department.

When calculating the usage ratios for the different operating departments, count only the usage of the
shared service departments by the operating departments. The usage of shared service departments
that takes place in the other service departments is excluded because service departments will not be
allocated any costs from other service departments.

102 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

The direct method is the simplest and most common method. A very short example follows (the calculations
of the allocations are not shown).

Maintenance Cafeteria Production 1 Production 2 Production 3


Departmental
100 120 300 400 800
Costs Incurred

Allocation of
Maintenance (100) 20 30 50
Costs

Allocation of
Cafeteria (120) 30 30 60
Costs
TOTAL COSTS 0 0 350 460 910

2) The Step-Down or Sequential Method – Recognizing SOME Services Provided Between


Service Departments
The step-down method is also called the step method or the sequential method. In the step-down
method the services the shared service departments provide to each other are included, but only one
allocation of the costs of each service department is made. After the costs of a particular service department
have been allocated, that service department will not be allocated any additional costs from other service
departments. The step-down method leads to a stair step-like diagram of cost allocations, as below. Be-
cause each service department’s costs, including costs allocated to it by other service departments, are
allocated in turn, all service department costs ultimately end up allocated to the operating departments.

In order to use the step-down method, an order must be established in which the service department costs
are allocated. The order can be any order management chooses. A popular method is to determine
the order according to the percentage of each department’s services provided to other shared service de-
partments, although that is not the only possible way. If, for example, the costs of the department that
provides the highest percentage of its services to other shared service departments are allocated first, then
the costs of the department that provides the next highest percentage of its services to other shared service
departments comes next, and so forth.

The first shared service department’s costs are allocated to the other shared service departments and the
operating departments. The second shared service department’s costs (which now include its share of the
first shared service department’s costs) are allocated next to the other shared service departments (but
not to the first shared service department that has already been allocated) and the operating departments.
Once a shared service department’s costs have been allocated, no costs will be allocated to it from other
shared service departments.

A problem on the exam will give the allocation order to use if it is not obvious.

Following is an example of the step-down method.

Maintenance Cafeteria Production 1 Production 2 Production 3


Departmental
100 120 300 400 800
Costs Incurred

Allocation of
Maintenance (100) 10 16 28 46
Costs

Allocation of
Cafeteria (130) 36 32 62
Costs

TOTAL COSTS 0 0 352 460 908

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 103
D.3. Overhead Costs CMA Part 1

The example of the step-down method is a bit different from the example of the direct method. Use of the
step-down method results in a slightly different result from the direct method, even though the operating
departments used the same amounts of services of the two shared service departments in both examples.
Again, the allocation calculations are not shown.

Note: In the step-down method, the costs allocated from the cafeteria include its own incurred costs
(120), plus the cafeteria’s share of the maintenance costs that were allocated to it from maintenance
(10). When the maintenance department’s costs are allocated, they are allocated on the basis of the
number of hours that the cafeteria utilized the services of the maintenance department. However, when
the cafeteria costs (including the cafeteria’s share of the maintenance costs) are allocated, none of its
costs are allocated to the maintenance department for its usage of the cafeteria.

3) The Reciprocal Method – Recognizing ALL Services Provided Between Service Depart-
ments
The reciprocal method is the most complicated and advanced of the three methods of shared services cost
allocation because it recognizes all of the services provided by the shared service departments to the other
shared service departments. Because of this detailed allocation between and among the shared service
departments, the reciprocal method is the most theoretically correct method to use. However, a company
will need to balance the additional costs of allocating costs this way against the benefits received. Graph-
ically the reciprocal method looks like the following.

Maintenance Cafeteria Production 1 Production 2 Production 3


Departmental
100 120 300 400 800
Costs Incurred

Allocation of
Maintenance (108) 10 20 29 49
Costs

Allocation of
Cafeteria 8 (130) 33 30 59
Costs

TOTAL COSTS 0 0 353 459 908

Notice that the shared service departments are each allocating some of their costs to the other shared
service department. As a result, the total amount allocated by each shared service department will be
greater than its own overhead costs, since each shared service department must allocate all of its own
costs plus some of the other shared service department’s costs that have been allocated to it. The math to
do this is the most complicated of the three methods, but even it is not too difficult.

To solve a problem using the reciprocal method, “simultaneous” or multiple equations are used. With two
shared service departments, the multiple equations are set up as follows:

Maintenance Costs (M) to Allocate = M’s Costs Incurred + M’s % of C’s Costs
Cafeteria Costs (C) to Allocate = C’s Costs Incurred + C’s % of M’s Costs

The first step is to solve for either “Maintenance Costs to Allocate” or “Cafeteria Costs to Allocate,” and
after that solve for the other number. These calculated amounts become the amounts that need to be
allocated from the maintenance department and cafeteria to all the other departments, including the other
service departments.

The process of solving the two equations is shown in detail in the example on the following pages.

104 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

Comprehensive Example of the Three Methods of Shared Service Cost Allocation

Example: The following information about Cubs Co. is used to demonstrate the three different methods
of allocating shared service costs. Cubs Co. has two shared service departments (A and B) and three
operating departments (X, Y and Z). Shared Service Department A allocates its overhead based on direct
labor hours and Shared Service Department B allocates its overhead based on machine hours. The fol-
lowing information is in respect to the service and operating departments:
Dept. A Dept. B Dept. X Dept. Y Dept. Z Total
Overhead $100,000 $50,000 $200,000 $300,000 $250,000 $900,000
Labor Hours -- 1,000 2,000 4,000 3,000 10,000
Machine Hours 2,000 -- 2,000 2,000 2,000 8,000

Direct Method
Under the direct method each shared service department allocates costs to only the operating depart-
ments. The operating departments used a total of 9,000 (2,000 + 4,000 + 3,000) labor hours of service
from Department A, so each department is allocated $11.11111 ($100,000 in costs ÷ 9,000 total service
hours provided) for each direct labor hour used. Department B provided 6,000 (2,000 + 2,000 + 2,000)
machine hours of service to the operating departments, so it allocates $8.33333 ($50,000 in costs ÷
6,000 machine hours provided) per machine hour to the operating departments. The services provided
by each service department to the other service department are ignored in the direct method. The
numbers that are ignored are shaded below.

Dept. A Dept. B Dept. X Dept. Y Dept. Z Total


Overhead $100,000 $50,000 $200,000 $300,000 $250,000 $900,000
Labor Hours -- 1,000 2,000 4,000 3,000 10,000
Machine Hours 2,000 -- 2,000 2,000 2,000 8,000

The next step is to determine how much of the costs of the shared service departments will be allocated
to each operating department. The costs of Department A are allocated first, on the basis of direct labor
hours. Since the 1,000 labor hours provided to Department B by Department A are ignored, only 9,000
total direct labor hours are used in the allocation. Of the total 9,000 direct labor hours used in the
allocation, Department X used 2,000 labor hours, Department Y used 4,000 labor hours, and Department
Z used 3,000 labor hours. Therefore, Department X is allocated $11.11111 × 2,000 hours = $22,222.22
of costs from Department A. Department Y is allocated $11.11111 × 4,000 hours = $44,444.45 (ad-
justed for rounding difference), and Department Z is allocated $11.11111 × 3,000 hours = $33,333.33.

Department B’s costs are allocated next on the basis of machine hours. Since 2,000 machine hours of
service were provided by Department B to Department A and those are ignored, 6,000 machine hours
in total are used in the allocation. Of the total 6,000 machine hours used in the allocation, Departments
X, Y and Z each used 2,000 machine hours. Therefore, each operating department is allocated $8.33333
per hour × 2,000 hours = $16,666.66. (Two of the amounts will be adjusted for rounding differences.)

Below are the total overhead costs for each department after the allocation:

Dept. A Dept. B Dept. X Dept. Y Dept. Z


Own Overhead $100,000.00 $50,000.00 $200,000.00 $300,000.00 $250,000.00
Allocated from A (100,000.00) 0.00 22,222.22 44,444.45 33,333.33
Allocated from B 0.00 (50,000.00) 16,666.67 16,666.66 16,666.67
Total OH $ 0.00 $ 0.00 $238,888.89 $361,111.11 $300,000.00

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 105
D.3. Overhead Costs CMA Part 1

Step-Down Method

The first requirement for the step-down method is to determine which shared service department’s costs
will be allocated first. 25% of the service provided by Department B was provided to Department A, but
only 10% of Department A’s service provided was provided to Department B. Therefore, Department B’s
costs are allocated first. Department B’s cost of $50,000 is allocated on the basis of machine hours used
by all the other departments, including Department A. In total 8,000 machine hours were used, so
Department B’s cost per machine hour is $6.25 ($50,000 ÷ 8,000).

Department B’s costs are allocated among all the other departments, including Department A, based on
machine hours used by each. Department A is allocated $6.25 × 2,000 hours, or $12,500. This $12,500
is added to the $100,000 of A’s own overhead, for a total $112,500 cost for Department A to allocate.
This $112,500 cost for Department A is allocated only to the operating departments based on the direct
labor hours used by each. The allocation of Department A’s costs is done the same way as in the direct
method, except that the amount of cost allocated by Department A is greater than it was in the direct
method because some costs allocated from Department B are included in Department A’s costs to allo-
cate.

Department B’s costs are allocated first. As mentioned previously, Department A is allocated $12,500 of
Department B’s costs. Since Departments X, Y and Z each also used 2,000 machine hours, each of those
departments is also allocated $12,500 from Department B ($6.25 × 2,000 machine hours).

Next, the costs of Department A are allocated to the operating departments only. The total costs to be
allocated for Department A are Department A’s own overhead of $100,000 plus the $12,500 allocated
overhead from Department B, for a total to allocate of $112,500. That $112,500 is allocated based on
direct labor hours, excluding the direct labor hours used by Department B. Therefore, a total of 9,000
direct labor hours (2,000 + 4,000 + 3,000) are used in allocating Department A’s total overhead of
$112,500. $112,500 ÷ 9,000 = $12.50 to be allocated per direct labor hour. Department X, which used
2,000 machine hours, is allocated $12.50 × 2,000 = $25,000. Department Y, which used 4,000 machine
hours, is allocated $12.50 × 4,000 = $50,000, and Department Z, which used 3,000 machine hours, is
allocated $12.50 × 3,000 = $37,500.

It is important to remember in the step-down method that the hours the second shared service depart-
ment provides to the first are ignored when allocating the second shared service department’s costs.
The ignored numbers are shaded below.
Dept. A Dept. B Dept. X Dept. Y Dept. Z Total
Overhead $100,000 $50,000 $200,000 $300,000 $250,000 $900,000
Labor Hours -- 1,000 2,000 4,000 3,000 10,000
Machine Hours 2,000 -- 2,000 2,000 2,000 8,000

Below are the total overhead costs for each department after the allocation:
Dept. A Dept. B Dept. X Dept. Y Dept. Z
Own Overhead $100,000.00 $50,000.00 $200,000.00 $300,000.00 $250,000.00
Allocated from A (112,500.00) 0.00 25,000.00 50,000.00 37,500.00
Allocated from B 12,500.00 (50,000.00) 12,500.00 12,500.00 12,500.00
Total OH $ 0.00 $ 0.00 $237,500.00 $362,500.00 $300,000.00

106 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.3. Overhead Costs

Reciprocal Method

Under the reciprocal method some of Department A’s costs are allocated to Department B and some of
Department B’s costs are allocated to Department A using simultaneous equations. Since Cubs Co. has
two shared service departments, two equations will be needed. (If the company had three shared service
departments, three equations would be needed, and so forth.) The equations express how much each
shared service department needs to allocate to all the other departments, including the other shared
service departments.

Department A used 25% of Department B’s services (2,000 MH used ÷ 8,000 total MH). Therefore, A is
allocated 25% of B’s costs. A’s costs to allocate equal the $100,000 in costs incurred by A plus 25% of
the costs incurred by B. Department B used 10% of Department A’s services (1,000 DLH used ÷ 10,000
total DLH). Therefore, B is allocated 10% of A’s costs. B’s costs to allocate equal the $50,000 in costs
incurred by B plus 10% of the costs incurred by A. The equations needed are:
A = $100,000 + 0.25B
B = $50,000 + 0.10A

Solve for A by substituting the right side of the second equation (B) into the first equation:

A = $100,000 + 0.25 ($50,000 + 0.10A)

The resulting equation has only one variable (A) in the equation and can be solved for Dept. A, as follows.

A = $100,000 + $12,500 + 0.025A


0.975A = $112,500
A = $115,384.62

The result, $115,384.62, is the total cost for Department A to be allocated to all the other departments,
including B, the other shared service department.

The next step is to put the value found for A into the second equation in place of the variable A:
B = $50,000 + 0.1 ($115,384.62)
B = $61,538.46

The total cost for Department B to be allocated to all the other departments, including A, the other
shared service department, is $61,538.46.

The total overhead cost for Department A, $115,384.61, is allocated among all the other departments
on the basis of the full 10,000 total direct labor hours used, or $11.53846 per hour. The cost of Depart-
ment A allocated to Department B is $11.53846 × 1,000 = $11,538.46. Department X is allocated
$11.52846 × 2,000 = $23,076.92 of Department A’s costs, Department Y is allocated $11.53846 ×
4,000 = $46,153.85, and Department Z is allocated $11.53846 × 3,000 = $34,615.38.

The total overhead cost for Department B, $61,538.46, is allocated among all the other departments on
the basis of the full 8,000 total machine hours used, or $7.69231 per machine hour. Departments A, X,
Y and Z each used 2,000 machine hours, so each will be allocated $7.69231 × 2,000 = $15,384.62.
(Two of the departments’ amounts will be adjusted for rounding differences.)

Below are the total overhead costs for each department after the allocations:
Dept. A Dept. B Dept. X Dept. Y Dept. Z
Own Overhead $100,000.00 $50,000.00 $200,000.00 $300,000.00 $250,000.00
Allocated from A (115,384.61) 11,538.46 23,076.92 46,153.85 34,615.38
Allocated from B 15,384.61 (61,538.46) 15,384.61 15,384.62 15,384.62
Total OH $ 0.00 $ 0.00 $238,461.53 $361,538.47 $300,000.00

Notice that in all of the methods of overhead allocation, the total of the overhead amounts allocated to
the operating departments is equal to the total amount of overhead that the company as a whole—
including the shared service departments—incurred during the period ($900,000).

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 107
D.3. Overhead Costs CMA Part 1

The following information is for the next five questions: The managers of Rochester Manufactur-
ing are discussing ways to allocate the cost of support departments such as Quality Control and
Maintenance to the production departments. To aid them, they were provided the following information:
Quality Control Maintenance Machining Assembly Total
Budgeted overhead costs
before allocation $350,000 $200,000 $400,000 $300,000 $1,250,000
Budgeted machine hours -- -- 50,000 -- 50,000
Budgeted direct labor hours -- -- -- 25,000 25,000
Budgeted hours of service:
Quality Control -- 7,000 21,000 7,000 35,000
Maintenance 10,000 -- 18,000 12,000 40,000
Question 35: If Rochester uses the direct method of allocating support department costs, the total
support costs allocated to the Assembly Department would be:
a) $80,000
b) $87,500
c) $120,000
d) $167,500
Question 36: If Rochester uses the direct method, the total amount of overhead allocated to each
machine hour at Rochester would be:
a) $2.40
b) $5.25
c) $8.00
d) $15.65
Question 37: If Rochester uses the step-down method of allocating support costs beginning with quality
control, the maintenance costs allocated to the Assembly Department would be:
a) $70,000
b) $108,000
c) $162,000
d) $200,000
Question 38: If Rochester uses the reciprocal method of allocating support costs, the total amount of
quality control costs to be allocated to the other departments would be:
a) $284,211
b) $336,842
c) $350,000
d) $421,053
Question 39: If Rochester decides not to allocate support costs to the production departments, the
overhead allocated to each direct labor hour in the Assembly Department would be:
a) $3.20
b) $3.50
c) $12.00
d) $16.00
(CMA Adapted)

108 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Estimating Fixed and Variable Costs

Question 40: A segment of an organization is referred to as a service center if it has:

a) Responsibility for combining the raw materials, direct labor, and other factors of production into
a final output.

b) Responsibility for developing markets and selling the output of the organization.

c) Authority to make decisions affecting the major determinants of profit, including the power to
choose its markets and sources of supply.

d) Authority to provide specialized support to other units within the organization.

(CMA Adapted)

Estimating Fixed and Variable Costs


Sometimes costs are mixed costs or the fixed costs are not segregated from the variable costs in the
historical information available, but the fixed costs need to be separated from the variable costs for analysis,
budgeting or costing purposes. The High-Low Points Method can be used to separate fixed costs from
variable costs.

High-Low Points Method


For the High-Low Points Method, the highest and lowest observed activity levels of the cost driver
within the relevant range and the costs associated with those values are used. The activity level of the
cost driver may be production level in units, direct labor hours, machine hours, or any other cost driver.

For example, to segregate fixed production costs from variable production costs when only a single total
cost amount is available, select the month of the highest production and the month of the lowest production.
Compare the difference in production levels with the difference in total costs for the two months to deter-
mine approximately the amount of costs that are variable and the amount of costs that are fixed. The steps
are as follows:

1) Estimate the variable portion of the total cost using the highest and lowest activity level values of
the cost driver. Divide the difference between the costs associated with the highest and the lowest
activity levels by the difference between the highest and lowest activity levels.

Difference in Associated Costs = Variable Cost per Unit


Difference in Activity Levels of the Cost Driver

Since fixed costs do not change with changes in activity, the difference between the costs for the
highest activity level and the costs for the lowest activity level divided by the difference in the two
activity levels is the estimated variable cost per unit of the cost driver.

2) Multiply the variable cost per unit of the cost driver by the activity level at either the highest or
the lowest activity level to calculate the total variable cost at that activity level.

3) Subtract the total variable cost at that activity level from the total cost at that activity level to
calculate the fixed cost.

Note: The “associated costs” used must always be the costs associated with the highest and lowest
activity levels of the cost driver, even if those are not the highest and lowest historical costs in the
data set.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 109
Estimating Fixed and Variable Costs CMA Part 1

Another way of estimating the variable cost per unit using the High-Low Points method is to set up two
equations in two variables, with one equation representing the highest level of the cost driver and one
equation representing the lowest level. The two variables are Fixed Costs and Variable Costs. Then, subtract
one equation from the other equation to eliminate the Fixed Cost as a variable and solve the remainder for
the Variable Cost.

Note: Both methods will be illustrated in the example that follows.

After the estimated total fixed costs and the estimated variable cost per unit of the cost driver have been
determined, the cost function can be written. Total costs at any activity level can be estimated by
using the cost function.

The cost function describes total costs in terms of total fixed cost plus total variable cost, where the total
variable cost is the variable cost per unit of the cost driver multiplied by the number of units. The cost
function expresses the relationship between fixed and variable costs.

The cost function can be written in either of two ways, as follows:

y = a + bx

Where: y = the dependent variable and the predicted value of y, which here is
total costs
a = the constant coefficient and the y-intercept, or the value of y when
the value of x is zero, which here is the fixed cost
b = the variable coefficient and the slope of the line, which here is the
variable cost per unit of the activity level
x = the independent variable, which here represents the number of units
of the activity level

OR
y = ax + b

Where: y = the dependent variable and the predicted value of y, or total costs
a = the variable coefficient and the slope of the line, or the variable cost
per unit of the activity level
x = the independent variable, or the number of units of the activity level
b = the constant coefficient and the y-intercept, or the fixed cost

Note: The terms may be in any order. The variable representing fixed costs may come first, or it may
come at the end. Whichever way the cost function is written, the constant coefficient (the fixed costs)
will always be by itself and the variable coefficient (the variable cost per unit of the cost driver) will
always be next to the independent variable (the x, or the number of units/hours).

Examples of both methods of calculating the fixed cost and the variable cost per unit of the cost driver
follow.

110 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D Estimating Fixed and Variable Costs

Example: Ray Corporation experienced the following total production costs during the past year:
Ray Corporation
Production Volumes and Costs
Production in Units Total Production Costs
January 6,257,000 $1,500,000
February 4,630,000 1,200,000
March 5,200,000 1,300,000
April 5,443,000 1,350,000
May 5,715,000 1,400,000
June 3,000,000 900,000
July 3,543,000 1,000,000
August 3,815,000 1,050,000
September 5,715,000 1,400,000
October 6,800,000 1,600,000
November 6,529,000 1,550,000
December 5,172,000 1,300,000

What is Ray Corporation’s fixed production cost and what is its variable production cost per unit?

The highest and the lowest production volumes and their associated production costs are:

Production in Units Total Production Costs


October 6,800,000 $1,600,000
June 3,000,000 900,000

Using the first method to estimate the variable cost per unit of the cost driver:

Difference in Associated Costs = $700,000 $0.1842105 variable cost/unit


Difference in Activity Levels 3,800,000

Using the second method (two equations) to estimate the variable cost per unit of the cost driver:

FC + 6,800,000 VC = $1,600,000
− FC + 3,000,000 VC = 900,000
0 + 3,800,000 VC = $ 700,000
VC = $0.1842105

The next step is to put the variable cost per unit into an equation to calculate fixed cost using either the
lowest or the highest activity level and its associated total cost. Using the highest volume, 6,800,000
units and its associated total cost of $1,600,000:

FC = Total Cost − Variable Cost


FC = $1,600,000 − ($0.1842105 × 6,800,000) = $347,369

Fixed cost is $347,369, and variable cost is $0.1842105 per unit.

Therefore, the cost function that expresses the relationship between fixed and variable costs and that
can be used to estimate total costs at any activity level is:

y = 347,369 + 0.1842105x

These fixed and variable cost amounts can be proved by substituting them for the variables in the cost
function using the lowest activity level of 3,000,000 units. The result should be the total historical cost
at that historical activity level, and it is.

$347,369 + ($0.1842105 × 3,000,000) = $900,000

Fixed cost is $347,369 and variable cost is $0.1842105 per unit.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 111
D.4. Supply Chain Management CMA Part 1

D.4. Supply Chain Management


What is Supply Chain Management?
Nearly every product that reaches an end user represents the coordinated efforts of several organizations.
Suppliers provide components to manufacturers, who in turn convert them into finished products that they
ship to distributors for shipping to retailers for purchase by the consumer. All of the organizations involved
in moving a product or service from suppliers to the end-user, the customer, are referred to collectively as
the supply chain.

Supply chain management is the active management of supply chain activities by the members of a supply
chain with the goals of maximizing customer value and achieving a sustainable competitive advantage. The
supply chain firms endeavor to develop and run their supply chains in the most effective and efficient ways
possible. Supply chain activities cover product development, sourcing, production, logistics, and the infor-
mation systems needed to coordinate the supply chain activities.

The organizations that make up the supply chain are linked together through physical flows of products and
through flows of information. Physical flows involve the movement, storage, and transformation of raw
materials and finished goods. The physical flows are the most visible part of the supply chain, but the
information flows are just as important. Information flows allow the various supply chain partners to coor-
dinate their long-term plans and to control the day-to-day flow of goods and material up and down the
supply chain.

By sharing information and by planning and coordinating their activities, all the members of the supply
chain can be in a position to respond quickly to needs without having to maintain large inventories. Retailers
share daily sales information with distributors and the manufacturer, so that the manufacturer knows how
much production to schedule and how much raw material to order and when, and the distributors know
how much to order. The trading partners share inventory information. The sharing of information reduces
uncertainty as to demand.

The result of effective supply chain management is fewer stockouts at the retail level, reduction of excess
manufacturing by the manufacturer and thus reduction of excess finished goods inventories, and fewer rush
and expedited orders. Each company in the supply chain is able to carry lower inventories, thus reducing
the amount of cash tied up in inventories for all.

Some supply chain management goes so far as the retailer allowing the distributor or the distributor allow-
ing the manufacturer to manage its inventories, shipping product to it automatically whenever its inventory
of an item gets low. Such a practice is called supplier-managed or vendor-managed inventory.

Along with the benefits of supply chain management, issues and problems can arise because of communi-
cations problems, trust issues, incompatible information systems, and the required increases in personnel
resources and financial resources.

112 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

Lean Resource Management


“Lean manufacturing,” the earliest form of lean resource management, was a philosophy and system of
manufacturing developed originally by Toyota following World War II. Toyota called the system the Toyota
Production System (TPS).

In 1990, a book titled The Machine That Changed the World by James P. Womack, Daniel T. Jones, and
Daniel Roos was published. The Machine That Changed the World articulated the Toyota principles and
coined the term “lean production” or “lean manufacturing” to describe the principles. The authors docu-
mented the advantages of lean production over the existing mass production model and predicted that the
method would eventually take over in place of the mass production model. That did occur, as manufacturers
throughout the world adapted the system to meet their own manufacturing needs. It has become a global
standard that virtually all manufacturing companies must adopt in order to remain competitive.

However, the authors of The Machine That Changed the World also predicted that the system would ulti-
mately spread beyond production and be applied to every value-creating function such as health care,
retail, and distribution. That is happening today.

Womack and Jones authored another book in 1996 that was updated and expanded in 2003, called Lean
Thinking. Lean thinking is a new way of thinking about the roles of firms and functions in managing the
flow of value creation—called the “value stream”—from concept to the final customer. The emphasis in lean
thinking is on adding value to the customer while cutting out waste, or muda in Japanese.

• “Adding value to the customer” is not the same thing as adding value to the product or service. If
extra options added to a product or service are not things the customer wants or is willing to pay
for, they may well add value to the product, but they do not add value to the customer.

• “Muda” or “waste” is anything that does not add value to the customer or anything the customer
is not willing to pay for.

Identifying and eliminating waste is a primary focus of lean resource management. For example, in manu-
facturing, the primary wasteful activities addressed by lean production include:

1) Over-production, that is, producing more items than can customers want.

2) Delay, or waiting for processing, and parts sitting in storage.

3) Transporting parts and materials from process to process or to various storage locations.

4) Over-processing, or doing more work on a part or product than is necessary.

5) Inventory, as in committing too much money and storage space to unsold parts or products.

6) Motion, or moving parts more than the minimum amount required to complete and ship them.

7) Making defective parts, that is, creating parts or products that are not sellable due to defects
and that must be discarded or reworked.

Waste inhibits throughput, the rate at which work proceeds through a manufacturing system. Machine
downtime, waiting for materials, out of stock supplies, operator errors, and poorly designed processes all
contribute to poor throughput in a manufacturing system. Eliminating waste increases throughput. Even
small reductions in waste can have a cumulative effect in increasing throughput.

In Lean Thinking, Womack and Jones defined “lean thinking” as five principles that incorporate the concepts
of the “lean enterprise” and “lean resource management.”

1) Specify value as the starting point for lean thinking. “Value” is defined by the ultimate customer
in terms of a specific good or service that meets the customer’s needs at a specific price and at a
specific time. The producer must provide the value customers actually want through dialogue with
specific customers. Providing a good or service that customers do not want is wasteful, even if it
is provided in the most efficient manner. Existing assets, technologies, and capabilities to supply

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 113
D.4. Supply Chain Management CMA Part 1

a product or service are irrelevant if the product or service they can supply is not what customers
want.

2) The next step is to identify the value stream. The “value stream” is all the specific actions
needed to provide a good or service. It involves problem-solving—developing a concept and de-
signing and producing the product or service; information management—including order-taking,
scheduling, and delivery; and the physical transformation process—converting raw materials to a
finished product. The value stream includes all the upstream and downstream participants in the
process, as well, not just a single company. Duplication of effort and inefficient processes can be
taking place at any point in the value stream. Mapping the value stream for a product or service
can expose many steps that create no value for the customer and thus are waste that can be
eliminated. All of the concerned parties working together to create a channel for the entire value
stream and removing the waste is called the lean enterprise.

3) The remaining steps are then lined up to create continuous flow. “Flow” is the antithesis to batch
processing. Doing work in batches, whether it is production work or other activities such as invoic-
ing or accounts receivable, leads to inefficiencies because it can mean long waits between batches
when volume is low. Unless volume is in the millions of units, it is actually more efficient to create
continuous flow from one process to the next without using assembly lines. Processing steps of
different types can be conducted immediately adjacent to each other so the object undergoing
conversion is kept in continuous flow if the tools can be quickly changed and if the equipment is
made smaller. Once production has begun, the product is kept moving through the value stream
without ever placing it into a holding area to wait for later processing. When flow is introduced,
time required for processing is dramatically reduced. Furthermore, shifting demand can be accom-
modated quickly because flow provides the ability to design, schedule, and make exactly what the
customer wants when the customer wants it. That leads to the next step.

4) Pull is the next step. “Pull” means the customer pulls the product from the producer as needed
rather than the producer’s pushing products onto the customers. When producers stop making
products customers don’t want, they can end periodic price discounting that is needed to move the
unwanted goods. Customer demand becomes more stable because customers know they can get
what they want when they want it. A good example of “pull” is the printing of books (such as this
one) in response to orders received and supplying electronic versions of the book that the user
can print as needed.

5) Perfection, or rather, “continuous improvement” is the final step. Product teams in dialogue with
customers find more ways to specify value and learn more ways to enhance flow and pull. As the
first four principles continuously interact, value flows faster and more waste is exposed in the value
stream. The lean enterprise seeks perfection by continuing to make improvements.

An additional principle has been added in more recent years—that of respect for people, where “people”
include employees as well as customers, shareholders, business partners, and communities.

Lean Concepts
Minimal machine downtime when changing from manufacturing one item to manufacturing a different
item is an important component of lean production. The changeover process generally involves removing
and replacing dies from machine beds and removing and replacing direct materials used. The more the
downtime involved in the changeover process can be reduced, the less waste of resources takes place.
“SMED,” or “Single Minute Exchange of Die” is a primary method of speeding up the changeover process.
The goal of SMED is to change dies and other components in less than 10 minutes. The most powerful
method of attaining SMED is to convert all “internal setup” procedures (procedures that can be completed
only while the machine is down) to “external setup” procedures (procedures that can be completed while
the machine is running). For example, preparation for the next setup can be done while the machine or
process is still running, resulting in less downtime because the machine needs to be stopped only during
the actual setup activities.

114 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management
efham CMA
Often, large batches are assumed to be more economical than small batches due to excessively long or
difficult changeover procedures. However, production of large batches results in large on-hand inventories.
Reducing setup time makes it practical to produce smaller batches. Smaller batches can reduce finished
goods inventory on hand, allow for more varieties of products that can be produced more quickly, and lead
to greater customer responsiveness.

Ideally, batching is not done at all in lean production. The goal of lean production is to maintain continuous
flow, meaning once production of a product has begun, the product is kept moving through the value
stream without ever placing it into a holding area for later processing.

In lean production, the plant layout is re-arranged by manufacturing cells or work cells. Each work cell
produces a specific product or product type. The work cells are generally laid out in a U-shape or horseshoe
shape, but the shape can be whatever works best. The configuration’s purpose is to enable workers to
easily move from one process to another and to pass parts from one worker to another with little effort.
The goal in the layout of the work cell is for the workers to be able to pass a part or product through every
needed process with a minimum amount of wasted motion and distance. Each worker in each cell knows
how to operate all the machines in that cell and can perform supporting tasks within that cell, reducing
downtime resulting from breakdowns or employee absences. Furthermore, a properly laid-out work cell can
produce a product with a staff of just one person moving from station to station, or fully staffed with a
worker at each station, or staffed somewhere in between. Product demand determines the staffing needed
in each work cell. The rate of production is matched to the demand to avoid creating excess inventory or
incurring excess costs.

Kaizen is part of the lean manufacturing philosophy. The term kaizen is a Japanese word that means
“improvement.” As used in business, it implies “continuous improvement,” or slow but constant incremental
improvements being made in all areas of business operations. Standard costs used in manufacturing may
be either ideal standards, attainable only under the best possible conditions, or practical, expected
standards, which are challenging to attain, but attainable under normal conditions. Toyota would say that
standards in manufacturing are temporary and not absolutes. Improvement is always possible, and the
goal is to attain the ideal standard. Even though practical standards are being attained, the ultimate goal
is still not being achieved. The concept of kaizen has extended to other business operations outside the
manufacturing function, and it will be discussed later in this volume in that context.

Kanban is also a component of lean manufacturing. Kanban refers to a signal that tells workers that more
work needs to be done. Kanban is covered in more detail in the next few pages.

Error and mistake-proofing means creating improvements on many different levels to make the products
correctly the first time. Tooling and processes are often reworked to produce error-free products or to catch
errors before they become products that become spoilage or require rework. Even the design of the product
may be changed to minimize errors in manufacturing.

Just-in-time (JIT) production and inventory management are also used in lean manufacturing. JIT is
a process for synchronizing materials, operators, and equipment so that the people and the materials are
where they need to be, when they need to be there, and in the condition they need to be in. Just-in-time
processes are discussed in more detail in the next topic.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 115
D.4. Supply Chain Management CMA Part 1

Benefits of Lean Resource Management

• Quality performance, fewer defects and rework.

• Fewer machine and process breakdowns.

• Lower levels of inventory.

• Less space required for manufacturing and storage.

• Greater efficiency and increased output per person-hour.

• Improved delivery performance.

• Greater customer satisfaction.

• Improved employee morale and involvement.

• Improved supplier relations.

• Lower costs due to elimination of waste leading to higher operating income.

• Increased business because of increased customer responsiveness.

Beyond Manufacturing
The concept of lean resource management can be and is being extended beyond manufacturing to every
enterprise. Examples include but are certainly not limited to the following:

• In health care, lean thinking can be used to reduce waiting time for patients, reduce waste in
hospitals’ inventories, and in some cases, it can lead to better quality care.

• In a warehouse, individual items can be organized by size and by frequency of demand, with
parts most frequently demanded moved closest to the beginning of the sorting and picking areas.
Appropriately-sized containers can be used to hold the items and pick them, and picking can take
place more quickly. Ordering can be done daily instead of weekly or monthly and include just what
was shipped that day. As a result, smaller inventories can be carried of a broader selection of items
and the investment in inventory can be minimized.

• Service organizations tend to have long cycle times, complex variables, multiple decision points,
interactions with various computer systems, and complex interactions with customers. A team that
does similar work can work to standardize the steps they take, even with complex, variable pro-
cesses, in order to maximize the consistency and quality of services provided to customers, subject
to continuous improvement efforts.

Lean principles create greater value for the customer while using fewer resources, that is increasing value
while reducing waste. Less waste also means lower costs. However, “fewer resources” should not be con-
fused with cost-cutting activities such as laying off employees. Short-term cost-cutting can lead to short-
term increases in profit. Without a focus on increased value, however, customers will figure out that corners
are being cut. On the other hand, if creating more value with fewer resources is done properly, there will
be plenty of work for everybody because satisfied customers will want more.

116 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

Just-in-Time (JIT) Inventory Management Systems


Just-in-time inventory management systems are used in lean manufacturing. They are based on a manu-
facturing philosophy that combines purchasing, production, and inventory control into one function. The
goal of a JIT system is to minimize the level of inventories held in the plant at all stages of production,
including raw materials, work-in-process, and finished goods inventories while meeting customer demand
in a timely manner with high-quality products at the lowest possible cost.

The advantage of a JIT system is reduction in the cost of carrying the inventory. In addition to reducing the
amount invested in the inventory, the cost savings include reduction in the risk of damage, theft, loss, or
a lack of ability to sell the finished goods.

One of the main differences between JIT and traditional inventory systems is that JIT is a “demand-pull
system” rather than a “push system.” In a push system, a department produces and sends all that it can
to the next step for further processing, which means that the manufacturer is producing something without
understanding customer demand. The result of a push system can be large, useless stocks of inventory.

The main idea of a demand-pull system such as JIT is that nothing is produced until the next
process in the assembly line needs it. In other words, nothing is produced until a customer orders it,
and then it is produced very quickly. The demand-pull feature of JIT requires close coordination between
and among workstations. Close coordination between and among workstations can keep the flow of goods
smooth in spite of the low levels of inventory.

To implement the JIT approach and to minimize inventory storage, the factory must be reorganized to
permit lean manufacturing, as discussed in the preceding topic. Thus, lean manufacturing and just-in-time
inventory management go together.

Elimination of defects is an important part of a JIT system. Because of the close coordination between and
among workstations and the minimum inventories held at each workstation, defects caused at one work-
station very quickly affect other workstations. JIT requires problems and defects to be solved by eliminating
their root causes as quickly as possible. Since inventories are low, workers are able to trace problems to
their source and resolve them at the point where they originated. In addition to the advantage of lower
carrying costs for inventory, other benefits of a JIT system include greater emphasis on improving quality
by eliminating the causes of rework, scrap,16 and waste.

Supply chain management is also an important part of just-in-time inventory management, and just-in-
time inventory management is an important part of supply chain management. Because inventory levels
are kept low in a JIT system, the company must have very close relationships with its suppliers to make
certain that the suppliers make frequent deliveries of smaller amounts of inventory. In a JIT system, in-
ventory is purchased so that it will be delivered just as needed for production (or just as needed for sales,
if the company is a reseller instead of a manufacturer). The inventory must be of the required quality
because the company has no extra inventory to use in place of any defective units that are delivered.
Because very little inventory is held, a supplier that does not deliver direct materials on time or delivers
direct materials that do not meet quality standards can cause the company to not be able to meet its own
scheduled deliveries. Therefore, a company that uses JIT purchasing must choose its suppliers carefully
and maintain long-term supplier relationships.

A goal of JIT is to reduce batch setup time because reduced setup time makes production of smaller batches
more economical. The smaller batches enable the inventory reductions and enable the company to respond
quickly to changes in customer demand. The reduced setup time also leads to lower manufacturing lead
times. As mentioned in the topic Lean Manufacturing, batch setup times can be reduced by separating the
required activities into preparation activities and actual setup activities. The preparation for the next setup
can be done while the machine or process is still running, resulting in less downtime because the machine
needs to be stopped only during the actual setup activities. Material handling for setups can be improved

16
Scrap is materials left over from product manufacturing such as small amounts of supplies or surplus materials. Unlike
waste, scrap usually has some monetary value, especially if it can be recycled.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 117
D.4. Supply Chain Management CMA Part 1

by moving the material closer to the machine. These are only two suggestions, and many other things could
be done to reduce setup times, depending on the actual process.

Furthermore, JIT systems typically require less floor space than traditional factories do for equivalent levels
of production because large amounts of inventories do not need to be stored. Reductions in square footage
can reduce energy use for heating, air conditioning, and lighting. Even more importantly, reducing the
needed floor space can reduce the need to construct additional production facilities, reducing the need for
capital investment and the associated environmental impacts that result from construction material use,
land use, and construction waste.

Just-in-time production also has costs and shortcomings. The reduced level of inventory carries with it an
increased risk of stockout costs and can lead to more frequent trips for parts and material inputs from sister
facilities or from suppliers. More frequent trips can contribute to traffic congestion and environmental im-
pacts associated with additional fuel use and additional vehicle emissions. If the products produced have
large or unpredictable market fluctuations, a JIT system may not be able to reduce or eliminate overpro-
duction and associated waste. Furthermore, JIT implementation is not appropriate for high-mix
manufacturing environments, which often have thousands of products and dozens of work centers.

Benefits of Just-in-Time Inventory Management

• JIT permits reductions in the cost of carrying inventory, including the investment in inventory and
reduction in the risk of damage, theft, loss, or a lack of ability to sell the finished goods.

• Because inventories are low, workers are able to trace problems and defects to their source and
resolve them at the point where they originated. Because problems with quality are resolved quickly,
the causes of rework, scrap, and waste are eliminated, leading to improved quality.

• Batch setup time is reduced, making production of smaller batches more economical and leading to
lower manufacturing lead times. The smaller batches and lower manufacturing lead times enable
inventory reductions and enable the company to respond quickly to changes in customer demand.

• JIT systems typically require less floor space than traditional factories do, leading to reduced facility
costs.

Limitations of Just-in-Time Inventory Management

• The demand-pull feature of JIT requires close coordination between and among workstations to keep
the flow of goods smooth despite the low levels of inventory held.

• Because of the close coordination between and among workstations and the minimum inventories
held at each workstation, defects caused at one workstation very quickly affect other workstations,
so problems and defects must be solved as quickly as possible.

• Since inventory levels are kept low, raw materials received must be of the required quality and de-
livered on time because the company has no excess inventory to use in place of any defective units
that are delivered or if a delivery is delayed. Therefore, the company must maintain very close rela-
tionships with its suppliers.

• The reduced inventory carries an increased risk of stockout costs.

• If the products produced have large or unpredictable market fluctuations, a JIT system may not be
able to reduce or eliminate overproduction and associated waste.

• JIT is not appropriate for high-mix manufacturing environments, which may have thousands of prod-
ucts and dozens of work centers.

118 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

Kanban
Kanban is a Japanese inventory system. The word “kanban” means “card” or “sign” or “visual record” in
Japanese. Kanban is an integral part of lean manufacturing and JIT systems. Kanban provides the physical
inventory control cues that signal the need to move raw materials from the previous process.

The core of the kanban concept is that components are delivered to the production line on an “as needed”
basis, the need signaled, for example, by receipt of a card and an empty container, thus eliminating storage
in the production area. Kanban is part of a chain process where orders flow from one process to another,
so production of components is pulled to the production line, rather than pushed (as is done in the traditional
forecast-oriented system).

A kanban can be a card, a labeled container, a computer order, or some other device that is used to signal
that more products or parts are needed from the previous production process. The kanban contains infor-
mation on the exact product or component specifications needed for the next process. Reusable containers
may serve as the kanban, assuring that only what is needed gets produced.

Kanban can be used to control work-in-process (WIP), production, and inventory flow, thus contributing to
eliminating overproduction.

However, if production is being controlled perfectly, kanban will not be needed because the necessary parts
will arrive where they are needed at just the right time. If the parts do not arrive where and when needed,
though, then the kanban is sent to request the needed parts so that the station can keep operating. As
production control is improved, fewer kanban will be needed because the parts will almost always be where
they are needed when they are needed.

The major kanban principles are:

• Kanban works from upstream to downstream in the production process, starting with the
customer’s order. At each step, only as many parts are withdrawn as the kanban instructs, helping
ensure that only what is ordered is produced. The necessary parts in a given step always accom-
pany the kanban to ensure visual control.

• The upstream processes produce only what has been withdrawn. Items are produced only
in the sequence in which the kanban are received and only in the amounts indicated on the kanban.

• Only products that are 100 percent defect-free continue on through the production line.
At each step in the production line, defects are recognized and corrected before any more defective
units can be produced.

• The number of kanban should be decreased over time. As mentioned above, kanban are
used when the needed components do not show up on time. As areas of needed improvement are
addressed, the total number of kanban is minimized. By constantly improving production control
and reducing the total number of kanban, continuous improvement is facilitated while the overall
level of stock in production is reduced.

Different types of kanban include supplier kanban (orders given to outside parts suppliers when parts are
needed for assembly lines); in-factory kanban (used between processes in the factory); and production
kanban (indicating operating instructions for processes within a line).

It should be mentioned that kanban can be extended beyond being a lean manufacturing and JIT technique
because it can also support industrial reengineering and HR management.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 119
D.4. Supply Chain Management CMA Part 1

Introduction to MRP, MRPII, and ERP


MRP, MRPII, and ERP systems are all integrated information systems that have evolved from early database
management systems.

• MRP stands for Material Requirements Planning

• MRPII refers to Manufacturing Resource Planning

• ERP stands for Enterprise Resource Planning

MRP and MRPII systems are predecessors of ERP systems, though MRP and MRPII are still used widely in
manufacturing organizations.

Material Requirements Planning (MRP) systems help determine what raw materials to order for production,
when to order them, and how much to order. Manufacturing Resource Planning (MRPII) systems followed
MRP and added integration with finance and personnel resources.

Enterprise Resource Planning (ERP) takes integration further by including all the systems of the organiza-
tion, not just the manufacturing systems. ERP systems address the problem of paper-based tasks that
cause information in organizations to be entered into systems that do not “talk” to one another. For exam-
ple, a salesperson takes an order and submits the order on paper to an order entry clerk, who prepares the
invoice and shipping documents. The shipping documents are delivered manually to the shipping depart-
ment, and the shipping department prepares the shipment and ships the order. After shipping, the sale is
recorded and the customer’s account is updated with the receivable due. If the organization maintains
customer relations management software, the order information is entered separately into that database,
so that if the customer calls about the order, the customer service person will be able to discuss the order
with knowledge, since the customer service person does not have access to the accounting records.

The above is only a minor example, and it does not even include the communication needed with production
to make certain the product ordered will be available to ship. Entering the same information into multiple
systems causes duplication of effort and leaves the organization more vulnerable to input errors.

Enterprise Resource Planning integrates all departments and functions across a company onto a single
computer system with a single database so that the information needed by all areas of the company will be
available to the people who need it for planning, manufacturing, order fulfillment, and other purposes.

MRP, MRPII, and ERP all provide information for decision-making by means of a centralized database.

Material Requirements Planning (MRP)


Material requirements planning, or MRP, is an approach to inventory management that uses computer
software to help manage a manufacturing process. It is a system for ordering and scheduling of dependent
demand inventories.

Dependent demand is demand for items that are components, or subassemblies, used in the production
of a finished good. The demand for them is dependent on the demand for the finished good.

MRP is a “push-through” inventory management system. In a push-through system, finished goods are
manufactured for inventory on the basis of demand forecasts. MRP makes it possible to have the needed
materials available when they are needed and where they are needed.

When demand forecasts are made by the sales group, the MRP software breaks out the finished products
to be produced into the required components and determines total quantities to be ordered of each com-
ponent and the timing for ordering each component, based on information about inventory of each
component already on hand, vendor lead times and other parameters that are input into the software.

Once the quantities and the timing have been worked out, the required cash to pay for the components can
be forecasted and arranged. MRP can be used to reduce the amount of cash needed by the organization,
which in turn improves profitability and ROI. MRP creates the antithesis of the situation often found in old
manufacturing organizations where large amounts of cash are tied up in inventory before products can be

120 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

assembled and sold. Instead, MRP aims to decrease the amount of cash tied up through careful planning
and management.

Although MRP is primarily a push inventory system, it can also be used in a “demand-pull” situ-
ation, for example if an unexpected order is received, to determine the components to be purchased and
when each should be purchased in order to produce the special order as efficiently and quickly as possible
using just-in-time (JIT) inventory management techniques.

MRP uses the following information in order to determine what outputs will be necessary at each stage of
production and when to place orders for each needed input component:

• Demand forecasts for finished goods.

• A bill of materials for each finished product. The bill of materials gives all the materials, compo-
nents, and subassemblies required for each finished product.

• The quantities of the materials, components, and product inventories to determine the necessary
outputs at each stage of production.

The need for management accountants to collect and maintain updated inventory records is a challenge in
using MRP. Accurate records of inventory and its costs are necessary. Management accountants also need
to estimate setup costs and downtime costs for production runs. When setup costs are high, producing
larger batches and thus incurring larger inventory carrying costs actually reduces cost because the number
of setups needed is reduced.

Manufacturing Resource Planning (MRPII)


Manufacturing Resource Planning (MRPII) is a successor to Material Requirements Planning. While MRP is
concerned mainly with raw materials for manufacturing, MRPII’s concerns are more extensive. MRPII inte-
grates information regarding the entire manufacturing process, including functions such as production
planning and scheduling, capacity requirement planning, job costing, financial management and forecast-
ing, order processing, shop floor control, time and attendance, performance measurement, and sales and
operations planning.

An MRPII system is designed to centralize, integrate and process information for effective decision making
in scheduling, design engineering, inventory management and cost control in manufacturing.

However, if a firm wants to integrate information on its non-manufacturing functions with the information
on its manufacturing functions, it needs an ERP system.

Enterprise Resource Planning (ERP)


Enterprise Resource Planning (ERP) is a successor to Manufacturing Resource Planning. An ERP system is
usually a suite of integrated applications that is used to collect, store, manage, and interpret data across
the organization. Often the information is available in real-time. The applications share data, facilitating
information flow among business functions.

In addition to integrating production information, ERP systems integrate the logistics, distribution, sales,
marketing, customer service, human resources, and all accounting and finance functions into a single sys-
tem. ERP systems track all of a firm’s resources (cash, raw materials, inventory, fixed assets, and human
resources) and forecast their requirements, track sales, shipping, invoicing, and the status of the firm’s
commitments (orders, purchase orders, and payroll, for example).

The main focus of an ERP system is tracking all business resources and commitments regardless of
where, when, or by whom they were entered into the system. For example, a customer support
representative using an ERP system would be able to look up a customer’s order, see that the product the
customer ordered is on backorder due to a production delay, and provide an estimate for the delivery based
on the expected arrival of the required raw materials. Without the sales, support, and production systems

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 121
D.4. Supply Chain Management CMA Part 1

being tightly integrated through an ERP system, such a level of customer service would be very difficult if
not impossible to achieve.

Writing software that serves the needs of finance as well as human resources and those in the warehouse
is not an easy task. Each of the individual departments in a company usually has its own computer system
and software to help perform its specific work. However, through ERP all of them are combined into a single,
integrated software program through business re-engineering.

All of the data for the entire company is also stored in a single location, called an enterprise-wide data-
base, also known as a data warehouse. By having all of the company’s information from different
departments in the same location, a company is able to more efficiently manage and access the information.
Through data warehousing and data mining facilities, individuals in the company can sort through and
utilize the company’s information more quickly and easily than if it were stored in separate locations. In
data mining, the data in the data warehouse are analyzed to reveal patterns and trends and discover new
correlations to develop business information.

Note: If two things are correlated with one another, it means there is a close connection between them.
It may be that one of the things causes or influences the other, or it may be that something entirely
different is causing or influencing both of the things that are correlated.

Note: The major components of an ERP system are:

 Production planning
 Integrated logistics
 Accounting and finance
 Human resources
 Sales, distribution and order management

Any subdivision of any of the above components is, by itself, not a component of an ERP system.

Early ERP systems ran on mainframe computers and could require several years and several million dollars
to implement, so only the largest companies were able to take advantage of them. As the systems evolved,
vendors created a new generation of ERP systems targeted to small and mid-sized businesses that were
easier to install, easier to manage, and required less implementation time and less startup cost. Many ERP
systems are now cloud-based, and the software is not purchased and is not installed at the user company
at all but is accessed over the Internet. Use of cloud-based ERP systems allows smaller and mid-sized
businesses to access only what they need and to reduce their investment in hardware and IT personnel.

Increasingly, ERP systems are being extended outside the organization as well, for example enabling supply
chain management solutions in which vendors can access their customers’ production schedules and ma-
terials inventory levels so they know when to ship more raw materials. Interconnected ERP systems are
known as extended enterprise solutions. ERP has also been adapted to support e-commerce applica-
tions.

122 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

Operational Benefits of ERP Systems

• Integrated back-office systems result in better customer service and production and distribution
efficiencies.

• Centralizing computing resources and IT staff reduces IT costs versus every department maintain-
ing its own systems and IT staff.

• Day-to-day operations are facilitated. All employees can easily gain access to real-time information
they need to do their jobs. Cross-functional information is quickly available to managers regarding
business processes and performance, significantly improving their ability to make business deci-
sions and control the factors of production. As a result, the business is able to adapt more easily to
change and quickly take advantage of new business opportunities.

• Business processes can be monitored in new and different ways, such as with dashboards.17

• Communication and coordination are improved across departments, leading to greater efficiencies
in production, planning, and decision-making that can lead to lower production costs, lower mar-
keting expenses, and other efficiencies.

• Data duplication is reduced and labor required to create, distribute, and use system outputs is
reduced.

• Expenses can be better managed and controlled.

• Inventory management is facilitated. Detailed inventory records are available, simplifying inventory
transactions. Inventories can be managed more effectively to keep them at optimal levels.

• Trends can be more easily identified.

• The efficiency of financial reporting can be increased.

• Resource planning as a part of strategic planning is simplified. Senior management has access to
the information it needs in order to do strategic planning.

Outsourcing
When a company outsources, an external company performs one or more of its internal functions. By
outsourcing certain functions to a specialist, management can free up resources within the company in
order to focus on the primary operations of the company. It may also be cheaper to outsource a function
to a company that specializes in an area than it is to operate and support that function internally. The
disadvantage of outsourcing is that the company loses direct control over the outsourced functions.

Theory of Constraints (TOC)


For a company to be competitive, it needs to be able to respond quickly to customer orders. Theory of
Constraints is an important way for a company to speed up its manufacturing time so it can improve its
customer response time and thus its competitiveness and its profitability. Theory of Constraints (TOC) was
developed by Eliyahu M. Goldratt in the 1980s as a means of making decisions at the operational level that
will impact a company’s profitability positively.

17
The word “dashboard” originally described the cluster of instruments used by drivers to monitor at a glance the major
functions of a vehicle. As used in business, it refers to a screen in a software application, a browser-based application,
or a desktop application that displays in one place information relevant to a given objective or process. For example, a
dashboard for a manufacturing process might show productivity information for a period, variances from standards, and
quality information such as the average number of failed inspections per hour. For senior management, it might present
key performance indicators, balanced scorecard data, or sales performance data, to name just a few possible metrics
that might be chosen. The dashboard may be linked to a database that allows the data presented to be constantly
updated.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 123
D.4. Supply Chain Management CMA Part 1

Manufacturing cycle time, also called manufacturing lead time or throughput time, is usually defined
as the amount of time between the receipt of an order by manufacturing and the time the finished good is
produced. However, different firms may define the beginning of the cycle differently. For some, it begins
when a customer places an order. For others, it can begin when a production batch is scheduled, when the
raw materials for the order are ordered, or when actual production on the order begins.

In addition to the actual production time, manufacturing cycle time includes activities (and non-activities)
such as waiting time (the time after the order is received by the manufacturing department and before
manufacturing begins, or time spent waiting for parts for the next process); time spent inspecting products
and correcting defects; and time spent moving the parts, the work-in-process, and the finished goods from
one place to another.

Manufacturing cycle efficiency, or MCE, is the ratio of the actual time spent on production to the total
manufacturing cycle time.

Value-Adding Manufacturing Time


Manufacturing Cycle Efficiency (MCE) =
Total Manufacturing Cycle Time

Notice that only actual manufacturing time—time when value is being added to the product—is included in
the numerator of the MCE calculation. Waiting time, time spent on equipment maintenance, and other non-
value-adding times are not included in the numerator (though they are included in the denominator). For
example, if the actual time spent on manufacturing is 3 days while the total manufacturing cycle time is 10
days (because the waiting time is 7 days), the MCE is 3 ÷ 10, which equals 0.30 or 30%. Companies would
like their MCE to be as close to 1.00 as possible, because that means very little time is being spent on non-
value-adding activities.

A diagram of the total customer response time, from the time the customer places an order until the order
is delivered to the customer, follows.

Receipt Manufacturing Delivery


Time Cycle Time Time

Waiting Manufacturing
Time Time

Order Received Order Received Manufacturing Finished Goods Order Delivered


From Customer By Manufacturing Begins Complete To Customer

Customer-Response Time

Theory of Constraints can be used to decrease a company’s manufacturing cycle time and its costs. If a
company is not using TOC, management might be devoting its time to improving efficiency and speed in
all areas of the manufacturing process equally. However, TOC stresses that managers should focus their
attention only on areas that are constraints or bottlenecks.

Constraints are the activities that slow down the product’s total cycle time while areas and people per-
forming other activities have slack time. If managers spend their time and effort speeding up activities that
are not constraints, they are wasting resources. Unnecessary efficiency just results in the buildup of work
waiting to be done at the constraint, while activities following the constraint do not have enough work to
do because work is held up in the constraint process. If activities that are not constraints are speeded up,

124 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

total production speed is not improved despite the extra cost incurred to improve efficiency. Managers’ time
and effort and the associated cost should be spent on speeding up the activities that cause production to
slow down.

Theory of Constraints says that constraint processes are the only areas where performance im-
provements will bring about a meaningful change in overall profitability. To improve profitability,
managers need to identify the constraints and focus on improving them. Theory of Constraints
focuses on measurements that are linked directly to performance measures such as net profit, return on
investment, and cash flow. It gives managers a method of making decisions on a day-to-day basis that will
truly affect the overall performance of the organization.

Throughput time, or manufacturing cycle time, is the time that elapses between the receipt of a cus-
tomer order by the manufacturing activity and the completion and shipment of the finished goods.
Throughput time is a rate. It is the rate at which units can be produced and shipped. For example, if it
takes 2 days to produce and ship 100 units, then the rate per day is 50 units per day.

Throughput contribution margin is the rate at which contribution margin is being earned in monetary
terms. Throughput contribution margin is the revenue earned from the sale of units minus the totally vari-
able costs only (usually only direct materials) for those units sold during a given period of time. If the sale
price for one unit is $500, the direct materials cost is $300 per unit, and throughput rate per day is 50 units
per day, then the throughput contribution margin per day is $200 × 50 = $10,000. Or, calculated
another way, if 50 units can be produced and shipped in one 8-hour day, then it takes 8 hours ÷ 50, or
0.16 of one hour, to produce and ship one unit. $200 ÷ 0.16 = $1,250 per hour. In an 8-hour day, through-
put contribution margin is $1,250 × 8, or $10,000.

Note: Throughput contribution margin is the amount earned for product produced and shipped during a
period such as one hour, one day, or one month, calculated using revenue for the period minus only the
strictly variable costs. Strictly variable costs are usually only direct materials costs.

Following are the steps in managing constrained operations through the use of TOC analysis:

1) Identify the constraint. Recognize that the constraint or bottleneck operation determines the
throughput contribution margin of the system as a whole, and identify the constraint by determin-
ing where total required hours exceed available hours. The management accountants work with
manufacturing managers and engineers to develop a flow diagram that shows the sequence of
processes, the amount of time each process requires given current demand levels, and the amount
of time available in terms of labor hours and machine hours. The flow diagram enables the man-
agement accountants, manufacturing managers, and engineers to identify the constraint.

2) Determine the most profitable product mix given the constraint. The most profitable prod-
uct mix is the combination of products that maximizes total operating income. Product profitability
is measured using the throughput contribution margin. The throughput contribution margin is
the product price less materials cost, including the cost of all materials used, all purchased com-
ponents, and all materials-handling costs. Direct labor and other manufacturing costs such as
overhead are excluded, because it is assumed they will not change in the short term. The through-
put contribution margin of each product is divided by the number of minutes required for one unit
at the constraint to calculate the throughput contribution margin per minute per unit in the
constraint activity for each product. The product with the highest throughput contribution margin
per minute in the constraint will be the most profitable, even though it may have a lower through-
put contribution margin for the manufacturing process as a whole.

3) Maximize the flow through the constraint. The management accountant looks for ways to
simplify the constraint process, reduce setup time, or reduce other delays due to non-value-adding
activities such as machine breakdowns, in order to speed the flow through the constraint.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 125
D.4. Supply Chain Management CMA Part 1

4) Add capacity to the constraint. Increase the production capabilities of the constraint by adding
capacity such as additional equipment and labor. Adding equipment and labor are longer-term
measures to consider if it is possible and profitable to do so.

5) Redesign the manufacturing process for flexibility and fast cycle time. Analyze the system
to see if improvements can be made by redesigning the manufacturing process, introducing new
technology, or revising the product line by eliminating hard-to-manufacture products or by rede-
signing products so they can be manufactured more easily. This final step in managing constrained
operations is the most strategic response to the constraint.

TOC Terms
Exploiting the constraint means taking advantage of the existing capacity at the constraint, because that
capacity can be wasted by producing the wrong products or by improper policies and procedures for sched-
uling and controlling the constraint. “Exploiting the constraint” means using it to its best advantage to
produce the product that will be most profitable to sell and by scheduling work so that the constraint is
kept busy all the time.

Elevating the constraint means adding capacity to the constrained activity or otherwise adjusting the
resources to increase the output possible from the constrained activity. Adding capacity means purchasing
an additional machine or using a new technology that makes better use of the existing machine.

126 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

Drum-Buffer-Rope
Drum-Buffer-Rope is the production planning methodology portion of Theory of Constraints. It is a tool that
can be used to balance the flow of production through the constraint. It minimizes the buildup of excess
inventory at the constraint while at the same time keeping the constraint producing at all times.

• Drum: The drum is the process that takes the longest time. It is the constraint. The constraint
is called the drum because it provides the beat that sets the pace for the whole production process.
All production flows must be synchronized to the drum.

• Rope: The rope consists of all of the processes that lead up to the drum, or the constraint. Activ-
ities preceding the drum must be carefully scheduled so that they do not produce more output
than can be processed by the constraint, because producing too much output in activities preceding
the constraint creates excess inventory and its associated costs without increasing throughput
contribution margin. At the same time, though, the constraint must be kept working with no down
time.

• Buffer: The buffer is a minimum amount of work-in-process inventory (a “buffer” inventory) of


jobs waiting for the constraint. The purpose of the buffer is to make sure the constraint process is
kept busy at all times. Production schedules are planned so that workers in the non-constrained
processes will not produce any more output than can be processed by the drum, the constraint
process; but at the same time, the non-constrained processes must produce enough to keep the
buffer full.

Example: A company manufactures garments. The manufacture of a jacket involves four separate pro-
cesses:
1) Cutting the fabric pieces
2) Stitching the fabric pieces together
3) Hemming the sleeves and the bottom of the jacket
4) Finishing the jacket, folding it, and packaging it in clear plastic.

Hemming the sleeves and the bottom of the jacket requires the most time and is the constraint.

The garment manufacturer sells only to wholesalers, who in turn sell to retailers. The time required for
each process for one jacket and the available times are as follows. The total hours available per month
are calculated by assuming 22 work days per person per month and 7 hours work per person per day.
Minutes Total Hours
Required Number of Available
per unit Employees Per Month
Cutting 18 20 3,080
Stitching 20 23 3,542
Hemming 30 28 4,312
Finishing, folding, and packaging 10 11 1,694

The demand per month averages 10,000 jackets per month. The constraint process is the hemming
process, the process for which the demand exceeds the hours available. Below are the total hours re-
quired to produce 10,000 jackets per month using the current number of employees and the current
equipment. The total hours required is 10,000 × minutes required per unit ÷ 60.
Difference
Total Total Between
Hours Hours Hrs. Required &
Required Available Hrs. Available
Cutting 3,000 3,080 (80)
Stitching 3,334 3,542 (208)
Hemming 5,000 4,312 688
Finishing, folding, and packaging 1,667 1,694 (27)
(Continued)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 127
D.4. Supply Chain Management CMA Part 1

Since the different jobs the employees are doing are not specialized, some of the employees currently
doing other jobs could be shifted to the hemming process, since they have some extra time and the
hemming process requires more time than is currently available. One employee currently doing cutting
could spend half time doing hemming instead, and one employee could be shifted from stitching to
hemming. The company has enough equipment to accommodate those changes. If those changes are
made, the number of employees per process will change to 19.5 for Cutting, 22 for Stitching, and 29.5
for Hemming. That would create the following changes in total hours available and the differences be-
tween hours required and hours available:
Difference
Total Total Between
Hours Hours Hrs. Required &
Required Available Hrs. Available
Cutting 3,000 3,003 (3)
Stitching 3,334 3,388 (54)
Hemming 5,000 4,543 457
Finishing, folding, and packaging 1,667 1,694 (27)
The production capability of the whole department is dependent on the production capability of the
constraint, which is the hemming process. The production line cannot move any faster than its slowest
process. After making these duty reassignments, the company still does not have the capacity to produce
10,000 units per month. With 4,543 hours available at the constraint, the company can produce only
9,086 jackets per month (4,543 hours available × 60 minutes per hour ÷ 30 minutes to hem one jacket).

The long-term answer to the problem is to add capacity to the constraint (elevate the constraint).
However, in the short-term, the challenge is to exploit the constraint by maximizing the flow through
the constraint—making sure that the hemming operation has no downtime during which it is waiting for
the prior processes to provide it with jackets to hem. To do that, the company needs to make sure it has
a small work-in-process inventory waiting to be hemmed at all times while not letting the inventory stack
up too much in front of the hemming process.

The Drum-Buffer-Rope system is one of the ways the size of the work-in-process inventory waiting to
be hemmed can be controlled.
CONSTRAINT
BUFFER
Small amount of
Input
work-in-process
inventory

Cutting Stitching Hemming Finishing

ROPE
Output

(Continued) DRUM

128 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

The work in cutting and stitching must be carefully controlled so that just the right amount of work-in-
process is in the buffer waiting to be hemmed at any one time so that the hemming process is always
busy and always working at its maximum while at the same time, too much work-in-process is not
allowed to build up in the buffer.

In the long term, if the company wants to meet the full demand, it will need to either hire more hemming
employees and invest in more equipment for them to use, or it will need to find some way to increase
the speed of the hemming process. The company finds it can invest in attachments for the hemming
machines that will reduce the time for the hemming from 30 minutes per jacket to 27 minutes per jacket.
After analyzing the costs of the two alternatives, management determines it will cost less to purchase
the attachments for the hemming machines in order to make better use of the 4,543 hours available
with current employees and equipment. With hemming requiring only 27 minutes per jacket, the time
required for 10,000 jackets will be 4,500 hours (10,000 jackets × 27 minutes to hem one jacket ÷ 60
minutes per hour). Now, the hours required for Hemming are decreased and are as follows:
Difference
Total Total Between
Hours Hours Hrs. Required &
Required Available Hrs. Available
Cutting 3,000 3,003 (3)
Stitching 3,334 3,388 (54)
Hemming 4,500 4,543 (43)
Finishing, folding, and packaging 1,667 1,694 (27)

The company now has the capability to manufacture 10,000 jackets per month.

When the Theory of Constraints is applied to production, speed in manufacturing is improved by increasing
throughput contribution margin while decreasing investments and decreasing operating costs.
Throughput (product produced and shipped) will be maximized, investments will be minimized, and oper-
ating costs will be minimized. TOC helps reduce throughput time, or cycle time, and therefore operating
costs. Furthermore, the use of TOC enables the company to carry a lower overall level of inventory, so
inventory investment is decreased.

In TOC terms, “investments” is the sum of costs in direct materials, work-in-process and finished goods
inventories; R&D; and costs of equipment and buildings. Inventory costs are limited to costs that are
strictly variable, called “super-variable.” Super-variable costs are usually only direct materials.

Note: Absorption costing for external financial reporting purposes is not done any differently when TOC
is being used. Inventory costs for internal TOC analysis purposes are simply different from inventory
costs for financial reporting purposes.

Also, for the purpose of using Theory of Constraints, operating costs are equal to all operating costs other
than direct materials or any other strictly variable costs incurred to earn throughput contribution margin.
In TOC, “operating cost” is defined as the cost of converting the inventory into throughput.

Thus, operating costs in TOC include salaries and wages, rent, utilities, depreciation, indirect materials, and
other overhead costs. For TOC purposes, all of these operating costs are treated as period costs that are
expensed as incurred, and they are not inventoried. Inventory for TOC includes only the direct material
costs. (Again, “inventory” for TOC purposes is used only for TOC management. It is not the same as “in-
ventory” for absorption costing.) All employee costs are considered operating costs, whether they are direct
labor or indirect labor. Direct labor is not included in the calculation of throughput contribution
margin, and thus it is considered an operating cost for TOC.

When work is properly scheduled, the constraint will achieve its maximum performance without interrup-
tions. The material is released only as needed without building up unneeded material (inventory) at the
non-bottleneck resources, enabling the factory to achieve optimal performance.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 129
D.4. Supply Chain Management CMA Part 1

Summary of Theory of Constraints:

• Throughput is product produced and shipped.

• Throughput time or manufacturing cycle time is the time that elapses between the receipt by
the manufacturing department of a customer’s order and the shipment of the order.

• Throughput contribution margin is revenue minus direct materials cost for a given period of time.

• Only strictly variable costs—which are usually only direct materials—are considered inventory
costs. All other costs, even direct labor, are considered operating, or fixed costs.

• Theory of Constraints assumes that operating costs are fixed costs because they are difficult to
change in the short run.

• Theory of Constraints focuses on short-run maximization of throughput contribution margin


by managing operations at the constraint in order to improve the performance of production as a
whole.

Some ways that operations at the constraint process can be relieved include:

• Eliminate any idle time at the constraint operation, such as time when the machine is not
actually processing products. For example, perhaps an additional employee could be hired to work
at the constraint operation to help unload completed units as soon as a batch is processed and to
help load the machine for the next batch. If doing that would increase throughput at the constraint
by a minimum of 2,000 units per year at an annual cost of $40,000, and if the throughput contri-
bution margin (selling price minus direct material cost per unit) is greater than $20 per unit
($40,000 ÷ 2,000), then hiring an additional employee at the constraint would increase operating
income.

• Process only products that increase throughput contribution margin and do not produce
products that will simply remain in finished goods inventory. Making products that remain in in-
ventory does nothing to increase throughout contribution margin.

• Move items that do not need to be processed on the constraint operation to other, non-
constrained machines or outsource their production.

• Reduce setup time and processing time at the constraint operation. If reducing setup time
costs an additional $10,000 per year but it enables the company to produce 500 additional units
per year, again, if the throughput contribution margin is greater than $20 per unit ($10,000 ÷
500), operating income will increase.

• Improve the quality of processing at the constrained resource. Poor quality is more costly
at a constraint operation than at a non-constraint operation. Since a non-constraint operation has
unused capacity, no throughput contribution is forgone when a non-constraint operation produces
product that cannot be sold, so the cost of the defective production is limited to the wasted mate-
rials. However, unsellable production at the constraint operation costs more than just the cost of
the wasted materials. The cost of poor quality at a constraint also includes the opportunity cost of
lost throughout contribution margin, because the constraint does not have any extra time to waste.
Lost time at the constraint is lost throughput contribution margin. Therefore, the constraint oper-
ation should not waste time processing units that were defective when received from the previous
process. Units in production should be inspected before they are passed on to the constraint
operation for processing.

If these actions are successful in increasing the capacity of the constraint operation, its capacity may even-
tually increase to the point where it exceeds the capacity of some other process, and so the other process
may become the constraint. The company would then focus its continuous improvement actions on increas-
ing the efficiency and capacity of the new constraint process, and so on.

130 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

Theory of Constraints Reports


A Theory of Constraints Report conveys throughput contribution margin and selected operating data.
It identifies each product’s throughput contribution margin per hour required for the constraint. It also
identifies the most profitable product or products and enables monitoring to achieve maximum prof-
itability given existing constraints. By identifying the most profitable products, it can assist in making
decisions about product mix.

Calculating Throughput Contribution Margin


The concept of throughput contribution margin as it is used in Theory of Constraints analysis is a variation
on the concept of contribution margin. The contribution margin is the difference between total revenues
and total variable costs, and contribution margin per unit is simply the sale price for one unit minus the
total variable costs for one unit. The variable costs included in the calculation of the contribution margin
include direct materials, direct labor, variable overhead, and variable selling expenses.

However, in Theory of Constraints analysis, everything except for direct materials and any other totally
variable cost is considered an operating cost and thus a period cost. The throughput contribution margin or
throughput contribution margin per unit in TOC analysis is the selling price minus only the totally var-
iable costs, as referred to in the above discussion of TOC analysis. The totally variable costs are usually
only the direct materials costs. Calculating the contribution margin in this way is called super-variable
costing. The assumption is made that labor and overheads are fixed costs because usually they cannot be
changed in the short term.

If an exam question requires calculation of the throughput contribution margin (or throughput contri-
bution or throughput margin—they all mean the same thing) for a period of time, calculate how many units
can be produced in that time. The throughput contribution margin will be the throughput contribution
margin per unit multiplied by the number of units that can be produced in the given time.

Example: Using the garment manufacturer again, the manufacturer has two different styles of jackets:
a down-filled jacket for winter and a light jacket for spring. For both jackets, the constraint is the hem-
ming operation. The winter jacket sells for $125, and the direct materials cost is $75. The spring jacket
sells for $75, and the direct materials cost is $30. The hemming process takes 30 minutes for the winter
jacket and 25 minutes for the spring jacket. Demand for the winter jacket is 6,000 jackets per month,
whereas demand for the spring jacket is 8,000 jackets per month. The company has 4,543 hours avail-
able per month for hemming. Which jacket should the company give priority to in scheduling production?

The company should give priority to the product with the higher throughput margin per minute,
calculated using the time required in the constraint process.

Winter Jacket Spring Jacket


Price $125.00 $75.00
Direct Materials cost 75.00 30.00
Throughput contribution margin $ 50.00 $45.00
Constraint time for hemming 30.00 25.00
Throughput margin per minute $ 1.67 $ 1.80

The company should manufacture the 8,000 spring jackets needed before manufacturing any winter
jackets, because the spring jacket’s throughput margin per minute in the constrained resource is higher.
That will mean using 3,334 hemming hours for the 8,000 spring jackets required (8,000 × 25 minutes
per jacket ÷ 60 minutes per hour). That will leave 1,209 hemming hours available for the winter jackets
(4,543 hours available – 3,334 hours used for the spring jackets). With those 1,209 hours, the company
can manufacture 2,418 winter jackets (1,209 hours × 60 minutes per hour ÷ 30 minutes per jacket).
That will maximize the company’s total contribution margin.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 131
D.4. Supply Chain Management CMA Part 1

Example: Below is another use for the throughput contribution margin.

A company has an opportunity to get a special order for 100,000 units. The company has excess capacity
(unused direct labor and equipment) and would really like to have this order. Because of the unused
capacity, the company would not need to hire any additional employees if it gets the order and would
not need to cut back on production of other orders. The company is bidding against several other com-
panies, so management knows in order to get this order, it must get the price down as low as possible.
The calculated costs per unit are as follows:
Direct materials $ 6
Direct labor 3
Fixed Overhead 2
Total $11

Management determines that if it bids $900,000 ($9 per unit), it will just cover the variable costs (direct
materials and direct labor), so it bids $900,000. To management’s amazement, however, the company
loses the bid to a company that bid $750,000.

If the company had bid $700,000, or $7 per unit, it would have gotten the job. The company would have
recovered its materials costs and its operating income would have been increased by $100,000 ($1 per
unit × 100,000 units). The direct labor cost would not have been covered; but then, the company was
paying those people anyway, even without the special order. Even though management did not have
work for them without the special order, it had no plans to lay them off. Therefore, the direct labor cost
is like a fixed cost because it would have been the same whether the company had the special order or
not. Likewise, the fixed manufacturing costs would have been no different whether the company had the
order or not.

To illustrate this incremental analysis, operating income is as follows without the special order, assum-
ing a selling price of $12 and a current volume of 1,000,000 units:
Sales revenue $12 × 1,000,000 $ 12,000,000
Variable costs:
Direct materials $6 × 1,000,000 6,000,000
Direct labor $3 × 1,000,000 3,000,000
Contribution margin $ 3,000,000
Fixed overhead $2 × 1,000,000 2,000,000
Operating income $ 1,000,000

If management had bid $700,000 and gotten the special order, operating income would have been as
follows:
Sales revenue for regular orders $12 × 1,000,000 $ 12,000,000
Sales revenue for special order $7 × 100,000 700,000
Total revenue $ 12,700,000
Variable costs:
Direct materials $6 × 1,100,000 6,600,000
Direct labor $3 × 1,000,000 3,000,000
Contribution margin $ 3,100,000
Fixed overhead $2 × 1,000,000 2,000,000
Operating income $ 1,100,000

If the company had calculated a throughput contribution margin using a price of $7 per unit ($7 −
$6 direct materials = throughput contribution margin of $1 per unit) in preparing the bid instead, it
would have gotten the special order and would have earned $100,000 more in operating income
(100,000 units × $1) than it earned without the special order.

Next time, management will use the throughput contribution margin in preparing its bid.

132 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

The following information is for the next four questions: EEK Industries produces thingamajigs.
The thingamajigs sell for $200 per unit. The product goes through three processes. The processes, costs
and volumes are as follows:

Molding Heat Transfer Trimming


Direct labor required
per unit @ $25/hr. 0.5 hrs. (2 units/hr.) 0.25 hrs. (4 units/hr.) 0.5 hrs. (2 units/hr.)
Direct labor hours available
per day 225 hrs. 100 hrs. 200 hrs.

Direct materials $100 per unit $20 per unit $0 per unit
Annual fixed manu-
facturing overhead $1,000,000 $750,000 $250,000
Fixed selling and admin.
exp. (allocated according $428,571 $190,476 $380,953
to available DL hrs.)
Daily capacity in units based
on DL hrs. available 450 units 400 units 400 units
Annual capacity (260
working days per year) 117,000 units 104,000 units 104,000 units
Annual production & sales 104,000 units 104,000 units 104,000 units

EEK has a policy of not laying off idle workers but instead finds other work for them to do such as
maintenance on equipment. The company can sell all it can produce.

Question 41: What is the throughput contribution per unit according to the theory of constraints?

a) $48.75

b) $80.00

c) $30.59

d) $29.52

Question 42: What is the throughput contribution per day?

a) $36,000

b) $19,500

c) $12,236

d) $32,000

Question 43: What are the annual operating costs?

a) $3,412,500

b) $1,000,000

c) $6,412,500

d) $5,412,500

(Continued)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 133
D.4. Supply Chain Management CMA Part 1

Question 44: The Heat Transfer and Trimming processes are very labor-intensive. How could EEK better
assign its existing workers to increase its operating income?

a) By reassigning an employee for 5 hours per day from Molding to Heat Transfer.

b) By reassigning an employee for 5 hours per day from Molding to Heat Transfer and reassigning
employees for 10 hours per day from Molding to Trimming.

c) By reassigning employees for 10 hours per day from Molding to Heat Transfer and reassigning an
employee for 5 hours per day from Molding to Trimming.

d) By reassigning an employee for 5 hours per day from Molding to Heat Transfer and reassigning
an employee for 5 hours per day from Molding to Trimming.

(HOCK)

Question 45: Urban Blooms is a company that grows flowering plants and sells them in attractively
designed container arrangements to upscale hotels, restaurants and offices throughout the greater New
York City metropolitan area. When first established, the organization produced every aspect of its prod-
uct on site and handled all business functions from its facility, in either the greenhouses, production
areas, or office. The only exception was importing expensive, large containers from Mexico. After five
years in business, Urban Blooms had become very profitable and increased its staff from 10 to 200
employees, including horticulturalists, production/design workers, business managers, and sales staff.
However, the owners found it increasingly difficult to keep up with the complexities and demands
brought about by the company’s continuing growth. It became apparent that several areas of the busi-
ness were not time- or cost-effective for Urban Blooms, such as mixing its own potting soil for the plants
and maintaining payroll accounting. Urban Blooms contacted Hampshire Farms to create a special pot-
ting soil mix for its plants and hired Lindeman Associates to handle the company’s payroll accounting
requirements.

This process is referred to as:

a) Materials resource planning (MRP).

b) Activity-based costing (ABC).

c) Outsourcing.

d) Lean production.

(HOCK)

134 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

Capacity Level and Management Decisions


Estimates about plant production capacity, called denominator level capacity, are used for various purposes.
Production activity levels are used in developing standards and the budget and allocating manufacturing
overhead, as previously discussed. However, production activity levels are also used in making manage-
ment decisions, particularly decisions about pricing, bidding, and product mix.

A company does not need to use the same denominator level capacity for management decisions as it uses
to set standards and allocate overhead for its external financial statements. The capacity level used in
making decisions should meet the need for the purpose for which it will be used. Recall that the various
choices of capacity levels are:

1) Theoretical, or ideal capacity, which assumes the company will produce at its absolutely most
efficient level at all times, with no breaks and no downtime.

2) Practical capacity, which is the most that a company can reasonably expect to produce in a year’s
time. Practical capacity is the theoretical level reduced by allowances for idle time and downtime,
but not reduced for any possible decreases in sales demand.

3) Master budget capacity, which is the amount of output actually expected during the budget period
based on expected demand.

4) Normal capacity, which is the level of annual activity achieved in the long run that will satisfy
average customer demand over a period of 2 to 3 years.

The level of plant capacity to use in decision-making is an important strategic decision that needs to be
made by management. In fact, that decision should really be made before any decisions are made about
how much plant capacity to provide. If the company needs to manufacture 15,000 units per year, then
management needs to determine how much it will cost per year to acquire the capacity to produce those
15,000 units. Since the capacity acquired will be the capacity needed for 15,000 units, 15,000 units is the
company’s practical capacity (theoretical capacity is not attainable).

In the short run, the capacity and its cost are fixed. If the company does not use all of the capacity it has
available, the fixed costs of that capacity do not decrease the way variable costs do. If the cost to acquire
and maintain the capacity to make 15,000 units per year is $1,125,000 per year, then at a production level
of 15,000 units, the fixed cost per unit is $75. However, if the company does not produce 15,000 units in
a given year but instead produces only 12,500 units, not all of the capacity supplied at $75 per unit will be
needed or used that year. The company will be paying for unused capacity. The fixed cost per unit produced
based on actual production would be $90 per unit ($1,125,000 ÷ 12,500 units); but in reality, the fixed
cost per unit should not change just because the number of units manufactured changes. The “real” fixed
cost per unit manufactured is still $75 per unit; and the company has unused capacity cost of $187,500
for the units not produced ($75 × [15,000 – 12,500]) that it needs to absorb as an expense.

Pricing decisions and bidding decisions should be made using the $75 per unit fixed cost that results from
using practical capacity to calculate the fixed cost per unit, regardless of whether that volume is being
produced or not.

Use of practical capacity as the denominator level for pricing and bidding decisions best expresses what the
true cost per unit of supplying the capacity should be, regardless of the company’s usage of its available
capacity. Recall that practical capacity is the absolute most that the company can reasonably expect to
produce in a year’s time using the capacity it has. It is the theoretical capacity level reduced by allowances
for unavoidable interruptions such as shutdowns for holidays or scheduled maintenance, though not de-
creased for any expected decrease in sales demand. Therefore, practical capacity is the best denominator
level to use in pricing and other management decisions.

Customers cannot be expected to absorb the unused capacity cost by paying higher prices that the company
charges per unit in order to cover the higher fixed cost it is allocating to each unit produced because
production is below the expected level. Customers will not absorb it. They will take their business elsewhere.
That will result in even lower production and even greater fixed cost per unit, even higher prices, and even

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 135
D.4. Supply Chain Management CMA Part 1

lower sales, called a downward demand spiral. A downward demand spiral can put a company out of
business. Customers expect a company to manage its unused capacity or else to bear the cost of the unused
capacity, not to pass it along to them.

Since the use of practical capacity excludes the cost of the unused capacity from the per unit fixed cost, it
gives management a more accurate idea of the resources needed to produce one unit of product and thus
the resources needed to produce the volume the company is actually producing. If the company does not
need all of its capacity, management should make adjustments. It may be that the company’s unneeded
capacity can be rented or sold. Alternatively, management might be able to make use of that unused
capacity by developing a new product or by producing a product for another company that is outsourcing
some of its manufacturing.

Highlighting the cost of the unused capacity gives management the opportunity to make strategic decisions
for its use.

Capacity Level and its Effect on Financial Statements


The company will have an under-application of manufacturing overhead any time the actual overhead in-
curred is greater than the amount of manufacturing overhead applied to the level of production achieved.
Under-application of overhead occurs when production is lower than anticipated and as a result, the man-
ufacturing overhead charged to the products produced was less than the actual incurred manufacturing
overhead. On the other hand, if production is higher than anticipated, the amount of manufacturing over-
head applied will be greater than the amount actually incurred, and the manufacturing overhead will be
over-applied.

At the end of the accounting period, variances that result from differences between the actual overhead
incurred and the overhead applied must be resolved as part of the closing activities. The variances should
be prorated among ending work-in-process inventory, ending finished goods inventory, and cost of goods
sold for the period according to the amount of overhead included in each that was allocated to the
current period’s production. Or, if the variances are not material, they may be 100% closed out to cost
of goods sold. A third approach (though not often used) is to restate all amounts using actual cost allocation
rates rather than the budgeted cost application rates.

• If a variance is closed out 100% to cost of goods sold, reported operating income and inventory
balances will vary depending on the activity level that was used to calculate the overhead applica-
tion rate.

• If a variance is pro-rated between inventories and cost of goods sold according to the amount of
that type of overhead allocated to the current period’s production for each, reported operating
income will be the same regardless of what activity level was used to calculate the fixed overhead
application rate. The combination of the amount allocated to production and the pro-ration of the
variance will cause inventories and cost of goods sold expense to be equal to the actual rate under
all activity levels.

Note: The pro-ration of under- or over-applied overhead should be done on the basis of the
overhead allocated to production during the current period only, not on the basis of the balances
in the inventories and cost of goods sold accounts at period end. Information on the amount of overhead
allocated to production during the current period should be available in the accounting system.

If variances are closed out 100% to cost of goods sold, the use of master budget capacity or normal capacity
will lead to higher operating income than if theoretical or practical capacity is used. Operating income
will be higher under master budget capacity or normal capacity whether the overhead is under-applied or
over-applied.

When master budget or normal capacity is used, the denominator level used will be lower than if theoretical
or practical capacity is used and the resulting application rate will be the higher. Therefore, more manufac-
turing overhead will be allocated to each product produced throughout the period than it will be if theoretical

136 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.4. Supply Chain Management

or practical capacity is used. Thus, more manufacturing overhead will be included in the finished goods and
work-in-process inventories on the balance sheet at the end of the period than would be the case with
usage of the other denominator levels.

To explain: When a variance is closed out 100% to cost of goods sold, no adjustment is made to inventories
as part of the closing entries, so inventories under master budget or normal capacity will remain higher
than under the other methods. Since inventories are higher, cost of goods sold will be lower. The result is
that operating income will be higher when master budget or normal capacity is used and 100% of the
variances are closed out to COGS.

However, if variances caused by differences between the actual manufacturing overhead incurred and the
overhead applied during the period are pro-rated among finished goods inventory, work in process inven-
tory and cost of goods sold on the basis of the amount of overhead allocated during the current period to
the units in each, the choice of the denominator level for allocating manufacturing costs during the period
will have no effect on the end-of-period financial statements.

After the variances have been resolved and closed out for the period, final operating income and inventory
balances will be the same no matter which capacity level was used to determine the overhead allocation
rate during the period if the variances were pro-rated. Under all of the capacity levels, operating income
and inventory balances will be same as if the actual incurred overhead had been allocated to the period’s
production.

Note: The above will be true only when the variance is pro-rated among inventories and cost of goods
sold on the basis of the amount of overhead allocated during the current period to the units in each. If
the variance is charged to cost of goods sold only, operating income and inventory balances will not be
the same as if the actual incurred overhead had been allocated to the period’s production.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 137
D.5. Business Process Improvement CMA Part 1

D.5. Business Process Improvement


A business process is a related group of activities encompassing several functions that produces a specific
product or service of value to a customer or customers. A business process is also activities that result in
higher quality or lowered costs of a product or service.

The Value Chain and Competitive Advantage


Michael Porter, a leading authority on competitive advantage from Harvard Business School, introduced the
concept of the value chain in his 1985 book, Competitive Advantage: Creating and Sustaining Superior
Performance.

Competitive Advantage
Competitive advantage, discussed in Strategic Planning in Section B of Vol. 1 of this textbook, is an ad-
vantage that a company has over its competitors that it gains by offering consumers greater value than
they can get from its competitors. Competitive advantage is derived from attributes that enable the organ-
ization to outperform its competitors, such as access to natural resources, highly-skilled personnel, a
favorable geographic location, high entry barriers, and so forth. The greater value offered to customers
may be in lower prices for the same product or service; or it may be in offering greater benefits and service
than its competitors do, thereby justifying higher prices; or it may be offering greater benefits at the same
or even at a lower price than its competitors charge.

A company that has competitive advantage will be more profitable than the average company it competes
with for customers. The higher its profits are in comparison to its competitors, the greater its competitive
advantage will be. Competitive advantage leads to increased profitability; and greater profitability leads to
increased competitive advantage. Competitive advantage makes the difference between a company that
succeeds and a company that fails.

Competitive Increased
Advantage Profits

In order to have competitive advantage, a company must have or create two basic things:

1) Distinctive competencies and the superior efficiency, quality, innovation, and customer
responsiveness that result.

2) The profitability that is derived from the value customers place on its products, the price that
it charges for its products, and the costs of creating those products.

Note: A company’s distinctive competencies are the things that it does better than the competition.

The four generic distinctive competencies that give the company competitive advantage are:

• Superior efficiency
• Superior quality
• Superior innovation
• Superior customer responsiveness

138 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

The Value Chain


Manufacturing companies create value for customers by transforming raw materials into something of value
to customers. Retailers create value by providing a range of products to customers in a way that is con-
venient for the customers, sometimes supported by services such as fitting rooms or personal shopper
advice. In service businesses, value is created when people use inputs of time, knowledge, equipment and
systems to create services of value to their customers.

The value that is created less its cost is called the profit margin:

Value Created − Cost of Creating that Value = Profit Margin

Anything that increases the value to the customer or decreases the cost of creating that value adds to the
profit margin. All of an organization’s functions play a role in lowering the cost structure and increasing the
perceived value of products and services through differentiation. The creation of value involves functions
from R&D through production to marketing and sales, and on to customer service, and includes all the
activities that play supporting roles.

The value chain describes a business unit’s chain of activities for transforming inputs into the outputs that
customers will value. The chronological sequence of the value chain activities adds value to the customer.

The transformation process includes primary activities (business functions), those that are essential for
adding value to the product or service and creating competitive advantage, and support activities, those
activities that support the primary activities. Support activities add value indirectly to the product or service.

Michael Porter’s 1985 Value Chain


The value chain as envisioned by Michael Porter in 1985 looked like the following:
Support Activities

Infrastructure

Human Resources Management


Technological Development

Procurement Value Added


Minus Cost
= Margin
Operations

Marketing
Outbound

and Sales
Logistics

Logistics
Inbound

Service

Primary Activities

Primary Activities:
• Inbound logistics is materials management and includes receiving and handling inputs to the
product, inventory control, warehousing, scheduling deliveries, and returns to the supplier.

• Operations activities are activities necessary to transform inputs into final products and services,
such as machining, assembly, equipment and facility maintenance, and packaging.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 139
D.5. Business Process Improvement CMA Part 1
efham CMA
• Outbound logistics involves handling of the finished products to distribute them to buyers or to
warehouses. It includes order processing and operation of delivery vehicles.

• Marketing and sales activities provide the means for buyers to purchase the firm’s products and
services and convince them to do so, such as by advertising, sales activities, marketing channel
selection, and making pricing decisions.

• Customer service includes activities that provide support to customers by enhancing or main-
taining the value of the product. Repair, parts supply, and employee training are service activities.

Support Activities:
Infrastructure includes the administrative activities such as management, accounting, finance, internal
audit, and legal counsel that are necessary to support the whole organization and enable it to maintain its
day-to-day activities. The infrastructure also includes the organizational structure, the organization’s con-
trol systems, and its culture. Through strong leadership, top management can shape the company’s
infrastructure and thus the performance of all the other value creation activities that go on within the
company.

Human resource management, through its involvement with recruiting activities, hiring, employee train-
ing and development, and compensation, aids the organization in obtaining and keeping the right mix of
skilled people needed to perform its value creation activities effectively. The human resources function
ensures that the people are properly trained, motivated, and compensated. An effective human resources
function leads to greater employee productivity, which lowers costs. A competent staff performs excellent
customer service, which increases value to the customer. Human resource management supports primary
activities, other support activities, and the entire value chain.

Technological development supports efforts to improve products, processes, and services throughout
the support activities and primary activities. Technology development that is related to product develop-
ment supports the entire value chain, and other technology development, such as investments in IT, are
associated with specific primary or support activities.

Procurement is the process of purchasing inputs that are used in the firm’s value chain. Inputs include
raw materials and supplies as well as fixed assets. Purchased inputs are used not only in the primary
activities, but they are present in every support activity, as well. Procurement may be performed by various
persons throughout the organization in addition to the purchasing department, for example office managers
who may purchase office supplies.

Value Chain Activities Today


Value chain activities vary from business to business and a specific activity may be a primary activity in
one business and a support activity in another business, depending on the business model of each firm.
Business models exist today that were unheard of in 1985 due to the rise of the Internet and on-line
business activities. Various generic versions of the value chain have been suggested by different authors,
and some versions, such as the following, do not separate what might be considered support activities from
primary activities or may not include some activities considered support activities.

Customer
R&D Design Production Marketing Distribution Service

140 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

• Research is the search for knowledge that can be used to create new or improved products,
services, or processes. Development uses those research findings in the planning process to
improve the products, services, or processes that the company intends to sell or use internally.

• Design is the detailed planning and engineering for development efforts.

• Production is the acquisition of raw materials and the coordination and assembly required to
produce a product or deliver a service. The costs of production include direct materials, direct
labor, and factory overhead (inventoriable costs). In this version of a value chain, inbound logistics
are part of the production phase.

• Marketing includes advertising, promotion, and sales activities.

• Distribution is the delivery of products or services to customers. It is the “outbound logistics” of


Porter’s value chain, but it can also be considered a part of marketing and selling.

• Customer service includes customer support and warranty services after a sale.

Another version of a value chain segregates activities into primary activities and support activities similar
to the way Porter’s value chain did. However, there are a few differences.

Primary Activities

Product Marketing Customer


INPUTS R&D Production OUTPUTS
Design and Sales Service

Information Materials Human


Infrastructure
Systems Management Resources

Support Activities

In the above version of a value chain, materials management is a support activity, and materials man-
agement includes both inbound and outbound logistics. The logistics function manages the movement of
physical materials through the value chain. Materials management controls procurement (inbound logis-
tics), which includes finding vendors and negotiating the best prices. Materials management also controls
the movement of the procured materials through production and distribution (outbound logistics).

Information systems in the above value chain is a support activity. Information systems are the electronic
systems for maintaining records of inventory, sales, prices, and customer service. Information systems can
add value to the customer by tracking inventory and sales so that the company can provide the proper mix
of goods to customers while eliminating items that do not sell well.

The actual activities in an individual company’s value chain will depend on the company’s type of business.
In a service industry, the focus will be on marketing, sales, and customer service rather than on manufac-
turing and raw materials. In a technology business, the emphasis will be on research and development.
Furthermore, an activity might be considered a primary value chain activity in one firm and a support value
chain activity in another firm. However, all activities in the value chain contribute to creating value.

Note: In any organization, the more value a company creates, the more its customers will be willing to
pay for the product or service, and the more likely they are to keep on buying.

Regardless of the way a company views its value chain, all of the value chain activities contain opportunities
to increase the value to the customer or to decrease costs without decreasing the value to the customer by
reducing non-value-adding activities. Some examples are given below in the topic Value Chain Analysis.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 141
D.5. Business Process Improvement CMA Part 1

The analysis should take place at a relatively detailed level of operations, at the level of processes that are
just large enough to be managed as separate business activities.

Value Chain Analysis


Value chain analysis can help an organization gain competitive advantage by identifying the ways in which
the organization creates value for its customers. Value chain analysis identifies the steps or activities that
do and do not increase the value to the customers. Once those areas are identified, the organization can
maximize the value created by increasing the related benefits or reducing (even eliminating) non-value-
adding activities. The resulting increase in value to the customer or decrease in costs will make the company
more profitable and competitive.

The firm should analyze each process in its operations carefully to determine how each activity contributes
to the company’s operating income and its competitiveness. The goal of value chain analysis is to provide
maximum value to the customer for the minimum possible cost.

For internal decision-making to achieve the goal of providing maximum value at minimum cost, value
chain financial statements may be utilized. In value chain financial statements, all the costs for primary
activities are considered product costs that are allocated to products and inventoried, whereas all the costs
for support activities are considered period costs that are expensed as incurred. Value chain financial state-
ments are significantly different from conventional financial statements because many costs that are period
costs in conventional financial statements (research and development, product design, marketing and sales,
customer service) would be inventoriable costs in value chain financial statements. Value chain financial
statements do not conform to any Generally Accepted Accounting Principles, so their use is limited to in-
ternal decision-making.

Steps in Value Chain Analysis


Value chain analysis involves three steps, as follows.

1) Identify the activities that add value to the finished product. Which activities are value-
adding activities depends on the industry the company is in and what the company does (manu-
facturing, resale, service). They will be whatever activities this firm and firms in its industry
perform in the processes of designing a product or service, manufacturing the product, marketing
it and providing customer service after the sale.

2) Identify the cost driver or cost drivers for each activity.

3) Develop a competitive advantage by increasing the value to the customer or reducing the costs of
the activity. For example:

• Research and development can add value to established products or services by finding
ways to improve them, in addition to developing new ones.

• The function of production is to assemble the raw materials into finished goods. When pro-
duction is done efficiently, high quality products can be manufactured while costs are lowered,
leading to higher operating income.

• Marketing adds value by informing customers about the products or services, which may
increase the utility that customers attribute to the product or service and enable the company
to charge a higher price. Through marketing research, marketing can also discover what cus-
tomers want and need and can then communicate that to the R&D group so the R&D group
can design products that match the customers’ needs.

• Customer service after the sale adds value by delighting customers with the responsive ser-
vice received, thus creating superior value for them. Increased utility for customers because
of excellent customer service can also enable the company to charge more for its products.

142 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

• Efficiency in inbound logistics can significantly lower costs, thus creating value. Controlling
the flow of goods from suppliers into a retailer’s stores and ultimately to the consumer can
eliminate the need to carry large inventories of goods. Lower inventories lead to lower inven-
tory costs and greater value creation.

• Efficient procurement for a manufacturer can lead to lower manufacturing costs.

• Information systems can add value to the customer by tracking inventory and sales so that
the company can provide the proper mix of goods to customers while eliminating items that do
not sell well. Staying current with technological advances and maintaining technical systems
excellence are other sources of value creation.

Value chain analysis can also be used to determine what place a firm should occupy in a complete value
chain that consists of multiple businesses. The main concept in such a value chain analysis is that each firm
occupies a selected part or parts of the entire value chain. Which part or parts of the value chain to
occupy is determined by comparative advantage of the individual firm, or where the firm can best provide
value at the lowest possible cost. Some firms manufacture parts that they sell to other firms for assembly
into another product entirely. Then those products may be sold to a wholesaler, who sells them to a retailer,
who sells them to the ultimate consumer. Each of those manufacturers, sellers, and resellers occupies a
place in the value chain.

Question 46: The primary activities or business functions that are considered in the value chain include:

a) Customer service, production, marketing and sales, and human resources.

b) Customer service, production, marketing and sales, and information systems.

c) Infrastructure, human resources, materials management, and R&D.

d) Customer service, production, marketing and sales, and R&D.

(HOCK)

Question 47: An outside consultant has been hired by a manufacturing firm to evaluate each of the
firm’s major products beginning with the design of the products and continuing through the manufac-
ture, warehousing, distribution, sales, and service. The consultant has also been requested to compare
the manufacturer’s major products with firms that are manufacturing and marketing the same or similar
products. The consultant is to identify where customer value can be increased, where costs can be
reduced, and to provide a better understanding of the linkages with customers, suppliers, and other
firms in the industry. The type of analysis that the consultant most likely has been asked to perform
for the manufacturing firm is called a

a) Balanced scorecard study.

b) Benchmarking analysis.

c) SWOT (strengths, weaknesses, opportunities, threats) analysis.

d) Value-chain analysis.

(ICMA 2013)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 143
D.5. Business Process Improvement CMA Part 1

Process Analysis
A business process is a related group of activities encompassing several functions that produces a specific
product or service of value to a customer or customers. A business process is also a group of activities that
result in higher quality or lowered costs of a product or service.

For an insurance company, settling a claim is a business process. For just about any company, fulfilling a
customer’s order is a business process. Though manufacturing is a process, it is really a sub-process of
order fulfillment—something that must take place in order to fulfill an order. The sales function is also a
sub-process of order fulfillment, because it is also something that must take place before any order is
fulfilled.

Any process has inputs and it has outputs. Inputs are materials, labor, energy, and capital equipment. The
process transforms the inputs into outputs that have value to the customer. In classic economics terms,
inputs are economic resources such as land, labor, and capital. The output of a process may be a manu-
factured product, or it may be a service provided.

The challenge to a business is to make its processes work effectively and efficiently, in order to accomplish
the most possible with the least waste. Process analysis is used to understand the activities included in a
process and how they are related to each other. A process analysis is a step-by-step breakdown of the
phases of the process that conveys the inputs, outputs, and operations that take place during each phase
of the process. A process analysis is used to improve understanding of how the process operates. The
process analysis usually involves developing a process flowchart 18 that illustrates the various activities and
their interrelationships.

Once a process has been analyzed, the information gained from the analysis can be used to make operating
decisions. Sometimes a process needs improvement, such as eliminating waste or increasing operating
efficiency, but sometimes the process needs to be completely reengineered.

Business Process Reengineering


The term reengineering was originally used to refer to the practice of disassembling a product in order to
redesign it, but it more recently has been applied to the restructuring of organizational processes brought
about by rapidly changing technology and today’s competitive economy. For instance, instead of simply
using computers to automate an outdated process, technological advances bring opportunities to funda-
mentally change the process itself. In applying the concept of business process reengineering, management
starts with a clean sheet of paper and redesigns processes to accomplish its objectives. Operations that
have become obsolete are discarded.

The philosophy of “reengineering” business processes was set forth by Michael Hammer and James Champy
in their book, Reengineering the Corporation: A Manifesto for Business Revolution. Hammer and Champy
maintained that businesses are still being operated on the assembly line model that led to the industrial
revolution in the 20th century. Specialization in work assignments and the concept of division of labor that
was introduced back then brought about major productivity gains. Productivity increased immensely be-
cause workers doing the same things over and over again became very adept at doing those things and
their speed increased tremendously. Furthermore, specialization saved the time that was required for a
worker to move from one task to the next task. The assembly line model worked very well in the 1900s.

However, the authors say that the world has changed since the 1900s and companies operated today on
the same model no longer perform well. The underlying problem, according to the authors, is fragmented
processes, not only in manufacturing but in other business processes as well such as order fulfillment,
purchasing and customer service. The division of labor model has led to fragmented processes where one
department performs a part of the process and then hands the work off to the next department to do its

18
A flowchart is a diagram showing a sequence of work or a process. Flowcharts are used to analyze, design, document,
or manage a process or a program. Flowcharts use geometric symbols to show the steps and arrows to show the sequence
of the steps.

144 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

part. Although this division of labor is modeled on the assembly line that worked so well in the 1900s,
passing work around an organization today slows down the completion of the process and creates additional
costs without adding any value to the product or service the customer receives. In fact, because it wastes
time, passing work around inhibits the corporation’s ability to respond to customers’ needs in a timely
manner.

Business process reengineering involves analyzing and radically redesigning the workflow. Radical redesign
means throwing out the old procedures and inventing new ways of getting the work done. Reengineering
is not about making incremental improvements but it is about making quantum leaps.

Hammer and Champy recommend that in reengineering, the work should be organized around outcomes,
not tasks. In other words, management should think about what it wants to accomplish and then think of
ways to accomplish it rather than thinking of tasks to be done.

The processes in the organization should first be identified, then prioritized for reengineering according to
three criteria: (1) which processes are the most dysfunctional, (2) which will have the greatest impact on
customers, and (3) for which processes reengineering is most feasible.

The authors emphasize the use of technology, not to make old processes work better but to break the rules
and create new ways of working. One of the primary “rules” in the use of technology is that information
should be captured only once, and it should appear simultaneously every place it is needed. Taking infor-
mation from one system and inputting it into another is an indicator of a broken process that needs to be
reengineered.

People within the organization involved in reengineering typically include:

• A reengineering project leader who makes a particular reengineering project happen.

• The process owner, the manager with specific responsibility for the process being reengineered.

• The reengineering team, a group of people who diagnose the current process and oversee its
redesign and the implementation of the redesign.

• A reengineering steering committee, a group of senior managers who determine the organization’s
overall reengineering strategy and oversee it.

• A reengineering czar, a person with overall responsibility for developing reengineering techniques
to be used throughout the organization and for coordinating the activities of the company’s various
reengineering projects.

The reengineering process begins with the customer, not with the company’s product or service. The reen-
gineering team asks itself how it can reorganize the way work is done to provide the best quality and the
lowest-cost goods and services to the customer.

Frequently, the answer to the question is that the company’s value-chain activities could be organized more
effectively. For example, several different people in different areas might be passing work from person to
person to perform a business process. Instead, the same process might be able to be performed by one
person or one group of people all working closely together, at lower cost. Individual job assignments might
become more complex and thus more challenging, and the grouping of people into cross-functional teams
can both reduce costs and increase quality.

After the business process reengineering has been completed and the value-chain activities have been
reorganized to get the product to the final customer more efficiently, quality management takes over and
focuses on continued improvement and refinement of the new process.

Furthermore, internal controls for the reengineered process must not be neglected. When a process is
reengineered, its internal controls must be reengineered, as well. If existing internal controls are disassem-
bled and not replaced with new ones, the process will be without any controls.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 145
D.5. Business Process Improvement CMA Part 1

Benchmarking Process Performance


One of the best ways to develop the distinctive competencies that lead to superior efficiency, superior
quality, superior innovation, and superior responsiveness to customers, all of which confer com-
petitive advantage to a firm, is to identify and adopt best practices. Best practices can be accomplished
through benchmarking.

A benchmark is a standard of performance; and benchmarking is the process of measuring the organiza-
tion against the products, practices and services of some of its most efficient global competitors or against
those of other segments of the firm. The company can use these standards, also called best practices, as
a target or model for its own operations. Through the application of research and sophisticated software
analysis tools, companies undertake best practice analysis and then implement improvements in the
firm’s processes to match or beat the benchmark. The improvements could include cutting costs, increasing
output, improving quality, and anything else that will aid the firm in achieving its strategic business goals
and objectives.

Benchmarking continuously strives to emulate (imitate) the performance of the best companies in the world
or the best units in the firm, and through meeting these higher standards, the organization may be able to
create a competitive advantage over its marketplace competitors. The benchmarked company does not
need to be in the same industry as the company that is trying to improve its performance.

The first thing a company must do is to identify the critical success factors for its business and the
processes it needs to benchmark. Critical success factors are the aspects of the company’s performance
that are essential to its competitive advantage and therefore to its success. Each company’s critical success
factors depend upon the type of competition it faces.

After identifying its critical success factors, a team is set up to do best practice analysis. Best practice
analysis involves investigating and documenting the best practices for the processes used in performing
the firm’s critical success factor activities. The team members should be from different areas of the business
and have different skills. The team will need to identify what areas need improvement and how it will
accomplish the improvement by utilizing the experience of the benchmarked company.

Certain companies are generally recognized as leaders in specific areas. Some examples are Nordstrom for
retailing, Ritz-Carlton Hotels for service, and Apple Computer for innovation. The American Productivity and
Quality Center (www.apqc.org), a non-profit organization, is a resource for companies that want to do
benchmarking. APQC is one of the world’s leading proponents of business process and performance im-
provement. It maintains a large benchmarking database and offers firms several ways to participate in
benchmarking and to access benchmarking information.

Activity-Based Management (ABM)


Activity-based management (ABM) is closely related to and draws upon data from activity-based costing.
Activity-based costing uses activity analysis to develop detailed information about the specific activities the
company uses to perform its operations. Activity-based costing improves tracing of costs to products and
can even be used to trace costs to individual customers. It answers the question, “What do things cost”?

Activity-based management, on the other hand, is a means of understanding what causes costs to occur.
Activity-based management uses ABC data to focus on how to redirect and improve the use of resources
to increase the value creation.19 Activity-based management is a means of performing value chain analysis
and business process reengineering.

19
Institute of Management Accountants, Implementing Activity-Based Management: Avoiding the Pitfalls (Montvale, NJ:
Statements on Management Accounting, Strategic Cost Management, 1998), 2.

146 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

The IMA’s Statement on Management Accounting, Implementing Activity-Based Management: Avoiding the
Pitfalls, defines activity-based management as

“a discipline that focuses on the management of activities as the route to improving the
value received by the customer and the profit achieved by providing this value. ABM in-
cludes cost driver analysis, activity analysis, and performance measurement, drawing on
ABC as its major source of data.”20

Activity-based management uses activity analysis and activity-based costing data to improve the value
of the company’s products and services and to increase the company’s competitiveness.

Activity-based management is subdivided into operational ABM and strategic ABM.

Operational ABM uses ABC data to improve efficiency. The goal is for activities that add value to the
product to be identified and improved, while activities that do not add value are reduced in order to cut
costs without reducing the value of the product or service.

Strategic ABM uses ABC data to make strategic decisions about what products or services to offer and
what activities to use to provide those products and services. Because ABC costs can also be traced to
individual customers, strategic ABM can also be used to do customer profitability analysis in order to identify
which customers are the most profitable so the company can focus more on serving their needs.

The Concept of Kaizen


The term kaizen is a Japanese word that means “improvement.” As used in business, it implies “continuous
improvement,” or slow but constant incremental improvements being made in all areas of business opera-
tions. Small-scale improvements are considered to be less risky than a major overhaul of a system or
process. The slow accumulation of small developments in quality and efficiency can, over time, lead to very
high quality and very low costs. Kaizen needs to be a part of the corporate culture. It requires conscious
effort to think about ways that tasks could be done better. Continuous improvement can be difficult to
maintain and takes years to show results, but if done properly, it confers a sustained competitive advantage.

Kaizen principles can also be used for a “blitz,” in which substantial resources are committed to a focused,
short-term project to improve a process. A blitz usually involves analysis, design, and reengineering of a
product line or area. The results can be immediate and dramatic.

Kaizen can be used along with activity-based management and activity-based costing in a business process
reengineering project to improve the quality and reduce the cost of a business process.

A company may use target costing along with kaizen principles to determine what its ideal standard costs
are. Target costing puts the focus on the market because it starts with a target price based on the market
price. The market determines the target price, and the company must attain the target cost in order to
realize its desired gross profit margin for the product. The ideal standard is thus defined as the target cost,
or the standard cost that will enable the company to attain its desired cost and desired gross profit margin.
Using kaizen principles, the company figures out how it can manufacture the product at the target cost.
The standard is achieved through development of new manufacturing methods and techniques that entail
continuous improvement or the ongoing search for new ways to reduce costs.

The heart of the kaizen concept is the implementation of ideal standards and quality improvements. Kaizen
challenges people to imagine the ideal condition and strive to make the necessary improvements to achieve
that ideal.

Organizations can also apply kaizen principles to the budgeting process. Because kaizen anticipates contin-
uous improvements in the production process, a budget developed on kaizen principles will incorporate
planned improvements, resulting in decreasing costs of production over the budget period.

20
Ibid.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 147
D.5. Business Process Improvement CMA Part 1

Note: Kaizen is the Japanese term for improvement and it is used in business to mean continuous
improvement.

The Costs of Quality


Management will also be closely interested in the costs of quality. The costs of quality include not only
the costs of producing quality products, but they also include the costs of not producing quality products.
Over the long term, not producing a quality product is more costly than producing a quality product because
lack of quality causes loss of customers.

The costs of quality are classified in two categories, costs of conformance and costs of nonconform-
ance. Each of those two categories is broken down into two different types of costs.

• The costs of conformance are the costs to produce a quality product, and they consist of preven-
tion costs and appraisal costs.

• The costs of non-conformance are the costs of not producing a quality product, and they consist
of internal failure costs and external failure costs.

Costs of Conformance
The costs of conformance are the costs the company incurs to assess internal quality with the purpose of
insuring that no defective products reach the consumer.

The two types of costs of conformance are:

1) Prevention costs are the costs incurred in order to prevent a defect from occurring in the first
place. Prevention costs include:

• Design engineering and process engineering costs, so the design of the product is not defective
and the process for manufacturing it produces a quality product;
• Quality training, both internal programs and external training to teach employees proper man-
ufacturing procedures, proper delivery procedures, and proper customer service procedures,
including salaries and wages for employee time spent in training;
• Preventive maintenance on production equipment so an equipment failure does not cause a
quality failure;
• Supplier selection and evaluation costs to ensure that materials and services received meet
established quality standards, and costs to train suppliers to conform to the firm’s require-
ments;
• Evaluation and testing of materials received from a new supplier to confirm their conformance
to the company’s standards;
• Information systems costs to develop systems for measuring, auditing, and reporting of data
on quality; and
• Planning and execution costs of quality improvement programs such as Six Sigma21 or Total
Quality Management.22

21
Six Sigma is an approach to quality that strives to virtually eliminate defects. To achieve Six Sigma, a process must
produce no more than 3.4 defects per million opportunities. “Opportunities” refers to the number of opportunities for
nonconformance or for not meeting the required specifications. “Opportunities” are the total number of parts, compo-
nents, and designs in a product, any of which could be defective. If a product has 10,000 parts, components, and designs,
for example, 3.4 defects per million would amount to 34 products out of every 1,000 that would have some defect. The
goal of Six Sigma is to improve customer satisfaction by reducing and eliminating defects, which will lead to greater
profitability.
22
Total Quality Management is covered later in this topic.

148 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

2) Appraisal costs are the costs incurred to monitor production processes and individual products
and services before delivery in order to determine whether all units of the product or service meet
customer requirements. Appraisal costs include:

• Costs to test and inspect manufacturing equipment, raw materials received, work-in-process,
and finished goods inventories;
• Cost for equipment and instruments to be used in testing and inspecting manufacturing equip-
ment, raw materials, work-in-process, and finished goods inventories; and
• Costs for quality audits.

Costs of Nonconformance
Nonconformance costs are costs incurred after a defective product has been produced. The costs of non-
conformance can be broken down into two types:

1) Internal failure occurs when a problem is detected before a product’s shipment to the customer
takes place. The costs associated with internal failure are:

• Cost of spoilage and scrap;

• Costs (materials, labor, and overhead) to rework and reinspect spoiled units;

• Tooling changes and the downtime required to do the tooling changes to correct a defective
product;

• Machine repairs due to breakdowns;

• Engineering costs to redesign the product or process to correct quality problems and improve
manufacturing processes if the problem is detected before the product is in the customers’
possession;

• Lost contribution margin due to reduction of output caused by spending time correcting defec-
tive units; and

• Expediting costs, the cost of rushing to re-perform and complete an order in time because of
a failure to complete it correctly the first time.

2) External failure occurs when a defect is not detected until after the product is already with the
consumer. The costs of external failure are:

• Customer service costs of handling customer complaints and returns;

• Warranty costs to repair or replace failed products that are returned;

• Product recall and product liability costs, including settlements of legal actions;

• Public relations costs to restore the company’s reputation after a high-profile external failure;

• Sales returns and allowances due to defects;

• Lost contribution margin on sales lost because of the loss of customer goodwill; and

• Environmental costs such as fines and unplanned cleanup fees caused by a failure to comply
with environmental regulations.

Note: Make certain to know the four subcategories of the costs of quality and the individual items that
go into these four types of costs.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 149
D.5. Business Process Improvement CMA Part 1

Opportunity Costs of Quality


The nonconformance costs include opportunity costs associated with poor quality. An opportunity cost
is the benefit of the next best use of the resource that was lost.

Production of defective products creates opportunity costs in several ways.

• The company must either repair or replace defective products that fail within the warranty period
and products that may be recalled if the defect is serious. If the product is repaired, the company
must use resources to fix or replace the defective parts. The people the company pays to do the
repair work could have been producing new items for sale instead of spending their time on work
that will not produce any revenue, and the replacement parts used could have been used in new
items instead. If defective products cannot be repaired and must be replaced, the products used
to replace the defective ones are products that could have been sold but instead will not generate
any revenue even though their production creates costs.

• The company will need to provide more customer service after the sale, so it will need to pay more
customer service employees. The extra people spending their time on customer service for the
defective products could have been doing something else for the company that would be more
productive.

• The company may gain a reputation of supplying poor-quality products and could lose future sales
as a result. The lost future sales lead to lost profits. The lost resource is the cash profits that could
have been earned from the lost sales. If the company had that cash, it could invest it and earn
more future profits with it.

• Another opportunity cost associated with poor quality management concerns design quality fail-
ures. Costs of design quality are costs to prevent poor quality of design or costs that arise as a
result of poor quality of design. Design quality failure costs include the costs of designing, produc-
ing, marketing and distributing a poorly designed product as well as costs to provide service after
the sale for it. In addition, design quality failures can cause lost sales because the product is not
what customers expect and want. Design quality failures can be a significant component of design
quality costs.

Note: In manufacturing, it is often said that the main causes of quality problems are the “Four Ms”:
machines, materials, methods, and manpower.

Calculating the Costs of Quality


The costs of quality (conformance and nonconformance) can be quantified and documented on a cost of
quality (COQ) report. A cost of quality report shows the financial impact of implementing processes for
prevention and appraisal and for responding to internal and external failures.

Activity-based costing simplifies preparation of a COQ report, because an ABC system identifies costs with
activities. The costs of activities that are required to prevent or to respond to poor quality can be much
more easily identified when ABC is used than when a traditional costing system is used, because traditional
costing systems accumulate costs according to function (such as production or administration), rather than
according to activities. If traditional costing is being used, additional analysis is required to locate and
segregate the costs of quality.

Various formats can be used to report the costs of quality. The format used should depend on how man-
agement wants to see the costs of quality reported. Usually such a report would be done separately for
each product or product line. Following is one example of a cost of quality report in a spreadsheet format.

150 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

A B C D E F
1 Cost of Quality – Widgets - Month of May 20X2

2 Cost Allocation Base Total


Cost Allocation Rate Quantity Costs
3 Prevention costs:
4 Design and process engineering $65 per hour 528 hours $ 34,320
5 Equipment preventive maintenance $60 per hour 528 hours 31,680
6 Training $55 per hour 160 hours 8,800
7 Supplier selection/evaluation $40 per hour 88 hours 3,520
8 Testing of materials $25 per hour 352 hours 8,800
9 Total prevention costs $ 87,120
10
11 Appraisal costs:
12 Inspection of manufacturing equipment $35 per hour 352 hours $ 12,320

13 Equipment for inspection of raw materials $30 per hour 176 hours 5,280

14 Equipment for inspection of work-in-process $30 per hour 528 hours 15,840
15 Equipment for inspection of finished goods $30 per hour 176 hours 5,280
16 Total appraisal costs $ 38,720
17
18 Internal failure costs:
19 Cost of spoilage $75 per def. unit 50 def. units $ 3,750
20 Cost of rework $95 per rewkd. unit 80 rewkd. units 7,600

21 Est. lost contribution margin due to rework $60 per rewkd. unit 80 rewkd. units 4,800

22 Total internal failure costs $ 16,150


23
24 External failure costs:
25 Customer service-complaints & returns $35 per hour 176 hours $ 6,160
26 Warranty costs $110 per def. unit 35 def. units 3,850
27 Estimated product liability costs $25 per unit mfd. 3,300 Units mfd. 82,500
28 Contribution margin on estimated lost sales $230 per unit lost 300 Est. lost sales 69,000
29 Total external failure costs $161,510
30
31 Total costs of quality $303,500

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 151
D.5. Business Process Improvement CMA Part 1

Total Quality Management (TQM)


Total Quality Management describes a management approach that is committed to customer satisfaction
and continuous improvement of products or services. The basic premise of TQM is that quality improvement
is a way of increasing revenues and decreasing costs. As such, a company should always strive for im-
provement by performing its service or producing its product correctly the first time. Total Quality
Management is a prevention technique. The costs of implementing a TQM program are classified on a Cost
of Quality Report as prevention costs.

At the heart of TQM is the definition of what quality is. Quality can mean different things to different people.
For a customer it is a product that meets expectations and performs as it is supposed to for a reasonable
price. For a production manager it is a product that is manufactured within the required specifications.
When a company is considering quality, it must be certain to include the different perspectives of quality
from all of the involved parties.

Total quality management programs are often developed to implement an overall low-cost or a differenti-
ation business strategy, because TQM’s goals are to both reduce costs and improve quality. Approximately
90% of manufacturing companies and 70% of service businesses have put into practice some form of a
TQM program, but success or failure revolves around involvement and leadership of senior management.
A TQM program requires a change in corporate culture by eliminating faulty processes, empowering em-
ployees and creating teamwork that focuses on quality.

The objectives of TQM include:

• Enhanced and consistent quality of the product or service

• Timely and consistent responses to customer needs

• Elimination of non-value-adding work or processes, which leads to lower costs

• Quick adaptation and flexibility in response to the shifting requirements of customers

Certain core principles, or critical factors, are common to all TQM systems:

• They have the support and active involvement of top management

• They have clear and measurable objectives

• They recognize quality achievements in a timely manner

• They continuously provide training in TQM

• They strive for continuous improvement (kaizen)

• They focus on satisfying their customers’ expectations and requirements

• They involve all employees

TQM is an organizational action. For it to be successful, the entire organization must strive for quality
improvement and pursue excellence throughout the organization. Part of this pursuit of excellence is a focus
on continuing education. Employees at all levels participate regularly in continuing education and train-
ing in order to promote and maintain a culture of quality.

One of the unique perspectives of TQM relates to customers. In a TQM system, it is important to remember
that people within the organization are also customers. Every department, process, or person is at
some point a customer of another department, process, or person and at some point, a supplier to another
department, process, or person.

Another feature of TQM is quality circles. A quality circle is a small group of employees who work together
and meet regularly to discuss and resolve work-related problems and monitor solutions to the problems.
The form of communication that takes place in quality circles is vital to a successful TQM program.

In TQM, the role of quality manager is not limited to a special department. Instead, every person in the
organization is responsible for finding errors and correcting any problems as soon as possible.

152 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

To achieve total quality management, the company must identify the relevant quality problems when
and where they occur, and then use some of the following, or similar, methods to analyze them.

Statistical Quality Control (SQC)


Statistical quality control (SQC), or statistical process control (SPC), is a method of determining whether a
process is in control or out of control. Some variations in quality are expected, but if too many units are
tested and found to be outside the acceptable range, the process may not be in control. A control chart
is used to record observations of an operation taken at regular intervals. The sample is used to determine
whether all the observations fall within the specified range for the operation, and the intervals are meas-
urable in time, batches, production runs or any other method of delineating an operation.

When statistics are used in determining the acceptable range, the control chart is a statistical control
chart. For example, the acceptable range might be plus or minus two standard deviations from the mean.
If no sample falls outside the limit of two standard deviations, the process is in statistical control, as long
as all the samples are randomly distributed with no apparent patterns, and if the numbers of observations
that are above and below the center of the specified range are approximately equal. If instead trends,
clusters, or many measurements near the limits are observed, the process may be out of control even
though all of the observations are within two standard deviations of the mean.

Below is an example of a statistical control chart showing that all observations are within two standard
deviations of the mean, with no trends, no clusters, nor many measurements near the limits. For a company
that considers two standard deviations to be acceptable, the control chart below indicates that the process
is in control.

+3σ

+2σ

+1σ

−1σ

−2σ

−3σ

JAN FEB MAR APR MAY JUN JUL AUG SEP OCT NOV DEC

If any of the observations were above +2 standard deviations or below −2 standard deviations from the
mean, or if they indicated a trend going one way or another, or if several observations were clustered
together near the ±2 standard deviations, the process might not be in control.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 153
D.5. Business Process Improvement CMA Part 1

Histograms
A histogram is a bar graph that represents the frequency of events in a set of data. Patterns that may not
be apparent when just looking at a set of numbers become clear in a histogram, which can pinpoint most
of the problem areas. For instance, if a particular production line is experiencing most of the difficulty, a
histogram can help determine what types of problems are causing the problems most often.

Here is an example of a histogram for a manufacturer of wood bookcases. The company has evaluated the
various types of problems that occur in the manufacturing process, and the most frequent problems are:

• Wood cracked, 15 out of 100 defective units

• Trim not attached correctly, 19 out of 100 defective units

• Uneven stain, 38 out of 100 defective units

• Shelves out of alignment, 21 out of 100 defective units

• Shelf or shelves missing, 7 out of 100 defective units

The company develops the following histogram to illustrate its findings:

Defective Bookcases
40

35

30

25

20

15

10

0
Wood Trim not Uneven Shelves out Shelf or
cracked attached stain of shelves
corrrectly alignment missing

The histogram illustrates that the problem of uneven stain is by far the greatest problem occurring.

154 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

Pareto Diagrams
A Pareto diagram is a specific type of histogram. Vilfredo Pareto, a 19th-century Italian economist, came
up with the now well-known 80-20 observation, or Pareto principle. Someone relying on the Pareto prin-
ciple would state, for example, that “20% of the population causes 80% of the problems” or “20% of the
population is doing 80% of all the good things,” and such a statement would usually be fairly accurate.
After management pinpoints which 20% of the causes are accounting for 80% of the problems, it can focus
efforts on improving the areas that are likely to have the greatest overall impact.

In addition to showing the frequency of the causes for the quality problems with bookcases, a Pareto dia-
gram puts them in order from the most frequent to the least frequent. Furthermore, it adds a curve to the
graph showing the cumulative number of causes, going from the most frequent to the least frequent.

Below is an example of a Pareto diagram for the manufacturer of wood bookcases.

Defective Bookcases
100

90

80

70

60

50

40

30

20

10

0
Uneven Shelves out Trim not Wood Shelf or
stain of attached cracked shelves
alignment corrrectly missing

The curve on the graph illustrates the incremental addition to total defective bookcases contributed by each
cause.

The histogram and the Pareto chart both show graphically that uneven stain is the most frequently occurring
manufacturing problem. The company will be able to maximize its quality control efforts by giving first
priority to finding ways to prevent the uneven stain problem.

Once the most frequent quality problems have been identified, the next step is to identify the cause or
causes of each quality problem.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 155
D.5. Business Process Improvement CMA Part 1

Cause-and-Effect (Ishikawa) Diagram


A cause-and-effect diagram, or Ishikawa, diagram, organizes causes and effects visually to sort out root
causes and identify relationships between causes and their effects. The idea of a cause-and-effect diagram
was identified by Karou Ishikawa, who noted that it is often difficult to trace the many causes leading to a
single problem. Ishikawa developed a way of diagramming the relationships to better trace them. An Ishi-
kawa diagram consists of a spine, ribs and bones, so it is also called a fishbone diagram. The right end
of the spine is the quality problem, and the spine itself connects the main causes—the ribs—to the effect,
the quality problem.

Ishikawa Diagram

Usually, causes of problems in manufacturing fall into the following categories, referred to as the “4M”s:

• Machines

• Materials

• Methods

• Manpower

To resolve the problem of uneven stain on bookcases, operating personnel and management of the book-
case manufacturer hold a series of meetings, called “brainstorming sessions,” in which they attempt to
figure out what is causing the problem of the uneven stain on the bookcases, using the 4Ms. They find the
following:

Machines:

No equipment is used. The stain and protective coating are applied by hand.

Materials:

Wood: Maple, a hardwood, is used in bookcase construction. Although hardwoods should absorb stains
evenly, maple sometimes does not. A wood conditioner used before the stain would alleviate the problem,
but management has not wanted to incur the additional cost for the wood conditioner.

The maple sometimes is received from the lumberyard with machine marks in it that cause the stain to be
absorbed unevenly.

Methods:

The stain is applied with a brush. As more pressure is applied to the brush, more stain is applied to the
wood. If the pressure is not kept consistent, the stain will have variations in color. After the stain has been
applied, it is wiped off. The longer the stain is left on, the darker is the resulting color. If the employee does
not work quickly enough, the color on the bookcase will be lighter in the first areas wiped off and darker in
the last areas wiped off.

Manpower:

The maple is inspected before it is made into bookcases. If boards with machine marks are not caught in
this inspection and are not pulled from production, little can be done at the staining point to prevent a poor
stain job. Recently, a significant number of boards with machine marks have been missed in the inspection.

156 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

The employees do not always keep a consistent pressure while applying the stain, and they do not always
get the color wiped off quickly enough. Both of these things are causing variations in the color. Some of
the employees are inexperienced, and some are simply not following procedures.

Using input from the brainstorming sessions, the following cause-and-effect diagram is prepared:

Machines Materials

Wood conditioner
not being used
Not applicable-ma-
chines not used for
staining Machine marks in wood

Problem:
Uneven
Machine marks in wood Stain
High level of skill required missed in inspections

Highly labor-intensive Inexperienced employees

Employees not following


instructions

Methods Manpower
4)

The above cause-and-effect diagram can lead to changes made that may mitigate the problem of uneven
stain. Management may decide to use a spray-on stain that is easier to apply; it may decide to begin using
a wood conditioner on the maple or may change to using a different hardwood that will take the stain more
evenly; it may increase employee training programs in both the staining department and in the wood
inspection department; and it may look into automating portions of the process.

Total Quality Management and Activity-Based Management


Activity-based management principles are widely used to recommend process performance improve-
ments. They are applied to help companies make better strategic decisions, enhance operating
performance, reduce costs, and benchmark performance. Activity-based management uses activity-based
costing data to trace costs to products or individual customers for the purpose of analyzing business process
performance.

A total quality management system is most compatible with an ABC system because the ABC system
makes the costs of quality more apparent. A firm with a good ABC system needs only to modify the system
to identify costs and activities relating to costs of quality. A company that utilizes ABC will also be continu-
ously identifying activities that can be eliminated, as well as ensuring that necessary activities are carried
out efficiently.

Unnecessary and inefficient activities are non-value-adding activities, and the cost drivers of those ac-
tivities need to be reduced or eliminated. Activity to rework defective products is a non-value-adding activity
and thus needs to be reduced or eliminated. On the other hand, the company needs to carry out the value-
adding activities as efficiently as possible. A focus on TQM and its central tenet of continuous improve-
ment will lead to achieving both goals.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 157
D.5. Business Process Improvement CMA Part 1

ISO 9000
The International Organization of Standardization introduced the ISO 9000 quality assurance standards.
The standards do not have the force of law, but in both the U.S. and the European Union, foreign companies
in some industries must be ISO 9000 certified in order to do business there.

The ISO 9000 standards are not set to assure the quality of an individual product but to assure that
the quality is the same throughout all of the company’s product of that type. However, due to the
wide acceptance of the standards, a company that fails to achieve them runs the risk of losing business
due to perceived lack of quality.

Note: Adherence to (that is, following) design specifications is one of the most important components
of quality control. ISO standards do not address whether design specifications have been followed.

Quality Management and Productivity


At first glance, it may seem that as a company’s commitment to quality increases, the productivity of the
company will decrease. Since productivity is measured as the level of output given an amount of input, it
would seem that by allocating resources to quality and spending resources in the quality process, the level
of outputs related to a given level of inputs would decline. However, that usually does not occur.

In fact, as a company’s commitment to quality increases, usually productivity also increases. Reasons for
this increase in productivity include:

• A reduction in the number of defective units. The reduction in the number of defective units
in turn reduces the amount of time, material and effort wasted on unusable output as well as time
spent fixing salvageable defective units. (The term “the hidden factory” refers to the time and
effort spent on rework and repair.)

• A more efficient manufacturing process. As a result of an emphasis on quality production, the


company may remove or change inefficient, unproductive or non-value-adding activities.

• A commitment to doing it right the first time. This commitment is related to the first item, a
reduction in the number of defective units. As the culture in the company focuses on doing things
right the first time, the employees of the company take a more conscientious approach to their
work, and a more conscientious approach may lead to greater productivity.

No matter the cause, the relationship between quality and productivity is a positive one: the more attention
that is paid to quality, the higher are the levels of production.

Other Quality Related Issues


With the development of a good TQM system, a company can also manage its time better and become
more productive. In today’s environment, it is ever more important to become the first company to get a
new product or service to the marketplace. The need to be first leads to a need to have shorter product
development time and shorter response times to changes in demand or the market.

Customer-response time is the length of time between receipt of an order from a customer and the
receipt of the product by the customer. The components of cycle time are order receipt time (from receipt
of the customer’s order until the order is received by manufacturing), manufacturing cycle time (con-
sisting of waiting time before beginning manufacturing of the order and manufacturing time), and order
delivery time.

158 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

The following graphic is used in the topic Theory of Constraints to illustrate Manufacturing Cycle Efficiency
(MCE) and it illustrates the elements of customer-response time, as well.

Elements of Customer-Response Time

Receipt Manufacturing Delivery


Time Cycle Time Time

Waiting Manufacturing
Time Time

Order Received Order Received Manufacturing Finished Goods Order Delivered


From Customer By Manufacturing Begins Complete To Customer

Customer-Response Time

Only the actual manufacturing time is time spent on value-adding activities. Total customer-response time
can be decreased by decreasing receipt time, waiting time, and delivery time.

Question 48: Of these statements, which is not relevant to the overall benchmarking process?

a) Management determines what processes the company uses to convert materials, capital and labor
into value-adding products.

b) Management uses target costing to set standard costs in order to put the focus on the market as
well as on what target price to charge because the company must attain the target cost to realize
its desired profit margin for the product.

c) Target costing utilizes kaizen to reduce costs in order to attain a desired profit margin.

d) Management puts into place a program to measure the organization against the products, prac-
tices and services of the most efficient competitors in its marketplace.

(HOCK)

Question 49: Which of the following is not an objective of a company’s TQM program?

a) To examine issues relating to creating value for the company’s customers.

b) To understand the company’s capacity for innovation.

c) To evaluate environmental issues that may affect the company’s profitability.

d) To utilize computers in product development, analysis, and design modifications.

(HOCK)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 159
D.5. Business Process Improvement CMA Part 1

Question 50: The cost of scrap, rework and tooling changes in a product quality cost system are cate-
gorized as a(n):

a) External failure cost.

b) Training cost.

c) Prevention cost.

d) Internal failure cost.

(CMA Adapted)

Question 51: Listed below are selected line items from the Cost of Quality Report for Watson Products
for last month.

Category Amount
Rework $ 725
Equipment preventive maintenance 1,154
Product testing equipment 786
Product repair 695

What is Watson's total prevention and appraisal cost for last month?

a) $2,665

b) $1,154

c) $786

d) $1,940

(CMA Adapted)

160 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

Accounting Process Redesign


While accounting and finance professionals are recommending process improvements for other areas of the
organization, they need to look at their own processes, too, for ways to make accounting and financial
operations more efficient.

Businesses are under pressure to reduce the cost of business transactions. At the same time, the financial
function needs to focus more of its resources on activities that add greater value to the organization such
as decision support and strategic analysis. Therefore, accounting and finance need to be able to provide
the same or even higher levels of service in transaction processing while using fewer resources for trans-
action processing. The resources freed by improvements made can be put to new uses, enabling the
financial function to add value to the organization.

In order to accomplish these changes, the role of finance needs to be re-thought, and financial processes
need to be redesigned to do things better, faster, and cheaper, in line with the philosophy of continuous
improvement. By redeploying its resources to decision support and other value-adding activities, the finan-
cial function can do more for the organization, possibly at a lower cost.

Creating a Future Vision for Finance


In developing a future vision for the financial function, benchmarking studies should be used to identify
best practices being used. Much can be learned from organizations using best practices. The project team
can make best-practice visits to organizations that are functioning at high levels of efficiency and effective-
ness in the processes being considered for redesigning. Benchmarking enables the project team to develop
standards against which its vision for the finance function can be measured.

A current use assessment is another technique for creating a future vision. A current use assessment is
a customer-centered approach based on the idea that every aspect of the financial function should be
traceable to internal customer needs. Internal customers are the internal users of the financial outputs. The
project team needs to ask, “What value do we currently provide to users?” Current use assessment identi-
fies the outputs and activities that users depend on and that must remain a part of the redesigned
processes.

A current use assessment determines what reports are being used and by whom, how often updates are
needed, and what information in the reports is most essential. Surveys can be utilized to find out what
reports users use, why they use them, and whether they are satisfied or dissatisfied with them. The surveys
can also be used to gather information for a “wish list” of information the users would like to receive.

In addition, moving toward a decision-support model for finance requires an understanding of what type of
decisions are being made and how the finance function can support the efforts. Interviews and surveys can
help in gaining that understanding.

Selecting and Prioritizing Areas for Improvement


Every process cannot be redesigned at the same time, so it is necessary to prioritize. Priority should be
given to the processes most central to the organization’s strategy. Any area that has a significant impact
on the effectiveness of the business drives total performance and should receive top priority because those
areas are critical to the success of the organization. Processes that have a high probability of successful
redesign should also receive priority.

Process Walk-Throughs
Once the decision has been made to redesign a specific accounting or finance function and the future vision
has been created, the first step is to gather information about how things are currently being done. The
current processes, assumptions, and the structures of the function need to be examined in order to identify
improvements that can be made in efficiency and effectiveness. Process walk-throughs are used to
gather the necessary information.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 161
D.5. Business Process Improvement CMA Part 1

A walk-through is a demonstration or explanation detailing each step of a process. To conduct a process


walk-through, a member of the process redesign team meets with the process owner and each participant
in the process in order to gain a thorough understanding of how the work gets done from beginning to end
for the purpose of uncovering opportunities for improvement. A process owner is a person who has the
ultimate responsibility for the performance of a process and who has the necessary authority and ability to
make changes in the process.

The existing processes need to be thoroughly documented before they can be streamlined. Documentation
of a process involves more than simply listing the sequence of events in the process. In addition to listing
the steps performed, the documentation should include:

• Time required to complete each step and information about what may be slowing down each step,
so that improvement can be directed toward shrinking the tasks that require the greatest amount
of time rather than those that require little time.

• Wait time between processing steps in order to discover areas requiring improvement.

• Forms used as inputs in order to locate those with duplicate, unused, or missing information.

• Reports used as outputs to identify unneeded reports or where information on several reports can
be merged into fewer reports.

• Identification of who performs each step, because handing off the process from employee to em-
ployee can cause waiting times that could be eliminated by assigning the task steps to fewer
employees to reduce the number of work queues.

• Informal communication paths being used, so they can be included in the new formal process.

• Controls currently in place, in order to know in advance what controls will be affected by changes
made in the system.

• Where errors typically occur in the process, because correction of errors is time-consuming and
decreasing the number of errors can decrease the amount of time required for the process.

Every step, every piece of paper, and every input and output should be challenged. For example, investi-
gate:

• Why each step is being done, whether it is necessary, and whether it adds value.

• Whether the step could be automated and whether the physical pieces of paper being used are
necessary.

• Whether there may be duplication of effort.

• Whether the same input is being keyed in more than once, such as into a database and also into
a spreadsheet.

• The accuracy of the inputs and outputs.

Process mapping may be used to provide a visual map of the way information, documents, and work is
routed through a process. Process maps can pinpoint problem areas such as bottlenecks within the system,
places where reports are being prepared manually due to fragmentation of data, and rework.

Identification of Waste and Over-capacity


The process walk-throughs are a good starting point for identifying waste and over-capacity, such as:

• Duplication of effort.

• Tasks being done that are not necessary or that do not add value.

• Output that is not being used.

162 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

Determination of personnel under- or over-capacity can be done in various ways:

• One way of determining personnel needs is through metrics. For example, if an accounts payable
clerk can process 100 invoices for payment per day, historical totals of invoices processed can be
used to determine how many payables clerks are needed.

• Another way of determining personnel usage is changes in the size of work backlogs. If work
backlogs increase, personnel capacity may be too low. If work backlogs decrease, personnel ca-
pacity may be too high.

• Employees can be asked to review their own needs for more or fewer employees.

• If overtime is regularly excessive, then more employees may be needed.

Identifying the Root Cause of Errors


The walk-through can help to identify the root cause of errors.

• If the same input is being keyed in multiple times, not only is the duplication of effort wasteful,
but the multiple inputs carry the inherent risk that the data will be keyed in differently each time.
Changing the process so that the input is keyed in just once and establishing controls such as
reconciliations of data can eliminate the cause of errors.

• Each error regularly made, such as an error on an invoice or an error in an inventory count, is an
indication of a control weakness. Controls should be devised to prevent or detect the errors.

Process Design
Once the current process is fully understood, process design can take place in line with the vision for the
new process. The design process builds on the process concept developed during the vision step. Every
process is different, and a creative project team is required in order to generate a range of alternative
solutions. The redesigned process needs to cover every aspect of the internal customers’ (users’) needs.

Several potential process designs should be generated for consideration.

Risk-Benefit Evaluation
After several potential process designs have been generated and before one is selected, the potential de-
signs need to be weighed carefully in terms of their potential risk impact as well as their potential benefits.
The greater the changes being made, the less the organization can be sure of a successful outcome. Risk
can be the deciding factor in selecting one approach to process redesign over another one. If the risks of
all the options are determined to be too great, a return to the process design step may be necessary to
generate another set of alternative designs.

Planning and Implementing the Redesign


A complete redesign of a process or processes has the potential to be very disruptive, and it requires careful
planning. The initiative for finance redesign often comes from senior management, so it is a top-down
implementation. It requires engaging people in the change, providing leadership, supporting the change,
and planning the change.

If the changes are extensive, they will need to be phased in to allow the employees and the rest of the
organization time to adjust to their impact. The impact of the changes on the people, their jobs, their sense
of security, and their attitudes must be considered. All those involved in the change effort need to actively
seek to reduce the stresses put on the human resources of the company, or else the people will limit the
effectiveness of the project. The project team needs to move the people involved from denial and resistance
to the change to acceptance and commitment to the new way of doing things.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 163
D.5. Business Process Improvement CMA Part 1

Process Training
Redesigning processes requires finding new ways to use the skills of existing employees and to further
enhance those skills through training.

• Training will be needed in the changes that have been made to the process. Everyone involved in
the revised process should be included in the training.
• The need for skills enhancement training should be evaluated on an individual basis. Skills training
should be individualized to the needs of each employee to eliminate weaknesses in each em-
ployee’s skills. For example, certain employees may need to attend a class on using spreadsheets
whereas others would not need that but would need something else.
• Training may be needed in how to fully take advantage of the systems being used.

Reducing the Accounting Close Cycle


Significant improvements can be made in the time required to close the general ledger at the end of each
month, quarter, and fiscal year. “Soft closes” can be used for month-end closes while reserving the more
detailed closing activities and allocations for quarter-end and year-end closes.

• Accuracy at the point of data entry needs to be heightened so that reconciliations can be done
more quickly.
• Perpetual inventory records should be used, and inventories can be estimated for the soft closes.
• Timing-related accruals, such as inventories in transit, can be eliminated in the soft closes.
• The materiality level for consolidation and elimination entries can be raised.
• Use of a standardized chart of accounts and general ledger application across all company locations
is important for speed in closing.
• Review transactions and a preliminary set of financial statements for errors prior to the end of the
period.
• Bank reconciliations can be done daily without waiting for the month-end statement by accessing
transactions online.
• Valuation accounts for obsolete inventory and credit losses should be reviewed for updating in
advance of the period end.
• If recurring invoices are sent out on a regular schedule, they can be printed in advance by setting
the accounting software date forward so the revenue is recorded in the proper months.
• If expenses incurred or employee hours worked are billable to customers, review them before the
end of the period so the billing can be done quickly and accurately at the appropriate time.
• Interest expense can usually be accrued prior to the end of the month. Unless a large change in
debt principal takes place at the very end of the period, last-minute changes in debt will not make
much impact on the total interest expense accrued.
• For accruing unpaid wages, a centralized timekeeping system can be used so the most current
information about hours worked is available.
• Depreciation can be calculated a few days before the end of the period. Calculating the depreciation
early does entail the risk that a fixed asset would be acquired or disposed of after the calculations
are done and before the end of the period. However, for month-end closings, the missing depre-
ciation can be recorded during the following month, so the total depreciation expense for the year
will still be correct. For year-end closings, the depreciation can be adjusted if necessary.
• To speed up payables closing, require purchase orders for all larger purchases so the company has
a record of all outstanding orders of a material size. The accounting staff can access the information
to accrue for invoices not yet received as of month end.

164 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section D D.5. Business Process Improvement

By completing some of the closing work before the end of the period, the accounting staff is less rushed
and may make fewer errors, leading to more accurate as well as more timely financial statements.

In order to maintain consistency, a standard checklist of journal entries needed in the closing should be
maintained. The checklist should include a place to initial for each month when the entry has been done.
Journal entry templates23 should be stored in the accounting system so the entry is standardized and only
the amount needs to be keyed in. The checklist should give the name of the stored template to be used for
each set of closing entries.

When divisional accounting staffs send summarized information to the corporate accounting staff, the in-
formation can contain errors that the corporate accounting staff must investigate and correct, which can
be very time-consuming. Furthermore, different accounting procedures used throughout the company re-
sult in inconsistent reporting that is difficult to reconcile and consolidate. Mandating standardized
accounting procedures can improve accuracy at the source, resulting in less time spent by the corporate
accounting staff in correcting errors.

Centralization of Accounting as a Shared Service


Centralization of all accounting processes using a single consolidated accounting system is the best way to
resolve closing problems created by accounting decentralization. Shared service centers are increasingly
being used to move accounting and finance, as well as other major business processes, out of individual
business units and to concentrate them in a single location.

When the processing of transactions is centralized, specific types of transactions such as accounts payable,
accounts receivable, and general ledger can be organized along functional lines, utilizing a smaller number
of highly-trained people. The result is usually fewer errors. By reorganizing responsibilities along functional
lines instead of geographical lines, responsibility for various closing tasks can be assigned to a smaller
number of managers who are in closer proximity, resulting in greater efficiency.

In a centralized accounting system, accounting errors can be researched more easily because of having a
single database of accounting information. Analysts can more easily drill down through the data to locate
problem areas when the numbers do not look right instead of having to contact the divisional accounting
office and asking them to research it and then waiting to receive a response. The time required to locate
and resolve errors is shortened as a result.

However, transaction processing is not the only finance and accounting function being transferred to shared
services departments. Increasingly, higher-level functions such as controllership, tax, treasury, internal
audit, and compliance are also being centralized, often in offshore locations. Some finance and accounting
functions such as planning, budgeting, and forecasting and review and approval of period-end results usu-
ally remain close to the headquarters office for control reasons, though.

Use of Cloud-Based Services


Smaller companies that may not be able to justify the cost of an ERP system can turn to the cloud to
integrate finance, sales, service and fulfillment. All the essential information is in one place, and the result
can be a much faster close.

23
A journal entry template is a blank journal entry stored in the accounting software for which the same account
numbers are used repetitively but the amounts of the entries vary. Users go to the journal entry checklist to find the
name of the stored template, enter the numbers into the template, and save the journal entry with the correct date.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 165
Section E – Internal Controls CMA Part 1

Section E – Internal Controls


Section E comprises 15% of the CMA Part 1 Exam. Section E is composed of two parts: (1) Governance,
Risk and Compliance and (2) Systems Controls and Security Measures.

Internal control examines the controls the company has developed and implemented to help achieve its
objectives. The purpose of internal controls is often perceived as being to prevent something from going
wrong, but controls are really set up to assist the organization in the achievement of its objectives. It is
important to be very familiar with the objectives of internal control.

Other important topics are the major internal control provisions of the Sarbanes-Oxley Act of 2002 in
Sections 201, 203, 204, 302, 404, and 407 of the Act and the role of the PCAOB (Public Company Accounting
Oversight Board), which was established by the Sarbanes-Oxley Act.

Two of the most important elements of internal control that candidates need to understand are the segre-
gation of duties and the elements that make up the components of internal control. It is important
to know these topics and the other internal control topics not only from an academic standpoint (definitions
and lists, for example) but also from a practical application standpoint. The answers to the application-
related questions can be very difficult because it may seem that all of the choices are good controls or none
of the duties are ones that can be performed by the same person. However, do not spend too much time
thinking about any particular question because each question has the same value, and therefore candidates
have nothing to gain from figuring out the answer to a hard question versus answering a simple one. Answer
the simple questions first and spend extra time on the hard ones only if time allows.

In addition, ExamSuccess includes a lot of questions from past exams that have covered specific situations
relating to internal control and systems control. The specific topics covered on these past exam questions
may not be specifically mentioned in this textbook because of the vast scope of potential topics that would
need to be explained in order to do that. Instead, these types of questions are included in ExamSuccess.
Candidates do not need to remember every specific detail from each question, but they should be familiar
with the concepts and issues covered in those questions. Candidates are advised to learn the overall con-
cepts and issues and then apply their best professional judgment in answering questions about them. The
actual exam questions will not be the same as the practice questions in these study materials, since the
practice questions are previous exam questions. The actual exam questions are always being updated and
changed, so it is not likely that past exam questions will be asked again. For that reason, this textbook
does not try to “teach to the study questions” in this section. Instead, the information in this textbook is
focused on the topics covered in the ICMA’s current Learning Outcome Statements, as questions asked on
an exam today are more likely to be from the current Learning Outcome Statements than they are likely to
duplicate past exam questions.

Most of the internal control concepts covered in Governance, Risk and Compliance are adapted from the
report Internal Control – Integrated Framework developed by COSO, the Committee of Sponsoring Organ-
izations of the Treadway Commission. It is the guide for all internal control systems.

The second topic in this section is Systems Controls and Security Measures. The most important thing to
learn in Systems Controls and Security Measures is the terminology. Some candidates may already be
familiar with the terminology from work or experience with computer systems, but if not, it is important to
learn it.

166 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.1. Governance, Risk, and Compliance

E.1. Governance, Risk, and Compliance


The internal controls of a company are an important part of its overall operations. A strong internal control
system will provide many benefits to a company including:

• Lower external audit costs.

• Better control over the assets of the company.

• Reliable information for use in decision-making.

A company with weak internal controls is putting itself at risk for employee theft, loss of control over the
information relating to operations, and other inefficiencies in operations and decision-making that can dam-
age its business.

Corporate Governance
Good corporate governance is basic to internal control. The term “governance” will be used frequently in
this section of the textbook. What is corporate governance, why is it important, and how is it related to
internal control, risk assessment, and risk management?

What is Corporate Governance?


Corporate governance includes all of the means by which businesses are directed and controlled, including
the rules, regulations, processes, customs, policies, procedures, institutions and laws that affect the way
the business is administered. Corporate governance spells out the rules and procedures to be followed in
making decisions for the corporation. Corporate governance is the joint responsibility of the board of direc-
tors and management.

Corporate governance also involves the relationships among the various participants and stakeholders in
the corporation, such as the board of directors, the shareholders, the Chief Executive Officer (CEO), and
the managers.

Corporate governance is very concerned with what is known as the “agency problem.” Agency issues arise
from the fact that the owners of the corporation (the shareholders) and the managers of the corporation
(the agents of the shareholders) are different people. The priorities and concerns of the managers are
different from the priorities and concerns of the shareholders. The managers are concerned with what will
benefit them personally and lead to increased salary, bonuses, power, and prestige. The shareholders’
priorities lie with seeing the value of their investments in the corporation increase. The priorities of the
shareholders and the priorities of the managers can easily be in conflict with one another, because what
benefits the managers may not benefit the owners.

Therefore, corporate governance specifies the distribution of rights and responsibilities among the
various parties with conflicting priorities and concerns in an effort to mitigate the agency problem and bring
about congruence between the goals of the shareholders and the goals of the agents.24 Incentives are
needed so the agents will take actions that are consistent with shareholder benefit. At the same time,
however, monitoring mechanisms are needed to control any activities of the agents that would benefit them
while hurting the shareholders.

For example, management compensation policies that tie managers’ bonuses to stock price increases can
lead to actions on the part of management that will cause the stock price to increase and will thus be good
for all shareholders. However, if managers try to conceal poor financial performance in an effort to keep
the stock price going up so their own bonuses remain intact, those same incentives can lead to fraudulent
financial reporting, which obviously is not good for the shareholders or any other stakeholders. Prevention

24
Goal congruence is defined as “aligning the goals of two or more groups.” As used in business, it refers to the aligning
of goals of the individual managers with the goals of the organization as a whole.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 167
E.1. Governance, Risk, and Compliance CMA Part 1

of unintended consequences such as fraudulent financial reporting is the responsibility of the board of
directors and should be implemented through internal controls.

Why is Corporate Governance Important?


Corporate governance has always been an important topic for shareholders, management, and the board
of directors. However, the topic took on greater importance following the dramatic downfalls of companies
such as Enron, WorldCom, Adelphia and others back in 2001-02. More recently, the world financial crisis
that began in 2008 raised again the issue of good corporate governance. AIG (American International
Group) went from being the 18th largest public company in the world in 2008 to needing an $85 billion U.S.
government bailout. The Lehman Bros. bankruptcy in September 2008 was the largest bankruptcy in U.S.
history. The lesson from these events is that good governance is not just a U.S. issue but it is a global
issue.

Good governance is not just a good idea for a company—it is an absolute must. Considering just Enron,
more than $60 billion of shareholder wealth was erased from investors’ books. Good corporate governance
is not only important for company shareholders but it is vital for the general health and well-being of a
country’s economy as well.

Corporate governance does not exist as a set of distinct and separate processes and structures. It is inter-
connected with the company’s internal control and enterprise risk management.

How is Corporate Governance Related to Risk Assessment, Internal Control and Risk Management?
Corporate governance specifies the distribution of rights and responsibilities among the various participants
in the corporation.

• The board of directors and executive management are responsible for developing and implement-
ing business strategies.

• In setting business strategies, the board and executive management must consider risk.

• In order to consider risk, the company must have an effective process for identifying, assessing
and managing risk.

• In order to have an effective risk management process, the company must have an effective in-
ternal control system, because an effective internal control system is necessary in order to
communicate and manage risk.

Therefore, governance, risk management and internal control all rely on each other.

The internal audit activity serves as the “eyes and ears” of management and the audit committee and thus
has an important role in the governance function of the organization.

Internal audit’s primary role is assessing internal controls over the reliability of financial reporting, the
effectiveness and efficiency of operations, and the organization’s compliance with applicable laws and reg-
ulations. According to IIA (Institute of Internal Auditors) Internal Auditing Standard 2110, this primary role
of internal audit includes assessing and making appropriate recommendations for improving the governance
process in its accomplishment of the following objectives:

• Promoting appropriate ethics and values within the organization.

• Ensuring effective organizational performance, management and accountability.

• Communicating risk and control information to appropriate areas of the organization.

• Coordinating the activities of and communicating information among the board, external and in-
ternal auditors, and management.

168 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.1. Governance, Risk, and Compliance

Principles of Good Governance


A set of governance principles, called 21st Century Governance Principles for U.S. Public Companies, was
published in 2007 by a group of leading academic experts from four universities. The principles were de-
veloped by Paul D. Lapides, Joseph V. Carcello, Dana R. Hermanson and James G. Tompkins of Kennesaw
State University; Mark S. Beasley of North Carolina State University, F. Todd DeZoort of The University of
Alabama; and Terry L. Neal of University of Tennessee. The authors stated that the purpose of the principles
was “to advance the current dialogue and to continue to promote investor, stakeholder and financial state-
ment user interests.”

The principles are (some explanatory footnotes have been added):

1) Board Purpose – The board of directors should understand that its purpose is to promote and
protect the interests of the corporation’s stockholders while considering the interests of other ex-
ternal and internal stakeholders such as (e.g. creditors, employees, etc.).

2) Board Responsibilities – The board’s major areas of responsibility should be monitoring the CEO
and other senior executives, overseeing the corporation’s strategy and processes for managing the
enterprise, including succession planning; and monitoring the corporation’s risks and internal con-
trols, including the ethical tone.25 Directors should employ healthy skepticism26 in meeting their
responsibilities.

3) Interaction – Sound governance requires effective interaction among the board, management,
the external auditor, the internal auditor, and legal counsel.

4) Independence – An “independent” director has no current or prior professional or personal ties


to the corporation or its management other than service as a director. Independent directors must
be able and willing to be objective in their judgments. The vast majority of the directors should be
independent in both fact and appearance.

5) Expertise and Integrity – The directors should possess relevant business, industry, company,
and governance expertise. The directors should reflect a mix of backgrounds and perspectives and
have unblemished records of integrity. All directors should receive detailed orientation and contin-
uing education to assure they achieve and maintain the necessary level of expertise.

6) Leadership – The roles of Board Chair and CEO should be separate.27 If the roles are not separate,
then the independent directors should appoint an independent lead director. The lead director and
committee chairs should provide leadership for agenda setting, meetings, and executive sessions.

7) Committees – The audit, compensation and governance committees of the board should have
charters, authorized by the board, that outline how each committee will be organized, the com-
mittees’ duties and responsibilities, and how they report to the board. Each of these committees
should be composed of independent directors only, and each committee should have access to
independent outside advisors who report directly to the committee.

25
Companies need to make sure that inappropriate and unethical behavior is not tolerated. A culture of integrity is
dependent on the “tone at the top,” or the overall ethical climate that originates at the top of the organization with the
board of directors, the audit committee of the board, and the CEO.
26
“Healthy skepticism” means having an attitude of doubt but not carrying it so far as to suspect wrongdoing every-
where. It means asking questions, gathering information, and making a decision based on objective analysis. In the
present context, it means directors should not just accept without question the information they are given by manage-
ment but should “dig a little deeper” and find out the facts, because management may have “forgotten” to include some
of the facts. Directors are not present on a day-to-day basis and that makes their jobs more difficult than they would be
if the directors were on site. However, they can talk to people within the organization at all levels and ask questions,
and they should do that. Directors should not just assume that what they are being told is true or is the whole truth.
27
Not too long ago, the CEO frequently served also as Chairman of the board of directors, and the dual role was not
questioned. However, since the board’s responsibilities include monitoring the CEO, the CEO should not serve as Chair-
man of the board, because that creates a conflict of interest. The CEO would be leading the body that would be monitoring
the CEO.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 169
E.1. Governance, Risk, and Compliance CMA Part 1

8) Meetings and Information – The board and its committees should meet frequently for extended
periods of time and should have unrestricted access to the information and personnel they need
to perform their duties. The independent directors and each of the committees should meet in
executive session on a regular basis.

9) Internal Audit – All public companies should maintain an effective, full-time internal audit func-
tion that reports directly to the audit committee of the board of directors through the Chief Audit
Executive.28 Companies also should consider providing an internal audit report to external stake-
holders to describe the internal audit function, including its composition, responsibilities, and
activities.

10) Compensation – The compensation committee and full board should carefully consider the com-
pensation amount and mix (e.g., short-term vs. long-term, cash vs. equity) for executives and
directors. The compensation committee should evaluate the incentives and risks associated with a
heavy emphasis on short-term performance-based incentive compensation for executives and di-
rectors.

11) Disclosure – Proxy statements29 and other communications (required filings and press releases)
should reflect board and corporate activities and transactions in a transparent and timely manner
(e.g., financial performance, mergers and acquisitions, executive compensation, director compen-
sation, insider trades, related-party transactions). Companies with anti-takeover provisions should
disclose why such provisions are in the best interests of their shareholders.

12) Proxy Access – The board should have a process for shareholders to nominate director candi-
dates, including access to the proxy statement for long-term shareholders with significant
ownership stakes.

13) Evaluation – The board should have procedures in place to evaluate on an annual basis the CEO,
the board committees, the board as a whole, and individual directors. The evaluation process
should be a catalyst for change in the best interests of the shareholders.

Hierarchy of Corporate Governance


Corporate governance includes all of the means by which businesses are directed and controlled. It spells
out the rules and procedures to be followed in making decisions for the corporation.

Formation, Charter, and Bylaws


U.S. corporations are formed under authority of state statutes. Application for a charter must be made to
the proper authorities of a state in order to form a corporation. A business usually incorporates in the state
where it intends to transact business but it may be formed in one state, while at the same time have its
principal place of business or conduct its business operations in another state or states. A company that
wants to have its principal place of business located in a different state from its incorporation files with the
other state for a license to do business in that state, and it is known as a “foreign” (that is, “out of state”)
corporation in that state. The corporation will owe state income tax, state franchise tax, state sales taxes
and any other state taxes imposed on businesses not only to the state where it is incorporated, but also to
every state where it is licensed as a foreign corporation.

Although the means of organizing a corporation may vary from state to state to some extent, each state
usually requires that articles of incorporation (the charter) be filed with the secretary of state or another
designated official within that state.

28
If the company’s board of directors does not have an audit committee, then the full board serves as the audit com-
mittee, and the internal audit function, through the Chief Audit Executive, reports to the full board.
29
A proxy statement is a document containing the information the SEC requires publicly-held corporations reporting to
it to provide to shareholders so they can make informed decisions about matters that will be brought up at an annual
stockholders’ meeting.

170 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.1. Governance, Risk, and Compliance

The corporation’s charter, also referred to as its “Articles of Incorporation” or “Certificate of Incor-
poration,” details the following:

• The name of the corporation. In many states, the corporate name must contain the word “cor-
poration,” or “incorporated,” or “company,” or “limited,” or an abbreviation thereof. A corporate
name cannot be the same as, or deceptively similar to, the name of any other domestic corporation
or any other foreign corporation authorized to do business within the state.

• The length of the corporation’s life, which is usually perpetual (meaning forever).

• Its purpose and the nature of its business.

• The authorized number of shares of capital stock that can be issued with a description of the
various classes of such stock.

• Provision for amending the articles of incorporation.

• Whether or not existing shareholders have the first right to buy new shares.

• The names and addresses of the incorporators, whose powers terminate upon filing.

• The names and addresses of the members of the initial board of directors, whose powers com-
mence (begin) upon filing.

• The name and address of the corporation’s registered agent for receiving service of process and
other notices.

The persons who sign the articles of incorporation are called the incorporators. Incorporators’ services
end with the filing of the articles of incorporation, and the initial board of directors, named in the articles
of incorporation, takes over.

State laws typically require that incorporators be natural persons (citizens of the U.S.), and over the
age of eighteen. State laws vary as to the number of incorporators required, but in most states, only one
incorporator is required.

Note: A corporation, being itself a legal entity, may act as an incorporator.

Most states provide standardized forms for articles of incorporation. A corporation can use the standardized
form or file another form as long as it complies with state requirements. The articles of incorporation are
filed with the designated state official for such filings, ordinarily the secretary of state. A corporation is
usually recognized as a legal entity as soon as the articles of incorporation are filed or when the
certificate of incorporation is issued by the state. However, some states may also require additional
filings in some counties before the corporation is recognized as a legal entity.

After the articles of incorporation have been filed and the certificate of incorporation has been issued by
the state, the following steps must be carried out by the new corporation:

1) The incorporators elect the directors if they are not named in the articles,

2) The incorporators resign,

3) The directors meet to complete the organizational structure. At this meeting they:

a. Adopt bylaws for internal management of the corporation. The bylaws specify:

o The requirements for annual meetings of shareholders;

o Specifications regarding what constitutes a quorum at a shareholders’ meeting and what


constitutes a majority vote on the part of shareholders;

o Methods of calling special shareholders’ meetings;

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 171
E.1. Governance, Risk, and Compliance CMA Part 1

o How directors are to be elected by the shareholders, the number of directors and the length
of their terms, specifications for meetings of the board of directors and for what constitutes
a quorum at a board meeting;

o How officers are to be elected by the board of directors, officer positions and the responsi-
bilities of each officer position;

o How the shares of the corporation shall be represented (for example, by certificates) and
how shares shall be issued and transferred;

o Specifications for payments of dividends; and

o How the bylaws can be amended. The directors ordinarily have the power to enact, amend
or repeal bylaws, but this authority may instead be reserved to the shareholders. Bylaws
must conform to all state laws and specifications in the articles of incorporation.

Note: Employees are not legally bound by the bylaws unless they have reason to be familiar
with them.

b. Elect officers.

c. Authorize establishment of the corporate bank account, designate the bank, and designate by
name the persons who are authorized to sign checks on the account.

d. Consider for ratification any contracts entered into before incorporation.

e. Approve the form of certificate that will represent shares of the company’s stock.

f. Accept or reject stock subscriptions.

g. Comply with any requirements for doing business in other states. For example, if a corporation
files with another state as a foreign corporation located in that state, it will need to appoint a
registered agent in that state.

h. Adopt a corporate seal to be used for corporate documents for which a seal is required.

i. Consider any other business as necessary for carrying on the business purpose of the corpo-
ration.

Amending the Articles of Incorporation


Most state corporation laws permit amendment of the articles of incorporation. An example of an amend-
ment is an increase in the number of authorized shares of common stock.

Any amendment to the articles of incorporation must be something that could have been included in the
original articles of incorporation. Thus, an amendment is not allowed for something that the corporation
cannot legally do.

The board of directors usually adopts a resolution containing the proposed amendment, and then the res-
olution must be approved by a majority of the voting shares. After shareholder approval, the articles of
amendment are filed with the state authorities. The amendments become are effective only upon the issu-
ance of a certificate of amendment.

Although the articles of incorporation specify the name and address of the corporation’s initial registered
agent, changing the registered agent can usually be done by the board of directors without the need for
shareholder approval.

172 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.1. Governance, Risk, and Compliance

Responsibilities of the Board of Directors


The board of directors of a company is responsible for ensuring that the company is operated in the best
interest of the shareholders, who are the owners of the company.

Thus, the members of the board of directors represent the owners of the company. The board’s responsi-
bility is to provide governance, guidance and oversight to the management of the company. The board has
the following specific responsibilities:

• Selecting and overseeing management. The board of directors elects the officers of the company
and the board of directors is responsible for overseeing the activities of the officers they elect.
• Because it elects the company’s management, the board determines what it expects from man-
agement in terms of integrity and ethics and it confirms its expectations in its oversight activities.
• The board has authority in key decisions and plays a role in top-level strategic objective-setting
and strategic planning.
• Because of its oversight responsibility, the board is closely involved with the company’s internal
control activities.
• Board members need to be familiar with the company’s activities and environment, and they need
to commit the time required to fulfill their board responsibilities, even though they may be outside,
independent directors.
• Board members should investigate any issues they consider important. They must be willing to
ask the tough questions and to question management’s activities. They must have access to the
necessary resources to do investigations and must have unrestricted communications with all com-
pany personnel—including the internal auditors—as well as with the company’s external auditors
and its legal counsel.
• Because board members are responsible for questioning and scrutinizing management’s activities,
it is important for the board members to be independent of the company. An independent director
has no material relationship with the company. In other words, an independent director is not an
officer or employee of the company and thus is not active in the day-to-day management of the
company. Boards of companies that are listed on secondary securities markets such as the New
York Stock Exchange are required to consist of a majority of independent directors.

Most boards of directors carry out their duties through committees. Committees of the board of directors
are made up of selected board members and are smaller, working groups of directors that are tasked with
specific oversight responsibilities. One of the committees whose membership is prescribed by SEC regula-
tions is the audit committee. Other usual committees are governance, compensation, finance, nominating
and employee benefits. All of the committees of the board of directors are important parts of the company’s
internal control system, as their members can bring specific internal control guidance in their specific areas
of responsibility.

Audit Committee Requirements, Responsibilities and Authority


The responsibilities of the audit committee are particularly critical. The requirements for serving on an audit
committee of a publicly-held company have been formalized in law and regulations, first by the Sarbanes-
Oxley Act of 2002 and then, as directed by Sarbanes-Oxley, in SEC regulations. Secondary securities mar-
kets also include audit committee requirements in their listing regulations.

According to the New York Stock Exchange, the audit committee of the board of directors of a corporation
“stands at the crucial intersection of management, independent auditors, internal auditors and the board
of directors.” The audit committee of the board of directors is made up of members of the board of directors
who are charged with overseeing the audit function of the corporation. The audit committee members’ audit
committee responsibilities are in addition to their responsibilities as members of the larger board.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 173
E.1. Governance, Risk, and Compliance CMA Part 1

The SEC first recommended that boards of directors of corporations have audit committees in 1972. Within
short order, stock exchanges began requiring or at least recommending that listed companies have audit
committees. The responsibilities of audit committees have been increased over the years.

The Sarbanes-Oxley Act of 2002 increased audit committees’ responsibilities to a great degree. It also
increased the qualifications required for members of audit committees, and it increased the authority of
audit committees. In response to the Sarbanes-Oxley Act, the stock exchanges and the SEC developed new
rules and regulations for the purpose of strengthening audit committees.

Under Section 3(a)(58) of the Exchange Act, as added by Section 205 of the Sarbanes-Oxley Act, the audit
committee is defined as:

• A committee (or equivalent body) established by and amongst the board of directors of an issuer
for the purpose of overseeing the accounting and financial reporting processes of the issuer and
audits of the financial statements of the issuer; and

• If no such committee exists with respect to an issuer, the entire board of directors of the issuer.

Accordingly, the SEC’s final rule on audit committees for issuers of securities states that an issuer either
may have a separately designated audit committee composed of members of its board or, if it fails to form
a separate committee or if it chooses, the entire board of directors will constitute the audit committee.

Thus, the requirements for and responsibilities of members of audit committees of public companies’ boards
of directors are highly regulated.

Requirements for Audit Committee and Audit Committee Members


1) The audit committee is to consist of at least three members. This requirement is a listing require-
ment of the New York Stock Exchange and other stock exchanges. The Sarbanes-Oxley Act and
the SEC do not prescribe a minimum number of members for the audit committee but do state
that if the corporation does not form an audit committee, the entire board of directors will be
responsible for the audit committee function.
2) All members of the audit committee must be independent per Section 10A 3(b)(3) of the Secu-
rities Exchange Act of 1934 (15 U.S.C. 78f), as amended. This requirement means that the
members of the audit committee may not be employed by the company in any capacity other
than for their service as board members and on any committee of the board.
3) In addition, the New York Stock Exchange requires a five-year “cooling-off” period for former em-
ployees of the listed company or of its independent auditor before they can serve on the audit
committee of a listed company.
4) One member of the audit committee must be a financial expert. This requirement is made by
stock exchanges. The Sarbanes-Oxley Act requires that if the audit committee does not include a
financial expert, this fact must be disclosed.30
5) All members of the audit committee must be financially literate at the time of their appointment
or must become financially literate within a reasonable period of time after their appointment to
the audit committee. Financial literacy of all members of the audit committee is a listing require-
ment of the New York Stock Exchange and other stock exchanges.

30
The SEC defines the audit committee financial expert as an individual who the board of directors determines possesses
the following:
1) An understanding of financial statements and generally accepted accounting principles (GAAP)
2) An ability to assess the application of GAAP in connection with the accounting for estimates, accruals, and reserves
3) Experience preparing, auditing, analyzing, or evaluating financial statements that present a breadth and level of
complexity of accounting issues generally comparable to what can reasonably be expected to be raised by the issuer‘s
financial statements, or experience actively supervising those engaged in such activities
4) An understanding of internal control over financial reporting
5) An understanding of the audit committee‘s functions.

174 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.1. Governance, Risk, and Compliance

Responsibilities of the Audit Committee


1) The audit committee is responsible for selecting and nominating the external auditor, approving
audit fees, supervising the external auditor, overseeing auditor qualifications and independence,
discussing with the auditors matters required under generally accepted auditing standards, and
reviewing the audit scope, plan, and results.

2) The New York Stock Exchange’s Listing Manual requires that listed companies have an audit com-
mittee charter that “addresses the committee's purpose—which, at minimum, must be to: (A)
assist board oversight of (1) the integrity of the listed company's financial statements, (2) the
listed company's compliance with legal and regulatory requirements, (3) the independent auditor's
qualifications and independence, and (4) the performance of the listed company's internal audit
function and independent auditors; and (B) prepare an audit committee report as required by the
SEC to be included in the listed company's annual proxy statement.”

3) Rule 10A 3(b)(4) of the Securities Exchange Act specifies that “each audit committee shall establish
procedures for (A) the receipt, retention, and treatment of complaints received by the issuer re-
garding accounting, internal accounting controls, or auditing matters; and (B) the confidential,
anonymous submission by employees of the issuer of concerns regarding questionable accounting
or auditing matters.” This rule relates to the “whistleblower” 31 requirement in the Sarbanes-Oxley
Act.

4) In addition, the New York Stock Exchange specifically requires the following for companies listed
on the New York Stock Exchange (known as “listed companies”):

o The audit committee is to review the annual and quarterly financial statements and the MD&A
(Management Discussion and Analysis) in the company’s annual report filed with the SEC and
discuss them with management and the independent auditors.

o The audit committee is to meet periodically and separately with management and with internal
auditors and independent auditors in order to uncover issues warranting committee attention.

o The audit committee is to review with the independent auditor any audit problems or difficul-
ties, including any restrictions on the scope of the independent auditor's activities or on access
to requested information, and any significant disagreements with management and manage-
ment's response.

o The audit committee is to set clear hiring policies for employees or former employees of the
independent auditors, taking into account the pressures that may exist for auditors consciously
or subconsciously when seeking a job with a company they audit.

5) The Blue Ribbon Committee report recommended that the audit committee monitor the company’s
internal control processes, and most audit committees do such monitoring. They oversee the in-
ternal audit function and monitor internal control systems for compliance with legal and regulatory
requirements.

31
A “whistleblower” is a person who informs on someone else or makes public disclosure of corruption or wrongdoing.
Section 301 of the Sarbanes-Oxley Act mandated that audit committees of public companies establish a system for
receiving, retaining, and treating whistleblower complaints regarding accounting, internal controls, or auditing matters.
Public companies are required to establish a means for confidential, anonymous submission by employees and others
about concerns they may have regarding questionable accounting and auditing matters. Furthermore, Section 806 of
Sarbanes-Oxley authorizes the U.S. Department of Labor to protect whistleblower complainants against employers who
retaliate and also authorizes the U.S. Department of Justice to criminally charge those responsible for any retaliation.
Section 1107 of the Act makes it a crime for a person to knowingly retaliate against a whistleblower for disclosing truthful
information to a law enforcement officer regarding an alleged federal offense.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 175
E.1. Governance, Risk, and Compliance CMA Part 1

Authority and Funding of the Audit Committee


Rule 10A 3(b)(5) of the Securities Exchange Act provides that “each audit committee shall have the au-
thority to engage independent counsel and other advisers, as it determines necessary to carry out its
duties.”

Rule 10A-3(b)(6) of the Securities Exchange Act provides that “each issuer shall provide for appropriate
funding, as determined by the audit committee, in its capacity as a committee of the board of directors, for
payment of compensation (A) to the registered public accounting firm employed by the issuer for the pur-
pose of rendering or issuing an audit report; and (B) to any advisers employed by the audit committee
under paragraph (5).”

The audit committee has the authority to investigate any matter.

Responsibilities of the Chief Executive Officer (CEO)


The responsibilities of the CEO are determined by the corporation’s board of directors. A CEO’s responsibil-
ities and authority can be extensive, or they can be very limited, depending on how much authority and
responsibility the board of directors delegates to the CEO.

The CEO should not serve as chairman of the board of directors. Since the board’s responsibilities include
monitoring the CEO, the CEO should not serve as Chairman of the Board, because that creates a conflict of
interest. The CEO would be leading the body that would be monitoring the CEO.

Election of Directors
The shareholders elect the members of the board of directors. Usually, each share of stock is allowed one
vote, and usually directors are elected by a plurality (whoever gets the most votes is elected, even if it is
not a majority).

The length of the directors’ term of office is set in the corporate bylaws. The term is usually one year, but
it may be longer, such as three years in staggered terms, with one-third of the board members up for
election at each annual shareholders’ meeting. Holding staggered elections provides for continuity on the
board as each year some of the members leave the board while the remainder of the members return.

Note: Power for the board to increase its size without shareholder approval can be reserved in the
articles of incorporation or the bylaws of the corporation.

176 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internal Control

Internal Control
Who Cares About Internal Control?
Ever since commercial organizations, nonprofit organizations and governments have existed, their leaders
have recognized the need to exercise control in order to ensure that their objectives were achieved. Today,
however, the leaders of an organization are not the only ones who care about its internal control policies
and procedures.

• For a public company, information on the effectiveness of its internal control system is important
to investors to enable them to evaluate management’s performance of its stewardship responsi-
bilities as well as the reliability of its financial statements.

• The company’s external auditors recognize that an audit of a company with effective internal
controls can be performed more efficiently.

• The potential for U.S. corporations to make illegal payments to foreign governments is of concern
to legislative and regulatory bodies. Compliance with legislation prohibiting such activities is
addressed through internal control policies and procedures.

• The development of larger organizations with increased numbers of employees has made it nec-
essary for management to limit and direct employees’ authority and discretion through
internal control policies and procedures.

• Even customers have an indirect interest in internal controls because a strong internal control
system may reduce the costs of production, and therefore also reduce products’ prices.

Internal Control Definition


According to the COSO publication, Internal Control – Integrated Framework,32

Internal control is a process, effected by33 an entity’s board of directors, management, and
other personnel, designed to provide reasonable assurance regarding the achievement of ob-
jectives relating to operations, reporting, and compliance.

Thus, internal control is a process that is carried out (effected) by an entity’s board of directors, manage-
ment, and other personnel that is designed to provide reasonable assurance that the company’s
objectives relating to operations, reporting, and compliance will be achieved.

1) Operations objectives relate to the effectiveness and efficiency of operations, or the extent
to which the company’s basic business objectives are being achieved and its resources are being
used effectively and efficiently. Operations objectives include operational and financial perfor-
mance goals and safeguarding of assets against loss. A company’s operations objectives will vary
depending on the choices management makes about structure and performance. As part of the
objective-setting process, management should specify its risk tolerance. For operations objectives,
risk tolerance might be expressed as an acceptable level of variance related to the objective.

2) Reporting objectives pertain to internal and external financial and non-financial reporting.
Reporting objectives include reliability, timeliness, transparency, or other requirements as set forth
by regulators, recognized standard setters, or the entity’s policies. External reporting objectives
are driven by rules, regulations, and standards as established by regulators and standard-setters

32
Internal Control – Integrated Framework, copyright 1992, 1994 and 2013 by the Committee of Sponsoring Organiza-
tions of the Treadway Commission. Used by permission. The Committee of Sponsoring Organizations of the Treadway
Commission includes the following five organizations: American Institute of Certified Public Accountants (AICPA), Amer-
ican Accounting Association (AAA), Institute of Internal Auditors (IIA), Institute of Management Accountants (IMA), and
Financial Executives International (FEI).
33
To “effect” something means to cause it to happen, put it into effect, or to accomplish it. Thus, “effected by” means
“put into effect by” or “accomplished by.”

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 177
Internal Control CMA Part 1

external to the organization. Internal reporting objectives are driven by the entity’s strategic di-
rection and by reporting requirements established by management and the board of directors.

3) Compliance objectives relate to the organization’s compliance with applicable laws and reg-
ulations, encompassing all laws and regulations to which the company is subject. These laws and
regulations establish minimum standards of behavior and may include marketing, packaging, pric-
ing, taxes, environmental protection, employee safety and welfare, and international trade as well
as many others. For a publicly-held corporation or any corporation that reports to the SEC, com-
pliance objectives also include the SEC’s reporting requirements. A company’s record of compliance
or noncompliance with laws and regulations affects its reputation in its communities and its risk of
being the recipient of disciplinary procedures.

Note: The three categories of company objectives with which internal control is concerned—operations,
reporting, and compliance—are very important to know.

The three categories address different needs and they may be the direct responsibilities of different man-
agers. But every internal control should be directed toward the achievement of objectives in at least one
and possibly more than one of the three categories.

The three categories of objectives are distinct, but they do overlap. Therefore, a specific control objective
for a specific company could fall under more than one category. For example, the reporting objective of
ensuring reliable external financial reporting in accordance with accounting standards also concerns the
compliance objective of adhering to accounting standards as established by standard-setters and, for pub-
licly-held corporations, complying with the SEC’s reporting requirements in accordance with that body’s
regulations.

Fundamental Concepts
The definition of internal control reflects several fundamental concepts, as follows:

1) The purpose of internal control is to help the company achieve its objectives. The focus is on
achieving objectives. The objectives that internal control applies to fall into the three categories
above: operations, reporting, and compliance.

2) Internal control is an ongoing process. It is not something that can be done once and be com-
pleted. It is a journey, not a destination. It consists of ongoing tasks and activities. It is a means
to an end, not an end in itself.

3) Internal control is effected (accomplished) by people. It is something that must be put into effect
by people—it is not policies and procedures. People are located throughout the organization at
every level, from the members of the board of directors to the staff. Simply writing policy manuals
that call for internal control procedures is not enough. To be effective, people must put the policies
and procedures into effect.

4) Internal control procedures can provide reasonable assurance only—not absolute assurance
and not a guarantee—to the entity’s board of directors and senior management that the com-
pany’s objectives will be achieved in the three named areas. This statement reflects the
fundamental concepts that (1) the cost of an internal control system should not exceed the ex-
pected benefits, and (2) the overall impact of a control procedure should not hinder operating
efficiency.

5) Internal control must be flexible in order to be adaptable to the entity’s structure. Internal control
needs to be adaptable to apply to an entire entity or just to a particular subsidiary, division, oper-
ating unit, or business process.

178 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internal Control

The Importance of Objectives


Since internal control’s purpose is to provide reasonable assurance regarding the achievement of objectives
relating to operations, reporting, and compliance, it stands to reason that internal control cannot operate
effectively unless objectives have been set. Setting objectives is part of the strategic planning process by
management and the board of directors. Objectives should be set with consideration given to laws, regula-
tions, and standards as well as management’s choices. Internal control cannot establish the entity’s
objectives.

Who Is Responsible for Internal Control?


The board of directors is responsible for overseeing the internal control system. The board’s oversight
responsibilities include providing advice and direction to management, constructively challenging manage-
ment, approving policies and major transactions, and monitoring management’s activities. Consequently,
the board of directors is an important element of internal control. The board and senior management es-
tablish the tone for the organization concerning the importance of internal control and the expected
standards of conduct across the entity.

The CEO is ultimately responsible for the internal control system and the “tone at the top.” The CEO
should provide leadership and direction to the senior managers and review the way they are controlling the
business. The “tone at the top” (part of the control environment) is discussed in more detail below.

Senior managers delegate responsibility for establishment of specific internal control policies and pro-
cedures to personnel responsible for each unit’s functions.

Financial officers and their staffs are central to the exercise of control, as their activities cut across as
well as up and down the organization. However, all management personnel are involved, especially in
controlling their own units’ activities.

Internal auditors play a monitoring role. They evaluate the effectiveness of the internal controls estab-
lished by management, thereby contributing to their ongoing effectiveness.

Virtually all employees are involved in internal control, because all employees produce information used
in the internal control system or carry out other activities that put the internal control systems into effect.
Furthermore, all employees are responsible for letting their managers know if they become aware of prob-
lems in operations or that rules, regulations or policies are being violated.

External parties provide information that is useful to effective internal control. For example, independent
auditors audit the financial statements and often provide other useful information as well to management
and the board. Other external parties that may provide useful information include legislators, regulators,
customers, financial analysts, bond rating agencies and the news media. However, external parties are
not part of the company’s internal control system, and they are not responsible for it.

Note: Internal auditors evaluate the effectiveness of the control systems and contribute to their ongoing
effectiveness, but they do not have responsibility for establishing or maintaining the control systems.

Note: Internal control should be an explicit or implicit part of everyone’s job description.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 179
Internal Control CMA Part 1

Components of Internal Control


According to the COSO report, Internal Control – Integrated Framework (2013 update), five interrelated
components comprise internal control. If the five components are present and functioning effectively, their
effective functioning provides reasonable assurance regarding achievement of the company’s objectives.
Thus, these components are all necessary for effective internal control to be present. They are:

1) Control Environment

2) Risk Assessment

3) Control Activities

4) Information and Communication

5) Monitoring Activities

Embedded within these five components are 17 principles.

Component 1: Control Environment


The control environment includes the standards, processes, and structures that provide the foundation for
carrying out internal control. The board of directors and senior management are responsible for establishing
the “tone at the top,” including expected standards of conduct that apply to all employees. Management is
responsible for reinforcing the expectations at all levels of the organization.

The control environment provides the organization’s ethical values. It influences the control consciousness
of all the people in the organization and sets the tone for the entire organization. If the control environment
does not include the necessary factors, none of the other components of internal control will be effective.

Organizations with effective control environments have the following characteristics, exemplified by these
five principles:

1) They demonstrate a commitment to integrity and ethical values. They set a positive “tone at
the top” by communicating, verbally and by example as well as formally, the organization’s ethical
values and commitment to integrity.
Every company should establish standards of conduct and formal policies regarding acceptable
business practices, conflicts of interest, and expected standards of behavior. However, these offi-
cial statements state only what management wants to have happen. Corporate culture, or the
“tone at the top,” determines what actually does happen. Top management, especially the CEO,
sets the ethical tone by modeling the ethical and behavioral standards expected of everyone in the
organization. Leadership by example is the most effective means of communicating that ethical
behavior is expected, because people imitate their leaders.
Management should foster a “control consciousness” by setting formal and clearly communicated
policies and procedures to be followed at all times, without exception, and that result in shared
values and teamwork.
Standards of integrity and ethical values extend to outsourced service providers and business
partners, as well. Management retains responsibility for the performance of processes it has dele-
gated to outsourced providers and business partners.

Processes should be in place to identify issues and evaluate the performance of individuals and
teams against the expected standards of conduct, and deviations need to be addressed in a timely
and consistent manner. The actions taken by management when violations occur send a message
to employees, and that message quickly becomes a part of the corporate culture.
2) The board of directors demonstrates independence from management and exercises
oversight over development and performance of internal control. The board of directors is re-
sponsible for setting corporate policy and for seeing that the company is operated in the best

180 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internal Control

interest of its owners, the shareholders. The attention and direction provided by the directors are
critical.

The board of directors should have a sufficient number of members who are independent from
management (not employed full-time by the company in management positions) to be independent
and objective in its evaluations and decision-making. Independence of the board from manage-
ment is critical, so that if necessary, difficult and probing questions will be raised.

Board and audit committee members should hold regular meetings with chief financial and ac-
counting officers and internal and external auditors. Sufficient and timely information should be
provided to board and audit committee members.

The board of directors has oversight responsibility for internal control, but the Chief Executive
Officer and senior management have direct responsibility for developing and implementing the
organization’s internal control system.

3) With the oversight of the board, management establishes structures, reporting lines, and
appropriate authorities and responsibilities to enable the corporation to pursue its objectives.

The company’s organizational structure should define the key areas of authority and responsibility
and delineate lines for reporting. The organizational structure is key to the company’s ability to
achieve its objectives, because the organizational structure provides the framework for planning,
executing, controlling and monitoring the activities it pursues to achieve its objectives.

Authority and responsibility should be delegated to the extent necessary to achieve the organiza-
tion’s objectives. The board of directors delegates authority and assigns responsibility to senior
management. Senior management delegates authority and assigns responsibility at the entity level
and to its subunits. The way management assigns authority and responsibility for operating activ-
ities affects the control environment because it determines how much initiative individuals are
encouraged to use in solving problems as well as the limits of their authority.

Delegation of authority means giving up centralized control of some of the business decisions and
allowing those decisions to be made at lower levels in the organization by the people who are
closest to the day-to-day operations of the business. Delegation of authority provides the organi-
zation with greater agility, but it also introduces complexity in risks to be managed. Senior
management with guidance from the board of directors needs to determine what is and is not
acceptable, in line with the organization’s regulatory or contractual obligations.

The challenge is to delegate only to the extent required to achieve the organization’s objectives.
The delegation should be based on sound practices for identifying and minimizing risk and on
weighing potential losses against potential gains from delegation.

4) The organization demonstrates a commitment to attract, develop, and retain competent in-
dividuals in alignment with objectives. In order for tasks to be accomplished in accordance
with the company’s objectives and plans for achievement of those objectives, the company needs
to have competent personnel. In order to have competent personnel, management should specify
the knowledge and skills required for each position. Formal or informal job descriptions should
specify the competence level needed for each job, and the company should make every effort to
hire and retain competent people and to train them when necessary.
Background checks should be thorough when hiring new employees. At a minimum, the applicant’s
work history and education should be confirmed and references checked. Any embellishment or
undisclosed history should be a red flag.

Individuals who are working in positions for which they are unqualified create a risk simply because
they are not capable of adequately performing the work they are supposed to do. Their lack of
capability provides an opportunity for someone else to take advantage of their lack of knowledge
or skills and perpetrate a fraud. Therefore, appropriate personnel policies and procedures are in-
tegral to an efficient control environment.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 181
Internal Control CMA Part 1

The board of directors should evaluate the competence of the CEO, and management should eval-
uate the competence across the organization and within outsourced providers in relation to
established policies and procedures and then act as necessary to address any shortcomings.

5) The organization holds individuals accountable for their internal control responsibilities in
pursuit of objectives.

The board of directors holds the CEO accountable for understanding the risks faced by the organ-
ization and for establishing internal controls to support the achievement of the organization’s
objectives. The CEO and senior management are responsible for establishing accountability for
internal control at all levels of the organization.

Increased delegation requires personnel to have a higher level of competence and requires the
company to establish accountability. Management should monitor results because the number of
undesirable or unanticipated decisions may increase with increased delegation. The extent that
individuals recognize that they will be held accountable for results greatly affects the control envi-
ronment. If a person does something that violates the company’s policies and standards, some
sort of disciplinary action should be taken against that person. If no penalty is assessed for violating
the company’s internal control policies, then other individuals will not see the need for compliance.

Management should regularly review the organization’s performance evaluation methods and in-
centives to ensure that they do not encourage inappropriate conduct. If increases in the bottom
line are the sole focus of performance evaluations, the organization is more likely to experience
unwanted behavior such as manipulation of accounting records and reports, offers of kickbacks,
and high-pressure sales tactics.

Internal controls are more likely to function well if management believes that the controls are important
and communicates that support to employees at all levels.

Component 2: Risk Assessment


Risk is the possibility that something will occur that will adversely affect the organization’s achievement of
its objectives. Within the control environment, management is responsible for the assessment of risk. The
questions should always be asked: “What could go wrong”? “What assets do we need to protect”?

Risk assessment is the process of identifying, analyzing, and managing the risks that have the potential to
prevent the organization from achieving its objectives, relative to the organization’s established risk toler-
ance. Assessment of risk involves determining the monetary value of assets that are exposed to loss as
well as the probability that a loss will occur. Management must determine how much risk it is willing to
accept and then work to maintain the risk within that level.

The 17 principles within the five components of Internal Control – Integrated Framework (2013) are num-
bered consecutively, so the principles relating to the risk assessment component begin with no. 6:

6) The company’s objectives must be specified clearly enough so that the risks to those ob-
jectives can be assessed. Objective setting is therefore the first step in management’s process
of risk assessment. Objectives may be explicitly stated or they may be implicit, such as to continue
a previous level of performance. Setting the company’s operations, reporting, and compliance
objectives is a strategic planning function of management.

Establishing objectives is a required first step to establishing effective internal control because it
forms the basis for assessing risk, in other words determining what could go wrong that could
prevent the company from achieving its objectives. If the objectives are not known, then it is not
possible to determine what could prevent the company from achieving them.

7) The organization identifies risks to the achievement of its objectives and analyzes them to
determine how the risks should be managed. The responsibility for risk identification and analysis
resides with management at the overall entity level and at the subunit level.

182 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internal Control

Risks can arise from both internal and external factors that can affect the company’s ability to
achieve its objectives. Change in objectives creates risk, especially. The greater the difference in
the current objectives from objectives of the past—the greater the amount of change—the greater
the amount of risk. Even the objective of maintaining performance as it has been in the past carries
both internal and external risks.

The risk assessment process should consider all risks that may occur. The risk assessment should
be comprehensive and consider all significant interactions between the company and external
parties, throughout the organization. External parties to include in the assessment are suppliers
(current and potential), investors, creditors, shareholders, employees, customers, buyers, inter-
mediaries, competitors, public bodies and the news media.

Once the risks have been identified, they should be analyzed in order to determine how best to
manage each one.

Risk Identification

Risk identification includes identification of risks at all levels of the organizational structure, includ-
ing the overall entity and its subunits. Entity level risk identification is conducted at a high level
and does not include assessing transaction-level risks. The identification of risks at the process
level is more detailed and includes transaction-level risks. Risks originating in outsourced service
providers, key suppliers, and channel partners need to be included in the risk assessment, as they
could directly or indirectly impact the organization’s achievement of its objectives.

Entity-level risks arise from external or internal factors. Following are a few examples:

• External factors include economic factors that impact financing availability, environmental
factors such as climate change that can require changes in operations, reduced availability of
raw materials, or loss of information systems, regulatory changes such as new reporting stand-
ards or new laws, changing customer needs, and technological changes that affect the
availability and the use of data.

• Internal factors can include decisions that affect operations and the availability of infrastruc-
ture, changes in management responsibilities that affect the way controls are implemented,
changes in personnel that can influence the control consciousness in the organization, em-
ployee access to assets that could contribute to misappropriation of assets, and a disruption in
information systems processing that could adversely affect the organization’s operations.

Transaction-level risks occur at the level of subsidiaries, divisions, operating units, or functions
such as sales, production, or marketing. The potential risks depend upon what the objectives are.
For example,

• An operational objective of maintaining an adequate raw materials inventory could lead to


identifying risks such as raw materials not meeting specifications, the failure of a key supplier,
supply disruptions in needed raw materials caused by weather conditions, or price increases
above acceptable levels.

• An objective of complying with existing laws and regulations leads to identifying risks associ-
ated with lack of compliance.

• The objective of protecting assets leads to identifying the risk of employee embezzlement ac-
companied by falsification of records to conceal the theft.

The number of potential risks is limitless, and practical limitations are needed in identifying them.
Some risks, such as a meteor falling out of the sky onto the company’s manufacturing facility, are
too minimal to be considered. But any situation that causes a change that could impact the system
of internal control should be included.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 183
Internal Control CMA Part 1

Risk Analysis

Risk analysis forms the basis for determining how the risks will be managed. After the company
has identified its entity-level and transaction-level risks, it should perform a risk analysis to (1)
assess the likelihood or frequency of each risk’s occurring; (2) estimate the impact of each risk;
and (3) consider how each risk should be managed by assessing what actions need to be taken.

Risks that do not have a significant impact on the company and that have a low likelihood of
occurring would not warrant serious concern. However, risks with a high likelihood of occurring
and that carry the possibility of significant impact usually require serious attention. Risks that are
in between these two extremes require judgment.

Once the likelihood and estimated impact of risks have been assessed, the following steps should
be taken to manage the identified risks. Risk responses fall into the following categories:

• Acceptance – No action is taken to affect the likelihood or impact of the risk.

• Avoidance – Exiting the activity or activities that give risk to the risk, such as exiting a product
line or selling a division.

• Reduction – Action is taken to reduce the likelihood or impact of the risk. The amount of the
potential loss from each identified risk should be estimated to the extent possible. Some risks
are indeterminate and can be described only as large, moderate or small.

• Sharing – Reducing the risk likelihood or impact by transferring or sharing the risk such as
purchasing insurance or forming a joint venture.

If the decision is to reduce or to share the risk, the organization determines what action to take
and develops appropriate control activities for the action. If the decision is to accept or avoid the
risk, typically no control activities are required.

8) In assessing the risks to the achievement of its objectives, management of the organization should
consider the potential for fraud. Fraud can include fraudulent financial reporting, possible loss
of assets, and corruption. Fraud can occur at any level and its possible impact needs to be consid-
ered as part of the risk identification and assessment. The potential for management fraud through
override of controls needs to be considered, and the directors’ oversight of internal control is nec-
essary to reduce that risk. Fraud can arise at the employee level, as well, for example if two
employees collude34 to defraud the organization. Furthermore, fraud can be perpetrated from the
outside, from someone hacking into the computer systems for example.

When management detects fraudulent financial reporting, inadequate safeguarding of assets, or


corruption, remediation may be necessary. In addition to dealing with the improper actions, steps
may need to be taken to make changes in the risk assessment process and in other components
of the internal control system such as control activities.

9) The organization identifies and assesses changes that could impact the organization’s sys-
tem of internal control.

Changes can occur in the external environment, such as in the regulatory, economic, and physical
environment in which the entity operates. Changes can also occur in the internal environment such
as new product lines, acquired or divested businesses and their impact on the internal control
system, rapid growth, changes in leadership and their attitudes toward internal control.

Note: Risk assessment, which is a function of internal control, is different from the actions taken by
management to address the risks. The actions taken by management to address the risks are a function
of management and not of the internal control system.

34
To “collude” is to act together, often in secret, to achieve some illegal or improper purpose.

184 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internal Control

Component 3: Control Activities


Control activities are actions established by policies and procedures that help ensure that management’s
instructions intended to limit risks to the achievement of the organization’s objectives are carried out.

Control activities may be preventive or detective and can include a range of activities such as authorizations
and approvals, verifications, reconciliations, and business performance reviews. Segregation of duties is
typically built in to the selection and development of control activities.

Principles relating to control activities include:

10) The organization selects and develops control activities that contribute to reducing to ac-
ceptable levels the identified risks to the achievement of its objectives.

Control activities should be integrated with risk assessment in order to put into effect actions
needed to carry out risk responses. The risk responses decided upon dictate whether control ac-
tivities are needed or not. If management decides to accept or avoid a risk, control activities are
generally not required. The decision to reduce or share a risk generally makes control activities
necessary. The control activities include a variety of controls including both manual and automated
and preventive and detective controls.

Segregation of duties should be addressed wherever possible and if segregation of duties is not
practical, management should develop alternate control activities.

A preventive control is designed to avoid an unintended event while a detective control is


designed to discover an unintended event before the ultimate objective has occurred (for example,
before financial statements are issued or before a manufacturing process is completed).

o Examples of preventive controls are segregation of duties, job rotation, enforced vacations,
training and competence of personnel, employee screening practices, physical control over
assets such as dual access controls, requirements for authorizations, and requirements for
approvals.

o Examples of detective controls are reconciliations, internal audits, periodic physical inventory
counts, variance analyses to detect ratios that might be out of line, random surprise cash
counts, supervisory review and approval of accounting work, management review and approval
of account write-offs, and exception reporting and review to identify unusual items.

11) The organization selects and develops general control activities over technology 35 to sup-
port the achievement of its objectives.

o Control activities are needed when technology is embedded in business processes to mitigate
the risk of the technology operating improperly.

o Control activities may be partially or wholly automated using technology. Automated controls
require technology general controls.

Control activities over technology designed to support the continued operation of technology and
to support automated control activities are called “technology general controls.” Technology gen-
eral controls include controls over the technology infrastructure, security management, and
technology acquisition, development and maintenance.

Activities designed and implemented to restrict technology access to authorized users to protect
the organization’s assets from external threats are a particularly important aspect of technology
general controls. Guarding against external threats is particularly important for an entity that de-
pends on telecommunications networks and the Internet.

35
See General Controls in Systems Controls and Security Measures later in this section for more information about
technology general controls.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 185
Internal Control CMA Part 1

12) The organization deploys control activities by developing policies that establish what is ex-
pected and procedures that put the policies into action. The control activities should be built
into business processes and employees’ day-to-day activities.

Whether a policy is in writing or not, it should establish clear responsibility and accountability. The
responsibility and accountability ultimately reside with the management of the entity and the sub-
unit where the risk resides. Procedures should be clear on what the responsibilities are of the
personnel performing the control activity. The procedures should be timely and should be per-
formed diligently and consistently by competent personnel.

Responsible personnel should investigate and if necessary take corrective action on matters iden-
tified as a result of executing control activities. For example, if a discrepancy is identified in the
process of doing a reconciliation, the discrepancy should be investigated. If an error occurred, the
error should be corrected and the correction reflected in the reconciliation.

Management should periodically review and reassess policies and procedures and related control
activities for continued relevance and revise them when necessary.

Component 4: Information and Communication


Relevant information must be identified, captured and communicated (shared) in a manner that enables
people to carry out the internal control responsibilities that support achievement of the organization’s ob-
jectives. Information and communication are both internal and external.

Principles relating to the information and communication component include:

13) The organization should obtain or generate and use relevant, quality information to support
the functioning of internal control.

Relevant information can be financial or non-financial. The information can be generated from
internal or external sources. Regardless of whether it is from internal or external sources and
whether it is financial or non-financial information, timely and relevant information is needed to
carry out internal control responsibilities supporting all three of the categories of objectives.

o Some examples of internal sources include emails, production reports regarding quality,
minutes of meetings discussing actions in response to reports, time reports, information on
number of units shipped during a period, and internal contacts made to a whistle-blower hot-
line. Other types of operating information, such as measurements of emissions generated, are
needed to monitor compliance with emissions standards.

o Some examples of external sources include data from outsourced providers, research reports,
information from regulatory bodies regarding new requirements, social media and blog post-
ings containing comments or opinions about the organization, and external contacts made to
a whistle-blower hotline. External information about economic conditions and actions of com-
petitors is needed for internal decision-making, such as decisions about optimum inventory
levels and inventory valuation.

14) The organization should internally communicate information, including objectives and respon-
sibilities for internal control, necessary to support the functioning of internal control.

Information systems must provide accurate, timely reports to appropriate personnel so they can
carry out their responsibilities. The people who deal with the customers every day are often the
first to know about a problem, and they should have a way to communicate that information
upward. Furthermore, people in the organization need to receive a clear message from top man-
agement that their internal control responsibilities must be taken seriously.

o Internal communication includes communications between the board of directors and manage-
ment so that both have the information they need to fulfill their roles in achieving the
organization’s objectives.

186 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internal Control

o Internal communication of information also takes place across the organization through, for
example, policies and procedures, individual authorities, responsibilities and standards of con-
duct, specified objectives, and any matters of significance relating to internal control such as
instances of deterioration or non-adherence.

Internal communication can take many forms, such as dashboards, email messages, training (ei-
ther live or online), one-on-one discussions, written policies and procedures, website postings, or
social media postings. Actions also communicate. The way managers behave in the presence of
their subordinates can communicate more powerfully than any words.

15) The organization should communicate with external parties regarding matters affecting the
functioning of internal control.

Relevant and timely information needs to be communicated to external parties including share-
holders, partners, owners, regulators, customers, financial analysts, and other external parties.

External communication also includes incoming communications from customers (customer feed-
back), consumers, suppliers, external auditors, regulators, financial analysts, and others. A
whistle-blower hotline should be available to external parties, as well.

o Communication from customers and suppliers can provide valuable input on the design and
quality of the company’s products and services.

o Communications from the external auditors inform management and the board about the or-
ganization’s operations and control systems.

o Regulators report results of examinations and compliance reviews that can inform manage-
ment of control weaknesses.

o If the company is a public company, communications to shareholders, regulators, financial


analysts, and others need to provide information relevant to those people so they can under-
stand the entity’s condition and risks.

o Customer complaints often are an indication of operating problems that need to be addressed.

o Any outsider dealing with the company must be informed that improper actions such as kick-
backs or other improper dealings will not be tolerated.

o Outgoing communication can take the form of press releases, blogs, social media, postings on
the company website, and emails.

Component 5: Monitoring Activities


Finally, management must monitor the entire internal control system. Monitoring is an activity overseen or
performed at the management level for the purpose of assessing the operation and effectiveness of existing
internal controls. Monitoring assesses the quality of the internal control system’s performance over time to
determine whether the components of internal control are present and are functioning. Management must
also revisit previously identified problems to make sure they have been corrected.

Monitoring ensures that the internal control system continues to operate effectively. Systems and proce-
dures change over time, and the way controls are applied need to change in order to continue to be
effective. Management needs to determine whether the internal control system is still relevant and whether
it is still able to address new risks that may have developed.

Information received from monitoring activities is used in management’s assessment of the effectiveness
of its internal control. Principles relating to the monitoring activities component are:

16) The organization selects, develops, and performs ongoing evaluations, separate evalua-
tions, or some combination of both to ascertain whether the five components of internal control
are present and functioning.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 187
Internal Control CMA Part 1

Monitoring can be done in two ways:

o Through ongoing evaluations that are built into business processes during normal operations
and provide timely information.

o Through separate evaluations conducted periodically by management with the assistance of


the internal audit function.

If monitoring is done regularly during normal operations, the need for separate evaluations is
lessened.

Note: Monitoring should be done on a regular basis. An advantage to ongoing monitoring is


that if operating reports are used to manage ongoing operations, exceptions to anticipated
results will be recognized quickly.

17) The organization evaluates and communicates internal control deficiencies in a timely man-
ner to parties responsible for taking corrective action, including senior management and the board
of directors, as appropriate.

Findings from monitoring activities are evaluated against established criteria and deficiencies are
communicated to management and the board of directors as appropriate. Remedial action should
be taken, and the results of the remedial action should also be monitored to be certain that the
situation has been corrected.

An example of evaluating a monitoring activity is reviewing a reconciliation to make sure it was


done properly and in a timely manner, that the sources of information used in the reconciliation
were appropriate, and to look for trends in the reconciling items. All the reconciling items should
have been investigated and resolved, and management should evaluate whether any new risks to
the operation have been caused by changes in the internal and external environments.

188 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internal Control

Summary: The 5 Components and the 17 Principles of Internal Control

Components Principles

Control Environment 1) There is a commitment to integrity and ethical values.

2) The board of directors exercises oversight responsibility for inter-


nal control.

3) Management establishes structures, authorities, and responsibili-


ties.

4) There is a commitment to competence.

5) Individuals are held accountable for their internal control responsi-


bilities.

Risk Assessment 6) Objectives are specified so risks to their achievement can be iden-
tified and assessed.

7) Risks are identified and analyzed.

8) Potential for fraud is considered.

9) Changes that could impact internal control are identified and as-
sessed.

Control Activities 10) Control activities to mitigate risks are selected and developed.

11) General control activities over technology are selected and devel-
oped.

12) Control activities are deployed through policies and procedures.

Information and
13) Relevant, quality information is obtained or generated and is used.
Communication

14) Information is communicated internally.

15) The organization communicates with external parties.

Monitoring Activities 16) Ongoing and separate evaluations are performed of the internal
control system.

17) Internal control deficiencies are evaluated and communicated for


corrective action.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 189
Internal Control CMA Part 1
efham CMA
What is Effective Internal Control?
An effective internal control system provides reasonable assurance regarding achievement of an entity’s
objectives by reducing to an acceptable level the risk of not achieving an entity objective. It requires each
of the five components and relevant principles of internal control to be present and functioning and for
the five components to be operating together in an integrated manner.

When an internal control system is effective, senior management and the board of directors have reason-
able assurance that the organization:

• Achieves effective and efficient operations or understands the extent to which operations are man-
aged effectively and efficiently.

• Prepares reports in conformity with applicable rules, regulations, and standards or with the entity’s
specified reporting objectives.

• Complies with all applicable laws and regulations.

However, the board of directors and management cannot have absolute assurance that the organization’s
objectives will be achieved. Human judgment in decision-making can be faulty, errors do occur, manage-
ment may be able to override internal controls, and management, other personnel, or third parties may be
able to circumvent internal controls through collusion.

Transaction Control Objectives


Commonly accepted transaction control objectives are:

• Authorization. All transactions are approved by someone with the authority to approve the spe-
cific transactions.

• Completeness. All valid transactions are included in the accounting records.

• Accuracy. All valid transactions are accurate, are consistent with the originating transaction data,
are correctly classified, and the information is recorded in a timely manner.

• Validity. All recorded transactions fairly represent the economic events that occurred, are lawful,
and have been executed in accordance with management’s authorization.

• Physical safeguards and security. Access to physical assets and information systems are con-
trolled and restricted to authorized personnel.

• Error handling. Errors detected at any point in processing are promptly corrected and reported
to the appropriate level of management.

• Segregation of duties. Duties are assigned in a manner that ensures that no one person is in a
position to both perpetrate and conceal an irregularity.

Types of Transaction Control Activities


• Authorization and approvals. Authorization confirms that the transaction is valid, in other words
that it represents an actual economic event. Authorization generally is in the form of an approval
by a higher level of management or of another form of verification, such as an automatic compar-
ison of an invoice to the related purchase order receiving report. When automated authorization
of payables is utilized, invoices within the tolerance level are automatically approved for payment,
while invoices outside the tolerance level are flagged for investigation.

• Verifications. Items are compared with one another or an item is compared with a policy, and if
the items do not match or the item is not consistent with policy, follow up occurs.

• Physical controls. Equipment, inventories, securities, cash, and other assets are secured physi-
cally in locked or guarded areas with physical access restricted to authorized personnel and are
periodically counted and compared with amounts in control records.

190 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internal Control

• Controls over standing data. Standing data, such as in a master file containing prices or inven-
tory items, is often used in the processing of transactions. Controls need to be put into place over
the process of populating, updating, and maintaining the accuracy, completeness, and validity of
the data in the master files.

• Reconciliations. Reconciliations compare two or more data elements and, if differences are found,
action is taken to make the data elements agree. For example, a bank reconciliation reconciles the
balance in the bank account according to internal records with the balance in the account according
to the bank. Reconciling items are items in transit (outstanding checks and deposits) and are to
be expected. However, differences that cannot be explained by items in transit must be investi-
gated and corrective action taken. Reconciliations generally address the completeness and
accuracy of processing transactions.

• Supervisory controls. Supervisory controls determine whether other transaction control activi-
ties are being performed completely, accurately, and according to policy and procedures. For
example, a supervisor may review a bank reconciliation performed by an accounting clerk to check
whether the bank balance as given on the reconciliation report matches the balance on the state-
ment and whether reconciling items have been followed up and corrected and an appropriate
explanation provided.

Safeguarding Controls
Physical safeguarding of assets against loss is an important part of a company’s operations objectives. Loss
to assets can occur through unauthorized acquisition, use, or disposition of assets or through destruction
caused by natural disasters or fire.

Prevention of loss through waste, inefficiency, or poor business decisions relates to broader operations
objectives and is not specifically considered part of asset safeguarding.

Physical protection of assets requires:

• Segregation of duties.

• Physical protection and controlled access to records and documents such as blank checks, purchase
orders, passwords, and so forth.

• Physical protection measures to restrict access to assets, particularly cash and inventory.

• Effective supervision and independent checks and verification.

Segregation of Duties
Duties need to be divided among various employees to reduce the risk of errors or inappropriate activities.
Separating functions ensures that no single individual is given too much responsibility so that no em-
ployee is in a position to both perpetrate and conceal irregularities.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 191
Internal Control CMA Part 1

Note: Different people must always perform the following four functions of related activities:

1) Authorizing a transaction.

2) Recordkeeping: Recording the transaction, preparing source documents, maintaining journals.

3) Keeping physical custody36 of the related asset: For example, receiving checks in the mail.

4) The periodic reconciliation of the physical assets to the recorded amounts for those assets.

In a question about an effective or ineffective internal control, keep in mind that the above four things
must be done by different people.

Following are a few examples of potential internal control failures that can result from inadequate segrega-
tion of duties:
• If the same person has custody of cash received and also has the authority to authorize account
write-offs, that person could receive a cash remittance on account from a customer, authorize a
fraudulent write-off of the customer’s (paid) receivable, and divert the cash collected to his or her
own use.
• If the same person who authorizes issuance of purchase orders is also responsible for recording
receipt of inventory and for performing physical inventory counts, that person could authorize the
issuance of a purchase order to a fictitious vendor using a post office box personally rented for the
purpose, then prepare a fictitious receiving record, and personally mail an invoice to the company
in the name of the fictitious vendor using the personal post office box. The accounts payable
department of the company would match the purchase order, the receiving report, and the invoice,
as it is supposed to do. Since all the documentation would match, the accounts payable department
would send a payment to the fictitious vendor for something the company never ordered or re-
ceived. Furthermore, during physical inventory counting that same person could easily mark the
item as being in inventory when it never was in inventory, thereby concealing the fraud.
• Also, if the same person who authorizes issuance of purchase orders is also responsible for record-
ing receipt of fixed assets and for performing physical fixed asset inventories, that person could
purchase office furniture, have it delivered to another address, and proceed to sell it and pocket
the revenue from its sale.

• If the same person prepares the bank deposit and also reconciles the checking account, that person
could divert cash receipts and cover the activity by creating “reconciling items” in the account
reconciliation report.
Be aware, however, that segregation of duties does not guarantee that fraud will not occur. Two or more
employees could collude with one another (work together to conspire) to commit fraud, covering for one
another and, presumably, sharing the proceeds.

Software tools are available to assist a business in identifying incompatible functions. An access manage-
ment application can help to assess segregation-of-duties and access risks and conflicts.

Note: Collusion occurs when two or more individuals work together to overcome the internal control
system and perpetrate a fraud. When two or more people work together, they are able to get around
the segregation of duties that may have been set out.

36
In the context of internal control, custody involves keeping, guarding, caring for, watching over, inspecting, preserv-
ing, and maintaining the security of an item that is within the immediate personal care and control of the person to
whose custody it is entrusted.

192 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internal Control

Question 52: The proper segregation of duties requires that:

a) The individual who records a transaction does not compare the accounting record of the asset
with the asset itself.

b) The individual who records a transaction must maintain custody of the asset resulting from the
transaction.

c) The individual who authorizes a transaction also records it.

d) The individual who maintains custody of an asset must have access to the accounting records for
the asset.

(CMA Adapted)

Question 53: In a well-designed internal control system, two tasks that should be performed by different
people are:

a) Posting of amounts from both the cash receipts journal and cash payments journal to the general
ledger.

b) Distribution of payroll checks and approval of sales returns for credit.

c) Approval of credit loss write-offs, and reconciliation of the accounts payable subsidiary ledger and
controlling account.

d) Reconciliation of bank account and recording of cash receipts.

(CMA Adapted)

Question 54: One characteristic of an effective internal control structure is the proper segregation of
duties. The combination of responsibilities that could be performed by the same person is:

a) Preparation of paychecks and check distribution.

b) Timekeeping and preparation of payroll journal entries.

c) Signing of paychecks and custody of blank payroll checks.

d) Approval of time cards and preparation of paychecks.

(CMA Adapted)

Physical Protection – Controlled Access to Records and Documents


Checks should be stored in a locked area, and access to them should be limited to personnel who have
responsibility for preparing checks, subject to authorization and approvals by other individuals. The checks
should be pre-numbered, and the check numbers should be recorded in a log as they are used. Any checks
discovered missing should be promptly reported to supervisory personnel.

Purchase orders should also be pre-numbered, numbers logged as used, and access to them similarly
restricted.

Corporate credit cards should be kept in a locked cabinet and access to them controlled.

Password access should be controlled by limiting the access granted by each employee’s password to just
the information needed by that employee to do his or her job. Employees should be instructed not to leave
their passwords in exposed areas (for example, taped to the computer).

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 193
Internal Control CMA Part 1

Physical Protection – Restricted Access to Assets


When cash must be stored until it can be deposited, it should be kept in a locked, fireproof file cabinet or
safe under controlled access.

Inventory can be a major portion of a company’s assets, and it is vulnerable to loss from theft, fire, and
natural disasters. The risk of loss can be at least partially transferred through purchase of insurance, but
internal controls are vital to protect as much as possible against loss.

• Inventory should be kept in a physically locked area, and the locks should be technically advanced
(not just simple combination locks).

• Requisitions for inventory should be approved by authorized personnel.

• The inventory area should be monitored by a gatekeeper who verifies proper authorization for
requests to move goods.

• Security cameras can be used to monitor activity in the inventory area and to help identify theft
and thieves. The very existence of cameras tends to discourage employee theft.

• Security alarms on doors and windows can alert local police in case of a break-in.

• A security guard may be employed during hours when employees are not present if the inventory
has high value.

• Regular physical inventories should be taken and missing items should be investigated.

Effective Supervision and Independent Checks and Verifications


Supervision over the performance of clerical functions is necessary. For example:

• Comparison of independent sets of records is necessary, such as comparing the report of the
physical count of inventory to the internal inventory records; or comparing the information on a
bank reconciliation with the actual bank statement to confirm the bank balance used in the recon-
ciliation is correct.

• Invoices should be prepared based on verified orders. The packing slip should be prepared at the
same time the invoice is prepared, and nothing should be shipped without a packing slip. Records
of actual shipments made should be compared with internal shipping documents, which should be
compared with invoices issued to verify that the procedures are being followed.

• The process of receiving inventory should be supervised to make sure the inventory clerk is actually
counting the items received before affirming their receipt.

194 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Legislative Initiatives About Internal Control

Legislative Initiatives About Internal Control


In the United States, various federal legislative initiatives have been created to require companies to im-
plement internal controls. CMA candidates should be able to do the following:

• Identify and describe the major internal control provisions of the Foreign Corrupt Practices Act.

• Describe the major internal control provisions of the Sarbanes-Oxley Act (Sections 201, 203, 204,
302, 404, and 407).

• Identify the role of the Public Company Accounting Oversight Board (PCAOB) in providing guidance
on the auditing of internal controls.

• Identify the PCAOB-preferred approach to auditing internal controls as outlined in Auditing Stand-
ard #5.

One significant issue with these federal laws is that some statutes apply only to publicly-traded companies
or only to companies that report to the SEC, and some statutes apply to all companies. Companies that are
not publicly traded and do not report to the SEC do not need to comply with many of these laws because
they do not fall under the jurisdiction of the SEC, which is the primary regulatory agency for these internal
control statutes. For example, the accounting and internal control provisions of the Foreign Corrupt Prac-
tices Act apply to all U.S. companies that are registered with and report to the SEC, not only those with
foreign operations, whereas the FCPA’s anti-bribery provisions apply to any company, public or private,
with significant operations in the United States, regardless of whether the corrupt act takes place inside or
outside the United States.

Foreign Corrupt Practices Act (FCPA)


The Foreign Corrupt Practices Act of 1977 (FCPA), substantially revised in 1988 and amended in 1998 by
the International Anti-Bribery and Fair Competition Act of 1998, was enacted in response to disclosures of
questionable payments that had been made by large companies. Investigations by the SEC had revealed
that over 400 U.S. companies had made questionable or illegal payments in excess of $300 million to
foreign government officials, politicians, and political parties. The payments were either illegal political
contributions or payments to foreign officials that bordered on bribery.

The FCPA has two main provisions: an anti-bribery provision, and an internal control provision.

Anti-Bribery Provision
The anti-bribery provisions of the FCPA apply to all companies, regardless of whether or not they are
publicly traded.

Under the FCPA, it is illegal for any company or anyone acting on behalf of a company to bribe any
foreign official in order to obtain or retain business. In addition, a firm, or an individual acting on
behalf of a firm, will be held criminally liable if it makes payments to a third party with the knowledge that
those payments will be used by the third party as bribes.

Note: This prohibition is against corrupt payments to a foreign official, a foreign political party or party
official, or any candidate for foreign political office only.

The entire company, not any one individual or position in the company, is responsible for ensur-
ing that all payments are legal and lawful, although individuals are personally liable for their own
actions. Furthermore, the company must ensure that all transactions are made in accordance with man-
agement’s general or specific authorization and are recorded properly.

Note: A corrupt payment is one that is intended to cause the recipient to misuse his or her official
position in order to wrongfully direct business to the payer, whether or not the payment leads to
the desired outcome.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 195
Legislative Initiatives About Internal Control CMA Part 1

Internal Control Provision


The fundamental premise of the internal control requirements of the FCPA is that effective internal control
acts as a deterrent to illegal payments. Therefore, under the Foreign Corrupt Practices Act corporate man-
agement is required to maintain books, records, and accounts that accurately and fairly reflect transactions
and to develop and maintain a system of internal accounting control.

Note: The internal control provision of the FCPA applies only to companies that are publicly traded.

Sarbanes-Oxley Act and the PCAOB


The Sarbanes-Oxley Act of 2002 (called SOX or SarbOx) contains provisions impacting auditors, manage-
ment, and audit committees of boards of directors. Sarbanes-Oxley was enacted in response to several
major incidents of financial reporting fraud and audit failures, and it applies to all publicly-held companies
in the U.S., all of their divisions, and all of their wholly-owned subsidiaries. It also applies to any non-U.S.
owned, publicly-held multinational companies that operate in the U.S. In addition, some provisions apply
also to privately-held companies; and a privately-held company may comply with SOX in preparation for
an initial public offering, in preparation for raising private funding for a sale of the company, or on a volun-
tary basis (for example, using it as a best-practices benchmark).

Title I: Public Company Accounting Oversight Board (PCAOB)


Title 1 of the Sarbanes-Oxley Act established the Public Company Accounting Oversight Board (PCAOB),
whose mandate is to oversee the auditing of public companies that are subject to the securities laws, to
protect the interests of investors, and to enhance the public’s confidence in independent audit reports. As
an independent, non-governmental board, the PCAOB is a non-profit corporation that operates under the
authority of the SEC, which oversees the approval of its Board’s rules, standards, and budget.

According to the Act, public accounting firms that serve as external auditors are required to register with
the PCAOB. Furthermore, the PCAOB is charged with developing auditing standards to be used by registered
public accounting firms in their preparation and issuance of audit reports. In addition, the PCAOB conducts
regular inspections of the registered public accounting firms to assess their degree of compliance with the
Act and it has procedures to investigate and discipline firms that commit violations.

The formation of the PCAOB constitutes the first time that auditors of U.S. public companies became subject
to external and independent oversight. Previously, the profession had been self-regulated through a formal
peer review37 program administered by the American Institute of Certified Public Accountants (AICPA). That
peer review program continues, and accounting and audit firms that are required to be inspected by the
PCAOB are also subject to peer review.

The responsibilities of the PCAOB include:

1) Registering public accounting firms that audit public companies. The Sarbanes-Oxley Act requires
all accounting firms (both U.S. and non-U.S. firms) that prepare or issue audit reports on or par-
ticipate in audits of U.S. public companies to register with the PCAOB.

2) Establishing auditing and related attestation, quality control, ethics, independence, and other
standards relating to the preparation of audit reports for securities issuers.

3) Conducting inspections of registered public accounting firms, annually for firms that audit more
than 100 issuers and every three years for others. In the inspections, the Board assesses the firm’s
compliance with the Sarbanes-Oxley Act, the rules of the Board, the rules of the Securities and

37
“Peer review” is a process of self-regulation used in professions to evaluate the work performed by one’s equals, or
peers, to ensure it meets certain criteria. Peer review is performed by qualified individuals within the same profession.
A peer review is performed for an accounting and audit firm by professionals from another accounting and audit firm.

196 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Legislative Initiatives About Internal Control

Exchange Commission (SEC), and its professional standards in connection with the firm’s perfor-
mance of audits, issuance of audit reports, and related matters involving issuers.

4) Enforcing compliance with the Act, the rules of the Board, professional standards, and securities
laws relating to audit reports and the obligations of accountants for them.

5) Conducting investigations and disciplinary proceedings and imposing appropriate sanctions for vi-
olations of any provision of the Sarbanes-Oxley Act, the rules of the Board, the provisions of the
securities laws relating to the preparation and issuance of audit reports, or professional standards.

6) Management of the operations and staff of the Board.

Title II: Auditor Independence

Section 201: Services Outside the Scope and Practice of Auditors


In order to maintain auditor independence, Section 201 lists specific non-audit services that may not be
provided by an external auditor to an audit client because their provision creates a fundamental conflict
of interest for the accounting firms. These services include:

1) Bookkeeping services or other services relating to keeping the accounting records or preparing the
financial statements of the audit client.

2) Financial information systems design and implementation.

3) Appraisal or valuation services, fairness opinions, or contribution-in-kind reports.

4) Actuarial services.

5) Internal audit outsourcing services.

6) Management functions.

7) Human resource services.

8) Broker/dealer, investment adviser, or investment banking services.

9) Legal services.

10) Expert services unrelated to the audit.

11) Any other service that the Public Company Accounting Oversight Board (PCAOB) determines, by
regulation, is not permissible.

Section 203: Audit Partner Rotation


A public accounting firm that is registered with the PCAOB may not provide audit services to a client if the
lead audit partner38 or the concurring review audit partner has performed audit services for that client in
each of the five previous fiscal years of the client. Therefore, the lead audit partner and the concurring
review audit partner must rotate off a particular client’s audit after five years. They must then remain off
that audit for another five years. Other audit partners who are part of the engagement team must rotate
off after seven years and remain off for two years if they meet certain criteria.

Specialty partners, (partners who consult with others on the audit engagement regarding technical or in-
dustry-specific issues), do not need to rotate off. Examples of specialty partners are tax or valuation
specialists. Other partners who serve as technical resources for the audit team and are not involved in the
audit per se are also not required to rotate off the audit.

38
An “audit partner” is defined as any partner on the audit engagement team with responsibility for decision-making on
any significant audit, accounting or reporting matter affecting the company’s financial statements or who maintains
regular contact with management and the audit committee of the audit client.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 197
Legislative Initiatives About Internal Control CMA Part 1

The purpose of the audit partner rotation requirement is to ensure that a “new look” is taken periodically
at the financial statements.

Section 204: Auditor Reports to Audit Committees


Section 204 requires each public accounting firm registered with the PCAOB that performs an audit for an
issuer of publicly-traded securities to report the following in a timely manner to the issuer’s audit commit-
tee:

1) All critical accounting policies and practices to be used;

2) All alternative treatments of financial information within generally accepted accounting principles
that have been discussed with the issuer’s management, the ramifications of the use of such al-
ternative disclosures and treatments, and the treatment preferred by the registered public
accounting firm; and

3) Other material written communication between the registered public accounting firm and the man-
agement of the issuer, such as any management letter or schedule of unadjusted differences.

If management of the company is using an alternative method of accounting for something, even though
it is within generally accepted accounting principles, the outside public accounting firm performing the audit
must report that fact to the company’s audit committee. The independent auditor must report all critical
accounting principles being used and any other material written communications between management and
themselves.

Title III: Corporate Responsibility


Section 302: Corporate Responsibility for Financial Reports
Sarbanes-Oxley requires that each annual (10K) or quarterly (10Q) financial report filed or submitted to
the SEC in accordance with the Securities Exchange Act of 1934 must include certifications by the company’s
principal executive officer or officers and its principal financial officer or officers. The certification must
indicate the following:

1) The signing officer has reviewed the report.

2) Based on the signing officer’s knowledge, the report does not contain any untrue material state-
ment or omit to state any material fact that could cause the report to be misleading.

3) Based on the signing officer’s knowledge, the financial statements and all the other related infor-
mation in the report fairly present in all material respects the financial condition and results of
operations of the company for all of the periods presented in the report.

4) The signing officers certify that they

o are responsible for establishing and maintaining internal controls;

o have designed the internal controls to ensure that they are made aware of all material
information relating to the company and all subsidiaries;

o have evaluated the effectiveness of the company’s internal controls within the previous
ninety days; and

o have presented in the report their conclusions about the effectiveness of their internal
controls, based on their evaluation as of the report date.

198 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Legislative Initiatives About Internal Control

5) The signing officers have disclosed to the company’s auditors and the audit committee of the board
of directors:

o all significant deficiencies in the design or operation of the company’s internal controls that
could adversely affect the company’s ability to record, process, summarize, and report financial
data and have identified for the company’s auditors any material weaknesses in its internal
controls; and

o any fraud, regardless of its materiality, that involves management or other employees who
have a significant role in the company’s internal controls.

6) The signing officers have stated in the report whether or not there were any significant changes in
internal controls or in any other factors that could significantly affect internal controls after the
date of their evaluation, including any corrective actions that have been taken with regard to
significant deficiencies and material weaknesses.

Note: Companies cannot avoid the requirements in the Sarbanes-Oxley Act by reincorporating outside
the United States or by transferring their company’s activities outside of the United States. The Act will
continue to have full legal force over any company that reincorporates outside the U.S. or that transfers
its activities outside of the U.S.

Title IV: Enhanced Financial Disclosures


Section 404: Management Assessment of Internal Controls and the Independent Auditor’s Attestation
to Management’s Assessment of Internal Controls
Section 404(a) requires each annual report required by the SEC to:

1) state the responsibility of management for establishing and maintaining an adequate internal con-
trol structure and procedures for financial reporting, and

2) contain an assessment by management of the adequacy of the company’s internal control over
financial reporting (or ICFR).

Section 404(b) requires the company’s independent auditor to report on and attest to management’s as-
sessment of the effectiveness of the internal controls.

In other words, according to Section 404(a) management is required to document and test its internal
financial controls and to report on their effectiveness. In many firms, the internal audit activity is very
involved in the management review and testing of the internal controls. Furthermore, according to Section
404(b) the external auditors are then required to review the supporting materials used by management
and/or internal auditing in developing their internal financial controls report. The external auditor’s report
is done in order to assert that management’s report is an accurate description of the internal control envi-
ronment.

The first step in a Section 404 compliance review is to identify the key processes. Here, the internal audit
activity can be of major assistance because it may already have defined the key processes during its annual
audit planning and documentation. The overall processes are generally organized in terms of the basic
accounting cycles, as shown here:

• Revenue cycle: processing of sales and service revenue

• Direct expenditures cycle: expenditures for material and direct production costs

• Indirect expenditures cycle: operating costs other than for production activities

• Payroll cycle: compensation of personnel

• Inventory cycle: processes for the management of direct materials inventory until it is applied
to production

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 199
Legislative Initiatives About Internal Control CMA Part 1

• Fixed assets cycle: processes for accounting for property and equipment, such as periodic re-
cording of depreciation

• General IT cycle: general IT controls that are applicable to all IT operations

The internal controls covering the key processes are reviewed and documented, and then the controls are
tested. The external auditor then reviews the work and attests to its adequacy.

The extent of internal audit’s involvement in the Section 404 testing of internal controls varies from firm to
firm. It can take three forms:

1) Internal auditors may act as internal consultants in identifying the key processes, documenting
the internal controls over these processes, and performing tests of the controls. Senior manage-
ment’s approval of internal audit’s work is necessary.

2) The company might designate some other internal or external consulting resource to perform the
Section 404 reviews. If an external party performs the Section 404 reviews, the internal auditors
may act as a resource to support the work. They may review and test internal control processes
as assistants or contractors for the outside entity doing the review.

3) Internal audit may work with and assist other corporate resources that are performing the Section
404 reviews without being directly involved with those reviews, allowing the internal audit activity
to concentrate its resources on other internal audit projects.

Management, in its assessment of internal controls, and the independent auditor, in its attestation to man-
agement’s assessment, can have different testing approaches because their roles are different and therefore
they are subject to different guidance.

• Guidance for the management assessment of internal controls is provided by the SEC.

• Guidance for the independent auditor’s attestation to management’s assessment is contained in


the Public Company Accounting Oversight Board’s Auditing Standard No. 5.

Although the sources of guidance are different for management and the independent auditor, the PCAOB
intentionally aligned its guidance in Auditing Standard No. 5 with the SEC’s guidance for management,
particularly with regard to prescriptive requirements, definitions, and terms. Therefore, the guidance to
management and the guidance to independent auditors are not in conflict.

Both SEC and the PCAOB’s guidance have the effect of efficiently focusing Section 404 compliance on
the most important matters affecting investors.

They both prescribe a top-down, risk-based approach to evaluating internal control over financial re-
porting. A top-down approach begins by identifying the risks that a material misstatement of the financial
statements would not be prevented or detected in a timely manner. Beginning with risk assessment allows
the auditor to focus on higher-risk areas while spending less time and effort on areas of lower risk. The
auditor should test those controls that are important to the auditor's conclusion about whether the compa-
ny's controls sufficiently address the assessed risk of misstatement to each relevant assertion. 39

It is important for the auditor to use a top-down approach, not a bottom-up approach. An auditor
who approaches the audit of internal controls from the bottom up would focus first on performing detailed
tests of controls at the process, transaction, and application levels. When the auditor uses a bottom-up
process, he or she often spends more time and effort than is necessary to complete the audit. Furthermore,

39
An assertion is a claim made. A management assertion is a claim made by management. Financial statement asser-
tions are claims made by management in presenting financial information. Examples of broad financial statement
assertions are “Total liabilities at December 31, 20XX were $150,000,000” or “Net income for the year ended December
31, 20XX was $5,000,000.” Financial statement assertions can be much narrower also, as in “Receivables due from
Customer X on December 31, 20XX totaled $50,000.” The auditor’s role is to determine whether the assertions being
made by management are correct. Most of the work of a financial audit involves evaluating and forming opinions about
management assertions.

200 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Legislative Initiatives About Internal Control

when an auditor takes a bottom-up approach, the auditor may spend relatively little time testing and eval-
uating entity-level controls but instead may rely almost exclusively on detailed tests of controls over
individual processes, transactions, and applications. Spending more effort than is necessary in lower-risk
areas can diminish the effectiveness of the audit because it may prevent a higher-risk area from receiving
the audit attention that it should.

A top-down approach ensures the proper testing of the controls for the assessed risk of mis-
statement to each relevant assertion. If a bottom-up approach is used, those controls that address the
risk of a material misstatement may not be tested.

Section 407: Disclosure of Audit Committee Financial Expert


Section 407 of the Sarbanes-Oxley Act required, and the SEC has issued rules requiring, each issuer of
publicly-traded securities to disclose whether or not the company’s audit committee consists of at least one
member who is a financial expert. If the audit committee does not have at least one member who is a
financial expert, the company must state the reasons why not.

The definition of a financial expert is a person who, through education and experience as a public account-
ant, auditor or a principal accounting or financial officer of an issuer of publicly-traded securities, has

1) an understanding of generally accepted accounting principles and financial statements and the
ability to assess the general application of GAAP in connection with accounting for estimates, ac-
cruals, and reserves;

2) experience in the preparation, auditing, or active supervision of the preparation or auditing of


financial statements of generally comparable issuers40 in terms of the breadth and level of com-
plexity of accounting issues;

3) experience and an understanding of internal accounting controls and procedures for financial re-
porting; and

4) an understanding of audit committee functions.

If the company discloses that it has one financial expert on its audit committee, it must disclose the name
of the expert and whether that person is independent. If the company discloses that it has more than one
financial expert serving on its audit committee, it may, but is not required to, disclose the names of those
additional persons and indicate whether they are independent.

40
“Issuers” means issuers of securities, generally publicly-traded securities.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 201
External Auditors’ Responsibilities and Reports CMA Part 1

External Auditors’ Responsibilities and Reports


The audit committee nominates the independent auditor and the shareholders ratify the independent audi-
tor’s appointment at the annual meeting of shareholders.

For publicly-traded companies, the auditor presents an opinion letter that is included in the company’s
annual report and contains the following:

• An opinion on whether the financial statements present fairly, in all material respects, the fi-
nancial position, results of operations, and cash flows of the company, in conformity with generally
accepted accounting principles. An identification of the country of origin of those generally accepted
accounting principles also is required.

• An opinion on how well the company’s management has maintained effective internal con-
trol over financial reporting. (This opinion is not required for companies that are not publicly
traded.)

The auditor’s only responsibilities are to express an opinion about the financial statements and to express
an opinion about the internal controls of the company (in the case of a publicly-traded company). In its
capacity as auditor, the external auditor has no responsibility to make recommendations or suggestions to
the client about the operation of the business. The external auditor has no responsibility to follow up any
audit findings from the previous year.

The auditor's opinion letter specifically states what the auditor’s responsibilities are and it also states that
it is the management of the company who is responsible for the financial statements. The auditor’s report
states:

1) Management is responsible for the financial statements, for maintaining effective control over fi-
nancial reporting, and for its assessment of the effectiveness of its internal control over financial
reporting.

2) The auditor’s responsibility is to express opinions on the financial statements and on the company’s
internal control over financial reporting based upon its audit.

Financial Statement Opinion


The auditor conducts an independent examination of the accounting data prepared and presented by man-
agement and expresses an opinion on the financial statements. Though the auditor will not use the word
“correct,” the opinion is in essence whether or not the financial statements are “correct.” Instead, the
auditor would assert that the financial statements “present fairly, in all material respects” the financial
position of the company.

The auditor’s opinion on the financial statements contains either an opinion on the financial statements as
a whole or an assertion that an opinion cannot be expressed and the reasons why an opinion cannot be
expressed.

The auditor’s opinion may be:

• Unqualified: Most audit opinions are unqualified, meaning that the results are “clean.” The auditor
expresses the opinion that the financial statements present fairly, in all material respects, the
financial position, results of operations, and cash flows of the company, in conformity with gener-
ally accepted accounting principles.

• Qualified: A qualified opinion contains an exception, meaning that the financial statements do
not present a true and fair picture. The exception is usually not significant enough to cause
the statements as a whole to be misleading to the point that they should not be used. However, it
does prevent the auditor from issuing an unqualified opinion. Usually, a qualified opinion is issued
under one of the following conditions:

202 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E External Auditors’ Responsibilities and Reports

o The scope of the auditor’s examination (that is, the work the auditor wanted to perform) was
limited or was affected by restrictions, or

o One element of the financial statements (for example, fixed assets or inventory) is not stated
properly in conformity with generally accepted accounting principles or because of inadequate
disclosures.

A qualified opinion states that, except for the specific, named, matter, the financial statements
present fairly in all material respects, the financial position, results of operations, and cash flows
of the company in conformity with generally accepted accounting principles.

• Adverse: An adverse opinion is issued when the exceptions are so material that, in the auditor’s
judgment, a qualified opinion is not appropriate. In other words, the financial statements, taken
as a whole, are not presented in conformity with generally accepted accounting princi-
ples. Adverse opinions are seldom issued because most companies change their accounting upon
the instructions of the auditor so that an adverse opinion is not warranted because their adjusted
financial statements do present a fair and true picture.

• Disclaimer: A disclaimer of opinion is used when the auditor has not been able to gather enough
information on the financial statements to express an opinion.

Going Concern Opinions and Other Modifications


The auditor must also evaluate whether substantial doubt exists about the company’s ability to continue as
a going concern. As part of the audit, the auditor considers several factors that might indicate that the
company will no longer be in existence by the time the auditor does the next annual audit. Some of the
factors are recurring operating losses, working capital deficiencies, loan defaults, unlikely prospects for
more financing, and work stoppages. The auditor also considers external issues, like legal proceedings and
the loss of a key customer or supplier.

If the auditor is not satisfied by management’s plans to overcome the problems and remain in business and
has a substantial doubt about the company’s ability to remain a going concern, the auditor will add an
explanatory paragraph to the opinion describing the problem. The explanatory paragraph follows the audi-
tor’s opinion. Below is an example of a going concern explanatory paragraph.

The accompanying financial statements have been prepared assuming that the Company
will continue as a going concern. As discussed in Note X to the financial statements, the
Company has suffered recurring losses from operations and has a net capital deficiency
that raise substantial doubt about its ability to continue as a going concern. Management's
plans in regard to these matters are also described in Note X. The financial statements do
not include any adjustments that might result from the outcome of this uncertainty.41

Doubt about the company’s ability to stay in business does not prevent the opinion from being unqualified,
if the financial statements present fairly, in all material respects, the financial position, results of operations
and cash flows of the company, in conformity with generally accepted accounting principles. Thus, the audit
opinion can still be an unqualified opinion even if the preceding explanatory note is added.

41
PCAOB AS 24.13.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 203
External Auditors’ Responsibilities and Reports CMA Part 1

The auditor may add an explanatory paragraph in certain other circumstances, as well, even though the
opinion expressed is an unqualified one. These opinions essentially are stating, “The financial statements
are correct, but there is something else we would like you to know.” Note that the auditor will not use the
term “correct” in respect to financial statements, but this is in essence what the auditor is saying. Some of
the circumstances in which an additional paragraph will be added to the unqualified opinion include:

• An uncertainty about something that will be resolved in the future, but until then, not enough is
known about the outcome of the matter to determine whether it might result in a loss.

• If the company has made a change in its accounting principles or in its method of applying the
principles that has had a material effect on the comparability of the financial statements from year
to year, the auditor should include an explanatory paragraph identifying the nature of the change
and referring readers to the note in the financial statements where the change is discussed in
detail. It is presumed that the auditor has agreed with the change in accounting principle unless
the auditor states otherwise.

• The auditor may simply want to emphasize something regarding the financial statements. An ex-
ample might be the fact that the company has had significant transactions with related parties.

Internal Control Opinion


The second opinion contained in the auditor’s opinion letter, as required by the Sarbanes-Oxley Act, is the
auditor’s opinion as to whether or not the company’s management has maintained effective internal control
over financial reporting. The company’s annual report, filed with the SEC and incorporated into the annual
report to shareholders, must be accompanied by a statement by management that management is respon-
sible for creating and maintaining adequate internal controls. Management’s statement must set forth its
assessment of the effectiveness of these controls. The company’s auditor must report on and attest to
management’s assessment of the effectiveness of the internal controls. The auditor’s internal control opin-
ion is considered to be part of the core responsibility of the auditor and an integral part of the audit report.

The criteria used by the independent auditor in assessing the company’s internal control over financial
reporting come from the document Internal Control—Integrated Framework issued by the Committee of
Sponsoring Organizations (COSO) of the Treadway Commission.

If the independent auditor is satisfied with the company’s internal control over financial reporting, it includes
the following paragraph in its opinion letter:

In our opinion, the Company maintained, in all material respects, effective internal control
over financial reporting as of [date], based on criteria established in Internal Control -
Integrated Framework (2013) issued by the Committee of Sponsoring Organizations of the
Treadway Commission (COSO).

Reviews and Compilations


Reviews and compilations are two other financial statement services performed by auditors.

A review involves some high-level inquiries and analytical procedures, but the accountant makes no tests
of transactions or balances, no evaluation of internal control, and no performance of any other audit pro-
cedures. A review report consists of a statement that the accountant has reviewed the financial statements;
that a review includes primarily applying analytical procedures to management’s financial data and making
inquiries of company management; that a review is substantially less in scope than an audit, the objective
of which is the expression of an opinion regarding the financial statements; and that the accountant does
not express such an opinion.

The review report includes a paragraph stating that the review was performed in accordance with the
Statements on Standards for Accounting and Review Services issued by the American Institute of Certified
Public Accountants and that the information is the representation of management.

204 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E External Auditors’ Responsibilities and Reports

A review report gives negative assurance. Negative assurance means that the accountant states in the
report that he or she is not aware of any material modifications that should be made to the financial state-
ments in order for them to be in conformity with accounting principles generally accepted in the country
where the client is located.

A compilation is simply a formatted financial statement presenting the assertions of management without
performing any procedures to verify, corroborate, or review the information. A compilation is less in scope
than a review and significantly less in scope than a full audit, and it provides no assurance whatsoever to
users regarding the accuracy of the financial statements.

Reviews and compilations are used by smaller, privately-owned entities. They are not acceptable for pub-
licly-owned entities.

Reports to the Audit Committee of the Board of Directors


An external auditor that is registered with the PCAOB under the Sarbanes-Oxley Act is obligated under
Section 204 to make reports to the company’s audit committee of the board of directors of a publicly-traded
company that include:

• The accounting principles being used.

• All alternative treatments being used.

• The ramifications of the use of such alternative treatments.

• The treatment preferred by the public accounting firm.

• All other material written communication between the registered public accounting firm and the
management of the company.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 205
E.2. Systems Controls and Security Measures CMA Part 1

E.2. Systems Controls and Security Measures


Introduction to Systems Controls
Extensive use of computers in operations and accounting systems can tend to increase the company’s
exposure to inaccuracies and fraud.

Since computers apply the same steps to similar transactions, processing of transactions should involve no
opportunity for clerical (human) error. However, if the program itself contains a mistake, every transaction
processed using that defective program will contain an error. Furthermore, if a clerical error is made in
input, the result will be an output error.

Potential for fraud is always present in organizations and is a serious problem, even without computer
processing of data. The automatic processing of data, the volume of the data processed and the complexity
of the processing are aspects of computer processing that can increase both the risk of loss and the potential
monetary loss from exposures that would exist anyway. The concentration of data storage creates expo-
sure, as well. The potential for fraud is further increased because of the fact that programs are used for the
processing. Fraud can potentially be committed within the program itself. Without proper controls, this type
of fraud may go undetected for a long period of time.

Further complicating the situation is the fact that because of the nature of the system, audit trails may
exist for only a short period of time, since support documents may be in electronic format and be periodically
deleted. An audit trail is a paper or electronic record that shows a step-by-step documented history of a
transaction. It enables an auditor or other examiner to trace the transaction from the general ledger back
to the source document such as an invoice or a receipt. The existence of an audit trail means that an
amount appearing in a general ledger account can be verified by evidence supporting all the individual
transactions that go into the total. The audit trail must include all of the documentary evidence for each
transaction and the control techniques that the transaction was subjected to, in order to provide assurance
that the transaction was properly authorized and properly processed. When an audit trail is absent, the
reliability of an accounting information system is questionable.

On the positive side, computer systems can provide large amounts of information to management in a very
short period of time, enabling management to maintain closer control over the activities of the company
and their results.

The objectives of controls for an information system are similar to the objectives of overall organizational
internal controls:

• Promoting effectiveness and efficiency of operations in order to achieve the company’s objectives.

• Maintaining the reliability of financial reporting through checking the accuracy and reliability of
accounting data.

• Assuring compliance with all laws and regulations that the company is subject to, as well as ad-
herence to managerial policies.

• Safeguarding assets.

In Internal Control – Integrated Framework, internal control is defined as a process designed to provide
reasonable assurance that the company’s objectives will be achieved in the areas of effectiveness and
efficiency of operations, reliability of financial reporting, and compliance with applicable laws and regula-
tions. The internal control system is the responsibility of the company’s board of directors, management
and other personnel.

206 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.2. Systems Controls and Security Measures

The internal control system consists of five interrelated components:

1) The control environment


2) Risk assessment
3) Control activities
4) Information and communication
5) Monitoring42

Common exposures to loss that result from a failure to implement controls include competitive disad-
vantage, deficient revenues, loss of assets, inaccurate accounting, business interruption, statutory
sanctions, erroneous management decisions, fraud and embezzlement, and excessive costs.

The ultimate responsibility for internal control lies with management and the board.

Controls should be subjected to cost/benefit analysis. “Cost/benefit analysis” means that management
should not spend more on controls than the amount the company can expect to receive in benefits from
the controls. Determining what is required to attain reasonable assurance that the control objectives are
being met without spending more on them than can possibly be gained from them is a matter of manage-
ment judgment.

Threats to Information Systems


Sources of threats to information systems and data are many. A few of them are:

• Errors can occur in system design.


• Errors can occur in input or input manipulation may occur.
• Data can be stolen over the Internet.
• Data and intellectual property, including trade secrets, can be stolen by employees who carry them
out on very small storage media or just email them.
• Unauthorized alterations can be made to programs by programmers adding instructions that divert
assets to their own use.
• Data and programs can be damaged or become corrupted, either deliberately or accidentally.
• Data can be altered directly in the data file, without recording any transaction that can be detected.
• Viruses, Trojan Horses, and worms can infect a system, causing a system crash, stealing data, or
damaging data.
• Hardware can be stolen.
• Physical facilities as well as the data maintained in them can be damaged by natural disasters,
illegal activity or sabotage.

High-profile incidents such as the theft of millions of peoples’ names and social security numbers from
databases have underscored the importance of protecting information systems and data. Effective sys-
tems controls are the first line of defense against events such as these incidents and against threats such
as those above. Controls must be a part of every system and application to preserve the integrity of the
data and reduce the risk of loss from inadequate records, inaccurate accounting, business interruption,
fraud, violations of the law, asset loss, and damage to the business’s competitive position. The controls
must not only exist, but they must also function effectively.

42
Internal Control – Integrated Framework, copyright 1992, 1994, and 2013 by the Committee of Sponsoring Organi-
zations of the Treadway Commission. Used by permission.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 207
E.2. Systems Controls and Security Measures CMA Part 1

The Classification of Controls


Controls for a computer system are broken down into two types.

• General controls relate to all systems components, processes, and data in a systems environ-
ment.

• Application controls are specific to individual applications and are designed to prevent, detect,
and correct errors and irregularities in transactions during the input, processing, and output stages.

Both general controls and application controls are essential.

General controls relate to the general environment within which transaction processing takes place. Gen-
eral controls are designed to ensure that the company’s control environment is stable and well managed.
A stable and well-managed control environment strengthens the effectiveness of the company’s application
controls. General controls include:

• Administrative controls, including segregation of duties

• Computer operations controls

• Controls over the development, modification, and maintenance of computer programs

• Software controls

• Hardware controls

• Data security controls

• Provision for disaster recovery

The general controls above are organized according to the following categories.

1) The organization and operation of the computer facilities and resources, including:

a. Administrative controls such as provision for segregation of duties within the data pro-
cessing function, segregation of the data processing function from other operations, and
supervision of personnel involved in control procedures to ensure that the controls are per-
forming as intended. The information systems activity should have a high enough level in the
organization and adequate authority to permit it to meet its objectives.

b. Computer operations controls. A systems control group should monitor production, keep
records, balance input and output, perform manual procedures designed to prevent and detect
errors, and see that work is completed on schedule.

2) General operating procedures, including:

a. Written policies, procedures, and manuals that establish standards for controlling infor-
mation systems operations should be available and should be kept up to date. They should
include controls over the setup of processing jobs, operations software and computer opera-
tions, and backup and recovery procedures, as well as instructions for running computer jobs.
The policies, procedures, manuals and instructions should be documented and reviewed and
approved by a manager.

b. Operating procedures also specify the process to be followed in system development and
system changes in order to provide reasonable assurance that development of, and changes
to, computer programs are authorized, tested, and approved prior to use.

3) Software, hardware, and access controls, including

a. Software controls that monitor the use of software and prevent unauthorized access to soft-
ware applications and system software, including program security controls to prevent
unauthorized changes to programs and systems in use.

208 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.2. Systems Controls and Security Measures

b. Hardware controls to keep computer hardware physically secure, to protect it from fires and
extremes of temperature and humidity, and to provide for disaster recovery. Hardware controls
also include monitoring for equipment malfunctions, such as controls installed in computers
that can identify incorrect data handling or malfunction of the equipment.

c. Access controls to equipment and data, such as controls over physical access to the computer
system that are adequate to protect the equipment from damage or theft. Access controls also
include data security controls that ensure that data files are not subject to unauthorized access,
change, or destruction, both when they are in use and when they are being stored.

General controls are discussed in greater detail below.

Application controls are specific to individual applications. They ensure that only authorized data are
processed by the application and that the data are processed completely and accurately. Thus, application
controls are designed to prevent, detect, and correct errors in transactions as they flow through the input,
processing, and output stages of work. They are organized into three main categories, as follows.

• Input controls, designed to provide reasonable assurance that input entered into the system has
proper authorization, has been converted to machine-sensible form, and has been entered accu-
rately.

• Processing controls, designed to provide reasonable assurance that processing has occurred
properly and no transactions have been lost or incorrectly added.

• Output controls, designed to provide reasonable assurance that input and processing have re-
sulted in valid output.

Application controls are discussed in greater detail below.

Question 55: Systems control procedures are referred to as general controls or application controls.
The primary objective of application controls in a computer environment is to:

a) Provide controls over the electronic functioning of the hardware.

b) Maintain the accuracy of the inputs, files and outputs for specific applications.

c) Ensure the separation of incompatible functions in the data processing departments.

d) Plan for the protection of the facilities and backup for the systems.

(CMA Adapted)

General Controls
Organization and Operation of the Computer Facilities

Note: This is the first of three categories of general controls.

An IT Planning or Steering Committee should oversee the IT function. Members should include senior man-
agement, user management and representatives from the IT function. The committee should have regular
meetings and report to senior management.

The IT function should be positioned within the organization so as to ensure its authority as well as its
independence from user departments.

Staffing requirements should be evaluated whenever necessary to make sure that the IT function has suf-
ficient, competent staff. Management should make certain that all personnel in the organization know their
responsibilities with respect to information systems and that they have sufficient authority to exercise their
responsibilities. Responsibilities should be delegated with appropriate segregation of duties, and duties
should be rotated periodically at irregularly scheduled times for key processing functions.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 209
E.2. Systems Controls and Security Measures CMA Part 1

Administrative Controls – Segregation of Duties


Although the traditional segregation practiced in accounting of separating the responsibilities of authoriza-
tion, record keeping, custody of assets, and periodic reconciliations may not be practiced in the same
manner in Information Systems (since the work is quite different), specific duties in the information systems
environment should still be separated from one another.

General guidelines for segregation of duties in information systems are as follows.

Separate Information Systems from Other Departments


Information Systems department personnel should be separated from the departments and personnel that
they support (called “users”). Examples of this segregation include:

• Users initiate and authorize all systems changes, and a formal written authorization is required.

• Asset custody remains with the user departments.

• An error log is maintained and referred to the user for correction. The data control group follows
up on errors.

Separate Responsibilities within the Information Systems Department


Responsibilities within the Information Systems department should be separated from one another. An
individual with unlimited access to a computer, its programs, and its data could execute a fraud and at the
same time conceal it. Therefore, effective segregation of duties should be instituted by separating the
authority for and the responsibility for the function.

Although designing and implementing segregation of duties controls makes it difficult for one employee to
commit fraud, remember that segregation of duties cannot guarantee that fraud will not occur because two
employees could collude to override the controls.

Segregation of duties should be maintained between and among the following functions:

• Systems analysts

• Information systems use

• Data entry

• Data control clerks

• Programmers

• Computer operators

• Network management

• System administration

• Librarian

• Systems development and maintenance

• Change management

• Security administration

• Security audit

Following are some of the various positions within a computer system, the responsibilities of each, and
incompatible responsibilities of each.

• Systems analysts are responsible for reviewing the current system to make sure that it is meet-
ing the needs of the organization, and when it is not, they provide the design specifications for the
new system to the programmers. Systems analysis consists of organizational analysis to learn

210 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.2. Systems Controls and Security Measures

about the current system’s strengths and weaknesses; identification of users’ requirements for the
new system; identifying the system requirements to fulfill the information needs of the users;
evaluating alternative designs using cost-benefit analysis; and preparing a systems analysis report
documenting the design of the proposed system and its specifications. Systems analysts should
not do programming, nor should they have access to hardware, software or data files.

• Programmers write, test, and document the systems. They are able to modify programs, data
files, and controls but should not have access to the computers and programs in actual use for
processing. For instance, if a bank programmer were allowed access to actual live data and had
borrowed money from the bank, the programmer could delete his or her own loan balance while
conducting a test. When programmers must do testing, they should work with copies of records
only and should not have the authority, opportunity, or ability to make any changes in master
records or files.

• Computer operators perform the actual operation of the computers for processing the data. They
should not have programming functions and should not be able to modify any programs. Their job
responsibilities should be rotated so no one operator is always overseeing the running of the same
application. The most critical segregation of duties is between programmers and com-
puter operators.

• The data control group receives user input, logs it, monitors the processing of the data, recon-
ciles input and output, distributes output to authorized users, and checks to see that errors are
corrected. The group personnel also maintain registers of computer access codes and coordinate
security controls with other computer personnel. They must keep the computer accounts and ac-
cess authorizations current at all times. They should be organizationally independent of
computer operations. Systems control personnel, not computer operators, should be responsible
for detecting and correcting errors.

Transaction authorization: No personnel in the Information Systems group should have author-
ity to initiate or authorize any entries or transactions. Users should submit a signed form with each
batch of input data to verify that the input has been authorized and that the proper batch control
totals43 have been prepared. Data control group personnel should verify the signatures and batch
control totals before submitting the input for processing to prevent a payroll clerk, for instance,
from submitting an unauthorized pay increase.

• Data conversion operators perform tasks of converting and transmitting data. They should have
no access to the library or to program documentation, nor should they have any input/output
control responsibilities.

• Librarians maintain the documentation, programs, and data files. The librarian should restrict
access to the data files and programs to authorized personnel at scheduled times. The librarian
maintains records of all usage, and those records should be reviewed regularly by the data control
group for evidence of unauthorized use. Librarians should have no access to equipment.

• Only authorized people should be able to call program vendor technical support departments. If
vendor-supplied systems are used, the vendors’ technical support area should have a means of
identifying callers and should give technical instructions for fixing problems only to employees who
are authorized to receive such instructions.

43
Batch control totals are any type of control total or count applied to a specific group of transactions, such as total
sales revenue in a batch of billings. Batch control totals are used to ensure that all input is processed correctly by the
computer. In batch processing, items are batched in bundles of a preset number of transactions. If a batch consists of
financial transactions, a batch control document that goes with the batch includes the bundle number, the date and the
total monetary amount of the batch. As the computer processes the batch, it checks the batch control total (the total
monetary amount) for the batch and compares the processed total with the batch control total. If they match, the batch
is posted. If they do not, the posting is rejected, and the difference must be investigated. Batch control totals can also
be calculated and used for non-financial fields in transactions. For instance, a batch control total might be the total hours
worked by employees.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 211
E.2. Systems Controls and Security Measures CMA Part 1

• The database administrator controls access to various files, making program changes, and mak-
ing source code details available only to those who need to know.

• The location of any off-site storage facilities should be known by as few people as possible.

• No Information Systems personnel should have access to any assets that are accounted for in the
computer system.

Question 56: The process of learning how the current system functions, determining the needs of users
and developing the logical requirements of a proposed system is referred to as:

a) Systems analysis.

b) Systems feasibility study.

c) Systems implementation.

d) Systems maintenance.

(CMA Adapted)

Question 57: The most critical aspect of separation of duties within information systems is between:

a) Programmers and computer operators.

b) Programmers and project leaders.

c) Programmers and systems analysts.

d) Systems analysts and users.

(CMA adapted)

Administrative Controls – Supervision of Control Procedures


Adequate supervision of personnel involved in systems control procedures is necessary. Supervisors can
spot weaknesses, correct errors, and identify deviations from standard procedures. Without adequate su-
pervision, controls may be neglected or deliberately by-passed.

General Operating Procedures

Note: This is the second of three categories of general controls.

Policies and procedures should be established formally in writing and approved by an appropriate manage-
ment level. Accountabilities and responsibilities should be clearly specified.

Procedures should be documented and kept up to date for all IT operations, including network operations.
Procedure documentation should include:

• The start-up process


• Job scheduling
• Setup of processing jobs
• Instructions for running jobs
• Processing continuity during operator shift changes
• Operations logs
• Backup and recovery procedures
• Procedures that ensure connection and disconnection of links to remote operations

212 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.2. Systems Controls and Security Measures

Task descriptions should be written for each job function; personnel should be trained in their jobs; assigned
duties should be rotated periodically for key processing functions.

Physical safeguards should be established over forms such as negotiable instruments and over sensitive
output devices such as signature cartridges. Sequential numbers on individual forms should be printed in
advance so missing forms can be detected.

Turnaround documents should be used whenever appropriate. A turnaround document is a document


created by a computer, then some additional information is added to it and it is returned to become an
input document to the computer. The documents are printed with Optical Character Recognition (OCR)
fonts so that they can be read by the computer when they are scanned and thus only the added information
needs to be keyed in. Some examples of turnaround documents are invoices with a top section that the
customer tears off and returns with payment (with the amount paid filled in); or magazine subscription
renewal notices that the subscriber returns with renewal instructions. The use of a turnaround document
limits the chances of input errors occurring and reduces the need for manual data entry.

General operating procedures also specify the process to be followed in system and program development
and changes. The process to follow should be documented in order to provide reasonable assurance that
development of, and changes to, computer programs are authorized, tested, and approved prior to the use
of the program.

An outline of system and program development and change controls is included later in this section.

Software, Hardware, and Access Controls

Note: This is the third of three categories of general controls.

Software Controls
Software controls monitor the use of the software and prevent unauthorized access to it. Program security
controls are used to prevent unauthorized changes to applications and systems. Control of system software
is particularly important as system software performs overall control functions for the application programs.
System software controls are used for compilers, utility programs, operations reporting, file handling and
file setup, and library activities.

Example: Programs are written in source code, the language the programmer uses for coding the
program, and they also exist in object code, the machine language the processor can understand. The
source code file is converted to object code by means of a program called a compiler, and the object
code, not the source code, is what actually runs on the computer.

The above fact is important because although in theory the source code and the object code
should correspond, the computer does not require them to correspond. It would be possible for
a knowledgeable person to make a copy of the source code, rewrite portions of the instructions, compile
the modified source code into a new object code file for use by the computer, and then destroy the
modified source code file, leaving the authorized source code file unchanged. The result is that the
executable object code—the actual instructions used by the computer—will not match the authorized
source code.

If controls over compiling and cataloging activities are not adequate, this security weakness can be used
to commit computer fraud. Despite the strongest internal controls over day-to-day operations in user
departments, a fraudulent change to a program could divert company funds to an individual, and the
fraud could continue for some time without being detected.

A reliable, automatically generated report of compilation dates of the programs being used should
be available and consulted regularly by management and system auditors.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 213
E.2. Systems Controls and Security Measures CMA Part 1

Hardware Controls
Hardware controls include keeping the computer equipment physically secure.

• Computer equipment in the computer center should be kept in a locked room and access limited
to computer operations personnel (see Access Controls, next).

• Computer equipment should be protected from extremes of temperature and humidity.

• The location of the computer center should be in a place where it is protected from fire and other
natural disasters as much as possible.

• The computer center should be equipped with smoke and water detectors, fire suppression devices,
burglar alarms, and surveillance cameras monitored by security personnel.

• A defined backup procedure should be in place, and the usability of the backups should be verified
regularly.

Insurance is the last resort, to be called upon only if all else fails, because it does not actually protect from
loss but rather compensates for loss after it occurs. Insurance policies for computer damages are usually
restricted to actual monetary losses suffered, and monetary loss is difficult to assess. For example, com-
puters may have a market value that is far less than the value of their services to the company.

Computer hardware frequently contains mechanisms to check for equipment malfunctions, another aspect
of hardware controls. For example:

• Parity checks are used to detect alteration of bits within bytes during processing caused by equip-
ment malfunctions.

• Echo checks verify that the data received is identical to the data sent. The data is retransmitted
back to the sending device for comparison with the original.

Computer networks require special controls due to the decentralized nature of the hardware.

• Checkpoint and rollback recovery processing should be used to enable recovery in case of a
system failure. Checkpoint control procedures are performed several times per hour, and during
that time, the network system will not accept posting. The system stops and backs up all the data
and other information needed to restart the system. The checkpoint is recorded on separate
media. Then, if a hardware failure occurs, the system reverts (“rolls back”) to the last saved copy,
restarts, and reprocesses only the transactions that were posted after that checkpoint.

• Routing verification procedures protect against transactions routed to the wrong computer
network system address. Any transaction transmitted over the network must have a header label
identifying its destination. Before sending the transaction, the system verifies that the destination
is valid and authorized to receive data. After the transaction has been received, the system verifies
that the message went to the destination code in the header.

• Message acknowledgment procedures can prevent the loss of part or all of a transaction or
message on a network. Messages are given a trailer label, which the receiving destination checks
to verify that the complete message was received.

Business continuity planning and disaster recovery are also part of hardware controls. Business con-
tinuity planning and disaster recovery are covered in a separate topic later.

Access Controls
Access controls include logical security and physical security.

Logical security consists of access and ability to use the equipment to protect it from damage or theft
and data security controls to ensure that data files are not subject to unauthorized access, change, or
destruction.

214 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.2. Systems Controls and Security Measures
efham CMA
Logical security includes access controls for users to limit actions they can perform; authentication pro-
cesses to verify the identity of users; and cryptographic techniques such as encryption of messages and
digital signatures.

• Unauthorized personnel, online connections and other system entry ports should be prevented
from accessing computer resources.

• Passwords should be changed regularly for all those authorized to access the data. Procedures
should be established for issuing, suspending, and closing user accounts, and access rights should
be reviewed periodically.

• All passwords should be issued with levels of authority that permit the users to access only the
data that they need to be able to access in order to do their jobs. For example, a person who does
invoicing needs access to the invoicing module of the accounting program but does not need access
to the general ledger module. The person who does receiving needs access to the purchase order
module but not to invoicing.

• Dual access and dual control should be established to require two independent, simultaneous ac-
tions before processing is permitted.

• Only authorized software from known sources should be allowed to be used in the system. Author-
ized software should be verified to be free of viruses and other malware before it is used.44

Data security controls are an important sub-set of logical security. Data security controls are established
to prevent access to data files without authorization and to prevent unauthorized or accidental change or
destruction.

• Online input and real-time systems are vulnerable because they can be accessed remotely. When
data can be input online, unauthorized input entry must be prevented. Systems that enable online
inquiries must have data files secured.

o Terminals can be physically restricted to permit access only to authorized personnel.

o Passwords should be assigned only to authorized personnel.

o Authorities should be assigned to passwords. For example, only individuals who are authorized
to update specific files should have the ability to do so, while others are able to read the files
or are denied access altogether.

• Data files can be controlled more easily in batch systems than in online input and real-time systems
because access in batch systems is limited to operators running the batch jobs. Data files that are
maintained offline should be secured in locked storage areas with tight controls for release only
for authorized processing. Usage logs and library records can be maintained for each offline storage
device.

File security and storage controls are an important part of data security controls.

File security control procedures include:

• Labeling the contents of discs (CDs, DVDs, external hard drives), tapes, flash drives or memory
cards, and any other removable media, both externally and internally as part of the data file.

• The read-only file designation is used to prevent users from altering or writing over data.

• Database Management Systems use lockout procedures to prevent two applications from
updating the same record or data item at the same time.

44
“Malware” is short for “malicious software.” It is software that is intended to damage or disable computers and
computer systems.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 215
E.2. Systems Controls and Security Measures CMA Part 1

Note: A deadly embrace occurs when two different applications or transactions each have a
lock on data needed by the other application or transaction. Neither process is able to proceed,
because each is waiting for the other to do something. In these cases, the system must have a
method of determining which transaction goes first, and then it must let the second transaction
be completed using the information updated by the first transaction.

• The librarian’s function is particularly critical, because documentation, programs and data files are
assets of the organization and require protection the same as any other asset would. The data files
contain information critical to the enterprise, such as accounting records. Although backup proce-
dures could reconstruct lost or damaged data, it is less costly to prevent a data loss than to repair
it. Furthermore, confidential information is contained in the data files and must be protected from
misuse by unauthorized individuals.

• Protection of program documentation is critical. Data can be changed within a file by someone who
knows how to do it, and technical manuals containing file descriptions are one way to get the
necessary information. Only authorized people who have the responsibility to repair data files that
may become corrupt should have access to technical manuals.

Logical security also includes Internet security (firewalls) and virus protection procedures. Internet security
is covered in a separate topic later.

Physical security involves protecting the physical assets of the computer center: the hardware, periph-
erals, documentation, programs, and data files in the library.

• The computer processing center should be in a locked area, and access to it should be restricted
to authorized persons.
• Servers and associated peripheral equipment should be kept in a separated, secure room with bars
on the windows and use of blinds or reflective film on the windows for heat blocking as well as
physical protection.
• Hardware components should be monitored to prevent them from being removed from the prem-
ises.
• Offsite backup media should be secured.
• The location of wiring that connects the system should be known.
• Uninterruptible power supplies should be maintained.

Media library contents should be protected. Responsibilities for storage media library management should
be assigned to specific employees. Contents of the media library should be inventoried systematically, so
any discrepancies can be remedied and the integrity of magnetic media is maintained. Policies and proce-
dures should be established for archiving. Some means of accomplishing access restrictions are:

• Have company personnel wear color-coded ID badges with photos. People authorized to enter the
computer area are assigned an ID badge of a particular color.
• An IT member should escort visitors when they enter the computer facilities, and a visitor’s log
should be kept and reviewed regularly.
• With magnetic ID cards, each employee’s entry into and exit from the computer center can be
automatically logged.
• Biometrics such as fingerprints, voice verification, and so forth may also be used to identify a
person based on physical or behavioral characteristics.
• The door can be kept locked, and a person can enter only if “buzzed” in by the control person, who
permits only authorized people to enter.
• Keys may be issued to authorized personnel, or combination locks can be used to limit access. If
keys are used, they should be keys that cannot be easily duplicated, and locks need to be changed
periodically. If a combination lock is used, the combination should be changed periodically.

216 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.2. Systems Controls and Security Measures

Application Controls
Application controls focus on preventing, detecting and correcting errors in transactions as they flow
through the input, processing, and output stages of work in an information system. Application controls
are specific controls within an individual computer application such as payroll or accounts receivable. Ap-
plication controls focus on the following objectives:

1) Input completeness and updating completeness. All authorized transactions reach the computer
and are recorded in data files.

2) Input accuracy and updating accuracy. Data is captured accurately and correctly recorded in data
files.

3) Validity. Data is authorized before being input and processed to approve the appropriateness of
the transactions. The transactions reflect the actual event that took place.

4) Data is processed in a timely manner.

5) Data files are accurate, complete, and current.

6) Output is accurate and complete.

7) Records are available that track the progress of data from input to storage to output.

Below are some things that can go wrong and that adequate controls can prevent, detect and correct:

• Input loss can occur when transaction information is transmitted from one location to another.

• Input duplication can occur if an input item is thought to be lost and is recreated, but the original
item is subsequently found or was never actually lost.

• Inaccurate input in the form of typographical errors in numbers or in spelling can occur.

• Missing information makes the input incomplete.

• Unrecorded transactions can occur as accidental failures or can be the result of theft or embezzle-
ment.

• In a volume-processing environment, management authorization of every individual transaction


may not take place, allowing improper transactions to slip through.

• Automated transactions may be set up for regular orders or payments to suppliers. However,
unusual situations can call for special transactions, and when a special transaction is needed, au-
tomated transactions can cause problems.

• Output can be sent to the wrong people or may be sent too late to be used.

• Programming errors or clerical errors can result in incomplete processing.

• Processing may be delayed.

• Files can be lost during processing.

• Poor documentation and a loss of knowledgeable people can result in errors and omissions.

Application controls are classified as input controls, processing controls, and output controls.

Input Controls

Note: This is the first of three categories of application controls.

Input controls are controls designed to provide reasonable assurance that input entered into the system
has proper authorization, has been converted to machine-sensible form, and has been entered accurately
and completely. Input controls can also provide some assurance that items of data (including data sent
over communications lines) have not been lost, suppressed, added, or changed in some manner.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 217
E.2. Systems Controls and Security Measures CMA Part 1

Input is the stage with the most human involvement and, as a result, the risk of errors is higher in the
input stage than in the processing and output stages. Most errors in systems result from input errors.
If information is not entered correctly, the output has no chance of being correct. Processing might be done
perfectly, but if the input into the system is inaccurate or incomplete, the output will be useless. Effective
input controls are vital.

Input controls are divided into three classifications, though there is some overlap and some controls appear
in more than one classification. The three classifications are:

1) Data observation and recording

2) Data transcription

3) Edit tests

Data Observation and Recording

Note: This is the first of three categories of application controls – input controls.

One or more observational control procedures may be practiced:

• Feedback mechanisms are manual systems that attest to the accuracy of a document. For in-
stance, a sales person might ask a customer to confirm an order with a signature, attesting to the
accuracy of the data in the sales order. Feedback mechanisms include authorization, endorse-
ment, and cancellation.

• Dual observation means more than one employee sees the input documents. In some cases,
“dual observation” will mean a supervisor reviews and approves the work.

• Point-of-sale devices such as bar codes that are scanned can decrease input errors substantially.
In addition, point-of-sale devices eliminate the need to manually convert the data to machine-
readable format.

• Preprinted forms such as receipt and confirmation forms can ensure that all the data required
for processing have been captured. For example, if a form utilizes boxes for each character in an
inventory part number, it is more likely that the correct number of characters will be entered.

• Batch control totals should be used in the input phase for transactions grouped in batches to
track input as it travels from place to place before it reaches the computer, to make sure no input
is lost. In batch processing, items are batched in bundles of a preset number of transactions. Batch
control totals are any type of control total or count applied to a specific group of transactions, such
as total sales revenue in a batch of billings. Batch control totals are used to ensure that all input
is processed correctly by the computer. The application recalculates the batch total and compares
it with the manual batch total. Batches for which the control totals do not balance are rejected.

Example: For a batch that consists of financial transactions, the batch control document that
goes with the batch includes the bundle number, the date and the total monetary amount of
the transactions in the batch. As the computer processes the batch, it checks the batch control
total (the total monetary amount) for the batch and compares the processed total with the
batch control total. If they match, the batch is posted. If they do not match, the posting is
rejected and the difference must be investigated.

Batch control totals can also be calculated and used for non-financial fields in transactions. For
instance, a batch control total might be the total hours worked by employees.

Batch control totals do not work as well with real-time systems, because input is entered at remote
terminals sporadically and by different people. Transactions cannot be easily batched. However,
entries can and should be displayed on a screen for visual verification and checked against source

218 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.2. Systems Controls and Security Measures

documents. In addition, control totals can be created to reconcile with the hard copy source doc-
uments used for a group of transactions input by an individual.

Furthermore, information input can be checked against the database, and edit programs can be
used to make sure that each field has the proper format (see following topics).

• Transaction trails should be created by the system that show the date, terminal ID, and individual
responsible for the input, especially in a real-time system. All inputs should be logged to a special
file that contains these identifying tags that identify the transactions. Including this additional,
audit-oriented information along with original transaction data is called tagging.

• Transaction logs also provide a source of control totals.

Data Transcription

Note: This is the second of three categories of application controls – input controls.

Data transcription is the preparation of the data for processing. If input is entered from source docu-
ments, the source documents should be organized in a way that eases the input process.

The actual data input usually takes place at a workstation with a display terminal. A preformatted input
screen can assist in the transcription process. For example, a date field to be filled in would be presented
onscreen as __/__/____.

Format checks are used to verify that each item of data is entered in the proper mode: numeric data in a
numeric field, a date in a date field, and so forth.

Edit Checks

Note: This is the third of three categories of application controls – input controls.

Edit programs or input validation routines are programs that check the validity and accuracy of input
data. They perform edit tests by examining specific fields of data and rejecting transactions if their data
fields do not meet data quality standards. As each transaction is input into a real-time system it can be
edited, and the operator is notified immediately if an error is found. The system can be designed to reject
any further input until the error has been corrected or to print an error report for review by supervisory
personnel.

Statistics on data input and other types of source errors should be accumulated and reviewed to determine
remedial efforts needed to reduce errors.

Edit tests include:

• Completeness, or field, checks, which ensure that input has been entered into all required fields
and that the input is in the proper format. For example, a field check would not permit numbers
to be input into a field for a person’s name.

• Limit checks, which ensure that only data within predefined limits will be accepted by the system.
For example, the number of days worked in a week cannot exceed seven.

• Validity checks, which match the input data to an acceptable set of values or match the charac-
teristics of input data to an acceptable set of characteristics.

• Overflow checks, which make sure that the number of digits entered in a field is not greater than
the capacity of the field.

• Check digits, which determine whether a number has been transcribed properly. A check digit is
a number that is a part of an account or other type of number. The check digit is a function of the

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 219
E.2. Systems Controls and Security Measures CMA Part 1

other digits within the number, determined by a mathematical algorithm.45 If a digit in the account
number is keyed in incorrectly, the check digit will be incorrect, and the system will generate an
error message and refuse to accept the input. Check digits are commonly used in credit card
account and bank account numbers, and they are especially helpful in detecting transposition er-
rors. If an operator keys in a number incorrectly, the operator will get an error message such as
“invalid account number.”

• Key verification, or keystroke verification, is the process of inputting the information twice
and comparing the two results. Key verification is often used when changing a password, to confirm
that the password has been typed correctly. Key verification can also be used to require input of
the same information twice by different people.

• Hash totals are another type of control total. Hash totals are totals of nonmonetary information.
For example, if a batch contains data on receipts from accounts receivable customers, the sum of
all the customers’ account numbers might be computed to create a hash total. The sum is, of
course, useful only for control purposes. A hash total can be run on a group of records to be input
before processing or transmission and again after processing. If the hash total changes during
processing, it indicates something has changed or some transactions may be lost.

• Format checks check whether the input has been entered in the proper mode and within the
proper fields.

• Reasonableness checks compare input with other information in existing records and historical
information to detect items of data that are not reasonable.

• Numerical checks assure that numeric fields are used only for numeric data.

• Reconciliations are used to determine whether differences exist between two amounts that
should be equal. If differences are found, the differences are analyzed to detect the reason or
reasons, and corrections are made if necessary.

Corrections of errors present additional problems. Often, attempts to correct an error result in addi-
tional errors. Error reports need to be analyzed, the action required to make the correction needs to be
determined, the incorrect information needs to be reversed and correct information needs to be input.
Often, corrections are needed in multiple data files.

Inquiries of data or master files need to be designed so that an incorrectly keyed-in inquiry cannot change
the information in the file. For example, when a user is displaying a customer’s record and then the user
needs to locate a second customer’s record, typing the second customer’s name into the “Name” field that
is displayed and submitting it as an inquiry should not cause the name on the first customer’s record to be
changed to the second customer’s name.

Question 58: Routines that use the computer to check the validity and accuracy of transaction data
during input are called:

a) Operating systems.

b) Compiler programs.

c) Edit programs.

d) Integrated test facilities.

(CMA Adapted)

45
An algorithm is a step-by-step procedure for solving a problem, usually a mathematical problem, in a finite number
of steps.

220 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.2. Systems Controls and Security Measures

Question 59: The online data entry control called preformatting is:

a) A check to determine if all data items for a transaction have been entered by the terminal oper-
ator.

b) A program initiated prior to regular input to discover errors in data before entry so that the errors
can be corrected.

c) The display of an input screen with blanks for data items to be entered by the terminal operator.

d) A series of requests for required input data that requires an acceptable response to each request
before a subsequent request is made.

(CMA Adapted)

Question 60: Data input validation routines include:

a) Passwords.

b) Terminal logs.

c) Backup controls.

d) Hash totals.

(CMA Adapted)

Question 61: An employee in the receiving department keyed in a shipment from a remote terminal
and inadvertently omitted the purchase order number. The best systems control to detect this error
would be:

a) Completeness test.

b) Batch total.

c) Reasonableness test.

d) Sequence check.

(CMA Adapted)

Question 62: Edit checks in a computerized accounting system:

a) Are preventive controls.

b) Must be installed for the system to be operational.

c) Should be performed on transactions prior to updating a master file.

d) Should be performed immediately prior to output distribution.

(CMA Adapted)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 221
E.2. Systems Controls and Security Measures CMA Part 1

Processing Controls

Note: This is the second of three categories of application controls.

Processing controls are controls designed to provide reasonable assurance that processing has occurred
properly and that no transactions have been lost or incorrectly added. Processing controls prevent or dis-
courage the improper manipulation of data and ensure satisfactory operation of hardware and software.

Processing controls overlap with general controls because they include the physical security of the equip-
ment. At one time, processing controls were limited to the computer room. But with more and more
distributed processing taking place, processing controls are moving outside the room where the computer
equipment is located.

Access to the computer should be permitted only to people who are authorized to operate the equipment,
and operators should be given access only to information they need to set up and operate the equipment.

Processing controls fall into two classifications:

1) Data access controls - processing controls at the time of data access

2) Data manipulation controls - controls involving data manipulation later in the processing

Data Access Controls

Note: This is the first of two categories of application controls – processing controls.

Transmittal documents such as batch control tickets are used to control movement of data from the
source to the processing point or from one processing point to another. Batch sequence numbers are
used to number batches consecutively to make sure all batches are accounted for.

Batch control totals were mentioned as input controls, but they are also processing controls. Batch control
totals are any type of control total or count applied to a specific group of transactions, such as total sales
revenue in a batch of billings. Batch control totals are used to ensure that all input is processed correctly
by the computer. In batch processing, items are batched in bundles of a preset number of transactions. If
a batch consists of financial transactions, a batch control document that goes with the batch includes
the bundle number, the date and the total monetary amount of the batch. As the computer processes the
batch, it checks the batch control total (the total monetary amount) for the batch and compares the
processed total with the batch control total. If they match, the batch is posted. If they do not match, the
posting is rejected and the difference must be investigated. Batch control totals can also be calculated and
used for non-financial fields in transactions. For instance, a batch control total might be the total hours
worked by employees.

A hash total is another control that is both an input and a processing control. For instance, if a batch
contains data on receipts from multiple accounts receivable customers, the sum of all the customers’ ac-
count numbers might be computed to create a hash total. The hash total is compared with the total of the
same numbers computed during processing to make sure no items were lost during processing. A hash
total is useful only for control purposes, because a total of all the account numbers in a batch, for example,
is meaningless by itself.

A record count utilizes the number of transaction items and counts them twice, once when preparing the
transactions in a batch and again when performing the processing.

Data Manipulation Controls

Note: This is the second of two categories of application controls – processing controls.

Standard procedures should be developed and used for all processing.

222 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.2. Systems Controls and Security Measures

Examining software documentation, such as system flowcharts, program flowcharts, data flow
diagrams, and decision tables, can also be a control, because it makes sure that the programs are
complete in their data manipulation.

Computer programs are error tested by using a compiler, which checks for programming language errors.

Test data can be used to test a computer program.

System testing can be used to test the interaction of several different computer programs. Output from
one program is often input to another, and system testing tests the linkages between the programs.

Other tests of processing include the following.

• Batch balancing is comparing the items actually processed against a predetermined control total.

• Cross-footing compares the sum of the individual components to a total.

• A zero-balance check is used when a sum should be zero. All of the numbers are added together
and the total is compared with zero.

• Run-to-run totals are output control totals from one process used as input control totals over
subsequent processing. Critical information is checked to ensure that it is correct. The run-to-run
totals tie one process to another.

• Default option is the automatic use of a predefined value when a certain value is left blank in
input. However, a default option may be correct, or it may be an incorrect value for a particular
transaction, so the default should not be automatically accepted.

Question 63: In an automated payroll processing environment, a department manager substituted the
time card for a terminated employee with a time card for a fictitious employee. The fictitious employee
had the same pay rate and hours worked as the terminated employee. The best control technique to
detect this action using employee identification numbers would be a:

a) Hash total.

b) Batch total.

c) Subsequent check.

d) Record count.

(CMA Adapted)

Output Controls

Note: This is the third of three categories of application controls.

Output controls are used to provide reasonable assurance that input and processing have resulted in valid
output. Output can consist of account listings, displays, reports, files, invoices, or disbursement checks, to
name just a few of the forms output can take. Controls should be in place to make sure that the output is
sent to the right people, that it is accurate and complete, it is sent in a timely manner, and that the proper
reports are retained for the appropriate time period.

The output of the system is supervised by the data control group. Output controls consist of:

• Validating processing results

• Controls over printed output

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 223
E.2. Systems Controls and Security Measures CMA Part 1

Validating Processing Results

Note: This is the first of two categories of application controls – output controls.

Activity, or proof, listings that document processing activity provide detailed information about all
changes to master files and create an audit trail. The proof listings should be compared with the batch
control totals that went along with the input and processing functions in order to confirm that all of the
transactions were processed correctly.

Transaction trails should be available for tracing the contents of any individual transaction record back-
ward or forward, and between output, processing, and source. Records of all changes to files should be
maintained.

Output totals should be reconciled with input and processing totals. A reconciliation is the analysis of
differences between two values that should be the same. The nature of the reconciling items is used to
identify whether differences are caused by errors or whether they are valid differences.

A suspense account is used as a control total for items awaiting further processing, such as a file of back-
ordered products awaiting receipt so they can be shipped to fulfill orders.

Output control also includes review of the processing and error logs by the control group to determine
whether all of the correct computer jobs were executed properly and for review of the output by the users.
A discrepancy report is a listing of items that have violated some detective control and need to be inves-
tigated, such as a list of all past-due accounts sent to the credit manager. End-of-job markers are printed
at the end of the report and enable the user to easily determine if the entire report has been received.

Upstream resubmission is the resubmission of corrected error transactions as if they were new transac-
tions, so that they pass through all the same detective controls as the original transactions.

Printed Output Controls

Note: This is the second of two categories of application controls – output controls.

Forms control, such as physical control over company blank checks, is one type of printed output control.
Checks should be kept under lock and key, and only authorized persons should be permitted access.

However, another control is needed with checks, because checks are pre-numbered. The preprinted
check number on each completed check must match the system-generated number for that
check, which may or may not be also printed on the check. The preprinted numbers on the checks
are sequential; the system-generated numbers also are sequential. The starting system-generated number
must match the pre-printed number on the first check in the stack, or all the check numbers in the whole
check run will be off. If the physical check number of each check in the check run does not match its check
number in the system, an investigation must be done because the starting number in the system should
be one more than the number of the last check legitimately issued. If the system number of the
first check printed does not match the preprinted number on the first check in the stack to be printed, one
or more blank checks could be missing.

Any form should be pre-numbered and controlled in the same manner as checks.

Companies are increasingly creating their own checks, using blank check stock and printers that print all of
their information as well as the MICR (Magnetic Ink Character Recognition) line as the check itself is printed.
The physical equipment used to create checks as well as the blank check stock must be strictly controlled.

Output control also concerns report distribution. For example, a payroll register with all the employees’
social security numbers and pay rates is confidential information and thus its distribution must be restricted.
Formal distribution procedures should be documented along with an authorized distribution list specify-
ing authorized recipients of output reports, checks, and other critical output. An adequate number of copies
of reports should be generated to permit only one report to be distributed to each person on the list. For a
confidential report, it is preferable to have a representative pick the report up personally and sign for it. If

224 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E E.2. Systems Controls and Security Measures

it is not possible for a representative to pick the report up personally, a bonded employee can be used to
hand deliver the reports. The employee’s supervisor should make random checks on the report distribution.

Confidential reports should be shredded when they are no longer needed.

Controls Classified as Preventive, Detective and Corrective


Just as financial controls can be classified as preventive, detective, and corrective, information systems
controls can be classified in the same manner.

• Preventive controls prevent errors and fraud before they occur. Examples of preventive con-
trols are segregation of duties, job rotation, training and competence of personnel, dual access
controls, authorization, approval, endorsement and cancellation, and preformatted input.

• Detective controls uncover errors and fraud after they have occurred. Examples of detective
controls are transmittal documents, batch control totals and other batch transmittal documents,
completeness checks, hash totals, batch balancing, check digits, limit checks, and validity checks.
The use of a turnaround document is also a detective control, because it checks on completeness
of input. Completeness-of-processing detective controls include run-to-run totals, reconciliations,
and use of suspense accounts. Error logs and exception reports are correctness of processing
detective controls and may also be completeness of processing controls.

• Corrective controls are used to correct errors. Upstream resubmissions are corrective controls.

Question 64: The reporting of accounting information plays a central role in the regulation of business
operations. The importance of sound internal control practices is underscored by the Foreign Corrupt
Practices Act of 1977, which requires publicly-owned U.S. corporations to maintain systems of internal
control that meet certain minimum standards. Preventive controls are an integral part of virtually all
accounting processing systems, and much of the information generated by the accounting system is
used for preventive control purposes. Which one of the following is not an essential element of a sound
preventive control system?

a) Documentation of policies and procedures.

b) Implementation of state-of-the-art software and hardware.

c) Separation of responsibilities for the recording, custodial and authorization functions.

d) Sound personnel practices.

(CMA Adapted)

Question 65: An advantage of having a computer maintain an automated error log in conjunction with
computer edit programs is that:

a) Less manual work is required to determine how to correct errors.

b) Better editing techniques will result.

c) The audit trail is maintained.

d) Reports can be developed that summarize the errors by type, cause and person responsible.

(CMA Adapted)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 225
E.2. Systems Controls and Security Measures CMA Part 1

Controls Classified as Feedback, Feedforward and Preventive


Another way of classifying information systems controls looks at them as feedback, feedforward, or
preventive controls.

• Feedback controls produce feedback that can be monitored and evaluated to determine if the
system is functioning as it is supposed to. Feedback controls are required in order to produce
usable information for end users. With the addition of feedback controls, a system becomes a self-
monitoring, self-regulating system.

A feedback loop is a part of a control system. It uses feedback to measure differences between
the actual output and the desired output. It then adjusts the operation according to those differ-
ences. Thus, it self-corrects. A self-monitoring system is sometimes called a cybernetic system.

For example, in a manufacturing situation where ingredients are being combined, computers may
monitor and control the mixing process, making adjustments as necessary to maintain the correct
proportions of each ingredient in the mix. In an accounting system, data entry displays or edit
sheets provide control of data entry activities, and accounting procedures such as reconciliations
provide feedback.

A report that summarizes variances from budgeted amounts us another example of a feedback
control.

• A feedforward control system may be used in addition to the feedback loop to provide better
controls. A feedforward system attempts to predict when problems and deviations will occur be-
fore they actually occur. It gives people guidance about what problems could occur, so they can
plan the necessary changes or actions to prevent the problem or deviation from occurring. Or, if it
is not possible to prevent the problem, a feedforward control can enable the company to minimize
the effects of the problem. A budget is a feedforward control. Policies, procedures, and rules are
also feedforward controls because they establish the way things are supposed to be done. When
people have detailed instructions and follow them, the chances of something going wrong may be
minimized.

• A preventive control attempts to stop a variance or problem from ever occurring, because it is
more cost effective to prevent a problem than it is to fix the problem after it occurs. Maintenance
is often given as an example of a preventive control. A preventive control is slightly different from
a feedforward control, in that the feedforward control simply tries to identify the potential problem,
whereas the preventive control attempts to prevent the problem from occurring. Segregation of
duties is a preventive control.

Question 66: Preventive controls are:

a) Usually more cost beneficial than detective controls.

b) Usually more costly to use than detective controls.

c) Found only in accounting transaction controls.

d) Found only in general accounting controls.

(CMA Adapted)

226 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E System and Program Development and Change Controls

System and Program Development and Change Controls


(A general operating procedure within general controls)
Systems development controls during the development stage of an information system enhance the ulti-
mate accuracy, validity, safety, security and adaptability of the new system’s input, processing, output and
storage functions.

Controls are instituted at the development stage for multiple reasons.

1) To ensure that all changes are properly authorized and are not made by individuals who lack
sufficient understanding of control procedures, proper approvals and the need for adequate testing.

2) To prevent errors in the resulting system that could cause major processing errors in data.

3) To limit the potential for a myriad of other problems during the development process and after its
completion.

Following are a few of the control considerations in systems development. This list is not exhaustive but is
presented to give candidates an idea of what is involved.

1) Statement of Objectives
A written proposal is prepared, including the need for the new system, the nature and scope of the project
and timing issues in terms of need and employee availability. A risk assessment is done to document
security threats, potential vulnerabilities, and the feasible security and internal control safeguards needed
to mitigate the identified risks.

• The written proposal should be reviewed by the IT Steering Committee.

• The nature and scope of the project should be defined clearly in writing.

A clear, written statement of objectives and a risk assessment at the Statement of Objectives stage can
limit the number of changes needed later on and shorten the time required to identify solutions and get
approvals.

2) Investigation and Feasibility Study of Alternative Solutions


Needs are identified and feasibility of alternative solutions are assessed, including the availability of required
technology and resources. A cost-benefit analysis is done, including both tangible costs and benefits (such
as hardware costs or increases in cash flow) and intangible costs and benefits (such as loss of customer
goodwill or better customer service).

• Questions to be answered in the cost-benefit analysis include whether the system will provide an
adequate payback; whether it will fit into the existing software environment; whether it will run
on existing hardware or whether a hardware upgrade will be needed; whether new storage media
will be required; whether the resources are available for the project; whether the application would
require extensive user or programmer training; and what effect it would have on existing systems.

• The feasibility study should include an analysis of needs, costs, implementation times, and poten-
tial risks.

• In evaluating possible solutions, criteria should be developed for consideration of in-house devel-
opment, purchased solutions and outsourcing options.

• The technological feasibility of each alternative for satisfying the business requirements should be
examined; and the costs and benefits associated with each alternative under consideration should
be analyzed.

• Key users should assist in the analysis and recommendations.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 227
System and Program Development and Change Controls CMA Part 1

• The development team should have a good knowledge of the solutions available and limit their
consideration to proven technology rather than experimenting with new technology, unless exper-
imentation is justified by the situation.

• Senior management should review the reports of the feasibility studies and approve or disapprove
proceeding with the project.

• For each approved project, a master plan should be created to maintain control over the project
throughout its life, which includes a method of monitoring the time and costs incurred.

The cost-benefit analysis done at the Investigation and Feasibility Study stage is extremely important as a
control tool. The cost-benefit analysis can reduce changes later on caused by the discovery of unexpected
costs. Furthermore, if the project is seriously flawed, it can be rejected at this stage before a major invest-
ment is made.

3) Systems Analysis
The current system is analyzed to identify its strong and weak points, and the information needs from the
new system, such as reports to be generated, database needs, and the characteristics of its operation are
determined.

• Business requirements satisfied by the existing system and those that the proposed system ex-
pects to attain should be clearly defined, including user requirements, specifications as to what the
new system is supposed to accomplish, and alternatives for achieving the specifications such as
in-house development versus a vendor package.

• Inputs, processing requirements, and output requirements should be defined and documented.

• All security requirements should be identified and justified, agreed upon, and documented.

• A structured analysis process should be used.

If the information required by the users is not clear, the new system cannot possibly support the business
process, leading to delays in implementation and additional costs to redesign the system.

4) Conceptual Design
The conceptual design is what the users are expecting the system to be. Systems analysts work with users
to create the design specifications and verify them against user requirements. System flowcharts and report
layouts are developed. The applications, network, databases, user interfaces, and system interfaces are
designed.

• Design specifications should be reviewed and approved by management, the user departments,
and senior management.

• Detailed program specifications should be prepared to ensure that program specifications agree
with system design specifications.

• Data elements are defined. Each field in each file is listed and defined.

• The file format should be defined and documented for the project to ensure that data dictionary
rules are followed.

• All external and internal interfaces should be properly specified, designed, and documented.

• An interface between the user and the machine that is easy to use and contains online help func-
tions should be developed.

• Mechanisms for the collection and entry of data should be specified for the development project.

• Adequate mechanisms for audit trails should be developed for the selected solution that provide
the ability to protect sensitive data.

228 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E System and Program Development and Change Controls

Benefits of properly-completed conceptual design include reduced interoperability problems, better inte-
gration with existing systems, and reduced costs for modifications later.

5) Physical Design
The physical system design involves determining the workflow, what and where programs and controls are
needed, the needed hardware, backups, security measures, and data communications.

• Hardware and software selection should identify mandatory and optional requirements. The po-
tential impact of new hardware and software on the performance of the overall system should be
assessed.

• An up-to-date inventory of hardware and software infrastructure should be available.

• Acquisition policies and practices should be clearly understood, and the selection process should
focus on using reusable components.

• Contracts with suppliers should include a definition of acceptance criteria and procedures, and
dependency on single-source suppliers should be managed.

• If a vendor package or packages are to be used, they should be evaluated rigorously. Factors to
consider include the stability of the vendor, how long the system has been on the market, whether
it has a base of users and the satisfaction level of current users of the system, the vendor’s quality
control standards, the adequacy of the documentation, the availability of vendor technical support,
and flexibility of the system such as whether it has a report writer that users can use to develop
reports themselves. Also, the system’s processing speed on the organization’s systems is a con-
sideration.

• Because only authorized people should be able to call vendor technical support departments, the
evaluation of a vendor system should include inquiries as to the means the vendor has to identify
callers to its technical support area and determine whether the caller has authority to receive
technical instructions for fixing problems.

• Performance and capacity requirements should be duly considered.

• Key requirements should be prioritized in case of possible scope reductions.

Benefits of proper physical design planning include reduced delays later due to inadequate infrastructure.

6) Development and Testing


The design is implemented into source code, the technical and physical configurations are fine-tuned, and
the system is integrated and tested. Data conversion procedures are developed.

• Programs are coded according to the specifications developed in the systems design and develop-
ment stage.

• Procedures should provide for a formal evaluation and approval, by management of the user de-
partment or departments and management of the IT function, of work accomplished and test
results in each phase of the cycle before work on the next phase begins.

• There should be a separation between development activities and testing activities.

• A formal process for handover from development to testing to operations should be defined.

• Resources should be available for a separate testing environment that reflects as closely as possi-
ble the live environment, and sufficient time should be allowed for the testing process.

• Parallel or pilot testing should be performed, and criteria for ending the testing process should be
specified in advance.

• Testing should be done both of the individual application and of the application within the system.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 229
System and Program Development and Change Controls CMA Part 1

• An independent group should do the testing and try to make the system fail.

• Both in-house systems and vendor packages should be tested.

• Software from a vendor should be tested on a stand-alone (non-networked) computer before being
distributed for general use because a vendor’s software could be infected with a virus.

Testing is the final check to make sure the system performs as it should.

7) System Implementation and Conversion


The site is prepared, equipment is acquired and installed, and conversion procedures, including data con-
version, are implemented. Documentation is completed and users are trained.

• An implementation plan should be prepared, reviewed and approved by relevant parties and also
used to measure progress. The plan should include site preparation, equipment acquisition and
installation, user training, installation of operating software changes and implementation of oper-
ating procedures and conversion procedures.

• Data conversion controls such as record counts, reviews of reports and other types of reconcilia-
tions are required.

• The degree and form of documentation required is agreed upon and followed in the implementa-
tion. Documentation will include

o System documentation, which is narrative descriptions, flowcharts, input and output


forms, file and record layouts, controls, authorizations for any changes and backup proce-
dures.

o Program documentation, or descriptions of the programs, program flowcharts, program


listings of source code, input and output forms, change requests, operator instructions, and
controls.

o Operating documentation, the information about the performance of the program.

o Procedural documentation, providing information about the master plan and the han-
dling of files.

o User documentation, including all the information a user will need in order to use the
program.

• Standard operating procedures should be documented, distributed, and maintained using


knowledge management, workflow techniques, and automated tools.

• Staff of the user departments and the operations group of the IT function should be trained in
accordance with the training plan.

• Formal evaluation and approval of the test results and the level of security for the system by
management of the user department and the IT function should cover all components of the infor-
mation system.

• Before the system is put into operation, the user should validate its operation as a complete prod-
uct under conditions similar to the application environment.

• The decision should be made as to whether the new system will be implemented using a parallel
conversion (running both the old and the new systems together for a period of time), a phased
conversion (converting only parts of the application at a time or only a few locations at a time),
pilot conversion (the new system is tested in just one work site before full implementation), or a
direct conversion (changing over immediately from the old system to the new).

Documentation provides a basis for effective operation, use, audit, and future system enhancements. A
detailed record of the system’s design is necessary in order to install, operate, or modify an application and

230 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E System and Program Development and Change Controls

for diagnosing and correcting programming errors. It provides a basis for reconstruction of the system in
case of damage or destruction.

Benefits of good documentation and other controls at this stage are more seamless integration of the new
system into existing business processes and greater user proficiency and satisfaction.

8) Operations and Maintenance


The system is put into a production environment and used to conduct business. Follow-up occurs to deter-
mine whether weaknesses in the previous system have been eliminated and whether or not any new
problems have arisen. Continuous monitoring and evaluation take place to determine what is working and
what needs improvement, to support continuous improvement of the system. Maintenance includes modi-
fying the system as necessary to adapt to changing needs, replacing outdated hardware as necessary,
upgrading software, and making needed security upgrades.

• The system is audited periodically to make sure it continues to operate properly.

• A maintenance process is utilized to correct errors.

• A process should be in place to manage coordination between and among changes, recognizing
interdependencies.

• Production duties should be segregated from development duties.

• Maintenance personnel should have specific assignments, their work should be monitored, and
their system access rights should be controlled.

• Any modifications to the system should be authorized by user management, made in accordance
with the same standards as are used for system development, and should be tested and approved
by the user and Information Systems management. Senior management should approve major
projects.

• When changes are being tested, they should be tested not only by using correct information, but
also by using incorrect information to make sure that the program will detect any errors and has
the necessary controls.

• Whenever system changes are implemented, associated documentation and procedures should be
updated accordingly.

• For a vendor package, maintenance procedures are of concern from a systems control standpoint.
Updates released by the vendor should be installed on a timely basis. For an organization with
integrated software, releases must be kept compatible. If one portion of the system is upgraded
and another part is not, the two systems may no longer interface properly.

• If vendor-supplied software has had custom changes made to the vendor’s source code and the
changes are not properly reinstalled on top of new releases, erroneous processing can result. The
organization should maintain change controls to verify that all custom changes are properly iden-
tified. A good audit trail of all program changes is necessary. Another concern with vendor update
releases when in-house changes have been made is that the changes may need to be not only
reinstalled, but completely rewritten. The changes made to the prior release of the program might
not work properly with the vendor’s new release.

• Heavy modification of vendor code with no intention of installing new vendor releases because of
the necessity to reinstall the modifications should be avoided, because the system becomes es-
sentially an in-house system without the benefit of vendor support.

The benefits include reduced errors and disruptions due to poorly-managed changes, reduced resources
and time required for changes, and reduced number of emergency fixes.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 231
Internet Security CMA Part 1

Internet Security
(A logical security access control within general controls)
Once a company is connected to an outside network (usually the Internet), a number of additional security
issues must be properly addressed. The company must make sure that the policies that it puts in place
allow the intended and authorized users to have access to the network as needed. However, accessibility
also creates vulnerability.

Electronic eavesdropping can occur if computer users are able to observe transmissions intended for
someone else. Therefore, organizations must ensure that information sent over a network is properly pro-
tected to maintain the confidentiality of company information. Furthermore, the company must ensure that
company files cannot be accessed or changed without authorization.

At a minimum, the system should include user account management, a firewall, anti-virus protection,
and encryption.

• User account management is the process of assigning people accounts and passwords. In order
for user account management to be as effective as possible, the company must keep the accounts
and passwords up-to-date. Inactive accounts should be eliminated and active passwords changed
frequently.

• A firewall serves as a barrier between the internal and the external networks and prevents unau-
thorized access to the internal network. A properly configured firewall makes a computer’s ports
invisible to port scans.46 In addition to protecting a computer from incoming probes, a firewall can
also prevent backdoor47 applications, Trojan horses,48 and other unwanted applications from send-
ing data from the computer. Most firewalls will usually prepare a report of Internet usage, including
any abnormal or excessive usage and attempts to gain unauthorized entry to the network. A fire-
wall can be in the form of software directly installed on a computer, or it can be a piece of hardware
installed between the computer and its connection to the Internet.

• Antivirus software, regularly updated with the latest virus definitions, is the best defense against
viruses, Trojan horses, and worms. Antivirus software recognizes and incapacitates viruses before
they can do damage. Users must keep their antivirus software up-to-date, however, because new
viruses appear constantly. Programs that specifically defend against Trojan horses are also avail-
able.

• Encryption is the best protection against traffic interception resulting in data leaks. Encryption
converts data into a code and then a key is required to convert the code back to data. Unauthorized
people can receive the coded information, but without the proper key they cannot read it. Thus,
an attacker may be able to see where the traffic came from and where it went but not the content.

46
A “port” is a place where information goes into and out of a computer. A “port scanner” is software that probes a
computer for open ports. Port scanners are used legitimately by administrators to verify security policies of their networks
and illegitimately by attackers to exploit vulnerabilities.
47
A “backdoor” is an undocumented means to gain access to a program, online service or an entire computer system
by bypassing the normal authentication process. A backdoor is a potential security risk.
48
A Trojan horse is “any program that does something besides what a person believes it will do.” A Trojan horse can
appear to be something desirable, but in fact it contains malicious code that, when triggered, will cause loss or even
theft of data. A typical example of a Trojan horse is a program hidden inside of a humorous animation that opens a
backdoor into the system. Another example of a Trojan horse is commercial software that collects data on the person
running the program and sends it back to the originating company without warning the target.

232 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internet Security

Viruses, Trojan Horses, and Worms


A computer virus is a program that alters the way another computer operates. Viruses can damage pro-
grams, delete files or reformat the hard disk. Other viruses do not do damage but replicate themselves and
present text, video, and audio messages. Although these other viruses may not cause damage directly,
they create problems by taking up computer memory and causing erratic behavior or system crashes that
can lead to data loss.

To be considered a virus, a virus must meet two criteria:

1) It must execute itself. A virus often places its own code in the path of the execution of another
program.

2) It must replicate itself. A virus can replace other executable files with a copy of the virus-infected
file.

A virus can be received from an infected disk, a downloaded file, or an email attachment, among other
places.

A Trojan horse is different from a virus. A very important distinction between Trojan horses and viruses
is that Trojan horses do not replicate themselves, whereas viruses do. The purpose of a Trojan horse is
not to spread like a virus, but to have a particular target — a particular computer — on which to run a
program. A strict definition of a Trojan horse is, “any program that does something besides what a person
believes it will do.” A Trojan horse can appear to be something desirable, but in fact it contains malicious
code that, when triggered, will cause loss or even theft of data. A typical example of a Trojan horse is a
program hidden inside of a humorous animation that opens a backdoor into the system. Another example
of a Trojan horse is commercial software that collects data on the person running the program and sends
it back to the originating company without warning the target.

A computer user can get a Trojan horse only by inviting it into his or her computer. Two examples are by:

1) Opening an email attachment.

2) Downloading and running a file from the Internet. Many mass-mailing worms are considered Trojan
horses because they must convince someone to open them. The SubSeven server, which is soft-
ware that lets an attacker remotely control any computer it is installed on, is an example of a
program typically embedded in a Trojan horse.

A worm is a program that replicates itself from system to system without the use of any host file. The
difference between a worm and a virus is that the worm does not require the use of an infected host file,
while the virus does require the spreading of an infected host file. Worms generally exist inside of other
files, often Word or Excel documents. However, worms use the host file differently from viruses. Usually
the worm releases a document that has the “worm” macro inside the document. The entire document
spreads from computer to computer, so the entire document is, in essence, the worm.

A virus hoax is an email telling the recipient that a file on his or her computer is a virus when the file is
not a virus. Such an email will tell recipients to look on their systems for a file with a specific name and, if
they see it, to delete it because the file contains a virus that is unrecognizable by anti-virus programs.
Everyone with the targeted operating system will find that file because it is a system file that is needed for
the computer to operate correctly. Someone who believes the virus hoax email and deletes the file will find
that the computer malfunctions afterward.

Note: The difference between a virus and a Trojan horse is that a virus replicates itself, but a Trojan
horse does not.

The difference between a virus and a worm is that the virus requires an infected host file in order to
replicate itself, while the worm can replicate itself without a host file.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 233
Internet Security CMA Part 1

Cybercrime
The Internet, online communications, and e-business are all subject to computer crime, and the threat is
growing every day.

Cyberattacks can be perpetrated by criminals, international adversaries, and terrorists. Adversaries target
the nation’s critical infrastructure and critical processes. Companies are targeted for competitive infor-
mation and sensitive corporate data, including consumer credit card information and personal identification
information. Universities are targeted to steal their research and development data. Individuals are targeted
by identity thieves and children are targeted by online predators.

In the U.S., the Cyber Division of the FBI is responsible for investigating cyberattacks. The FBI maintains
specially trained cyber squads at FBI headquarters and in each of its 56 field offices. In addition, Cyber
Action Teams are available to travel around the world to assist in computer intrusion cases, providing
investigative support and helping to answer critical questions that can move a hacking case forward. In
addition, the FBI has 93 Computer Crimes Task Forces nationwide that work with federal, state, and local
counterparts.

The FBI’s priorities are:

• Computer and network intrusions

• Ransomware

• Identity theft

• Online predators

Some specific computer crimes include:

• Copyright infringement such as the illegal copying of copyrighted material, whether intellectual
property, such as computer programs or this textbook, or entertainment property such as music
and movies.

• Denial of Service (DOS) attacks in which a website is accessed repeatedly so that other, legiti-
mate users cannot connect to it.

• Intrusions and theft of personal information, including emails, for gain.

• Phishing, a high-tech scam that uses spam email to deceive consumers into disclosing their credit
card numbers, bank account information, Social Security numbers, passwords or other sensitive
personal information.

• Installation of malware on a computer without the user’s knowledge. An example of malware is


a keylogger that records every keystroke and sends it back to the hacker. Keylogging software has
been used to gather bank information, credit card information, and passwords. Other malware
turns a PC into a “zombie,” giving hackers full control over the machine. Hackers set up “botnets”
— networks consisting of millions of zombies — that can be made to each send out tens of thou-
sands of spam emails or emails infected with viruses, and the computer users do not even know
it is happening.

• In tech support scams, scammers use fake warning popup boxes on users’ computers to
convince potential victims that their computers suffer from security problems and to offer a toll-
free number to call for a “Microsoft” or “Apple” support line. Users who believe it and call the
number are usually told to visit a website and download a “remote administration tool” that gives
the scammer access to their computer so the scammer can “run tests” on the computer. The
scammer then presents supposed “results” showing the computer infected with malware or having
been breached by hackers and offers to fix it for a fee of several hundred dollars. Users who fall
for the scheme are out the “support fee.” Domain hosting companies usually discover the fraud
and remove the websites within a few days, and the victim has nowhere to turn for compensation.

234 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internet Security

Cybercriminals’ Tools
Port scanners. A port scanner is software that can probe a server or a computer for ports that are open.
Although port scanners have legitimate uses, they are also used by hackers to locate and exploit vulnera-
bilities. Using port scans, hackers can look for a particular make of computer or a particular software
program, because they know of weaknesses in those computers or programs that they can exploit. Once a
hacker has identified a vulnerable computer or software application, the hacker can leave a back door
open in the computer in order to re-enter it at any time. If the original entry point is detected and closed,
the back door functions as a hidden, undetected way back in.

Sniffers. A sniffer is a piece of software that grabs all of the traffic flowing into and out of a computer
attached to a network. Like port scanners, sniffers have legitimate as well as illegitimate uses. Intrusion
Detection Systems (IDS) use sniffers legitimately to match packets against a rule set designed to flag things
that appear malicious or strange. Network utilization and monitoring programs often use sniffers to gather
data necessary for metrics and analysis. However, sniffers are also used to gather information illegally.

Many personal computers are on Local Area Networks (LANs), meaning they share a connection with several
other computers. If a network is not switched (a switch is a device that filters and forwards packets
between segments of the LAN), traffic intended for any machine on a segment of the network is broadcast
to every machine on that segment. Thus, every computer actually sees the data traveling to and from each
of its neighbors but normally ignores it. The sniffer program tells a computer to stop ignoring all the traffic
headed to other computers and instead pay attention to that traffic. The program then begins a constant
read of all information entering the computer.

Anything transmitted in plain text over the network is vulnerable to a sniffer — passwords, web pages,
database queries and messaging, to name a few. Once traffic has been captured, hackers can quickly
extract the information they need. The users will never know their information has been compromised,
because sniffers cause no damage or disturbance to the network environment.

Other tools of hackers include:

• Password crackers, which is software that creates different combinations of letters and numbers
in order to guess passwords.

• Logic bombs or errors in the logic of computer programs that result in the destruction of computer
data or a malicious attack when specific criteria are met.

• Buffer overflow, which sends too much data to the buffer in a computer’s memory, crashing it
or enabling the hacker to gain control over it.

Some computer crime tactics involve efforts in person as well as computer activities. Tactics involving
personal effort include social engineering and dumpster diving. Social engineering involves deceiving
company employees into divulging information such as passwords, usually through a fraudulent email but
it may be through something as simple as a telephone call. Dumpster diving is sifting through a company’s
trash for information that can be used either to break into its computers directly or to assist in social
engineering.

Another online scam is directed against companies that advertise on search engines on a “pay-per-click”
basis. Google is probably the best-known example of a search site that charges advertisers each time a
visitor clicks on an ad link. In one version of this scam, a competitor will write a software program that
repeatedly clicks on a business’s online ads in order to run up its advertising charges. Ultimately, after too
many clicks within a 24-hour period, the ad is pushed off the search engine site, resulting in lost business
for the company along with the inflated advertising fees.

However, outsiders are not the only ones who commit computer crimes against a company. Insiders—or
company employees—are a primary source of trouble. Employees who are planning to leave one employer
and go to work for a competitor can use their company email to transmit confidential information from the
current employer to the future employer or can carry it out on a flash drive.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 235
Internet Security CMA Part 1

Insider crime can also include using the company computer for private consulting, personal financial busi-
ness, playing video games on company time or browsing pornography sites. A legitimate use of sniffers,
described earlier, is monitoring network usage to reveal evidence of improper use. Many businesses install
software that enables them not only to monitor their employees’ access to websites but also to block access
to certain websites. Improper use of the Internet and email at work can get an employee fired immediately.

Defenses Against Cybercrime


The best defense against port scans is a good firewall. A firewall serves as a barrier between the internal
and the external networks and prevents unauthorized access to the internal network. A properly configured
firewall makes a computer’s ports invisible to port scans. In addition to protecting a computer from incoming
probes, a firewall can also prevent backdoor applications, Trojan horses and other unwanted applications
from sending data from the computer. Most firewalls will usually prepare a report of Internet usage, includ-
ing any abnormal or excessive usage and attempts to gain unauthorized entry to the network. A firewall
can be in the form of software directly installed on a computer, or it can be a piece of hardware installed
between the computer and its connection to the Internet.

An organization may also use a proxy server, which is a computer and software that creates a gateway
to and from the Internet. The proxy server contains an access control list of approved web sites and handles
all web access requests, limiting access to only those sites contained in the access control list. A proxy
server enables an employer to deny its employees access to sites that are unlikely to have any productive
benefits. The proxy server also examines all incoming requests for information and tests them for authen-
ticity, thus functioning as a firewall. The proxy server can also limit the information stored on it to
information the company can afford to lose, and if the proxy server is broken into, the organization’s main
servers remain functional.

Tools called anti-sniffers are available to defend against sniffers. When a sniffer program is active on a
computer, the computer’s network interface card (NIC) is placed in a state called promiscuous mode. The
anti-sniffer scans networks to determine if any network interface cards are running in promiscuous mode.
Anti-sniffers can be run regularly to detect evidence of a sniffer on the network. A switched network is
also a deterrent, because it eliminates the broadcasting of traffic to every machine, although hackers do
have programs they can use to get around a switched network.

The best defense against phishing and other scams is in the hands of the recipient. Recipients need to know
not to respond to any email that requests personal or financial information and not to click on any link given
in such an email that could take them to a spoofed website. Similarly, recipients of unexpected email
attachments need to know not to open them, even if a virus scan has not identified any virus in the attach-
ment. New viruses appear every day and one could slip past an antivirus program, even an antivirus
program that is updated regularly. Recipients of popup “notifications” about malware on their computers
should know that there is no legitimate scenario in which a computer will tell them it is infected and ask
them to call a toll-free number. Thus, employee education is a vital part of Internet security.

Encryption
Encryption is the best protection against traffic interception resulting in data leaks. Encryption converts
data into a code and then a key is required to convert the code back to data. Unauthorized people can
receive the coded information, but without the proper key, they cannot read it. Thus, an attacker may be
able to see where the information came from and where it went but not the content.

Encryption can be accomplished by two methods: secret key and public key/private key.

• In a secret key encryption system, each sender and recipient pair has a single key that is used
to encrypt and decrypt the messages. The disadvantage to a secret key system is that every pair
of senders and receivers must have a separate set of keys that match. If several pairs all used the
same set, then anyone having the key could decrypt anyone else’s message and it would not be a
secret. Use of a secret key system is impractical over the Internet, because any one company

236 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Internet Security

could have thousands of potential customers as well as others from whom it would need to receive
messages.

• The public key/private key encryption system is a better system for companies to use. In a
public-key/private-key encryption system, each entity that needs to receive encrypted data pub-
lishes a public key for encrypting data while keeping a private key to itself as the only means for
decrypting that data. Anyone can encrypt and send data to the company using its published public
key, but only the company’s private key can be used to decrypt the data, and only the company
that published the public key has the private key.

A company obtains a public key and the private key to go with it by applying to a Certificate
Authority, which validates the company’s identity and then issues a certificate and unique public
and private keys. The certificate is used to identify a company, an employee or a server within a
company. The certificate includes the name of the entity it identifies, an expiration date, the name
of the Certificate Authority that issued the certificate, a serial number and other identification. The
certificate always includes the digital signature of the issuing Certificate Authority, which permits
the certificate to function as a “letter of introduction” from the Certificate Authority. One example
of public/private encryption keys is SSL (Secure Sockets Layer), used on secure web sites.

Public key/private key encryption can be illustrated with an analogy. Imagine a door with two
locks, one keyed to the public key, and the other keyed to the private key. Both locks are unlocked,
and a pile of keys that fit the public key lock is available on a table nearby. When someone wants
to leave something in the room, they take a public key and lock the door. Upon locking the door
with the public key, the private key lock is locked automatically. Only the person with the private
key can open the door and look at what is in the room. Therefore, the person who left the message
in the room knows the message is safe from being read by anyone else and that only the intended
recipient can read it.

The most important point to remember here is that only someone with the private key (which
should be closely guarded) can open the door and have access to the contents of the room. While
it is possible to "pick the lock," doing so requires a large amount of time, skill and determination,
just as in real life.

Businesses can use public key cryptography when sending information. If Ronnie Retailer needs to
order from Smith Supply, Ronnie can use Smith Supply’s public key to encrypt the information it
wants to send via the Internet. Smith Supply can then use its private key to decrypt the message
and process whatever transaction was requested. Anyone else who happens to intercept Ronnie
Retailer’s message will see only gibberish.

Among other uses for it, public key/private key encryption is used to transmit credit card infor-
mation securely over the Internet to online merchants.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 237
Business Continuity Planning CMA Part 1

Business Continuity Planning


(A hardware control within application controls)
Business continuity planning involves defining the risks facing a company in the event of a disaster, as-
sessing those risks, creating procedures to mitigate those risks, regularly testing those procedures to ensure
that they work as expected, and periodically reviewing the procedures to make sure that they are up to
date. In the context of an information technology system, it is essential that the company have plans for
the backup of data and the recovery of data, especially in the context of disaster recovery.

Several different processes and backup plans function as part of the backup and recovery plan.

• Program files, as well as data files, should be backed up regularly.


• Copies of all transaction data are stored as a transaction log as they are entered into the system.
Should the master file be destroyed during processing, computer operations will roll back to the
most recent backup; recovery takes place by reprocessing the data transaction log against the
backup copy.
• Backups should be stored at a secure, remote location, so that in the event of data destruction
due to a physical disaster, the records can be reconstructed. It would do very little good to have
backup media in the same room as the computer if that area were destroyed by fire, for example.
• Backup data can be transmitted electronically to the backup site through a process called elec-
tronic vaulting, or backing up to the cloud. Backing up to the cloud carries its own control
considerations. Some considerations include
o Data security – A cloud backup provider should be certified compliant with international secu-
rity standards.
o Data privacy – Encryption alone is no guarantee of privacy. If encryption keys are held with
data in the cloud, the cloud provider may be required to turn over data if subpoenaed. For
maximum privacy, the encryption key can be further encrypted, called digital envelope en-
cryption.
o Dependability of the remote location – Management considering a cloud solution should un-
derstand the backup providers’ Service Level Agreements to determine whether they offer
automatic data redundancy across multiple data centers, so if something happens to one cen-
ter, other centers can take over. Without such data redundancy, data can be permanently lost.
o Legal data requirements of the country where the data is resident – Multinational organizations
with employees all over the world need to understand the data regulations in each country in
which they are located. Using a provider with a single location may result in a violation of local
data residency laws. If a cloud backup center has international locations, multinational entities
can choose which data centers to use for their data backups to ensure they are in compliance
with local data regulations.
• Grandparent-parent-child processing is used because of the risk of losing data before, during
or after processing work. Data files from previous periods are retained and if a file is damaged
during updating, the previous data files can be used to reconstruct a new current file. Like backup
media, these files should also be stored off-premises.
• Computers should be on Uninterruptible Power Supplies (UPS) to provide some protection in
the event of a power failure. Software is available that works in tandem with the UPS to perform
an orderly shutdown of the system during that short period of power maintenance that the UPS
can give the computer.
• Fault-Tolerant Systems are systems designed to tolerate faults or errors. They often utilize
redundancy49 in hardware design, so that if one system fails, another one will take over.

49
Redundancy is the duplication of critical components or functions of a system in an effort to increase the reliability of
the system. If one component fails, the duplicate component can take over the processing.

238 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E Business Continuity Planning

Disaster Recovery
Not many firms could survive for long without computing facilities. Therefore, an organization should have
a formal disaster recovery plan to fall back on in the event of a hurricane, fire, earthquake, flood, or criminal
or terrorist act. A disaster recovery plan specifies:

1) Which employees will participate in disaster recovery and what their responsibilities will be. One
person should be in charge of disaster recovery and another should be second in command.

2) What hardware, software, and facilities will be used.

3) The priority of applications that should be processed.

Arrangements for alternative facilities as a disaster recovery site and offsite storage of the company’s
databases are also part of the disaster recovery plan. An alternative facility might be a different facility
owned by the company, or it might be a facility contracted by a different company. The different locations
should be a significant distance away from the original processing site.

A disaster recovery site may be a hot site, a cold site, a warm site, or a mobile site.

A hot site, or a mirrored data center, is a backup facility that has a computer system similar to the one
used regularly. The hot site must be fully operational and immediately available, with all necessary tele-
communications hookups for online processing. A hot site also has current, live data being replicated to it
in real time from the live site by automated data communications.

A cold site is a facility where space, electric power, and heating and air conditioning are available and
processing equipment can be installed, though the equipment and the necessary telecommunications are
not immediately available. If an organization uses a cold site, its disaster recovery plan must include ar-
rangements to quickly get computer equipment installed and operational.

A warm site is in between a hot site and a cold site. It has the computer equipment and necessary data
and communications links installed, just as a hot site does. However, it does not have live data. If use of
the warm site is required because of a disaster, current data will need to be restored to it.

A mobile site is a disaster recovery site on wheels. It can be a hot site, a warm site, or a cold site. It is
usually housed in a trailer and contains the necessary electric power, heat and air conditioning. If it is a
warm or a hot site, it also contains the computer equipment; and if it is a hot site, it also contains current
data.

A mobile site may be company-owned or it may be available to the company on a contracted basis. Several
companies operate mobile recovery centers on a contract basis. In the event of a disaster that destroys
operations facilities, they arrive within hours in a tractor-trailer or van that is fully equipped with their
client’s platform requirements, 50 to 100 workstations, and staffed with technical personnel to assist in
recovery.

Personnel should be trained in emergency procedures and re-training should be done regularly to keep their
knowledge fresh. The disaster recovery plan should be tested periodically by simulating a disaster in order
to reveal any weaknesses in the plan. This test should be conducted using typical volumes, and processing
times should be recorded. The disaster recovery plan should be reviewed regularly and revised when nec-
essary because organizational and operational changes made that have not been incorporated into the
recovery plan could make the recovery plan unusable.

All members of the disaster recovery team should each keep a current copy of the plan at home.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 239
System Auditing CMA Part 1
efham CMA
Question 67: A critical aspect of a disaster recovery plan is to be able to regain operational capability
as soon as possible. To accomplish this, an organization can have a fully operational facility available
that is configured to the user's specific needs. This is best known as a(n):

a) Uninterruptible power system.

b) Parallel system.

c) Cold site.

d) Hot site.

(CMA Adapted)

System Auditing
Assessing Controls by Means of Flowcharts
A flowchart is one of the methods used by an auditor to understand the flow of information through pro-
cesses and systems. A flowchart depicts the flow of work: how procedures are initiated, authorized, and
processed, and when and how information flows between and among multiple systems. Flowcharts use
geometric symbols to show the steps and arrows to show the sequence of the steps.

A flowchart assists in properly identifying risks at each point in the process or system, identifying controls
necessary to address the risks, and assessing the effectiveness of existing controls. It helps to identify the
preventive controls that have been implemented, the detective controls that are in place, and the cor-
rective controls that are being practiced. It also helps to identify gaps in existing controls and risk areas
that may have not been previously identified.

The main elements shown in a flowchart are:

• Data sources (where the information comes from).

• Data destinations (where the information goes).

• Data flows (how the data gets there).

• Transformation process (what happens to the data).

• Data storage (how the data are stored for the long term).

A system flowchart shows the different departments or functions involved in a process horizontally across
the top. It documents the manual processes as well as the computer processes and the input, output and
processing steps. A systems flowchart also clearly shows the segregation of duties in the system.

The flowchart identifies specific control points in the system. A control point is a point in a process where
an error or irregularity is likely to occur, creating a need for control.

A flowchart can be used to determine whether the system ensures that authorizations, verifications, rec-
onciliations, and physical control activities are properly designed, documented, and operating effectively.
For example, in the payroll cycle a flowchart can help to illustrate whether access to employees’ personal
information in the payroll records is properly secured at every point in the process.

240 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E System Auditing

The auditor’s understanding of the controls is accomplished through:

• Inquiries made of employees.

• Analytical procedures.

• Observation of processes, or “walk-throughs.”

• Inspection of documents and documentation.

Finally, the auditor will use a system to assign a value to the probability that the controls will or will not
prevent or detect and correct an error or a fraudulent activity. Professional judgment is necessary at this
step to determine whether the overall assessment represents a pass or a fail of the IT control system.

Computerized Audit Techniques


Internal auditors use the computer to evaluate the processing being done by the computer and the controls
in place. Auditors can use a variety of tools to audit information systems, as follows.

Generalized Audit Software


Generalized audit software (GAS) permits the computer to be used by auditors as an auditing tool. The
computer can select, extract, and process sample data from computer files. Generalized audit software can
be used on mainframe computers and also on PCs. Generalized audit software can check computations,
search files for unusual items, and perform statistical selection of sample data. It can also prepare confir-
mation requests.

Test Data
To use test data, an auditor prepares input containing both valid and invalid data. The input is processed
manually to determine what the output should look like. The auditor then processes the test data elec-
tronically and compares the manually-processed results with the electronically-processed results. Test data
are used not only by auditors but also by programmers to verify the processing accuracy of the programs
they write and the programming changes they make.

Data used as test data might be real or fictitious. Since test data should not actually be processed, it is
important to ensure that the test transactions do not actually update any of the real data files maintained
by the system.

Test data can evaluate only the operation of programs. Other tests that verify the integrity of input and
output are required as well. Furthermore, the test data usually cannot represent all possible conditions that
a computer program might encounter in use. Another limitation is that test data can be run only on a
specific program at a specific time; because the test data must be processed separately from other data,
the auditor cannot be sure that the program being tested is the same program actually being used for
processing.

Integrated Test Facility


An Integrated Test Facility (ITF) involves the use of test data and also creation of test entities that do not
really exist, such as vendors, employees, products, or customers. The fictitious entities are actually included
in the system’s master files, and the test data are processed concurrently with real transactions. The trans-
actions are processed against live master files that contain the real records as well as the fictitious records.

The major difference between test data and an ITF is that the test data used in an ITF are pro-
cessed along with real data, whereas test data are not actually processed. No one knows that the
data being processed in the ITF includes the fictitious entries to fictitious records. Therefore, the auditor

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 241
System Auditing CMA Part 1

can be sure that the programs being checked are the same programs as those being used to process the
real data.

The difficulty with using the ITF approach is that the fictitious transactions must be excluded from the
normal outputs of the system in some way. They may be excluded manually, or they may be excluded by
designing or modifying the application programs. Either way, the fictitious transactions must be identified
by means of special codes so they can be segregated from the real data. Careful planning is required to
make sure that the ITF data do not become mixed in with the real data, corrupting the real data.

If careful planning is done, the costs of using an ITF are minimal, because no special processing is required
and thus no interruption of normal computer activity takes place. Costs are involved in developing an ITF,
both while the application is being developed and as later modifications are made to it. However, once the
initial costs are past, the ongoing operating costs are low.

An ITF is normally used to audit large computer systems that use real-time processing.

Parallel Simulation
Parallel simulation is an audit technique that uses real data rather than simulated data but processes it
through test or audit programs. The output from the parallel simulation is compared with the output from
the real processing. Parallel simulation is expensive and time-consuming and is usually limited to sections
of an audit that are of major concern and are important enough that they require an audit of 100% of the
transactions.

Since parallel simulation is done using test programs, the parallel simulation can be done on a computer
other than the one used for the real processing. However, the auditor should make sure that the system
used for the real processing of the output that is used for comparison is the same one that is used all the
time.

Embedded Audit Routines


Embedded audit routines involve modifying a regular production program by building special auditing rou-
tines into it so that transaction data can be analyzed. Embedded audit data collection is one type of
embedded audit routine, and it uses specially programmed modules embedded as inline code within the
regular program code. The embedded routine selects and records data as it is processing the data for
normal production purposes, for later analysis and evaluation by an auditor.

Transactions are selected by the embedded audit routine according to auditor-determined parameters for
limits and reasonableness. Transactions that violate those parameters are written to a file as exceptions.
Alternatively, transactions might be selected randomly. If transactions are selected randomly, the objective
is to create a statistical sample of transactions for auditing.

The approach that selects transactions that violate established limits is called a system control audit
review file (SCARF). The approach that selects random transactions is called a sample audit review file
(SARF).

It is easier to develop embedded audit routines when a program is initially developed than to add them
later.

Extended Records and Snapshot


Many different processing steps may be combined in a program, and therefore the audit trail for a single
transaction may exist in several different files. Extended records refer to modifying a program to tag
specific transactions and save all their processing steps in an extended record, permitting an audit trail to
be reconstructed from one file for those transactions. Transactions might be selected randomly, or they
might be selected as exceptions to edit tests.

242 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section E System Auditing

The snapshot technique “takes a picture” of a transaction as it is processed. Program code is added to the
application to cause it to print out the contents of selected memory areas when the snapshot code is
executed. A snapshot is used commonly as a debugging technique. As an audit tool, snapshot code can be
used for transactions that exceed predetermined limits.

Extended records and snapshot are very similar. The difference is that snapshot generates a printed audit
trail, whereas extended records incorporates snapshot data in the extended record file rather than on hard
copy.

Tracing
Tracing provides a detailed audit trail of all the instructions executed by a program. A single trace can
produce thousands of output records, so the auditor must take care to limit the number of transactions
tagged for tracing. Tracing might be used to verify that internal controls in an application are being executed
as the program is processing data, either live data or test data.

A trace may also reveal sections of unexecuted program code, which can indicate incorrect or unauthorized
modifications made to the program.

Mapping
Mapping involves using special software to monitor the execution of a program. The software counts the
number of times each program statement50 in the program is executed. Mapping can help determine
whether program application control statements that appear in the source language listing of the program
are actually executed when the program runs and have not been bypassed. It can also locate “dead” pro-
gram codes and can flag codes that may be being used fraudulently.

Mapping can be used with a program running test data. The output of the mapping program can indicate
whether the program contains unused code. All unused codes are investigated, and the purpose of the
unused code is evaluated to determine whether it should stay in the program or be removed.

Mapping originally was a technique used for program design and testing, but auditors also use it to deter-
mine whether program statements are being executed.

50
A “program statement” is an instruction to the computer. Each programming language has its own acceptable format
for program statements. The program statements are translated by the compiler into object code, the machine language
that the processor can understand.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 243
Section F – Technology and Analytics CMA Part 1

Section F – Technology and Analytics


Introduction to Technology and Analytics
Section F constitutes 15% of the CMA Part 1 Exam. Section F includes Information Systems, Data Govern-
ance, Technology-enabled Finance Transformation, and Data Analytics.

• The focus of Information Systems is on accounting information systems, inputs to them, and the
use of their outputs.

• Data Governance involves the management of an organization’s data assets and data flows. The
objective of data governance is to enable reliable and consistent data so that management is able
to properly assess the organization’s performance and make decisions.

• Technology-enabled Finance Transformation covers issues of importance for management account-


ants such as robotic process automation, artificial intelligence, cloud computing, and blockchains.

• Data Analytics also covers current technological and analytical issues that management account-
ants need to be familiar with. Business intelligence, data mining in large datasets, the use of
statistics and regression analysis, and visualization of data by charting are covered.

F.1. – Information Systems


The Value Chain and the Accounting Information System
The goal of any organization is to provide value to its customers. A firm’s value chain is the set of business
processes it uses to add value to its products and services.

The more value a company creates and provides to its customers, the more its customers will be willing to
pay for its products or services, and the more likely they are to keep buying those products or services. A
business will be profitable if the value it creates for its customers is greater than its cost of producing the
products and services it offers. All of the activities in the value chain contain opportunities to increase the
value to the customer or to decrease costs without decreasing the value the customer receives.

The value chain was discussed earlier in this volume. To review, the value chain as envisioned by Michael
Porter looks like the following:
Support Activities

Infrastructure

Human Resources Management


Technological Development

Procurement Value Added


Minus Cost
= Margin
Operations

Marketing
Outbound

and Sales
Logistics

Logistics
Inbound

Service

Primary Activities

244 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems

Though various organizations may view their value chains differently depending on their business models,
the preceding graphic encompasses the business processes used by most organizations. The primary ac-
tivities include the processes the organization performs in order to create, market, and deliver products
and services to its customers and to support the customers with service after the sale. The support activities
make it possible for the organization to perform the primary activities.

An organization’s accounting information system (AIS) interacts with every process in the value chain. The
AIS adds value to the organization by providing accurate and timely information so that all of the value
chain activities can be performed efficiently and effectively. For example:

• Just-in-time manufacturing and raw materials inventory management is made possible by an ac-
counting information system that provides up-to-date information about inventories of raw
materials and their locations.

• Sales information can be used to optimize inventory levels at retail locations.

• An online retailer can use sales data to send emails to customers suggesting other items they
might be interested in based on items they have already purchased.

• Allowing customers to access accounting information such as inventory levels and their own sales
orders can reduce costs of interacting with customers and increase customer satisfaction.

• A variance report showing a large unfavorable variance in a cost indicates that investigation and
possibly corrective action by management is needed.

• An AIS can provide other information that improves management decision-making. For instance:

o It can store information about the results of previous decisions that can be used in making
future decisions.

o The information provided can assist management in choosing among alternative actions.

The Supply Chain and the Accounting Information System


Parts of a company’s value chain are also parts of its supply chain. A company’s supply chain describes
the flow of goods, services, and information from the suppliers of materials and services to the organization
all the way through to delivery of finished products to customers. In contrast to the value chain, the activ-
ities of a company’s supply chain also take in outside organizations.

Nearly every product that reaches an end-user represents the coordinated efforts of several organizations.
Suppliers provide components to manufacturers who in turn convert them into finished products that they
ship to distributors for shipping to retailers for purchase by the consumer. All of the organizations involved
in moving a product or service from suppliers to the end-user (the customer) are referred to collectively as
the supply chain.

A well-designed accounting information system can improve the efficiency and effectiveness of a company’s
supply chain, thus enhancing the company’s profitability.

Automated Accounting Information Systems (AIS)


Automated accounting information systems are computer-based systems that transform accounting data
into information using the fundamental elements of paper-based accounting systems, but with electronic
processing.

Elements of Automated Accounting Information Systems


Any financial accounting information system, automated or otherwise, captures transactions or business
events that affect an entity’s financial condition. The transactions are recorded in journals and ledgers.
Journals are used to record accounting transactions. A journal contains a chronological record of the events

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 245
F.1. – Information Systems CMA Part 1

that have impacted the business’s finances. Journal entries show all the information about a transaction,
including the transaction date, the accounts debited and credited, and a brief description. Journal entries
are then posted to the general ledger.

The general ledger contains a separate account for each type of transaction. The list of general ledger
accounts used by an organization is called its chart of accounts. In a paper-based accounting system,
only an account name is needed, but in an automated accounting information system, the accounts need
to have numbers so that input is done in a consistent manner.

An automated accounting information system stores information in files. Master files store permanent
information, such as general ledger account numbers and history or customer account numbers and his-
torical data for each customer. Transaction files are used to update master files, and they store detailed
information about business activities, such as detail about sales transactions or purchase of inventory. For
each general ledger account number, the general ledger master file stores the transactions that have ad-
justed that account’s balance along with some basic information such as date and source. The detail needed
to maintain an audit trail is stored in transaction files. The detail in the transaction files may be needed in
future years if it becomes necessary to adjust or restate prior period financial statements, so it is a good
idea to maintain the transaction files for a number of years. After the transaction details are no longer
needed, though, the transaction files can be deleted.

In an automated accounting information system, accounts in the general ledger chart of accounts are num-
bered using block codes. Block codes are sequential codes that have specific blocks of numbers reserved
for specific uses. For example, a general ledger account numbering system is used to organize the accounts
according to assets, liabilities, equity, incomes, and expenses. An entity can use any numbering scheme it
wants, but one method of organizing account numbers uses account numbers beginning with “1” for assets,
“2” for liabilities, “3” for equity, “4” for incomes, and “5” for expenses. Thus, if 4-digit account numbers are
being used, then all asset accounts would be in the 1000 block, all liability accounts in the 2000 block, and
so forth. The numbers following the number in the first position subdivide the types of accounts more finely.
For example, in the asset section, current asset accounts might begin with “11” while noncurrent asset
accounts begin with “12.” Within the current asset section, then, cash might be 1110, accounts receivable
might be 1120, and inventory might be 1130.

Special journals are used for specific kinds of transactions, and in a computerized system, the journals are
known as modules. For example, an order entry module may be used to record sales. When a credit sale
is recorded in the order entry module, the order entry module updates the customer’s account in the ac-
counts receivable module (to add the charge to the account); it updates the sales module (to increase the
sales revenue); and it updates the inventory module (to reduce the items on hand by the items sold). It
also updates the general ledger to record the increase to accounts receivable, the increase to sales revenue,
the decrease in the value of inventory by the cost of the items sold, and the equivalent increase in cost of
goods sold expense. If sales tax or value-added tax is invoiced, those items are recorded as well in the
general ledger. When a customer pays, the receipt is recorded in the cash receipts module, which updates
the customer’s account in the accounts receivable module at the same time. The journal transactions update
the general ledger to record the increase to the cash account and the decrease to the accounts receivable
account.

In an automated accounting system, transactions are recorded electronically. They may be input by em-
ployees, but they may just as easily be recorded automatically. For example, a transaction may be created
automatically in the order entry module when a customer places an order over the Internet.

When a transaction is created in a module, the input includes a transaction code that identifies, for
example, a transaction in the order entry module as a sale transaction. The code causes the data entered
with that transaction code to be recorded in the other modules and in the proper general ledger accounts.
The specific transaction code for a sale may be set as the default code in the transaction code input field
within the order entry module, although the code can be changed if instead, for example, a sales return is
being processed.

246 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems

Codes, both numeric and alphanumeric, are used elsewhere in an automated AIS, as well. For example,
sequence codes are used to identify customer sales invoices created in the order entry module and may
be used for new customer accounts in the accounts receivable master file. Sequence codes are simply
assigned consecutively by the accounting system. For example, when a new customer account is added to
the accounts receivable master file, the AIS may automatically assign the next consecutive unused account
number to the new account. The number has no other meaning.

In a responsibility accounting system, a code is used to identify the responsibility center that is the source
of the transaction, and that code is part of transactions input to the AIS, as well.

Example: General ledger expense account numbers used for advertising expenses include a code for
the department that initiated the cost. The expense account number identifies the type of advertising
medium, for example television advertising, followed by codes indicating the type of advertising expense
and the product advertised.

Thus, a production expense for a television commercial advertising a specific kitchen appliance such as
a blender is coded to the television advertising account for commercial production for blenders. As the
cost is recorded in the AIS, the cost is directed to the correct responsibility center code and expense
account, as follows:
120 = 5= 16 = 2= 7= 31 =
Advertising Production
Department Expense Advertising Television Costs Blender
120 5 16 2 7 31

Thus, the full expense account number charged is 5162731 in department 120. That account number in
that responsibility center accumulates only expenses for television advertising production costs for
blender advertising that have been committed to by the advertising department.

As a result, the different types of advertising expenses are clearly delineated in the general ledger ac-
cording to responsibility center, type of expense, advertising medium, type of cost, and product
advertised, enabling easier analysis of the data.

Output of an Automated Accounting Information System


The data collected by an AIS is reported on internal reports. The internal reports are used by accountants
to prepare adjusting entries, by management for analysis and decision-making, and by both accountants
and management to produce the external financial reports needed.

An AIS needs to be designed so that it will be able to produce the reports that users will need. In the
preceding example, for instance, before the expense codes were developed, management decided it needed
to know not only total advertising expense, but also how much was spent for advertising in each advertising
medium, within each medium how much was spent for advertising production and how much for media
charges, and within those subdivisions, how much was spent to advertise each product.

Reports from an AIS may be paper reports, screen reports, or reports in various other forms such as audio
reports. They could be regularly scheduled reports or they could be produced on demand. Good reports
should have the following characteristics:

1) The report should include a date or dates. For example, a current customer list should show the
date the report is “as of.” A balance sheet should also show the “as of” date of the report. An
income statement for a period of time such as the year to date should indicate the period covered
by the report, such as “January 1 through April 30, 20X9.”

2) The report should be consistent over time so managers can compare information from different
time periods, consistent across segments so management can compare segment performance, and
consistent with generally accepted accounting principles so the report can be understood and used.

3) The report should be in a convenient format and should contain useful information that is easy to
identify. Summary reports should contain financial totals and comparative reports should provide
related numbers such as actual versus budgeted amounts in adjacent columns.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 247
F.1. – Information Systems CMA Part 1

Accounting Information System Cycles


Transaction cycles are grouped business processes for which the transactions are interrelated. They include:

• Revenue to cash cycle.

• Purchasing and expenditures cycle.

• Production cycle.

• Human resources and payroll cycle.

• Financing cycle.

• Fixed asset cycle (property, plant, and equipment).

• General ledger and reporting systems.

Revenue to Cash Cycle


The revenue to cash cycle involves activities related to the sale of goods and services and the collection of
customers’ cash payments. The cycle begins with a customer order and ends with the collection of the cash
from the customer.

To include the collection of customers’ cash payments for sales, the company needs to maintain accurate
records of customers’ outstanding invoices. Thus, the accounts receivable subsidiary ledger is an important
function of the AIS. Customer records also include payment history, assigned credit limit, and credit rating.

The accounting information system is used for:

• Tracking sales of goods and services to customers.

• Recording the fulfilling of customer orders.

• Maintaining customer records.

• Billing for goods and services.

• Recording payments collected for goods and services provided.

• Forecasting sales and cash receipts using the outputs of the AIS.

The process of entering an order and sale into the AIS also updates the inventory module so that items
sold are deducted from on-hand inventory and their costs charged to cost of goods sold.

The AIS should include a means to invoice customers as products are shipped. The production of the invoice
and the production of the packing slip should occur simultaneously, and nothing should be shipped without
a packing slip.

The AIS needs to be able to produce analytical reports of sales orders, sales terms, and payment histories
for customers for use in predictive analytics. Sales orders can be used to predict future sales, and the sale
terms can be used to make cash flow forecasts.

248 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems

Activities in and inputs to the revenue to cash cycle include:

• Receipt of customer orders and creation of sales orders. The sales order serves as input to the AIS
and contains full customer information, payment method, item or items sold, prices, and terms.

• Inventory is checked to determine whether the item or items ordered are in stock. A process needs
to be in place to track and follow up on backordered items.

• An order confirmation may be emailed to the customer, particularly for orders received on the
Internet.

• If the order includes a request for credit, the customer’s credit is checked. A process needs to be
in place to screen for fraudulent orders, as well, and much of that can be automated in the AIS.

• If the order is to be paid by credit card, preliminary credit card authorization is obtained for the
amount due.

• Input is created to the AIS for the invoice and the packing slip. (If the order was received over the
Internet, the input to the AIS may be automatically created.)

• Notice is sent to the warehouse to print the packing slip, pick the order, and prepare it for shipping.
For a service, the service is scheduled.

• Shipping information is processed, the package is moved to the shipping dock to be picked up by
the carrier, or the service is provided. For a credit card sale, the credit card charge is released. For
a sale made on credit, the invoice is sent to the customer.

• The accounts receivable module, inventory module, and general ledger are updated and reports
are processed.

• A process needs to be in place to handle approval and processing of returns. If a return is approved,
the process needs to include making sure the returned item is physically received and, if it can be
resold, that it is physically delivered back to the warehouse and added to inventory on hand in the
inventory module at the correct cost. A credit is created to the customer’s account and an appro-
priate journal entry is made to the general ledger.

• When payment is received for a credit sale, a remittance advice (usually detached from the invoice
and sent back by the customer) should accompany the customer’s payment. The remittance advice
is used to input the information about the payment received to the correct customer account in
the accounts receivable module, including the amount received and the invoice or invoices the
payment is to be applied to. The AIS also updates the cash receipts journal and the general ledger
to record the increase to cash and the decrease to accounts receivable.

Inputs to the revenue cycle can be made by desktop computer, but they may also be voice inputs, input
with touch-tone telephones, or input by wireless capability using tablet computers. Sales made on the
Internet may automatically update the subledgers and the general ledger in the AIS. Outside sales people
may use laptops, mobile devices, portable bar code scanners, and other types of electronic input devices
to enter sales orders.

Outputs of the revenue to cash cycle include:

• Internal management reports such as sales reports and inventory reports.

• Customer statements summarizing outstanding sales invoices by customer, payments received,


and balance owed.

• Customer aging reports, showing the total accounts receivable outstanding balance broken down
according to ranges of time outstanding, such as amounts outstanding for 0 to 30 days, 31 to 60
days, and so forth.

• Collection reports, showing specific accounts that need follow-up for overdue balances.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 249
F.1. – Information Systems CMA Part 1

• Information from sales reports, receivables aging reports, and receipts reports is used as input to
a cash receipts forecast.

• Various outputs are used in preparing closing entries and producing external financial statements.
For example, the aging report and other information is used in estimating credit loss expense for
the period.

Purchasing and Expenditures Cycle


The purchasing and expenditures process involves obtaining items and services in a timely manner at the
lowest price consistent with the quality required, managing the inventory, and seeing that payment is made
for items purchased and services received. The responsibilities of the purchasing function include:

• Locating and approving reputable vendors who offer quality goods and services at reasonable
prices. Vendor shipping and billing policies, discount terms, and reliability are important concerns
in approving vendors.

• Maintaining relationships with suppliers and other partners in the company’s supply chain.

The purchasing and expenditures cycle begins with a request for goods or services and ends with payment
to the vendor for the goods or to the provider of the service.

The accounting information system is used for:

• Tracking purchases of goods or services.

• Tracking amounts owed and making timely and accurate vendor payments.

• Maintaining vendor records and a list of authorized vendors.

• Managing inventory to ensure that all goods purchased are received, properly recorded in the AIS,
and properly dispensed from inventory.

• Forecasting purchasing needs and cash outflows.

The inventory control function in the AIS interfaces with production departments, purchasing, vendors, and
the receiving department.

Activities and inputs to the purchasing and expenditures cycle include:

• An internal purchase requisition is prepared by the requesting manager or department. Requests


for raw materials may be automated in the AIS when inventories fall below pre-determined levels,
or employees may key in the requests.

• The request is transmitted electronically to the purchasing department.

• The purchasing department selects the vendor, prepares input to the AIS for a purchase order,
and transmits the purchase order to the vendor.

• The purchasing department also transmits the information from the purchase order to the com-
pany’s receiving department so it will have the record of the order when the order is delivered.
However, as a control, the information the receiving department has access to should not include
the quantities ordered of each item. Thus, the receiving clerk must actually count and record the
items received rather than simply assuming the quantities ordered were the quantities received.

• When items are received, the receiving department creates an electronic receiving report. The
receiving report may be integrated in the AIS with the purchase order input as long as the receiving
personnel cannot access the quantities ordered. In other words, the receiving clerk may create the
receiving report by accessing each item on the purchase order electronically and inputting the
quantity counted that was received, but the receiving clerk should not be able to see the quantity
ordered.

250 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems

• Recording the receipt of the items electronically in the AIS updates the inventory on hand in the
inventory module, increases inventory in the general ledger by the amount of the cost, and creates
a payable in the accounts payable module and in the general ledger.

Note: Items shipped FOB shipping point belong to the purchaser as soon as they are shipped.
Items shipped FOB shipping point that have been shipped but not yet received as of a financial
statement date should be accrued as payables and the items should be included in ending in-
ventory. The AIS should have a means of properly handling those transactions.

• The invoice received from the vendor is compared with the purchase order and the receiving report.
A packing slip and a bill of lading from the freight carrier may also be received and they should be
included in the review.

• A process should be in place to investigate any differences in the items, quantity, and prices be-
tween and among the purchase order, receiving report, invoice, packing slip, and bill of lading.

• The invoice information is input into the accounts payable module to complete the information that
was added to the accounts payable module by the receiving report.

• The AIS should include controls that limit the potential for duplicate invoices to be paid. For exam-
ple, if an invoice is input that matches one from the same vendor with the same invoice number
or amount that has already been paid, it should be flagged for investigation before payment.

• If the item “received” was a service, approval should be received from a manager above the level
of the requesting manager before payment is prepared and sent. The higher-level approval is a
control to limit the opportunity for a manager to create a fictitious company, give fictitious service
business to that company, and approve payment to be sent to that company (that is actually sent
to him- or herself).

• If everything on the purchase order was received and if the items, quantities, and prices on the
invoice match the items, quantities, and prices on the purchase order, payment is prepared in the
AIS and sent. The payment may be sent as a paper check printed by the AIS. However, payments
are increasingly being sent by electronic funds transfer (EFT) through the automated clearing
house (ACH) whereby the funds are deducted from the purchaser’s bank account and sent elec-
tronically to the vendor.

• The accounts payable module and the general ledger are updated and reports are processed.

• Petty cash or procurement cards may be used for smaller purchases.

Inputs to the purchasing and expenditures cycle can be made with desktop computers, bar code scanners,
radio or video signals, magnetic ink characters on checks, scanned images, or entered into a tablet com-
puter and transmitted wirelessly.

Outputs of the purchasing and expenditures cycle include:

• The check (if a paper check is used) or payment advice (for an ACH transfer) and the payment
register.

• A cash requirements forecast.

• Discrepancy reports that note differences in items, quantities, or amounts on the purchase order,
the receiving report, and the vendor’s invoice or duplicate invoice numbers or duplicate amounts
to a vendor.

As noted previously, the discrepancy report is needed to prevent authorization of payment to a vendor until
any differences between or among the items, quantities, or prices on the purchase order, the receiving
report, and the purchase invoice or any potential duplicate invoices have been investigated and resolved.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 251
F.1. – Information Systems CMA Part 1

Using unpaid vendor invoices, outstanding purchase orders, and reports of items received for which the
invoices have not yet been received, the AIS can predict future cash payments that will be needed and the
dates they will be needed.

Production Cycle
The production cycle involves conversion of raw materials into finished goods and the goal is to do that as
efficiently as possible. Computer-assisted design technology and robotics are often used.

The production process begins with a request for raw materials for the production process. It ends with the
completion of manufacturing and the transfer of finished goods inventory to warehouses.

The accounting information system is used for:

• Tracking purchases of raw materials.

• Monitoring and controlling manufacturing costs.

• Managing and controlling inventories.

• Controlling and coordinating the production process.

• Providing input for budgets.

• Collecting cost accounting data for operational managers to use in making decisions.

• Providing information for manufacturing variance reports, usually using job costing, process cost-
ing, or activity-based costing systems.

Activities of and inputs to the production cycle include:

• Production managers issue materials requisition forms when they need to acquire raw material
from the storeroom.

• Physical inventory is taken periodically and reconciled to inventory records. The number of items
physically counted is input to the raw materials inventory module for reconciliation, and the ac-
counting records are updated.

• If the level of raw materials inventory in the storeroom falls below a predetermined level, a pur-
chase requisition is issued to the purchasing department. The issuance of the purchase requisition
may be automated in the accounting information system so that it occurs automatically when the
inventory reaches the reorder point.

• A bill of materials for each product is used to show the components needed and the quantities
of each needed to manufacture a single unit of product.

• The master production schedule shows the quantities of goods needed to meet anticipated sales
and when the quantities need to be produced in order to fulfill sales projections and maintain
desired inventory levels.

• Labor time needs to be tracked for costing purposes. Job time cards may be used to capture the
distribution of labor to specific orders.

• Enterprise Resource Planning systems are used in most large- and medium-sized firms to collect,
store, manage, and interpret data across the organization. ERP systems can help a manufacturer
to track, monitor, and manage production planning, raw materials purchasing, and inventory man-
agement, and are also integrated with tracking of sales and customer service.

Data entry is accomplished with automated technology such as bar code scanners, RFID (radio frequency
identification systems), GPS locators, and other input technologies that can reduce input errors and support
fast and accurate data collection for production.

252 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems

RFID can be used to track components to products and the products themselves along the production
process and the supply chain. Tags containing electronically-stored information are attached to components
and products.

Example: An RFID tag attached to an automobile can track its progress through the assembly line, and
RFID tags attached to individual components of the automobile can be used to make sure everything is
assembled properly.

Outputs of the production cycle include:

• Materials price lists showing prices paid for raw materials, kept up to date by the purchasing de-
partment and used by cost accountants to determine actual and standard costs for production.

• Usage reports showing usage of raw materials by various production departments, used by cost
accountants and managers to detect waste by comparing actual raw material usage to standard
raw material usage for the actual production.

• Inventory reconciliations comparing physical inventories with book balances.

• Inventory status reports that enable purchasing and production managers to monitor inventory
levels.

• Production cost reports detailing actual costs for cost elements, production processes, or jobs.
These reports are used by cost accountants to calculate variances for materials, labor, and over-
head.

• Manufacturing status reports, providing managers with information about the status of specific
processes or jobs.

• Reports output from the production process are used in developing financial statements.

Human Resources and Payroll Cycle


The human resources management and payroll cycle involves hiring, training, paying, and terminating
employees.

The accounting information system is used for:

• Recording the hiring and training of employees.

• Processes associated with employee terminations.

• Maintaining employee earnings records.

• Complying with regulatory reporting requirements, including payroll tax withholdings.

• Reporting on payroll benefit deductions such as for pensions or medical insurance.

• Making timely and accurate payroll payments to employees, payments for benefits such as pen-
sions and medical insurance, and payroll tax payments to taxing authorities.

Inputs to the human resource management and payroll cycle include forms sent by the human
resources department to payroll processing:

• Personnel action forms documenting hiring of employees and changes in employee pay rates and
employee status.

• Time sheets or time cards that record hours worked.

• Payroll deduction authorization forms.

• Tax withholding forms.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 253
F.1. – Information Systems CMA Part 1

Outputs of the human resource management and payroll cycle include:

• Employee listings showing current employees and information about them such as home ad-
dresses.

• Payroll payment registers listing gross pay, deductions, and net pay for each employee, used to
make journal entries to the general ledger. The journal entries may be automated in the AIS.

• Preparing paychecks or, for direct deposits to employees’ bank accounts, electronic funds trans-
fers.

• Deduction reports containing company-wide or individual segment employee deduction infor-


mation.

• Payroll summaries used by managers to analyze payroll expenses.

• Tax reports, used for remitting payroll taxes to the taxing authorities, both amounts withheld from
employees’ pay and employer taxes.

• Reporting to employees the information on their income and taxes paid that they need for their
personal tax reporting.

The payroll process is outsourced by many companies.

Financing Cycle
The financing process is responsible for acquiring financial resources by borrowing cash or selling stock and
for investing financial resources. It involves managing cash effectively, minimizing the cost of capital,51
investing in a manner that balances risk and reward, and making cash flow projections in order to make
any adjustments necessary to have cash on hand when it is needed for operations, investing, and repaying
debt.

Minimizing the cost of capital requires determining how much capital should be in the form of debt and how
much in the form of equity. Financial planning models are often used by management to help them
determine the optimum strategies for acquiring and investing financial resources.

Managing cash effectively includes collecting cash as quickly as possible, and many firms use lockbox ar-
rangements to reduce the collection time for payments received.52 Managing cash effectively also means
paying invoices as they are due and taking advantage of discounts for prompt payment when the discounts
offered are favorable. Electronic funds transfer through the automated clearing house is often used to pay
accounts payable and also to pay employees by depositing the funds directly to the payees’ bank accounts.
Use of electronic funds transfer enables a business to closely control the timing of funds being deducted
from its operating and payroll accounts.

Projecting cash flows involves using a cash receipts forecast—an output of the revenue cycle—and cash
disbursements forecasts—outputs of the purchasing and expenditures and the human resources and payroll
cycles.

51
“Capital” as used in the context of the “cost of capital” is the term used for the long-term funding used by firms that
is supplied by its lenders or bondholders and its owners (its shareholders). A company’s capital consists of its long-term
debt and its equity. A company’s cost of capital is the return expected by investors on a portfolio consisting of all the
company’s outstanding long-term debt and its equity securities. The cost of capital is tested on the CMA Part 2 exam
and is covered in more depth in the study materials for that exam.
52
With a lockbox system, a company maintains special post office boxes, called lockboxes, in different locations. Invoices
sent to customers contain the address of the lockbox nearest to each customer as that customer’s remittance address,
so customers send their payments to the closest lockbox. The company then authorizes local banks with which it main-
tains deposit relationships to check these post office boxes as often as is reasonable, given the number of receipts
expected. Because the banks are making the collections, the funds that have been received are immediately deposited
into the company’s accounts without first having to be processed by the company’s accounting system, thereby speeding
up cash collection. Cash management and the use of lockboxes are tested on the CMA Part 2 exam and are covered in
more depth in the study materials for that exam.

254 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems

Although the finance department utilizes the accounting information system, finance responsibilities should
be segregated from accounting responsibilities. A common approach is to have a controller who manages
the accounting function and a treasurer or CFO who manages the finance function.

The accounting information system is used for:

• The AIS can provide information about how quickly customers are paying their bills and can show
trends in cash collections for use in managing the collection of cash.

• An AIS with EFT capability can be used to make payments by electronic funds transfer through the
automated clearing house, or a separate EFT application that interfaces with the AIS can be used.

• Estimates of interest and dividend payments and receipts are used to develop cash flow forecasts.

Activities of and inputs to the financing cycle mostly originate outside the organization, as follows:

• Paper checks and remittance advices returned by customers with their payments are used to apply
the payments to customers’ accounts.

• Deposit receipts issued by the bank are used to document bank deposits.

• Bank statements are used to reconcile the cash balance according to the company’s ledger with
the cash balance in the bank account.

• Economic and market data, interest rate data and forecasts, and financial institution data are used
in planning for financing.

Outputs of the financing cycle

The production of periodic financial statements draws on general ledger information about the financing
processes, as follows:

• Interest revenue and expense.

• Dividend revenue and dividends paid.

• Summaries of cash collections and disbursements.

• Balances in investment accounts and in debt and equity.

• Cash budget showing projected cash flows.

Reports about investments and borrowings produced by the AIS for the financing cycle include:

• Changes in investments for a period.

• Dividends paid.

• Interest earned.

• New debt and retired debt for the period, including payments of principal and interest made for
the period and information about lending institutions and interest rates.

• Significant ratios such as return on investment for the organization as a whole and for individual
segments of it can be calculated by a financial planning model to help management make decisions
regarding investing and borrowing.

Property, Plant, and Equipment (Fixed Asset) System


The fixed asset management system manages the purchase, valuation, maintenance, and disposal of the
firm’s fixed assets, also called property, plant, and equipment.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 255
F.1. – Information Systems CMA Part 1

The accounting information system is used for:

• Recording newly-purchased fixed assets in the fixed asset module and the general ledger.

• Maintaining depreciation schedules and recording depreciation in order to calculate the book values
of fixed assets.

• Tracking differences between depreciation as calculated for book purposes and depreciation as
calculated for tax purposes in order to maintain deferred tax records.

• Maintaining records of the physical locations of fixed assets, as some of them may be moved
frequently.

• Tracking repair costs and distinguishing between repair costs that are expensed and repair costs
that are capitalized.

• Recording impairment of fixed assets.

• Tracking disposal of fixed assets and calculating the amount of gain or loss on the sale.

Activities and inputs to the fixed asset management system include:

• A request for a fixed asset purchase. The individual making the request uses a purchase requisition
form, usually input electronically. The request usually requires approval by one or more higher-
level managers.

• The purchasing department usually gets involved in vendor selection and issues a purchase order,
similar to the way it is done for other purchases. Receiving reports and supplier invoices are han-
dled the way they are for other purchases.

• If the company builds the fixed asset rather than acquiring it, a work order is used that details the
costs of the construction.

• Repair and maintenance records need to be maintained for each fixed asset or for categories of
fixed assets. Activities should be recorded on a repair and maintenance form so either the asset
account can be updated or an expense account can be debited in the AIS, as appropriate, for the
cost.

• When an existing fixed asset is moved from one location to another, those responsible should
complete a fixed asset change form so that its location can be tracked.

• A fixed asset change form should also be used to record the sale, trade, or retirement of fixed
assets.

Outputs of the fixed asset management system include:

• A list of all fixed assets acquired during a particular period.

• A fixed asset register listing the assigned identification numbers of each fixed asset held and each
asset’s location as of the register date.

• A depreciation register showing depreciation expense and accumulated depreciation for each fixed
asset owned.

• Repair and maintenance reports showing the current period’s repair and maintenance expenses
and each fixed asset’s repair and maintenance history.

• A report on retired assets showing the disposition of fixed assets during the current period.

All of the reports are used in developing information for the external financial statements.

256 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems

The General Ledger and Reporting Systems


The general ledger contains accounts for all the assets, liabilities, equity, revenues, and expenses. The
individual modules (or journals) support the general ledger and provide the detail behind the activity rec-
orded in the general ledger.

In a responsibility accounting system, a journal such as accounts receivable can be subdivided according
to responsibility center using responsibility center codes. Subsidiary ledgers are maintained for each re-
sponsibility center containing only the accounts used by that responsibility center.

In an automated accounting information system, most day-to-day transactions are initially recorded in the
individual modules such as the accounts receivable or accounts payable module. Recording transactions in
a module updates that module and possibly other modules and automatically creates transactions to the
general ledger. Thus, the general ledger obtains data from the other cycles and processes it so that financial
reports may be prepared. To be in accordance with generally accepted accounting principles, however,
many valuation and adjusting entries are also required.

Financial Reporting Systems


The primary purpose of the financial reporting system is to produce external financial statements for the
company’s stakeholders and other external users such as analysts. The reports include the statement of
financial position (balance sheet), income statement, statement of cash flows, statement of comprehensive
income, and statement of changes in stockholders’ equity.

Various internal reports are used by accountants who are reviewing data and making adjusting entries in
order to produce the external financial statements. Two of them are:

• A trial balance is a columnar report that lists each account in the general ledger in the first
column, followed by its balance as of a certain date and time in either the second or the third
columns from the left. Debit balances are shown in the second column from the left and credit
balances are shown in the third column from the left. All the debits are totaled and all the credits
are totaled, and the total debits and total credits on the trial balance must balance. A trial balance
is used to check preliminary balances before adjusting entries are made. It is printed after the
adjustments are recorded to confirm that the adjusted balances are correct.

• A general ledger report is a report covering a specific period of time that shows either all the
individual general ledger accounts and the transactions that adjusted them during that period, or
it may be printed for only one or for only a few accounts. A general ledger report is used to analyze
transactions posted to a specific account.

Management Reporting Systems


The information in the external financial statements is not what managers need to know for their decision
making. Managers need the detailed internal statements produced by the AIS.

Cost accounting systems collect labor, material, and overhead costs that are used to determine the
inventory costs of manufactured goods. Their output is used to determine the value of the inventories
reported on the balance sheet, but cost accounting systems are also used to report variances from antici-
pated costs that production managers need in order to control production, as described in Section C,
Performance Management, in Volume 1 of this textbook.

Profitability reporting systems and responsibility reporting systems involve comparisons between
actual and planned amounts and variances. They are usually prepared according to responsibility center.
Responsibility reporting traces events to the responsibility of a particular responsibility center or a particular
manager. Each significant variance should be explained by the manager who has knowledge of what caused
the variance. A variance may be a favorable variance, and an explanation of how it occurred may be useful
information for other responsibility centers in the organization. If a variance is unfavorable, knowing how
it occurred can be the first step to taking corrective action.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 257
F.1. – Information Systems CMA Part 1

Databases
A database is a collection of related data files, combined in one location to eliminate redundancy, that can
be used by different application programs and accessed by multiple users.

Basic Data Structure


The most commonly-used type of database is a relational database, which is a group of related tables.
When the database is developed, specific data fields and records are defined. The data must be organized
into a logical structure so it can be accessed and used. Data is stored according to a data hierarchy, and
the data is structured in levels.

A data field is the first level in the data hierarchy. A field is information that describes one attribute of an
item, or entity, in the database such as a person or an object. In an employee file, for example, one data
field would be one employee’s last name. Another field would be the same employee’s first name. A field
may also be called an “attribute,” or a “column.”

A database record is the second level of data. A database record contains all the information about one
item, or entity, in the database. For example, a single database record would contain information about
one employee. Each item of information, such as the employee’s Employee ID number, last name, first
name, address, department code, pay rate, and date of hire, is in a separate field within the employee’s
record. The data fields contained in each record are part of the record structure. The number of fields in
each record and the size of each field is specified for each record.

A file, also called a table, is the third level of the data hierarchy. A table is a set of common records, such
as records for all employees.

A complete database is the highest level. Several related files or tables make up a database. For example,
in an accounting information system, the collection of tables will contain all the information needed for an
accounting application.

Database
File or Table
Record
Field

Example: Consider the example of worker at a company who, over time, will have received numerous
monthly paychecks. The relational database will contain at least two database files, or tables, for
employees. The first table, the “Employees” table, contains all the records of the individual employees’
IDs and their names. The second table, the “Paychecks” table, contains data on all the paychecks that
have been issued and, for each paycheck, the Employee ID of the employee to whom it was issued.

A database management system can be used to locate all of the paychecks issued for one particular
employee by using the employee ID attached to the person’s name in that employee’s record in the
Employees table and locating all of the individual paycheck records for that same employee ID in the
Paychecks table.

The Employee ID ties the information in the Paychecks table to the information in the Employees table.

258 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems

Database Keys
The primary key is a data field in a record that distinguishes one record from another in the table. Every
record in a database has a primary key, and each primary key is unique. The primary key is used to find a
specific record, such as the record for a specific employee. A primary key may consist of one data field or
more than one data field. For example, in an Employees table, each employee record contains an Employee
ID. The Employee ID is the primary key in the Employees table.

Every record will have a primary key, and some records will also have foreign keys. Foreign keys connect
the information in a record to one or more records in other tables.

Example: Using the example of employees, the first table, the “Employees” table, contains all the rec-
ords of the individual employees’ IDs and their names. The second table, the “Paychecks” table, contains
records of all the paychecks that have been issued and, for each paycheck, the Employee ID of the
employee to whom it was issued. In the Employees table, the Employee ID in each employee record
serves as the primary key. In the Paychecks table, the Employee ID in each paycheck record is a
foreign key.

Entity-Relationship Modeling
Database administrators use the Entity-Relationship Model to plan and analyze relational database files and
records. An entity-relationship diagram utilizes symbols to represent the relationships between and
among the different entities in the database. The three most important relationship types are one-to-one,
one-to-many, and many-to-many. These relationship types are known as database cardinalities and
show the nature of the relationship between the entities in the different files or tables within the data-
base.

Example: Using the example of employees again, the connection between each employee (one person)
and each employee’s paychecks (which are many) is an example of a one-to-many relationship.

Database Management System (DBMS)


A database management system is a software package that serves as an interface between users and the
database. A database management system manages a set of interrelated, centrally-coordinated data files
by standardizing the storage, manipulation, and retrieval of data. It is used to create the database, maintain
it, safeguard the data, and make the data available for applications and inquiries. Because of its standard-
ized format, a database can be accessed and updated by multiple applications.

Database management systems perform four primary functions:

1) Database development. Database administrators use database management systems to develop


databases and create database records.
2) Database maintenance. Database maintenance includes record deletion, alteration, and reor-
ganization.
3) Database interrogation. Users can retrieve data from a database using the database manage-
ment system and a query language in order to select subsets of records to extract information.
4) Application development. Application development involves developing queries, forms, reports,
and labels for a business application and allowing many different application programs to easily
access a single database.

Note: A database management system is not a database but rather a set of separate computer programs
that enables the database administrator to create, modify, and utilize database information; it also en-
ables applications and users to query the database.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 259
F.1. – Information Systems CMA Part 1

A database management system provides languages for database development, database maintenance,
and database interrogation, and it provides programming languages to be used for application development.
The languages use statements, which is another word for commands. For example, a database adminis-
trator uses statements to create a database. A database administrator uses the DBMS not only to create a
database, but also sometimes to create applications that will access the data in the database.

Database Development
When a relational database is developed, the data fields and records to be used in the database must be
structured. The database administrator uses a database management system and a Data Definition
Language (DDL) to create a description of the logical and physical structure or organization of the data-
base and to structure the database by specifying and defining data fields, records, and files or tables. The
database administrator also specifies how data is recorded, how fields relate to each other, and how data
is viewed or reported.

The structure of the database includes the database’s schema, subschemas, and record structures.

• The schema is a map or plan of the entire database—its logical structure. It specifies the names
of the data elements contained in the database and their relationships to the other data elements.

• A particular application or user may be limited to accessing only a subset of the information in the
database. The limited access for an application or a user is called a subschema or a view. One
common use of views is to provide read-only access to data that anyone can query, but only some
users can update. Subschemas are important in the design of a database because they determine
what data each user has access to while protecting sensitive data from unauthorized access.

Note: The schema describes the design of the database, while the subschemas describe the
uses of the database. The database schema should be flexible enough to permit creation of all
of the subschemas required by the users.

• In defining the record structure for each table, the database administrator gives each field a
name and a description, determines how many characters the field will have, and what type of
data each field will contain (for example, text, integer, decimal, date), and may specify other
requirements such as how much disk space is needed.

The database administrator also defines the format of the input (for example, a U.S. telephone number
will be formatted as [XXX] XXX-XXXX).

The input mask for a data field creates the appearance of the input screen a user will use to enter data
into the table so that the user will see a blank field or fields in the style of the format. For example, a date
field will appear as __ __ /__ __ / __ __ __ __. The input mask helps ensure input accuracy.

Once the record structure of the database table is in place, the records can be created.

Database Maintenance
A data manipulation language (DML) is used to maintain a database and consists of “insert,” “delete,”
and “update” statements (commands). Databases are usually updated by means of transaction processing
programs that utilize the data manipulation language. As a result, users do not need to know the specific
format of the data manipulation commands.

Database Interrogation
Users can retrieve data from a database by using a query language. Structured Query Language (SQL)
is a query language, and it is also a data definition language and a data manipulation language. SQL has
been adopted as a standard language by the American National Standards Institute (ANSI). All relational
databases in use today allow the user to query the database directly using SQL commands. SQL uses the

260 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems

“select” command to query a database. However, business application programs usually provide a graph-
ical user interface (GUI) that creates the SQL commands to query the database for the user, so users
do not need to know the specific format of SQL commands.

Application Development
Database management systems usually include one or more programming languages that can be used
to develop custom applications by writing programs that contain statements calling on the DBMS to perform
the necessary data handling functions. When writing a program that uses a database that is accessed with
a DBMS, the programmer needs only the name of the data item, and the DBMS locates the data item in
the storage media.

Note: One of the key characteristics of a database management system is that the applications that
access the database are programmed to be independent of the data itself, meaning the programs
do not refer to a specific number or item, but rather to the name of the data item.

This independence is similar to changing a number in a spreadsheet cell that is referenced in a formula
in another cell elsewhere in the spreadsheet. It is not necessary to change the formula because the
formula relates to the cell and not to the number itself. Whatever number appears in that cell will be
used in the formula in the other cell.

Enterprise Resource Planning Systems


An accounting information system utilizes a database specific to the accounting, finance, and budgeting
functions. Information systems are used throughout organizations for much more than financial applica-
tions, though, and they all require their own databases. For example, materials requirements planning
systems are used to determine what raw materials to order for production, when to order them, and how
much to order. Personnel resource systems are used to determine personnel needs and to track other
personnel data.

However, problems arise when an organization’s various systems do not “talk” to one another. For instance:

• Production is budgeted based on expected sales, and raw materials need to be ordered to support
the budgeted production. If the materials requirements planning system cannot access the finan-
cial and budgeting system, someone needs to manually input the planned production of every
product into the materials requirements planning system after the sales have been budgeted and
the budgeted production has been set.

• A salesperson takes an order. If the items ordered are in stock, the salesperson submits the order
to an order entry clerk, who prepares the invoice and shipping documents. The documents are
delivered manually to the shipping department, and the shipping department prepares the ship-
ment and ships the order. After shipping, the sale is recorded in the accounting information system
and the customer’s account is updated with the receivable due. The order information is entered
separately into the database used by the customer relations management system, so that if the
customer calls about the order, the customer service personnel will be able to locate the order
information, because the customer service agents do not have access to the accounting records.

Inputting what is basically the same information multiple times is duplication of effort. Not only does it
waste time, but the multiple input tasks cause delays in making the information available to those who
need it. Furthermore, each time information is input manually into a separate system, the opportunity for
input errors increases. Thus, when the information finally is available, it may not even be accurate.

Enterprise Resource Planning (ERP) can help to overcome the challenges of separate systems because it
integrates all aspects of an organization’s activities—operational as well as financial—into a single system
that utilizes a single database.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 261
F.1. – Information Systems CMA Part 1

ERP systems consist of the following components:

• Production planning, including determining what raw materials to order for production, when to
order them, and how much to order.

• Logistics, both inbound (materials management) and outbound (distribution).

• Accounting and finance.

• Human resources.

• Sales, distribution, and order management.

Features of ERP systems include:

1) Integration. The ERP software integrates the accounting, customer relations management, busi-
ness services, human resources, and supply chain management so that the data needed by all
areas of the organization will be available for planning, manufacturing, order fulfillment, and other
uses. The system tracks all of a firm’s resources, including cash, raw materials, inventory, fixed
assets, and human resources, forecasts their requirements, and tracks shipping, invoicing, and the
status of commitments such as orders, purchase orders, and payroll.

2) Centralized database. The data from the separate areas of the organization flows into a secure
and centralized database rather than several separate databases in different locations. All users
use the same data that has been derived through common processes.

3) Usually require business process reengineering. An ERP system usually forces organizations
to reengineer or redesign their business processes in order to use the system. Because ERP soft-
ware is “off-the-shelf” software, customization is usually either impossible or prohibitively
expensive. Thus, business processes used must accommodate the needs of the system and many
may need to be redesigned.

When budgeted production is set, the information will immediately be available to determine what raw
materials should be ordered, how much, and when. When a salesperson enters a customer’s order into the
system (or when it is automatically entered as a result of an online order), inventory availability can be
immediately checked. If the order is a credit order, the customer’s credit limit and payment history is
checked. If everything is OK, the warehouse is notified to ship, the accounting information system is auto-
matically updated, and if the order is a credit card order, the customer’s credit card is charged. Information
on the order and its status is immediately visible to customer service personnel so they can answer ques-
tions about the order if they receive an inquiry from the customer.

Information about exceptions is also immediately available and can be addressed automatically. For exam-
ple:

• If the customer’s credit card charge is declined, shipment of the order can be automatically held
until an investigation can be performed and the order possibly cancelled.

• If the item ordered is not in stock, a backorder report is automatically generated for follow-up with
the customer, and the ERP system can trigger the production system to manufacture more product.
The production system can revise the production schedules accordingly, and human resources may
be involved if additional employees will be required.

Extended ERP Systems


Extended enterprise resource planning systems include customers, suppliers, and other business part-
ners. The systems interface with customers and suppliers through supply chain management
applications that give partners along the supply chain access to internal information of their suppliers and
customers. Suppliers can access the company’s internal information such as inventory levels and sales
orders, enabling the company to reduce its cycle time for procuring raw materials for manufacturing or
goods for sale. Customers can also view their supplier’s information about their pending orders.

262 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems

Advantages of ERP Systems

• Integrated back-office systems result in better customer service and production and distribution
efficiencies.

• Centralizing computing resources and IT staff reduces IT costs versus every department maintain-
ing its own systems and IT staff.

• Centralization of data provides a secure location for all data that has been derived through common
processes, and all users are using the same data.

• Day-to-day operations are facilitated. All employees can easily gain access to real-time information
they need to do their jobs. Cross-functional information is quickly available to managers regarding
business processes and performance, significantly improving their ability to make business deci-
sions and control the factors of production. As a result, the business is able to adapt more easily to
change and quickly take advantage of new business opportunities.

• Business processes can be monitored in new and different ways, such as with dashboards.

• Communication and coordination are improved across departments, leading to greater efficiencies
in production, planning, and decision-making that can lead to lower production costs, lower mar-
keting expenses, and other efficiencies.

• Data duplication is reduced and labor required to create inputs and distribute and use system out-
puts is reduced. Potential errors caused by inputting the same data multiple times are reduced.

• Expenses can be better managed and controlled.

• Inventory management is facilitated. Detailed inventory records are available, simplifying inventory
transactions. Inventories can be managed more effectively to keep them at optimal levels.

• Trends can be more easily identified.

• The efficiency of financial reporting can be increased.

• Resource planning as a part of strategic planning is simplified. Senior management has access to
the information it needs in order to do strategic planning.

Disadvantages of ERP Systems

• Business re-engineering (developing business-wide integrated processes for the new ERP system)
is usually required to implement an ERP system and it is time-consuming and requires careful
planning.

• Converting data from existing systems into the new ERP system can be time-consuming and costly
and, if done incorrectly, can result in an ERP system that contains inaccurate information.

• Training employees to use the new system disrupts existing workflows and requires employees to
learn new processes.

• An unsuccessful ERP transition can result in system-wide failures that disrupt production, inventory
management, and sales, leading to huge financial losses. Customers who are inconvenienced by
the implementation may leave. Because the entire business relies on the new ERP system, it is
critical that it be completely functional and completely understood by all employees before it “goes
live.” No opportunities are available to “work out the bugs” or “learn the ropes” when the entire
business relies on the one system.

• Ongoing costs after implementation include hardware costs, system maintenance costs, and up-
grade costs.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 263
F.1. – Information Systems CMA Part 1

Data Warehouse, Data Mart, and Data Lake

Data Warehouse
A copy of all of the historical data for the entire organization can be stored in a single location known as a
data warehouse, or an enterprise data warehouse. A data warehouse is separate from an ERP system
because a data warehouse is not used for everyday transaction processing. By having all of the company’s
information from different departments in one location for analysis, a company is able to more efficiently
manage and access the information for data mining to discover patterns in the data for use in making
decisions. For example, the marketing department can access production data and be better able to inform
customers about the future availability of products.

Managers can use business intelligence tools to extract information from the data warehouse. For in-
stance, a company can determine which of its customers are most profitable or can analyze buying trends.

Note: Business intelligence is the use of software and services to collect, store, and analyze data pro-
duced by a firm’s business activities. Business intelligence tools are used to access and analyze data
generated by a business and present easy-to-understand reports, summaries, dashboards, graphs, and
charts containing performance measures and trends that provide users with detailed information about
the business that can be used to make strategic and tactical management decisions. Business intelligence
is covered in more detail later in this section.

To be useful, data stored in a data warehouse should:

1) Be free of errors.

2) Be uniformly defined.

3) Cover a longer time span than the company’s transactions systems to enable historical research.

4) Allow users to write queries that can draw information from several different areas of the database.

Note: The data in a data warehouse is a copy of historical data, and therefore is not complete with the
latest real-time data. Furthermore, information in a data warehouse is read-only, meaning users cannot
change the data in the warehouse.

Because the data stored in a data warehouse exists in different formats in the various sources from which
it is copied, all differences need to be resolved to make the data available in a unified format for analysis.
The process of making the data available in the data warehouse involves the following.

1) Periodically, data is uploaded from the various data sources, usually to a staging server before
going to the data warehouse. The data upload may occur daily, weekly, or with any other estab-
lished frequency.

2) The datasets from the various sources are transformed to be compatible with one another by
adjusting formats and resolving conflicts. The transformation that must take place before the data
can be loaded into a data warehouse is known as Schema-on-Write because the schema is ap-
plied before the data is loaded into the data warehouse.

3) The transformed data is loaded into the data warehouse to be used for research, analysis, and
other business intelligence functions.

264 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.1. – Information Systems
efham CMA
Data Mart
A data mart is a subsection of a data warehouse that provides users with analytical capabilities for a
restricted set of data. For example, a data mart can provide users in a department such as accounts re-
ceivable access to only the data that is relevant to them so that the accounts receivable staff do not need
to sift through unneeded data to find what they need.

A data mart can provide security for sensitive data because it isolates the data certain people are authorized
to use and prevents them from seeing data that needs to be kept confidential. Furthermore, because each
data mart is used only by one department, the demands on the data servers can be distributed; one de-
partment’s usage does not affect other departments’ workloads.

Data marts can be of three different types:

1) A dependent data mart draws on an existing data warehouse. It is constructed using a top-down
approach and withdraws a defined portion of the data in the data warehouse when it is needed for
analysis. Several dependent data marts, used by different areas of the organization, can draw on
a single data warehouse.

2) An independent data mart is created without the use of a data warehouse through a bottom-up
approach. The data for just one data mart for a single business function is uploaded, transformed,
and then loaded directly into the data mart. If an organization needs several data marts for differ-
ent areas of the organization, each one would need to be created and updated separately, which
is not optimal. Furthermore, the idea of independent data marts is antithetical to the motivation
for developing a data warehouse.

3) A hybrid data mart combines elements of dependent and independent data marts, drawing some
data from an existing data warehouse and some data from transactional systems. Only a hybrid
data mart allows analysis of data from a data warehouse with data from other sources.

Data Lake
Much of the data captured by businesses is unstructured, such as social media data, videos, emails, chat
logs, and images of invoices, checks, and other items. Such data cannot be stored in a data warehouse
because the types of data are so disparate and unpredictable that the data cannot be transformed to be
compatible with the data in a data warehouse. A data lake is used for unstructured data.

A data lake is a massive body of information fed by multiple sources for which the content has not been
processed. Unlike data warehouses and data marts, data lakes are not “user friendly.” Data lakes have
important capabilities for data mining and generating insights, but usually only a data scientist is able to
access it because of the analytical skills needed to make sense of the raw information. The information in
a data lake is transformed and a schema is applied by the data scientist when it is analyzed, called Schema-
on-Read.

Data lakes are discussed in more detail later in the topic of Data Mining.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 265
F.1. – Information Systems CMA Part 1

Enterprise Performance Management


Enterprise Performance Management (EPM), also known as Corporate Performance Management (CPM) or
Business Performance Management (BPM), is a method of monitoring and managing the performance of an
organization in reaching its performance goals. It is the process of linking strategies to plans and execution.

The organization’s strategic goals and objectives must be clearly communicated to managers and be incor-
porated into their budgets and plans. Then, periodically the performance of the organization is reviewed
with respect to its progress in attaining the strategic goals and objectives. Key Performance Indicators
(KPIs), Balanced Scorecards, and Strategy Maps are frequently used and are monitored and managed. If
the organization or a segment of it is not performing as planned, adjustments are made, either in the
strategy or in the operations.

Enterprise Performance Management software is available that integrates with an organization’s accounting
information system, ERP system, customer relations management system, data warehouse, and other sys-
tems. It is designed to gather data from multiple sources and consolidate it to support performance
management by automating the collection and management of the data needed to monitor the organiza-
tion’s performance in relation to its strategy. Users can create reports and monitor performance using the
data captured and generated in the other systems. Some examples of an EPM’s capabilities include:

• Reports comparing actual performance to goals.

• Reports on attainment of KPIs by department.

• Balanced scorecards, strategy maps, and other management tools.

• Creating and revising forecasts and performing modeling.

• Generating dashboards presenting current information customized to the needs of individual users.

EPM software can also automate budgeting and consolidations. Tasks that in the past may have required
days or weeks can now be completed very quickly.

Note: EPM software can be on premises or it can be deployed as Software as a Service (SaaS), otherwise
known as “the cloud.” Cloud computing is covered in this section in the topic Technology-enabled Finance
Transformation.

266 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.2. Data Governance

F.2. Data Governance


Definition of Data Governance
Corporate governance includes all of the means by which businesses are directed and controlled, includ-
ing the rules, regulations, processes, customs, policies, procedures, institutions, and laws that affect the
way the business is administered. Corporate governance spells out the rules and procedures to be followed
in making decisions for the corporation.

Data governance is similar, but it is specific to data and information technology. Data governance en-
compasses the practices, procedures, processes, methods, technologies, and activities that deal with the
overall management of the data assets and data flows within an organization. Data governance is a process
that helps the organization better manage and control its data assets. In a sense, data governance is quality
control for data. It enables reliable and consistent data, which in turn makes it possible for management to
properly assess the organization’s performance and make management decisions.

Data governance includes the management of the following.

• Data availability, or the process of making the data available to users and applications when it
is needed and where it is needed.

• Data usability, including its accessibility to users and applications, its quality, and its accuracy.

• Data integrity, or the completeness, consistency, reliability, and accuracy of data.

• Data security, meaning data protection, including prevention of unauthorized access and protec-
tion from corruption and other loss, including backup procedures.

• Data privacy, that is, determining who is authorized to access data and which items of data each
authorized person can access.

• Data integration, which involves combining data from different sources (which can be both in-
ternal and external) and providing users with a unified view of all the data.

• System availability, that is, maximizing the probability that the system will function as required
and when required.

• System maintenance, including modifications of the system done to correct a problem, to im-
prove the system’s performance, to update it, or to adapt it to changed requirements or a changed
environment.

• Compliance with regulations, such as laws regulating privacy protections.

• Determination of roles and responsibilities of managers and employees.

• Internal and external data flows within the organization.

IT Governance and Control Frameworks


IT governance and control frameworks have been developed to provide models, or sets of standardized
guidelines, for the management of IT resources and processes. Frameworks provide numerous benefits to
an organization.

• They identify specific roles and responsibilities that need to be met.


• They provide a benchmark for assessing risks and controls.
• Following a framework provides a higher likelihood of implementing effective governance
and controls.
• Frameworks break down objectives and activities into groups.
• Regulatory compliance may be easier to achieve by following effective governance and con-
trol frameworks.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 267
F.2. Data Governance CMA Part 1

Following is an overview of two of the most prominent IT governance frameworks currently in use: COSO’s
internal control framework and ISACA’s COBIT.

Internal Control – Integrated Framework by COSO, the Committee of Sponsoring Organiza-


tions
The Committee of Sponsoring Organizations (COSO)53, which consists of five professional organiza-
tions, created one of the first internal control frameworks in 1992 with the publication of Internal Control—
Integrated Framework and it introduced the concept of controls for IT systems. Internal Control—Integrated
Framework was updated in 2013. Internal Control—Integrated Framework is covered in this volume in
Section E, Internal Control, so it will be reviewed only briefly here.

Internal Control—Integrated Framework defines internal control as “a process, effected by54 an entity’s
board of directors, management, and other personnel, designed to provide reasonable assurance regarding
the achievement of objectives relating to operations, reporting, and compliance.” According to the Inte-
grated Framework, the internal control system should consist of the following five interrelated components.

1) The control environment: the standards, processes, and structures that provide the foundation
for carrying out internal control.

2) Risk assessment: the process of identifying, analyzing, and managing the risks that have the
potential to prevent the organization from achieving its objectives, relative to the organization’s
established risk tolerance.

3) Control activities: the actions established by policies and procedures that help ensure that man-
agement’s instructions intended to limit risks to the achievement of the organization’s objectives
are carried out.

4) Information and communication: obtaining, generating, using, and communicating relevant,


quality information necessary to support the functioning of internal control. Communication needs
to be both internal and external.

5) Monitoring: overseeing the entire internal control system to assess the operation of existing in-
ternal controls to ensure that the internal control system continues to operate effectively.

COBIT® by ISACA
COBIT® is an I & T (Information and Technology) framework for the governance and management of en-
terprise information and technology. “Enterprise information and technology” refers to all the technology
and information processing used by the whole enterprise to achieve its goals, no matter where the tech-
nology and information processing occurs in the enterprise. Thus, while enterprise I & T includes the
organization’s IT department, it is not limited to the IT department.

The COBIT® Framework was first introduced for information systems in 1996 by ISACA, an independent,
nonprofit, global association dedicated to the development, adoption, and use of globally-accepted
knowledge and practices for information systems. ISACA was previously known as the Information Systems
Audit and Control Association. However, ISACA now serves a wide constituency of other IT professionals,
including consultants, educators, security professionals, risk professionals, and chief information officers.
To reflect the fact that it now serves such a broad range of IT professionals, the organization is now known
simply by its acronym, ISACA.

53
The Committee of Sponsoring Organizations sponsored the Treadway Commission, the National Commission on Fraud-
ulent Financial Reporting, that was created in 1985 to identify the causal factors of fraudulent financial reporting and to
make recommendations to reduce its incidence. The sponsoring organizations included the AICPA (American Institute of
Certified Public Accountants), the AAA (American Accounting Association), the IIA (Institute of Internal Auditors), the
FEI (Financial Executives International), and the IMA (Institute of Management Accountants).
54
To “effect” something means to cause it to happen, put it into effect, or to accomplish it. So “effected by” means “put
into effect by” or “accomplished by.”

268 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.2. Data Governance

In line with the early identity of the organization, the early focus of COBIT® was on information systems
audit and control, and COBIT® was an acronym for Control OBjectives for Information and Related Tech-
nology. However, in the intervening years the focus of the Framework has changed to IT governance and
management in recognition of the needs of the wide range of IT professionals that ISACA serves. When
COBIT® 5 was introduced in 2012, ISACA dropped the Framework’s full name entirely, and like ISACA,
COBIT® is now known simply by its acronym.55

ISACA published an updated version of the Framework, COBIT® 2019, in November 2018. The information
that follows is from COBIT® 2019.

COBIT® 2019 draws a distinction between governance and management. According to the Framework,
governance and management are different disciplines that involve different activities, different organiza-
tional structures, and serve different purposes.

Governance is usually the responsibility of the board of directors under the leadership of the chair of the
board of directors. The purpose of governance is to ensure that:

• Stakeholder56 needs are considered and conditions and options are evaluated in order to determine
balanced, agreed-on enterprise objectives.

• Prioritization and decision-making are used to set direction.

• Performance and compliance are monitored in terms of the agreed-on direction and enterprise
objectives.

Management is usually the responsibility of the executive management under the chief executive officer’s
(CEO’s) leadership. The purpose of management is to plan, build, run, and monitor activities, in accordance
with the direction set by the body responsible for governance such as the board of directors, in order to
achieve the enterprise objectives.57

The guidance in the COBIT Framework is generic in nature so that users can customize it to focused guid-
ance for the enterprise.58

Components of a Governance System


Each enterprise needs to establish and sustain a governance system that includes the following compo-
nents, or factors, that contribute to the operations of the enterprise’s governance system over information
and technology (I & T).

• Processes. A process is a set of practices and activities needed to achieve a specific objective and
produce outputs that support achievement of IT-related goals.

• Organizational structures. Organizational structures are the primary decision-making entities


within the enterprise.

• Principles, policies, and frameworks. Principles, policies, and frameworks provide practical
guidance for the day-to-day management of the enterprise.

• Information. Information includes all the information produced and used by the enterprise.
COBIT® 2019 focuses on the information needed for effective governance of the enterprise.

55
ISACA also promulgates separate IS auditing and IS control standards, so IS audit and control have not been left
behind.
56
Stakeholders for enterprise governance of information and technology include members of the board of directors,
executive management, business managers, IT managers, assurance providers such as auditors, regulators, business
partners, and IT vendors. (COBIT® 2019 Framework: Introduction and Methodology, p. 15, © 2018 ISACA. All rights
reserved. Used by permission.) Stakeholders are discussed in more detail later in this topic.
57
COBIT® 2019 Framework: Introduction and Methodology, p. 13, © 2018 ISACA. All rights reserved. Used by permis-
sion.
58
Ibid., p. 15.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 269
F.2. Data Governance CMA Part 1

• Culture, ethics, and behavior. The culture of the enterprise and the ethics and behavior of both
the enterprise and the individuals in it are important factors in the success of governance and
management activities.

• People, skills, and competencies. People and their skills and competencies are important for
making good decisions, for corrective action, and for successful completion of activities.

• Services, infrastructure, and applications. These include the infrastructure, technology, and
applications used to provide the governance system for information and technology processing.59

Components of a Governance System


According to COBIT® 2019

Processes
Organizational Principles,
Structures Policies, and
Frameworks

Governance
System Culture, Ethics,
Information
and Behavior

Services,
People, Skills,
Infrastructure,
and
and
Competencies
Applications

59
Ibid., pp. 21-22.

270 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.2. Data Governance

Governance and Management Objectives


Certain governance and management objectives need to be achieved in order for information and technol-
ogy to contribute to enterprise goals. COBIT® 2019 organizes the governance and management objectives
into five domains: one domain for governance and four domains for management.

Governance objectives are organized in one domain: EDM ‒ Evaluate, Direct and Monitor. The gov-
ernance body—usually the board of directors—evaluates strategic options, directs senior management in
achieving the chosen strategies, and monitors the achievement of the strategies.

Management objectives are organized in the following four domains.

1) APO ‒ Align, Plan, and Organize includes the overall organization, strategy, and supporting
activities for information and technology.

2) BAI ‒ Build, Acquire, and Implement covers defining, acquiring, and implementing of I & T
solutions and integrating them into business processes.

3) DSS ‒ Deliver, Service, and Support addresses operations for delivery and support of infor-
mation and technology services. DSS includes security.

4) MEA ‒ Monitor, Evaluate, and Assess involves performance monitoring of I & T and its con-
formance with internal performance targets, internal control objectives, and external requirements
such as laws and regulations.60

Governance
Objectives Management Objectives

EDM APO BAI DSS MEA


Evaluate, Direct, Align, Plan, Build, Acquire, Deliver, Service, Monitor, Evaluate,
and Monitor and Organize and Implement and Support and Assess

Design Factors for a Governance System


Design factors are factors that can impact the governance system, influence its design, and position the
enterprise for success in its use of information and technology. Design factors include any combination of
the following:

1) Enterprise strategy. Enterprise strategy is developed through the strategic planning process as
described in Section B, Planning, Budgeting, and Forecasting, in Volume 1 of this textbook. For
example, an enterprise might focus on differentiation, cost leadership, a particular marketing niche
or segment, or a number of other specific strategies. The strategies suggested in the COBIT® 2019
Framework are growth/acquisition, innovation/differentiation, cost leadership, and client ser-
vice/stability.

2) Enterprise goals. Enterprise goals support the enterprise strategy, and their achievement is nec-
essary in order to realize the enterprise strategy. The COBIT® 2019 Framework defines these goals
in terms of the four perspectives used in the Balanced Scorecard—Financial, Customer, Internal
Process (called simply “Internal” in COBIT® 2019), and Learning and Growth (called simply
“Growth” in COBIT® 2019). The Balanced Scorecard is covered in Section C, Performance Man-
agement, in Volume 1 of this textbook.

3) Risk profile. The enterprise’s risk profile identifies the types of information and technology risks
the enterprise is exposed to and indicates which risks exceed the enterprise’s risk appetite.

60
Ibid., p. 20.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 271
F.2. Data Governance CMA Part 1

4) I & T-related issues. I & T-related issues are I & T-related risks that have materialized.

5) Threat landscape. The “threat landscape” is the general environment of threats under which the
enterprise operates. For example, a political situation may be causing a high-threat environment,
or the particular industry in which the enterprise operates may be experiencing challenges.

6) Compliance requirements. The compliance requirements the enterprise is subject to may be


low, normal, or high, depending on the extent to which the industry within which the enterprise
operates is regulated.

7) Role of IT (information technology). The role of IT within the enterprise may be any of the
following:

a. Support – IT is not critical for the running and continuity of the business processes and ser-
vices or for innovation of business processes and services.

b. Factory – Any failure in IT would have an immediate impact on the running and continuity of
business processes and services, but IT is not a driver for innovation in business processes
and services.

c. Turnaround – IT is a driver for innovating business processes and services, but the enterprise
does not have a critical dependency on IT for the current running and continuity of its business
processes and services.

d. Strategic – IT is critical for running business processes and services and also for innovation
in the enterprise’s business processes and services.

8) Sourcing model for IT. Choices of sourcing include outsourcing (using services of a third party
to provide IT services), the cloud, insourced (the enterprise employs its own IT staff and provides
for its services), and hybrid (using a combination of outsourcing, the cloud, and insourcing).

9) IT implementation methods. Methods of implementing IT can be any of the following:

a. Agile – Agile development methods61 are used for software development.

b. DevOps – DevOps methods62 are used for software building, deployment, and operations.

c. Traditional – A traditional approach is used for software development, and software develop-
ment is separated from operations.63

d. Hybrid, or bimodal, IT – A mix of traditional and modern methods of IT implementation are


used.

61
Agile software development refers to a set of frameworks and software development practices based on a docu-
ment called the “Manifesto for Agile Software Development,” which says: “We are uncovering better ways of developing
software by doing it and helping others do it. Through this work we have come to value:
Individuals and interactions over processes and tools
Working software over comprehensive documentation
Customer collaboration over contract negotiation
Responding to change over following a plan
“That is, while there is value in the items on the right, we value the items on the left more.” (© 2001, the Agile Manifesto
authors, https://www.agilealliance.org/agile101/the-agile-manifesto/, accessed April 9, 2019.)
62
DevOps is the practice of operations and development engineers participating together in the entire service lifecycle,
from design through the development process to production support. It is also characterized by operations staff making
use of many of the same techniques as developers for their systems work. (https://theagileadmin.com/what-is-devops/,
accessed April 9, 2019.)
63
The traditional method of developing systems and programs is covered in System and Program Development and
Change Controls in Section E of this volume, Systems Controls and Security Measures. It is also reviewed briefly as the
Systems Development Life Cycle (SDLC) in this Section as part of Section F.3., Technology-enabled Finance Transfor-
mation.

272 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.2. Data Governance

10) Technology strategy – Methods of adopting new technologies followed by the enterprise may be
any of the following:

a. First mover – The enterprise tries to gain first-mover advantage by generally adopting new
technologies as early as possible.

b. Follower - The enterprise usually waits to adopt new technologies until they have been proven
and have become mainstream technologies.

c. Slow adopter – The enterprise is usually exceptionally slow to adopt new technologies.

11) Enterprise size. An enterprise may be considered a large enterprise—one with more than 250
full-time equivalent employees—or a small and medium enterprise—one with 50 to 250 full-time
equivalent employees. (Micro-enterprises, that is, those with fewer than 50 employees, are not
addressed by COBIT® 2019.)64

Goals Cascade
Enterprise goals, one of the design factors for a governance system, are involved in transforming stake-
holder needs into actionable strategy for the enterprise.

Stakeholders for enterprise governance of information and technology (EGIT) and the target audience for
COBIT® include the following:

Internal stakeholders:

• Members of the board of directors, for whom COBIT® provides insight into how to obtain value
from the use of I & T and explains relevant board responsibilities.

• Executive management, for whom COBIT® provides guidance in organizing and monitoring per-
formance of I & T.

• Business managers, for whom COBIT® helps understanding in how to obtain the I & T solutions
that the enterprise requires and how best to use new technology and exploit it for new strategic
opportunities.

• IT managers, for whom COBIT® provides guidance in how best to structure and operate the IT
department and manage its performance.

• Assurance providers such as auditors, for whom COBIT® helps in managing assurance over
IT, managing dependency on external service providers, and ensuring an effective and efficient
system of internal controls.

• Risk management, for whom COBIT® helps with identification and management of IT-related
risk.

External stakeholders:

• Regulators, in relation to which COBIT® helps ensure an enterprise is compliant with applicable
rules and regulations and has an adequate governance system in place to manage compliance.

• Business partners, in relation to which COBIT® helps the enterprise to ensure that a business
partner’s operations are secure, reliable, and compliant with applicable rules and regulations.

• IT vendors, in relation to which COBIT® helps the enterprise to ensure that IT vendors’ operations
are secure, reliable, and compliant with applicable rules and regulations.65

64
COBIT® 2019 Framework: Introduction and Methodology, pp. 23-27, © 2018 ISACA. All rights reserved. Used by
permission.
65
Ibid., p. 15.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 273
F.2. Data Governance CMA Part 1

Management objectives are prioritized based on prioritization of enterprise goals, which in turn are priori-
tized based on stakeholder drivers and needs. Alignment goals emphasize the alignment of the IT efforts
with the goals of the enterprise.66

COBIT® 2019 Goals Cascade

Stakeholder
Drivers and
Needs

Cascade to
Enterprise
Goals

Cascade to
Alignment
Goals

Cascade to Governance
and
Management
Objectives

The Enterprise Goals and the Alignment Goals are organized according to the Balanced Scorecard perspec-
tives:

• Financial

• Customer

• Internal Process (called “Internal” in COBIT® 2019).

• Learning and Growth (called “Growth” in COBIT® 2019).

The Governance and Management Objectives are organized according to the five domains:

• EDM ‒ Evaluate, Direct and Monitor for governance objectives

• APO ‒ Align, Plan, and Organize for management objectives

• BAI ‒ Build, Acquire, and Implement for management objectives

• DSS ‒ Deliver, Service, and Support for management objectives

• MEA ‒ Monitor, Evaluate, and Assess for management objectives

66
Ibid., p. 28.

274 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.2. Data Governance

Performance Management in COBIT® 2019


Performance management includes the activities and methods used to express how well the governance
and management systems and the components of an enterprise work, and if they are not achieving the
required level, how they can be improved. Performance management utilizes the concepts capability lev-
els and maturity levels.67

Performance management is organized in COBIT® 2019 according to the components that make
up the enterprise’s governance system over information and technology. Although not all of the
components are specifically addressed in COBIT® 2019 as to performance management issues at this time,
to review, the components include:

• Processes

• Organizational structures

• Principles, policies and frameworks

• Information

• Culture, ethics, and behavior

• People, skills, and competencies

• Services, infrastructure, and applications.

Those components for which performance management issues have been addressed include the following.

Managing the Performance of the Processes Component of the I & T Governance System

Governance and management objectives consist of several processes, and a capability level is assigned
to all process activities. The capability level is an expression of how well the process is implemented and is
performing. A process reaches a certain capability level when all the activities of that level are performed
successfully. Capability levels range from 0 (zero) to 5, as follows:

Level General Characteristics

0 Lack of any basic capability; incomplete approach to address governance and manage-
ment purpose; may or may not be meeting the intent of any Process practices.

1 The process more or less achieves its purpose through the application of an incomplete
set of activities that can be characterized as initial or intuitive—not very organized.

2 The process achieves its purpose through the application of a basic, yet complete, set of
activities that can be characterized as performed.

3 The process achieves its purpose in a much more organized way using organizational
assets. Processes typically are well defined.

4 The process achieves its purpose, is well defined, and its performance is (quantitatively)
measured.

5 The process achieves its purpose, is well defined, its performance is measured to improve
performance, and continuous improvement is pursued.68

67
Ibid., p. 37.
68
COBIT® 2019 Framework: Governance and Management Objectives, pp. 19-20, © 2018 ISACA. All rights reserved.
Used by permission.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 275
F.2. Data Governance CMA Part 1

Performance Management of the Organizational Structures Component of the I & T Governance System
Organizational structures (the primary decision-making entities within the enterprise) can be less formally
assessed according to criteria such as the following:

• Successful performance of processes for which the organizational structure or role has accounta-
bility or responsibility.

• Successful application of good practices for organizational structures.69

Performance Management of the Information Component of the I & T Governance System


No generally accepted method exists for assessing Information items. However, according to COBIT® 2019,
Information items can be less formally assessed using the information reference model from the previous
version of COBIT®, COBIT® 5: Enabling Information.70

According to COBIT® 5, the goals of the information component are organized into three dimensions of
quality: Intrinsic, Contextual, and Security/Accessibility.

1) Intrinsic – The extent to which data values are in conformance with the actual or true values.

2) Contextual – The extent to which information is applicable to the task of the information user and
is presented in an intelligible and clear manner, recognizing that information quality depends on
the context of use.

3) Security/Accessibility – The extent to which information is available or obtainable.

Each dimension is subdivided into several quality criteria. For example, a criterion of the Intrinsic dimension
is “accuracy,” a criterion of the Contextual dimension is “relevancy,” and a criterion of the Security/Acces-
sibility dimension is “availability.”71

The performance of an information item is assessed by evaluating the extent to which the relevant quality
criteria are achieved.72

Performance Management of the Culture, Ethics, and Behavior Component of the I & T Governance
System
To manage the performance of the Culture, Ethics, and Behavior component, a set of desirable behaviors
for good governance and management of IT should be defined and levels of capability should be assigned
to each.73

Some examples of desirable behaviors from the COBIT® 2019 Framework: Governance and Management
Objectives publication are:

• The culture and underlying values should fit the overall business strategy of the enterprise.

• Find ways to speed up processes. Introduce a culture and behavior that support moving faster,
perhaps by having more frequent strategy leadership meetings or by automating activities. 74

• Leaders must create a culture of continuous improvement in IT solutions and services. 75

69
COBIT® 2019 Framework: Introduction and Methodology, p. 40, © 2018 ISACA. All rights reserved. Used by permis-
sion.
70
Ibid., p. 41.
71
COBIT® 5: Enabling Information, p. 31, © 2013 ISACA. All rights reserved. Used by permission.
72
COBIT® 2019 Framework: Introduction and Methodology, p. 42, © 2018 ISACA. All rights reserved. Used by permis-
sion.
73
Ibid., p. 43.
74
COBIT® 2019 Framework: Governance and Management Objectives, p. 71, © 2018 ISACA. All rights reserved. Used
by permission.
75
Ibid., p. 196.

276 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.2. Data Governance

Focus Areas
A focus area is a governance topic, domain, or issue such as small and medium enterprises, information
security, DevOps, and cloud computing. A focus area can be addressed by governance and management
objectives.76

Managing the Performance of Focus Areas


Focus areas such as information security or DevOps are associated with maturity levels. Using the capa-
bility levels associated with the Processes component as a basis, a given maturity level is achieved if all
the processes within the focus area have achieved that particular capability level. Like the capability levels
associated with the Processes component, maturity levels range from 0 (zero) to 5.

The maturity levels used in managing the performance of focus areas are as follows:

Level General Characteristics

0 Incomplete — Work may or may not be completed toward achieving the purpose of gov-
ernance and management objectives in the focus area.

1 Initial — Work is completed, but the full goal and intent of the focus area are not yet
achieved.

2 Managed — Planning and performance measurement take place, although not yet in a
standardized way.

3 Defined — Enterprise-wide standards provide guidance across the enterprise.

4 Quantitative — The enterprise is data driven, with quantitative performance improve-


ment.

5 Optimizing — The enterprise is focused on continuous improvement.77

76
COBIT® 2019 Framework: Introduction and Methodology, p. 22, © 2018 ISACA. All rights reserved. Used by permis-
sion.
77
Ibid., p. 40.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 277
F.2. Data Governance CMA Part 1

Data Life Cycle


The data life cycle encompasses the period from creation of data and its initial storage through the time
the data becomes out of date or no longer needed and is purged. The stages of the data life cycle include
data capture, data maintenance, data synthesis, data usage, data analytics, data publication, data archival,
and data purging.

The stages do not describe sequential data flows, because data may pass through various stages several
times during its life cycle. Furthermore, data does not have to pass through all of the stages. However,
data governance challenges exist in all of the stages and each stage has distinct governance needs, so it is
helpful to recognize the various stages and some of the governance challenges associated with each.

• Data capture is the process of creating new data values that have not existed before within the
organization. Data can be captured through external acquisition, data entry, or signal reception.

o External acquisition. Data can be acquired from an outside organization, often through con-
tracts governing how the data may be used. Monitoring performance with the contracts is a
significant governance challenge.

o Data entry. Data can be created through entry by human operators or by devices that gen-
erate data. Data governance requires monitoring the accuracy of the input.

o Signal reception. Data may be received by transmission, for example from sensors. Data
governance challenges include monitoring the function of the devices and the accuracy of the
data received.

• Data maintenance is the processing of data before deriving any value from it, such as performing
data integration. A governance issue is determining how best to supply the data to the stages at
which data synthesis and data usage occur.

• Data synthesis is the creation of new data values using other data as input. To “synthesize”
something means to combine different things to make something new. Data synthesis is therefore
combining data from different sources to create new data. Governance issues include concerns
about data ownership and the need for citation, the quality and adequacy of the input data used,
and the validity of the synthesized data.

• Data usage, which is the application of the data to tasks, whether used in support of the organi-
zation or used by others as part of a product or service the organization offers. A governance
challenge with respect to data usage is whether the data can legally be used in the ways the users
want to use it. Regulatory or contractual constraints on the use of the data may exist, and the
organization must ensure that all constraints are observed.

• Data analytics, or the process of gathering and analyzing data in a way that produces meaningful
information to aid in decision-making. As businesses become more technologically sophisticated,
their capacity to collect data increases. However, the stockpiling of data is meaningless without a
method of efficiently collecting, analyzing, and utilizing it for the benefit of the company. A gov-
ernance issue with respect to data analytics is ensuring that the company’s data is accurately
recorded, stored, evaluated, and reported.

• Data publication occurs when data is sent outside of the organization or leaves it in other ways.
Periodic account statements sent to customers are an example of data publication, but data
breaches also constitute data publication. The governance issue is that once data has been released
it cannot be recalled, because it is beyond the reach of the organization. If published data is in-
correct or if a data breach has occurred, data governance is needed in deciding how to deal with
it.

• Data archival is the removal of data from active environments and its storage in case it is needed
again. No maintenance, usage, or publication occurs while the data is archived, but if it is needed,
it can be restored to an environment where it will again be maintained, used, or published.

278 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.2. Data Governance

• Data purging occurs at the end of the data’s life cycle. Data purging is the removal of every copy
of a data item from all locations within the organization. Purging should generally be done only of
data that has been previously archived. Data governance considerations include development and
maintenance of data retention and destruction policies that comply with all laws and regulations
regarding record retention; conformance with established policies; confirmation that purging has
been done properly; and documentation of data purged.

Records Management
Every organization should have a documented records management policy that establishes how records are
to be maintained, identified, retrieved, preserved, and when and how they are to be destroyed. The policy
should apply to everything defined by the organization as a “record,” which includes both paper documents
and data records. The concern for information systems is, of course, data records.

Although not every item of data will be determined to be a “record,” consideration in the policy should also
be given to management of data that is not actually “records,” such as drafts of documents. The records
management policy should identify the information that is considered records and the information that is
not considered records but that nevertheless should be subject to the guidance in the policy.

Factors to be considered in developing a records management policy include:

• Federal, state, and local document retention requirements. U.S. federal requirements in-
clude Internal Revenue Service requirements for retaining income tax information and various
employment laws and laws governing employee benefits. State and local requirements may also
apply. Regulations and laws provide minimum records retention requirements, but those should
not be regarded as guidance on when records must be destroyed. Decisions may be made to retain
specific records longer than their required minimum periods due to other factors such as ongoing
business use or internal audit requirements.

• Requirements of the Sarbanes-Oxley Act of 2002. Section 802 of the Sarbanes-Oxley Act
prohibits altering, destroying, mutilating, concealing, or falsifying records, documents, or tangible
objects with the intent to obstruct, impede, or influence a potential or actual federal investigation
or bankruptcy proceeding. Violation is punishable by fines and/or imprisonment for up to 20 years.
Furthermore, accountants must maintain certain audit records or review work papers for a period
of five years from the end of the fiscal period during which the audit or review was concluded.
Section 1102 of the Act states that corruptly altering or destroying a record or other object with
the intent to impair its integrity or availability for use in an official proceeding is also punishable
with fines and/or imprisonment for up to 20 years.

• Statute of limitations information. A statute of limitations is the time period during which an
organization may sue or be sued or the time period within which a government agency can conduct
an examination.

• Accessibility. An important consideration with electronic records is hardware, software, and me-
dia obsolescence. Records can become inaccessible if they are in an obsolete format, and the
records management policy must include a means to either migrate the records to new versions,
or the old hardware and software must be retained so the records can be accessed. If the records
are to be migrated to new formats, quality control procedures must be in place to ensure none of
the content is lost or corrupted.

• Records of records. The records management policy should establish the practice of maintaining
an index of active and inactive records and their locations and of maintaining logs containing rec-
ords of all purged data.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 279
F.2. Data Governance CMA Part 1

Needless to say, it is not enough to simply have a documented records management policy. The policy
must be followed consistently. However, if a lawsuit is pending or anticipated, all document and data de-
struction should cease. All documents and data should be preserved, even if they would otherwise be purged
under the policy.

Electronic data management should be supported by top management, and responsibility for custody of the
records should be assigned.

Benefits of a Documented and Well-Executed Records Management Policy

• Locating documents that are needed is easier and more efficient.

• In the event of an examination, investigation, or lawsuit, having a documented records manage-


ment policy that is consistently followed demonstrates a legitimate and neutral purpose for
having previously destroyed any requested documents or data purged in accordance with the policy.

• Having a documented policy that is consistently followed increases the probability of the organ-
ization’s compliance with all federal, state, and local regulations relating to document and data
retention and destruction.

• Records will be adequately protected and their accessibility maintained.

• Records that are no longer needed or that are no longer of value will be destroyed at the appropriate
time.

Cyberattacks
Cybersecurity is the process or methods of protecting Internet-connected networks, devices, or data from
attacks. Cyberattacks are usually made to access, change, or destroy data, interrupt normal business op-
erations, or, as with ransomware, they may involve extortion.

Some specific cybersecurity risks include the following:


• Copyright infringement is the theft and replication of copyrighted material, whether intellectual
property, such as computer programs or textbooks, or entertainment property such as music and
movies.

• Denial of Service (DOS) attacks occur when a website or server is accessed so frequently that
legitimate users cannot connect to it. Distributed Denial of Service (DDOS) attacks use multiple
systems in multiple locations to attack one site or server, which makes stopping or blocking the
attack difficult. Hackers gain access to unsecured Internet of Things (IoT)78 devices on which
the default passwords have not been changed and use malware tools to create a botnet made up
of innumerable devices. A botnet is a network of devices connected through the Internet that are
all infected with the malware. A hacker or a group of hackers control the botnet without the owners’
knowledge. The hacker directs the botnet to continuously and simultaneously send junk Internet
traffic to a targeted server, making it unreachable by legitimate traffic. Sophisticated firewalls and
network monitoring software can help to mitigate DOS and DDOS attacks.

• Buffer overflow attacks are designed to send more data than expected to a computer system,
causing the system to crash, permitting the attacker to run malicious code, or even allowing for a
complete takeover of the system. Buffer overflow attacks can be easily prevented by the software
programs adequately checking the amount of data received, but this common preventative meas-
ure is often overlooked during software development.

78
Internet of Things (IoT) devices are products used in homes and businesses that can be controlled over the Internet
by the owner. Examples are door locks, appliances, lights, energy- and resource-saving devices and other devices that
can be controlled either remotely or on-premises using voice commands.

280 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.2. Data Governance

• Password attacks are attempts to break into a system by guessing passwords. Brute force
attacks use programs that repeatedly attempt to log in with common and/or random passwords,
although most modern systems effectively prevent brute force attacks by blocking login attempts
after several incorrect tries. Two-factor authentication can also prevent brute force attacks from
being successful because a password alone will not allow access to the system. Systems should
include sophisticated logging and intrusion-detection systems to prevent password attacks, and
password requirements should be in place to reject short or basic passwords such as “password”
or “123456.”

• Phishing is a high-tech scam that uses spam email to deceive people into disclosing sensitive
personal information such as credit card numbers, bank account information, Social Security num-
bers, or passwords. Sophisticated phishing scams can mock up emails to look like the information
request is coming from a trusted source, such as state or local government, a bank, or even a
coworker. The best defense against phishing is awareness and common sense. Recipients should
be wary about any email that requests personal or financial information and should resist the
impulse to click on an embedded link.

• Malware broadly refers to malicious software, including viruses. Spyware can secretly gather
data, such as recording keystrokes in order to harvest banking details, credit card information,
and passwords. Other types of malware can turn a PC into a bot or zombie, giving hackers full
control over the machine without alerting the owner to the problem. Hackers can then set up
“botnets,” which are networks consisting of thousands or millions of “zombies,” which can be made
to send out spam emails, emails infected with viruses, or as described above, to cause distributed
denial of service attacks.

• Ransomware is particularly dangerous malware that encrypts data on a system and then de-
mands a ransom (a payment) for decryption. If the ransom is not paid, the data is lost forever.
The most common way that ransomware is installed is through a malicious attachment or a down-
load that appears to come from a trusted source. The primary defenses against ransomware are
to avoid installing it in the first place and having data backups.

• “Pay-per-click” abuse refers to fraudulent clicks on paid online search ads (for example, on
Google or Bing) that drive up the target company’s advertising costs. Furthermore, if there is a set
limit on daily spending, the ads are pushed off the search engine site after the maximum-clicks
threshold is reached, resulting in lost business as well as inflated advertising costs. Such scams
are usually run by one company against a competitor.

Some cybercrime is conducted on a more personal and in-person fashion. Through social engineering an
individual may pose as a trustworthy coworker, perhaps someone from the company’s IT support division,
and politely ask for passwords or other confidential information. Dumpster diving is the act of sifting
through a company’s trash for information that can be used either to break into its computers directly or
to assist in social engineering.

Outsiders are not the only threat to the security of a company’s systems and data. Insiders can also be a
source of security risks. For example, disgruntled employees or those who are planning to work for a
competitor can sabotage computer systems or steal proprietary information.

Defenses Against Cyberattack


Encryption is an essential protection against hacking. Encryption protects both stored data and data that
could be intercepted during transmission. If a hacker gains access to encrypted files, the hacker is not able
to read the information.

Ethical hackers are network and computer experts with hacking skills who attempt to attack a secured
system. They use the same methods as are used by malicious hackers, but if they find vulnerabilities that
could be exploited by a malicious hacker, they report them to the owner or manager so they can be reme-
died. Ethical hacking is called intrusion testing, penetration testing, and vulnerability testing.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 281
F.2. Data Governance CMA Part 1

Advanced firewalls are firewalls that perform traditional firewall protection but have other capabilities,
as well. Traditional firewalls use packet filtering to control network access by monitoring outgoing and
incoming packets. Packets are used in networking to carry data during transmission from its source to its
destination. They have a header that contains information that is used to help them find their way and to
reassemble the data after transmission. Traditional firewalls permit or deny access based on the information
in the header about the source and the destination Internet Protocol (IP) addresses, protocols, and ports.

Advanced firewalls are called Next Generation Firewalls (NGFW). In addition to the traditional firewall
protection, advanced firewalls can filter packets based on applications and can distinguish between safe
applications and unwanted applications because they base their detection on packet contents rather than
on information in packet headers. Thus, they are able to block malware from entering a network, which is
something that traditional firewalls cannot do. However, both traditional and advanced firewalls rely on
hardware. For cloud-based (Software as a Service, or SaaS) applications, a cloud-based (Firewall as a
Service) application is needed.

Access Controls
Access controls provide additional defenses against cyberattack. Access controls include logical access
controls and physical access controls.

Logical Access Controls


Companies need to have strict controls over access to their proprietary data. Poor data oversight can leave
a company vulnerable to accidents, fraud, and malicious parties who manipulate equipment and assets.
Logical security focuses on who can use which computer equipment and who can access data.
Logical access controls identify authorized users and control the actions that they can perform.

To restrict data access only to authorized users, one or more of the following strategies can be adopted:

1) Something you know

2) Something you are

3) Something you have

Something You Know


User IDs and passwords are the most common “something you know” way of authenticating users. Security
software can be used to encrypt passwords, require changing passwords after a certain period of time, and
require passwords to conform to a certain structure (for example, minimal length, no dictionary words, re-
stricting the use of symbols). Procedures should be established for issuing, suspending, and closing user
accounts, and access rights should be reviewed periodically.

Something You Are


Biometrics is the most common form of “something you are” authentication. Biometrics can recognize
physical characteristics such as:

• Iris or retina of the eyes

• Fingerprints

• Vein patterns

• Faces

• Voices

Biometric scanners can be expensive and are generally used only when a high level of security is required.

282 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.2. Data Governance

Something You Have


Some very high-security systems require the presence of a physical object to certify an authorized user’s
identity. The most common example of this “something you have” authentication is a fob, a tiny electronic
device that generates a unique code to permit access; for increased security, the code changes at regular
intervals. A lost fob may be inconvenient but not a significant problem because the fob by itself is useless.
Furthermore, a stolen fob can be remotely deactivated.

Two-Factor Authentication
Two-factor authentication requires two independent, simultaneous actions before access to a system is
granted. The following are examples of two-factor authentication:

• In addition to a password, some systems require entering additional information known only to the
authorized user, such as a mother’s maiden name or the answer to another security question
chosen by the authorized person. However, this security feature can be undermined if the second-
ary information can be obtained easily by an unauthorized third party.

• Passwords can be linked to biometrics.

• In addition to a password, a verification code is emailed or sent via text message that must be
entered within a few minutes to complete the login.

• A biometric scan and a code from a fob are combined to allow access.

Considerations when evaluating the effectiveness of a logical data security system include:

• Does the system provide assurance that only authorized users have access to data?

• Is the level of access for each person appropriate to that person’s needs?

• Is there a complete audit trail whenever access rights and data are modified?

• Are unauthorized access attempts denied and reported?

Other User Access Considerations


Besides user authentication, other security controls related to user access and authentication to prevent
abuse or fraud include:

• Automatic locking or logoff policies. Any login that is inactive for a specified period of time can
automatically be logged out in order to limit the window of time available for someone to take
advantage of an unattended system.

• Logs of all login attempts, whether successful or not. Automatic logging of all login attempts
can detect activities designed to gain access to an account by repeatedly guessing passwords.
Accounts under attack can then be proactively locked in order to prevent unauthorized access.

• Accounts that automatically expire. If a user needs access to a system only for a short period
of time, the user’s access to the system should be set to automatically expire at the end of that
period, thus preventing open-ended access.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 283
F.2. Data Governance CMA Part 1

Physical Access Controls


Physical access controls are used to secure equipment and premises. The goal of physical access controls
is to reduce or eliminate the risk of harm to employees and of losing organizational assets. Controls should
be identified, selected, and implemented based on a thorough risk analysis. Some common examples of
general physical security controls include:

• Walls and fences

• Locked gates and doors

• Manned guard posts

• Monitored security cameras

• Guard dogs

• Alarm systems

• Smoke detectors and fire suppression systems

Physical access to servers and networking equipment should be limited to authorized persons. Keys are
the least expensive way to manage access but also the weakest way because keys can be copied. A more
effective method is card access, in which a magnetically encoded card is inserted into or placed near a
reader. The card access also provides an audit trail that records the date, time, and identity of the person
(or at least of the card of the person) who entered. One significant limitation of card access, however, is
that a lost or stolen card can be used by anyone until it is deactivated.

Biometric access systems, discussed above as logical access controls, also serve as physical access con-
trols. They can be used when physical security needs to be rigorous. Biometric access systems use physical
characteristics such as blood vessel patterns on the retina, handprints, or voice authentication to authorize
access. In general, such systems have a low error rate. That said, no single system is completely error-
free, so biometric access systems are usually combined with other controls.

Controls can also be designed to limit activities that can be performed remotely. For example, changes
to employee pay rates can be restricted to computers physically located in the payroll department. Thus,
even if online thieves managed to steal a payroll password, they would be prevented from changing pay
rates because they would not have access to the premises.

284 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.3. Technology-enabled Finance Transformation

F.3. Technology-enabled Finance Transformation


Systems Development Life Cycle (SDLC)
The systems development life cycle was described previously in System Controls and Security Measures as
System and Program Development and Change Controls, and it will be briefly reviewed here.

1) Statement of objectives. A written proposal is prepared, including the need for the new system,
the nature and scope of the project and timing issues. A risk assessment is done to document
security threats, potential vulnerabilities, and the feasible security and internal control safeguards
needed to mitigate the identified risks.

2) Investigation and feasibility study of alternative solutions. Needs are identified and feasi-
bility of alternative solutions are assessed, including the availability of required technology and
resources. A cost-benefit analysis is done.

3) Systems analysis. The current system is analyzed to identify its strong and weak points, and the
information needs from the new system, such as reports to be generated, database needs, and
the characteristics of its operation are determined.

4) Conceptual design. Systems analysts work with users to create the design specifications and
verify them against user requirements.

5) Physical design. The physical design involves determining the workflow, what and where pro-
grams and controls are needed, the needed hardware, backups, security measures, and data
communications.

6) Development and testing. The design is implemented into source code, the technical and phys-
ical configurations are fine-tuned, and the system is integrated and tested. Data conversion
procedures are developed.

7) System implementation and conversion. The site is prepared, equipment is acquired and in-
stalled, and conversion procedures, including data conversion, are implemented. System
documentation is completed, procedures are developed and documented, and users are trained.

8) Operations and maintenance. The system is put into a production environment and used to
conduct business. Continuous monitoring and evaluation take place to determine what is working
and what needs improvement. Maintenance includes modifying the system as necessary to adapt
to changing needs, replacing outdated hardware as necessary, upgrading software, and making
needed security upgrades.

If follow-up studies indicate that new problems have developed or that previous problems have recurred,
the organization begins a new systems study.

Business Process Analysis


Business process reengineering and redesign, discussed in Section D, Cost Management,79 relies on tech-
nology to accomplish its objectives. When a business process needs to be reengineered or redesigned, new
information systems will be needed.

For example, if the same input for a process is being keyed in more than once, such as into a database and
also into a spreadsheet, that process needs to be redesigned so that the input is keyed in only one time.
Not only is the duplication of effort wasteful, but inputting the information multiple times opens the process
to the risk that the data will be keyed in differently each time. However, in order to successfully accomplish
the redesign of the process, the information system being used will need to be redesigned, as well.

79
See Business Process Reengineering and Accounting Process Redesign.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 285
F.3. Technology-enabled Finance Transformation CMA Part 1

Any business operation, or process, consists of a group of related tasks or events that has a particular
objective. Business process analysis is used to analyze a business process to determine the specific way
the process is presently being accomplished from beginning to end. Business process analysis can provide
information needed to monitor efficiency and productivity, locate process weaknesses, pinpoint potential
improvements, and determine whether the potential improvements should be carried out.

Business process analysis involves the following steps:

1) Determine the process to be analyzed. Processes that influence revenue, expenses, the end
product, and other critical functions are processes that should be analyzed periodically, as are
processes that appear to be underperforming. A new process just implemented might also be
analyzed to determine whether it is working as intended. Determine the beginning and ending
points of the process to be analyzed.

2) Collect information about the process that will be needed to analyze it. Go through the
documentation, interview the people involved, and do any other research necessary to answer any
questions that arise.

3) Map the process. Business process mapping is visualizing the whole process from start to finish
to better understand the various roles and responsibilities involved. Mapping makes it easier to
see the big picture, what is working and what is not working, and where the risks are.

A flowchart can be created for this step, or several software solutions, called business process
analysis tools, are available for business process mapping.

4) Analyze the process. For example, determine the most important components and whether they
could be improved. Look for any delays or other problems and determine whether they can be
fixed. Look for ways to streamline the process so it uses fewer resources.

5) Determine potential improvements. Make recommendations for ways to improve the process.
For example, determine whether incremental changes are needed and if so, what they are, or
whether the process needs to be completely reengineered. Business process analysis tools can be
an important part of this step, because they can be used to model changes to the process and
prepare visuals.

Robotic Process Automation (RPA)


Robotic process automation (RPA), a type of artificial intelligence (see next topic), is not the same thing as
the use of industrial robots. Robotic process automation software automates repetitive tasks by interacting
with other IT applications to execute business processes that would otherwise require a human. RPA soft-
ware can communicate with other systems to perform a vast number of repetitive jobs, and it can operate
around the clock with no human errors. The automation of the repetitive part of a job frees up employees
to do other things. RPA software cannot replace an employee, but it can increase the employee’s produc-
tivity.

For example, RPA can be used when a change in policy necessitates a change in processing that would
otherwise require additional employee time to implement or when sales growth causes changes in a system
that would require either costly integration with another system or employee intervention.

The RPA software is not part of the organization’s IT infrastructure. Instead, it interacts with the IT infra-
structure, so no change is needed to the existing IT systems. Thus, RPA allows the organization to automate
what would otherwise be a manual process without changing the existing systems.

Note: Robotics process automation allows users to create their own robots that can perform high-
volume, repeatable tasks of low complexity faster and more accurately than humans can.

286 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.3. Technology-enabled Finance Transformation

The software robots, also called “clients” or “agents,” can log into applications, move files, copy and paste
items, enter data, execute queries, do calculations, maintain records and transactions, upload scanned
documents, verify information for automatic approvals or rejections, and perform many other tasks.

• RPA can be used to automate portions of transaction reporting and budgeting in the accounting
area.

• RPA can automate manual consolidations of financial statements, leaving the accountants more
time to follow up on items that require investigation, perhaps because of significant variances.

• Financial institutions can use RPA to automate account openings and closings.

• Insurance companies can use it to automate claims processing.

• RPA can be used in supply chain management for procurement, automating order processing and
payments, monitoring inventory levels, and tracking shipments.

Any task that is high volume, rules-driven, and repeatable qualifies for robotic process automation. The
RPA software can create reports and exception alerts so employees or management know when to get
involved. The result can be cost savings, a higher accuracy rate, faster cycle times, and improved scalability
if volumes increase or decrease. Employees can focus on more value-added activities.

Benefits of Robotic Process Automation

• Developing a process in RPA software does not require coding knowledge. RPA software usually has
“drag-and-drop” functionality and simple configuration wizards that the user can employ to create
an automated process. Some RPA solutions can be used to define a process by simply capturing a
sequence of user actions.

• It enables employees to be more productive because they can focus on more advanced and engag-
ing tasks, resulting in lower turnover and higher employee morale.

• It can be used to ensure that business operations and processes comply with regulations and stand-
ards.

• The tasks performed can be monitored and recorded, creating valuable data and an audit trail to
further help with regulatory compliance as well as to support process improvement.

• Once an RPA process has been set up, the process can be completed much more rapidly.

• Robotic process automation can result in cost savings.

• It can help provide better customer service by automating customer service tasks. Customer service
personnel can make use of it, or in some cases, RPA can even be used to converse with customers,
gathering information and resolving their queries faster and more consistently than a person could.

• Low-volume or low-value processes that would not be economical to automate via other means can
be automated using RPA.

• Business process outsourcing providers can use RPA tools to lower their cost of delivery or to offer
“robots-as-a-service.”

• Robots follow rules consistently, do not need to sleep, do not take vacations, do not get sick, and
do not make typographical errors.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 287
F.3. Technology-enabled Finance Transformation CMA Part 1

Limitations of Robotic Process Automation

• Robots are not infallible. Like any machine, their reliability is not 100%. Poor quality data input can
cause exceptions, and their accuracy can be affected by system changes.

• Robots cannot replicate human reasoning. RPA software can mimic human behavior in the way it
interacts with application user interfaces, but it can only follow highly methodical instructions.

• Robots have no “common sense.” If a flaw in the instructions creates an error that would be obvious
to a human, the robot will continue to follow the instructions provided without deviation and the
error may be replicated hundreds or thousands of times before it is recognized by a human. Then,
correcting all the incidents of the error could be very difficult (unless the errors could be corrected
using the same automated tools).

• Because RPA can be used to automate processes in a “noninvasive” manner (in other words, without
changing the IT system), management may be tempted to deploy RPA without relying on assistance
from the IT department. However, although RPA can be deployed without involving the IT depart-
ment, doing so may lead to unexpected problems. The IT department needs to be involved in the
effort so the deployment is stable.

Artificial Intelligence (AI)


Artificial intelligence is a field in computer science dedicated to creating intelligent machines, especially
computers, that can simulate human intelligence processes. AI uses algorithms, which are sets of step-
by-step instructions that a computer can execute to perform a task. Some AI applications are able to learn
from data and self-correct, according to the instructions given.

Artificial intelligence is categorized as either weak AI, also called narrow AI, or strong AI, also called
artificial general intelligence (AGI).

• Weak AI is an AI system that can simulate human cognitive functions but although it appears to
think, it is not actually conscious. A weak AI system is designed to perform a specific task, “trained”
to act on the rules programmed into it, and it cannot go beyond those rules.

o Apple’s Siri80 voice recognition software is an example of weak AI. It has access to the whole
Internet as a database and is able to hold a conversation in a narrow, predefined manner; but
if the conversation turns to things it is not programmed to respond to, it presents inaccurate
results.

o Industrial robots and robotic process automation are other examples of weak AI. Robots can
perform complicated actions, but they can perform only in situations they have been pro-
grammed for. Outside of those situations, they have no way to determine what to do.

• Strong AI is equal to human intelligence and exists only in theory. A strong AI system would be
able to reason, make judgments, learn, plan, solve problems, communicate, create and build its
own knowledge base, and program itself. A strong AI system would be able to find a solution for
an unfamiliar task without human intervention. It could theoretically handle all the same work that
a human could, even the work of a highly-skilled knowledge worker.

Artificial intelligence is increasingly being used in administrative procedures and accounting. Robotic pro-
cess automation, covered in the previous topic, is one application of AI. Other applications are digital
assistants powered by AI and speech recognition (such as Siri), machine vision, and machine learning.

80
Siri is a trademark of Apple Inc., registered in the U.S. and other countries.

288 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.3. Technology-enabled Finance Transformation

Digital assistants have become standard in smartphones and for controlling home electronics, and their
use has expanded into enterprise applications, as well. Some Enterprise Resource Planning systems incor-
porate digital assistants. The Oracle Cloud application includes functionality for enterprises to create
chatbots and virtual assistants. A chatbot is software that can conduct a conversation via auditory or text
methods. Examples of the use of chatbots are customer service and for acquisition of information.

Machine vision includes cameras, image sensors, and image processing software. It can automate indus-
trial processes such as quality inspections by enabling robots to “see” their surroundings. Machine vision is
also used in non-industrial settings such as surveillance and medical applications. It is increasingly being
used in administrative and accounting applications, as well.

• Machine vision can be used to analyze satellite imagery for several purposes.

o Insurance agents can use it to verify property information provided by existing clients or iden-
tify physical features of properties such as the roof condition and validate property features
such as building size prior to providing an insurance quote to a new client, thereby reducing
inspection costs.

o Investment firms can use it to determine economic trends and forecast retail sales based on
the number of cars in a retail parking lot on an hourly basis.

o Financial institutions can monitor the status of construction on projects for construction lending
purposes.

o Businesses making investments in projects can use it to assess the degree to which a project
is complete for accounting purposes and for monitoring and management.

• Machine vision can be used to automate document data extraction.

o Businesses can assess large numbers of incoming paper documents or forms to extract the
information from them. When documents are fed to the system, the software can identify each
document as to its type and sort the documents to be forwarded to the appropriate processing
group.

o Incoming paper documents can be digitized for review by human employees, eliminating the
need for manual data entry. Instead of doing the manual data input, the human employees
can instead spend their time reviewing and ensuring the accuracy of fields entered by the
machine learning software. The machine vision can even “read” handwritten text. When the
data extraction is manually done, in many cases the full amount of the data is not input but
only the most important data points are extracted. With machine vision, the data that results
is complete and organized, and greater insights can be gained from data analytics.

Machine learning is another aspect of artificial intelligence being put to use in the accounting area. In
machine learning, computers can learn by using algorithms to interpret data in order to predict outcomes
and learn from successes and failures. Computers can “learn” to perform repeatable and time-consuming
jobs such as the following.

• Checking expense reports. Computers can learn a company’s expense reimbursement policies,
read receipts, and audit expense reports to ensure compliance. The computer can recognize ques-
tionable expense reimbursement claims and forward them to a human to review.

• Analyzing payments received on invoices. When a customer makes a payment that needs to
be applied to multiple invoices or that does not match any single invoice in the system, accounts
receivable staff might need to spend time figuring out the proper combination of invoices to clear
or may need to place a call to the customer, requiring considerable time and effort. However, a
“smart machine” can analyze the possible invoices and match the payment to the right combination
of invoices. Or, if the payment is short (is less than the amount due), the computer can apply the
short payment and automatically generate an invoice for the remaining amount without any human
intervention.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 289
F.3. Technology-enabled Finance Transformation CMA Part 1
efham CMA
• Risk assessment. Machine learning can be used to compile data from completed projects to be
used to assess risk in a proposed project.

• Data analytics. Using available data, machines can learn to perform one-time analytical projects
such as how much the sales of a division have grown over a period of time or what the revenue
from sales of a specific product was during a period of time.

• Bank reconciliations. Machines can learn to perform bank reconciliations.

AI-enabled robots will not replace accountants, but they will substantially transform what accountants do.
When machines are able to do the repetitious work of calculating, reconciling, transaction coding, and
responding to inquiries, accountants can focus less on tasks that can be automated and more on work such
as advisory services that can be done only by humans, thereby increasing their worth in an organization.
Accountants will need to monitor the interpretation of the data processed by AI to ensure that it continues
to be useful for decision making. Accountants will need to embrace AI, keep their AI and analytical skills
current, and be adaptive and innovative in order to remain competitive.

Considerations in Instituting Artificial Intelligence


• Processes should be re-imagined where possible, rather than just using the AI to replicate existing
processes.

• Activities to be performed by AI should be those that are standardized and not often changed.

• Processes that are automated should be fully documented.

• Data quality, both input and output, must be reviewed. Potential exceptions and errors requiring
human intervention must be identified and investigated.

Cloud Computing
Cloud computing is a method of essentially outsourcing the IT function. It is a way to increase IT capacity
or add capabilities without having to invest in new infrastructure or license new software.

The National Institute of Standards and Technology (NIST) of the U.S. Department of Commerce defines
cloud computing as follows:

Cloud computing is a model for enabling ubiquitous, convenient, on-demand network ac-
cess to a shared pool of configurable computing resources (e.g., networks, servers,
storage, applications, and services) that can be rapidly provisioned and released with min-
imal management effort or service provider interaction.81

Thus, cloud computing means the use of business applications offered over the Internet. Cloud computing
resources include data storage, infrastructure and platform (that is, hardware and operating system), and
application software. Cloud service providers offer all three types of resources.

81
The NIST Definition of Cloud Computing, Special Publication (NIST SP) Report Number 800-145, Computer Security
Division, Information Technology Laboratory, National Institute of Standards and Technology, U.S. Department of Com-
merce, Gaithersburg, MD, September 2011, p. 2, https://nvlpubs.nist.gov/nistpubs/Legacy/SP/nistspecialpublication800-
145.pdf, accessed April 22, 2019.

290 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.3. Technology-enabled Finance Transformation

Software as a Service (SaaS) is defined by NIST as follows:

The capability provided to the consumer is to use the provider’s applications running on a
cloud infrastructure. The applications are accessible from various client devices through
either a thin client interface, such as a web browser (e.g., web-based email), or a program
interface. The consumer does not manage or control the underlying cloud infrastructure
including network, servers, operating systems, storage, or even individual application ca-
pabilities, with the possible exception of limited user-specific application configuration
settings.82

In other words, SaaS is software that has been developed by a cloud provider for use by multiple businesses
(called multi-tenant use), and all business customers use the same software. Applications available as
SaaS applications include enterprise resource planning (ERP), customer relationship management (CRM),
accounting, tax and payroll processing and tax filing, human resource management, document manage-
ment, service desk management, online word processing and spreadsheet applications, email, and many
others.

Cloud computing also includes Platform as a Service (PaaS), and Infrastructure as a Service (IaaS).

NIST’s definition of Platform as a Service is

The capability provided to the consumer is to deploy onto the cloud infrastructure con-
sumer-created or acquired applications created using programming languages, libraries,
services, and tools supported by the provider. The consumer does not manage or control
the underlying cloud infrastructure including network, servers, operating systems, or stor-
age, but has control over the deployed applications and possibly configuration settings for
the application-hosting environment.83

If a company uses Platform as a Service, the company deploys its own applications to the cloud using the
cloud provider’s operating systems, programming languages, libraries, services, and tools. PaaS services
include operating systems, database solutions, Web servers, and application development tools.84

Infrastructure in the context of cloud computing is both the hardware resources that support the cloud
services being provided, including server, storage, and network components, and the software deployed. 85

The definition of Infrastructure as a Service (IaaS) according to NIST is

The capability provided to the consumer is to provision processing, storage, networks, and
other fundamental computing resources where the consumer is able to deploy and run
arbitrary software, which can include operating systems and applications. The consumer
does not manage or control the underlying cloud infrastructure but has control over oper-
ating systems, storage, and deployed applications and possibly limited control of select
networking components (e.g., host firewalls).86

A company using Infrastructure as a Service is provided with physical and virtual processing, storage,
networks, and other computing resources. The company can use the infrastructure to run software and
operating systems. Although the company does not manage or control the cloud infrastructure, it does have
control over the operating systems, storage, and deployed applications it uses, and it may have some
control over things like configuration of a host firewall. Examples of Infrastructure as a Service include
storage servers, network components, virtual machines, firewalls, and virtual local area networks. 87

82
Ibid.
83
Ibid., pp. 2-3.
84
Moving to the Cloud, Joseph Howell, Strategic Finance magazine, June 2015, © Institute of Management Accountants,
https://sfmagazine.com/post-entry/june-2015-moving-to-the-cloud/, accessed April 22, 2019.
85
The NIST Definition of Cloud Computing, Special Publication (NIST SP) Report Number 800-145, Computer Security
Division, Information Technology Laboratory, National Institute of Standards and Technology, U.S. Department of Com-
merce, Gaithersburg, MD, September 2011, p. 2, note 2,
https://nvlpubs.nist.gov/nistpubs/Legacy/SP/nistspecialpublication800-145.pdf, accessed April 22, 2019.
86
Ibid., p. 3.
87
Moving to the Cloud, Joseph Howell, Strategic Finance magazine, June 2015, © Institute of Management Accountants,
https://sfmagazine.com/post-entry/june-2015-moving-to-the-cloud/, accessed April 22, 2019.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 291
F.3. Technology-enabled Finance Transformation CMA Part 1

Benefits of Cloud Computing, SaaS, PaaS, and IaaS

• Users pay for only what they use, either on a periodic basis or on a usage basis. Thus, cloud com-
puting is scalable. A firm can quickly increase or decrease the scale of its IT capability.

• Since the provider owns and operates the hardware and software, a user organization may be able
to decrease its investment in its own hardware and software.

• The provider keeps the software updated, so the user organizations do not need to invest in up-
grades or be concerned with applying them.

• Applications and data resident in the cloud can be accessed from anywhere, from any compatible
device.

• Technology available in the cloud can be leveraged in responding to new and existing requirements
for external compliance reporting, sustainability and integrated reporting, internal management
reporting, strategic planning, budgeting and forecasting, performance measurement, risk manage-
ment, advanced analytics, and many others.

• Cloud technology can be used to free up accountants so they can handle more higher-value activities
and streamline lower-value processes.

• The cloud can enable the CFO to move into a more strategic role instead of spending time on
transactional activities.

• The cloud can provide greater redundancy of systems than an on-site IT department may be able
to offer, particular for small to medium-sized entities that may not be able to afford backup systems.

• The cloud can offer to companies of all sizes the advanced computing power needed for advanced
analytics, something that otherwise only the largest companies would be able to afford due to cost.
As a result, small to medium-sized businesses can be better positioned to compete with much larger
competitors.

• Although security is a concern with the cloud, security is a concern with on-site IT, as well. The
cloud frequently can provide stronger infrastructure and better protection than an on-site IT de-
partment may be able to.

Limitations, Costs, and Risks of Cloud Computing, SaaS, PaaS, and IaaS

• Reliability of the Internet is a concern. If the Internet goes down, operations stop.

• The quality of the service given by the provider needs to be monitored, and the service contract
needs to be carefully structured.

• Loss of control over data and processing introduces security concerns. Selection of the cloud vendor
must include due diligence. The cloud vendor must demonstrate that it has the proper internal
controls and security infrastructure in place, and the vendor’s financial viability as a going concern
needs to be ascertained. Furthermore, the vendor’s internal controls over data security and its in-
frastructure, as well as its continued viability as a going concern, need to be monitored on an
ongoing basis.

• Contracting with overseas providers may lead to language barrier problems and time-zone problems
as well as quality control difficulties.

• The ability to customize cloud solutions is limited, and that may hamper management from achiev-
ing all that it hopes to achieve.

• Service provided by automatic backup service providers may be problematic because timing of au-
tomatic backups may not be controllable by the user and may not be convenient for the user.
(Continued)

292 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.3. Technology-enabled Finance Transformation

• The cloud cannot overcome weak internal controls. People are the greatest area of weakness with
both internal IT and with cloud technologies. Security awareness training, proper hiring procedures,
good governance, and protection from malware continue to be necessary after a company moves
to the cloud, just as they are when the IT is on-site.

• The company’s data governance must be structured to cover the cloud and the risks inherent in it,
such as employees downloading new applications without authorization.

• Expected cost savings may not materialize. An organization may find that managing its own IT
internally, even with all of its attendant problems, is less expensive than using the cloud.

Blockchains, Distributed Ledgers, and Smart Contracts


The concept of the blockchain and the first cryptocurrency, bitcoin, was first presented in 2009 in a paper
titled Bitcoin: A Peer-to-Peer Electronic Cash System, ostensibly written by Satoshi Nakamoto. No one
knows who Satoshi Nakamoto is, and it may be a pseudonym for one or more people.88

The blockchain was initially envisioned as a peer-to-peer system for sending online payments from one
party to another party without using a financial institution. While online payments are still important, block-
chain technology has expanded and is now used in many other areas.

Blockchain Terminology
Blockchain – A blockchain is a public record of transactions in chronological order, more specifically “a
way for one Internet user to transfer a unique piece of digital property to another Internet user, such that
the transfer is guaranteed to be safe and secure, everyone knows that the transfer has taken place, and
nobody can challenge the legitimacy of the transfer.” 89 Thus, a blockchain is a continuously growing digital
record in the form of packages, called blocks, that are linked together and secured using cryptography. A
blockchain is a system of digital interactions that does not need an intermediary such as a financial institu-
tion to act as a third party to transactions. Many users write entries into a record of information, the
transactions are timestamped and broadcast, and the community of users controls how the record of infor-
mation is updated. The digital interactions are secured by the network architecture of blockchain
technology. The blocks are maintained via a peer-to-peer network of computers, and the same chain of
blocks, called a ledger, is stored on many different computers. A blockchain can be public, private, or a
hybrid, which is a combination of public and private.

Public blockchain - A public blockchain is open to anyone, anyone can contribute data to the ledger, and
all participants possess an identical copy of the ledger. A public blockchain is also called a permissionless
ledger. The public blockchain has no owner or administrator, but it does have members who secure the
network, and they usually receive an economic incentive for their efforts.

Private blockchain – A private blockchain, also called a permissioned ledger, allows only invited partic-
ipants to join the network. Permissioned ledgers are controlled by one or more network administrators. All
of the members—but only the members—have copies of the ledger. Private blockchains can be used by a
single entity, for example to manage its value chain.

Hybrid blockchain – A hybrid blockchain is a mix of a public and private blockchain. Some processes are
kept private and others are public. Participants in public or private networks are able to communicate with
each other, enabling transactions between them across networks. A hybrid blockchain can be used by a
supply chain group to control the supply chain.

88
For more information, see https://en.wikipedia.org/wiki/Satoshi_Nakamoto.
89
Marc Andreessen, quoted on https://www.coindesk.com/information/what-is-blockchain-technology, accessed April
25, 2019.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 293
F.3. Technology-enabled Finance Transformation CMA Part 1

Encryption - Encryption is used to keep data on a blockchain private. Encryption makes use of an algorithm
that transforms the information into an unreadable format. A key is required to decrypt the data into its
original, readable format. A key for electronic/digitally encrypted information is bits of code that use an
algorithm to lock and unlock information.

Private key encryption - Private key encryption is a form of encryption in which a single private key can
both encrypt and decrypt information. While private key encryption is fast because it uses a single key, the
privacy of that secret key is essential because anyone with that key can decrypt the information. Because
of the risk of the private key being stolen or leaked, the private keys must be changed frequently.

Public key infrastructure – Public key infrastructure uses two keys: private and public. The public key is
distributed, but the private key is never shared. Each public key has a corresponding private key, and the
two keys are mathematically linked based on very large prime numbers.

Private key/public key encryption – A blockchain uses public key/private key encryption. On a block-
chain, the private key of a user is converted into a private address, and the user’s public key is converted
into a public address. On a blockchain, private key cryptography fulfills the need to authenticate users. A
private key is used to create a digital signature. Possession of a private key equals ownership and
also prevents its owner from having to share any more personal information than is needed for the trans-
action, which protects the user from hackers.

When a user wants to send funds, the sender “signs” the transaction with his or her private key, and the
software on the user’s computer creates the digital signature, which is sent out to the network for validation.
Validation includes confirming that the sender owns the funds being sent and that the sender has not
already sent those same funds to someone else. The network validates the transaction by entering the
digital signature and the sender’s public key (which is publicly known) into the software. If the signature
that was created with the user’s private key corresponds to the user’s public key, the program validates
the transaction.

Node – A node is a powerful computer running software that keeps the blockchain running by participating
in the relay of information. Nodes communicate with each other to spread information around the network.
A node sends information to a few nodes, which in turn relay the information to other nodes, and so forth.
The information moves around the network quickly.

Mining nodes, or “miners” – Miners are nodes (computers) on the blockchain that group outstanding
transactions into blocks and add them to the blockchain.

Distributed ledger – A distributed ledger is a database held by each node in a network, and each node
updates the database independently. Records are independently constructed and passed around the net-
work by the various nodes—they are not held by any central authority. Every node on the network has
information on every transaction and then comes to its own conclusion as to whether each transaction is
authentic (that is, whether the people are who they say they are) and, if the transaction is a payment
transaction, whether the sender has enough funds to cover the payment. The data sent in a transaction
contains all the information needed to authenticate and authorize the transaction. When there is a consen-
sus among the nodes that a transaction is authentic and should be authorized, the transaction is added to
the distributed ledger, and all nodes maintain an identical copy of the ledger.

Hash – Hashing is taking an input string of any length and giving it an output of a fixed length using a
hashing algorithm. For example, bitcoin uses the hashing algorithm SHA-256 (Secure Hashing Algorithm
256) on Bitcoin networks. Any input, no matter how big or small, always has a fixed output of 64 symbols,
which is made up of 256 bits, the source of the “256” in its name. The fixed output is the hash.

Block – A block is a record in a blockchain that contains and confirms many waiting transactions. It is a
group of cryptocurrency transactions that have been encrypted and aggregated into the block by a miner.
Each block has a header that contains (1) the details of the transactions in the block, including the senders’
and receivers’ addresses for each transaction and the amount of funds to be transferred from each sender
to each receiver, (2) the hash of the information in the block just preceding it (which connects it to the

294 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.3. Technology-enabled Finance Transformation

blockchain), (3) a “nonce” (see below), and (4) the hash of the information in the block, including the
nonce.

Nonce - The nonce in a block is a random string of characters that is appended to the transaction infor-
mation in the block before the block is hashed and it is used to verify the block. After a nonce is added to
the block, the information in the block, including the nonce, is hashed. The nonce needs to be a string of
characters that causes the hash of the whole block, including the nonce, to conform to a particular require-
ment: the hash of the block that results after the nonce is included must contain a certain number of leading
zeroes. If it does not, the nonce must be changed and the block hashed again.

The powerful mining nodes on the network all compete to determine the block’s nonce by trying different
values and re-hashing the block multiple times until one miner determines a nonce that results in the hash
of the block conforming to the requirement for the number of leading zeroes. The mining node on the
network that is the first to “solve the puzzle”—that is, calculate a nonce that results in a hash value for the
block that meets the requirement for the number of leading zeroes—receives a reward of a certain number
of units of the digital currency. (That is the source of the term “mining nodes.” Receiving the currency
reward is called “mining” the block because new digital currency is created and received.)

The blocks are hard to solve but easy to verify by the rest of the network once they are solved. Therefore,
after one mining node solves the block, the other nodes on the network check the work to determine
whether the block’s nonce and its hash have been correctly calculated. If the other nodes agree that the
calculations have been done correctly, the block is validated.

The determination of a nonce that fulfills the requirement for a new block’s hash usually takes about 10
minutes and the nonce that fulfills the requirement is called “Proof of Work.”

Note: “Proof of Work” is the consensus algorithm used on the Bitcoin blockchain. A proof of work is
a piece of data that satisfies certain requirements, is costly and time-consuming to produce, but is easy
for others to verify.

A consensus algorithm is a set of rules and number of steps for accomplishing a generally accepted
decision among a group of people. For blockchains, consensus algorithms are used to ensure that a
group of nodes on the network agree that all transactions are validated and authentic. The Proof of Work
(PoW) algorithm is the most widely used consensus algorithm, but other blockchains may use different
algorithms to accomplish the same thing.

Consensus algorithms are used to prevent double spending by a user, that is, spending the same digital
currency more than once. In a traditional means of exchange, financial intermediaries ensure that cur-
rencies are removed from one account and placed into another, so that the same money is not spent
more than once. However, digital currencies are decentralized and so there are no financial intermedi-
aries. Consensus algorithms are used to make sure that all the transactions on the blockchain are
authentic and that the currency moves from one entity to another entity. All the transactions are stored
on the public ledger and can be seen by all participants, creating full transparency.

Confirmation – When a block has been validated on a blockchain, the transactions processed in it are
confirmed. Confirmation means the transactions in the block have been processed by the network. Trans-
actions receive a confirmation when they are included in a block and when each subsequent block is linked
to them. Once a transaction has received just one confirmation it is not likely to be changed because
changing anything in the block would change the block’s hash value. Since each block’s hash value is part
of the hash of the following block on the blockchain, changing anything in one block would require re-
hashing all of the blocks following the changed block. Re-hashing all of the following blocks would necessi-
tate recalculating a proper nonce for each subsequent block in turn.

The general rule is that after six confirmations (in other words, six subsequent new blocks have been
created following an added block, requiring work of about 60 minutes, that is, 6 blocks at 10 minutes per
block), the work involved to make a change in the previously-added block makes doing so prohibitive. Each
added confirmation exponentially decreases the risk that a transaction will be changed or reversed.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 295
F.3. Technology-enabled Finance Transformation CMA Part 1

If any change to a confirmed transaction is needed, a new transaction must be created.

Note: It is not impossible to change a previously-recorded transaction, but it would be very, very difficult
to do so. Because of the time that would be required to recalculate all the nonces in all the subsequent
blocks, it would be next to impossible to “catch up” to the most recent block, since new blocks would be
being added to the chain all the time. Thus, transactions on a blockchain are considered immutable,
which means “not subject or susceptible to change.”

Immutability is important because if a transaction can be changed, a hacker could change the receivers’
addresses and divert the payments, thus stealing a significant amount of the currency. That has actually
happened to some of the smaller, less active, cryptocurrencies.

Bitcoin and Other Cryptocurrencies


The first usage of a blockchain was to transfer virtual currency, or cryptocurrency. A virtual currency is a
digital representation of value that functions as a medium of exchange, a unit of account, and/or a store of
value.90 It is a piece of computer code that represents ownership of a digital asset.

Virtual currencies that have an equivalent value in real currency or that can act as a substitute for real
currency are called “convertible virtual currency.” Bitcoin was the first cryptocurrency and it is a convertible
virtual currency. It can be digitally traded between users and can be purchased for and exchanged into U.S.
dollars, Euros, and other currencies.91

The term “Bitcoin” is also used to refer to the protocol for the distributed ledger, that is, the distributed
network that maintains a ledger of balances held in bitcoin tokens by users on the network. The word
“bitcoin” with a lower-case “b” refers to the tokens, while the word “Bitcoin” with a capital “B” refers to the
Bitcoin protocol and the Bitcoin network.

Note: Do not confuse the “distributed ledger” of a blockchain with an accounting ledger in which double-
entry accounting is performed. When a blockchain is used to make payments, the transactions in the
blockchain are single entry transactions for each entity and they represent the amount of cryptocurrency
to be paid by the payor and received by the receiver. Although the transactions are not double-entry
accounting entries for each entity, in a sense each currency transaction will be double entry since a
sender’s payment is equal to the receiver’s receipt.

Bitcoin tokens can be used to make payments electronically if both parties agree to use it. The system
enables payments in bitcoin tokens to be sent from one user to another user without having to pass through
a central authority. When a transaction is finalized by being added to the blockchain, the sender’s balance
of bitcoin is reduced and the receiver’s balance of bitcoin is increased.

Bitcoin was the first cryptocurrency, but over a thousand other cryptocurrencies are being used now, as
well—ether, bitcoin cash, litecoin, XRP, and binance coin, to name just a few.

Bitcoin and other cryptocurrencies are not backed by any government or commodity such as gold. They are
backed by technology and cryptographic proof and do not rely on a trusted third party such as a financial
institution to act as intermediary. Users of cryptocurrencies put their trust in a digital, decentralized, and
basically anonymous system that maintains its integrity through cryptography and peer-to-peer network-
ing. Cryptocurrencies are unregulated, and no central bank can take corrective actions to protect the value
of a cryptocurrency in the event of a crisis. In the U.S., cryptocurrencies do not currently have legal tender
status. Nevertheless, many cryptocurrencies can be purchased, sold, and exchanged on cryptocurrency
exchanges such as Bitstamp, Wirex, and Coinbase. The values of cryptocurrencies in terms of other cur-
rencies such as the U.S. dollar or the Euro are extremely volatile, however.

90
A CFTC Primer on Virtual Currencies,” October 17, 2017, https://www.cftc.gov/LabCFTC/Primers/Index.htm, p. 4,
accessed April 29, 2019.
91
Ibid.

296 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.3. Technology-enabled Finance Transformation

Other uses of blockchain include:

• Private, permissioned blockchains can be used by financial institutions for trading, payments, clear-
ings, settlements, and repurchase agreement transactions (short-term borrowing of securities).92

• Intercompany transactions where different ERP systems are in use can be streamlined using a
blockchain.

• Procurement and supply chain operations on blockchain can be used to optimize accounts payable
or accounts receivable functions.

Note: Investor Alert

On April 26, 2019, the U.S. CFTC (Commodity Futures Trading Commission) and the U.S. SEC (Securities
and Exchange Commission) jointly issued an investor alert warning about websites purporting to operate
advisory and trading businesses related to digital assets. In some cases, the fraudsters claim to invest
customers’ funds in proprietary crypto trading systems or in “mining” farms, promising high guaranteed
returns with little or no risk. After an investor makes an investment, usually using a digital currency such
as bitcoin, sometimes the fraudsters stop communicating. Sometimes the fraudsters require the investor
to pay additional costs to withdraw fake “profits” earned on the investment, “profits” that are nonexist-
ent.

Investors should beware of “guaranteed” high investment returns, complicated language that is difficult
to understand, unlicensed and unregistered firms, and any offer that sounds too good to be true.

Other investor alerts issued recently by the CFTC have including warnings about virtual currency “pump-
and-dump” schemes,93 “IRS approved” virtual currency IRAs, fraudulent “Initial Coin Offerings,” (ICOs)
and various other fraudulent promotions.

For more information, see:

https://www.cftc.gov/PressRoom/PressReleases/7917-19

https://www.cftc.gov/Bitcoin/index.htm

92
A CFTC Primer on Virtual Currencies,” The U.S. Commodity Futures Trading Commission, October 17, 2017, p. 8,
https://www.cftc.gov/LabCFTC/Primers/Index.htm, accessed April 29, 2019.
93
A “pump-and-dump” scheme involves aggressively promoting an investment by falsely promising outsize gains to
come (“pumping” it), thus inflating demand for the investment. The inflated demand inflates the investment’s market
price. When the price reaches a certain point, the promoters sell their holdings (“dump” them) on the open market, the
market price crashes, and the investors are left with large losses on a nearly worthless investment while the promoters
make off with the capital gain.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 297
F.3. Technology-enabled Finance Transformation CMA Part 1

Smart Contracts
Blockchain technology can be used for more than payments. Blockchains are being used to store all kinds
of digital information and to execute contracts automatically. A contract that has been digitized and up-
loaded to a blockchain is called a smart contract.

According to Nick Szabo, a computer scientist who envisioned smart contracts in 1996,

“A smart contract is a set of promises, specified in digital form, including protocols within
which the parties perform on these promises.”94

A smart contract is created by translating the terms and conditions of a traditional agreement into a com-
putational code written by blockchain developers in a programming language. It is basically a set of coded
computer functions with a set of rules. The computer code is self-executing and performs an action at
specified times or based on the occurrence or non-occurrence of an action or event such as delivery of an
asset or a change in a reference rate. A simple example is a translation of “If X occurs, then Y makes a
payment to Z.”

A smart contract may include the elements of a binding contract (offer, acceptance, and consideration), or
it may simply execute certain terms of a contract. When a smart contract is uploaded to a blockchain, the
validity of the contract is checked and the required steps are enabled. After that, it is automatically exe-
cuted.

A blockchain executes smart contracts in basically the same way as it executes transactions, and when the
contract calls for payments to be made, they are made automatically in the cryptocurrency of the particular
blockchain system being used. As with payments, the information in the contract is encrypted and hashed
to a standardized size. A smart contract in a block of data is linked to the previous block. The smart contract
is executed and payments are transferred according to its terms within the blockchain. The contract is
funded by the payor so it can perform transactions automatically according to the agreement.

Specific technologies are available, and a smart contract is executed on a blockchain-based platform specific
to the technology being used. For example, an Ethereum smart contract is executed on an Ethereum block-
chain using the Ethereum Virtual Machine (EVM). Ethereum is a blockchain-based platform that was
developed with smart contracts in mind and the smart contracts on Ethereum are self-executing. The cryp-
tocurrency used to transfer funds on Ethereum is called Ether. Other platforms and cryptocurrencies are
also available for smart contracts, but Ethereum is one of the more well-known platforms.

Following are some examples of the uses of smart contracts on blockchains.

• A blockchain can be used to ensure the authenticity of a product, so a purchaser of the product
can be assured that the product he or she is buying is genuine and not a counterfeit. The infor-
mation stored on the blockchain is unchangeable, so it is easier to prove the origins of a given
product.

• It can be used to protect intellectual property. For example, artists can protect and sell music on
a blockchain system. Artists who are due royalties from each sale of their material can receive the
payments due them automatically through a smart contract as sales are made. They do not need
to wait until the end of a period or wonder whether the publisher is being truthful about the number
of sales made during the period.

• Blockchain and smart contracts have an important place in supply chain management, freight, and
logistics, particularly in international transactions. Blockchain supply chain management does not
rely on freight brokers, paper documents, or banks around the world to move goods and payments.
The blockchain can provide secure digital versions of all documents that can be accessed by all the
parties to a transaction. Defined events cause execution of the contract when the requirements

94
Nick Szabo, Smart Contracts: Building Blocks for Digital Markets, 1996, http://www.fon.hum.uva.nl/rob/Courses/
InformationInSpeech/CDROM/Literature/LOTwinterschool2006/szabo.best.vwh.net/smart_contracts_2.html, accessed
April 29, 2019.

298 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.3. Technology-enabled Finance Transformation

are fulfilled. The smart contract can manage the flow of approvals and make the transfers of
currency upon approval. For example, after goods have been received, the payment to the shipper
takes place automatically. The transaction has complete transparency from beginning to end.

• On-demand manufacturing can be performed by machines that are automated and running on a
blockchain network. A design would be sent to a locked machine along with an order for a specific
number of units. The contract would be executed, the machine would be unlocked to produce the
correct number of units, the goods would be produced, and the machine would be locked again.

• An insurance contract can be in the form of a smart contract. For example, an orchard owner is
concerned about a freeze that could destroy the year’s fruit crop. An insurance company offers
insurance against a freeze through a self-executing smart contract. The orchard owner and the
insurance company agree to the contract terms and digitally sign a smart contract that is uploaded
to a blockchain. The orchard owner’s periodic premium payments are automated, and the block-
chain checks a third party source such as the National Weather Service daily for a possible freeze
event. If a freeze event occurs, payment is sent automatically from the insurance company to the
orchard owner.

Note: The third party source for a smart contract is called an oracle.

• Fish are being tracked from their sources to consumers in world markets and restaurants to pre-
vent illegal fishing.

• The title to real property can be transferred using a blockchain. The whole history of the property
owners and all the buy and sell transactions can be maintained in a blockchain dedicated to that
piece of property. The blockchain can provide consensus regarding the current owner of the prop-
erty and the historic record of property owners. The blockchain technology makes changing the
historical records prohibitively difficult and costly. As a result, a title agency—a third party—is not
needed to research the property records each time a piece of property is sold. The owner of the
property can be identified using public key cryptography.

• Various other public records are being maintained by state and local governments on blockchains.

• The provenance of high-value investment items such as art can be tracked on a blockchain to help
protect against forgeries and fraud.

• Blockchains can be used to store data on archaeological artifacts held in museums. If an artifact
or artifacts are stolen, the museum could release data using the hash codes to law enforcement
so they could prevent the export or sale of artifacts matching the descriptions of the stolen items.
Furthermore, if an artifact has a blockchain record, each time it crossed a border or was sold to a
new collector the record would be updated to show it was legitimate. Looted antiquities would have
no record, and faked records could not be manufactured.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 299
F.3. Technology-enabled Finance Transformation CMA Part 1

Benefits of Smart Contracts

• Smart contracts can authenticate counter-party identities, the ownership of assets, and claims of
right by using digital signatures, which are private cryptographic keys held by each party.

• Smart contracts can access outside information or data to trigger actions, for example, commodity
prices, weather data, interest rates, or an occurrence of an event.

• Smart contracts can self-execute. The smart contract will take an action such as transferring a pay-
ment without any action required by the counter-parties. The automatic execution can reduce
counter-party risk and settlement risk.

• The decentralized, distributed ledger on the blockchain prevents modifications not authorized or
agreed to by the parties.

• Smart contracts can enhance market activity and efficiency by facilitating trade execution.

• Use of standardized code and execution may reduce costs of negotiations.

• Automation reduces transaction times and manual processes.

• Smart contracts can perform prompt regulatory reporting.

Limitations and Risks of Smart Contracts

• The operation of a smart contract is only as smart as the information it receives and the computer
code that directs it.

• Existing laws and regulations apply to all contracts equally regardless of what form a contract takes,
so contracts or parts of contracts that are written in code are subject to otherwise applicable law and
regulation. However, if a smart contract unlawfully circumvents rules and protections, to the extent
it violates the law, it is not enforceable. For example, if a U.S. derivative contract is traded on or
processed by a facility that is not appropriately registered with the U.S. Commodity Futures Trading
Commission (CFTC) or is executed by entities required to be registered with the CFTC but which are
not and do not have an exception or exemption from registration, the contract is prohibited.95

• A smart contract could introduce operational, technical, and cybersecurity risk.


Operational risk: For example, smart contracts may not include sufficient backup and failover mech-
anisms in case of operational problems, or the other systems they depend on to fulfill contracts terms
may have vulnerabilities that could prevent the smart contract from functioning as intended.

Technical risk: For example, humans could make a typographical error when coding, Internet service
can go down, user interfaces may become incompatible, the oracle (the third party source used by
the smart contract to authorize payments) may fail or other disruptions can occur with the external
sources used to obtain information on reference prices, events, or other data.

Cybersecurity risk: Smart contract systems may be vulnerable to hacking, causing loss of digital
assets. For example, an attacker may compromise the oracle, causing the smart contract to improp-
erly transfer assets. There may be limited or no recourse if hackers transfer digital assets to
themselves or others.96

• A smart contract may be subject to fraud and manipulation. For example, smart contracts can include
deliberately damaging code that does not behave as promised or that may be manipulated. Oracles
may be subject to manipulation or may themselves be fraudulent and may disperse fraudulent infor-
mation that results in fraudulent outcomes.97

95
A Primer on Smart Contracts, The U.S. Commodity Futures Trading Commission, Nov. 27, 2018, p. 25,
https://www.cftc.gov/sites/default/files/2018-11/LabCFTC_PrimerSmartContracts112718_0.pdf, accessed April 29,
2019.
96
Ibid, pp. 27-29.
97
Ibid., p. 30.

300 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Governance for Smart Contracts


Good governance is important for smart contracts, whch require ongoing attention and may require action
and possible revision. Governance standards and frameworks are needed and appear to be in the early
stages of development.

• Governance standards may assign responsibility for smart contract design and operation and es-
tablish mechanisms for dispute resolution.

• Standards may incorporate terms or conditions that smart contracts need to have in order to be
enforceable.

• Standards could create presumptions regarding the legal character of a smart contract, depending
on its attributes and manner of use.

• Good governance standards may help address the risks that smart contracts present.98

F.4. Data Analytics


Data analytics is the process of gathering and analyzing data in a way that produces meaningful
information that can be used to aid in decision-making. As businesses become more technologically
sophisticated, their capacity to collect data increases. However, the stockpiling of data is meaningless with-
out a method of efficiently collecting, aggregating, analyzing, and utilizing it for the benefit of the company.

Data analytics can be classified into four types: descriptive analytics, diagnostic analytics, predictive
analytics, and prescriptive analytics.

1) Descriptive analytics report past performance. Descriptive analytics are the simplest type of
data analytics and they answer the question, “What happened”?

2) Diagnostic analytics are used with descriptive analytics to answer the question, “Why did it
happen”? The historical data is mined to understand the past performance and to look for the
reasons behind success or failure. For example, sales data might be broken down into segments
such as revenue by region or by product rather than revenue in total.

3) Predictive analytics focus on the future using correlative99 analysis. Predictive analytics answer
the question, “What is likely to happen”? Historical data is combined with other data using rules
and algorithms. Large quantities of data are processed to identify patterns and relationships be-
tween and among known random variables or data sets in order to make predictions about what
is likely to occur in the future. A sales forecast made using past sales trends is a form of predictive
analytics.

4) Prescriptive analytics answer the question “What needs to happen?” by charting the best course
of action based on an objective interpretation of the data. Prescriptive analytics make use of
structured and unstructured data and apply rules to predict what will happen and to prescribe how
to take advantage of the predicted events. For example, prescriptive analytics might generate a
sales forecast and then use that information to determine what additional production lines and
employees are needed to meet the sales forecast. In addition to anticipating what will happen and
determining what needs to happen, prescriptive analytics can help determine why it will happen.
Prescriptive analytics can incorporate new data and re-predict and re-prescribe, as well. Prescrip-
tive analytics is most likely to yield the most impact for an organization, but it is also the most
complex type of analytics.

98
Ibid., p. 31.
99
If two things are correlated with one another, it means there is a close connection between them. It may be that one
of the things causes or influences the other, or it may be that something entirely different is causing or influencing both
of the things that are correlated.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 301
F.4. Data Analytics CMA Part 1

Business Intelligence (BI)


Business intelligence is the combination of architectures, analytical and other tools, databases, applications,
and methodologies that enable interactive access—sometimes in real time—to data such as sales revenue,
costs, income, and product data. Business intelligence provides historical, current, and predicted values for
internal, structured data regarding products and segments. Further, business intelligence gives managers
and analysts the ability to conduct analysis to be used to make more informed strategic decisions and thus
optimize performance.

The business intelligence process involves the transformation of data into information, then to knowledge,
then to insight, then to strategic decisions, and finally to action.

• Data is facts and figures, but data by itself is not information.

• Information is data that has been processed, analyzed, interpreted, organized, and put into con-
text such as in a report, in order to be meaningful and useful.

• Knowledge is the theoretical or practical understanding of something. It is facts, information, and


skills acquired through experience or study. Thus, information becomes knowledge through expe-
rience, study, or both.

• Insight is a deep and clear understanding of a complex situation. Insight can be gained through
perception or intuition, but it can also be gained through use of business intelligence: data analyt-
ics, modeling, and other tools.

• The insights gained from the use of business intelligence lead to recommendations for the best
action to take. Strategic decisions are made by choosing from among the recommendations.

• The strategic decisions made are implemented and turned into action.

A Business Intelligence system has four main components:

1) A data warehouse (DW) containing the source data.

2) Business analytics, that is, the collection of tools used to mine, manipulate, and analyze the data
in the DW. Many Business Intelligence systems include artificial intelligence capabilities, as well as
analytical capabilities.

3) A business performance management component (BPM) to monitor and analyze performance.

4) A user interface, usually in the form of a dashboard.

Note: A dashboard is a screen in a software application, a browser-based application, or a desktop


application that displays in one place information relevant to a given objective or process, or for senior
management, it may show patterns and trends in data across the organization.

For example, a dashboard for a manufacturing process might show productivity information for a period,
variances from standards, and quality information such as the average number of failed inspections per
hour. For senior management, it might present key performance indicators, balanced scorecard data, or
sales performance data, to name just a few possible metrics that might be chosen.

A dashboard for a senior manager may show data on manufacturing processes, sales activity, and current
financial metrics.

A dashboard may be linked to a database that allows the data presented to be constantly updated.

302 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Big Data and the Four “V”s of Big Data


Big Data refers to vast datasets that are too large to be analyzed using standard software tools and so
require new processing technologies. Those new processing technologies are data analytics.

Big Data can be broken down into three categories:

1) Structured data is in an organized format that enables it to be input into a relational database
management system and analyzed. Examples include the data in CRM or ERP systems, such as
transaction data, customer data, financial data, employee data, and vendor data.

2) Unstructured data has no defined format or structure. It is typically free-form and text-heavy,
making in-depth analysis difficult. Examples include word processing documents, email, call center
communications, contracts, audio and video, photos, data from radio-frequency identification
(RFID) tags, and information contained on websites and social media.

3) Semi-structured data has some format or structure but does not follow a defined model. Exam-
ples include XML files, CSV files, and most server log files.

Big Data is characterized by four attributes, known as the four V’s: volume, velocity, variety, and veracity.

1) Volume: Volume refers to the amount of data that exists. The volume of data available is in-
creasing exponentially as people and processes become more connected, creating problems for
accountants. The tools used to analyze data in the past—spreadsheet programs such as Excel and
database software such as Access—are no longer adequate to handle the complex analyses that
are needed. Data analytics is best suited to processing immense amounts of data.

2) Velocity: Velocity refers to the speed at which data is generated and changed, also called its
flow rate. As more devices are connected to the Internet, the velocity of data grows and organi-
zations can be overwhelmed with the speed at which the data arrives. The velocity of data can
make it difficult to discern which data items are useful for a given decision. Data analytics is de-
signed to handle the rapid influx of new data.

3) Variety: Variety refers to the diverse forms of data that organizations create and collect. In the
past, data was created and collected primarily by processing transactions. The information was in
the form of currency, dates, numbers, text, and so forth. It was structured, that is, it was easily
stored in relational databases and flat files. However, today unstructured data such as media files,
scanned documents, Web pages, texts, emails, and sensor data are being captured and collected.
These forms of data are incompatible with traditional relational database management systems
and traditional data analysis tools. Data analytics can capture and process diverse and complex
forms of information.

4) Veracity: Veracity is the accuracy of data, or the extent to which it can be trusted for decision-
making. Data must be objective and relevant to the decision at hand in order to have value for use
in making decisions. However, various distributed processes—such as millions of people signing
up online for services or free downloads—generate data, and the information they input is not
subject to controls or quality checks. If biased, ambiguous, irrelevant, inconsistent, incomplete, or
even deceptive data is used in analysis, poor decisions will result. Controls and governance over
data to be used in decision-making are essential to ensure the data’s accuracy. Poor-quality data
leads to inaccurate analysis and results, commonly referred to as “garbage in, garbage out.”

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 303
F.4. Data Analytics CMA Part 1

Some data experts have added two additional Vs that characterize data:

5) Variability: Data flows can be inconsistent, for example, they can exhibit seasonal peaks. Fur-
thermore, data can be interpreted in varying ways. Different questions require different
interpretations.

6) Value: Value is the benefit that the organization receives from data. Without the necessary data
analytics processes and tools, the information is more likely to overwhelm an organization than to
help the organization. The organization must be able to determine the relative importance of dif-
ferent data to the decision-making process. Furthermore, an investment in Big Data and data
analytics should provide benefits that are measurable.

Data Science
Data science is a field of study and analysis that uses algorithms and processes to extract hidden
knowledge and insights from data. The objective of data science is to use both structured and unstructured
data to extract information that can be used to develop knowledge and insights for forecasting and strategic
decision making.

The difference between data analytics and data science is in their goals.

• The goal of data analytics is to provide information about issues that the analyst or manager either
knows or knows he or she does not know (that is, “known unknowns”).

• On the other hand, the goal of data science is to provide actionable insights into issues where
the analyst or manager does not know what he or she does not know (that is, “unknown un-
knowns”).

Example: Data science would be used to try to identify a future technology that does not exist today
but that will impact the organization in the future.

Decision science, machine learning (that is, the use of algorithms that learn from the data in order to
predict outcomes), and prescriptive analytics are three examples of means by which actionable insights
can be discovered in a situation where “unknowns are unknown.”

Data science involves data mining, analysis of Big Data, data extraction, and data retrieval. Data science
draws on knowledge of data engineering, social engineering, data storage, natural language processing,
and many other fields.

The size, value, and importance of Big Data has brought about the development of the profession of data
scientist. Data science is a multi-disciplinary field that unifies several specialized areas, including statistics,
data analysis, machine learning, math, programming, business, and information technology. A data scien-
tist is a person with skills in all the areas, though most data scientists have deep skills in one area and less
deep skills in the other areas.

Note: “Data mining” involves using algorithms in complex data sets to find patterns in the data that can
be used to extract usable data from the data set. Data mining is discussed in more detail below.

304 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Data and Data Science as Assets


Data and data science capabilities are strategic assets to an organization, but they are complementary
assets.

• Data science is of little use without usable data.

• Good data cannot be useful in decision-making without good data science talent.

Good data and good data science, used together, can lead to large productivity gains for a company and
the ability to do things it has never done before. Data and data science together can provide the following
opportunities and benefits to an organization:

• They can enable the organization to make decisions based on data and evidence.

• The organization can leverage relevant information from various data sources in a timely manner.

• When the cloud is used, the organization can get the answers it needs using any device, any time.

• The organization can transform data into actionable insights.

• The organization can discover new opportunities.

• The organization can increase its competitive advantage.

• Management can explore data to get answers to questions.

The result can be maximized revenue, improved operations, and mitigated risks. The return on investment
from the better decision-making that results from using data and data science can be significant.

As with any strategic asset, it is necessary to make investments in data and data science. The investments
include building a modern business intelligence architecture using the right tools, investing in people with
data science skills, and investing in the training needed to enable the staff to use the business intelligence
and data analytics tools.

Challenges of Managing Data Analytics


Some general challenges of managing data analytics include data capture, data curation (that is, the or-
ganization and integration of disparate data collected from various sources), data storage, security and
privacy protection, data search, data sharing, data transfer, data analysis, and data visualization.

In addition, some specific challenges of managing data analytics include:

• The growth of data and especially of unstructured data.

• The need to generate insights in a timely manner in order for the data to be useful.

• Recruiting and retaining Big Data talent. Demand has increased for data engineers, data scientists,
and business intelligence analysts, causing higher salaries and creating difficulty filling positions.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 305
F.4. Data Analytics CMA Part 1

Data Mining
Data mining is the use of statistical techniques to search large data sets to extract and analyze data in
order to discover previously unknown, useful patterns, trends, and relationships within the data that go
beyond simple analysis and that can be used to make decisions. Data mining uses specialized computational
methods derived from the fields of statistics, machine learning, and artificial intelligence.

Data mining involves trying different hypotheses repeatedly and making inferences from the results that
can be applied to new data. Data mining is thus an iterative process. Iteration is the repetition of a
process in order to generate a sequence of outcomes. Each repetition of the process is a single iteration,
and the outcome of each iteration is the starting point of the next iteration. 100

Data mining is a process with defined steps, and thus it is a science. Science is the pursuit and application
of knowledge and understanding of the natural and social world following a systematic methodology based
on evidence.101

Data mining is also an art. In data mining, decisions must be made regarding what data to use, what tools
to use, and what algorithms to use. For example, one word can have many different meanings. In mining
text, the context of words must be considered. Therefore, instead of just looking for words in relation to
other words, the data scientist looks for whole phrases in relation to other phrases. The data scientist must
make thoughtful choices in order to get usable results.

Data mining differs from statistics. Statistics focuses on explaining or quantifying the average effect of an
input or inputs on an outcome, such as determining the average demand for a product based on some
variable like price or advertising expenditures. Statistical analysis includes determining whether the rela-
tionships observed could be a result of the variable or could be a matter of chance instead. A simple example
of statistical analysis is a linear regression model102 that relates total historical sales revenues (the depend-
ent variable) to various levels of historical advertising expenditures (the independent variable) to discover
whether the level of advertising expenditures affects total sales revenues. Statistics may involve using a
sample from a dataset to make predictions about the population as a whole. Alternatively, it may involve
using the entire dataset to estimate the best-fit model in order to maximize the information available about
the hypothesized relationship in the population and predict future results.

In contrast, data mining involves open-ended exploring and searching within a large dataset without putting
limits around the question being addressed. The goal is to predict outcomes for new individual records. The
data is usually divided into a training set and a validation set. The training set is used to estimate the
model, and the validation set is used to assess the model’s predictive performance on new data.

Data mining might be used to classify potential customers into different groups to receive different mar-
keting approaches based on some characteristic common to each group that is yet to be discovered. It may
be used to answer questions such as what specific online advertisement should be presented to a particular
person browsing on the Internet based on their previous browsing habits and the fact that other people
who browsed the same topics purchased a particular item.

Thus, data mining involves generalization of patterns from a data set. “Generalization” is the ability to
predict or assign a label to a “new” observation based on a model built from past experience. In other
words, the generalizations developed in data mining should be valid not just for the data set used in ob-
serving the pattern but should also be valid for new, unknown data.

Software used for data mining uses statistical models, but it also incorporates algorithms that can “learn”
from the patterns in the data. An algorithm is applied to the historical data to create a mining model, and
then the model is applied to new data to create predictions and make inferences about relationships. For

100
Definition of “iteration” from Wikipedia, https://en.wikipedia.org/wiki/Iteration, accessed May 8, 2019.
101
Definition of “science” from the Science Council, https://sciencecouncil.org/about-science/our-definition-of-science/,
accessed May 8, 2019.
102
In statistics, a “model” is the representation of a relationship between variables in the data. It describes how one or
more variables in the data are related to other variables. “Modeling” is building a representative abstraction from the
observed data set.

306 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

example, data mining software can help find customers with common interests and determine which prod-
ucts customers with each particular interest typically purchase, in order to direct advertising messages
about specific products to the customers who are most likely to purchase those products.

Data mining is used in predictive analytics. Basic concepts of predictive analytics include:

• Classification – Any data analysis involves classification, such as whether a customer will pur-
chase or not purchase. Data mining is used when the classification of the data is not known. Similar
data where the classification is known is used to develop rules, and then those rules are applied
to the data with the unknown classification to predict what the classification is or will be. For
example, customers are classified as predicted purchasers or predicted non-purchasers.

• Prediction – Prediction is similar to classification, but the goal is to predict the numerical value of
a variable such as the amount of a purchase rather than (for example) simply classifying custom-
ers as predicted purchasers or predicted non-purchasers. Although classification also involves
prediction, “prediction” in this context refers to prediction of a numerical value, which can be an
integer (a whole number such as 1, 2, or 3) or a continuous variable.103

• Association rules – Also called affinity analysis, association rules are used to find patterns of
association between items in large databases, such as associations among items purchased from
a retail store, or “what goes with what.” For example, when customers purchase a 3-ring notebook,
do they usually also purchase a package of 3-hole punched paper? If so, the 3-hole punched paper
can be placed on the store shelf next to the 3-ring notebooks. Similar rules can be used for bundling
products.104

• Online recommendation systems – In contrast to association rules, which generate rules that
apply to an entire population, online recommendation systems use collaborative filtering to
deliver personalized recommendations to users. Collaborative filtering generates rules for “what
goes with what” at the individual user level. It makes recommendations to individuals based on
their historical purchases, online browsing history, or other measurable behaviors that indicate
their preferences, as well as other users’ historical purchases, browsing, or other behaviors.

• Data reduction – Data reduction is the process of consolidating a large number of records into a
smaller set by grouping the records into homogeneous groups.

• Clustering – Clustering is discovering groups in data sets that have similar characteristics without
using known structures in the data. Clustering can be used in data reduction to reduce the number
of groups to be included in the data mining algorithm.

• Dimension reduction – Dimension reduction entails reducing the number of variables in the data
before using it for data mining, in order to improve its manageability, interpretability, and predic-
tive ability.

• Data exploration – Data exploration is used to understand the data and detect unusual values.
The analyst explores the data by looking at each variable individually and looking at relationships
between and among the variables in order to discover patterns in the data. Data exploration can
include creating charts and dashboards, called data visualization or visual analytics (see next item).
Data exploration can lead to the generation of a hypothesis.

• Data visualization – Data visualization is another type of data exploration. Visualization, or visual
discovery, consists of creating graphics such as histograms and boxplots for numerical data in

103
A continuous variable is a numerical variable that can take on any value at all. It does not need to be an integer such
as 1, 2, 3, or 4, though it can be an integer. A continuous variable can be 8, 8.456, 10.62, 12.3179, or any other number,
and the variable can have any amount of decimal points.
104
Product bundling occurs when a seller bundles products, features, or services together and offers the bundle at a
price that is lower than the price of the items if purchased individually. For example, a software vendor may create a
suite of software applications and offer the applications together at a reduced price.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 307
F.4. Data Analytics CMA Part 1

order to visualize the distribution of the variables and to detect outliers.105 Pairs of numerical
variables can be plotted on a scatter plot graph in order to discover possible relationships. When
the variables are categorized, bar charts can be used. Visualization is covered in more detail later
in this section.

Supervised and Unsupervised Learning in Data Mining


Supervised learning algorithms are used in classification and prediction. In order to “train” the algo-
rithm, it is necessary to have a dataset in which the value of the outcome to be predicted is already known,
such as whether or not the customer made a purchase. The dataset with the known outcome is called the
training data because that dataset is used to “train” the algorithm. The data in the dataset is called
labeled data because it contains the outcome value (called the label) for each record. The classification
or prediction algorithm “learns” or is “trained” about the relationship between the predictor variables and
the outcome variable in the training data. After the algorithm has “learned” from the training data, it is
tested by applying it to another sample of labeled data for which the outcome is already known but is
initially hidden (called the validation data) to see if it works properly. If several different algorithms are
being tested, additional test data with known outcomes should be used with the selected algorithm to
predict how well it will work. After the algorithm has been thoroughly tested, it can be used to classify or
make predictions in data where the outcome is unknown.

Example: Simple linear regression is an example of a basic supervised learning algorithm. The x
variable, the independent variable, serves as the predictor variable. The y variable, the dependent var-
iable, is the outcome variable in the training and test data where the y value for each x value is known.
The regression line is drawn so that it minimizes the sum of the squared deviations between the actual
y values and the values predicted by the regression line. Then, the regression line is used to predict the
y values that will result for new values of x for which the y values are unknown.106

Unsupervised learning algorithms are used when there is no outcome variable to predict or classify.
Association rules, dimension reduction, and clustering are unsupervised learning methods.

Neural Networks in Data Mining


Neural networks are systems that can recognize patterns in data and use the patterns to make predictions
using new data. Neural networks derive their knowledge from their own data by sifting through the data
and recognizing patterns. Neural networks are used to learn about the relationships in the data and combine
predictor information in such a way as to capture the complicated relationships among predictor variables
and between the predictor variables and the outcome variable.

Neural networks are based on the human brain and mimic the way humans learn. In a human brain, neurons
are interconnected and humans can learn from experience. Similarly, a neural network can learn from its
mistakes by finding out the results of its predictions. In the same way as a human brain uses a network of
neurons to respond to stimuli from sensory inputs, a neural network uses a network of artificial neurons,
called nodes, to simulate the brain’s approach to problem solving. A neural network solves learning prob-
lems by modeling the relationship between a set of input signals and an output signal.

The results of the neural network’s predictions—the output of the model—becomes the input to the next
iteration of the model. Thus, if a prediction made did not produce the expected results, the neural network
uses that information in making future predictions.

105
Outliers are data entries that do not fit into the model because they are extreme observations.
106
Regression analysis is covered in more detail later in this section.

308 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Neural networks can look for trends in historical data and use it to make predictions. Some examples of
uses of neural networks include the following.

• Picking stocks for investment by performing technical analysis of financial markets and individual
investment holdings.

• Making bankruptcy predictions. A neural network can be given data on firms that have gone bank-
rupt and firms that have not gone bankrupt. The neural network will use that information to learn
to recognize early warning signs of impending bankruptcy, and it can thus predict whether a par-
ticular firm will go bankrupt.

• Detecting fraud in credit card and other monetary transactions by recognizing that a given trans-
action is outside the ordinary pattern of behavior for that customer.

• Identifying a digital image as, for example, a cat or a dog.

• Self-driving vehicles use neural networks with cameras on the vehicle as the inputs.

The structure of neural networks enables them to capture complex relationships between predictors and an
outcome by fitting the model to the data. It calculates weights for the individual input variables. The weights
allow each of the inputs to contribute a greater or lesser amount to the output, which is the sum of the
inputs. Depending on the effect of those weights on how well the output of the model—the prediction it
makes—fits the actual output, the neural network then revises the weights for the next iteration.

Goodness of fit refers to how closely the predicted values resulting from a model match the actual, ob-
served values.

One of the weaknesses of neural networks is that they can “overfit” the data. Overfitting the data means
the model fits the training data perfectly, but the model does not generalize well and thus does not do a
good job of making predictions using new data. Overfitting occurs because of “noise” in the training data.
Noise is outliers (extreme observations) and randomness.

Overfitting can be detected by examining the performance of the model on the validation dataset. The
errors in the model’s performance on the validation dataset decrease in the early iterations of the training,
but after a while, the errors begin to increase. Therefore, it is important to limit the number of training
iterations to the number where the validation errors are minimized. The weights for the variables used at
that stage are probably the best model to use with new data.

Underfitting can occur as well. Underfitting happens when the model is too simple, because if the model
is too simple, it will not be flexible enough in learning from the data.

Neural Networks and Sensitivity Analysis


Neural networks are called black boxes. They work best when solving problems where the input data and
the output data are fairly simple but the process that relates the input to the output is extremely complex.
Even though neural networks can approximate a function that relates the input to the output, studying the
structure of the network does not produce any insights into the function being approximated. In other
words, there is no simple link between the weights used for the relative importance or frequency of the
input variables and the resulting output.

However, in some cases it is possible to learn something about the relationships being captured by a neural
network by conducting sensitivity analysis on the validation dataset. Sensitivity analysis estimates the
amount of change in the output of a model that is caused by a single change in the model inputs. It can be
used to determine which input parameters are more important for achieving accurate output values.

First, all the predictor values are set to their means107 and the network’s prediction is obtained. Then the
process is repeated over and over with each predictor set sequentially to its minimum and then to its

107
A mean is an average.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 309
F.4. Data Analytics CMA Part 1

maximum value. By comparing the predictions created by the different levels of the predictors, the analyst
can gain a sense of which predictors affect the predictions the most and in what way. For example:

• Which model parameters contribute the most to output variability?

• Which parameters are insignificant and can be eliminated from the model?

• Which parameters interact with each other?

Sensitivity analysis is also used to understand the behavior of the system being modeled, to determine
whether the model is doing what it was developed to do, to evaluate whether the model is applicable to the
data, and to determine the model’s stability.

Challenges of Data Mining


Some of the challenges inherent in data mining include the following.

• Poor data quality. Data stored in relational databases may be incomplete, out of date, or incon-
sistent. For example, mailing lists can contain duplicate records, leading to duplicate mailings and
excess costs. Poor quality data can lead to poor decisions.

Furthermore, use of inaccurate data can cause problems for consumers. For example, when credit
rating agencies have errors in their data, consumers can have difficulty obtaining credit.

• Information exists in multiple locations within the organization and thus is not centrally located,
for example Excel spreadsheets that are in the possession of individuals in the organization. Infor-
mation that is not accessible cannot be used.

• Biases are amplified in evaluating data. The meaning of a data analysis must be assessed by a
human being, and human beings have biases. A “bias” is a preference or an inclination that gets
in the way of impartial judgment. Most people tend to trust data that supports their pre-existing
positions and tend not to trust data that does not support their pre-existing positions. Other biases
include relying on the most recent data or trusting only data from a trusted source. All such biases
contribute to the potential for errors in data analysis.

• Analyzed data often displays correlations.108 However, correlation does not prove causation.
Establishing a causal relationship is necessary before using correlated data in decision-making. If
a causal relationship is assumed where none exists, decisions made on the basis of the data will
be flawed.

• Ethical issues such as data privacy related to the aggregation of personal information on millions
of people. Profiling according to ethnicity, age, education level, income, and other characteristics
results from the collection of so much personal information.

• Data security is an issue because personal information on individuals is frequently stolen by


hackers or even employees.

• A growing volume of unstructured data. Data items that are unstructured do not conform to
relational database management systems, making capturing and analyzing unstructured data more
complex. Unstructured data includes items such as social media posts, videos, emails, chat logs,
and images, for example images of invoices or checks received.

108
A “correlation” is a relationship between or among values in multiple sets of data where the values in one data set
move in relation to the values in one or more other data set or sets.

310 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

NoSQL, Data Lakes, and the Challenge of Unstructured Data


Unstructured data cannot be stored in a data warehouse because the data in a data warehouse must be
uniformly defined so that users can write queries using Structured Query Language to extract information.
Thus, the data in a data warehouse must be transformed, that is, put into a proper format before being
added to the data warehouse. This transformation before data can be loaded to a data warehouse is known
as Schema-on-Write.109

However, unstructured items of data do not conform to any relational database schema, so they cannot be
transformed to be used in a data warehouse. Therefore, a data lake can be used for data storage and
analysis when unstructured data must be included. Because unstructured data items do not conform to
relational database management systems that use Structured Query Language (SQL), a data lake utilizes
a non-relational database management system, called NoSQL.

NoSQL stands for “Not only SQL.” A NoSQL database management system can be used to analyze high
volume and disparate data, including unstructured and unpredictable data types. Unlike relational data-
bases, NoSQL databases do not require SQL to analyze the data and most do not use a schema and thus
they are more flexible.110 A NoSQL database management system can be used with a data lake that contains
both structured and unstructured data.

Note: SQL can be used as a query language with a NoSQL database management system, but SQL is
not the main query language used because its usage is limited to structured data.

Items of structured and unstructured data from all sources are stored in a data lake as “raw data,” that is,
untransformed data items that have not been uniformly defined. The data is transformed and a schema
applied only when specific analysis is needed. Such data transformation is called Schema-on-Read.

The primary users of a data lake and a NoSQL database management system are data scientists who have
the analytical and programming skills to do deep analysis of both structured and unstructured data.

109
See the topic Data Warehouse, Data Mart, and Data Lake in Section F.1. – Information Systems in this volume for
more information about data warehouses.
110
A database’s schema is a map or plan of the entire database, that is, the database’s logical structure. The schema
specifies the names of the data elements contained in the database and their relationships to the other data elements.
For more information about relational databases, please see the topic Databases in this volume in Section F.1. – Infor-
mation Systems in this volume.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 311
F.4. Data Analytics CMA Part 1

What Data Mining is NOT


Data mining is not simple querying of a database using Structured Query Language. SQL searches of
databases are descriptive analytics and are based on rules developed by the user, such as “find all credit
card holders in a given ZIP code who charge more than $25,000 annually and who have had no delinquent
payments in the last two years.” The database query displays information from the database, but the
information displayed is not hidden information. Data mining, on the other hand, involves statistical mod-
eling and automated algorithmic methods of discovering previously hidden, unknown relationships.

Data mining is also not “discovering” historical patterns in data that are random relationships (rather than
real relationships) and assuming they will repeat. The results of analysis must be assessed with intelligence
to determine whether the relationships are meaningful and can be generalized to new data.

Steps in Data Mining


A typical data mining project will include the following steps.

1) Understand the purpose of the project. The data scientist needs to understand the user’s
needs and what the user will do with the results. Also, the data scientist needs to know whether
the project will be a one-time effort or ongoing.
2) Select the dataset to be used. The data scientist will take samples from a large database or
databases, or from other sources. The samples should reflect the characteristics of the records of
interest so the data mining results can be generalized to records outside of the sample. The data
may be internal or external.
3) Explore, clean, and preprocess the data. Verify that the data is in usable condition, that is,
whether the values are in a reasonable range and whether there are obvious outliers. Determine
how missing data (that is, blank fields) should be handled. Visualize the data by reviewing the
information in chart form. If using structured data, ensure consistency in the definitions of fields,
units of measurement, time periods covered, and so forth. New variables may be created in this
step, for example using the start and end dates to calculate the duration of a time period.
4) Reduce the data dimension if needed. Eliminate unneeded variables, transform variables as
necessary, and create new variables. The data scientist should be sure to understand what each
variable means and whether it makes sense to include it in the model.
5) Determine the data mining task. Determining the task includes classification, prediction, clus-
tering, and other activities. Translate the general question or problem from Step 1 into the specific
data mining question.
6) Partition the data. If supervised learning will be used (classification or prediction), partition the
dataset randomly into three parts: one part for training, one for validation, and one for testing.
7) Select the data mining techniques to use. Techniques include regression, neural networks,
hierarchical clustering, and so forth.
8) Use algorithms to perform the task. The use of algorithms is an iterative process. The data
scientist tries multiple algorithms, often using multiple variants of the same algorithm by choosing
different variables or settings. The data scientist uses feedback from an algorithm’s performance
on validation data to refine the settings.
9) Interpret the results of the algorithm. The data scientist chooses the best algorithm and tests
the final choice on the test data to learn how well it will perform.
10) Deploy the model. The model is run on the actual records to produce actionable information that
can be used in decisions. The chosen model is used to predict the outcome value for each new
record, called scoring.111
A data mining project does not end when a particular solution is deployed, however. The results of the data
mining may raise new questions that can then be used to develop a more focused model.

111
Shmueli, Galit, Bruce, Peter C., Yahav, Inbal, Patel, Nitin R., and Lichtendahl Jr., Kenneth C., Data Mining for Business
Analytics: Concepts, Techniques, and Applications in R, 1st Edition, John Wiley & Sons, Hoboken, NJ, 2018, pp. 19-21.

312 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Analytic Tools

Linear Regression Analysis


Regression analysis measures the extent to which an effect has historically been the result of a specific
cause. If the relationship between the cause and the effect is sufficiently strong, regression analysis using
historical data can be used to make decisions and predictions.

Time Series Analysis

Note: Time series analysis was introduced in Section B in Volume 1 of this textbook, topic B.3. Fore-
casting Techniques. Candidates may wish to review that information before proceeding. The trend
pattern in time series analysis was introduced in Forecasting Techniques and will be further explained in
this topic. Additional patterns will be discussed in this topic, as well.

A time series is a sequence of measurements taken at equally-spaced, ordered points in time. A time series
looks at relationships between a variable and the passage of time. The variable may be sales revenue for
a segment of the organization, production volume for a plant, expenses in one expense classification, or
anything being monitored over time. Only one set of historical time series data is used in time series analysis
and that set of historical data is not compared to any other set of data.

A time series can be descriptive or predictive. Time series analysis is used for descriptive modeling, in
which a time series is modeled to determine its components, that is, whether it demonstrates a trend
pattern, a seasonal pattern, a cyclical pattern, or an irregular pattern. The information gained from a time
series analysis can be used for decision-making and policy determination.

Time series forecasting, on the other hand, is predictive. It involves using the information from a time
series to forecast future values of that series.

A time series may have one or more of four patterns (also called components) that influence its behavior
over time:

1) Trend

2) Cyclical

3) Seasonal

4) Irregular

Trend Pattern in Time Series Analysis


A trend pattern is the most frequent time series pattern and the one most amenable to use for predicting
because the historical data exhibits a gradual shifting to a higher or lower level. If a long-term trend exists,
short-term fluctuations may take place within that trend; however, the long-term trend will be apparent.
For example, sales from year to year may fluctuate but overall, they may be trending upward, as is the
case in the graph that follows.

A trend projection is performed with simple linear regression analysis, which forecasts values using
historical information from all available past observations of the value. The regression line is called a trend
line when the regression is being performed on a time series.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 313
F.4. Data Analytics CMA Part 1

Note: The x-axis on the graph of a time series is always the horizontal axis and the y-axis is always the
vertical axis. The x-axis represents the independent variable, also known as the predictor variable, and
the y-axis represents the dependent variable, also known as the outcome variable.

In a time series regression analysis, the passage of time is the independent variable and is on the x-
axis.

The equation of a simple linear regression line is:

ŷ = a + bx

Where:

ŷ= the predicted value of ŷ on the regression line corre-


sponding to each value of x, the dependent variable.
a= the y-intercept, or the value of ŷ on the regression
line when x is zero, also called the constant coeffi-
cient.
b= the slope of the line and the amount by which the ŷ
value of the regression line changes (increases or de-
creases) when the value of x increases by one unit, also
called the variable coefficient.
x= the independent variable, the value of x on the x-
axis that corresponds to the predicted value of ŷ on the
regression line.

Example of a Trend Pattern in Time Series Analysis


Sales for each year, 20X0 through 20X9, are as follows:

Year Sales

20X0 $2,250,000

20X1 $2,550,000

20X2 $2,300,000

20X3 $2,505,000

20X4 $2,750,000

20X5 $2,800,000

20X6 $2,600,000

20X7 $2,950,000

20X8 $3,000,000

20X9 $3,200,000

314 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics
efham CMA
The following chart illustrates the trend pattern of the sales. It indicates a strong relationship between the
passage of time (the x variable) and sales (the y variable) because the historical data points fall close to
the regression line.

Trend Pattern
Sales 20X0 - 20X9
$4,000,000

$3,500,000

$3,000,000

$2,500,000

$2,000,000

Sales
$1,500,000
Regression Line: ŷ = 2,273,636 + 92,636x

$1,000,000

Years
20X0 is x = 0, 20X1 is x = 1, and so forth

The regression equation, ŷ = 2,273,636 + 92,636x, means the regression line begins at 2,273,636 and
increases by 92,636 each succeeding year.

On the chart, 20X0 is at x = 0, 20X1 is at x = 1, and so forth.

The symbol over the “y” in the formula is called a “hat,” and it is read as “y-hat.” The y-hat indicates
the predicted value of y, not the actual value of y. The predicted value of y is the value of y on the
regression line (the line created from the historical data) at any given value of x.

Thus, in 20X4, where x = 4, the predicted value of y, that is, ŷ, is

ŷ = 2,273,636 + (92,636 × 4)
ŷ = 2,273,636 + 370,544
ŷ = 2,644,180
In 20X7, where x = 7, the predicted value of y is

ŷ = 2,273,636 + (92,636 × 7)
ŷ = 2,273,636 + 648,452
ŷ = 2,922,088
Those values for ŷ (the value on the regression line) for 20X4 and 20X7 can be confirmed by looking at the
chart.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 315
F.4. Data Analytics CMA Part 1

Note: It would be a good idea to calculate the predicted value of y, that is, ŷ, at various other values of
x and confirm the values on the graph. The predicted value of y at a given value of x may be
required in an exam question.

The actual equation of the regression line as shown above may not be given in an exam question.
The y-intercept and the slope of the regression line may be given instead. The constant coefficient,
2,273,636 in the above equation, is the y-intercept, and the variable coefficient, 92,636 in the above
equation, is the slope of the line.

Thus, candidates need to know that the y-intercept of a regression line is the constant coefficient (the
number that stands by itself) and the slope of the regression line is the variable coefficient (the number
next to the x in the equation).

Furthermore, an exam question may use different letters to represent the variables and the coefficients,
so candidates should be able to recognize the form of the equation and the meaning of each component
of the formula based on its usage in the formula.

Trends in a time series analysis are not always upward and linear like the preceding graph. Time series
data can exhibit an upward linear trend, a downward linear trend, a nonlinear (that is, curved) trend, or no
trend at all. A scattering of points that have no relationship to one another would represent no trend at all.

Note: The CMA exam tests linear regression only.

Cyclical Pattern in Time Series Analysis


Any recurring fluctuation that lasts longer than one year is attributable to the cyclical component of the
time series. A cyclical component in sales data is usually due to the cyclical nature of the economy.

A long-term trend can be established even if the sequential data fluctuates greatly from year to year due
to cyclical factors.

Example of a Cyclical Pattern in Time Series Analysis


Sales for each year, 20X0 through 20X9, are as follows:

Year Sales

20X0 $1,975,000

20X1 $2,650,000

20X2 $2,250,000

20X3 $2,450,000

20X4 $2,250,000

20X5 $2,250,000

20X6 $2,600,000

20X7 $2,450,000

20X8 $3,000,000

20X9 $2,750,000

The following chart illustrates the cyclical pattern of the sales. The fluctuations from year to year are greater
than they were for the chart containing the trend pattern. However, a long-term trend is still apparent.

316 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Cyclical Pattern
Sales 20X0 - 20X9
$4,000,000

$3,500,000

$3,000,000
Sales

$2,500,000

$2,000,000

Sales
$1,500,000
Regression Line: ŷ = 2,139,091 + 65,758x

$1,000,000

Years
20X0 is x = 0, 20X1 is x = 1, and so forth

Seasonal Pattern in Time Series Analysis


Usually, trend and cyclical components of a time series are tracked as annual historical movements over
several years. However, a time series can fluctuate within a year due to seasonality in the business. For
example, a surfboard manufacturer’s sales would be highest during the warm summer months, whereas a
manufacturer of snow skis would experience its peak sales in the wintertime. Variability in a time series
due to seasonal influences is called the seasonal component.

Note: Seasonal behavior can take place within any time period. Seasonal behavior is not limited to
periods of a year. A business that is busiest at the same time every day is said to have a within-the-
day seasonal component. Any pattern that repeats regularly is a seasonal component.

Seasonality in a time series is identified by regularly spaced peaks and troughs with a consistent direction
that are of approximately the same magnitude each time, relative to any trend. The graph that follows
shows a strongly seasonal pattern. Sales are low during the first quarter each year. Sales begin to increase
each year in the second quarter and they reach their peak in the third quarter, then they drop off and are
low during the fourth quarter. However, the overall trend is upward, as illustrated by the trend line.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 317
F.4. Data Analytics CMA Part 1

Example of a Seasonal Pattern in Time Series Analysis


Sales for each quarter, March 20X6 through December 20X8, are as follows:

Year Sales

Mar. 20X6 $1,200,000

Jun. 20X6 $2,500,000

Sep. 20X6 $3,200,000

Dec. 20X6 $1,500,000

Mar. 20X7 $1,400,000

Jun. 20X7 $2,800,000

Sep. 20X7 $3,800,000

Dec. 20X7 $1,400,000

Mar. 20X8 $1,700,000

Jun. 20X8 $2,500,000

Sep. 20X8 $3,900,000

Dec. 20X8 $1,600,000

The chart that follows contains historical sales by quarter for three years and forecasted sales by quarter
for the fourth year. The fourth year quarterly forecasts were calculated in Excel using the FORECAST.ETS
function, which is an exponential smoothing algorithm.

Note: Exponential smoothing is outside the scope of the CMA exams, so it is not covered any further in
these study materials.

The chart illustrates that sales volume begins to build in the second quarter of each year. The sales volume
reaches its peak in the third quarter and is at its lowest in the fourth quarter of each year.

318 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Seasonal Pattern
Historical Sales By Quarter, 20X6-20X8
with 20X9 Quarterly Forecasts

Sales Forecast Trend Line


$4,000,000
Sales

$3,000,000

$2,000,000

$1,000,000

Quarters

Irregular Pattern in a Time Series


A time series may vary randomly, not repeating itself in any regular pattern. Such a pattern is called an
irregular pattern. It is caused by short-term, non-recurring factors and its impact on the time series
cannot be predicted.

Example of an Irregular Pattern in Time Series Analysis


Sales for each year, 20X0 through 20X9, are as follows:

Year Sales

20X0 $1,500,000

20X1 $3,200,000

20X2 $2,100,000

20X3 $2,500,000

20X4 $1,400,000

20X5 $1,600,000

20X6 $3,600,000

20X7 $2,000,000

20X8 $2,500,000

20X9 $1,700,000

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 319
F.4. Data Analytics CMA Part 1

The following chart exhibits the irregular pattern of the sales:

Irregular Pattern
Sales 20X0 - 20X9
$4,000,000

$3,500,000

$3,000,000
Sales

$2,500,000

$2,000,000

$1,500,000 Sales

$1,000,000

Years

Time Series and Regression Analysis


A time series that has a long-term upward or downward trend can be used to make a forecast. Simple
linear regression analysis is used to create a trend projection and to forecast values using historical infor-
mation from all available past observations of the value.

Note: Simple regression analysis is called “simple” to differentiate it from multiple regression analysis.
The difference between simple linear regression and multiple linear regression is in the number of inde-
pendent variables.

• A simple linear regression has only one independent variable. In a time series, that independent
variable is the passage of time.

• A multiple linear regression has more than one independent variable.

Linear regression means the regression equation graphs as a straight line.

Simple linear regression analysis relies on two assumptions:

• Variations in the dependent variable (the value being predicted) are explained by variations in
one single independent variable (the passage of time, for a time series).

• The relationship between the independent variable and the dependent variable (whatever is being
predicted) is linear. A linear relationship is one in which the relationship between the independent
variable and the dependent variable can be approximated by a straight line on a graph. The re-
gression equation, which approximates the relationship, will graph as a straight line.

The equation of a simple linear regression line is:

ŷ = a + bx

320 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Where:

ŷ= the predicted value of the dependent variable, ŷ, on the


regression line corresponding to each value of x.
a= the constant coefficient, or the y-intercept, the value
of ŷ on the regression line when x is zero.
b= the variable coefficient and the slope of the regres-
sion line, which is the amount by which the ŷ value of the
regression line changes (either increases or decreases)
when the value of x increases by one unit.
x= the independent variable, or the value of x on the x-
axis that corresponds to the predicted value of ŷ on the
regression line.

Note: The equation of a simple linear regression line graphs as a straight line because none of the
variables in the equation are squared or cubed or have any other exponents. If an equation contains any
exponents, the graph of the equation will be a curved line.

The line of best fit as determined by simple linear regression is a formalization of the way one would fit a
trend line through the graphed data just by looking at it. To fit a line by looking at it, one would use a ruler
or some other straight edge and move it up and down, changing the angle, until it appears the differences
between the points and the line drawn with the straight edge have been minimized. The line that results
will be a straight line located at the position where approximately the same number of points are above the
line as are below it and the distance between each point and the line has been minimized (that is, the
distance is as small as possible).

Linear regression is used to calculate the location of the regression line mathematically. Linear
regression analysis is performed on a computer or a financial calculator, using the observed values of x and
y.

On a graph, the difference between each actual, observed point and its corresponding point on the calcu-
lated regression line is called a deviation. When the position of the regression line is calculated
mathematically, the line will be in the position where the deviations between each graphed value and
the regression line have been minimized. The resulting regression line is the line of best fit. That line
can then be used to predict the value of y for any given value of x.

Note: The statistical method used to perform simple regression analysis is called the Least Squares,
also known as the Ordinary Least Squares method or OLS. The regression line is called the least squares
regression line.

Simple linear regression was used to calculate the regression line and the forecast on the graph presented
earlier as an example of a trend pattern. The regression line was extended out for one additional year to
create a forecast for that year.

Before Developing a Prediction Using Regression Analysis


Before using regression analysis to predict a value, determine whether regression analysis even can be
used to make a prediction.

1) The dependent variable, y, must have a linear relationship with the independent varia-
ble, x.

To determine whether a linear relationship exists, make a scatter plot of the actual historical
values in the time series and review the results. Plotting the x and y coordinates on a scatter plot
will indicate whether or not there is a linear relationship between them.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 321
F.4. Data Analytics CMA Part 1

Note: The x-axis on a scatter plot is on the horizontal and the y-axis is on the vertical, and
each observation is plotted at the intersection of its x-value and its y-value.

If the long-term trend appears to be linear, simple linear regression analysis may be able to
be used (subject to correlation analysis as described below) to determine the location of the linear
regression line, and that linear regression line can be used to make a prediction.

Below is a scatter plot that exhibits no correlation between the x-variable, time, and the y-
variable, sales. The chart below is also an example of the irregular pattern described previously.

Scatter Plot
Sales Year 1-Year 10
$4,000,000

$3,000,000
Sales

$2,000,000

$1,000,000
0 1 2 3 4 5 6 7 8 9 10 11
Years

Historical sales like the above indicate that regression analysis using a time series would not be a
good way to make a prediction.

On the other hand, the following scatter plot does display a linear relationship between the x-
values and y-values, and the use of regression analysis to make a prediction could be helpful.

Scatter Plot
Sales Year 1-Year 10
$4,000,000

$3,000,000
Sales

$2,000,000

$1,000,000
0 1 2 3 4 5 6 7 8 9 10 11
Years

322 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

2) Correlation analysis should indicate a high degree of correlation between the independ-
ent variable, x, and the dependent variable, y.

In addition to plotting the points on a scatter plot graph, correlation analysis should be per-
formed before relying on regression analysis to develop a prediction.

Correlation analysis determines the strength of the linear relationship between the x-values
and their related y-values. The results of the correlation analysis tell the analyst whether the
relationship between the independent variable (the passage of time, for a time series) and the
dependent variable (sales, for example) is reasonable.

Note: Correlation describes the degree of the relationship between two variables. If two
things are correlated with one another, it means there is a close connection between them.

 If high measurements of one variable tend to be associated with high measurements of the
other variable, or low measurements of one variable tend to be associated with low meas-
urements of the other variable, the two variables are said to be positively correlated.

 If high measurements of one variable tend to be associated with low measurements of the
other variable, the two variables are said to be negatively correlated.

 If there is a close match in the movements of the two variables over a period of time, either
positive or negative, it is said that the degree of correlation is high.

However, correlation alone does not prove causation. Rather than one variable causing the
other variable to occur, it may be that some other, entirely different, factor is affecting both vari-
ables.

In a time series, the only independent variable is the passage of time. Many factors in addition to
time can affect the dependent variable. For example, if sales are being predicted, economic cycles,
promotional programs undertaken, and industry-wide conditions such as new government regula-
tions can cause changes in sales volume. If time series regression analysis is used to develop a
prediction, the prediction should be adjusted for other known factors that may have affected the
historical data and that may affect the prediction.

Note: Correlation does not prove causation.

Correlation Analysis
Correlation analysis is used to assess how well a model can predict an outcome.

Correlation analysis involves several statistical calculations, all done with a computer or a financial calcula-
tor, using the observed values of x and y. Correlation analysis is used to determine how well correlated the
variables are in order to decide whether the independent variable or variables can be used to make decisions
regarding the dependent variable used in the analysis.

Some of the most important statistical calculations for determining correlation are:

1) The correlation coefficient, R

2) The standard error of the estimate, also called the standard error of the regression

3) The coefficient of determination, R2

4) The T-statistic

The Correlation Coefficient (R)


The correlation coefficient measures the relationship between the independent variable and the depend-
ent variable. The coefficient of correlation is a number that expresses how closely related, or correlated,

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 323
F.4. Data Analytics CMA Part 1

the two variables are and the extent to which a variation in one variable has historically resulted in a
variation in the other variable.

Mathematically, the correlation coefficient, represented by R, is a numerical measure that expresses both
the direction (positive or negative) and the strength of the linear association between the two variables.
In a time series using linear regression analysis, the period of time serves as the independent variable (x-
axis) while the variable such as the sales level serves as the dependent variable (y-axis).

When a time series (such as sales over a period of several years) is graphed, the data points on the graph
may show an upsloping linear pattern, a downsloping linear pattern, a nonlinear pattern (such as a curve),
or no pattern at all. The pattern of the data points indicates the amount of correlation between the values
on the x-axis (time) and the values on the y-axis (sales).

The amount of correlation, or correlation coefficient (R), is expressed as a number between −1 and
+1. The sign of the correlation coefficient and the absolute value of the correlation coefficient describe
the direction and the magnitude of the relationship between the two variables.

• A correlation coefficient (R) of +1 means the linear relationship between each value for x and its
corresponding value for y is perfectly positive (upsloping). When x increases, y increases by the
same proportion; when x decreases, y decreases by the same proportion.

• A correlation coefficient (R) of −1 means the linear relationship between each value for x and its
corresponding value for y is perfectly negative (downsloping). When x increases, y decreases
by the same proportion; when x decreases, y increases by the same proportion. In a time series,
the x value (time) can only increase, so if the correlation coefficient is negative, the level of the y
value decreases with the passage of time.

• A correlation coefficient (R) that is close to zero usually means there is very little or no relation-
ship between each value of x and its corresponding y value. However, a correlation coefficient that
is close to zero may mean there is a strong relationship, but the relationship is not a linear one.
(Candidates do not need to know how to recognize a non-linear relationship. Just be aware that
non-linear relationships occur.)

A high correlation coefficient (R), that is, a number close to either +1 or −1, means that simple linear
regression analysis would be useful as a way of making a projection. Generally, a correlation coefficient of
±0.50 or higher indicates enough correlation that a linear regression can be useful for forecasting. The
closer R is to ±1, the better the forecast should be.

• If R is a positive number close to +1 (such as 0.83), it indicates that the data points follow a
linear pattern fairly closely and the pattern is upsloping. In a time series analysis of sales, it means
sales are increasing as the time moves forward. A forecast made from this data using simple
regression analysis should be fairly accurate.

• If R is a negative number close to −1 (such as −0.77), it indicates in a time series that the data
points follow a linear pattern, although less closely than in the previous example, and the pattern
is downsloping instead of upsloping (for example, sales are decreasing as the time moves forward).
A forecast made from this data using simple regression analysis would also be fairly accurate,
though not as accurate as the previous example of an upsloping pattern, because the absolute
number of 0.77 is lower than the absolute number of 0.83. In other words, −0.77 is further from
−1 than 0.83 is from +1.

A moderate correlation coefficient (R), generally defined as ±0.30 to ±0.49, indicates a lower amount of
correlation and questionable value of the historical data for forecasting.

A low correlation coefficient (R), around ±0.10, indicates that a forecast made from the data using
simple regression analysis would not be useful.

The correlation coefficient (R) does not indicate how much, that is, the proportion, of the variation in the
dependent variable that is explained by changes in the independent variable. The correlation coefficient

324 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

indicates only whether there is a direct (upsloping) or inverse (downsloping) relationship between the pairs
of x and y variables and the strength of that relationship.

Note: The correlation coefficient, R, can be calculated in Excel by entering the x values in one column
(for example, Column A, rows 1-10), the y values in another column (for example, Column B, Rows 1-
10), and, for the current example, entering the following formula in any blank cell:

=CORREL(A1:A10,B1:B10)

The correlation coefficient is also a part of the output of a regression analysis performed in Excel.

Since Excel is not available on the CMA exam, candidates do not need to know how to calculate either a
correlation coefficient or a regression analysis in Excel for the exam. The information will be given in any
exam question where it is needed, and the candidate needs to know only how to interpret it.

Note: It is important to first look at the plotted data points on the scatter plot graph when determining
whether or not there is a relationship between the independent variable and the dependent variable. Do
not rely on the value of the correlation coefficient alone to indicate whether or not there is a relationship
between the two variables because the correlation coefficient will not detect non-linear relationships.

Question 68: Correlation is a term frequently used in conjunction with regression analysis, and is meas-
ured by the value of the coefficient of correlation, R. The best explanation of the value R is that it

a) is always positive.

b) interprets variances in terms of the independent variable.

c) ranges in size from negative infinity to positive infinity.

d) is a measure of the relative relationship between two variables.

(CMA Adapted)

The Standard Error of the Estimate, also called the Standard Error of the Regression (S) and the Error
Term (e)
Recall that the equation of a linear regression line, or the “line of best fit” on a graph is:

ŷ = a + bx

Where:

ŷ= the predicted value of the dependent variable, ŷ, on the regression


line corresponding to each value of x.
a= the constant coefficient, or the y-intercept, the value of ŷ on the
regression line when x is zero.
b= the variable coefficient and the slope of the regression line, which
is the amount by which the ŷ value of the regression line changes (ei-
ther increases or decreases) when the value of x increases by one unit.
x= the independent variable, also called the predictor variable, or the
value of x on the x-axis that corresponds to the value of ŷ on the re-
gression line.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 325
F.4. Data Analytics CMA Part 1

Note: In a time series, the value of x only increases because time moves in only one direction: forward.

The equation of the simple linear regression model results in the average, or predicted value of ŷ (the
response) for any given value of x (the predictor). However, the actual observed data has responses that
are not on the line itself, but rather they are scattered around the regression line. Thus, on a graph of
a regression of historical data, there are two y values for each value of x: one is the actual historical value
of y for that value of x, and the other is the estimated, or predicted, value of ŷ for that value of x, repre-
sented by the y value on the trend line aligned with each value of x.

The scatter, that is, the difference between the actual value of y and the estimated value of ŷ for each
value of x, is called the error term or the residual for that value of x. The error term—the scatter of the
data around the line—is represented by e in the linear regression model. If all of the historical data fell on
the regression line, the error term for each value of x on the graph would be zero.

In algebra, an equation such as y = 2,000 + 300x means that y is exactly 2,000 + 300x. However, with
regression data, the equation ŷ = 2,000 + 300x is true on average but is not true for any given value of
x.

Therefore, the equation that describes a given actual, historical, value of x used in a regression is as
follows:

y = a + bx + e

Where:

y= the dependent variable (its actual, historical value, not its predicted
value) corresponding to each value of x.
a= the constant coefficient, or the y-intercept, the value of ŷ on the
regression line when x is zero.
b= the variable coefficient and the slope of the regression line, or the
average amount of change in y as a result of one unit of change in x.
x= the independent variable, or the value of x on the x-axis that corre-
sponds to the value of ŷ on the regression line.

e= the error term, also called the residual, which for each value of x is
the difference between the estimated ŷ value on the regression line for
that value of x and the actual value of y for that value of x. The error
term will be different for each value of x used in the regression
function.

The standard error of the estimate (S) represents the average distance that the observed values
fall from the regression line. In other words, it describes how wrong the regression model is on aver-
age, using the units of the dependent variable, y. The standard error of the estimate can give an indication
of how well a linear regression works. It provides a comparison of the actual values of y (that is, the values
that did occur historically) to the estimated values of ŷ on the regression line. The estimated values of ŷ on
the regression line are the values that result from putting the various values for x into the regression
function and calculating the resulting value of ŷ at each value of x.

Each value of x has one residual, or error term. For any given value of x,

e=y−ŷ

326 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Some of the residuals for a dataset that has been regressed will be positive and some will be negative.
However, in a regression that has a constant term, the mean of the residuals will be exactly zero.

The standard error of the estimate (S) gives an indication of the precision, that is, the predictive ability of
the regression model. The lower the standard error is, the more accurate will be the predictions made using
the regression model.

Note: Smaller values for the standard error of the estimate are better because that indicates the obser-
vations are closer to the fitted line. However, the size of the standard error of the estimate must
be interpreted in relationship to the average size of the dependent variable. If the average size
of the dependent variable is 5,000,000 and the standard error is 250,000, the percentage of error is
fairly small: 250,000 is only 5 percent of 5,000,000. If the size of the standard error of the estimate is
less than 5 to 10 percent of the average size of the dependent variable, the regression analysis is fairly
precise.

The inclusion of an error term in the regression model recognizes that:

• The regression model is imperfect.

• Some variables that help to “explain” the behavior of the dependent variable might not be included.

• The included variables may have been measured with error.

There is always some component in the variation of the dependent variable that is completely random.

The Coefficient of Determination (R2)


The coefficient of determination is the percentage of the total variation in the dependent variable
(y) that can be explained by variations in the independent variable (x), as depicted by the re-
gression line.

In a simple linear regression with only one independent variable, the coefficient of determination is the
square of the correlation coefficient. The coefficient of determination is represented by the term R2.

R2 is expressed as a number between 0 and 1.

• If R2 is 1, then 100% of the variation in the dependent variable is explained by variations in the
independent variable.

• If R2 is 0, then none of the variation in the dependent variable is explained by variations in the
independent variable.

In a regression analysis with a high coefficient of determination (R2), the data points will all lie close to the
trend line. In a regression analysis with a low R2, the data points will be scattered at some distances above
and below the regression line. The higher the R2, the better the predictive ability of the linear regression.

The T-Statistic
The t-statistic, or t-value, measures the degree to which the independent variable has a valid, long-term
relationship with the dependent variable. The t-value for the independent variable used in a simple regres-
sion analysis should generally be greater than 2. A value below 2 indicates little or no relationship
between the independent variable and the dependent variable, and thus the forecast resulting from the
regression analysis should not be used.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 327
F.4. Data Analytics CMA Part 1

Note: The equation of a linear regression line may be written in different ways, but x will usually repre-
sent the independent variable and y will usually represent the dependent variable.

The constant coefficient is the letter that stands by itself. The constant coefficient represents the y-
intercept because it is the value of y when x is zero.

The coefficient of the independent variable, or the variable coefficient, is whatever term is next
to the x in the formula. That term represents the amount of change in y for each unit of increase
in x, or the slope of the regression line.

The regression equation may be written in various ways, though the standard form of the equation used
in statistics is

ŷ = a + bx
However, the equation ŷ = ax + b is exactly the same as the equation ŷ = a + bx. The coefficients
have just been expressed differently and the order of the terms on the right side of the equation is
reversed. The right side of the equation may present the terms in any order.

If the regression line has a negative slope, that is, the values of ŷ decrease for each increase in the
value of x, the equation will be ŷ = a – bx.

Remember to look for the x, or the independent variable. The term next to it will be the variable
coefficient, or the amount of change in y for each unit of increase in x. The term that is all by itself will
be the constant coefficient and the y-intercept, or the value of y when x is zero.

The symbol over the y in the formula is called a “hat,” and it is read as “y-hat.” The y-hat indicates the
predicted value of y, not the actual value of y. The predicted value of y is the value of y on the
regression line (the line created from the historical data) at any given value of x.

Multiple Regression Analysis


When more than one independent variable is known to impact a dependent variable and each independent
variable can be expressed numerically, regression analysis using all of the independent variables to forecast
the dependent variable is called multiple regression analysis.

Note: Remember that there must be a reasonable basis to assume a cause-and-effect relationship be-
tween the independent variable(s) and the dependent variable. If there is no reason for a connection,
any correlation found through the use of regression analysis is accidental. A linear relationship does
not prove a cause-and-effect relationship, and correlation does not prove causation.

The equation of a multiple regression function is usually written with either all “a”s or all “b”s as the coef-
ficients, with a subscripted zero to indicate the constant coefficient and subscripted subsequent numerals
to indicate the variable coefficients, such as the following, although any letters could be used:

ŷ = a0 + a1x1 + a2x2 + ... + akxk

328 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Note: The variables and the coefficients in a multiple regression equation could be identified using any
letters. To identify the various components, look for the form of the equation rather than the specific
letters.

• The equation will have one component that stands by itself, and that will be the constant coefficient.

• The variable coefficients will be next to the independent variables.

• The independent variables may or may not be identified by “x”s.

Evaluating the Reliability of a Multiple Regression Analysis


In simple regression analysis, the coefficient of determination, R2, is the proportion of the total variation
in the dependent variable (y) that can be explained by variations in the independent variable (x). Thus, R2
is an indicator of the reliability of a simple regression analysis.

R2 is used as an indicator of the reliability of a multiple regression analysis, as well. In multiple regression
analysis, though, the R2 value evaluates the whole regression, including all of the independent variables
used. If R2 is above 0.50 or 50%, then the regression is fairly reliable because the regression can be used
to predict that percentage of the total variation in the dependent variable. The higher the R2 is, the better.

Multiple regression analysis uses the t-value—actually t-values (plural)—to evaluate the reliability of each
individual independent variable as a predictor of the dependent variable. The t-value for each inde-
pendent variable evaluates the contribution of that independent variable to the multiple regression
analysis. A separate t-value is calculated for each of the individual independent variables in the multiple
regression, and each independent variable is evaluated individually.

The t-value for every independent variable used in a multiple regression should generally be
greater than 2, the same as in simple regression analysis. A t-value below 2 for a given independent
variable indicates little or no relationship between that independent variable and the dependent variable;
thus, that independent variable should not be used in the multiple regression.

As with simple regression, the R2 and the t-values are based on the input supplied for the independent
variables and the dependent variable.

Before using each independent variable in the multiple regression analysis, each independent variable
being considered should be evaluated individually. Examine the output of the regression analysis for each
independent variable, including the coefficient of correlation, R, the coefficient of determination, R2,
and the t-value. The evaluation will be similar to the evaluation done when with simple regression analysis
with one independent variable.

If evaluation of each independent variable is done before running the final multiple regression, and if
only those independent variables that are highly correlated with the dependent variable are used in the
final multiple regression, the t-values that result from the final multiple regression should all be greater
than 2. If, however, one or more of the t-values are below 2, then those independent variables should be
eliminated and the multiple regression run again, even if the R2 for the whole regression is high (for exam-
ple, 0.85).

Note: Remember, correlation does not prove causation. In addition to a strong correlation between each
independent variable and the dependent variable, there must be a logical cause-and-effect relationship
between them before the independent variable can be used effectively to predict the dependent variable.

If the coefficient of determination, R2, is only 0.50 for a multiple regression analysis (which is low but
acceptable), the multiple regression might be used to predict 50% of the variation in the dependent varia-
ble, but only if the t-values for all the independent variables are greater than 2.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 329
F.4. Data Analytics CMA Part 1

Goodness of Fit in Regression Analysis


The term goodness of fit describes how close the actual values used in a statistical model are to the
expected values, that is, the predicted values, in the model.

In regression analysis, the regression equation is the model used to predict future values based on the
behavior of the actual observations in response to a predictor. Thus, the correlation analysis described in
this topic leads to the measurement of the regression equation’s goodness of fit.

When the independent variable or variables used in the regression are not well correlated with the obser-
vations of the dependent variable used in the regression, the regression line is said to have a low goodness
of fit.

The chart that follows exemplifies low goodness of fit for the regression equation, ŷ = $2,352,273 +
$15,000x. Sales of soda (the dependent variable on the y-axis) are regressed on sales of beans (the inde-
pendent variable on the x-axis). Most of the points on the regression line are very far from the observed
values of the dependent variable at that value for the independent variable, indicating very little correlation
between the two variables.

Sales of Soda as a Function of Sales of Beans


$4,000,000

$3,500,000

$3,000,000
Sales of Soda

$2,500,000

$2,000,000

$1,500,000

$1,000,000

Sales of Beans
Sales of Soda Regression Line: y = $2,352,273 + 15,000x

On the other hand, when the independent variable or variables used in the regression are highly correlated
with the observations of the dependent variable used in the regression, the regression equation is said to
have a high goodness of fit.

330 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Following is a chart showing sales revenue (on the y-axis) regressed on historical advertising expense (on
the x-axis) for a representative period. The regression equation, ŷ = $2,062,045 + $97,682x, has a high
goodness of fit. The points representing historical sales as a function of advertising expense are very close
to the regression line.

Sales as a Function of Advertising Expense


$4,000,000

$3,500,000
y-axis: Sales

$3,000,000

$2,500,000

$2,000,000

$1,500,000 Sales
Regression Line: ŷ = $2,062,045 + $97,682x
$1,000,000

x-axis: Advertising Expense


$50,000 represents x = 0, $100,000 represents x = 5,
$150,000 represents x = 10, and so forth

Remember, when calculating the predicted value of ŷ at a given value of x, reassign values to the x-axis
beginning with zero. For example, the value on the x-axis that represents advertising expense of $110,000
in the formula is x = 6. Thus, the predicted value of ŷ when advertising expense is $110,000 is

ŷ = $2,062,045 + ($97,682 x 6)

ŷ = $2,648,137

In the above example, using $110,000 as x in the formula will result in an incorrect value for ŷ.

Confidence Interval in Regression Analysis


The confidence interval is used in regression analysis to describe the amount of uncertainty caused by
the sampling method used when drawing conclusions about a population based on a sample. If several
samples are drawn from a population using the same sampling method and a confidence interval at a
confidence level of 95% is used, 95% of the interval estimates in the samples can be expected to
include the true parameter of the population.

The following chart contains information on number of sales made by an Internet retailer on its website
regressed against the number of unique visitors to the website each day during a representative calendar
period, January 1 through January 31, 20X1. The chart includes the upper and lower bands of a confidence
interval at a 95% confidence level.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 331
F.4. Data Analytics CMA Part 1

Number of Sales Versus Number of Unique Website


Visitors Per Day, January 1 through January 31, 20X1

1,000
Sales
900
Regression Line: ŷ = 38.0667 +
0.0158x
Number of Sales Per Day

800
Upper Confidence Interval Boundary,
700 95% confidence level
Lower Confidence Interval Boundary,
95% confidence level
600

500

400

300

200

100

Number of Unique Visitors to Website Per Day

Note that several of the observations are actually outside the 95% confidence interval. That fact illustrates
what the confidence interval is and highlights what it is not.

This sample’s confidence interval at a confidence level of 95% does not mean that 95% of the observations
in this sample, in any other sample, or in the population will be within the confidence interval, nor does it
mean that the true value of sales as a function of website visitors will be within that interval 95% of the
time. Instead, a confidence interval of 95% means that if several periods are sampled and analyzed
using the same 95% confidence interval, the proportion of those sample intervals that would
contain the true number of sales to website visitors in the population would be equal to 95%.

Example: If the same sampling of sales versus website visitors were performed for each month of 20X1,
20X2, and 20X3 (36 months) using the same 95% confidence interval, the location of the confidence
interval bands would be slightly different for each month. However, for 34 out of the 36 sampled
months (95%), the bands would contain the true value of sales related to website visitors for all past,
present, and future periods (that is, the population).

332 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Benefits and Limitations of Regression Analysis

Benefits of Regression Analysis

• Regression analysis is a quantitative method and as such it is objective. A given data set generates
specific results. The results can be used to draw conclusions and make predictions.

• Regression analysis is an important tool for drawing insights, making recommendations, and deci-
sion-making.

Limitations of Regression Analysis

• To use regression analysis, historical data are required. If historical data are not available, regression
analysis cannot be used.

• Even when historical data are available, the use of historical data is questionable for making predic-
tions if a significant change has taken place in the conditions surrounding that data.

• The usefulness of the data generated by regression analysis depends on the choice of independent
variable(s). If the choice of independent variable(s) is inappropriate, the results can be misleading.

• The statistical relationships that can be developed using regression analysis may be valid only for the
range of data in the sample.

Sensitivity Analysis
Sensitivity analysis can be used to determine how much the prediction of a model will change if one input
to the model is changed. It can be used to determine which input parameter is most important for achieving
accurate predictions. Sensitivity analysis is known as “what-if” analysis.

To perform sensitivity analysis, define the model and run it using the base-case assumptions to determine
the predicted output. Next, change one assumption at a time, leaving the other assumptions unchanged
and run the model again to determine what effect changing that one assumption has on the predictions of
the model. The amount of sensitivity of the prediction to the change in the input is the percentage of change
in the output divided by the percentage of change in the input. Sensitivity analysis may reveal some area
of risk that the company had not been aware of previously.

Monte Carlo Simulation Analysis


Whereas sensitivity analysis involves changing one input variable at a time, a Monte Carlo simulation
analysis can be used to find solutions to mathematical problems that involve changes to multiple variables
at the same time. Monte Carlo simulation can be used to develop an expected value when the situation is
complex and the values cannot be expected to behave predictably. Monte Carlo simulation uses repeated
random sampling and can develop probabilities of various scenarios coming to pass that can be used to
compute a predicted result.

Adding a Monte Carlo simulation to a model allows the analyst to assess various scenario probabilities
because various random values for the probabilistic inputs can be generated based on their probability
distributions. The analyst determines ranges for the probabilistic inputs and also their probability distribu-
tions, means, and standard deviations. The application then generates the random values for the
probabilistic inputs based on their ranges, probability distributions, means, and standard deviations as
determined by the analyst.

The values for the probabilistic inputs are used to generate multiple possible scenarios, similar to performing
statistical sampling experiments, except that it is done on a computer and over a much shorter time span
than actual statistical sampling experiments. Enough trials are conducted (indeed, hundreds or thousands)
with different values for the probabilistic inputs in order to determine a probability distribution for the
resulting scenario, which is the output. The repetition is an essential part of the simulation.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 333
F.4. Data Analytics CMA Part 1

For example, if the simulation is run to evaluate the probability that a new product will be profitable, the
output may include a prediction for average profit and the probability of a loss.

Benefits of Sensitivity Analysis and Simulation Models


• Sensitivity analysis can identify the most critical variables, that is, the variables that are most likely
to affect the end result if they are inaccurate. Since those are the variables that will make the most
difference, those are the variables that should receive the most attention in making predictions.
• Simulation is flexible and can be used for a wide variety of problems.
• Both sensitivity analysis and simulation analysis can be used for “what-if” situations, because they
enable the study of the interactive effect of variables.
• Both sensitivity analysis and simulation analysis are easily understood.
• Many simulation models can be implemented without special software packages because most
spreadsheet packages provide useable add-ins. For more complex problems, simulation applications
are available.
Limitations of Sensitivity Analysis and Simulation Models

• The results of sensitivity analysis can be ambiguous when the inputs used are themselves predic-
tions.
• The variables used in a sensitivity analysis are likely to be interrelated. Changing just one variable
at a time may fail to take into consideration the effect that variable’s change will have on other
variables.
• Simulation is not an optimization technique. It is a method that can predict how a system will
operate when certain decisions are made for controllable inputs and when randomly generated val-
ues are used for the probabilistic inputs.
• Although simulation can be effective for designing a system that will provide good performance,
there is no guarantee it will be the best performance.
• The results will be only as accurate as the model that is used. A poorly developed model or a model
that does not reflect reality will provide poor results and may even be misleading.
• There is no way to test the accuracy of assumptions and relationships used in the model until a
certain amount of time has passed.

Benefits of Data Analytics in General


• The process of cleaning the data preparatory to processing it can detect errors, duplicate infor-
mation, and missing values. If the errors and duplicate information can be corrected and the missing
values supplied, the data quality can be improved.
• The results of data analytics done correctly can lead to improved sales revenues and profits.
• It can help to reduce fraud losses by recognizing potentially fraudulent transactions and flagging
them for investigation.
• Some easy-to-use data analytics tools are available that average users with little knowledge of data
science are able to make use of to access data, perform queries, and generate reports. As a result,
data scientists can be freed up to do more critical data analysis projects.
• Forecasting can be vastly improved through the use of data analytics.

334 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Limitations of Data Analytics in General

• Big Data is used in data analytics to find correlations between variables. However, correlation does
not prove causation. The fact that two variables are correlated does not mean that one variable
caused the other. Both variables could have been caused by a third, unidentified, factor.
• Big Data can be used to find correlations and insights using an endless number of questions. But if
the wrong questions are asked of the data, the answer will be meaningless even though it may be
the “right” answer.
• Failure to take into consideration all relevant variables can lead to inaccurate predictions.
• Data breaches are a risk of using Big Data.
• Customer privacy issues and the risk of the misuse of data obtained from data analytics are matters
for concern.
• In addition to the cost of the data analytics tools themselves, training on the use of the tools so
they are used to their best advantage may entail costs, as well.
• Some easy-to-use data analytics tools are available that average users with little knowledge of data
science are able to make use of to access data, perform queries, and generate reports. Use of the
tools by those without a background in statistical analysis and data science and without adequate
training, though, can cause risks such as data inconsistency, a lack of knowledgeable verification of
the results, a lack of proper data governance, and ultimately, poor decisions.
• Selection of the right data analytics tools can be difficult.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 335
F.4. Data Analytics CMA Part 1

Visualization
Data visualization is used for better understanding data and predictions from data. Charts, tables, and
dashboards can be used to explore, examine, and display data. Interactive dashboards allow users to access
and interact with real-time data and give managers a means to quickly see what might otherwise not be
readily apparent. The choice of information to include in a dashboard depends on what a manager needs
to see and can include visual presentations such as colored graphs showing, for instance, current customer
orders.

In the data mining process, visualization is primarily used in exploration and cleaning of the data in the
preprocessing step of data mining and in the reduction of the data dimensions step of the process.112 For
example:

• Visualization used in data exploration can help the analyst determine which variables to include in
the analysis and which variables might be unnecessary.

• Visualization is used in data cleaning to find erroneous values in the historical data that need to
be corrected (such as a sale recorded with a date 10 years in the future or a patient aged 250
years because his birth date is incorrect), missing values, duplicate records, columns that may
have the same values in all the rows, and so forth.

• In data reduction, visualization can help in determining which categories can be combined.

Some visualization options are presented in the following pages, and information on how each can be used
is provided.

Tables Used in Visualization


A table can be in any form and include all of the data available or only certain data.

The data table below will be used in all the chart examples that follow. The table contains data on the
number of pounds of strawberries sold on each day of the week by a grocery store over a twelve-week
period from June 3 through August 25. This information is used in placing orders, so that enough straw-
berries are purchased to meet the anticipated demand each day without over-buying and having excess
strawberries that will need to be thrown away because they spoil.

In addition to daily strawberry sales for each of the days of the week for twelve weeks, the data table below
contains the mean (average) for each day of the week over the twelve-week period.

Day
of the Jun. Jun. Jun. Jun. Jul. Jul. Jul. Jul. Jul.29- Aug. Aug. Aug.
Week 3-9 10-16 17-23 24-30 1-7 8-14 15-21 22-28 Aug. 4 5-11 12-18 19-25 Mean

Mon. 15 10 28 39 48 25 12 20 30 23 28 22 25

Tues. 35 25 45 40 46 49 30 60 38 32 22 58 40

Wed. 68 42 57 74 84 55 30 55 60 75 88 32 60

Thur. 60 80 65 90 65 85 50 70 110 75 45 105 75

Fri. 95 60 85 90 70 80 105 85 50 80 105 55 80

Sat. 110 75 85 98 75 102 85 50 120 100 65 115 90

Sun. 11 7 14 10 40 18 20 25 35 20 22 18 20

112
See Steps in Data Mining earlier in this section.

336 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Scatter Plot
A scatter plot can be used to show all the values for a dataset, typically when there are two variables. One
variable may be independent and the other value dependent, or both variables may be independent. The
independent variable is generally plotted on the horizontal (x) axis and the dependent variable is plotted
on the vertical (y) axis.

A scatter plot can reveal correlations between variables or alternatively, a lack of correlation. For example,
do sales of strawberries correlate with days of the week? A scatter plot can answer that question.

In this case, it appears that strawberry sales do correlate with days of the week. Sales build from Monday
through Saturday and then they drop off on Sunday each week.

Scatter Plot Chart


Number of Pounds of Strawberries Sold Per Day
June 3 through August 25

120
Number of Pounds Sold Per Day

100

80

60

40

20

Jun. 3-9 Jun. 10-16 Jun. 17-23 Jun. 24-30 Jul. 1-7 Jul. 8-14
Jul. 15-21 Jul. 22-28 Jul. 29-Aug. 4 Aug. 5-11 Aug. 12-18 Aug. 19-25

Charts Containing Summarized Statistics


Several charts are used to present summarized statistics such as means, maximum values, and minimum
values.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 337
F.4. Data Analytics CMA Part 1

Dot Plot
A dot plot provides information in the form of dots. A dot plot can be used to visualize several data points
for each category on the x-axis. For example, the following dot plot shows the minimum, the maximum,
and the mean number of pounds of strawberries sold for each day of the week during the twelve-week
period.

Dot Plot
Strawberry Sales By Day of the Week
June 3 through August 25

Minimum Maximum Mean


Number of Pounds Sold Per Day

140

120

100

80

60

40

20

Day of the Week

338 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Bar Chart
A bar chart is useful for comparing a statistic across groups. The height of the bar or the length of it, if the
bar is displayed horizontally, displays the value of the statistic.

The following bar chart shows the mean number of pounds of strawberries sold per day over the twelve-
week period. Thus, the Monday sales figure is the average of twelve Mondays, and so forth for each of the
days of the week. This chart can be used to easily visualize which are the heaviest days of the week for
selling strawberries in order to place orders at the appropriate times.

Vertical Bar Chart


Strawberry Sales by Day of the Week
June 3 through August 25
100
Mean Number of Pounds Sold Per Day

90

80

70

60

50

40

30

20

10

Day of the Week

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 339
F.4. Data Analytics CMA Part 1
efham CMA
A bar chart can also be used to portray values horizontally, and when it does, it becomes the exception to
the rule that the independent variable is generally on the x-axis (horizontal) and the dependent variable is
generally on the y-axis (vertical). When a bar chart portrays the values horizontally, the independent var-
iable is on the vertical axis and the bars extend to their values on the horizontal axis, representing the
dependent variable.

The following horizontal bar chart is used to show not only the mean sales in pounds for each day of the
week but also the minimum sales and maximum sales for each day, which are also important information.

Horizontal Bar Chart


Strawberry Sales by Day of the Week
June 3 through August 25

Sun.
Days of the Week

Sat.

Fri.

Thurs.

Wed.

Tues.

Mon.

0 20 40 60 80 100 120 140

Number of Pounds Sold Per Day


Minimum Mean Maximum

Here is the same information presented vertically:

Vertical Bar Chart


Strawberry Sales by Day of the Week
June 3 through August 25

120
Number of Pounds Sold Per Day

110
100
90
80
70
60
50
40
30
20
10
0

Day of the Week


Maximum Mean Minimum

340 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Pie Chart
A pie chart is in the form of a circle that portrays one value for each category, marked as pieces of a pie.
In the example that follows of the mean pounds of strawberries sold per day for each day of the week over
the twelve-week period, the pieces of the pie represent the days of the week and each one is sized to
represent the mean sales for that day. The size of the “pieces” helps the user to visualize the relative sizes
of the mean sales for each day.

A pie chart does not have axes. Therefore, including values on the chart as labels helps users to interpret
and use the information. The mean pounds of strawberries sold on each day have been added to each pie
piece in the chart that follows. Values can be added as labels to data in other types of charts as well, but
in some charts, doing so tends to make the chart hard to read.

A limitation of the pie chart is that it can present only one value for each category.

Pie Chart
Mean Number of Pounds of Strawberries Sold
by Day of the Week
June 3 through August 25

Sun. Mon.
20 25
Tues.
Sat. 40
90
Wed.
60

Fri.
Thurs.
80
75

Mon. Tues. Wed. Thurs. Fri. Sat. Sun.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 341
F.4. Data Analytics CMA Part 1

Line Chart
A line chart can be used to visualize several observations for each category, using one line for each obser-
vation. The strawberry sales data are shown on the following line chart as the minimum, the maximum,
and the mean values for each day of the week for the twelve-week period.

Line Chart
Strawberry Sales by Day of the Week
June 3 through August 25
Maximum Minimum Mean
140
Number of Pounds Sold Per Day

120

100

80

60

40

20

Day of the Week

Bubble Chart
A bubble chart replaces data points with bubbles that vary in size according to the size of the values they
depict, thus adding an additional dimension to the chart: the relative sizes of the values plotted on the
chart.

The following bubble chart shows the means of the strawberry sales for each day of the week.

Bubble Chart
Strawberry Sales by Day of the Week
Mean Number of Pounds Sold

June 3 through August 25


120

100
90
80 80
75
60 60

40 40

20
25
20

Mon. Tues. Wed. Thurs. Fri. Sat. Sun.

Day of the Week

342 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Charts Containing the Entire Distribution of Values


Although summaries and averages are very useful, much can be gained by looking at additional statistics
such as the median of a set of data or by examining the full distribution of the data. Histograms and
boxplots can be used to display the entire distribution of a numerical variable.

Histogram
A histogram shows the frequencies of a variable using a series of vertical bars. The values of the variable
may occur over a period of time, or they may be as of a moment in time.

The following histogram shows how many days during the twelve-week period the sales of strawberries
were between 1 and 20 pounds, how many days between 21 and 40 pounds were sold, and so forth.

A histogram looks similar to a bar graph. However, it is different from a bar graph in that a bar graph
relates two variables to one another, whereas a histogram communicates only one variable.

To construct a histogram, the range of values of the variable must be first divided into intervals, or bins,
and then the number of values that fall into each interval are counted. For the histogram that follows, six
bins are used. The bins and the values in each bin are:

Number of Pounds Frequency


Sold Per Day (Number of Days)
1-20 pounds 12
21-40 pounds 20
41-60 pounds 18
61-80 pounds 14
81-100 pounds 12
101-120 pounds 8

Histogram
Pounds of Strawberries Sold Per Day
June 3 through August 25
25
Frequency (Number of Days)

20

15

10

Pounds Sold Per Day

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 343
F.4. Data Analytics CMA Part 1

Boxplot
A boxplot is another type of chart that is used to display the full distribution of a variable. A boxplot shows
the minimum, the first quartile, the median, the mean, the third quartile, and the maximum for
each day of the week, as well as individual observations for each day of the week.

Note:

• The minimum is the smallest value in a distribution.

• The median is the middle value in a distribution. If the distribution contains an odd number of values,
the median is the value with an equal number of values below it and above it. If the distribution
contains an even number of values, the median is the mean (average) of the middle two numbers.

• The first quartile (Q1) is the middle value between the minimum value and the median of a distri-
bution; or, if two values are in the middle, it is the mean of those two values. If the median falls
between two values, the first quartile is calculated as one-quarter of the difference between the
middle number (or the mean of the two middle numbers) and the next largest value when the values
are ordered from the smallest to the largest.

• The second quartile (Q2) is the same as the median of the distribution.

• The third quartile (Q3) is the middle value between the median and the maximum value in the
distribution; or, if two values are in the middle, it is the mean of those two values. If the median falls
between two values, the third quartile is calculated as one-quarter of the difference between the
middle number (or the mean of the two middle numbers) and the next smallest value when the
values are ordered from the smallest to the largest.

• The maximum is the largest value in the distribution.

In the table that follows, the data on daily strawberry sales for the twelve weeks have been re-ordered by
day from the smallest value to the largest value for that day. Ordering the data in that way makes apparent
the minimum, median, and maximum values for each day of the week and the approximate values for the
first and third quartiles.

Note: The first quartile and the third quartile are approximately equal to the values in the table col-
umns marked ≈Q1 and ≈Q3 that follow. The first quartile for each day is the middle value between the
minimum value for that day and the median. Since the median is halfway between two values in this
dataset, one-quarter of the difference between the number in the column marked ≈Q1 and the next
largest value is added to the number in the column marked ≈Q1 to calculate the actual first quartile.
The third quartile is the middle value between the median and the maximum value. One-quarter of the
difference between the value in the column marked ≈Q3 and the next smallest value is subtracted from
the number in the column marked ≈Q3 to calculate the actual third quartile.

344 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
Section F F.4. Data Analytics

Median=the
Day of mean of Mean Median
the Mini- these two Maxi- (calcu- (calcu-
Week mum ≈Q1 columns ≈Q3 mum lated) lated)

Mon. 10 12 15 20 22 23 25 28 28 30 39 48 25 24.0

Tues. 22 25 30 32 35 38 40 45 46 49 58 60 40 39.0

Wed. 30 32 42 55 55 57 60 68 74 75 84 88 60 58.5

Thur. 45 50 60 65 65 70 75 80 85 90 105 110 75 72.5

Fri. 50 55 60 70 80 80 85 85 90 95 105 105 80 82.5

Sat. 50 65 75 75 85 85 98 100 102 110 115 120 90 91.5

Sun. 7 10 11 14 18 18 20 20 22 25 35 40 20 19.0

A boxplot chart of the data follows.

Each day’s box encloses the first quartile through the third quartile of values for that day. For example, on
Wednesday, the bottom of the box is at the calculated first quartile of 45.25, calculated as 42 + (0.25 ×
[55 − 42]). The top of the box is at the calculated third quartile of 74.75, calculated as 75 − (0.25 × [75
− 74]).

The horizontal bar through each box marks the location of the median, which for Wednesday is the average
of 57 and 60, or 58.5.

The “X” in each box marks the location of the mean for that day, which for Wednesday is 60.

The lines that extend vertically above and below each box with horizontal lines at the ends (called “whisk-
ers”) mark the minimum and maximum for each day. For Wednesdays, the minimum is 30 and the
maximum is 88.

The circles mark values, and those outside the boxes are considered “outliers.” “Outliers” are values that
are far away from most of the other values in the distribution.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 345
Answers to Questions CMA Part 1

Answers to Questions
1 c – When normal costing is used, the difference between the actual variable overhead incurred and the
variable overhead applied will consist of only the variable overhead spending variance. Under normal cost-
ing, overhead is applied based on the amount of the allocation base actually used for the actual units of
output rather than based on the standard amount of the allocation base allowed (as in standard costing).
Therefore, there can be no variable overhead efficiency variance under normal costing. Also under normal
costing, direct materials and direct labor costs are applied to production at their actual rates multiplied by
the actual amount of the direct inputs used in production. Therefore, no direct material or direct labor
variances can arise because the actual costs incurred are equal to the costs applied to production.
2 c - Cost of goods sold consists of cost of goods manufactured adjusted for the change in finished goods
inventory, as follows:
Beginning FG inventory + Cost of goods manufactured – Ending FG inventory = Cost of goods sold
From the information given in this question, cost of goods sold can be calculated as Sales $160,000 − Gross
Profit $48,000 = $112,000. Beginning and ending finished goods inventory are given as $60,190 and
$58,300, respectively. Therefore, using the formula above and letting X stand for cost of goods manufac-
tured, cost of goods manufactured is calculated as follows.
$60,190 + X − $58,300 = $112,000
X = $110,110.
3 b – Under variable costing, the profits that the company reports will fluctuate with the level of sales.
When sales increase, reported profits will increase. When sales decrease, reported profits will decrease.
With variable costing, manufacturing fixed costs are treated as period costs and are expensed as incurred
instead of being capitalized in inventory. Thus, cost of goods sold consists of variable costs only. Operating
income will increase by the amount of the unit contribution margin multiplied by any increase in units sold,
and it will decrease by the amount of the unit contribution margin multiplied by any decrease in units sold.
4 b – Conversion costs include direct labor and manufacturing overhead. In this question the direct labor
is $20 per unit and the manufacturing overhead is $21 per unit for a conversion cost of $41 per unit.
5 d – Prime costs include direct materials and direct labor. In this question, the direct materials and direct
labor are $32 and $20, respectively, for a prime cost per unit of $52.
6 b – Variable costs are direct materials ($32), direct labor ($20), variable manufacturing overhead ($15),
and variable selling expense ($3), for estimated total variable costs of $70 per unit.
7 c – Total estimated costs include both production and nonproduction (here, selling) costs and both vari-
able and fixed costs. Total estimated manufacturing costs include manufacturing costs for the total number
of units planned to be produced, not sold. Total estimated costs also include the fixed and variable selling
costs that are expensed as incurred, but for selling costs, the total number of units planned to be sold is
used, not the number of units planned to be produced.
Total estimated fixed manufacturing overhead must be calculated by multiplying the estimated fixed man-
ufacturing overhead per unit by the number of units planned to be produced per month. The estimated
total fixed manufacturing overhead was determined before the estimated unit cost was determined in the
planning process, and it does not change in total with changes in production level. The fixed manufacturing
overhead per unit was then determined by dividing the total estimated fixed manufacturing overhead by
the number of units planned to be produced. So to find the estimated total fixed manufacturing overhead,
multiply the estimated fixed manufacturing overhead per unit ($6) by the number of units planned to be
produced (12,000).
Estimated total costs are:
Variable manufacturing cost of $67 per unit in total × 12,000 units planned to be produced: $ 804,000
Fixed manufacturing overhead of $6 per unit × 12,000 units planned to be produced: 72,000
Variable selling expense of $3 per unit × 8,000 units sold: 24,000
Fixed selling expense (12,000 planned sales × the estimated fixed selling cost of $4/unit): 48,000
$948,000
Note that total fixed selling expense does not change if only 8,000 units are sold instead of 12,000. That
is true because fixed expenses are fixed—they do not change in total with changes in the level of activity.
8 c – When absorption costing is being used and sales are greater than production, operating income will
be lower than it would be under variable costing, because some of the fixed costs incurred in a previous
period and capitalized in inventory at that time will be expensed in cost of goods sold along with the current
period’s fixed costs. Under variable costing, only the current period’s fixed costs would have been expensed.

346 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
CMA Part 1 Answers to Questions

9 d – If the managers decrease production of any item, fewer costs will be put on the balance sheet. By
putting fewer costs on the balance sheet, more costs will be on the income statement, reducing operating
income and bonuses. All of the choices that include an increase in production will cause operating income
to be higher since some of the costs of production will be held on the balance sheet as part of unsold
finished goods inventory.
10 b – The difference between the two methods relates only to the fixed manufacturing overhead costs.
Because this was the company’s first year of operations and thus beginning inventory was zero, the amount
of the difference can be calculated as follows:
Fixed manufacturing costs per unit × Change in the level of inventory
Fixed manufacturing costs divided by 50,000 units produced equals fixed manufacturing cost per unit of
$5. The company produced 50,000 units and sold 45,000 units, so inventory increased by 5,000 units.
Thus, the increase of fixed manufacturing costs in inventory was $5 × 5,000 = $25,000. Because this
$25,000 is put into inventory under absorption costing but expensed under variable costing, operating
income under the absorption method would be $25,000 higher because cost of goods sold would be $25,000
lower than under variable costing.
11 b – Under absorption costing, the fixed manufacturing costs are allocated to the products produced.
The variable costs of production are $30 per unit and the fixed costs per unit are $3 per unit ($600,000 ÷
200,000 units produced). In total, the cost per unit is $33. Since the sales price was $40 per unit, this is a
gross profit of $7 per unit. With 120,000 units sold, gross profit is $840,000 (120,000 × $7). Subtracting
from this the selling and administrative costs of $400,000 results in operating income of $440,000.
12 a – Under variable costing, the fixed manufacturing costs are expensed. With a selling price of $40 per
unit and a variable cost of $30 per unit, the manufacturing contribution margin is $10 per unit. With sales
of 120,000 units, the total contribution margin is $1,200,000. Subtracting from this $1,200,000 the fixed
manufacturing costs of $600,000 and the selling and administrative costs of $400,000 results in operating
income of $200,000.
13 c – In this question the first thing needed is to reduce the joint cost to be allocated by the sales value
of the byproduct. The sales value of the byproduct is $120,000 (60,000 pounds × $2 selling price per
pound), reducing the joint cost to be allocated to $2,400,000. The joint cost is to be allocated based on the
physical measure (pounds) of the two products. The weight of the first main product is 90,000 pounds and
the weight of the second main product is 150,000 pounds. In total this is 240,000 pounds, and since the
second product is 150,000 pounds, the allocation to it is as follows: $2,400,000 × 150,000 ÷ 240,000 =
$1,500,000.
14 a – In order to answer this question, the total sales value of all of the produced units must be deter-
mined, then the percentage of the total sales value contributed by the MSB is calculated. That percentage
is the percentage of the $300,000 of joint costs that will be allocated to the MSB. The sales value of the
MSB is $120,000 ($2 per unit × 60,000 units), and the sales value of the CBL is $360,000 ($4 per unit ×
90,000 units). The total sales value of both products is $480,000, and the MSB is 25% of the total ($120,000
$480,000). Therefore, 25% of the joint costs, or $75,000, is allocated to the MSB.
15 c – To solve this question, use the following formula:
Units in Beginning WIP + Units Started/Transferred In = Units in Ending WIP + Units Completed
Letting X stand for the number of units started, the formula is as follows:
4,000 + X = 5,000 + 15,000
X = 16,000.
In this question, the information about the percentage complete of BWIP and EWIP is unnecessary since
the question is about physical units and not equivalent units.
16 c – In order to solve this, use the three stages of work that can be done on a unit. To finish BWIP, Ben
had to do 80% of the work on 200 units, resulting in 160 EUP to finish BWIP. 900 units (1,100 units
completed − 200 units in BWIP) were started and completed during the period. 400 units in EWIP had 80%
of the work done, or 320 EUP. Adding these three EUP figures together results in 1,380 EUP for May.
17 a – In this question the equivalent units of production need to be calculated separately for materials
and conversion costs because materials are added at the beginning of the process whereas conversion costs
are added proportionately. Since the materials are added at the beginning of the process, all units started
have had all of their materials added and the equivalent units of production for materials are equal to the
number of units started, which was 5,000.
The number of equivalent units of production must be calculated for conversion costs, using the EUP to
finish BWIP, the number of units that were started and completed, and the number of EUP to start EWIP.

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 347
Answers to Questions CMA Part 1

To finish BWIP required 800 EUP in conversion costs because the 2,000 units in BWIP were already 60%
complete, leaving 40% of the work to be done this period. 4,000 units were both started and completed
(6,000 units completed less 2,000 units in BWIP) and the equivalent units of production for units started
and completed is always equal to the number of physical units, or 4,000 EUP. To start the EWIP required
400 EUP because the 1,000 units in EWIP were 40% complete. The total number of EUP for conversion
costs was therefore 800 + 4,000 + 400, or 5,200.
Below are the calculations for total equivalent units.
Materials Conversion Costs
Complete Beginning WIP Already have been added 2,000 × 0.4 = 800
Started and Completed 4,000 4,000 × 1 = 4,000
Start Ending WIP 1,000 1,000 × 0.4 = 400
Total EUP 5,000 5,200
18 a – Direct materials EUP is very simple under FIFO when all materials are added at the beginning of the
process. The EUP is equal to the number of units started during the period, which is given as 5,000.
19 b – In order to finish BWIP, 400 EUP of conversion costs were required (40% of 1,000, because BWIP
consisted of 1,000 units that were already 60% complete). A total of 3,000 units were started and com-
pleted (given in the question). 400 EUP were required to start EWIP (20% of 2,000 units). 400 + 3,000 +
400 = 3,800 EUP of conversion costs under FIFO.
20 d – Under the weighted average method, the assumption is made that the work that was in BWIP was
done during the current period. Therefore, the EUP for materials will include the total units completed, both
those in BWIP (1,000) and those started and completed (3,000), plus the equivalent units in EWIP (2,000).
This is 6,000 in total.
21 d – Again, under the weighted average method, the work that was in BWIP as well as the work actually
done during the period are included in the calculation of total equivalent units. As a result, the calculation
of EUP has only two parts—units completed and units started that were in EWIP at the end of the period.
The number of units completed was 4,000 (1,000 in BWIP and 3,000 started and completed) and the EUP
to start EWIP was 400 (20% of 2,000). Total equivalent units are 4,000 + 400, or 4,400 EUP. The EUP
under the weighted average method will never be lower than the EUP under FIFO.
22 d – Note that the question does not say whether FIFO or WAVG is being used. However, it makes no
difference which is being used because the beginning work-in-process inventory was zero. Therefore, only
the costs incurred will be allocated between completed units and ending work-in-process since costs in
BWIP were zero.
The worksheet for calculating equivalent units of production and the cost per equivalent units is as follows:
Equivalent Units of Production:
Direct Ma- Conversion
terials Costs
Units Completed 8,000 8,000
+ Starting of EWIP 2,000 500
= Total EUP this period 10,000 8,500
Costs:
Costs in BWIP 0 0
+ Costs Added $33,000 $17,000
= Total costs to allocate
$33,000 $17,000

Total Cost
Per EUP
Costs per EUP: $3.30 $2.00 $5.30

A total of 8,000 units were completed, so the total cost of the units transferred out was $42,400 (8,000
units × $5.30).
23 d – The cost of the EWIP must be calculated separately for materials and conversion costs since the
number of equivalent units in each is different for EWIP. Using the worksheet developed for the previous
question, the cost per equivalent unit of direct materials is $3.30 and 2,000 equivalent units of direct
materials were in EWIP, for a total direct materials cost of $6,600. The cost per equivalent unit of conversion

348 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
CMA Part 1 Answers to Questions

costs is $2.00 and 500 equivalent units of conversion costs were in EWIP, for a total conversion cost of
$1,000. Thus, the total cost of the work in EWIP was $6,600 + $1,000, or $7,600.
Note that the total cost of the units transferred out ($42,400, see previous question) plus the total cost of
the work in EWIP ($7,600) equals $50,000, the total costs to be allocated. Thus, all the costs are accounted
for.
24 d – Since the company began operations in January, work-in-process inventory on January 1 was zero.
When BWIP is zero, as it was in January, the FIFO and the weighted average cost flow assumptions result
in the same total number of EUP. However, when BWIP is greater than zero, as it was in February, the
weighted average method will always result in a greater total number of EUP than will the FIFO method.
25 c – To calculate the equivalent units under the weighted average method, simply add together the
number of units completed and the number of equivalent units of production that were done in order to
start the EWIP. In this question 92,000 units were completed. The 24,000 units in EWIP are 90% complete
for materials, so 21,600 EUP of direct materials were in EWIP (24,000 × 0.90). The total EUP for direct
materials was therefore 92,000 + 21,600, or 113,600. The total costs to allocate for materials under the
weighted average method include the costs incurred during the period ($468,000) as well as the costs of
the materials in BWIP ($54,560). The total material cost is thus $522,560. The total direct material cost is
divided by the 113,600 EUP for materials to calculate the cost per equivalent unit for materials of $4.60.
26 c – To calculate the equivalent units under the weighted average method, simply add together the
number of units completed and the number of EUP that were done in order to start the EWIP. In this
question 92,000 units were completed. The 24,000 units in EWIP are 40% complete for conversion costs,
so 9,600 EUP of conversion costs were in EWIP (24,000 × 0.40). The total EUP for conversion costs was
therefore 92,000 + 9,600, or 101,600. The total conversion costs under the weighted average method
include the costs incurred during the period ($182,880 + $391,160) as well as the conversion costs in BWIP
($20,320 + $15,240). The total conversion cost is thus $609,600. The total conversion cost is divided by
the 101,600 EUP for conversion costs to calculate the cost per equivalent unit for conversion costs of $6.00.
27 d – Since Larkin just completed its first month of operations, its beginning work-in-process inventory
must have been zero. Ending WIP contained 2,000 physical units that were 25% complete as to conversion
costs, so the equivalent units in ending WIP were 500 EUP. When BWIP is zero but EWIP is greater than
zero, the total number of EUP will have to be less than the units placed into the process since not all of the
units placed into the process were completed.
28 a – The additional direct materials are added in Department 2 when the units have reached the 25%
stage of completion with respect to conversion costs. Since the beginning work-in-process is 40% complete,
the BWIP is past the 25% point with respect to conversion costs. Therefore, 100% of the materials were
already added to the beginning work-in-process inventory during the previous period, so none remained to
be added during the current period. However, as the ending work-in-process is only 20% complete (less
than 25% complete), none of the materials have been added to EWIP. As a result, the materials were added
this period only to the units that were started and completed. 85,000 units were completed during the
period, but because 15,000 units were in beginning work-in-process, only 70,000 units were started and
completed during the period. Therefore, 70,000 equivalent units for direct materials were added in De-
partment 2 for the current month.
29 d – The abnormal spoilage of $50,000 is charged directly to the income statement in this period. The
$20,000 of normal spoilage is allocated to the 50,000 good units produced. This normal spoilage will then
be on the income statement only when the units are sold. Since only half of the good units were sold, only
half of the normal spoilage, or $10,000, has been charged against revenue. Therefore, the total spoilage
costs on the income statement in May are $60,000.
30 d – The cost of goods sold includes direct materials of $13,700, direct labor of $4,800 and overhead of
$20,000 (800 hours × $25 per hour), for a total of $38,500. Allocated to the 7,000 units, the cost per unit
is $5.50. Administrative and selling costs are not part of the cost of goods sold.
31 b – The actual overhead incurred was $45,000 and $3,000 of overhead was over-applied. This means
that a total of $48,000 of overhead was applied. Given that the overhead was applied at a rate of $3 per
unit produced [(40,000 + 20,000) ÷ 20,000 units], 16,000 units must have been produced in order to
have $48,000 applied ($48,000 ÷ $3).
32 d – Cost drivers are activities, events, or factors that cause costs to be incurred. Whenever a cause-
and-effect relationship exists between the activity and the incurrence of a cost or costs, the activity is a
cost driver for the cost or costs.
33 a – If the overhead costs are allocated based on direct labor hours, the mirrors and the windows will
each receive an equal amount of overhead because both products require 200 direct labor hours per unit
and the same number of units are planned to be produced of each product. Therefore, the mirrors will

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 349
Answers to Questions CMA Part 1

receive 1/2 of the budgeted overhead, or $25,000. Since the company plans to produce 25 mirrors, the
overhead allocated per mirror is $1,000.
34 b – Under ABC, the materials handling costs would be allocated based on the number of material moves.
Each mirror requires 5 material moves and each window requires 15 material moves. Given that the budg-
eted number of mirrors and windows is the same, 25% (5 ÷ [5 + 15]) of the material moves will be for the
mirrors. Therefore, the mirrors should be allocated 25% of the budgeted overhead costs, or $12,500. This
total for mirrors is allocated to each of the 25 mirrors to be produced at a rate of $500 per mirror.
35 d – Under the direct method of allocating support costs, service department costs are not allocated to
other service departments. Thus, when the ratio used to allocate the costs to the production departments
is determined, the usage by the other service departments must be excluded. Total support costs allocated
to the Assembly department include costs allocated from both Quality Control and Maintenance. To allocate
the costs of the Quality Control department to Assembly, determine what percentage of the Quality Control
time used by the production departments (Machining and Assembly only) was used by Assembly. As-
sembly used QC 7,000 hours and Machining used QC 21,000 hours, for a total of 28,000 hours. Thus,
Assembly used 7,000 hours out of 28,000 hours total, or 25% of the total hours, so Assembly will be
allocated 25%, or $87,500, of the $350,000 in QC costs. Assembly used the Maintenance department a
total of 12,000 hours and Machining used Maintenance for 18,000 hours, for a total of 30,000 hours. As-
sembly used 12,000 hours out of the 30,000 total hours, or 40% of the hours, so Assembly will be allocated
40% of the $200,000 Maintenance costs, or $80,000. In total, therefore, the Assembly department was
allocated $87,500 + $80,000, or $167,500 in cost from the two service departments.
36 d – Again using the direct method, the support cost allocated to the Machining department needs to be
calculated. This can be done in either of two ways. The first way is the same way as it was calculated in the
previous question. The second way is to subtract the costs that were allocated to Assembly as calculated
in the previous question from the total service costs. Total service costs were $550,000 ($350,000 for QC
and $200,000 for Maintenance). Since $167,500 was allocated to Assembly, then $382,500 in total service
costs was allocated to Machining. However, the question asks for the total amount of overhead allocated
to each machine hour. Therefore, the Machining department’s own overhead costs of $400,000 need to be
included. Therefore, $782,500 in total costs ($382,500 allocated service costs + $400,000 overhead in-
curred) needs to be allocated to the 50,000 machine hours. Dividing $782,500 by 50,000 machine hours
results in an overhead application rate of $15.65 per machine hour.
37 b – The allocations begin with the Quality Control department’s costs. Some of the Quality Control costs
will first be allocated to Maintenance, the other support department. In total, Quality Control provided
35,000 hours of service and Maintenance used 7,000 of the hours, or 20%. Therefore, Maintenance will be
allocated 20% of the $350,000 in Quality Control costs, or $70,000. This $70,000 allocated to Maintenance
from Quality Control is added to the $200,000 of the Maintenance department’s own incurred costs, and so
the Maintenance department has $270,000 to allocate to the production departments. The total number of
maintenance hours used by the two production departments was 30,000 (18,000 hours by Machining and
12,000 hours by Assembly). The Assembly department used 40% of the maintenance hours that were used
by production departments (12,000 ÷ 30,000), so 40% of $270,000, or $108,000, will be allocated to the
Assembly department.
38 d – Because there are two support departments, the reciprocal method requires the use of two formulas,
each in two variables. The two formulas are based on the total costs that each support department incurred,
as follows: QC = $350,000 + 0.25M
M = $200,000 + 0.20QC
Now, the formulas can be solved for either variable by substituting the right side of one equation into the
other in place of the variable. In other words, since according to the first formula, “QC” is equal to $350,000
+ 0.25M, then “$350,000 + 0.25M” can be substituted into the second formula in place of the “QC” in that
formula. Likewise, since according to the second formula, “M” is equal to $200,000 + 0.20QC, then
“$200,000 + 0.20QC” can be substituted into the first formula in place of the “M” in that formula. The result
of either is a formula in only one variable that can be solved.
To solve for QC (because this is what the question asks for), the formula is as follows:
QC = $350,000 + 0.25($200,000 + 0.20QC)
QC = $350,000 + $50,000 + 0.05QC
0.95QC = $400,000
QC = $421,053
$421,053 is the amount of costs that the Quality Control department will allocate to the Maintenance de-
partment and the two production departments.

350 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
CMA Part 1 Answers to Questions

The total amount to be allocated by the Maintenance department is found (though the question does not
ask for it) by substituting $421,053 for the “QC” in the equation M = $200,000 + 0.20QC and solving for
“M.”
39 c – This is a very simple question, but after the previous four, it seems too easy. Since support costs
are not being allocated to the production departments, simply divide the Assembly department’s overhead
cost of $300,000 by the budgeted number of direct labor hours. $300,000 ÷ 25,000 = $12 per direct labor
hour in the Assembly department.
40 d – A service department or service center provides specialized support services to other units and
departments in the organization.
41 b – The throughput contribution per unit is the selling price minus the strictly variable costs. Since the
company does not lay off idle workers, labor is not a variable cost for this company. The only strictly variable
cost is direct materials. Thus, the throughput contribution per unit is the $200 selling price less $100 direct
materials used in Molding and less $20 direct materials used in Heat Transfer = $80 per unit. All of the
other information given is irrelevant.
42 d – The company can produce 400 units per day because that is the output of the two lowest-producing
departments (104,000 units per year divided by 260 working days per year). The throughput contribution
per unit is the $200 selling price less $100 direct materials used in Molding and less $20 direct materials
used in Heat Transfer = $80 per unit. Therefore, the throughout contribution per day is $80 × 400 units,
or $32,000.
43 c – Because direct labor is not a variable cost since the company does not lay off its employees, the
annual operating costs are:
Direct labor: 525 hours /day (225 hours in Molding + 100 hours in Heat Transfer + 200 hours in Trimming)
@ $25/hour × 260 working days/year = $3,412,500
Fixed manufacturing overhead: $2,000,000 ($1,000,000 in Molding + $750,000 in Heat Transfer +
$250,000 in Trimming)
Fixed selling and administrative expense: $1,000,000 ($428,571 + $190,476 + $380,953)
Total: $3,412,500 + $2,000,000 + $1,000,000 = $6,412,500
44 b – Production capacity is the capacity of the lowest-producing department or departments, since that
represents the maximum number of units that can be produced, even though other department(s) may be
able to produce more. When other department(s) can produce more than the lowest-producing depart-
ment(s), those departments that can produce more have excess unusable resources. Production capacity
will be maximized at the point where the production capacity of every department is the same number of
units and thus none of the departments have excess unusable resources. At that point, operating income
will be maximized.
By reassigning an employee for 5 hours per day from Molding to Heat Transfer and by reassigning employ-
ees for 10 hours per day from Molding to Trimming, the production capacities in each department will be
the same: 420 units per day and 109,200 units per year. Thus, the overall company’s capacity would
increase from 104,000 units to 109,200 units per year.
This answer was calculated by trial and error. Since this is a multiple-choice question, trial and error means
trying each possible combination to see which one results in the highest daily and annual production ca-
pacities, signified by the fact that the capacities are the same in every department. On an essay question,
trial and error would mean trying different combinations until the production capacities of the various de-
partments are the same, or as close to being the same as possible.
Here are the calculations of what the direct labor hours available, the volumes and the costs would be after
the usage of the available labor has been optimized:
Molding Heat Transfer Trimming
Direct labor @ $25/hr. 0.5 hrs. (2 units/hr.) 0.25 hrs. (4 units/hr.) 0.5 hrs. (2 units/hr.)
Direct labor hours availa-
210 hrs. 105 hrs. 210 hrs.
ble/day
Direct materials $100 per unit $20 per unit $0 per unit
Annual fixed manufactur-
$1,000,000 $750,000 $250,000
ing overhead
Fixed selling and admin.
exp. (allocated according $400,000 $200,000 $400,000
to available DL hrs.)
Daily capacity in units 420 units 420 units 420 units
Annual capacity (260
109,200 units 109,200 units 109,200 units
working days per year)

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 351
Answers to Questions CMA Part 1

Annual production & sales 109,200 units 109,200 units 109,200 units
The company can now produce 420 units per day or 109,200 units per year, compared with 400 units per
day and 104,000 units per year previously. Since the Heat Transfer and Trimming processes are labor-
intensive, no new equipment would need to be purchased, so the increase in operating income is the con-
tribution per unit (under TOC) multiplied by the increase in number of units produced. And since the
company can sell all it produces, the increase in operating income would be (109,200 − 104,000) × $80
per unit, which equals $416,000 per year.
The other options do not result in equal capacities in each department. Thus, they do not increase capacity
and operating income as much (or at all).
45 c – A company outsources a particular function or functions to an outside specialist to free up its own
resources, allowing the company to focus on its primary operations. It may be cheaper to outsource a
function, but the disadvantage can be that the company loses direct control over the function or situation.
46 d – Customer service, production, marketing and sales, and R&D are the primary activities or business
functions that add value to a company’s product or service. Information systems, infrastructure, human
resources, and materials management are activities that support the primary activities as defined in the
value chain.
47 d – Because the analysis is to include identifying where customer value can be increased, the type of
analysis the consultant most likely has been asked to perform is a value-chain analysis.
48 a – Determining what processes the company uses to convert materials, capital and labor into value-
adding products is not part of benchmarking. Rather, benchmarking involves measuring the company’s
products, practices and services against successful competitors, using the competitors’ standards as goals
for its own operations. This “best practices” process helps the company to cut costs, boost output and
achieve competitive advantage. Companies use target costing to set standard costs that will create the
desired gross profit margin for their products sold at the target price. Target costing utilizes kaizen, or
continuous improvement, to reduce costs and attain desired gross profit margins.
49 d – TQM focuses on enhancing the quality of the product or service with consistency, providing a high
level of customer service, eliminating non-value-adding work, and responding quickly and flexibly to the
shifting requirements of customers. Computer-aided design (CAD) utilizes computers in the product devel-
opment, analysis and design modification stages, but CAD is not an objective of TQM.
50 d – Internal failure occurs when a problem is detected before the product is shipped to any customer.
Internal failure costs include rework, scrap, tooling changes, downtime required to do the tooling changes,
and costs associated with rushing to complete an order because of failure to do the work properly the first
time.
51 d – Equipment preventive maintenance ($1,154) is a prevention cost and product testing equipment
($786) is an appraisal cost, for total prevention and appraisal cost of $1,940.
52 a – Segregation of duties requires that the following four functions be done by different people: (1)
authorizing a transaction; (2) recording the transaction; (3) keeping physical custody of the related asset;
and (4) periodic reconciliation of the physical asset with the recorded amount for the asset. If the individual
recording the transaction were to also compare the accounting record of the asset with the asset itself, that
person would be performing two of the above functions: recording the transaction and periodic reconciliation
of the physical asset with the recorded amount for the asset.
53 d – If the same person is recording cash receipts and also reconciling the bank account, that person
could post a cash receipt to a customer’s account, divert the cash from the bank deposit, and cover it up
by reporting an outstanding deposit or a reconciling item on the bank reconciliation.
54 b – Both timekeeping and preparation of payroll journal entries should take place outside of the payroll
department. As long as a person performing those two functions had no payroll department function, that
person could do both timekeeping and prepare journal entries without violating segregation of functional
responsibilities.
55 b – Application controls are related to the inputs, processing of the files, and outputs of specific appli-
cations.
56 a – Systems analysis consists of organizational analysis to learn about the current system’s strengths
and weaknesses; identification of users’ requirements for the new system; identifying the system require-
ments to fulfill the information needs of the users; evaluating alternative designs using cost-benefit
analysis; and preparing a systems analysis report documenting the design of the proposed system and its
specifications.
57 a – The most important segregation in computer systems is between the programmers and the opera-
tors.

352 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.
CMA Part 1 Answers to Questions

58 c – Edit programs check the validity and accuracy of input data.


59 c – Preformatted input screens present a blank field in the format the input should take.
60 d – Hash totals are a method of validating the input of data.
61 a – A completeness test would not let the processing proceed if an item is not complete.
62 c – Edit tests are an input control. They are used to check whether data have been input correctly.
Thus, they should be performed on transaction files before those files are used to update the master file in
a posting run, because it is much easier to correct errors before posting has taken place than it is after-
wards.
63 a – A hash total is a meaningless sum of numbers in a batch, such as the sum of all the employee I.D.
numbers. A hash total would detect a substituted employee time card, because the employee I.D. number
of the substituted employee would be different from the employee I.D. number of the original employee.
64 b – State-of-the-art software and hardware are not part of a sound preventive control system.
65 d – Computer generated reports can be designed to provide more specific, and sorted, information
about the errors.
66 a – Preventive controls are the most cost-effective controls.
67 d – A hot site is a backup facility with a computer system similar to the one used regularly that is fully
operational and immediately available.
68 d – The correlation coefficient measures the strength of the linear relationship between two variables:
the value of x (the independent variable) and the value of y (the dependent variable).

© 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited. 353
Answers to Questions CMA Part 1

354 © 2019 HOCK international, LLC. For personal use only by original purchaser. Resale prohibited.

Вам также может понравиться