Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
BEC
Business Environment & Concepts
SECTIONS 5000–5500
2019 (EFFECTIVE JANUARY 1)
Surgent CPA Review:
Editor-in-Chief: Liz Kolar, CPA, CGMA
Senior Editor: Susan Cox, M.Acc., CPA
This book contains material copyrighted © 1953 through 2019 by the American Institute of Certified Public
Accountants, Inc., and is used or adapted with permission.
Portions of various FASB and GASB documents, copyrighted by the Financial Accounting Foundation, 401 Merritt 7,
P.O. Box 5116, Norwalk, CT 06856-5116, are reprinted with permission. Complete copies of these documents are
available from the Financial Accounting Foundation.
Material from Uniform CPA Examination Questions and Unofficial Answers, copyright © 1976 through 2019,
American Institute of Certified Public Accountants, Inc., is used or adapted with permission.
This book is written to provide accurate and authoritative information concerning the covered topics. It is not meant
to take the place of professional advice. The content of this book has been updated to reflect relevant legislative
and governing body modifications as of January 2019.
No part of this work may be reproduced or transmitted in any form or by any means, electronic or mechanical,
including photocopying and recording, or by any information storage or retrieval system, except as may be expressly
permitted by the 1976 Copyright Act or in writing by the Publisher.
Surgent CPA Review was developed by a team of professionals who are experts in the fields of accounting, business
law, and computer science, and are also experienced teachers in CPA Review programs and continuing professional
education courses.
Surgent CPA Review expresses its sincere appreciation to the many individual candidates, as well as accounting
instructors, who took time to write to us about previous editions. The improvements in this edition are attributable
to all of these individuals. Of course, any deficiencies are the responsibilities of the editors and authors. We very
much appreciate and solicit your comments and suggestions about this year’s edition.
The editors and authors are also indebted to the American Institute of Certified Public Accountants, the Financial
Accounting Standards Board, and the Governmental Accounting Standards Board for permission to quote from their
pronouncements. In addition, the AICPA granted us permission to use material from previous Uniform CPA
Examination Questions and Answers. AICPA codification numbers are used throughout the Auditing portion of the
Review to indicate the source of materials.
We recognize the work and dedication of our team of software designers and developers. Their vision has made this
the best product possible. They contributed countless hours to deliver this package and are each fully dedicated to
helping you pass the exam. Our thanks go out to the many individuals who have made contributions to both the
software and textbook portions of the CPA Review. We extend our gratitude to our team of software testers who
ensure that you receive only the highest quality product. Finally, we express appreciation to the editorial teams who
have devoted their time to review this product. They have provided invaluable aid in the writing and production of
the Surgent CPA Review.
Paul Brown, CPA, is the Director of Technical Services for a large state CPA Society. One of Paul’s main duties is to
serve as the technical reviewer in the state society’s American Institute of Certified Public Accountants (AICPA)
Peer Review Program. The program administers approximately 450 reviews annually and oversees approximately
125 peer reviewers. Paul has previously been an instructor and author of continuing education programs, for
which he had received several outstanding discussion leader and author awards. He also served on the AICPA’s
Technical Reviewers Advisory Task Force to the Peer Review Board and serves as staff liaison to two committees
and one section at the state society level. Prior to joining the state society, Paul was an audit manager with a
regional firm in Florida. He holds a BS degree in Accounting and Finance from Florida State University.
Annhenrie Campbell, PhD, CPA, CMA, CGFM, is a professor of Accounting at California State University, Stanislaus.
A former governmental accountant, her teaching and research interests have included governmental and not-for-
profit accounting throughout her career. She completed her MBA at Humboldt State University in California and
her PhD at the University of Colorado in Boulder. After earning a CPA and CMA and starting her university career,
Dr. Campbell became a Certified Government Financial Manager at the start of the CGFM program. In addition to
numerous presentations, Dr. Campbell and her colleagues continue to publish research on accounting education
issues.
Susan Cox, M.Acc., CPA (FL) has worked in various fields of accounting for over 30 years. She began her career with
Arthur Andersen & Co. after graduating from Florida State University with a master’s degree in accounting. In
addition to her experience in public accounting, she has worked as a senior internal auditor for GTE and as a full-
time accounting instructor at the University of South Florida, where she developed and taught the first online
accounting course. Susan also has previous experience in valuation accounting and as an editor for a national CPA
review publication.
Joshua Dixon, CPA, is President of J Dixon Accounting Services, PC, a full-service tax, accounting, and business
consulting firm located in Frisco, Texas. He graduated from Metropolitan State University of Denver with dual
Bachelor of Science degrees in Accounting and Finance. Josh is a CPA licensed in both Colorado and Texas. He is a
member of the American Institute of Certified Public Accountants (AICPA) and the Texas Society of Certified Public
Accountants (TSCPA).
Tim Firch, JD, LLM, is an Adjunct Professor of accounting and business law. A graduate of the University of
California, Davis, King Hall School of Law, he is a member of the California State Bar, taxation section (inactive in
good standing). His research and teaching interests center on writing skills development for upper-division
accounting students. His writing skills courses include a major research component. He also teaches business law
along with financial, managerial, and tax accounting.
David P. Flanders, MBA, CPA, is an independent technical editor for CPA Examination reviews and tax and
accounting CPE courses. A native of Colorado, he received his MBA and undergraduate degrees from the University
of Colorado at Boulder. He is a CPA in Colorado. For over 20 years he was a corporate accountant and controller
for a variety of corporations, including subsidiaries of NYSE-listed companies. More recently he spent 14 years as a
senior editor of accounting reviews and CPE courses for some of the largest publishing companies in the United
States.
LaMarion N. Green-Hughey, CPA, MBA currently serves as Director of Finance for the DeKalb County, Georgia,
Sheriff’s Office. Previously, she served as a Field Auditor for the Office of the Comptroller of the Currency.
LaMarion received both her BBA and MBA from Georgia Southern University and holds an active CPA license in the
state of Georgia.
Janel Greiman is an Assistant Professor at the Monfort College of Business at the University of Northern Colorado.
She teaches tax at the undergraduate level as well as in the Master of Accountancy program. Ms. Greiman earned
her Master of Taxation at the Sturm College of Law at the University of Denver. She is a CPA in Colorado with
recent experience practicing in taxation. She is a member of the AICPA, the Colorado Society of Certified Public
Accountants (CSCPA), the tax section of the CSCPA, and the American Accounting Association. Her research
includes recent publication in Practical Tax Strategies and the American Journal of Business Education.
Jeff Helton, PhD, CMA, CFE, FHFMA, is an Assistant Professor of Health Care Management at Metropolitan State
University of Denver, where he teaches courses in health care finance. He has over 25 years of financial
management experience in hospitals, health plans, and government entities. Jeff is a Certified Management
Accountant, Certified Fraud Examiner, Fellow of the Healthcare Financial Management Association, and a member
of the Board of Examiners for the Healthcare Financial Management Association. He has published articles on
health care accounting and budgeting in Strategic Finance, Healthcare Financial Management, and the American
Journal of Management. Jeff has a PhD in Health Care Management from the University of Texas, an MS in Hospital
Administration from the University of Alabama at Birmingham, and a BBA from Eastern Kentucky University.
Cynthia Hollenbach, BBA, MS, CPA, CGMA, has been teaching at the collegiate level for the last 10 years. She has
also been and is currently a financial consultant to many corporate clients for over 25 years, involved with both
private and public companies. She received her BBA from Baylor University and her MS in Taxation from the
University of North Texas. Cynthia is licensed as a CPA in Texas and recently licensed as a CGMA, a new designation
by the AICPA to recognize skills as a global management accountant.
Tracy Hunt, Esq., is a founding partner of Timby Hunt, LLC. He received his Bachelor of Arts from the University of
North Carolina, Chapel Hill, and his law degree from Widener University School of Law in Wilmington,
Delaware. For the past 15 years, Mr. Hunt has concentrated his legal practice in all areas of civil litigation, including
business-related contract issues, employment, land use, and personal injury. Mr. Hunt also heads up the Wills and
Estates practice for the firm, and assists his business clients in incorporation, drafting of contracts, trademark
infringement, and personnel matters.
Liz Kolar, CPA, CGMA, has been teaching CPA Review for more than 25 years in the United States, has personally
taught more than 2,500 live sessions, and has helped thousands of candidates pass the CPA Exam. She founded
Pinnacle CPA Review and co-founded Surgent Kolar CPA Review. She is a recipient of the ASWA Business Woman of
the Year Northeast Region, Distinguished Faculty Member of the Year, and PICPA Outstanding Educator of the
Year. Liz began her career in Public Accounting with a Big Four accounting firm auditing financial service clients
after graduating from Pace University with an MBA in Public Accounting. Liz’s teaching career spans almost 30
years. She has taught at the undergraduate and graduate levels at Pace University and Seton Hall University, and is
currently a professor at Delaware Valley University.
Lola Neudecker is an Internal Auditor employed at the University of New Mexico. Lola has over 20 years of
financial accounting experience in subjects including auditing, fraud investigations, taxation, contracts and grants,
purchasing, budgeting, payroll, and banking. She has worked in public accounting, higher education, and
government. Lola graduated from the University of Texas at El Paso with a Bachelor’s degree in Business
Administration with a major in Accounting and a Master of Accountancy degree. Lola is a Certified Public
Accountant (CPA), Certified Management Accountant (CMA), Certified Financial Manager (CFM), Certified
Government Auditing Professional (CGAP), Certified Internal Auditor (CIA), Certified Fraud Examiner (CFE), and
Chartered Global Management Accountant (CGMA) and is Certified in Financial Forensics (CFF). She is a member of
the American Institute of Certified Public Accountants (AICPA), The Institute of Internal Auditors (IIA), the
Association of Certified Fraud Examiners, and the Institute of Management Accountants (IMA).
Paul E. Pierson, CPA, is Director of Professional Standards and Peer Review for the Illinois CPA Society. In this
capacity, he oversees the administration of the Peer Review Program for approximately 1,500 CPA firms in Illinois
and Iowa. Paul has served as a discussion leader at the AICPA’s Annual Peer Review Conference and Advanced
Reviewer Training Course as well as several state CPA society conferences and chapter events. He is also
responsible for monitoring the continuing professional education and licensing rules in Illinois and Iowa and
responding to member inquiries regarding those matters. Paul has served on the Technical Reviewers Advisory
Task Force to the AICPA Peer Review Board, and currently serves as staff liaison to the Peer Review Report
Acceptance and Employment Benefits Committees of the Illinois CPA Society. He also oversees the Society’s
Accounting Principles, Audit and Assurance Services, Not-for-Profit, Governmental, and Ethics Committees. Paul
graduated from Illinois State University with a BS in Accounting and was an audit manager with a large, local CPA
firm in East Peoria, Illinois, prior to joining the Society.
Darlene A. Pulliam, PhD, CPA, joined the faculty of West Texas A&M in 1997. A native of eastern New Mexico, Dr.
Pulliam received a BS in Mathematics in 1976 and an MBA in 1978 from Eastern New Mexico University and joined
the accounting firm of Peat, Marwick, Mitchell and Co. (now KPMG) in Albuquerque. After five years in public
accounting, she completed her PhD at the University of North Texas and joined the faculty of the University of
Tulsa in 1987. As a Regents Professor and McCray Professor of Business at WTAMU, her responsibilities include
teaching tax and financial accounting, and advising the students in the Master of Professional Accounting program.
Her publications include many articles appearing in Tax Advisor; the Journal of Accountancy; Practical Tax
Strategies; Oil, Gas and Energy Quarterly; CPA Journal; and the Journal of Forensic Accounting as an author or
coauthor.
Guy Schmitz, JD, LLM, MA, has practiced tax law for over 40 years. He has been a lecturer in tax law at Fontbonne
University and Washington University in St. Louis, Missouri, and is the author of articles for the Journal of the
Missouri Bar and St. Louis Bar Journal. He received his JD from Cornell Law School, where he was editor-in-chief of
the Cornell International Law Journal; received his LLM in Taxation from New York University; and served as clerk
to the U.S. Tax Court.
Ken Smith, PhD, CPA-Retired, is the President of Accounting Institute Seminars, for which he has written and
presented CPA Review seminars in many locations around the United States since 1972. He received his BS in
Accounting from the University of Montana and his PhD from the University of Texas–Austin. Ken is a professor of
accounting at Idaho State University, where he has taught since 1970. He has served as accounting department
chair, associate dean of the College of Business, and College of Business dean. He is actively involved with the
AICPA and the American Accounting Association.
John (Jack) M. Surgent, MS in Taxation, CPA, is Chairman of Surgent Professional Education, the largest provider of
tax and financial-planning seminars to CPAs in the United States. Mr. Surgent has presented over 1,800 live
seminars in the past 25 years and has been named Outstanding Discussion Leader by various professional
organizations. He has worked in the tax department of a Big Four accounting firm and was an assistant professor at
St. Joseph's University in Philadelphia. Mr. Surgent received his Bachelor's degree in accounting from Villanova
University and a Master's in Taxation from Villanova University School of Law.
Geri B. Wink, BBA, MBA, CPA, joined the faculty at Colorado State University–Pueblo in 2004 after serving over 20
years on the accounting faculty at the University of Texas at Tyler. She received both her BBA and MBA from Sam
Houston State University and holds CPA licenses in both Texas and Colorado. She maintained a tax practice for
over 20 years while in Texas. Her publications include articles and cases appearing in Oil and Gas Quarterly,
Management Accounting, The Business Review, Journal of International Society of Business Disciplines, and Journal
of International Academy of Case Studies. She also has written a book with a coauthor, Intermediate Accounting
DeMystified. She is very active in the Colorado Society of CPAs.
Section 5000 Overview of the Business Environment and Concepts Examination ........................................... 1
Skill Levels
The examination or assessment of problems, and use of judgment to
Evaluation draw conclusions.
The examination and study of the interrelationships of separate
Analysis areas in order to identify causes and find evidence to support
inferences.
Application The use or demonstration of knowledge, concepts, or techniques.
5010.05 Blueprint
The Blueprint is organized by content AREA, content GROUP, and content TOPIC. Each topic
includes one or more representative TASKS. The purpose of the Blueprint is to:
a. document the minimum level of knowledge and skills necessary for initial licensure.
b. assist candidates in preparing for the Exam by outlining the knowledge and skills that
may be tested.
c. apprise educators of the knowledge and skills candidates need to function as newly
licensed CPAs.
d. guide the development of Exam questions.
The tasks in the Blueprint are representative; they are not intended to be viewed as an all-
inclusive list of tasks that may be tested. It is important to note that the number of tasks
associated with a particular content group or topic is not indicative of the extent such
content group, topic, or related skill level will be assessed on the Exam.
Source: AICPA
5020.04 References for the Business Environment and Concepts section include the following:
1. The Committee of Sponsoring Organizations of the Treadway Commission (COSO):
a. Internal Control – Integrated Framework
b. Enterprise Risk Management – Integrating with Strategy and Performance
2. Sarbanes-Oxley Act of 2002:
a. Title III – Corporate Responsibility
b. Title IV – Enhanced Financial Disclosures
c. Title VIII – Corporate and Criminal Fraud Accountability
d. Title IX – White-Collar Crime Penalty Enhancements
e. Title XI – Corporate Fraud Accountability
3. Current business periodicals
4. Current textbooks on BEC topics
5110.01 A financial transaction control is a procedure that is developed to discover and/or prevent
errors, misappropriations, or policy noncompliance in a financial transaction process. Such a
control will aid an organization in achieving specific goals and objectives. It is an internal
control that allows for a firm’s resources to be properly used, monitored, and measured. It is
hoped that such controls will detect fraud and provide adequate protection to the
company’s assets. Internal control objectives will include such items as:
a. providing reliable financial reporting.
b. providing timely feedback related to operational and/or strategic goals.
c. providing compliance with laws and regulations.
d. assuring such transactions as prompt payment of obligations or effective collection of
receivables.
e. reducing process variation.
5110.02 Since the introduction of the Sarbanes-Oxley Act of 2002 (SOX), companies have needed to
meet certain financial reporting requirements. Per Section 404 of SOX, the management of a
company is required to have an internal control report included in the annual Exchange Act
Report. The responsibility of establishing and maintaining an adequate internal control
structure related to financial reporting rests on management. Management is also required
to assess the effectiveness of their internal control structure and procedures.
5110.03 In 2007, the Public Company Accounting Oversight Board (PCAOB) issued Auditing Standard
(AS) 5 in an attempt to provide guidelines for management when developing effective and
efficient internal control structure and procedures. This standard requires management to:
a. develop an understanding of the flow of transactions through the company so that
points where misstatement could occur can be identified.
b. examine the risks related to the outline and the effectiveness of internal controls
associated with significant financial accounts from the standpoint of the possibility of
misstatement.
c. perform a fraud risk assessment and carefully examine controls meant to prevent or
detect fraud during normal transactions as well as during the period-end financial
reporting process.
d. examine internal controls that correspond to the components of the COSO (the
Committee of Sponsoring Organizations of the Treadway Commission) framework.
e. balance the assessment steps taken to the size and complexity of the company and its
operations.
f. approach the assessment tasks related to internal control of financial reporting with
competency and objectivity, and with understanding of potential risks.
g. provide a conclusion on the firm’s adequacy of internal control of financial reporting.
5111.02 The Securities and Exchange Commission (SEC) defined controls as particular policies,
procedures, and activities meant to meet a stated objective. Controls can:
a. be automated or manual.
b. relate to reconciliations.
c. provide segregation of duties.
d. be designed to safeguard assets.
e. be designed to prevent or detect error or fraud.
5111.03 Objectives. Internal controls are designed to provide reasonable (but not absolute)
assurance that objectives are achieved and compliance to laws and regulations is obtained.
Control objectives related to financial reporting, operational efficiency, and law and
regulation compliance include the following:
a. Valid: Only transactions that are valid, authorized, and legal are processed.
b. Timely recorded/processed: Only transactions that occurred within the period are
recorded, and all necessary transactions within a given period are executed.
c. Accurately recorded: Transactions are accurately recorded free of omissions,
accounting categorization errors, and/or mathematical errors.
d. Supportable: The goods and services received and provided are recorded properly and
supported by documentation considered to be standard for normal practices.
e. Reasonable: Transactions are recorded using standard methodology to determine a
value considered to be representative of the transaction.
f. Adequate representation of rights and obligations: The assets presented on the
balance sheet fairly represent the rights of the company, and the liabilities presented
fairly represent the obligations.
g. Funded: Sufficient funds are on hand to meet current obligations.
h. Appropriateness: The transactions approved directly relate to the goals of the company
and seem reasonable when compared to other transactions or trends.
5111.04 Effective and efficient control activities should be developed throughout all levels and
functions of an organization; however, the cost of the control activity should not exceed the
cost of the risk that is potentially being avoided by that control. Care should be taken,
however, not to have so many controls that productivity is negatively impacted. As a general
rule, it is less expensive to develop control procedures as a new process is being developed.
5111.05 There are numerous types of control activities, such as the following:
a. Segregation of duties: Dividing the duties of authorization, custody, and record keeping
to multiple individuals. Such separations will reduce the risk of fraud and error.
b. Authorization: Review of appropriate transactions by designated individuals for
approval. Such activities will help assure that company polices are adhered to and any
exceptions have been reviewed by individuals with the necessary oversight.
c. Review and verification: Review of transactions for accuracy and completeness. Such an
examination will help assure that all transactions have been correctly identified and
recorded.
d. Information processing controls: The use of hash totals, sequential documents (checks,
invoices, etc.), limited access to particular files, etc. allows for the protection that
unauthorized transactions will occur.
e. Retention of records: Retention of transaction records for an appropriate period of
time. This will allow any particular transaction to be substantiated if necessary.
5111.06 Internal controls can provide reasonable assurance that the objectives of the company will
be met and that external laws and regulations will be adhered to; however, even the most
effective internal control system cannot guarantee these results.
a. The reasonable assurance model is closely tied to the cost/benefit concept. Since such
items as competition, economic conditions, and innovation are outside the control of
management, internal controls cannot ensure the attainment of company goals; it can
only provide measurement and timely feedback, thus allowing management the ability
to fine-tune specific activities.
b. Limitations related to the control process include the following:
(1) Limitations on staff size can result in the inability to properly have a segregation of
duties.
(2) The judgment of employees may be affected by lack of time and pressure to
provide a rapid response.
(3) Breakdowns in communication, training, and technology can result in controls not
performing as designed.
(4) Two or more employees could collude to alter financial information or underlying
data.
(5) Management could override various controls for illegitimate purposes.
5111.07 Two important definitions related to the internal control process are those for control
precision and control sufficiency:
a. Control precision is the alignment between a risk and the control activity designed to
mitigate that risk. In other words, a control activity that has a direct influence on the
achievement of a stated objective is considered to be more precise than one that only
has an indirect influence.
b. Control sufficiency is a group of controls with a variety of degrees of precision necessary
to achieve a control objective. For example, there would potentially be a number of
control activities such as segregation of duties, reconciliation of bank statements, and
daily deposits of receipts in order to protect all incoming receivable payments from
theft or fraud.
5111.08 Section 404 of the Sarbanes-Oxley Act of 2002 (SOX) requires that all publicly traded firms
establish internal controls related to financial reporting that are documented, tested, and
maintained. The purpose of these controls is to reduce the probabilities of corporate fraud.
In order to be in compliance with SOX 404, a company needs to:
a. develop documentation of existing internal controls and procedures associated with
financial reporting.
b. test the effectiveness of those controls and procedures.
c. provide details on any deficiencies in the controls and/or documentation.
5111.09 A top-down risk assessment (TDRA) is done in order for a company to be in compliance with
SOX 404. A TDRA is a set of steps used to identify and assess financial reporting elements,
related risks, and internal control procedures meant to limit those risks. Such steps would
include the following:
a. The identification of important financial reporting elements
b. The identification of material risks related to the important financial reporting elements
c. The identification of entity-level controls that would mitigate the risks with adequate
precision. These would include such items as the company policy and procedure
manual, controls related to the internal auditors, controls related to the period-ending
procedures, variance analysis procedures, and oversight by senior management.
d. The identification of transaction-level controls that would mitigate the risks in the
absence of precise entity-level controls
e. The analysis of the nature, extent, and timing of evidence collected about the internal
controls as part of the assessment process
5111.10 The purpose of the TDRA (top-down risk assessment) is for the company to analyze the
internal controls currently in place and to assess the effectiveness of those controls to avoid
material misstatement in the firm’s financial reporting.
a. In the TDRA, the focus of the assessment of internal controls will deal with significant
(material) accounts as defined by the magnitude of the dollar amount and the likelihood
of financial statement error.
b. The TDRA will focus on the identification and analysis of pertinent risks related to the
achievement of the company’s objectives. These objectives are hierarchical in nature
related to the entity as a whole, a particular account, an assertion, a process, and a
transaction. The higher levels are examined first in the assessment process.
(1) A company-wide control (higher level) might be that all employees have received,
read, and understood the firm’s conduct guidelines.
(2) An assertion-level control might be that ownership of inventory is recognized upon
delivery of items.
(3) A process-level control might be the separation of duties and steps required when
issuing a credit for a return and/or allowance.
c. As part of the TDRA, management will develop a list of risks related to a particular
account that would have a reasonable likelihood of material misstatement. In particular,
such risks as what problems/errors have occurred in the past, how procedures/
circumstances have changed since the problem/error, and who might have the
motivation and ability to commit fraud would be considered.
d. The existing controls meant to manage the identified risks should be examined,
including the ability of senior management to mitigate such controls. This does not
mean that such controls need to completely eliminate the possibility that fraud and/or
errors could occur; such controls need to reduce the possibility that fraud and/or errors
might occur.
e. Based upon the identification and analysis of risks and the associated internal control to
mitigate those risks, management needs to conclude whether the danger of an internal
control failure is low, medium, or high. Corrective action can then be taken as deemed
necessary. Risk is inherent in all business transactions; however, that risk can be
moderated to an acceptable level based upon a cost benefit analysis.
5112.02 COSO’s internal control framework is designed to aid firms when assessing control systems,
includes the following five components:
1. Control Environment. The foundation of a business is its people and the environment in
which it operates.
2. Risk Assessment. Organizations must set objectives to identify, analyze, and manage
their risks.
3. Control Activities. Policies and procedures are needed to make sure control objectives
are effectively carried out.
4. Information and Communication. Organizations should create and use information and
communication systems to plan, conduct, manage, evaluate, and control their
operations.
5. Monitoring. Information systems and internal control policies and procedures must be
monitored so that needed modifications can be made.
Exhibit 1:
Source: COSO
5112.03 Principles. Under COSO, the internal control environment consists of 17 principles spread
over the five major components listed below:
a. Control Environment
(1) Demonstrates commitment to integrity and ethical values
(2) Exercises oversight responsibility
(3) Establishes structure, authority, and responsibility
(4) Demonstrates commitment to competence
(5) Enforces accountability
b. Risk Assessment
(6) Specifies suitable objectives
(7) Identifies and analyzes risk
(8) Assesses fraud risk
(9) Identifies and analyzes significant change
c. Control Activities
(10) Selects and develops control activities
(11) Selects and develops general controls over technology
(12) Deploys through policies and procedures
d. Information and Communication
(13) Uses relevant information
(14) Communicates internally
(15) Communicates externally
e. Monitoring
(16) Conducts ongoing and/or separate evaluations
(17) Evaluates and communicates deficiencies
5113.02 Exercise of oversight responsibility. The board of directors is charged with the responsibility
of overseeing management’s role in establishing and maintaining the organization’s system
of internal controls. The board should challenge management, ask the tough questions, and
seek input and support from informed personnel (e.g., internal and external auditors). The
board must be apprised of:
a. risks and deficiencies of the internal control.
b. assessments of internal control deficiencies.
c. assessments of internal control effectiveness.
5113.03 Establishes structure, authority, and responsibility. An organizational structure defines lines
of authority, responsibility, and reporting. An overly complex or unclear organizational
structure may indicate more serious problems. Important aspects of organizational structure
include:
a. what the company’s overall framework is for planning, directing, and controlling its
operations.
b. how a company defines lines of authority and responsibility.
c. whether authority is centralized or decentralized.
d. how responsibility is assigned for specific tasks.
Risk Assessment
5113.06 An entity’s risk assessment process for financial reporting purposes is its identification,
analysis, and management of risks relevant to the preparation of financial statements that
are fairly presented in conformity with GAAP. Objectives should be specified for the
identification and assessment of risk in each of the four objective areas:
1. Operations objectives
a. Reflects management’s policies
b. Factors risk tolerance
c. Indicates goals for both operational and financial activities
d. Establishes a foundation for allocation of resources
2. External financial reporting objectives
a. Adheres to accounting standards
b. Scrutinizes materiality
c. Reflects organization’s activities
3. Internal reporting objectives
a. Reflects management’s decisions
b. Examines the required level of accuracy
c. Reflects organization’s activities
4. Compliance objectives
a. Complies with external laws and regulations
b. Factors risk tolerance
5113.07 Identifies and analyzes risk. Identify risks related to achievement of the objectives and
determine how to manage the risks.
a. Includes entity, subsidiary, division, operating unit, and functional levels
5113.08 Assess fraud risk. The organization should consider potential for fraud.
a. Considers various types of fraud
b. Assesses incentive and pressures
c. Assesses opportunities
d. Assesses attitudes and rationalizations
5113.09 Identifies and analyzes significant change. The organization should identify changes that
could affect the current control system.
a. Assesses changes in the external environment
b. Assesses changes in the business model
c. Assesses changes in leadership
5113.10 Types of risk: Inherent risk is the risk that exists before management takes any steps to
control the likelihood or impact of a risk. Residual risk is the risk that remains after
management reacts to the risk, such as by implementing internal controls.
5113.11 An internal control risk assessment has six steps and is performed as follows.
Step 1: Identify threats. Companies face the following types of threats:
a. Strategic (doing the wrong things)
b. Operating (doing the right things, but in the wrong way)
c. Financial (financial resources are lost, wasted, or stolen; inappropriate liabilities are
incurred)
d. Information (incorrect input data, faulty or irrelevant stored information, an unreliable
system, incorrect or misleading reports)
Step 2: Estimate risk, which is a measure of how likely an error is to occur. Errors are more
likely than fraud.
Step 3: Estimate exposure, which is a measure of the magnitude of an error. For example,
there is a small risk of an earthquake, but the exposure is enormous if it destroys the
information system.
Step 4: Identify controls. One or more controls should be identified to protect the company
from each threat. Preventive, detective, and corrective controls complement each other,
and a good internal control system employs all three.
a. A preventive control is superior to a detective one.
b. If preventive controls fail, detective controls are essential for discovering the problem.
c. Corrective controls help recover from the problem.
Step 5: Estimate costs and benefits. The following principles should govern this analysis:
a. An internal control system provides reasonable assurance that control problems do not
take place; no system can provide foolproof protection, as the cost would be
prohibitive.
b. Too many controls slow the system down and negatively affect operational efficiency.
c. The benefits of an internal control procedure must exceed its costs. Otherwise, the
procedure is not worth implementing.
d. The benefit of a control procedure is the difference between the expected loss with the
control procedure(s) and the expected loss without it.
e. Expected loss is the mathematical product of risk and exposure. (Expected loss =
Risk × Exposure)
Step 6: Respond to risk.
a. Reduce the risk. Implementing an effective system of internal controls is the most
effective way to reduce the likelihood and impact of risk.
b. Accept the risk by not acting to prevent the risk or lessen its impact.
c. Share some of the risk or transfer it to another party by outsourcing an activity,
acquiring insurance, or hedging transactions.
d. Avoid the risk by not engaging in the activity that produces the risk.
Exhibit 2:
Control Activities
5113.12 Control activities are the policies and procedures that help ensure that management
directives are carried out and necessary actions are taken to address risks that threaten the
achievement of the entity’s objectives. The organization should develop activities to mitigate
identified risks.
a. Integrates with risk assessment
b. Considers entity-specific factors
c. Determines relevant business processes
5113.13 Selects and develops general control over technology. The organization should develop
general control activities over technology.
a. Determines dependency between the use of technology in business process and
technology general controls
b. Establishes relevant technology infrastructure control activities
c. Establishes relevant security management process control activities
d. Establishes relevant technology acquisition, development, and maintenance process
control activities
5113.14 Deploys through policies and procedures. The organization should deploy control activities
with policies about what is expected and procedures to implement those policies.
a. Establishes policies and procedures to support deployment of management’s directives
b. Establishes responsibility and accountability for executing policies and procedures
c. Performs in a timely manner
d. Takes corrective action
e. Performs using competent personnel
f. Reassesses policies and procedures
5113.15 The information system relevant to financial reporting objectives, which includes the
accounting system, consists of the procedures, whether automated or manual, and records
established to initiate, record, process, and report entity transactions and to maintain
accountability for the related assets, liabilities, and equity. The organization generates and
uses relevant, quality information to support the functioning of the control system.
a. Identifies information requirements
b. Captures internal and external sources of data
c. Processes relevant data into information
d. Maintains quality throughout processing
e. Considers costs and benefits
Monitoring Activities
5113.18 Monitoring of controls is a process to assess the quality of internal control performance over
time. It involves assessing the design and operation of controls on a timely basis and taking
necessary corrective actions. The organization should evaluate whether the components of
internal control are present and functioning.
a. Considers a mix of ongoing and separate evaluations
b. Considers rate of change
c. Establishes baseline understanding
d. Uses knowledgeable personnel
e. Integrates with business processes
f. Adjusts scope and frequency
g. Objectively evaluates
5113.19 Evaluates and communicates deficiencies. The organization should communicate control
deficiencies to those responsible for taking corrective action.
a. Assesses results
b. Communicates deficiencies
c. Monitors corrective actions
5113.20 The Control Self-Assessment (CSA) framework helps improve the efficiency and
effectiveness of financial statements. The process consists of all employees taking part in the
responsibility of internal controls by lower-level employees and managers completing risk
assessment and internal control evaluations.
5115.03 Board of directors (BOD). A board of directors (BOD) is an elected or appointed committee,
required for corporations that issue publicly traded equity or debt securities, and
responsible for supervising and overseeing the activities of a corporation. The board should
include executive and non-executive people. This will bring knowledge from outside and
inside the organization. The board is able to monitor the organization’s management
regarding financial decision making. The board of directors should maintain an attitude of
independence from top management in order to help with the checks and balances on the
actions taken by the CEO (chief executive officer) and CFO (chief financial officer). The board
is usually organized into committees to:
a. approve compensation for board members and corporate officers.
b. establish and communicate a code of conduct.
c. review significant management decisions.
d. design and enforce effective internal control procedures.
e. manage internal and external audits.
5115.04 The board of directors (BOD) provides governance, guidance, and oversight for
management. The BOD has a fiduciary duty to act in the best interests of the corporation. As
fiduciaries, board members are held to a higher standard of care than would be exercised in
discharge of those people’s personal affairs. Directors may not put themselves in a position
where their interests and duties conflict with the duties that they owe to the company. If a
conflict exists, the board member should disclose it to the BOD so that it may take
appropriate and timely action. Fiduciary duties fall within three categories:
1. Duty of care. Directors and officers must use care and be diligent when making
decisions on behalf of the corporation and its shareholders. Directors and officers must
act in good faith, with the care of a reasonable person in a similar position, and with
reasonable belief their decisions are in best interest of the corporation
2. Duty of loyalty. Directors and officers must have an undivided loyalty to the corporation
and shareholders; they must put the interests of shareholders and the corporation
above their own interests.
3. Duty of due diligence. Due diligence refers to the care a reasonable person should take
before entering into an agreement or a transaction with another party; it is essentially a
way of preventing unnecessary harm to either party involved in a transaction
5115.05 Audit committee. An audit committee is an operating committee of the board of directors,
independent of the corporation. This committee should include independent directors (not
stockholders or officers) and should be chaired by someone with significant financial
reporting qualifications and experience. Typically audit committees oversee the financial
reporting process; monitor the choice of accounting policies and principles; monitor the
internal control process; appoint, compensate, and oversee the external auditors; and
receive communications and audit reports directly from the external auditors. The audit
committee periodically meets with the internal and external auditors (without management
present) to discuss the financial reporting process, the system of internal control, and any
significant issues or recommendations.
5115.06 Officers. Corporate officers are appointed by the board of directors and serve at the board’s
pleasure. Officers are agents of the corporation and have a fiduciary responsibilities similar
to those of the board.
a. The president is usually known as the chief executive officer (CEO) and is responsible for
the day-to-day management of the company. The CEO is ultimately responsible for the
development and implementation of the internal control of the company. It is expected
that this individual will set the ethical tone for the firm by providing direction and
leadership to the senior staff reporting to him/her.
b. The Sarbanes-Oxley Act of 2002, Section 302, requires that CEOs and CFOs of a
corporation include certifications that:
(1) the signing officers have reviewed the reports.
(2) the signing officers are evaluating the internal controls within 90 days and reporting
their findings.
(3) all deficiencies in internal controls are being reported.
(4) negative impacts on internal controls are being reported and corrected.
(5) the financials do not contain untrue statements or material misstatements.
(6) the financial statements present fairly the financial condition of the company.
c. The officers are not permitted to reincorporate the activities of a company to attempt
to avoid these requirements. They are also not permitted to move the activities outside
of the United States to attempt to avoid these requirements.
d. Senior staff is responsible for the specific internal control functions employed in the
various areas. Without properly implemented internal controls, members of
management have the ability to override necessary controls, enabling potential
dishonest dealings or recording of transactions.
5115.07 Internal audit: Internal auditors are required to assist management in maintaining adequate
internal controls by evaluating their effectiveness and efficiency and by promoting their
continuous improvement. The internal audit activity must evaluate the adequacy and
effectiveness of controls in regards to the:
a. reliability and integrity of financial and operational information.
b. effectiveness and efficiency of operations and programs.
c. safeguarding of assets.
d. compliance with laws, regulations, policies, procedures, and contracts.
e. potential for the occurrence of fraud and how the organization manages fraud risk.
5115.08 The internal audit function must also determine whether risk management processes are
effective; this judgment results from the internal auditor's assessment that:
a. organizational objectives support and align with the organization’s mission;
b. significant risks are identified and assessed;
c. appropriate risk responses are selected that align risks with the organization's risk
appetite; and
d. relevant risk information is captured and communicated in a timely manner across the
organization, enabling staff, management, and the board to carry out their
responsibilities.
5115.09 Both internal and external auditors continually assess the effectiveness of the internal
control system in place. These individuals would:
a. assess whether internal controls are correctly designed.
b. assess whether internal controls are properly implemented.
c. assess whether interest controls are effective.
d. make recommendations for improvements in the internal control procedures.
5115.10 In evaluating the design of the entity’s control environment, the external auditor should
consider the following elements and how they have been incorporated into the entity’s
processes:
a. Communication and enforcement of integrity and ethical values
b. Commitment to competence
c. Participation by those charged with governance
d. Management’s philosophy and operating style
e. Organizational structure
5121.02 Purpose. The COSO Enterprise Risk Management (ERM) framework takes a risk-based, rather
than a controls-based, approach. It expands on the elements of the internal control
integrated framework and is much more comprehensive. ERM assists management in
effectively dealing with uncertainty and its related risk and opportunity, thus building
stakeholder value in the entity. Management, particularly the chief executive officer, is
charged with the responsibility of finding a balance between growth and profit while using
resources in an efficient and effective manner. ERM helps ensure that reporting and
compliance laws and regulations are met, and assists in protecting the entity’s reputation.
5121.03 Objectives. The objective of the ERM framework is to achieve all the goals of the control
framework and help the organization to:
a. attain reasonable assurance that company objectives and goals are achieved and
problems and surprises are minimized,
b. continuously assess risks and identify the appropriate action to take and the resources
to allocate to overcome or mitigate risk,
c. achieve its financial and performance targets, and
d. avoid adverse publicity and damage to the entity’s reputation.
5121.04 The COSO ERM (Enterprise Risk Management) framework is designed to help an entity’s
management achieve its objectives, grouped into four overlapping categories:
1. Strategic: Ensure that high-level goals support the entity’s mission.
2. Operations: Ensure resources are used effectively and efficiently.
3. Reporting: Ensure reliable reporting.
4. Compliance: Ensure compliance with laws, rules, and regulations.
5121.05 Limitations. Even the strongest enterprise risk management (ERM) system has its
limitations, resulting from faulty human judgment in decision making. Decisions on
responding to risk and establishing controls must consider the relative costs and benefits,
and breakdowns can occur because of simple errors or mistakes. Additionally, controls can
be circumvented by collusion of two or more people, and management can also override
enterprise risk management decisions.
5122.01 Components. Enterprise risk management (ERM) components help the entity achieve its
objectives. These interrelated components need to be both present and functioning
effectively (i.e., no material weaknesses) in order to have an effective ERM system. ERM
consists of eight components:
1. Internal environment. Essentially this is the “tone at the top,” consisting of risk
management philosophy, risk appetite, integrity and ethical values, accountability, and
commitment to competence.
2. Objective setting. Management must put into place a process to formulate objectives in
order to help the company assess and respond to risks.
3. Event identification. Certain events can affect the company’s ability to implement its
strategy and achieve its objectives. Management must identify these positive and
negative events and determine whether they represent risks or opportunities.
a. It is important to remember that events represent uncertainty, may or may not
occur, are difficult to foretell, can have an unforeseeable impact on a company,
may trigger other events, and can occur individually or concurrently.
b. Managers need to understand that events may interrelate (e.g., one event might
trigger another event). Some of the more common techniques companies use to
identify potential events include using comprehensive lists of potential events,
performing an internal analysis, monitoring leading events and trigger points,
conducting workshops and interviews, performing data mining and analysis, and
analyzing business processes.
4. Risk assessment. Identified risks are evaluated from a likelihood perspective and an
impact perspective to determine how they affect the company’s ability to achieve and
manage its objectives. Both qualitative and quantitative methods are used to assess
risks.
5. Risk response. Management can choose to avoid, reduce, share, or accept risks after
careful analysis.
a. Risk avoidance. Choosing not to engage in an activity
b. Risk reduction. Implementing some compensating or mitigating control to offset
the risk of an activity
c. Risk sharing. Sharing the risk with another organization such as establishing a joint
venture
d. Risk acceptance. Assuming all of the risk because it is deemed acceptable
6. Control activities. To ensure that management’s risk responses are effectively carried
out, policies and procedures should be implemented.
5122.02 ERM cube. There is a direct relationship between objectives, which are what an entity strives
to achieve, and enterprise risk management components, which represent what is needed
to achieve them. The relationship is depicted in a three-dimensional matrix, in the form of a
cube.
Exhibit 3:
Source: COSO
5122.03 Enterprise risk management (ERM) philosophies, practices, and procedures should be
integrated into the entity’s strategic planning process to enable corporate governance and
management to benchmark their risk management practices and processes against
attainment of organizational goals and objectives. ERM can enhance and augment the role
of corporate governance by clearly defining the entity’s risk-taking and risk-avoidance
tolerances, aligning those tolerances with management’s behavior, and maintaining
accountability for deviations from those tolerances.
5122.04 COSO issued an update to the ERM framework in 2017, Enterprise Risk Management—
Integrating with Strategy and Performance, which addresses the evolution of enterprise risk
management (ERM) and the need for entities to improve their approach to managing risk to
meet the demands of an evolving business environment. The updated framework focuses on
the importance of considering risk in both the strategy-setting process and in driving
performance. Specifically, the update:
a. provides greater insight into the value of ERM when setting and carrying out strategy.
5122.05 Although management holds overall responsibility for managing risk to the entity, the board
of directors and stakeholders need to understand that using enterprise risk management
(ERM) capabilities should be an integral part of selecting and refining a strategy and gaining
a competitive edge.
a. ERM adds perspective to the strengths and weaknesses of a given strategy as conditions
change, and helps in determining whether that strategy fits within the organization’s
mission and vision.
b. In the long run, ERM assists entities in identifying factors that represent change and in
assessing how that change could affect performance, potentially resulting in the
decision to revise existing strategies.
5122.06 The revised ERM framework is designed to assist boards of directors in fulfilling their risk
oversight role, which includes the following:
a. Reviewing, challenging, and concurring with management on:
(1) proposed strategy and risk appetite
(2) alignment of strategy and business objectives with the entity’s stated mission,
vision, and core values
(3) significant business decisions, including mergers acquisitions, capital allocations,
funding, and dividend-related decisions
(4) response to significant fluctuations in entity performance or the portfolio view of
risk
(5) responses to instances of deviation from core values
b. Approving management incentives and remuneration
c. Participating in investor and stakeholder relations
5122.07 The updated ERM framework’s Executive Summary lists the following benefits that can be
achieved when entities integrate ERM throughout the organization:
a. Increased range of opportunities
b. Improved identification and management of risk entity-wide
c. Increased positive outcomes; reduced negative surprises
5122.08 Although risk is already a consideration in many strategy-setting processes, enterprise risk
management (ERM) encourages management to further consider:
a. the possibility that the entity’s chosen strategy may not align with their mission, vision,
and core values. A misaligned strategy increases the possibility that the organization
may not realize its mission and vision, or may compromise its values, even if a strategy is
successfully carried out.
b. the implications from the strategy chosen. Every alternative strategy has its own risk
profile, and developing a given strategy requires trade-offs. The board of directors and
management need to ensure that the chosen strategy works in conjunction with the
organization’s risk appetite, guiding the organization to set objectives and ultimately
allocate resources efficiently.
5122.09 The ERM framework itself consists of five interrelated components. The components are
supported by a set of 20 principles:
Component 1: Governance and Culture: Governance sets the organization’s tone,
reinforcing the importance of, and establishing oversight responsibilities for, enterprise risk
management (ERM). Culture pertains to ethical values, desired behaviors, and
understanding of risk in the entity.
1. Exercises board risk oversight: The board of directors provides oversight of the strategy
and carries out governance responsibilities to support management in achieving
strategy and business objectives.
2. Establishes operating structures: The organization establishes operating structures in
the pursuit of strategy and business objectives.
3. Defines desired culture: The organization defines the desired behaviors that
characterize the entity’s desired culture.
4. Demonstrates commitment to core values: The organization demonstrates a
commitment to the entity’s core values.
5. Attracts, develops, and retains capable individuals: The organization is committed to
building human capital in alignment with the strategy and business objectives.
Component 2: Strategy and Objective-Setting: Enterprise risk management, strategy, and
objective-setting work together in the strategic-planning process. A risk appetite is
established and aligned with strategy; business objectives put strategy into practice while
serving as a basis for identifying, assessing, and responding to risk.
6. Analyzes business context: The organization considers potential effects of business
context on risk profile.
7. Defines risk appetite: The organization defines risk appetite in the context of creating,
preserving, and realizing value.
8. Evaluates alternative strategies: The organization evaluates alternative strategies and
potential impact on risk profile.
9. Formulates business objectives: The organization considers risk while establishing the
business objectives at various levels that align and support strategy.
Component 3: Performance: Risks that may impact the achievement of strategy and
business objectives need to be identified and assessed. Risks are prioritized by severity in the
context of risk appetite. The organization then selects risk responses and takes a portfolio
view of the amount of risk it has assumed. The results of this process are reported to key risk
stakeholders.
10. Identifies risk: The organization identifies risk that impacts the performance of strategy
and business objectives.
11. Assesses severity of risk: The organization assesses the severity of risk.
12. Prioritizes risks: The organization prioritizes risks as a basis for selecting responses to
risks.
13. Implements risk responses: The organization identifies and selects risk responses.
14. Develops portfolio view: The organization develops and evaluates a portfolio view of
risk.
Component 4: Review and Revision: By reviewing entity performance, an organization can
consider how well the ERM components are functioning over time and in light of substantial
changes, and what revisions are needed.
15. Assesses substantial change: The organization identifies and assesses changes that may
substantially affect strategy and business objectives.
16. Reviews risk and performance: The organization reviews entity performance and
considers risk.
17. Pursues improvement in ERM: The organization pursues improvement of enterprise risk
management.
Component 5: Information, Communication, and Reporting: ERM requires a continual
process of obtaining and sharing necessary information, from both internal and external
sources, which flows up, down, and across the organization.
18. Leverages information systems: The organization leverages the entity’s information and
technology systems to support ERM.
19. Communicates risk information: The organization uses communication channels to
support ERM.
20. Reports on risk, culture, and performance: The organization reports on risk, culture,
and performance at multiple levels and across the entity.
5122.10 The update to the ERM framework cites several trends that will continue to have an effect
on enterprise risk management (ERM):
a. Adapting to the proliferation of data: As more and more data becomes available and
the speed at which new data can be analyzed increases, ERM will need to adapt. The
data will come from both internal and external sources, and it will be structured in new
ways. Advanced analytics and data visualization tools will need to evolve and be used to
better understand risk and its impact.
b. Leveraging artificial intelligence and automation: It is important for ERM practices to
consider the impact of current and future technologies, and to leverage their
capabilities. Previously unrecognizable relationships, trends, and patterns can provide a
rich source of information critical to managing risk.
c. Managing the cost of risk management: As ERM practices evolve, it will become
important that activities spanning risk, compliance, control, and even governance be
5131.02 Administration. The Sarbanes-Oxley Act (SOX) is administered by the Securities and
Exchange Commission (SEC), and created a new agency, the Public Company Accounting
Oversight Board (PCAOB), which is in charge of overseeing, regulating, inspecting, and
disciplining accounting firms in their roles as auditors of public companies.
a. The financial statements are required to be audited by an independent CPA firm that is
certified by the PCAOB to perform audits of corporations under the jurisdiction of the
SEC.
b. The independent auditor is hired by and interacts directly with the board of directors’
audit committee. This committee should include independent directors (not
stockholders or officers) and should be chaired by someone with significant financial
reporting qualifications and experience.
5131.03 Voluntary compliance. Nonpublic entities and not-for-profit entities are exempt from SOX
requirements. These entities may adopt SOX requirements voluntarily to achieve a quality of
internal control with a recognized standard. Plans for going public or future sale of a
company also may include SOX compliance as buyers generally are willing to pay a premium
for companies that already have SOX-quality internal controls in place.
5131.04 Title III of SOX mandates senior executives take individual responsibility (and liability) for the
accuracy and completeness of corporate financial reports, defines the interaction between
external auditors and corporate audit committees, and limits the permissible behavior of
corporate officers.
5131.05 Audit committees. The audit committee is directly responsible for the appointment,
compensation, and oversight of the work of any registered public accounting firm (firm)
employed by that public company (issuer); external CPA firms report directly to the audit
committee.
a. Procedures. Each audit committee will establish procedures for the receipt, retention,
and treatment of complaints about internal controls or auditing matters; and accepting
confidential, anonymous employee comments concerning questionable accounting or
auditing matters.
b. Members. Each member of the audit committee must be a member of the board of
directors, and must otherwise be independent. Audit committee members may not
accept any consulting, advisory, or other compensation from the issuer or be an
affiliated person of the issuer.
c. Financial expert. The audit committee should contain at least one financial expert. If an
issuer does not have an audit committee financial expert, that issuer must disclose the
reason why the role is not filled. Financial expertise includes an understanding of
generally accepted accounting principles and financial statements, experience in the
preparation or auditing of financial statements of generally comparable issuers,
experience with internal accounting controls, and understanding of audit committee
functions.
5131.06 Corporate responsibility for financial reports. The chief executive officer (CEO) and the chief
financial officer (CFO) must certify that, for certain annual and quarterly reports:
a. they have reviewed the report(s) and, based on their knowledge, the reports do not
contain any material misstatements or material omissions; and the financial statements
fairly present, in all material respects, the financial condition and results of operations
of the issuer.
b. they are responsible for establishing and maintaining internal controls to ensure that
material information is made known to the officers of the entity; the effectiveness of
the internal controls has been evaluated within 90 days prior to the report; and their
report includes their conclusions about the effectiveness of the internal controls.
c. they have disclosed to the auditors and the audit committee all significant deficiencies
in the design or operation of internal controls which could adversely affect the financial
statements, and any fraud, material or not, that involves management or other
employee with a significant role in internal controls.
The officers are not permitted to reincorporate the activities of a company to attempt to
avoid these requirements. They are also not permitted to move the activities outside of the
United States to attempt to avoid these requirements.
5131.07 Improper influence on conduct of audits. It is unlawful for any officer or director of an
issuer, or any other person acting under the direction thereof, to take any action to
fraudulently influence, coerce, manipulate, or mislead any independent public or certified
accountant engaged in the performance of an audit of the financial statements of that issuer
for the purpose of making the financial statements materially misleading.
5131.08 Forfeiture of certain bonuses and profits. If an issuer is required to prepare an accounting
restatement due to material noncompliance with any financial reporting requirement under
the securities laws, the chief executive officer (CEO) and chief financial officer (CFO) may
have to reimburse the issuer for certain bonus, incentive-based, or equity-based payments
and/or any profits realized from the sale of issuer securities during the 12-month period
following the first public issuance or filing with the Securities and Exchange Commission
(SEC).
5131.09 Title IV of the Sarbanes-Oxley Act describes the enhanced reporting requirements and
disclosures for financial transactions, and requires internal controls for assuring the accuracy
of financial reports and disclosures, as well as mandates audit and reporting controls.
5131.10 Disclosures in periodic reports. Each financial report that contains financial statements
prepared in accordance with GAAP must reflect all material correcting adjustments that have
been identified by the external auditors.
a. Off-balance sheet transactions. Annual and quarterly financial reports should disclose
all material off-balance sheet transactions, arrangements, obligations (including
contingent obligations), and other relationships of the issuer with unconsolidated
entities or other persons, that may have a material current or future effect on financial
condition, changes in financial condition, results of operations, liquidity, capital
expenditures, capital resources, or significant components of revenues or expenses.
b. Pro forma financial information. Pro forma financial information cannot contain an
untrue statement of a material fact or omit to state a material fact necessary in order to
make the pro forma financial information not misleading.
5131.11 Enhanced conflict of interest provisions. In general, personal loans to executives are
prohibited, except for certain loans that are made in the ordinary course of the consumer
credit business of such issuer. The loans must be of a type that is generally made available to
the public, on market terms or terms that are no more favorable than those offered by the
issuer to the general public.
5131.12 Directors, officers, and principal stockholders. An individual who is directly or indirectly the
beneficial owner of more than 10% of a registered equity security, or who is a director or an
officer of the issuer of such security, shall file a statement with the Securities and Exchange
Commission (SEC) within 10 days after he/she becomes such beneficial owner, director, or
officer, and before the end of the second business day following the day on which there has
been a change in such ownership.
5131.13 Management assessment of internal control. Each required annual report must contain an
internal control report that states the responsibility of management for establishing and
maintaining an adequate internal control structure and procedures for financial reporting;
and contain an assessment, as of the end of the most recent fiscal year of the issuer, of the
effectiveness of the internal control structure and procedures.
5131.14 Code of ethics for senior financial officers. Each issuer must disclose in its periodic reports if
they have adopted a code of ethics for senior financial officers; if there is no code, the issuer
must disclose the reasons. Any change in or waiver of the code of ethics for senior financial
officers must be promptly disclosed.
5131.15 Real-time issuer disclosures. Each issuer must disclose to the public, in a timely manner, any
additional information concerning material changes in the financial condition or operations
of the issuer.
5131.16 Title VIII of the Sarbanes-Oxley Act describes the criminal penalties associated with being
guilty of manipulating, destroying, or altering of financial records or otherwise interfering
with investigations. The title also provides certain protections for whistleblowers.
5131.17 Criminal penalties for altering documents. Individuals who knowingly alter, destroy,
mutilate, conceal, cover up, falsify, or make a false entry in any record, document, or
tangible object with the intent to impede, obstruct, or influence the investigation or proper
administration of any matter within the jurisdiction of the United States can be fined,
imprisoned for not more than 20 years, or both.
5131.18 Criminal penalties. Audit or review workpapers must be retained for a period of 5 years
from the end of the fiscal period in which the audit or review was concluded; violations will
result in fines, imprisonment for not more than 10 years, or both. Individuals who execute a
scheme to defraud, or who obtain, by false or fraudulent pretenses, any money or property
in connection with the purchase or sale of securities of an issuer, can be fined, imprisoned
for not more than 25 years, or both.
5131.19 Statute of limitations. Violations of security laws may be brought no later than the earlier of
two years after the discovery of the facts constituting the violation, or five years after the
violation.
5131.20 Whistle-blower protection. Companies may not discharge, demote, suspend, threaten,
harass, or in any manner discriminate against an employee that provided information or
assisted in an investigation of security law violations. An employee who alleges discharge or
other discrimination can file a complaint with the U.S. Secretary of Labor or bring action in
the appropriate U.S. district court. Compensatory damages may include reinstatement with
the same seniority, back pay with interest, and compensation for special damages, including
litigation costs.
5131.21 Penalties. The penalties prescribed in the Sarbanes-Oxley Act for the attempt and conspiracy
of an offense shall be the same as the penalties for the offense itself. This title increases
penalties as follows: mail and wire fraud carries a maximum sentence of 20 years, up from 5
years; violations of the Employee Retirement Income Security Act of 1974 carry an increased
maximum fine of $500,000 from $100,000 and an imprisonment sentence of up to 10 years,
up from 1 year.
5131.22 Certification violations. A criminal punishment is implemented for corporate officers who
fail to certify corporate financial reports filed with the Securities and Exchange Commission
(SEC). Penalties include fines of not more than $1,000,000, imprisonment for not more than
10 years, or both for certification violations; and fines of not more than $5,000,000,
imprisonment for not more than 20 years, or both for willful certification violations.
5131.23 Tampering. Individuals who alter, destroy, mutilate, or conceal records, documents, etc.,
with the intent to impair objectivity or availability of use, or otherwise obstruct, influence, or
impede any official proceeding, will be fined, imprisoned for not more than 20 years, or
both.
5131.24 Prohibit persons from serving as officers/directors. In any cease-and-desist proceeding, the
Securities and Exchange Commission (SEC) can issue an order to prohibit any person who has
violated certain security laws, rules, and regulations, from serving as an officer or director of
the issuer.
5131.25 Retaliation against informants. Any individual who knowingly retaliates (i.e., takes any
harmful action) against any person who has provided truthful information to law
enforcement can be fined, imprisoned for not more than 10 years, or both.
5132.01 The Dodd-Frank Wall Street Reform and Consumer Protection Act (Dodd-Frank) was passed
into law in 2010 and was designed to prevent the recurrence of events that caused the 2008
financial crisis. Although the law largely focuses on regulations for financial institutions,
there are several sections applicable to most U.S. public companies.
a. Proxy disclosures regarding information on the following:
(1) The board of directors’ role in risk oversight
(2) The background and qualifications of directors and nominees
(3) Legal actions involving executive officers, directors, and director nominees
(4) The company’s leadership structure
(5) Diversity considerations for nominating director candidates
b. Corporate governance provisions require:
(1) advisory votes of shareholders about executive compensation (i.e., say on pay) and
golden parachutes.
(2) compensation committee members to be independent; members must assess and
report on the independence, and any conflicts of interest, of their compensation
consultants and other advisors.
(3) additional disclosures about certain compensation matters, intended to promote
transparency and “fairness,” including pay-for-performance and the ratio between
the CEO’s total compensation and the median total compensation for all other
company employees.
(4) listed companies to implement a claw-back policy to recoup from current or former
executives any financial performance-based incentive compensation paid during a
three-year look-back period, after a company restates financial results due to a
material error, even if there was no misconduct. It also requires the Securities and
Exchange Commission (SEC) to direct the exchanges to prohibit the listing of
securities of issuers that have not developed and implemented compensation claw-
back policies.
(5) additional disclosure about whether directors and employees are permitted to
hedge any decrease in market value of the company’s stock.
c. Whistleblower provisions:
(1) Section 922 of the Dodd-Frank Act requires the SEC to pay awards to
whistleblowers who voluntarily provide original information to the SEC that leads to
the successful enforcement of a violation of the federal securities laws resulting in
monetary sanctions exceeding $1,000,000. By statute, the awards must range from
at least 10% to a maximum of 30% of the monetary sanctions collected in the case.
(2) Section 922 of the Dodd-Frank Act also prohibits retaliation by employers against
individuals who provide the SEC with information about possible securities law
violations. Unlike the Sarbanes-Oxley Act anti-retaliation provisions, the Dodd-Frank
provisions do not require an employee to first file a complaint with the Department
of Labor and instead permit retaliation claims to be initiated in federal court.
d. Empowering shareholders by:
(1) prohibiting uninstructed broker voting in director elections, say-on-pay votes, and
votes on other “significant” matters (i.e., retail investors who hold stock in street
name will not be able to rely on their brokers to vote for them without specific
instruction)
(2) authorizing the SEC to adopt proxy access rules permitting shareholders to use
company proxy solicitation materials to nominate directors. Additionally,
shareholders or groups who hold 3% of a company’s voting power for at least three
years have the right to contest 25% of a company’s board seats using the company
proxy.
(3) requiring companies to explain their board leadership structures to shareholders
e. New governmental agencies
5211.03 Economic growth: The trend or the path of economic growth measures the level of output
(potential output) that could be produced if all the resources were fully employed over time.
a. The business cycle measures cyclical deviations from the potential output.
b. If the current level of output is below trend output, there is a recessionary gap.
c. If the current level of output is above trend output, there is an inflationary gap.
d. Economic growth can occur if a society has access to more or better resources and/or
the productivity of existing resources increases, allowing the existing resources to be
used more efficiently.
(2) Surveys indicate the voters are worried about both issues and perceive rising
unemployment to be a problem. Voters are also concerned about the rate of
inflation, the rate of increase in the inflation rate, and the effect that inflation could
have on their buying power.
(3) Politicians desire to have the economy “moving in the right direction” as an election
approaches.
(4) Politicians tend to run restrictive fiscal policies early in their terms and to blame the
previous administration for current economic problems.
(5) Fiscal policy tends to become more expansionary as an election year approaches.
(6) There is little doubt that most administrations tend to pursue certain of these
policies; however, it is very difficult for any administration to “fine-tune” fiscal
policy since the executive branch does not control the major policy tools. Monetary
policy is controlled by the Federal Reserve and fiscal policy measures are enacted
by the legislative branch.
d. The insufficient aggregate expenditure model is based on the premise that a business
cycle is caused by inadequate spending. The key components of aggregate expenditure
are personal consumption (C), business investment (I), government expenditures (G),
and the net difference between exports (X) and imports (M). It is represented by the
formula:
GDP = C + I + G + (X - M)
Each of these expenditure categories is assumed to be a function of other economic
variables:
(1) The level of personal consumption depends primarily on consumer disposable
income and wealth. In addition, the demand for consumer durable goods such as
housing and automobiles is influenced by the impact of interest rate changes on
monthly payments.
(2) Business investment depends primarily on the level of interest rates and business
expectations concerning the net present value of the cash flows from potential
projects.
(3) Government expenditures are determined by the fiscal policy that is developed by
the executive and legislative branches of government.
(4) Exports are driven by consumer income and wealth in foreign nations as well as
foreigners’ tastes and preferences for foreign (in this case U.S.) goods. The demand
for imports depends primarily on the same factors as personal consumption plus
the tastes and preferences U.S. citizens have for foreign goods.
e. The accelerator model assumes that the business cycle is caused by the volatility of
investment spending.
(1) Expenditures on capital goods and investment in inventory are related to the rate of
change in GDP.
(2) It assumes that a given level of capital goods is required to produce a given level of
output. Also, firms desire a given level of inventory at a given level of demand.
(3) If the economy is operating at full capacity and there is an increase in aggregate
demand, then there would be an increase in the demand for capital goods, which
stimulates increases in aggregate demand, causing a secondary increase in the
demand for capital goods.
(4) With the increase in aggregate demand, inventory levels will fall below desired
levels and firms will increase their investment in inventory, which will cause an
increase in production and employment at the wholesale level.
5211.05 Unemployment
a. The unemployment rate measures the percentage of the labor force that is not
working. It is calculated as the number of unemployed workers divided by the number
of workers in the labor force multiplied by 100. The labor force equals all U.S. residents
16 years of age or older who are not institutionalized, who are working, or who are
unemployed and actively looking for work. Part-time workers are counted as being
employed.
b. Discouraged workers are unemployed workers who are not actively seeking
employment, and they are not considered to be part of the labor force. Discouraged
workers are also defined as workers who are able and willing to work, but who have not
made specific efforts to find work during the past four weeks.
c. Example: The data below indicates what happened in the labor market in the United
States during the period June 2009 to June 2010. The unemployment rate stayed
constant at 9.5%, but the average duration of unemployment rose more than 10 weeks
to 32.8 weeks and more than 400,000 additional workers gave up looking for work, and
became classified as discouraged workers.
June 2009 June 2010
Civilian Labor Force 154,759,000 153,741,000
Employed Workers 140,038,000 139,119,000
Unemployed Workers 14,721,000 14,623,000
Unemployment Rate 9.5% 9.5%
Average Duration of
Unemployment (weeks) 22.5 32.8
Discouraged Workers 793,000 1,207,000
b. There is a social dimension to unemployment, as the loss of a job causes problems with
self-esteem. Generally, there is also an increase in domestic violence as the
unemployment rate increases and the average time workers remain unemployed
increases. Bankruptcy filings also increase as the duration of unemployment increases.
5211.09 Inflation
a. Inflation is a sustained increase in the average level of prices and is measured by using a
fixed-weight price index. In measuring the consumer price index (CPI), consumers are
assumed to buy a specified group of goods and services—a “market basket of goods.”
The items in the market basket are “weighted” based on their relative importance in
terms of consumer spending. The rate at which the weighted-average price for these
goods increases is the inflation rate.
b. The inflation rate is determined by comparing the prices paid for the market basket of
goods in one period to that paid in another. The CPI is reported by the government
every month and is used to adjust Social Security benefits annually and to adjust income
tax brackets for inflation.
c. Biases in the CPI calculation include the following:
(1) New goods bias. New goods constantly replace old goods, and the index does not
compare both the price and quality between the old and new goods. This problem
is handled by periodically revising the market basket of goods used to calculate the
CPI. This factor creates an upward bias in the CPI.
(2) Quality change bias. The quality of many items improves each year and the
improvement must be compared to the increase in price of the good. This factor
generally creates an upward bias in the CPI.
(3) Commodity substitution bias. Consumers have a tendency to cut back on the
consumption of relatively more expensive goods and substitute relatively cheaper
goods as prices rise. This factor tends to have increases in the CPI overstate the
overall rate of inflation.
(4) Outlet substitution bias. As prices increase, there is a tendency for more people to
shop at discount stores, and in recent years, online shopping has allowed
consumers to search for lower prices for many products.
d. Other commonly reported measures of inflation include the producer price index (PPI)
that measures the rate of increase in wholesale prices and the GDP deflator that
measures the price changes for all goods and serviced included in GDP. It is used when
calculating real GDP.
by 2%. A simple formula provides an approximation of how much real income will
change:
% change in real income = % change in nominal income - % change in price level
c. Who is hurt by inflation: A key group that is hurt by unanticipated inflation is those
individuals who receive a fixed income. They see their real income fall as prices rise. This
group would include individuals who receive fixed pension payments, firms that receive
fixed lease payments, workers with fixed-step raises that do not keep pace with
inflation, minimum wage workers, and families living on welfare. A second group that is
hurt by unanticipated inflation is savers. As prices increase, the purchasing power of
accumulated savings declines. To the extent that the savings earn interest, part of the
impact of inflation can be offset. The third major group hurt by unanticipated inflation is
creditors. Ignoring interest payments, the same number of dollars will be repaid as are
borrowed. As prices go up, the purchasing power of these dollars goes down. This fact
strongly suggests that one component of nominal interest rates would be a premium
designed to compensate lenders for the risk of anticipated inflation.
d. Illustrations: Assume that a saver has $5,000 deposited in a savings account and is
earning 3% on the funds. The nominal value of the account would be $5,150 at the end
of the year. If the inflation rate was 2%, the purchasing power of that $5,150 at the end
of the year would be approximately $5,050; and if inflation ran at 4%, the purchasing
power would be approximately $4,950.
e. Illustration: If the bank lent Jane $2,000 that is to be repaid in three years and inflation
averages 5% above the loan interest rate during the period, a rough approximation of
the purchasing power of the principal amount that is repaid would be $1,700. The bank
gets repaid in “cheap dollars.”
f. Who is helped (or unaffected) by inflation: Individuals who are fortunate enough to
have flexible income may be able to be less affected by inflation. Some payments are
indexed to the consumer price index (CPI). For example, payments to Social Security
recipients are adjusted annually for increases in the CPI. Historically, union contracts
have cost-of-living adjustments (COLA) that provide partial protection from the impact
of rising prices on the purchasing power of the income received. Property owners are
often able to increase rents when inflation is high. Firms with pricing power often are
often able to pass through a portion of cost increases on to consumers in the form of
higher prices.
Unanticipated inflation provides a benefit to debtors. They are able to pay back the
principal and fixed-rate interest with “cheaper dollars.” Real income is redistributed
from the bank’s owners to the borrowers. Inflation during the 1980s helped fuel a rapid
increase in housing prices, providing a windfall capital gain to many existing
homeowners and reducing the real cost of their fixed-mortgage payments.
The federal government’s outstanding debt was $13 trillion in mid-2010. With the
government operating at a deficit, the government tends to pay off maturing bonds by
issuing new debt, thereby repaying the debt with dollars that had less purchasing power
than the dollars originally borrowed.
g. Anticipated inflation: If inflation is anticipated, some individuals and firms may be able
to reduce the impact of inflation on their purchasing power. For example, lenders can
add an inflation premium to the interest rates they charge, and even may begin to offer
variable-rate loans to shift the impact that inflation has on interest rates to the
borrower. Savers might choose to hold their funds in short-term instruments,
reinvesting at higher rates as they mature.
h. Impact of inflation on output: The impact of inflation on the nation’s level of real
output depends on the type of inflation that is occurring and its severity. Cost-push
inflation reduces the level of real output and also redistributes the resulting lower level
of real income.
Many economists suggest that having full employment and economic growth depends
on a high level of consumer spending. This spending generates profits, leads to low
unemployment, and provides an incentive for firms to expand capacity. However, at
some point the resulting level of aggregate expenditures might lead to demand-pull
inflation.
A commonly held belief is that a low level of inflation is a small price to pay if it supports
full employment and continued growth. Moderate inflation also makes it easier for a
firm to adjust real wages downward if demand for their product falls. In this case, simply
holding nominal wages constant would reduce the firm’s real wage costs. From a public
relations perspective, reducing costs in this manner would be preferable to highly visible
decreases in nominal wages.
i. Impact of inflation related to deflation: A relatively low rate of inflation serves as a
buffer against deflation if there is a decline in the level of aggregate demand.
5211.12 Deflation
a. Deflation is a sustained decline in general price levels.
b. The deflation rate is measured by comparing prices paid for a market basket of goods in
one period to that paid in another.
(2) The cost to existing borrowers would increase as well, causing an increase in
delinquencies and defaults, and place more financial stress on financial institutions.
d. There is an implicit tax on borrowers and holders of liquid assets and, as the risk-
adjusted rate of return becomes negative, it encourages investors and consumers to
hoard money rather than purchase hard assets.
e. There is a tendency for consumers to postpone consumption in anticipation of further
price declines.
money supply, everything else being equal, interest rates would rise, and this would
reduce the incentive for business investment and the spending on consumer
durables. (Leading indicators such as money supply are forecasts, not
measurements.)
(9) Interest rate spread: Increases in short-term nominal interest rates generally
indicate the monetary authorities wish to slow the level of economic activity. These
moves tend to reduce the spread between short-term and long-term interest rates
that have a negative impact on business investment. Thus, a restrictive monetary
policy is likely to cause a decline in future real GDP.
(10) Consumer expectations: Declining consumer optimism, as measured by a consumer
survey such as that conducted by the Survey Research Center at the University of
Michigan or the Conference Board, suggests declines in consumption and possibly
future declines in real GDP.
c. While none of these indicators in isolation forecasts an economic slowdown, historically
a decline in the index of leading economic indicators for three consecutive months has
provided advance notice of a decline in real GDP. It is important to note, however, that
these forecasts have not been infallible and, at times, have not provided sufficient
advance warning for policymakers to be able to respond in a timely manner.
(2) M2: A broader definition of the money supply that is equal to M1 plus
noncheckable savings deposits, small time deposits (less than $100,000), and
individual money market mutual fund balances. It consists of near-monies and
highly liquid financial assets that do not serve as a medium of exchange but are
readily convertible into currency or checkable deposits.
(3) M3: The broadest definition of the money supply that is equal to M2 plus time
deposits in excess of $100,000. This includes the most illiquid components of the
money supply. While these assets can be converted to cash, they often would not
be able to be disposed of quickly at their current market value.
b. Credit cards are a convenient way to make purchases, but these purchases are not
included as part of the money supply since when credit cards are used, the transaction
creates a short-term loan that allows the consumer to defer the payment for some
period of time.
c. Debit cards are the equivalent of an “electronic check” when funds are transferred
immediately from the purchaser’s account to the merchant’s account. Many brokerage
firms offer a product to their customers that will allow the payment for the purchases
(that is, a debit to their account) to be deferred with the account being debited on a
particular day of each month.
In our example, the money multiplier = 1 ÷ 0.10 = 10. Thus, since the original increase in
the banking system’s excess reserves increase was $90,000, as a result of the deposit
the potential growth in the money supply would equal $900,000 (10 × $90,000).
d. Illustration: Assume that the Federal Reserve wishes to expand the money supply and
purchases $10 million in government securities from Bank XYZ. These funds are
deposited in Bank XYZ’s reserve account.
Bank XYZ now has $10 million in excess reserves with the potential to increase loans by
$10M, and if the reserve requirement is 8%, then the potential increase in the money
supply for the banking system is:
Potential money creation = Excess reserves × Money multiplier
= $10M × (1 ÷ 0.08)
= $10M × 12.5
= $125M
c. Discount rate
(1) The Fed serves as the “lender of last resort” for commercial banks and thrifts, and
the discount rate is the rate the Fed charges when it lends reserves to member
institutions. By providing a loan, the Fed would be creating excess reserves that
would allow the financial intermediaries to extend additional credit.
(2) Raising the discount rate would make borrowing money from the Fed less
attractive, and it also represents a “signal” from the Fed to the banking system that
they would prefer to slow the rate of growth of the money supply.
(3) The Fed generally utilizes this tool of monetary control only sparingly.
d. Open market operations
(1) Open market operations are transactions involving buying and selling of U.S.
Treasury or federal agency securities by the Fed in the open market. They purchase
(or sell) the securities from (to) financial institutions and individuals. If purchased
from financial institutions, the funds are deposited in the institution’s reserve
account at the Fed and represent both an increase in the amount of excess reserves
in the banking system and the potential for a multiple expansion of the money
supply. Funds purchased from individuals result in a deposit of funds into a
checkable deposit that would increase the level of required reserves as well as the
amount of excess reserves and thus be less expansionary than if the securities were
purchased from a bank.
(2) If the Fed wishes to decrease bank reserves, they would sell securities in the open
market, draining excess reserves from the system.
(3) This is the most commonly used tool of monetary policy.
e. Term Asset-Backed Securities Loan Facility: The Term Asset-Backed Securities Loan
Facility (TALF) is a funding facility to help market participants meet the credit needs of
households and small businesses by supporting issuance of asset-backed securities (ABS)
that will be collateralized by loans of various types to consumers and businesses of all
sizes. Under the TALF, the Federal Reserve Bank of New York will issue up to $200 billion
on a nonrecourse basis to holders of certain newly originated AAA-rated loans.
f. Term deposit facility: Term deposits will facilitate the implementation of monetary
policy by providing a tool that allows Federal Reserve Banks to accept term deposits
from depositary institutions. These deposits are removed from the institution’s account
for the term of the deposit and thereby will drain reserves from the banking system to
the degree that they are purchased by the depository institutions.
c. When a bank borrows in the “Fed funds market” the funds are simply transferred from
one bank’s account to another’s on the books of the Fed. These loans are generally
short-term, often only overnight to allow the borrowing institution to raise their reserve
deposits to the required level.
only begin to change after the level of economic activity has been changing for some
period.
(1) Average duration of unemployment (measured in weeks)
(2) Changes in per-unit labor costs
(3) Average prime rate charged by banks
(4) Ratio of manufacturing and trade inventories to sales
(5) Outstanding volume of commercial and industrial loans
(6) Ratio of consumer credit outstanding to personal income
(7) Change in the consumer price index (CPI) for services
GDP = C + I g + G + X n
(1) Personal consumption expenditures (C) are the personal consumption expenditures
on durable consumer goods, nondurable consumer goods, and services. Primary
determinants of personal consumption are disposable income, consumer wealth,
and interest rates on consumer credit. Personal consumption accounts for
approximately two-thirds of GDP.
(2) Gross private domestic investment (Ig) are all final purchases of machinery and
equipment, all construction, all changes in inventory, and purchases of new
residential housing. This can be summarized as:
Net investment = Gross investment - Capital consumption allowance
Primary determinants of investment spending are interest rates, capacity utilization
rates, and a firm’s ability to develop projects with a positive net present value.
(3) Government purchases (G) are the spending for goods and services used in
providing government services, including spending on social capital such as
buildings and highways.
(4) Net exports (Xn) is equal to exports minus imports or (X – M). Primary determinants
for exports are similar to those for consumption, and those for exports would
include similar measures for citizens of foreign nations.
c. The income approach for calculating GDP deals with the income derived or created to
produce the output. National income (NI) is the sum of all payments to factors of
production. However, since all the expenditures noted above do not flow directly to
factors of production in the form of income, it is necessary to make certain adjustments
to (NI) to derive GDP. The process is as follows:
(1) National income (NI) = Compensation of employees + Rental income + Interest
income + Proprietor’s income + Corporate profits
(2) GDP = NI + Indirect business taxes + Capital consumption allowance + Net foreign
factor income
d. There are several other national accounts that provide useful information concerning
the economy’s performance. Key among them are the following:
(1) Net domestic product (NDP) = Gross domestic product - Capital consumption
allowance
(2) National income (NI) = NDP - Net foreign factor income - Indirect business taxes
(3) Personal income (PI) = NI - Social Security contributions - Corporate income tax -
Undistributed corporate profits + Transfer payments
(4) Disposable income (DI) = PI - Personal taxes
Nominal GDP
Real GDP =
GDP deflator
b. There are a number of price indices that are used in economic analysis, including the
consumer price index (CPI), producer price index (PPI), and GDP deflator.
(1) The CPI is used to calculate real income, to adjust Social Security payments, and as a
base for many cost-of-living adjustments (COLA).
(2) The PPI is a measure of the increase in prices paid by producers for their inputs.
(3) The GDP deflator is an index that was developed to calculate real GDP given
nominal GDP.
c. Possible shortcomings of the use of GDP as a measure of economic well-being include
the following:
(1) It does not account for non-market transactions; for example, services of
homemakers and individuals who do their own repairs.
(2) It fails to account for changing attitudes toward leisure, as reflected in the fact that
the average workweek has declined during the 1990s as legislation increased
vacation time, the number of holidays, and the amount of available leave time. It
also does not account for any “psychological income” that some workers derive
from their work.
(3) It does not account for improved product quality. This is dealt with by periodically
revising the market basket of goods and re-indexing the base year for the index
used to calculate real GDP.
(4) It does not account for activities in the underground economy. This would include
illegal activities such as the sale of illicit drugs and gambling, and fencing stolen
goods, as well as income legally earned but not reported as income to the IRS.
(5) It does not account for most negative externalities, with the primary example being
its failure to account for the economic and social costs of pollution.
e. An increase in inflation expectations will cause nominal interest rates to increase. If the
rate of inflation increases beyond some point, people tend to spend their money more
rapidly as it tends to lose its value as a medium of exchange and a store of value. At the
extreme of hyperinflation, money tends to be spent as soon as it is received.
than from within the organization. Many firms specialized, and partnership
arrangements varied according to the particular market or distribution channel.
(2) The next evolutionary step was specialization within the supply chain into aspects
of providing supply planning, collaboration, execution, and performance
management. Changing market forces required rapid response, and specialization
allowed for honing of particular competencies with an outsourcing technology
hosting for supply chain solutions.
(3) The ever-increasing pace of change and complexity and speed of supply chains
resulting from rapid price fluctuations, ever shorter product life cycles, and more
offshoring have led to a need for enhanced creativity in information sharing and
often a talent scarcity within individual organizations necessary to accomplish the
needed changes.
b. Demand schedules are simply tables of information that relate price and quantity
demanded. Demand curve is a graphical representation of the information found in the
demand schedule.
c. It is key to understand the difference between change in quantity demanded (moving
along a given demand curve) and a change in demand (where one of the demand
determinants other than the price of the good changes and causes a shift to a new
demand curve).
d. The law of demand states that there is an inverse relationship between price and
quantity demanded.
e. Firms try to forecast the demand for their products by using surveys, focus groups and
test marketing, and information systems and inventory control systems to monitor
consumer buying patterns.
5223.03 Supply
a. Supply is the quantity that producers are willing and able to offer for sale at various
alternative prices at a given moment in time.
b. Supply schedules are simply tables of information that relate price and quantity
supplied. The supply curve is nothing more than the graphical representation of the
information found in the supply schedule.
b. A change in supply is caused by a change in one of the supply determinants other than
the price of the good. Examples of supply determinants are:
(1) the price of the resources needed to produce the product.
(2) technology.
(3) taxes and subsidies.
the product at that price ceiling. On the other hand, at a price below the equilibrium
price (a price ceiling), the quantity demanded is greater than quantity supplied and
there is a shortage of the product at that price. For example, looking at the graph below,
if the price were fixed at a price (P2) that is above the equilibrium price, then there
would be a surplus (equal to s2 – d2) in the market. Alternatively, if the price ceiling were
set at a price (P1) that is below the equilibrium price, then there would be a shortage (d1
– s1) in the market.
d1 – s1 = Shortage at a price below the market equilibrium price
s2 – d2 = Surplus at a price above the market equilibrium
f. The table below provides a summary of the impact on equilibrium price and quantity
demanded for the various situations where there are changes in both demand and
supply at the same time.
Effect on
Change in Change in Effect on
Equilibrium
Supply Demand Equilibrium Price
Quantity
1 Increase Decrease Decrease Indeterminate
2 Decrease Increase Increase Indeterminate
3 Increase Increase Indeterminate Increase
4 Decrease Decrease Indeterminate Decrease
%Q∆ d
Ed =
%P∆
d. The law of demand states that price and quantity move in opposite directions. In other
words, as price increases, quantity demanded falls. By convention, economists use the
absolute value of the measure rather than using the minus sign when stating price
elasticity of demand.
e. A problem arises depending upon which value is used as the base value when
performing the calculation. The following situation could arise. If the base price is $10
and the price increases to $12, the price has increased by 20%. If the base were $12 and
the price fell to $10, then the percentage change would be 16.67%. To solve this
problem, the formula is modified to use a mid-point and is calculated as follows:
(Q0 − Q1 ) ÷ ((Q0 + Q1 ) ÷ 2)
Ed =
(P0 − P1 ) ÷ ((P0 + P1 ) ÷ 2)
QO = Original quantity
Q1 = New quantity
Po = Original price
P1 = New price
f. Summary of price elasticity of demand (Ed):
Condition
Elastic Ed > 1
Inelastic Ed < 1
Unit elastic Ed = 1
Perfectly elastic Ed = ∞
Perfectly inelastic Ed = 0
5223.08 Price elasticity and total revenue (the total revenue rule)
Firms are interested in being able to estimate their revenue from sales. The concept of
elasticity provides one method to facilitate making such measurement. In the simplest case,
revenue is defined as the price of the product times the quantity sold.
a. If demand is elastic at the current price, then reducing the price will increase the total
revenue and vice versa.
b. If demand is inelastic at the current price, then increasing the price will increase total
revenue and vice versa.
c. If demand is unitarily elastic, then changing the price will have no impact on total
revenue.
This set of relationships is known as the “total revenue rule” and allows a firm to determine
the general impact that a price change would have on the firm’s revenue if the product’s
elasticity of demand can be estimated.
h. If EPy > 0, then the goods are substitutes. If EPy < 0, then the goods are complements. A
summary of elasticity:
Condition
Complements Epy < 0
Substitutes Epy > 0
Unrelated Epy = 0
the demand for the other good rises. The cross-elasticity of demand of complementary
goods is negative since as the price of product A declines, the more of product B will be
purchased at any price. For example, if the price of hot dogs falls, consumers will
purchase more hot dogs, and they will also purchase more hot dog buns, a
complementary good.
b. Inferior goods are goods or services whose demand decreases as a consumer’s income
increases, assuming prices remain constant. The “inferiority” being described has
nothing to do with the actual quality of the item but with the consumer’s ability to
purchase needed items. For example, when consumers have limited incomes, canned
vegetables will often be purchased rather than fresh vegetables; however, as incomes
rise, less canned vegetables will be purchased as the consumer substitutes fresh
vegetables for the “inferior product” that was purchased only because of the
consumer’s limited income. Inferior goods have a negative income elasticity of demand.
c. Normal goods are good or services whose demand increases as consumer’s income
increases, assuming prices remain constant. Normal goods have a positive income
elasticity of demand.
d. Substitute goods are goods and services that can be used in place of each other in at
least some of their uses. When the price of one good falls, the quantity demanded of
that good increases and the demand for the other good without a price decrease falls. If
goods are perfect substitutes, the consumer is indifferent as to which product to
purchase and will purchase the one with the lower price. If goods are imperfect
substitutes, the purchase decision would be based on the price change of one good
relative to trade-offs being made such as differences in quality between the two goods.
For example, a consumer may believe a Ford truck is of better quality that a Toyota
truck, but at some difference in price, some consumers would substitute the Toyota
truck. A substitute good is the opposite of a complementary good.
e. Law of diminishing marginal utility is a proposition that describes how an individual’s
satisfaction (or marginal utility) derived from the consumption of each additional unit of
the good or service declines as consumption of that good or service increases. For
example, a guest may desire a piece of pie at the end of dinner on Thursday evening and
derives a level of “satisfaction” from consuming it. The host offers the guest a second
piece of pie and the “satisfaction” the guest receives from consuming that piece of pie is
less than that received from consuming the first.
If marginal revenue is greater than marginal costs, then producing more will increase profits.
However, if marginal revenue is less than marginal cost, then producing more will lower
profit. Thus, profit is at a maximum when marginal revenue equals marginal cost: MR = MC.
c. The market equilibrium occurs where price equals marginal cost at the minimum point
on the average cost curve. This means that in equilibrium, a firm is operating at its most
efficient level of output. The following graph represents the equilibrium position in a
perfectly competitive industry. Marginal revenue equals marginal cost as minimum
average cost. At this point, the firm is making only a normal profit.
PO = Price
Q0 = Quantity demanded
D0 = Demand
MR0 = Marginal revenue
MC = Marginal cost
ATC = Average total cost
P = Price
Q = Quantity demanded
D = Demand
MC = Marginal cost
MR = Marginal revenue
ATC = Average total cost
5225.03 Oligopoly
a. Oligopoly is a market structure characterized by a few firms that sell either a
standardized or differentiated product and where entry into the industry is difficult.
b. There are a number of assumptions for an oligopoly:
(1) There are a small number of relatively large firms.
(2) Firms may produce either a standardized or differentiated product.
(3) Firms in the industry are interdependent.
(4) Firms tend to engage in non-price competition.
c. In oligopolies:
(1) firms face a downward-sloping demand curve and have some degree of control
over price.
(2) there generally are barriers to the entry for new firms into the industry, including
economies of scale (that is, the firm must be large to be efficient), ownership of raw
materials, patents, and brand image.
(3) firms can benefit from collusion. A cartel is one form of collusion where producers
create formal agreements specifying how much each member will produce and
charge. OPEC is an example of a cartel. Such collusion is illegal in the United States.
d. A kinked demand curve is one type of demand curve that a firm operating in an
oligopoly might face. One strategic assumption is that competitors will follow a price
reduction but will not follow a price increase. Under these circumstances prices would
tend to remain relatively stable.
(1) A firm will not gain market share by lowering prices and thus revenue would fall.
(2) A firm would lose market share if it raises prices, and thus revenue would fall if the
firm were operating on the elastic portion of its demand curve.
(3) If competitors fail to understand this logic, there is a strong possibility of a price
war as firms engage in successive rounds of price cutting in an attempt to maintain
market share.
e. If a firm in an oligopoly faces a kinked demand curve, demand is relatively more elastic
in response to a price increase and more inelastic if the firm reduces price. The theory
indicates that in the range of costs between points A and B on the graph, the firm would
not change price or output even though marginal costs increased.
P = Price
q = Quantity
MR = Marginal revenue
d = Demand
MC = Marginal cost
f. Since firms in an oligopoly prefer to engage in non-price competition, advertising wars
often occur. Product improvements and successful advertising campaigns cannot be
easily replicated by the competition. Since many oligopolies have sufficient resources to
finance advertising, an increase in advertising expenditures on the part of one firm may
force competitors to respond or lose market share. This would be one example of a
barrier to entry for competitors.
This often is akin to the inverse of a price war in that, in the end, advertising is self-
canceling, and firms have increased cost without increasing revenue.
Price leadership is another form of competition in oligopoly. It often results in an
implicit agreement that coordinates prices without having the firms engage in illegal
collusion. Basic tactics of leadership include the following:
(1) Infrequent price changes. The firm does not respond immediately to day-to-day
fluctuations in demand. Price tends to change only when costs have increased
significantly (an industry-wide wage increase or increases in energy costs).
(2) Potential impending price changes are usually “announced” by the price leader.
(3) The goal is not to maximize short-term profit but rather to discourage the entry of
new firms into the industry and protect long-term profitability. The price leader
often has a cost advantage over other firms in the industry and can still be
profitable at a price level that smaller, higher-cost firms could not.
5225.04 Monopoly
a. In a monopoly, there is a single seller of a product for which there are no close
substitutes. The firm has considerable control over price and is a price maker.
b. A monopolist:
(1) has strong barriers to entry that keep competitors out of the industry. These
barriers might include economies of scale, legal barriers such as patents and
licenses, advertising or competitive pricing strategies, and control of existing
resources.
(2) faces a downward-sloping demand curve with marginal revenue less than price.
(3) may have its monopoly position derived from government action. In this instance,
the firm may go to great limits to maintain that monopoly by engaging in rent-
seeking activities such as lobbying legislators.
(4) will tend to set price in the elastic range of their demand curve.
(5) tends to have less incentive to innovate and improve efficiency than firms in
monopolistically competitive or oligopolistic industries.
A public utility is a common example of a regulated monopoly.
c. Natural monopolies arise when economies of scale extend beyond the market’s size and
total costs are minimized by having only one firm in the industry. Conditions often exist
for local public utilities where competition is impractical.
(1) Often the government regulates the industry with a key form of regulation being
the setting of prices that can be charged.
(2) In some instances in the United States, natural monopolies have been moved to
private ownership. For example, the U.S. Postal Service has been privatized.
(3) Some regulation (for example, early regulation of the airline industry) was designed
to prevent creation of a natural monopoly.
5225.05 Consumer surplus and producer surplus: The demand curve and supply curve show what
consumers and producers are willing and able to buy or offer for sale at various alternative
prices at a given moment in time. The interaction of buyers and sellers in the marketplace
determines an equilibrium price. If a consumer is able to buy the product at the equilibrium
price, and that price is below what the consumer would be willing to pay for the product, a
“consumer surplus” exists. The same would be true for producers who are able to sell their
product above the price they would have been willing to use. This difference is known as a
“producer surplus.” The graph below demonstrates the degree of consumer and producer
surplus at a market equilibrium.
p = Price
q = Quantity
D = Demand
S = Supply
b. Companies that are successful in creating sustainable competitive advantages are likely
to have above-average profits.
c. A variation on the differentiation strategy is for a firm to make a product that requires
replacement parts that are slightly different than standard parts available made by
other companies. Once the product is sold, the consumer is “locked into” purchasing the
specific replacement part produced by the original firm. Often these replacement parts
have high mark-ups.
5227.01 Globalization
a. Globalization involves an ongoing process by which national and regional economies,
societies, and cultures become integrated through global networks of communication
and trade, which has resulted in growing connectivity and interdependence between
businesses in the broader world market.
b. Globalization often specifically relates to economic globalization: the integration of
national economies into the international economy through trade, foreign direct
investment, capital flow migration, and the spread of technology.
c. Globalization is usually recognized as a process being driven by a combination of
economic, technological, socio-cultural, political, and ecological factors, as well as the
circulation of ideas, languages, and popular culture through acculturation, or adoption
of the behavior patterns of the surrounding culture.
d. The term globalization is used by some to talk of global integration through trade and
investment flows as well and the movement of goods and services in order to enhance
international competitiveness.
e. At the most basic level, globalization is an extension of outsourcing at an international
level, which simply refers to a company that contracts with another company to provide
services that might otherwise be performed by in-house employees. The first wave of
outsourcing took advantage of labor arbitrage resulting from significant wage
differentials, and saw the outsourcing of jobs such as call center, e-mail services, and
payroll. A second wave of globalization was in manufacturing related to the dispersion
of lean production methods and flexible computer-aided manufacturing systems. A third
wave that encompasses moving to value-added services such as market research, IT,
and product inspection is now underway.
(a) During this period, Britain engaged in a forced drug trade with China.
(b) The United States “forced open” Japanese markets with gunboat diplomacy.
(2) There was voluntary trans-ocean mass migration of peoples.
(3) Globalization was sustained by new and faster means of transportation and
communication.
(4) The period ended with German colonies being taken by other world powers under
the Treaty of Versailles. By the late 1920s there was a negative period, with
retrenchment from globalization involving shrinking world trade, national
abandonment of the gold standard, the growth of state action versus market
expansion, and national and ethnic clashes.
c. The third wave of globalization followed World War II, with the costs of communication
and transportation again declining rapidly.
(1) The main thrust was political, with the ongoing struggle between the United States
and the USSR.
(2) Cold War globalization peaked in the 1970s to mid-1980s.
d. Today’s globalization is characterized by a massive increase in trade between emerging
nations such that there is now a new “Silk Road.”
(1) Historically, the old Silk Road was a set of trade routes that crisscrossed Asia some
2,000 years ago, linking merchants in China with India, Arabia, and the Middle East.
(2) The new Silk Road connects countries in Asia, Latin America, the Middle East, and
Africa, with current data indicating trade on the order of $2.8 trillion among and
between these regions.
e. A relatively peaceful world that induces economic agents to exploit the opportunities
presented by globalization
f. The main effects of rapid technological change included the following:
(1) Disaggregation of production, integrating manufacturing decisions including
activities such as CAD-CAM (computer-aided design and computer-aided
manufacturing) and using activity-based costing
(2) Dispersion of lean production methods, which call for the use of limited resources,
in comparison to the more traditional mass production methods with a
combination of innovations within the productive organization, supply chain
management, product development, and customer interaction
(3) Migration of labor-intensive standard production for components, sub-assemblies,
and finished goods production to low-wage countries
(4) The “brain drain” or immigration of highly educated and skilled labor to countries
with advanced IT
(5) When immigration limited this migration, then technology (e.g., “research centers”)
began to migrate to the location of skilled workers.
g. The use of flexible computer-aided manufacturing systems has allowed the production
of custom output where systems combine the low costs of mass production processes
with the flexibility to meet the tastes and preferences of ever smaller groups, i.e., mass
customization.
h. There was a shift that caused economies of scope to replace economies of scale as the
focus for reducing costs for large-scale production.
(1) Economies of scope states that the average cost of production declines as the result
of increasing the number of different goods produced and can be achieved not only
by a single firm that produces a variety of related goods and services, but also by a
group of firms located very near each other in a district that produces related goods
and services.
(2) Firms take advantage of the available inputs rather than specializing in the
production and delivery of a single product. Sharing or joint utilization of inputs
among similar products is the main reason for economies of scope leading to lower
costs.
i. The underlying strength of many emerging economies is based on an undervalued
currency, which stimulates exports, low-cost labor, high savings rates, and strong
investment in infrastructure.
j. Two key socioeconomic elements have enhanced globalization and the rise of emerging
market economies which has reached a tipping point in their favor:
(1) There was significant labor force growth coupled with declining birth rates in
virtually every emerging market economy. The result is a declining dependency
ratio, which is arbitrarily defined as the ratio of the elderly (those 65 years and
over) plus the young (those under 15 years of age) to the population in the
“working ages” (those 15–64 years of age). Declining dependency ratios have
historically been associated with dynamic positive economic change in the now-
developed economies.
(a) Over the next 40 years, Japan and Europe will see their population fall by 30
million and 37 million respectively, with the over-65 population climbing from
the current 1 in 5 to 1 in 3 respectively by 2060.
(b) China’s population will continue increasing for the next 15 years, but given
current low birth rates, will result in there being 100 million fewer workers
than today by 2050, with the over-80 population rising from a current 1.4% of
the population to 7.2% by that date.
(c) India’s population will increase by 300 million in the next 40 years, and the U.S.
population will increase slightly during the same period, primarily due to
immigration.
(2) The largest urban migration in history, with nearly 1-1/2 million people moving to
cities each week in the emerging economies, led to an economic output gain
enhanced by rapid labor productivity growth supported by technological
enhancements.
(1) Labor arbitrage opportunities that allow short-term wage savings will shift
geographically over time.
(2) With globalization creating a higher probability of financial shocks, managers will
need to become more adept at risk management. (Risk management is the process
of determining the maximum allowable overall level of risk a firm is willing to
accept, using risk assessment techniques to determine the original level of risk, and
developing a strategy to control individual risks such that the firm does not engage
in excessively risky behavior.)
e. Scenario planning causes a firm to shift focus from strategic planning to the analysis of
potential future events and related outcomes. For example, a firm might forecast the
possible economic outcomes if the economy was growing rapidly, moderately, or slowly.
When dealing with potential scenarios involving globalization, firms might do scenario
planning for situations such as the following:
(1) Rapid shifts in currency or commodity prices
(2) Accepting that emerging economies are now becoming “drivers” in consumption as
well as their current position as “drivers in production”
(3) Shifts in currency prices that will make purchasing power parity more likely as the
relative prices of goods and services in various countries change (Purchasing power
parity is the economic theory that indicates that the long-term equilibrium
exchange rate for two currencies will be that which equalizes their purchasing
power for goods and services.)
(4) The fact that sophisticated labor arbitrage will replace the earlier “easy pickings” of
simply moving to a low-labor-cost country, which will require firms to:
(a) build better products, find employees, deliver services, and reach consumers
efficiently
(b) capture a winning position in global supply chains by creating chokepoint
technologies on which larger industries depend and are not able to obtain
elsewhere (An example of a choke point would be the ability to control unlike
sales channels that would allow a firm to control distribution and/or that
influences customer relationships.)
(c) find business applications that use scenario planning that conceptualizes a
changing mindset about the factors related to the outside world prior to
formulating specific strategies. (Scenario planning involves aspects of systems
thinking, specifically the recognition that many factors may combine in
complex ways to create sometimes surprising futures due to nonlinear
feedback loops.)
f. The rebalancing that gives more power to emerging economies will require a significant
shift in strategic focus. Emerging markets become the locus of growth in consumption,
production, and most of all innovation.
g. More and more, global leadership will depend on winning in the emerging markets first
as:
(1) more than 70 million people per year are crossing the threshold to middle class in
the emerging markets.
(2) companies that can reduce cost structures by 20% or 30% of developed world levels
or lower will be able to capture the new demand.
(3) the business model must be reinvented to include factors such as:
5227.07 Tariffs
Tariffs are essentially taxes or duties placed on an imported good or service by a domestic
government, making domestic goods cheaper and imported goods more expensive. There
are a number of negative effects from import tariffs:
a. Tariffs protect inefficient domestic industries at the expense of efficient industries,
creating an internal misallocation of resources. In other words, land, labor, and capital
will flow toward inefficient industries that are subsidized by protectionist policies and
thus not as readily available to those industries that are competitive.
b. The increased cost of imported products can lead to an increase in domestic inflation.
c. Domestic producers will take advantage of the situation and raise their prices due to
less competitive pressure to keep their prices low.
d. Tariffs decrease the amount of imported goods, thus sending less cash to the exporting
nation. The end result is that the exporting nation will have less money to spend and
there will be a decrease in exports of domestically produced goods.
(a) China has a very robust supplier network that can easily serve other Asian
nations and has a shipping infrastructure that is able to move products
throughout the region.
(b) There is a growing consideration toward developing a “common market” with
investment in rail infrastructure throughout the Asian region to allow these
nations to become more competitive with China.
b. One of the outcomes of the recent globalization has been a shift in the pattern of
foreign direct investment (FDI), with funds now increasingly flowing from emerging
market economies to the more developed economies. This represents an
unprecedented shift from the traditional patterns of FDI in the post-World War II era,
which saw capital flow from slower-growing rich economies to emerging economies.
(Foreign direct investment (FDI) is long-term participation by one county in another in
such areas as management, joint ventures, transfers of technology, and business
expertise.)
c. Emerging market economies are capturing an increasing amount of foreign exchange
earnings, either from the sale of raw materials or rising trade surpluses.
d. Emerging market economies tend to have the following:
(1) Low debt-to-GDP ratios
(2) Large currency reserves and rising trade surpluses
(3) Low-cost labor
(4) High savings rates
(5) Investment in infrastructure
e. In addition, the level of trade among and between emerging market economies is
increasing significantly.
f. Developed economies will need to generate strong productivity gains in order to have
reasonable levels of economic growth.
g. The changing pattern of trade flows are dominated by transnational corporations,
where vertical specialization has caused a marked increase in the use of imported
components in the production process that have been exported from emerging
economies.
(1) A key driver is expected to be that commodity prices are likely to remain relatively
strong due to the growth of emerging economies and the fact that they have a very
high-income elasticity of demand for commodities. (Income elasticity of demand is
the measurement of the responsiveness of the proportional change in the demand
of a good to the change in the income of the people desiring the good. It is
calculated as the ratio of the percent change in demand to the percent change in
income.)
(2) The surging price for commodities has shifted the terms-of-trade (the relative price
of exports compared to the price of imports) in favor of any country exporting from
extractive and commodity-based industries.
b. A key determinant of the export earnings is generally a weak currency that makes a
country’s goods and services lower priced in terms of the foreign currency.
c. Developed economies are running fiscal policy deficits and borrowing funds from
emerging economies to finance them. In addition, developed economies are increasingly
running current account deficits as their consumers desire products produced by
emerging market economies. To finance these twin deficits, they borrow funds from the
emerging economies.
d. Central Banks in emerging economies often do not have the policy tools necessary to
absorb or sterilize all the domestic currency created to support the trade surplus.
(1) The classic form of sterilizing the excess monetary base is issuing local currency
bonds, and Central Banks in many emerging economies do not have that tool
available.
(2) High savings rates and a lack of well-functioning capital markets, coupled with
insufficient local investment opportunities, will create a pool of funds available for
investment.
(3) Many Central Banks are now accepting a myriad of currencies in addition to dollars,
which compounds the foreign exchange risk problems they are dealing with.
those who are critical suggest that due to lack of enforceable property rights,
small farmers in these nations will be forced off land they have utilized for
years.
(b) These governments find investing in their own domestic agriculture to be
problematic due to a scarcity of arable land and, more importantly, often a
shortage of water.
(c) The governments use the SWFs to buy or lease farmland abroad, thus
outsourcing food production, growing crops abroad, and shipping them back to
the home country.
(d) The investment is designed to prevent wide price swings from impacting their
economies.
(i) The governments could afford the higher prices, but cannot afford the
possibility of trade bans when crops would be in short supply worldwide.
This would curtail their ability to import such crops. Food security is the
primary driver.
(ii) Several countries, including Pakistan, which is the largest recipient of food
aid, are letting investors export a significant portion of their agricultural
output to help reverse a long-term underinvestment in agriculture and to
obtain increasing amounts of hard currency.
c. China is investing to acquire technologies, brands, resources, and better access to
international markets as they use technology to enhance productivity and augment
earnings. China uses two primary organizational structures for making FDI. The first is
the Chinese Investment Corporation (CIC), which is generally seen as an arm of the
government, and the second are state-owned enterprises (SOEs).
(1) The primary goal of CIC investment is to diversify the use of foreign exchange
reserves. The Chinese government is concerned that the nature of their
investments will appear to others as if the government is attempting to control the
management of target companies and/or purchase politically sensitive raw
materials.
(2) SOEs, on the other hand, have a goal of “going global” and establishing a Chinese
presence. They are attracted by the low valuations of increasingly desperate foreign
operations (particularly in strategically important sectors such as energy and raw
materials) attempting to achieve energy security and access to strategic materials at
a known contract price.
(3) To facilitate their investments, the Chinese government has eased and
decentralized regulatory procedures and made financing more available for such
ventures, while also making no-strings-attached loans to governments rather than
meddling directly in their affairs, such as in the case with World Bank lending.
(4) There is an attempt to learn how U.S. companies operate and transfer the
information to improve the operation of SOEs.
(5) A significant portion of Chinese FDI involves investing in Hong Kong, with the funds
then immediately reinvested in China, where these investments receive special
treatment as “FDI.”
(4) the historic subsidies of fossil fuel energy and support for various defense
industries.
g. Many countries are attempting to capture an increasing portion of linkages in the supply
chain, offering tax holidays, investing in capital grants, and providing free infrastructure
construction.
5231.02 Financial risk management is one component of the concept of enterprise risk management
of the firm, which includes items such as the following:
a. Business risk: the uncertainty associated with the ability to forecast EBIT due to factors
such as sales variability and operating leverage
b. Operations risk: the risk of loss resulting from inadequate or failed internal processes,
people, and systems, or from external events
c. Supply-chain risk: the potential disruptions to continued manufacturing production and
thereby commercial financial exposure
d. Product liability risk: the responsibility of the firm or vendor of goods to compensate
for injury caused by defective merchandise that it has provided for sale
e. Political and economic risk: the risk that a government buyer or a country prevents a
transaction from being completed, or fails to meets its payment obligations; and/or the
risk associated with the overall health of the economy
e. Default risk: the risk that the counterparty will default on clearing obligations. This risk
is often related to the liquidity risk and/or credit risk inherent in the transaction.
f. Systemic risk: the risk that payment system failures will lead to one market participant
being unable to meet its obligation when due that will lead to additional participants
being unable to meet commitments as well
g. Counterparty risk: the inability of a counterparty to meet its commitments
5231.04 The risk management process includes the strategies that allow a firm to manage its risks
related to participating in financial markets.
5231.06 The risk management process includes an understanding the requirements for valuation of
derivative securities and applicable accounting regulations per FASB ASC 815 or IAS 39.
Often a firm will attempt to manage the risk through the use of some form of hedging
program. (A hedge is a strategy to insulate a firm from exposure to price, interest rate, or
foreign exchange fluctuations.)
5231.07 The risk management process includes both internal and external controls, and involves:
a. identifying and prioritizing risks and understanding their relevance.
b. understanding the stakeholder’s objectives and their tolerance for risk.
c. developing and implementing appropriate strategies in the context of a risk
management policy.
5231.08 The process requires understanding the interconnectedness among and between the basic
components of economics, accounting, and control.
a. Economics requires being able to measure the risk before attempting to manage it.
b. The control function provides the needed guidelines for the hedging activity and any
necessary oversight.
c. The accounting function requires understanding the requirements set forth in FASB ASC
815. As per the FASB, the following points relate to FASB ASC 815:
(1) Under FASB ASC 815, an entity that elects to apply hedge accounting is required to
establish at the inception of the hedge the method it will use for assessing the
effectiveness of the hedging derivative and the measurement approach for
determining the ineffective aspect of the hedge. Those methods must be consistent
with the entity's approach to managing risk.
(2) For a derivative designated as hedging, the exposure to changes in the fair value of
a recognized asset or liability, or a firm commitment (referred to as a fair value
hedge), the gain or loss is recognized in earnings in the period of change together
with the offsetting loss or gain on the hedged item attributable to the risk being
hedged. The effect of that accounting is to reflect in earnings the extent to which
the hedge is not effective in achieving offsetting changes in fair value.
(3) For a derivative designated as hedging, the exposure to variable cash flows of a
forecasted transaction (referred to as a cash flow hedge), the effective portion of
the derivative's gain or loss is initially reported as a component of other
comprehensive income (outside earnings) and subsequently reclassified into
earnings when the forecasted transaction affects earnings. The ineffective portion
of the gain or loss is reported in earnings immediately.
(4) For a derivative designated as hedging and the foreign currency exposure of a net
investment in a foreign operation, the gain or loss is reported in other
comprehensive income (outside earnings) as part of the cumulative translation
adjustment. The accounting for a fair value hedge described above applies to a
derivative designated as a hedge of the foreign currency exposure of an
unrecognized firm commitment or an available-for-sale security. Similarly, the
accounting for a cash flow hedge described above applies to a derivative designated
as a hedge of the foreign currency exposure of a foreign-currency-denominated
forecasted transaction.
(5) For a derivative not designated as a hedging instrument, the gain or loss is
recognized in earnings in the period of change.
Interest Rates
information that would allow them to be able to compare the effective yield on
alternative instruments.
Illustration: Assume that an individual deposits $10,000 in a one-year certificate of
deposit paying 6% interest. If interest is compounded annually, he would receive
$600 in interest. If the interest were compounded semi-annually, there would be
two payments received. The first payment would be $300 and the second payment
would also include interest on that $300 and the consumer would receive $609.
Quarterly compounding would result in interest earnings in the amount of $614,
and daily compounding $618.
(2) 360-day vs. a 365-day year. Many credit card issuers compute the daily periodic
rate using a 360-day as opposed to a 365-day year, yet charge interest for 365 days.
This has the result of increasing the effective annual rate on a loan. (The effective
annual interest rate is the investment’s annual rate of interest when compounding
occurs more often than once a year.)
Illustration: Assume that you have an 18% loan. With a 365-day year, the daily
periodic rate would be 0.0493%, while with a 360-day year the daily periodic rate
would be 0.05%.
(a) Effective interest rate using a 365-day year with a 365-day daily periodic rate:
((1 + 0.000493)365 - 1) = 19.71%
(b) Effective interest rate using a 365-day year with a 360-day daily periodic rate:
((1 + 0.0005) 365 - 1) = 20.02%
funds at lower rates, which gives them an advantage over other firms that might have
more profitable projects but a relatively weak financial position.
Lower interest rates tend to make R&D expenditures more profitable. Thus, lower rates
would cause firms to increase such expenditures.
Bond Value = Coupon Payment 1 -
1
(
t ÷ r + Par Value ÷ (1 + r )
t
)
(1 + r )
Illustration: Assume that there is a bond with a 6% coupon, 20-year remaining maturity with
a $1,000 par value, paying 40 semiannual payments of $30 each. If the market interest is
equal to the coupon rate (6%), the value of the bond can be calculated as:
Bond Value = $301 −
1
(
÷ 3% + $1,000 ÷ (1 + 3% )40
(1 + 3% )40
)
= 693.44 + 306.56
= $1,000
The interest payments represent an annuity, and the final payment is a lump-sum payment.
If the market interest rate were equal to the coupon rate, the price of the bond would be
equal to the par value as shown above. If, however, the market rate was 8%, the value of the
bond would decrease due to the fact that the bondholder would still be getting only the $30
semiannual payments. Given these circumstances, the bond’s value could be calculated as:
Bond Value = $30 1 −
1
(
÷ 4% + $1,000 ÷ (1 + 4% )40
(1 + 4% )40
)
= 593.78 + 208.29
= $802.07
5232.02 There is growing consensus that the regulatory framework should apply a set of
comprehensive risk management concepts that include key concepts used by insurance
regulators and actuaries that use a “control cycle” approach:
a. Incorporating allowances (risk margin) for extreme or outlier events (i.e., black swan
events):
(1) Traditional regulatory approaches previous to the 2008 financial crisis did not
identify or lessen critical risk concentrations (e.g., the level of collateralized debt
obligations (CDOs)).
(2) Historically, there has been inadequate capital support at both the system and
individual institution level for accepting such risks that resulted in a significant
underpricing of the risks as the 2008 financial crisis began.
(3) There must be an ability to track risk in the unregulated financial sector.
(4) There should be development of models that track the expected level of loss from
an expected event rather than the current “value at risk” (VAR) approach that only
measures the minimum amount of loss from a low-probability event.
Note: A “black swan” event has a high impact, is hard to predict, and is a rare event that
is beyond the realm of normal expectations in history, science, finance, and technology.
The result is that there is a noncomputability of the probability of the consequential rare
events using scientific methods owing to their very nature of small probabilities. A
collateral debt obligation (CDO) is a type of structured asset-backed security (ABS)
whose value and payments are derived from a portfolio of fixed-income underlying
assets. The portfolio of fixed-income assets are rated at different levels of risk with
some collateralized debt obligations invested in subprime mortgage bonds.
b. There should be requirements for specific financial condition reporting.
c. There should be an independent sign-off of liability and provisioning for regulatory
purposes by someone such as an actuary who is subject to professional codes of
conduct, and that individual should have the freedom and capability to take an objective
view that may differ from management. This would be made possible by enforceable
professional codes of ethics with disciplinary standards.
The incentive structures within the industries should not distort the proper evaluation
of risk:
(1) Capital requirements should be increased for any market participant where
remuneration and incentives focus excessively on short-term results.
(2) Regulators should develop a deep understanding of the risk culture within a
particular enterprise that often might mitigate against timely reporting of critical
information.
(3) The concept of “risk margin” developed by actuaries to measure the value of illiquid
assets should be used.
b. increasing the actual level of capital available from time to time after excluding all
interests that are not at “arm’s length.”
c. looking at what could happen in the future that requires sensitivity testing as to the
potential value of assets and/or liabilities as market conditions change.
periodically must be converted to another currency. The analysis can only be completed
when the estimated cash flows are ultimately measured in the parent’s local currency.
e. Long-term financing decisions: Corporations that issue bonds may wish to consider
denominating the bonds in a foreign currency. Their preference would be to have the
bonds denominated in a currency that would depreciate over time against the currency
they are receiving from the sale. Exchange rate forecasts are needed to estimate the
expected cost differential from borrowing in different currencies.
f. Earnings assessment: When earnings are reported, subsidiary earnings are consolidated
and translated into the currency of the parent’s home country.
unbiased estimate of the future spot rate, then corporations can monitor the
forward rate as they develop their exchange rate expectations. Arbitrage activity
would tend to eliminate any differential.
Types of Exposure
5232.13 The rationale for forecasting foreign exchange rates is to help the firm make decisions as to
whether they should remain exposed to exchange rate fluctuations or if they should attempt
to hedge their foreign exchange positions. Forecasting methodologies can be evaluated over
time by comparing the actual values of the currencies to those predicted by the forecast
over long periods of time.
5232.14 While exchange rates cannot be forecasted with perfect accuracy, a reasonable forecast can
allow a firm to estimate its exposure to exchange rate fluctuations. To the degree that a firm
discovers that it is highly exposed, it then can begin to consider approaches to reduce the
exposure.
5232.15 Transactions exposure is simply defined as the degree to which the value of a firm’s future
cash transactions can be affected by exchange rate fluctuations. It can be measured by
determining (1) the projected net amount of inflows or outflows in a particular foreign
currency and then (2) the overall risk of exposure for each of the currencies.
a. Exposure to “net” cash flows: Forecasts of net cash flows tend to be for relatively short
periods where the anticipated cash flows can be predicted with reasonable accuracy.
Such predictions require a reasonably sophisticated information system to track the
positions in the various currencies.
The firm needs to project a consolidated “net” position in the currency. These positions
must be converted into dollars so that the firm can measure the exposure in each
currency using a standardized measure. Since the actual level of period cash flows and
end-of-period exchange rates are unknown, it would be reasonable to construct a range
of estimates for both cash flows and exchange rates.
b. The firm may not have its own method for forecasting end-of-period exchange rates,
but at a minimum the firm can use historical data to measure the degree of currency
variability. The standard deviation measure is a common way to define the variability of
a particular currency. Currency variability can change over time, and use of this method
would require periodic updating of the measure.
The key is to determine how each individual currency could affect the firm by assessing
the standard deviations and correlations between the various currencies over the period
of the transaction.
5232.17 Currency correlations: The correlation coefficient measures the degree to which the two
currencies move in relation to each other. Over the long run, the correlations tend to be
positive, which indicates that currencies tend to move in the same direction against the
dollar, albeit not in the same magnitude. This information would help the firm determine the
degree to which they have transaction risk among and between various currencies to which
they might be exposed. For example, if the firm has exposure with the Swiss franc and the
Japanese yen, and those two currencies are correlated vis-a-vis the dollar, then an expected
inflow of Swiss francs can serve as a partial offset against an expected outflow of Japanese
yen.
5232.18 An alternative method to assess exposure would be the value-at-risk (VAR) measure that
would combine the methods above to predict the potential maximum one-day loss for a firm
that is exposed to exchange rate fluctuations. In this instance, the firm might have a policy
that would trigger hedging activities if the VAR reaches a certain magnitude.
5232.19 Economic exposure measures the degree to which a firm’s present value of expected future
cash flows can be impacted by exchange rate fluctuations. As an example, a firm’s local sales
in its home country would be expected to decrease if its home currency appreciates because
the firm will face increased foreign competition as foreign substitute products will now be
cheaper in terms of the local currency. At the same time, cash inflows on exports
denominated in the local currency will decline due to the appreciation of the currency since
foreign importers will need more of their currency to pay for their imports.
5232.20 We must note that if a firm’s local currency depreciates, it would be impacted in the
opposite direction from when the currency appreciates. Local sales would be expected to
increase due to reduced foreign competition, and their exports denominated in local
currency will appear cheaper to importers, which will lead to increased foreign demand for
their products.
On the cost side with regard to cash outflows, imported supplies denominated in the local
currency will not be directly affected by exchange rate movements. However, the cost of
supplies denominated in foreign currencies will rise because more of the depreciated
currency will be needed to acquire the foreign currency needed to make the purchase.
5232.21 Even firms that are not involved in foreign sale of goods or purchase of foreign supplies may
be subject to economic exposure due to the impact of exchange rate movements on the
price of substitute goods offered by foreign competitors in their local markets. The degree of
economic exposure is likely to be much greater for multinational corporations (MNCs). An
MNC needs to assess the degree to which economic exposure exists and then make a
determination as to whether or not they wish to attempt to insulate themselves against it.
Economic exposure can be measured by the following:
a. Sensitivity of earnings to exchange rates: This method involves classifying cash flows
into specific income statement items and then making a subjective prediction about the
value of these items based on an exchange rate forecast. The base forecast then can be
compared with alternative exchange rate forecasts. Comparing the resulting income
statements would allow the firm to assess the impact of exchange rate movements on
earnings and cash flows.
b. Sensitivity of cash flows to interest rates: This method involves using regression analysis
of historic cash flows and comparable exchange rates. The following formula illustrates:
PCFt = α o + a1et + µ t
5232.22 Translation exposure is the exposure of the MNC’s consolidated financial statements to
foreign exchange fluctuations. A subsidiary’s earnings are translated into the reporting
currency and, therefore, are subject to exchange rate fluctuations.
There are some questions concerning the relevance of translation exposure, with arguments
in favor including the following:
a. Cash flow perspective: The key consideration is the degree to which a subsidiary would
remit earnings to the parent. If the earnings are retained by the subsidiary and
reinvested if feasible investment opportunities exist, then the earnings would not be
converted and sent to the parent. If a portion of the earnings are remitted, or expected
to be remitted in the future, then it could be expected that future cash flows could be
adversely affected if the foreign currency is weakening.
b. Stock price perspective: Investors tend to use expected future earnings forecasts to
derive an expected value per share. If an MNC conducts a large portion of its business in
foreign countries that have changing exchange rates that are expected to negatively
affect the translation of future earnings, then it is likely that the market stock price for
the MNC will decline.
There is one possible non-hedging possibility that might be available to the firm to manage
the exposure. Depending on the negotiating strength of the two parties involved in the
transaction, it might be possible for the firm to invoice the transaction in the currency that
will be used to pay for the import, effectively shifting the transaction exposure to the
exporter.
Techniques to eliminate all or part of transaction exposure:
a. Future hedges: A firm buys a currency futures contract that gives the firm the right to
receive a specified amount of a specified currency for a given price on a specific date. By
buying such a contract, the firm is able to lock-in the amount of the foreign currency
needed to make the expected payment. These contracts are designed to hedge currency
needs in the short term.
A firm could sell a futures contract if it desired to sell a specific amount of a specified
currency for a stated price on a specific date. This type of hedge might be used to hedge
future receivables denominated in a foreign currency.
b. Forward hedge: A forward hedge is very similar to a futures hedge except that it is
designed to be used by large corporations who have relatively large positions to hedge.
These hedges are negotiated between the corporation and a commercial bank with the
contract specifying the currency, the exchange rate, and the settlement date. These
contracts may be used to either purchase or sell foreign currency with a company that
needs foreign currency in the future. A company expecting to need to sell currency in
the future would negotiate a contract to sell the currency forward to lock in the rate at
which the currency can be sold.
c. Money market hedge: This type of hedge involves the firm taking a position in domestic
or foreign money markets to hedge a payables or receivables position.
d. Currency option hedge: This type of hedge ideally would insulate the firm from adverse
foreign exchange movements, but also allow the firm to benefit from favorable
exchange rate movements if the currency does not move in the expected manner during
the hedging period. This could create a situation where not hedging a position could
produce a better outcome than hedging. A currency option hedge resolves this problem
by allowing the firm to hedge the exchange risk, and at the same time participate in any
gains that might result from unexpected currency fluctuations.
5232.25 The techniques described in section 5232.24 could be used in the following fashions to
hedge specific transaction exposures:
Hedging
Hedging Receivables Hedging Payables
Technique
Futures Hedge A currency futures contract could A currency futures contract could be
be sold in the local currency in the purchased in the local currency in the
amount of the particular receivable. amount of the particular payable.
Money Market Borrow funds in the denominating Borrow funds in the local currency of
Hedge currency of the receivable, convert the payables, and invest the funds
the loan to the local currency, and until they are needed to meet the
invest it. When the receivables are payable obligation.
received, the loan is then paid off
with the cash inflows.
Currency Option A currency put option is purchased A currency call option is purchased in
Hedge in the local currency for the amount the local currency for the amount
related to the receivables. related to the payables.
exposure to the variability of cash flow attributable to a particular risk associated with a
recognized asset or liability (IAS 39.86)).
5232.27 An example of a cash flow hedge would be to use an interest rate swap to convert variable-
rate interest exposure to a fixed interest rate.
a. The swap is an instrument that, in its usual form, transforms one kind of interest stream
to another, such as floating to fixed or fixed to floating. Each swap has two
counterparties and, therefore, in each swap one party pays fixed and receives floating,
while the other party receives fixed and pays floating.
b. There are two basic forms for a swap:
(1) First, a floating-rate borrower converts to a fixed rate. In this case, a borrower has
floating-rate bank debt and carries out a pay-fixed swap, converting the debt to a
fixed rate.
(2) Second, a fixed-rate borrower converts to a floating rate. In this case, a borrower
has fixed-rate bond debt and undertakes a receive-fixed swap, converting the debt
to a floating rate.
c. From an accounting perspective, the changes in the fair value of the interest rate swap
would accumulate first in the statement of comprehensive income, and a portion of the
gains or losses would be transferred from comprehensive income to the income
statement whenever interest is paid on the hedged debt.
5232.28 There is a price risk related to the sensitivity of the market value of assets and liabilities to
the level of changes in market interest rates that could be covered by a fair value hedge.
a. A derivative would be used to hedge changes in the fair value (market value) of a fixed-
rate bond that would occur when interest rates move up or down. Hedging the risk
associated with a bond’s price risk would be a fair value hedge.
b. An example would be that a firm issues fixed-rate debt and then uses a swap for
floating-rate debt.
c. The swap would be marked to market. Any change in value of the swap would be
recognized as current earnings, and the change in the value of the debt instrument
related to change in market interest rates would be recognized on the income
statement as earnings.
5232.29 One must be aware of the differences between interest rate-fixing products (such as a swap)
and options:
a. A swap binds its user to the rate that is set when it is transacted. An option gives the
buyer the right to walk away if it would be less expensive to do so.
b. So, a firm taking out a pay-fixed swap after which rates decline is left paying the higher
rates. The risk is transformed rather than transferred. Exposure to rising rates has
become an exposure to falling rates.
c. Options offer an alternative, but with a cost. The two key factors determining the cost of
the option are time (the longer an option has until expiration, the higher the premium)
and volatility (the higher the volatility in the underlying risk being hedged, the higher
the premium).
d. Both of these factors tend to deter firms from using options, and for longer-term risk-
transfer alternatives, interest rate swaps are usually chosen.
(4) Could produce relays for the Motor division for $10 of variable costs per relay since
there would be no selling or marketing expenses.
(5) There is only capacity to produce 50,000 relays a year, so production can either be
sold to the outside or internally, but not both.
Motor division:
(1) Needs 50,000 relays a year.
(2) Can purchase relays from the outside for $15 each.
(3) Adds $25 of additional variable costs to each motor.
(4) Sells the motors for $60 each.
(5) Has variable costs of $40 per relay.
Relay Motor
Total
Division Division
Sales ($20, $60) $1,000,000 $3,000,000 $4,000,000
Less Variable ($12, $40) 600,000 2,000,000 2,600,000
Contribution Margin $ 400,000 $1,000,000 $1,400,000
If the Motor division purchases the relays from the Relay division for $12 each (the
Relay division’s variable costs), the contribution margins of the divisions and the total
organization will be changed to:
Relay Motor
Sale of 50,000 units Total
Division Division
Sales ($12, $60) $600,000 $3,000,000 $3,600,000
Less Variable ($12, $35) 600,000 1,750,000 2,350,000
Contribution Margin $ -0- $1,250,000 $1,250,000
Decrease in Contribution Margin $ 150,000
This situation, however, would potentially be different if the Relay division had excess
capacity, and the sales to the Motor division at $12 each would not have an adverse
effect on the outside sales.
d. Another problem with the above illustration is that the only division to show a profit is
the one that makes the final sale to an outside party. Thus, there is little incentive for
the first division to control costs since all of the costs are simply passed on to another
division. Unless costs are subject to some type of competitive pressure at transfer
points, waste and inefficiency almost invariably develop.
e. Despite the shortcoming of setting transfer prices at cost, this method is frequently
used. It is easy to understand and highly convenient to use. If, however, transfer prices
are based upon costs, standard costs should be used. This will avoid passing on
inefficiencies from one division to another.
f. Transfer prices at cost frequently use a cost that includes at least some level of fixed
costs. However, there are numerous variations of the definition of fixed costs:
(1) Only selling division fixed costs
(2) Only manufacturing fixed costs for the selling division
(3) Selling division’s fixed costs and corporate costs allocated to the selling division
Obviously, there would need to be an annual allocation rate for the fixed costs in
question applied to the units transferred from one division to another if some level of
defined fixed costs were to be included in the transfer price.
additional outside sales prospects at the current $20 per unit, then the organization will
benefit from the transfer. Using the transfer price formula:
Transfer price = Variable costs per unit + Lost contribution margin per unit)
It can be determined that the minimum transfer price should be $12, the variable costs
to produce the relays, since there is no lost contribution margin for sales that would not
otherwise have occurred. The organization as a whole will benefit if the transfer price is
set anywhere within the range of $12 and $15. If the Relay division (selling division)
refuses to sell to the Motor division (purchasing division), the company as a whole will
lose up to $3 of contribution margin on each unit. This can be summarized as follows:
Sale of 100,000 units (50,000 to the outside Relay Motor
and 50,000 to the Motor division) Division Division Total
Original sales ($20, $60) $1,000,000 $3,000,000 $4,000,000
Additional internal sales 500,000 -0- 500,000
Total sales 1,500,000 3,000,000 4,500,000
Less outside sales variable costs ($12) (600,000) (600,000)
Less variable costs for internal sales ($10, $35) (500,000) (1,750,000) (2,250,000)
Contribution margin $ 400,000 $1,250,000 $1,650,000
Original contribution margin $1,400,000
Increase in contribution margin $ 250,000
If the transfer price is changed to favor Subsidiary B (only charging $150,000 for the
component part), the total net income will increase due to the fact that Subsidiary B has
a lower tax rate.
Subsidiary A Subsidiary B Total
Sales $150,000 $400,000 $550,000
COGS 100,000 150,000 250,000
Gross profit 50,000 250,000 300,000
Operating expenses 40,000 150,000 190,000
Earnings before taxes 10,000 100,000 110,000
Income tax (50%, 30%) 5,000 30,000 35,000
Net income $ 5,000 $ 70,000 $ 75,000
Increase in total net income $ 10,000
5240.02 Many different ratios are computed and used to analyze the operating characteristics of
enterprises, primarily by investors and creditors, both present and potential. Several of the
frequently encountered ratios are presented in the next paragraph in the illustration of the
Ratio Company, whose income statement (year 20X2) and comparative balance sheet (years
20X1 and 20X2) are presented as follows. The ratios are computed as of December 31, 20X2.
Ratio Company
Income Statement
For Year Ended December 31, 20X2
a
($75,000 - $15,000)/20,000 shares = $3
Ratio Company
Comparative Balance Sheets
At December 31, 20X1 and 20X2
12/31/X2 12/31/X1
Assets:
Cash and marketable securities $ 250,000 $ 200,000
Net accounts receivable 200,000 100,000
Inventories 100,000 150,000
Total current assets 550,000 450,000
Net plant and equipment 400,000 450,000
Goodwill 100,000 100,000
Total assets $1,050,000 $1,000,000
Ratio Company
Statement of Cash Flows
For Year Ended December 31, 20X2
b. Quick (acid Current assets less Measures ability to $550,000 − $100,000 = 2.25 Receivables may be subject to a
test) ratio Inventories and discharge currently $200,000 lengthy collection period and
prepaid assets maturing might have to be factored at
Current liabilities obligations based less than carrying value if cash is
on most liquid needed immediately. Securities
(quick) assets are subject to fluctuating
market conditions which affect
their liquidation amount.
c. Inventory Cost of goods sold Measures relative $600,000 Different inventory cost flow
turnover Average inventory control over
inventory
investment. May
( $150,000 + $100,000
2 ) assumptions can produce
widely different inventory
valuations and thus turnover
= 4.80
provide basis for ratios.
determining the
presence of
obsolete inventory
or pricing problems
(in case of low
turnover).
d. Receivables Net credit sales Confirms the $1,000,000 Affected by significant seasonal
turnover Average receivables fairness of the
receivable balance.
May indicate
( $200,000 + $100,000
2 ) fluctuations unless denominator
is a weighted-average. Poor
collection policy can understate
= 6.67
presence of this ratio by increasing average
possible collection receivables.
problems (in case
of low turnover).
e. Cash from Net cash provided Shows extent to $100,000 Current liabilities may
= 0.50
operating by operating which a company $200,000 understate short-term demands
activities to activities has covered its on cash.
current Current liabilities current liabilities by
liabilities generating cash
from normal
operations
g. Rate of Net income Measures the $75,000 Similar to total asset turnover
return on
total assets
Average total assets productivity of
assets in terms of
producing income
(
$1,050,000 + $1,000,000
2 ) ratio. Accrual net income
subject to estimates and does
not reflect actual cash return.
= 0.073 or 7.3% Rate of return is based on
historical asset cost which does
not reflect current values.
h. Return on Net income less Measures return to $75,000 − $15,000 Book value of common
( )
common preferred dividends common $300,000 + $250,000 stockholders’ equity reflects
stock- Average common stockholders in 2 historical cost and not current
holders’ stockholders’ equity aggregate value. Accrual net income
equity = 0.22 or 22% involves estimations and does
not reflect actual cash return.
i.(1) Debt Total liabilities Indicates extent of $200,000 + $450,000 Denominator reflects historical
ratio Total assets leverage used and $1,050,000 costs and not current values.
creditor protection
= 0.62 or 62%
in case of
insolvency
i.(2) Debt/ Total liabilities Determine the $200,000 + $450,000 There is a lack of uniformity in
equity Stockholders’ equity equity’s long-term $100,000 + $200,000 + calculating this ratio.
ratio debt-paying ability $100,000
= 1.62
j. Equity ratio Total stockholders’ Measures total $100,000 + $200,000 Assets are recorded at historical
equity asset investment + $100,000 cost and not at current value.
Total assets provided by $1,050,000
stockholders = 0.38
k. Times Income before Measures ability $175,000 = 3.50 Accrual income does not
interest interest expenses to cover interest $50,000 necessarily indicate availability
earned and taxes charges of cash to pay interest charges.
Interest expense
l. Price- Market price per Indicates $17.00 Earnings per share computation
= 5.67
earnings common share relationship of $3.00 subject to arbitrary assumptions
ratio Earnings per share common stock to and accrual income. EPS is not
net earnings the only factor affecting market
prices.
o. Book value Common stock- Measures net $200,000 + $100,000 Assets are recorded at historical
per common holders’ equity assets applicable to 20,000 shares cost and not at current value.
share Common shares each common
outstanding share = $15.00
p. Cash flow Income plus noncash Measures resources $75,000 + $25,000 This ratio is the least
= $5.00
per common adjustments (i.e., cash) 20,000 shares understood and therefore the
share Common shares generated per most likely to mislead an
outstanding common share investor. FASB ASC 230-10-45-3
strongly recommends against
the isolated disclosure of this
ratio.
q. Payout ratio Common dividends Measures portion $10,000 = 0.17 Income does not necessarily
to common Net income less of net income to $75,000 − $15,000 measure cash available for
share- preferred dividends common dividend payment. Heavily
holders shareholders paid influenced by management
out in dividends policy, nature of business, and
stage of development, all of
which diminish comparability.
r. Cash from Net cash provided by Shows cash flow $100,000 = 1.33 Net income includes noncash
operating operating activities effects of the $75,000 revenues and expenses
activities to Net income company’s net recognized by accrual
net income income for the accounting principles.
period
5240.04 For analytical purposes, the ratios presented in this table can be grouped as follows:
a. Measures of liquidity (i.e., ability to meet current debt):
(1) Current ratio
(2) Quick ratio
(3) Inventory turnover
(4) Receivables turnover
(5) Cash from operating activities to current liabilities
b. Measures of return on investment:
(1) Total asset turnover
(2) Rate of return on total assets
5240.05 Care must be taken in the use of ratios such as those described for several reasons, as can be
seen by the specific limitations stated. In summary, several of the critical considerations of
which you must be continually mindful of in ratio analysis include the following:
a. Accounting methods used to state assets, liabilities, stockholders’ equity, revenues, and
expenses are not necessarily designed to produce the most useful numbers for the
purposes for which the ratios are intended.
b. Differences in the underlying economic events, the types of enterprises involved, the
stage of development of enterprises, and other factors affect comparability between
enterprises.
c. Several ratios include the use of external market values of stock which are influenced by
numerous variables over which management has little influence or control.
d. Management policy may influence many ratios.
The potential drop in market value is the risk associated with holding a long-term fixed
interest-bearing instrument.
c. If a short-term Treasury bill is held during a period of rising interest rates, the loss in
market value related to that increase in interest rates is much less significant. For
example, given the same interest rate scenario described above, the price of a
$100,000, 180 T-bill would decline to $99,069.
Credit Risk
5311.06 Credit risk is the risk that receivables will not be collected in full on a timely basis. Increased
risk includes costs related to bad debt losses, higher receivable balances resulting in higher
carrying costs, higher customer investigation fees, and collection costs.
Leverage
A DOL of 2 means that the percentage change of EBIT will be twice the size of the
percentage change in sales. If sales increase by 20%, then EBIT will increase by 40%;
however, if sales decrease by 20%, then EBIT will decrease by 40%.
c. A firm can change the DOL by changing the proportion of fixed costs to variable costs.
The larger the proportion of fixed costs, the higher the DOL and the higher the
breakeven point. The higher the proportion of variable costs, the lower the DOL and the
lower the breakeven point.
Also, notice that after the breakeven point is reached with a high DOL, the operating
profit increases quickly; thus, the additional risk associated with high fixed costs also
contains the possibility of higher profits in the event of a positive sales outcome. The
following diagram shows how income can increase rapidly after reaching the breakeven
point if high fixed costs and low variable costs are used in the production process.
The lower the fixed costs, the lower the DOL and breakeven point will be. Given the
same sales as in the above diagram, a lower percentage of fixed cost will cause the
breakeven point to be reached at an earlier sales level; however, the operating profit
will increase slowly in relation to increased sales. The operating profit (loss) is the
difference between the sales and total costs.
d. Illustration: Tally Company has the following information available regarding sales and
costs from the last period:
Sales in units 10,000
Selling price per unit $10.00
Variable costs per unit $6.00
Fixed costs for period $20,000
earn enough to cover the fixed interest and preferred dividend payments. The more
leverage, the greater the risk, and the higher the cost of capital.
b. Degree of financial leverage (DFL) is the percentage change in earnings available to
common stockholders related to a given percentage change in EBIT.
% change in net income EBIT
DFL = =
% change in net operating income EBIT – Interest
When there is no debt financing used, the DFL will equal 1. When debt does exist, a
change in EBIT will result in a greater proportional change in the earnings per share.
If the DFL were 2, then a 10% increase in EBIT would result in a 20% (10% × 2) increase
in earnings per share.
If the organization has preferred stock, the DFL equation is modified to:
EBIT
DFL = Preferred dividends
EBIT – Interest expense –
1 – Tax rate
c. If a firm has no debt and begins to substitute debt for equity, at first the value of the
firm will rise, reach a peak, and then begin to fall due to the increased risk of additional
debt. This peak in maximized value is the target capital structure.
d. Illustration: Tally Company has the following information available about sales and costs
from the last period:
Sales in units 10,000
Selling price per unit $10.00
Variable costs per unit $6.00
Fixed costs for period $20,000
Interest costs for period $15,000
b. This ratio can be used to predict the effect of a change in sales on income and ultimately
the earnings available to common stockholders.
c. This ratio is determined by the relationship between operating and financial leverage.
For example, if risk is lowered by reducing the degree of operating leverage, then the
firm would be in a better position to increase the use of debt, thus increasing the
financial leverage.
d. The financial statement effects of DOL and DFL can be shown by using the income
statement format:
Percentage Resulting
Previous Increase Change
Sales $100,000 10% $110,000
DOL Variable costs 60,000 66,000
of
2 Fixed costs 20,000 20,000
DTL EBIT 20,000 20% 24,000
of
8 Fixed financing costs 15,000 15,000
DFL EBT 5,000 9,000
of
4 Taxes (40%) 2,000 3,600
Net income $ 3,000 80% $ 5,400
Please note that the results were based on the assumptions that the cost relationships
remain unchanged (fixed and variable) as well as the firm’s capital structure, including
the amount of the fixed financing payments. As the relationships change, the DOL, DFL,
and DTL would also change.
e. Illustration: Tally Company has the following information available about sales and costs
from the last period:
Sales in units 100,000
Selling price per unit $10.00
Variable costs per unit $6.00
Fixed costs for period $20,000
Interest costs for period $15,000
What is the DTL?
Solution: DTL = DOL × DFL
10,000 ($10 – $6) 10,000 ($10 – $6) – $20,000
= × 10,000 ($10 – $6) – $20,000 – $15,000
10,000 ($10 – $6) – $20,000
= 2×4=8
(2) Firms often use ratio analysis to aid in capital structure decisions. For example, an
average company in the meat processing industry has a debt-to-equity ratio of 2.17.
If ABC Meat Processing Company has a debt-to-equity ratio of 1.75, management
may decide to use debt financing for a proposed expansion. If, however, that same
company has a debt-to-equity ratio of 2.5, management may decide to use equity
financing instead. The whole idea here is for an organization to keep their debt-to-
equity ratio in line with the industry average.
g. Items that management may wish to consider when making capital structure decisions
would include the following:
(1) Financial flexibility: As the proportion of debt increases, the less likely that
additional debt will be available at a reasonable interest rate; therefore,
management tends to act on the conservative side and use proportionally more
debt increasingly sparingly.
(2) Tax position: Unlike dividends, interest expense provides a tax shield for companies
and will generally make debt financing less expensive than equity financing.
(3) Interest rates: When interest rates are low, companies will look more favorably on
long-term borrowing than when rates are high.
(4) Industry comparison: Lenders often will use industry averages as a criterion when
evaluating the creditworthiness of a firm, as will potential investors.
(5) Internal funds: Companies will generally use internally generated funds before
seeking additional debt or issuing additional equity.
(6) Risk: The more debt a company has, the greater the financial distress costs and the
higher the risk for bankruptcy. Also, the higher the debt level, the lower the
potential credit rating. As a firm’s credit rating drops, the cost of borrowing
increases.
(7) Stock values: If a company’s stock is undervalued, management is not likely to issue
additional shares, since such an action could potentially have a negative effect on
an already low stock price. Using an additional stock issue to raise necessary funds
when a company’s stock is overvalued stock, however, would be desirable.
kd = Cost of debt
T = Marginal tax rate
(2) Preferred stock provides the investor with a constant stream of dividends; however,
these dividends do not provide a tax shield since they are not a deductible expense
for the issuing company. The cost of preferred stock to the issuing company can be
calculated by dividing the annual return (interest payment) by the net issuance
price for the preferred stock.
Dp
kp =
Pp
Where:
kp = Cost of preferred stock
Dp = Preferred dividend
Pp = Price of preferred stock
Generally, however, when looking at the cost of issuing new preferred stock, the
company would have to pay flotation costs to an investment house; thus, the
proceeds received by the company would be less than the issue price. This would
increase the cost of this form of financing. Taking flotation costs into consideration,
the formula to calculate the cost of preferred stock to the issuing company would
be:
Dp
Kp =
(1 – f)Pp
Where:
kp = Cost of preferred stock
Dp = Preferred dividend
Pp = Price of preferred stock
f = Flotation costs
(3) The cost of retained earnings is the opportunity cost that stockholders of a firm
could earn elsewhere if they made investments of comparable risk. This figure
would need to be imputed.
(4) The cost of equity (ke) is more expensive than the cost of debt since stockholders
are subject to more risk than debt holders. There are a number of ways to estimate
the cost of equity, and one commonly used method to obtain the estimate is the
dividend growth model. The formula used is:
ke = (D1 ÷ Po (1 – F)) + g
This formula assumes that the firm pays a dividend equal to D1, would issue new
stock at price Po, and would pay a flotation cost (F) that is equal to some percentage
of the value of the stock being issued. It is assumed that the firm’s dividend will
grow at a constant rate (g). The result is the cost of using internally generated
equity. If new stock is going to be issued, the cost of new equity would be the result
obtained for internally generated equity divided (1 – Percent of flotation costs).
Illustration: The firm’s dividend was $1.50 and is expected to grow 6% per year.
They expect to be able to sell new stock at a price $30 per share and flotation costs
are expected to be 10% of the value of the stock issue. What is the firm’s cost of
equity?
ke = ($1.50 ÷ $30 (1 – 0.10)) + 6.0%
= ($1.50 ÷ $27.00) + 6.0%
= 5.56% + 6.0%
= 11.56%
c. Most analysts use the CAPM (capital asset pricing model) to estimate the required
return on a firm’s cost of equity. The basic equation for estimating the required return
on equity (Ri) is:
Ri = RF + βi (Rm – RF)
d. The primary conclusion of the capital-asset pricing model (CAPM) is that the relevant
risk on any security is its contribution to the risk of a well-diversified portfolio. A
commonly used benchmark portfolio for the market portfolio is the S&P 500.
e. Illustration: Assume that the current T-bill rate is 3.35% and the market risk premium is
7% for a diversified market portfolio.
Ri = RF + β i (Rm – RF)
= 3.35% + 1.0 (7.0%)
= 10.35%
Under this scenario, investors would expect to earn 10.35% on a diversified market
portfolio. (Note: The beta coefficient for the diversified market portfolio is 1.0.)
Given the information in the example above and assuming that the beta coefficient for
an individual security is 1.4, what return would investors require to hold this security?
Solution: Using the formula from (c) above, that is, Ri = RF + βi (Rm – RF), the return
required by investors would be 13.15%. Again, this is the cost of internally generated
equity. The cost of issuing new stock would be the cost of internally generated equity
divided by (1 – Percentage of flotation costs).
Ri = RF + βi (Rm – RF)
= 3.35% + 1.4 (7.0%)
= 13.15%
f. The weighted-average cost of capital can be computed as follows:
WACC = (wtd × kd) + (wtpf × kpf) + (wte × ke)
Example: Assume that the firm can issue new debt at 5% and that the marginal tax rate
is 40%. The firm has $1,000 par preferred stock that pays a dividend of 4%. Using the
CAPM, it is estimated that investors require a 16% return on equity investments. The
firm has a target capital structure of 30% debt, 10% preferred stock, and 60% equity.
What is the firm’s WACC?
Solution:
WACC = (.30 × [.05 × (1 – .40)]) + (.10 × .04) + (.60 × .16)
= .009 + .004 + .096
= .109 or 10.9%
b. The tax shield for debt makes debt a very attractive component of the capital structure.
However, increasing the amount of debt in the structure beyond some point causes
increasing financial distress costs that would cause the cost of additional debt to climb.
Theoretically, the firm should continue to issue debt until the increasing financial
distress costs offset the positive value of the tax shield.
c. It is very difficult to accurately determine the optimal capital structure, and in practice a
firm attempts to find an optimal range of debt-to-equity that would minimize the
weighted-average cost of capital.
d. Key elements in making capital structure decisions include the following:
(1) Sales stability: A firm whose sales are stable can take on more debt than a firm with
unstable sales.
(2) Asset structure: Firms with assets that are suitable to be pledged as security for a
loan can use debt more heavily than firms with special purpose assets.
(3) Operating leverage: Firms with less operating leverage are better able to employ
financial leverage.
(4) Growth rate: Faster-growing companies are more likely to rely more heavily on
debt.
(5) Profitability: Firms with high profitability are better able to support more of their
financing needs with internally generated funds.
(6) Taxes: The higher a firm’s marginal tax rate, the greater the advantages of using
debt.
(7) Management attitude: Management can exercise judgment as to the appropriate
capital structure.
e. As shown by the following graphs, the lower the firm’s weighted-average cost of capital
(WACC), the greater the value of the firm since the WACC is used to discount the firm’s
expected cash flows. Thus, many firms are concerned about creating a capital structure
that minimizes its cost of capital. As a firm moves from being an all-equity firm (that is,
with debt equal to zero) to adding debt to the capital structure, the firm’s WACC will
begin to decline. This is true because debt is less risky than equity (demanding a lower
return) and debt has a “tax shield” in that the interest cost is tax deductible. Therefore,
adding debt to the capital structure mix reduces WACC. This decline will continue until
financial markets perceive that the firm’s debt-to-asset (or equity) ratio has reached a
point where additional debt will increase the firm’s “financial distress costs” (that is, the
possibility of bankruptcy) and require a higher return on the debt to compensate for the
increased risk. At this point, the firm’s WACC will begin to rise. A firm’s “optimal capital
structure” is the ratio that minimizes WACC; in the example found in Graph B it is found
to be 40%.
Graph A:
Graph B:
(b) A small firm may be reluctant to use debt as a source of funds due to the
extreme risk that failure of one product or a slight down-turn in the economy
may have on cash flow. Default is more likely in a small firm than a large one.
(2) the industry norms.
Debt service is related to sales volume and the control of costs leading to
profitability. Growth industries tend to have high profit margins. Mature industries
tend to have stable sales and profit margins. Lenders are concerned about a specific
company’s ability to make necessary payments. Generally, debt holders will not
lend funds to a firm that has a debt to equity ratio above the normal range for the
industry or a times interest earned ratio below the normal range for the industry
without higher interest rates to compensate for the additional risk. Lenders are
particularly interested in an organization’s solvency ratios as compared to the
industry norms when deciding whether additional funds should be lent to the firm
in question.
(3) the anticipated future growth rate.
If new financing is used to expand, it is important to estimate the future growth of
sales in order to determine whether cash flows resulting from the new productive
assets will create a cash flow large enough to service the debt.
(4) lenders’ attitudes toward the industry and the specific firm.
Lenders’ attitudes play an important role in capital structure. As lenders become
concerned about rising debt levels, they will either demand higher interest rates to
compensate for the additional risk, or they may refuse to lend the necessary funds.
A firm’s ability to grow may often be limited to the funds available through the use
of debt.
c. Whenever additional long-term funds are needed, management can choose between
debt and equity. Long-term debt is likely to be used when:
(1) sale and profits are estimated to be stable or increasing.
(2) anticipated profits are sufficient to make good use of leverage.
(3) control (through voting privileges) is important.
(4) the existing capital structure has a low use of debt.
(5) requirements or covenants of debt issues are not arduous.
seasonal needs. Spontaneous credit (accounts payable) is used to finance the remaining
seasonal needs. A firm using this approach would hold liquid assets in marketable
securities during the low points of the seasonal cycle.
c. The maturity matching approach to financing matches assets to liability maturities. This
approach minimizes default risk and is considered to be a moderate approach to
financing.
Illustration: If a new factory is financed with short-term funds, it is unlikely that the new
facility would have been built and produced a large enough cash flow at the end of one
year in order to repay the loan at maturity. In this scenario, if the lender refused to
renew the loan, the company would be in a difficult position. If the new facility had
been financed with long-term debt, the cash flows provided by the new operation
would have more closely matched the repayment cash outflows, and the need for
refinancing is less likely to arise.
d. The difficulty with maturity matching is that imperfect estimates are used in
determining the lives of assets. In reality, firms tend to attempt to develop a reasonable
approximation of matching. Also, it is necessary to have some percentage of the
financing come from equity that has no maturity.
e. The aggressive approach to financing short-term debt would be to finance part of the
permanent current assets with spontaneous credit such as accounts payable. The
aggressive approach can be carried to a high degree by also financing part of the fixed
assets with short-term borrowing. Such a firm would be at risk if interest rates increase
as well as face possible difficulties with loan renewals. The degree of aggressiveness is
defined by the degree to which permanent NOWC (net operating working capital) and
fixed assets are funded by spontaneous credit and/or short-term debt.
c. Commercial paper
d. Secured loans
million on the revolving credit agreement, the commitment fee would drop to a percent
of the unused portion of the agreement ($62,500 in this illustration), and the
corporation would pay interest on the outstanding borrowed balance as well. The
biggest difference between a line of credit and a revolving credit agreement is that
there is a legal obligation with a revolving credit agreement that does not exist with a
simple line of credit. This legal obligation guarantees the company access to the funds
over the life of the agreement.
d. A letter of credit is an international financing tool that guarantees payment to an
international supplier upon the safe arrival of the goods by issuing a loan to the
purchaser. A letter of credit can be irrevocable (not subject to cancellation if the specific
conditions are met) or revocable.
e. Commercial paper is an unsecured promissory note that is issued by large banks and big
corporations to meet short-term cash needs. It is a promise to repay the borrowed
funds at a future given date with a maturity of no more than 270 days. Rates for these
loans are normally lower than for bank loans. This financing option is considered to be
quite safe due to the short time frame and the strength of the borrowing companies;
however, there is no flexibility in the repayment date as might be received with a bank
loan.
Solution: The factor fee will be $1,500 (3% of $50,000). The reserve will be $4,000 (8%
of $50,000). The interest will be calculated on 10% of the balance available to the seller
and adjusted for a monthly period length.
This can be summarized as follows:
Sales: $50,000
Less:
Factor fee (3% × $50,000) $1,500
Reserve (8% × $50,000) 4,000 5,500
Amount due firm 44,500
Less interest on amount for
30 days (10% × $44,500 × 30/360) 371
Amount paid to seller $44,129
After 30 days, any remaining part of the reserve (maximum of $4,000 in this case) that
was not absorbed by returns and allowances will be remitted to the seller.
The cost of factoring needs to be compared to the costs of running a credit department
when making a decision as to which alternative to pursue.
d. Inventory financing entails the use of inventory as security for a short-term loan. Such a
loan can attach to all inventory held or a particular portion of inventory identified by
serial numbers or placed in a particular location.
5313.10 A term loan is a legal agreement between a borrower and lender where the borrower
promises to make interest and principal payments at specific times to the lender for the use
of borrowed funds. It is a form of funded debt (long-term debt).
a. Term loans have advantages over the issuance of bonds and equity.
b. Term loans can be drawn up quickly since the borrower and lender work directly
together.
(1) The terms of the loan can have more flexibility since they are not registered with
the Securities and Exchange Commissions (SEC). Also, modifications can potentially
be made during the life of the loan.
(2) Term loans have low issuance costs.
c. Generally, term loans are paid off in equal installments. This is a protection to both the
lender and the borrower that funds will be available for repayment. When funds from
the loan are used to purchase equipment, the repayment schedule is frequently
matched to the productive life of the equipment.
d. Interest rates can be either fixed or variable.
5313.11 A bond is generally a publicly offered form of long-term debt where the borrower agrees to
makes payments of interest and principal on specific dates to the bond holder. Most bond
issues are advertised and held by many different investors as opposed to the one lender of a
term loan. The bonds have a coupon rate (rate of interest) that is generally fixed.
5313.12 An indenture is the legal document that contains the terms of the bond issue. This
document is approved by the SEC before the bond issue is offered to the public. It will clearly
state the provisions of the issue. Common provisions include the following:
a. Call provision: The right of the issuer to redeem the bond issue prior to the maturity
date under certain circumstances
b. Sinking fund: Requires the issuer to retire a portion of the bond issue each year or
deposit funds into a restricted account to be accumulated for the retirement of the
bonds
c. Restrictive covenants: Conditions that must be met by the issuer during the life of the
bond issue, such as maintaining a debt ratio no higher than a specified amount,
restricting dividend payments, maintaining a minimum current ratio, or other similar
requirements
5313.13 Bond ratings are determined using a rating system such as Standard & Poor’s, based upon
the probability that the issuing corporation will go into default. These ratings range from
AAA (very strong) to D (in default). AAA and AA are considered to be of a high investment
quality, and they along with A and BBB rated bonds are considered to be investment grade,
BB and B are considered to be substandard, and CCC to D are considered to be speculative.
5313.14 The bond rating is based upon the financial health of the issuing organization (capital
structure as well as stable profitability), bond provisions, pending legal actions, and time to
maturity. The rating is done by rating agencies using subjective judgment. The bond rating
has a direct effect on the bond’s coupon rate, and thus the firm’s cost of capital if a firm
wishes to issue new debt. Since many bonds are purchased by institutional investors who
are generally restricted to purchasing no lower than investment-grade bonds, this rating is
extremely important. The issue here is that there is an additional impact in the event a firm’s
rating is downgraded. Then there would be a decrease in the market price of outstanding
bonds.
5313.15 The market value of bonds is based upon the present value of discounted future cash flows,
comprised of an annuity plus a lump sum. The bond’s market value fluctuates with changes
in the market interest rates. If the coupon rate equals the market interest rate, then the
market value of the bond will be equal to the face value (par value). If the market interest
rate is above the coupon rate, the bond will be selling at a discount to par. If the market
interest rate is lower than the coupon rate, the bond will sell at a premium to par.
As the bond approaches maturity, its market value will approach par. Changing interest rates
will have an impact on the market value of the bond, the impact being greater the longer the
remaining maturity of the bond. The issuing corporation, however, will be required to make
the same semi-annual interest payment no matter what changes occur in the interest rate
over the life of the bond. When bonds are issued, the coupon rate on the bond generally is
set close to the current market rate.
Illustration: A $1,000 bond with a 7% coupon rate, maturing in 10 years, is selling in a 6%
market. What is the current market value of the bond?
Solution: The bond is valued by determining the PV of the future cash flows (maturity value
of $1,000 and semi-annual interest payments of $35 each ($1,000 × 3.5%)). Using present
value tables for 3% (due to semi-annual interest payments), 20 periods, the value of the
bond can be calculated:
PV of maturity ($1,000 × .5537) $ 553.70
PV of interest payments ($35 × 14.8775) 520.71
Present value $1,074.41
Due to the fact that the coupon rate is higher than the current market rate, the bond will sell
at a premium.
c. Long-term debt has a maturity date at which time the principal needs to be repaid.
d. Interest payments are a fixed cost that increase the breakeven point of the firm. Interest
must be paid whether a profit is made or not. During a downturn, making interest
payments can become difficult.
e. Debt increases the financial leverage (risk) of the company, and the higher the leverage,
the less flexibility a company has for raising future financing.
f. Long-term debt is often obtained in larger amounts, and since predicting actual cash
needs is not an exact science, long-term borrowing is often out of tune with actual
needs.
g. The amount of debt financing is limited, and as the percentage of debt in the capital
structure increases, so does the cost of issuing additional debt.
Lease Financing
5313.18 The FASB defines a lease in ASU 2016-02 to be “a contract, or part of a contract, that
conveys the right to control the use of identified property, plant, or equipment (an identified
asset) for a period of time in exchange for consideration.” An entity will determine whether
a contract contains a lease by assessing whether the use of an identified asset is either
explicitly or implicitly specified and whether the customer controls the use of the identified
asset.
5313.19 There are two sets of lease criteria: the first set is for both lessees and lessors, and the
second set is for lessors only.
5313.20 The first set of criteria result in a finance lease for a lessee or sales-type lease for the lessor if
the lease meets any of the following criteria at commencement:
a. Title (ownership) transfers to the lessee by the end of the lease term.
b. The lease contains a purchase option that the lessee is reasonably certain to exercise.
c. The lease term is for the major part of the remaining economic life of the underlying
asset. This criterion shall not be used if the lease commencement date is near the end of
the asset’s economic life.
d. The present value of the sum of the lease payments and any lessee guaranteed residual
value not already in the lease payments equals or exceeds substantially all of the fair
value of the underlying asset.
e. The underlying asset is specialized and is not expected to have an alternative use to the
lessor at the end of the lease term.
5313.21 For most leases of property (e.g., land or building), a lessee would classify the lease as an
“operating” lease if none of the five criteria listed in section 5313.20 are met and will
recognize the following:
a. At commencement, a right-of-use (ROU) asset and a lease liability, initially measured at
the present value of lease payments, using the interest rate implicit in the lease if
known; otherwise, the lessee uses its incremental borrowing rate.
b. Subsequent to commencement, a single lease cost, combining the unwinding of the
discount on the lease liability (i.e., interest expense) with the amortization of the ROU
asset, on a straight-line basis. In other words, the lease liability is reduced by the
amount of the periodic payment less the amount of that payment attributable to
interest. The lessee then “plugs” the ROU asset amortization at whatever amount is
needed for interest plus amortization to equal the straight-line payment in the lease
contract. Impairment, if any, is recorded.
5313.22 If any of the five criteria in section 5313.20 are met (control is deemed to have transferred),
the lessor accounts for the lease as a sales-type lease with or without a selling profit and will:
a. at commencement, recognize the net investment in the lease [lease receivable (lease
payments plus guaranteed residual values) plus the unguaranteed residual asset];
derecognize the carrying amount of the underlying asset; and recognize sales revenue
and cost of goods sold at the beginning of the lease such that selling profit (usually
when the fair value of the asset exceeds the cost or book value of the asset) if any, is
immediately recognized. Expense initial direct costs if a selling profit exists and defer
initial direct costs by increasing the net investment in the lease if there is zero profit.
b. subsequent to commencement, recognize interest income (interest on the lease
receivable and interest from the unguaranteed residual asset accretion) using the
effective interest method and reduce the net investment in the lease.
5313.23 When none of the first set of criteria in section 5313.20 are met, a lessor shall classify the
lease as an operating lease unless both of the following second set of criteria are met to
classify the lease as a direct financing lease:
a. The present value of the sum of the lease payments and any lessee or other third party
guaranteed residual value not already in the lease payments equals or exceeds
substantially all of the fair value of the underlying asset.
b. It is probable that the lessor will collect the lease payments plus the residual value
guarantee.
5313.24 The availability of credit is sometimes a major part of the decision to purchase or lease. If an
organization wishes to maximize financial leverage while maintaining a high level of
(superficial) solvency, operating leases may be chosen.
b. Although the leased equipment is not owned, frequently maintenance costs for the
asset are the responsibility of the lease holder.
c. Over the life of a lease, the payments made on the lease are greater than if the asset
were purchased for cash. The additional amounts of the basic cost of the asset are
similar to the interest payments made on borrowed funds.
d. There is no ownership during the lease period, thus no effective equity related to
ownership.
Equity Financing
Preferred Stock
5313.27 Preferred stock is often referred to as a hybrid in that it has some of the characteristics of
debt and some of the characteristics of equity. Preferred stock is more risky than debt;
however, it is less risky than common stock.
a. Similarities to debt:
(1) There is a fixed dividend similar to the fixed interest charge for debt.
(2) Preferred stockholders have priority over common stockholders in the case of
liquidation and dividend payments.
(3) Preferred stockholders have no voting rights.
b. Similarities to equity:
(1) Preferred stock increases equity.
(2) The original investment does not need to be paid back.
(3) There is no maturity date.
(4) Preferred dividends are not a deductible expense.
(5) The preferred dividend is not required to be paid.
(6) The liability for preferred stockholders is limited to their initial investment.
c. There are a number of varieties of preferred stock.
(1) Participating preferred stock allows for a limited participation in dividend increases
based on a stated formula.
(2) Dividends on preferred stock can be cumulative or noncumulative.
d. Preferred stock generally does not have a maturity date, and some issues are
convertible into common stock and some have a callable feature that provides a level of
flexibility for management.
e. The equity position of the company is increased, thus improving the creditworthiness of
the firm.
f. The convertible feature coupled with the callability feature can be used to force
conversion. This gives management some degree of flexibility.
Common Stock
5313.30 Common shareholders are owners of the company and have all of the rights given to them in
the corporate charter. A common stock investment is the most risky of the investments
presented in this section since the common stockholders are not guaranteed any type of
return in either dividends or stock appreciation, and these investors are the last to receive
anything in a liquidation. As a matter of fact, it is the common stockholders who provide a
cushion for the debt holders should a loss occur in the liquidation process.
5313.31 Common shareholders have a right to the earnings of a company after obligations to
creditors and preferred stockholders are met. This residual income can be paid to the
shareholders in the form of dividends or reinvested (retained) in the business. Shareholders
can benefit from the earnings that are retained due to the fact that the value of the firm is
increased, and the firm’s earnings power can increase due to the additional investment
resulting in future dividends and a potential appreciation in the value of each share of stock.
5313.32 Although common shareholders hold ownership interests in a company, they have limited
liability in that the only amount that they have at risk is the amount of their investment. This
limited liability provides potential safety in the case of legal proceedings against the firm
and/or liquidation. It is unlikely that most investors would purchase stock without the
limited liability provision.
Retained Earnings
5313.36 Retained earnings are GAAP profits that have not been paid out to shareholders. The “cost”
associated with this source of funding is the opportunity cost for the common shareholders.
a. The directors of a company have two choices when a firm has profits:
(1) Pay those profits out in the form of dividends.
(2) Retain those earnings and reinvest them in the organization.
b. The cost of retained earnings is the opportunity cost to the investors who would have
had free choice in how to use those profits in whatever investment possibility deemed
as to hold the highest return possibility had dividends been paid.
c. If earnings are retained in the company as opposed to issuing new common stock,
flotation costs associated with a new issue are avoided; thus, the cost of retained
earnings is lower than the cost of a new issue of common stock.
Thus cash is tied up in non-cash current items for 50 days. This represents a permanent
financing need.
b. As long as the cash conversion cycle can be shortened without encroaching on sales or
increasing costs, a firm should attempt to do so since that will mean that a lower
amount of permanent financing will be necessary, and there is a cost to any financing.
This can be done by:
(1) selling products more quickly.
(2) collecting receivables more quickly.
(3) slowing down payments.
d. The net float is the difference between the firm’s checkbook balance and the balance of
the account in the bank’s records. The difference is due to the collection float and the
disbursement float. The disbursement float is the amount of the checks written but that
have yet to clear the bank. The disbursement float allows a firm to have lower cash
balances.
e. Techniques such as the use of zero balance accounts allow a firm to not hold funds in
non-interest-bearing accounts to meet anticipated disbursement needs.
Marketable Securities
(4) Collection policy: how tough or lax the firm is in terms of attempting to collect
slow-paying accounts. Again, there is a trade-off between attempting to speed up
the collection of an account and possibly losing a customer.
c. The credit manager is generally responsible for administering a firm’s credit policy.
However, credit policy is generally set at an administrative level within the firm.
d. A change in credit policy involves the interaction between several variables. Below are
some of the issues management would need to analyze when considering a change in
the credit policy.
(1) Sales: What is the projected increase in the level of gross sales that would result
from the change?
(2) Account receivable: There is generally a proportional relationship between an
increase in sales and an increase in accounts receivable. An increase in accounts
receivable would mean that additional funds would be tied up until collection.
Increasing accounts receivable would require increasing funding (debt or equity)
with a possible increase in interest costs.
(3) Discounts and bad debt losses: What proportion of customers would be expected to
take the discount? What proportion of the additional receivable would become bad
debts?
(4) Costs of running a collection department: Would more staff (higher costs) be
needed to run collection department due to higher volume? Would there be fixed
costs associated with the expansion of the credit collection efforts?
(5) Management would have to compare the increased revenues expected with the
increased costs to determine whether the firm would be better off on an
incremental profit basis.
e. A firm’s credit policy includes the credit standards, credit period, available discounts,
and collection procedures.
f. The quality of credit is defined as the ability to collect receivables in full and in a timely
manner.
g. A firm’s credit policy is often dictated or strongly influenced by competitors since unless
the firm’s policies are competitive, sales will often be lost.
b. If the credit period is extended, trade accounts receivable will increase due to the
increased holding time of receivables as well as increased credit sales.
c. The cost of extending the credit period can be calculated by multiplying the increase in
accounts receivable by the marginal cost of capital.
Illustration: If a firm increases its credit period from 30 to 45 days, average accounts
receivable are predicted to increase from $400,000 to $500,000. If the cost of capital is
10%, then the marginal cost of changing the credit period is $10,000 ($100,000 × 10%).
d. In poor economic times, a firm needs to be concerned that customers will unilaterally
extend the credit period.
Inventory Management
5321.14 Just-in-time
a. A just-in-time (JIT) production system purchases raw materials and component parts
just as they are needed in the production process, thus reducing raw material inventory
close to zero. It is part of a manufacturing philosophy that promotes the simplest, least
costly means of production. Under ideal conditions, the company would receive raw
materials just in time to go into production, manufacture parts just in time to be
assembled into products, and complete products just in time to be shipped to
customers. JIT shifts the production philosophy from a push approach to a pull
approach.
(1) The push approach begins with the acquisition of raw materials that are pushed
into production through the various stages of production, and finally into finished
goods inventory. The goal is to have finished goods ultimately sold to customers
without deep discounting.
(2) The pull approach starts with the demand for finished goods from a customer. The
customer’s order triggers the production of the goods. Inventories are acquired
only to meet production needs.
b. Under ideal conditions, the company would attempt to create a situation where both
raw material inventory as well as finished goods inventory would be eliminated and
work-in-process inventory kept to a minimum.
c. Raw materials are delivered just in time to meet the production schedule, with a
guaranteed high level of product quality so that they can bypass inspection and go
directly to production. This requires long-term purchase agreements with suppliers and
coordination between purchasing and production. The result is a virtual elimination of
raw material inventory. The cost of carrying inventory is shifted to the supplier.
d. The rate of production for individual departments is determined by the needs of each
succeeding department. Inventory buffers with a predetermined size are used to signal
when a department should be working. The department works as long as the buffer is
less than the desired level. This eliminates most work-in-process inventory.
e. As soon as products are finished, they are immediately shipped to the customer, thus
eliminating finished goods inventory and immediately converting the product into cash
or accounts receivable. Close supplier/customer relations can allow both supplier and
customer to use JIT systems.
f. Generally there are changes required in several areas of the manufacturing process to
make JIT successful.
(1) Activities that do not add value need to be eliminated. Activities are identified as
value-adding activities and nonvalue-adding activities. A nonvalue-adding activity is
any activity that could be eliminated without detracting in any way from customers’
satisfaction with the final product. Moving time, inspection time, and times in
queues are all nonvalue-adding activities and should be eliminated as much as
possible. Machining, milling, and polishing of products are all value-adding
activities, as customer satisfaction would be reduced without them.
(2) Supplier relationships are of key importance. It is important to select a small
number of suppliers who can guarantee both the quality and timely delivery of raw
materials. Raw materials are usually delivered in small quantities as needed for
production. Responsibility for raw material quality is also shifted to the supplier.
(3) Manufacturing cells are developed. Traditional departments and processes are
replaced with manufacturing cells that contain families of machines typically spread
through several departments. A component or a complete product can be produced
within one cell rather than being transferred between several departments.
(4) The workforce is highly trained and flexible. The cellular manufacturing
environment requires a labor force that is capable of operating several different
machines. The workforce must be capable of servicing the machines and
performing inspections on finished products.
(5) Total quality control is implemented at the inventory and production level. The
streamlined flow of raw materials and components through the production process
does not allow for product defects. No defects can be allowed in parts or raw
materials received from suppliers, in work-in-process, or in finished goods.
Emphasis is placed on doing things right the first time and avoiding rework or waste
of any type. Products are inspected for quality at each stage of the process. When a
defect is discovered, the production process is halted until the problem is
corrected. Responsibility for inspecting products is shifted from inspectors to
production workers. There is a commitment for continuous improvement in all
aspects of the company.
(6) Set-up time is reduced. Set-up time is the time required to change equipment,
move materials, and obtain forms needed to shift production from one product to
another. Meeting customer demand on short notice without any finished goods
inventory requires the ability to shift production from one product to another with
a short set-up time.
g. When using JIT methods, the differences between net operating income using the
absorption method or variable costing method will be reduced since the differences
relating to the costing of inventory are eliminated with the elimination of the finished
goods inventory. If inventories are almost non-existent, then differences in inventory
valuations will be extremely small, bringing the results of net income using absorption
or variable costing extremely close.
h. Advantages of JIT production systems:
(1) Funds are no longer tied up in inventories and can be more effectively used
elsewhere.
(2) Warehouses formerly used to store inventory can be put to more productive use.
(3) As defects are reduced, less waste occurs, and customers’ satisfaction is increased.
(4) As production time is decreased, a greater output potential is created.
(5) Customers’ satisfaction is increased due to the shorter delivery time from the point
the order is received (order lead time).
i. Disadvantage of JIT production systems: Increased possibility of stockouts if delivery of a
critical part is delayed due to strikes, disasters, or other unexpected occurrences
2 DS
EOQ =
Ci
b. Relevant items in determining the economic order quantity:
D = The demand per year in units
S = Setup or ordering cost per order or batch
C = The cost per unit
i = The carrying costs expressed as a percentage of inventory cost
c. Inventory decision models contain costs that move in opposite directions. Carrying
costs increase as the size of the order increases. Set-up or ordering costs, however,
decrease as the size of the production run or order increases. The EOQ model seeks
to minimize the combined total of these two costs, as illustrated in the following
graph.
d. Assuming that there is constant demand and no change in lead time, the reorder point
is equal to the sum of expected units of demand during the lead time plus any required
safety stock. This relationship can be illustrated as follows:
2 DS
EOQ =
Ci
=
(2 )(40,000 )($115)
($50)(20% )
= 960 units
RP = D LT + SS
= [(40,000 units / 300 days )x 3 days ] + 500 = 900 units
and lost sales when the time to replace inventory exceeds the anticipated period. The
level of safety stock is determined by a balancing of the cost of holding additional
inventory items and the cost, or lost revenue, associated with not having goods
available for the customer when they wish to purchase them.
There is a cost to this additional credit—the lost $40,000 in discounts. The lost discounts
represent increased cost for the merchandise; therefore, the price of the additional
$108,889 of trade credit is $40,000. The cost of this credit is computed as follows:
Lost discount $40,000
= = 36.7%
Additional credit supplied $108,889
c. If a firm can borrow funds at an interest rate lower than the cost of the foregone
discount, then the funds should be borrowed and the discounts taken. If the firm above
can borrow $108,889 at 10% for an annual interest cost of $10,889, then $29,111 would
be saved on an annual basis ($40,000 - $10,889).
d. The return on taking discount can also be calculated using the following formula:
360 Percentage of discount
×
Total credit period – Discount period 100% - Percentage of discount
Since the discounts exceed the cost of the interest expense, net income is increased by
taking the discounts. In this case, income is increased by $17,467 or approximately 6%.
c. Assuming that the firm goes to the bank to finance a note payable so that funds are
available to take advantage of the discounts, the liability/equity sections of the balance
sheet will have the following comparison when taking versus not taking trade discounts:
Do Not Take
Take Discounts
Discounts
Accounts payable $163,333 $54,444
Note payable 108,889
Accruals 50,000 50,000
Long-term debt 75,000 75,000
Equity 250,000 250,000
Total liabilities and equity $538,333 $538,333
Thus, the total liabilities and equity will be the same. The only difference is that the firm
will be able to choose how to finance the amount needed to take advantage of the
discount for early payment and reduce the cost of financing—short-term loan, long-
term loan, or equity. This example was presented using a short-term note. The nature of
the spontaneity of the trade credit financing often causes these cost-saving options to
be missed.
b. Ratio analysis tends to be more useful for narrowly focused firms as it is difficult to
develop industry averages for large firms that have different divisions in different
industries.
c. Since the goal for most firms is to be better than average, merely attaining average
performance generally is not deemed to be satisfactory. One alternative is to attempt to
benchmark performance against industry leaders’ ratios.
d. Inflation may cause balance sheet distortion making book values significantly different
from market values. Inflation may also have a significant impact on depreciation
charges, inventory costs and/or profitability.
e. Seasonal factors might also distort ratio analysis. This problem can be minimized by
using monthly averages for the variables that are subject to seasonal fluctuation.
f. Analysts need to be aware that firms may use “window dressing” to attempt to improve
the appearance of their financial position. This could cause a firm to manipulate data to
show positive short-term effects that would, in reality, produce negative long-term
results.
g. Different accounting practices can distort ratios and/or comparisons between
companies. For example, a firm that leases assets rather than purchasing them can
artificially improve fixed asset turnover and debt ratios. Also, different fiscal years,
different markets, different locations (affecting taxation, labor costs, governmental
regulation, etc.), different sizes, different stages of development (growth vs. mature),
different dividend payment policies, use of LIFO vs. FIFO, differences in the use of
operational and financial leverage, and different industries can make intercompany
comparisons difficult.
h. There is often disagreement among analysts as to what constitutes a “good” or “bad”
value for a ratio. For example, underlying circumstances such as the availability of an
unused line of credit may allow for a particular company to have a low current ratio that
is not considered to be “inadequate” if the line of credit was not available.
i. Firms often have a mix of ratios, some of which would present a “good” picture and
others that might present a “bad” picture. This requires an analyst to use some form of
statistical technique such as discriminate analysis to assess the probability that the firm
would have financial trouble based upon the particular mix.
j. Some balance sheet and income statement accounts are based on estimates and
judgments (receivables and inventory), and some ratios are based upon arbitrary
assumptions (EPS).
k. For some ratios, there is a lack of uniformity in the formula used in calculating the ratio.
Different formulas for identical ratios may be used by different firms, as well as before-
versus after-tax figures as ratio components, thus leading to an inability to make inter-
firm comparisons. The use of a 360-day year (for computational convenience) as
opposed to a 365-day year also can lead to different results; however, the actual sales
days of the year for many nonretail companies is usually less than 300.
l. Accrual income does not necessarily indicate the availability of cash, and current
liabilities may understate short-term demands on cash due to such items as operating
leases.
m. Unless data from an adequate number of years is used, performance measures and
trends may be misinterpreted and/or misstated.
n. Some ratios include the use of external market values such as the price of stock and are
influenced by numerous variables over which management has little influence or
control.
12/31/20X2 12/31/20X1
Assets
Cash $250,000 $200,000
Net Accounts Receivable 200,000 100,000
Inventories 100,000 150,000
Total Current Assets 550,000 450,000
Net Plant and Equipment 400,000 450,000
Goodwill 100,000 100,000
Total Assets $1,050,000 $1,000,000
Sample Company
Income Statement
For the Year Ended December 31, 20X2
Liquidity
5322.04 Liquidity measures a firm’s ability to satisfy short-term obligations. It implies that the
company has the ability to turn assets into cash or the ability to obtain cash. It is also related
to the speed with which given assets can be converted into cash.
Lack of liquidity:
a. is a matter of degree as opposed to being strictly liquid or non-liquid.
b. may mean that an organization is unable to take advantage of favorable discounts.
c. may mean that an organization might be unable to take advantage of profitable
business opportunities.
d. limits management’s flexibility.
e. may cause lower profitability.
increasingly risky position, and banks might become hesitant to lend to the firm except
at higher rates.
d. The current ratio can easily be manipulated in the short term (“window dressing”);
however, a higher manipulated ratio is difficult to sustain over the long term. This is
done by using cash to pay current liabilities, thus increasing the current ratio.
e. In determining the adequacy of the current ratio, some consideration must be given to
the relationship between the current ratio and the operating cycle. If the cycle is
relatively short, a lower current ratio may be adequate. Conversely, a lower accounts
receivable turnover and/or inventory turnover would necessitate the need of a higher
current ratio to provide the same level of safety.
f. If a line of credit is available, a lower current ratio would be acceptable.
g. If LIFO is used and there are old LIFO layers in inventory, the current ratio will be
understated if inflation has occurred.
h. Calculation using data from the Sample Company for December 31, 20X2:
Current assets
Current ratio =
Current liabilities
$550,000
= = 2.75 to 1
$200,000
b. Since inventory is not included, the problems with inventory valuation are avoided as
well as the additional business cycle issues related to turning inventory into cash.
c. This ratio might still need to be adjusted if receivables are subject to a lengthy collection
period (dollars may not be available as needed), or if they might be factored at less than
the carrying value (fewer dollars will be available).
d. Calculation using data from the Sample Company for December 31, 20X2:
b. Very few companies would have a 1:1 cash ratio or higher; however, this is not
necessarily a bad thing since not all current liabilities need to be met today and more
cash will be available as additional sales are made and receivables are collected.
c. Calculation using data from the Sample Company for December 31, 20X2:
Cash + Cash equivalent + Short-term investments
Cash ratio =
Current liabilities
$250,000 + $0 + $0
= = 1.25 to 1
$200,000
5322.10 Number of days sales in trade accounts receivable (average collection period)
a. The number of days sales in trade accounts receivable measures the number of days on
average that it takes to collect accounts receivable.
360
Number of days sales in accounts receivable =
Accounts receivable turnover
Or
Average accounts receivable
=
Credit sales
360
b. This ratio can easily be compared to the credit terms in order to evaluate the
performance of the receivables department.
c. Significance is gained when comparing the ratio to credit terms or to industry averages.
b. Since inventories are purchased or produced for the purpose of obtaining a return, it is
extremely important to determine if corrective action needs to be taken. A low turnover
can be caused by a variety of factors, some of which are anticipated and based upon
management’s actions, and others that are unanticipated and require corrective action.
These factors include the following:
(1) Obsolete items
(2) Unanticipated weak demand
(3) Build-up to meet expected future demand or firm commitments
(4) Build-up to avoid future price increases
(5) Build-up in anticipation for a strike or material shortages
c. Excessive inventories can lead to:
(1) high storage and insurance costs.
(2) a greater risk of obsolescence, physical deterioration, or theft.
(3) tying up funds that could be used in a more productive fashion, thus leading to
decreased profitability.
d. A high turnover may be an indication that:
(1) the firm is managing its inventory well and thereby achieving higher productivity.
(2) the inventory that is held is highly marketable.
b. This ratio converts the inventory turnover into days and is also a good indicator of how
long it takes a company to turn inventory into cash.
c. Generally, a lower (shorter) DSI is preferred.
d. Calculation using data from the Sample Company for 20X2:
Number of days sales in = 360
inventory Inventory turnover
360
= = 75 days
4.8 times
Or:
= Average inventory = ($150,000 + $100,000) ÷ 2
Cost of sales ÷ 360 $600,000 ÷ 360
= 75 days
Solvency
5322.13 Solvency is the organization’s ability to meet long-term obligations (repayment of principal
as well as interest) through the use of cash or by converting non-cash assets into cash.
a. Solvency ratios are related to questions about the firm’s long-term viability.
b. Solvency is closely related to financial risk. As the financing of the organization shifts
from equity financing toward debt financing, the risk that the firm could default on its
long-term obligations increases.
Total debt
Total debt to total assets =
Total assets
= $200,000 + $450,000
= 62%
$1,050,000
b. The debt to equity ratio measures the company’s ability to meet long-term obligations.
As with the previous ratio, care must be used when comparing the results to other
organizations since there are a variety of definitions for the variables used in both the
numerator and denominator. The conservative approach is to use total debt as the
numerator.
c. From a creditor’s point of view, the lower the ratio, the less the risk of insolvency.
d. Calculation using data from the Sample Company for December 31, 20X2:
Total debt
Debt to equity =
Total stockholders’ equity
$200,000 + $450,000
= = 1.63 to 1
$100,000 + $200,000 + $100,000
b. A higher ratio is associated with greater solvency and demonstrates a greater margin of
safety for meeting fixed interest payments.
c. This ratio measures the ability to add new debt in the future.
d. A concern is that accrual income is not necessarily representative of the cash available
to pay the interest obligations.
e. Since interest is tax deductible, both interest and taxes are added back to net income
indicating the amount available to cover the fixed interest payments.
f. Calculation using data from the Sample Company for 20X2:
Times interest earned = Net income before tax + Interest expense
Interest expense
= $125,000 + $50,000
= 3.5 times
$50,000
d. Calculation using data from the Sample Company for the year ended December 31,
20X2, assuming that the depreciation expense for the period was $40,000:
EBIT + Depreciation
Cash coverage =
Interest
= $175,000 + $40,000
= 4.3 times
$50,000
b. As with TIE, a higher fixed charge coverage ratio is associated with greater solvency and
demonstrates a greater margin of safety for meeting fixed payments.
c. This ratio measures the organization’s ability to add new debt in the future.
d. Calculation using the data from the Sample Company for the year ended December 31,
20X2, assuming that the lease expenses are $10,000:
EBIT + Lease payments
Fixed charge coverage =
Interest + Lease payments
= $175,000 + $10,000
= 3.1 times
$50,000 + $10,000
Profitability
5322.19 Profitability (asset utilization) is the measure of success in terms of income defined in a
variety of ways over a period of time. In analyzing profitability, earnings are compared to a
base such as sales, productive assets, or equity. Increased profitability provides the potential
for increased dividend payments as well as from stock appreciation. Increased profitability
also provides greater security for debt holders.
b. This ratio is generally more useful to management than to creditors or investors since
the data necessary to analyze why changes occurred in the ratio are only available
internally. Changes are caused by one of the following elements or a combination of the
following items.
(1) Increase/decrease in unit sales price
(2) Increase/decrease in cost per unit
c. Calculation using data from the Sample Company for 20X2:
Gross margin in ratio = Gross margin
Net sales revenue
$400,000
= = 40%
$1,000,000
b. In alternative definitions, the numerator for this ratio can take a variety of forms—
operating income, net income, income available to stockholders, or income from
recurring operations.
c. The use of accrual income includes estimates and items over which management has
little or no control.
d. Calculation using data from the Sample Company for 20X2:
Net income
Profit margin ratio =
Net sales
$75,000
= = 7.5%
$1,000,000
b. As in all turnover ratios, a high asset turnover is a desirable outcome since it results
from an effective use of available assets; however, it is not unusual for a company that
has recently expanded to see a drop in asset turnover until the new facility is fully
utilized.
c. For internal use, it is desirable to calculate this ratio comparing divisional sales to
divisional assets.
d. Comparison between firms within an industry can be difficult due to the fact that the
age of the assets used in the production of the sales may vary. This is a limitation caused
by the use of historical costs as opposed to fair market values for the productive assets.
There is also a problem when comparing a capital-intensive firm to a labor-intensive
firm.
e. There are also problems created due to the fact that the generally used formula does
not take into account that certain assets make no tangible contribution to sales. It is
assumed that an asset’s participation in generating sales is related to its recorded
amount. Adjustments to the asset figure are used in attempts to compensate for this
problem. For example, long-term investments are not involved in the production and
sale of the product and are often not included in the calculation.
f. Calculation using data from the Sample Company for 20X2:
Net sales revenue
Asset turnover =
Average total assets
$1,000,000
= = 0.98
$1,050,000 + $1,000,000
2
b. The DuPont equation breaks ROA down to the asset turnover and profit margin.
Net income Net sales
Return on assets = ×
Net sales Average total assets
Through the use of the DuPont equation, ROA is broken down to the profit margin
showing the effective control of cost (Net income/Net sales) and the asset turnover
showing the efficient use of assets (Net sales/Average total assets).
c. Since neither profit margin nor asset turnover can increase indefinitely, at some point
the only way to create further increases in earnings is to increase the asset base by
adding to production capacity.
d. Limitations specific to this ratio include the use of accrual net income that is subject to
estimates and does not reflect actual cash return. The rate of return is also based upon
historical asset costs that do not reflect current market values.
e. The numerator can use a variety of income figures, including EBIT. The philosophy
behind the use of EBIT is that the earnings available to three parties, stockholders,
creditors, and the government (taxes) is created through the use of the assets available
to the organization. The denominator is often restricted to only income-producing
assets by excluding long-term investments and intangibles.
f. Calculation using data from the Sample Company for 20X2:
Net income
Return on assets = Average total assets
$75,000
= 7.3
($1,050,000 + $1,000,000) ÷ 2
b. The denominator consists of the average of the total equity less preferred shares and
minority interest. One of the major problems with this ratio is that the historical issue
price is used as opposed to the current market value.
c. The common stockholders’ leverage ratio (Average assets/Average stockholders’ equity
used above) measures the portion of assets that are financed through common equity
(also called the equity multiplier). The larger this ratio, the greater the financial leverage
will be. Various organizations within an industry will have potentially dramatic
differences in ROE resulting from financing decisions—debt vs. equity. If ROA is greater
than the cost of borrowing, then ROE will be higher than ROA due to the impact of
financial leverage.
d. Calculation using data from the Sample Company for 20X2:
Market Measures
$17.00
= = 5.67
$3.00
$10,000
= = 16.67%
$75,000 – $15,000
In other words, if the required rate of return is 12% and the company earns 16%, then
value has been created for the shareholders; therefore, it could be said that the
investors value the firm above the amount of capital that was originally invested.
b. Another way of looking at EVA is that a cost is assigned to the equity capital in addition
to the cost of debt. On the traditional income statement, interest expense represents
the cost of debt; however, there is nothing to denote the cost of equity. Since EVA
focuses on economic profits, the opportunity cost of the equity capital must be taken
into consideration. For example, if the shareholders could earn 12% on an investment of
similar risk, then this opportunity cost of 12% must also be accounted for. The cost of all
capital employed in the first would be considered to be the weighted-average cost of
capital (WACC). NOPAT is the after-tax profit of operating income and is equal to the
operating profit times (1 – Tax rate).
EVA = NOPAT – (Cost of all capital × Total assets)
c. The EVA concept is often used by firms during the decision-making process when
determining whether a large investment in plant and equipment would be in the best
interest of the shareholder.
d. Calculation using the data from the Sample Company for the year ended December 31,
20X2, assuming that the WACC is 10%:
EVA = NOPAT – (Cost of all capital × Total assets)
= ($175,000 × (1 – 0.40)) – (10% × $1,000,000)
= $5,000
financing of $300,000 has been spontaneously supplied to that firm providing the payment
terms are met.
b. A lockbox system can significantly reduce the time required to receive funds and make
them available for use, often by two or three days, since multiple lockboxes can be
located throughout the United States and internationally.
c. Today, the bank generally notifies the firm with a daily record of receipts collected via
an electronic data transmission system and wires funds to the firm’s cash concentration
account.
(3) Ability of the valuation object to stand alone or the need to function as part of a
unit
(4) Location of the valuation object
(5) Restriction placed on the use or sale of the valuation object
c. In determining the price under a fair value assumption, the conditions under which the
hypothetical sale are assumed to occur would be orderly and not under conditions of
liquidation or distress. The seller would be able to take advantage of normal market
activities and would have the ability to take enough time prior to the date of
measurement to take advantage of normal exposure to the market given the item being
valued.
d. Under fair value assumptions, the hypothetical transaction is considered to have
occurred in either the principal market for the subject transaction or, in the absence of
such a market, in the most advantageous market for that transaction. The principal
market is considered to be where the holder of the transaction asset or liability could
locate the greatest volume of similar transfers. The most advantageous market is
considered to be where either the sales price for an asset can be maximized or the
transfer costs for a liability can be minimized. Note that these hypothetical transfers are
considered from the perspective of the holder of the asset or liability. Also, if there is a
principal market for the subject item, then the fair value measurement is determined
using that market even if the most advantageous market would yield a more beneficial
transfer for the holder.
e. The market participants to be considered when determining fair market value are the
actual holder of the asset or liability being valued and the hypothetical buyer in the
principal (or most advantageous) market. Characteristics for both the buyer and seller
would include:
(1) buyer and seller are not related parties (independence).
(2) both parties are knowledgeable about the item being transferred and have access
to usual and customary information.
(3) both parties have the ability to complete the transaction.
(4) both parties are willing to complete the transaction but are not being forced to do
so.
f. A fair value measurement assumes that the item being valued will be put to the highest
and best use resulting in the maximization of value that would be physically possible,
legally permissible, and financially feasible. Note that the highest and best use is not
determined from the perspective of the current holder but by the hypothetical market
participants.
g. The fair value determination of an asset uses either an in-use or an in-exchange
premise. An in-use premise assumes that the maximum value from the purchaser’s
perspective of the subject asset is when it is used in conjunction with other assets as a
group. An in-exchange premise assumes that the maximum value from the purchaser’s
perspective of the subject asset is when it is used alone.
h. The fair value determination of a liability assumes that the liability will continue to exist
after the transfer and maintain the same risk of nonperformance after the transfer as
before.
obtained from sources independent of the reporting entity used in pricing an asset or
liability. When possible, it is expected that more observable inputs will be used than
unobservable inputs.
c. Unobservable inputs are the least desirable ones to use when determining fair value.
Unobservable inputs are those based upon assumptions developed using the best
information available in the circumstances reflecting the reporting entity’s own
speculations used in pricing assets or liabilities.
d. Level 1 inputs are observable and considered to be the highest level and most desirable
when determining fair value. Level 1 inputs would include unadjusted quoted active
market prices of identical assets. An active market is where there is adequate trading to
provide constant market data.
(1) Generally accepted business valuation techniques often apply a blockage discount
to the quoted market price when valuing a large block of stock, since it is assumed
that the market price would drop if a large block of stock was put on the market on
a particular date. FASB ASC 820 specifically disallows the use of such a blockage
discount.
(2) It is not unusual for a company to announce information that will have a negative
impact on market price at the end of the day or shortly after the close of the
market. If the quoted market price is adjusted for the new information, then the
fair value should be listed at the lower price.
e. Level 2 inputs are considered to be the middle level when determining fair value, and
include such items as prices of assets or liabilities that can be either directly or indirectly
observed but do not include the Level 1 quoted market prices. Often market values of
similar assets or liabilities (as opposed to identical) are used to determine fair value. In
such cases, the degree of comparability must be determined, and often adjustment
must be made when determining fair value. A higher degree of adjustment could cause
the fair value measure to fall to a Level 3 category. Level 2 items would include such
items as the following:
(1) Active market prices for similar assets or liabilities
(2) Available market prices for identical or similar assets or liabilities when markets are
not active
(3) Inputs such as interest rates, default rates, yield curves, and credit risks
(4) Other market-corroborated inputs
f. Level 3 inputs are unobservable and considered to be the lowest and least desirable
level when determining fair value.
(1) Level 3 inputs should be used to develop an exit price to represent fair value from
the perspective of the reporting entity.
(2) Level 3 inputs should reflect the reporting entity’s market assumptions.
(3) Level 3 inputs should be based upon the best available information that can be
obtained without undue cost and effort.
(2) provides enough information so that the use of significant unobservable inputs
(Level 3) can be assessed.
b. When a fair value measurement is done on a recurring basis, the following information
needs to be disclosed:
(1) The fair value measurement as of the reporting date
(2) The level within the hierarchy (Level 1 market information, Level 2 other significant
observable inputs, and/or Level 3 significant unobservable inputs) of the inputs
used in the fair value measurement
(3) Reconciliation of the beginning and ending balances when using significant
unobservable inputs (Level 3)
(4) The amount of the total gains or losses that are attributable to the change in
unrealized gains or losses still held at the date of measurement
(5) The valuation method(s) used
(6) A discussion of any changes in valuation techniques, if any, within the period
c. When a fair value measurement is done on a nonrecurring basis, the following
information needs to be disclosed:
(1) The fair value measurement as of the reporting date and the reason for the
measurement
(2) The level within the hierarchy (Level 1 market information, Level 2 other significant
observable inputs, and/or Level 3 significant unobservable inputs) of the inputs
used in the fair value measurement and a description of any Level 3 inputs used
(3) Reconciliation of the beginning and ending balances when using significant
unobservable inputs (Level 3)
(4) A discussion of the valuation method(s) used
(2) Fair market value defines the seller as hypothetical, whereas there is a specific
seller when using fair value.
(3) Fair market value takes advantage of an unrestricted market, whereas fair value
uses the principal or most advantageous market.
d. Per the International Glossary of Business Valuation Terms, intrinsic value is defined as
follows:
“The value that an investor considers, on the basis of an evaluation of available facts, to
be the ‘true’ or ‘real’ value that will become the market value when other investors
reach the same conclusion.”
e. Per the International Glossary of Business Valuation Terms, replacement cost new is
defined as “the current cost of a similar new property having the nearest equivalent
utility to the property being valued.” Reproduction cost new is defined as “the current
cost of an identical new property.”
f. Per the International Glossary of Business Valuation Terms, investment value is defined
as “the value to a particular investor based on individual investment requirements and
expectations.”
use any approach or method available. The results are expressed in terms of a
calculated valued and can be either a single number or a range.
(3) When in liquidation, equipment on hand is not likely to be of the most recent
technology; therefore, any disposal price would tend to be low. A going concern
would continue to use the equipment on hand to create products to sell.
(4) When in liquidation, leasehold improvements would be of no further use and could
not be sold. A going concern would continue to use leasehold improvements.
(5) When in liquidation, no goodwill would exist. A going concern would potentially
continue to have goodwill associated with its good name, reputation, and brand.
(6) When in liquidation, the costs of asset disposal and the administration of the
liquation process can be costly. A going concern would not have any costs related to
a liquidation process.
d. When the premise of value is liquidation, the valuator will assume that historical records
will provide little valuable information needed to predict future inflows and outflows for
the organization.
e. When the premise of value is a going concern, the valuator will generally base future
predictions at least in part on the firm’s historical performance.
f. The assumption related to the premise of value must be translated into estimates of
future inflows and outflows.
negative impacts on the subject of the valuation such as general market risk, labor
efficiency, attrition, availability risk, supplier risk, and competitive risk. The higher the
risk level, the lower the value of the subject of the valuation.
(3) items that are considered to be part of normal operations are not included in the
financial statement of the subject company.
(4) an unusual item not considered to be recurring is included in the historical
statements used to predict the future.
(5) discretionary items not considered to be normal business expenses are included in
the financial statements. These would include such items as personal expenses run
through the business, such as family members on the payroll being paid above-
market wages for the work done; personal vacations; and other items of a private
as opposed to a business nature.
d. The valuator may need to make normalization adjusts for one or more of the following
items:
(1) Nonoperating adjustments: The removal of nonoperating items included in the
historical financial statements that are not assumed to be part of normal operations
(2) Nonrecurring adjustments: The removal of unusual, unexpected, or infrequent
items not likely to occur again from the historical financial statements
(3) Comparability adjustments: Adjustment of the historical financial statements to
match GAAP choices of potential guideline companies. For example, a company that
used a LIFO inventory could not be compared to one using a FIFO inventory.
(4) Discretionary adjustments: Adjustments to the historical financial statements to
include or remove items considered to be part of normal operations. For example, it
is not unusual for owners of small, private companies to pay either excessive or
below-average salaries to themselves. It is also common for personal expenditures
to be run through the business.
company. The further out the projections go, the less likely it is that those projections
will be accurate. Many valuators do not like to use increasingly speculative projections
going out more than five years.
d. When using an income approach, the present value of the terminal value at the end of
the projection period used needs to be calculated. Here the valuator must decide
whether the company is expected to still be a going concern at the end of the projection
period before calculating terminal value. If not, terminal value would be the present
value of the liquidation value of the net assets at the end of the projection period. If the
company would be expected to continue as a going concern at the end of the projection
period, then terminal value would be based upon the projected cash flows during the
first year of the terminal period, the expected growth rate into perpetuity, and the
discount rate. Obviously, there are a number of assumptions that need to be examined
closely.
e. A discount rate is the total required rate of return that an investor would expect on an
investment given the risk level inherent in that investment. The discount rate (also
known as the cost of capital) is extremely critical in the valuation process since it is used
to convert projected income streams to present value. The higher the risk, the higher
the discount rate, and the lower the present value of the subject company. Discount
rates reflect the risk involved in a particular investment. Techniques used to develop a
discount rate include the capital asset pricing model (CAPM), build-up methods, and the
weighted-average cost of capital (WACC). Again, many assumptions related to the
development of discount rates need to be examined closely.
Black-Scholes
5333.03 Capital asset pricing model (CAPM): The capital asset pricing model, also called the security
market line (SML), attempts to quantify the relationship between investment risk and the
rate of return on the investment. The CAPM model assumes that the individual investment
contains two types of risk, and is primarily concerned with measuring the former:
1. Systematic risk: Market risks that cannot be diversified away; examples include interest
rates, recessions and wars.
2. Unsystematic risk (also known as specific risk): This risk is specific to individual stocks
and can be diversified away. In other words, it represents the component of a stock's
return that is not correlated with general market moves.
5333.04 The return on an individual stock, or a portfolio of stocks, should equal its cost of capital. The
CAPM formula is as follows:
a. Risk-free rate of return: A default-free rate of return is estimated as the rate of return
on long-term (i.e., 10 years) U.S. Treasury bonds.
b. Market risk premium: The market risk premium is the additional return over the
default-free rate required for the average stock.
c. Rate of return estimates: The required rate of return is the default-free rate of return
plus the beta times the market risk premium.
5333.05 Beta coefficients: CAPM assumes that the beta coefficient is the only relevant measure of a
stock’s risk. It measures a stock’s volatility (i.e., how much a stock’s price increases or
decreases compared to movement in the stock market). If a given stock moves precisely as
the overall market, it is said to have a beta of 1.
5333.06 Dividend discount model (DDM): A procedure for valuing the price of a stock by using the
predicted dividends and discounting them back to the present value. If the value obtained
from the DDM is higher than what the shares are currently trading at, then the stock is
undervalued.
a. There are many variations in this procedure. The most common DDM is the Gordon
growth model (GGM), computed as follows:
b. Limitation: The model assumes a constant dividend growth rate in perpetuity. This
assumption is generally true for very mature companies, but newer companies have
fluctuating dividend growth rates in their earlier years.
5334.02 In a finance context, asset management is the act or practice of an investment advisory firm
making investment decisions on behalf of a client. The professional manager of various
securities and other assets works to meet the specified investment goals of the client. The
purpose for choosing to use a firm to act in an advisory capacity is to obtain access to a
broader range of investment vehicles and acquire the knowledge and expertise of the asset
managers. Due to the cost of this service, it is generally limited to high-wealth individuals.
This is also sometimes known as investment management.
Portfolio Management
5334.03 Portfolio management moves to the next level and involves deciding which assets to include
in the portfolio given the goals of the investor and the changing economic conditions within
which the portfolio is being managed. Portfolio management decisions are often judged by
some performance measure relative to the goals and circumstances of the investor, and
investors differ in the degree to which they are risk adverse.
Investment Process
5334.06 Execution of the choices by trading in the market is also a major consideration. Questions
that must be addressed include the following:
a. How often should one trade? A key element here would be commission costs.
b. How large will the trades be? What are the bid/ask spreads on the trades, and what
would be the price impact of the size of the trade?
c. Should derivatives such as options, futures, or swaps be used to hedge the portfolio?
5334.08 The Treynor index was the first index developed as an objective measure that could apply to
all investors whatever their risk preferences. Treynor suggests that there are two
components of risk: those produced by fluctuations in the market, and the risk of fluctuation
of individual stocks.
a. Treynor developed the concept of the security market line, which defines the
relationship between portfolio returns and security returns where the slope of the line
measures the volatility between the portfolios and the market (represented by β). Beta
(β) represents the systematic risk of the portfolio, the risk inherent to the entire market
or entire market segment (also known as nondiversifiable risk or market risk).
b. The β coefficient is the volatility measure of a stock portfolio or the market itself.
c. The greater the slope of the line, the better the risk-return trade-off.
d. Treynor index = (Portfolio return – Risk-free rate) ÷ β, with the resulting value
representing the portfolio return per-unit risk.
5334.09 The Sharpe measure is very similar to the Treynor index, except that the risk measure is the
standard deviation of the portfolio. The ratio follows the capital asset pricing model and
compares portfolios to the capital market line.
Sharpe measure = (Portfolio Return – Risk-free rate) ÷ Standard deviation
(The capital asset pricing model (CAPM) is the risk of price change due to the unique
circumstances of a specific security, as opposed to the overall market that can be virtually
eliminated from a portfolio through diversification. The capital market line is used in the
CAPM to illustrate the rates of return for efficient portfolios depending on the risk-free rate
of return and the level of risk (standard deviation) for a particular portfolio.)
5334.10 The Jensen measure is a measure of absolute performance on a risk-adjusted basis and
measures the portfolio manager’s predictive ability to achieve higher return than expected
for the level of risk accepted in the portfolio.
Return on the portfolio = Risk-free rate + (Return on market index – Risk-free rate) × β
(β represents the systematic risk of the portfolio, the risk inherent to the entire market or
entire market segment, also known as nondiversifiable risk or market risk.)
f. Cost of capital refers to an overall cost of obtaining investment funds. While different
techniques may be used in computing cost of capital, it is generally agreed that a cost
factor should be assigned to both debt and all stockholders’ equity (including retained
earnings). Some authorities suggest using a weighted-average cost of capital for capital
budgeting purposes. Cost of capital is the desired or target rate of return used in the net
present value method of discounted cash flow computations. It is also the minimum
acceptable (cutoff) rate used in choosing among projects employing the time-adjusted
rate of return method, described below (section 5335.07-.08).
g. There are several techniques for capital budgeting. These approaches are often used,
first to screen project possibilities, and second to rank existing choices. Those commonly
used techniques include the following:
(1) Payback method
(2) Discounted payback
(3) Accounting rate of return
(4) Net present value
(5) Internal rate of return
(6) Profitability index
(7) CVP analysis/margin of safety
F = P(1 + i)n
Factors have been computed for various combinations of i and n where the interest
factor is specified as fin = (1 + i)n and are contained in future value tables. A portion of a
future value table is shown below.
Period 8% 10% 12%
1 1.0800 1.1000 1.1200
2 1.1664 1.2100 1.2544
3 1.2597 1.3310 1.4049
4 1.3605 1.4641 1.5735
5 1.4693 1.6105 1.7623
When using the factors from the table, the equation F = P(fin) is used.
Illustration: How much will $10,000 be worth in four years if the interest rate is 10%?
Formula solution: F = P (1 + i)n
= $10,000 (1 + .10)4
= $14,641
Table solution: F = P(fin) when i = 10% and n = 4
= $10,000 (1.4641)
= $14,641
c. Present value of a $1 sum: The method for computing the present value or discounted
value of a $1 sum is just the opposite of future value. The question is, what amount (P)
must be invested today at a given interest rate (i) to be worth a specified amount (F) in a
given number of years (n)? The present value equation can be solved for P to obtain the
following:
1
P = F
(1 + i )n
…where the PV factor for a sum received in the future is specified as:
1
Pin =
(1 + i )n
The present value tables contain various combinations of i and n. Note that Pin is the
reciprocal of fin. A portion of the PV tables is shown below.
Period 8% 10% 12%
1 0.9259 0.9091 0.8929
2 0.8573 0.8264 0.7972
3 0.7938 0.7513 0.7118
4 0.7350 0.6830 0.6355
5 0.6806 0.6209 0.5674
Illustration: An investment will pay $15,000 in four years. What is the present value of
the payout if the interest rate is 10%? In other words, what amount of money would
have to be invested today to pay $15,000 in four years?
Formula solution:
1
P=F x
(1 + i )n
1
= $15, 000 x
1.104
= $15, 000 x (0.6830)
= $10, 245
Table solution:
P = F(Pin) where i = 10% and n = 4
= $15,000(0.6830)
= $10,245
d. Annuity: An ordinary annuity (annuity in arrears) is a series of equal cash flows received
at the end of equal intervals of time. An annuity due (annuity in advance) is a series of
equal cash flows received at the beginning of equal intervals of time. The difference
between these two cash flows is that no interest is computed on the last payment of an
annuity in arrears. When dealing with future value/present value tables, care must be
taken that the correct annuity table is used.
An example of an annuity in arrears (regular, ordinary, or deferred annuity) with three
annual payments of $1,000 each would be:
Time Period Payment
0 0
1 $1,000
2 $1,000
3 $1,000
An example of an annuity due (annuity in advance) with three annual payments of
$1,000 each would be:
Time Period Payment
0 $1,000
1 $1,000
2 $1,000
3 0
e. Future value of an ordinary annuity: The future value of an ordinary annuity represents
equal annual payments received at the end of the period that are accumulated and earn
interest. In other words, what will be the value of a series of payments that are
compounded at a given rate of interest at a point in the future? There will be no interest
earned until the first payment is made, and no interest will be earned on the last
payment. (FVA = future value of an annuity, PMT = the payment)
(1 + i )n − 1
FVA = PMT
i
As with future value and present value of a $1 sum, there are tables available to speed
the calculation of the future value of an annuity. A portion is shown below:
Period 8% 10% 12%
1 1.0000 1.0000 1.0000
2 2.0800 2.1000 2.1200
3 3.2464 3.3100 3.3744
4 4.5061 4.6410 4.7793
5 5.8666 6.1051 6.3528
Illustration: What is the future value of a three-year, ordinary $1,000 annuity when
money is worth 10%?
Formula solution:
(1 + i )n − 1
FVA = PMT
i
(1 + .10 ) − 1
3
= $1, 000
.10
= $1, 000(3.3100)
= $3, 310
Table solution:
FVA = PMT(FVIFAin) where i = 10% and n = 3
= $1,000(3.3100)
= $3,310
The future value of an annuity due has a similar formula to the one above. Remember,
an annuity due has the cash flow occurring at the beginning of the period.
(1 + i )n − 1
FVAdue = PMT (1 + i )
i
There are also annuities due tables available in order to make calculations easy.
f. Present value of an annuity: The present value of an annuity is the value today of a
future series of payments discounted at a particular interest rate. The present value of
an annuity can be computed by restating each of the annuity amounts to the present
time period using the present value of an annuity formula or the factor from a table of
present values of an annuity of $1. Remember to be careful to determine whether an
annuity in arrears (ordinary annuity) or an annuity in advance (annuity due) is
appropriate.
The formula for the present value of an ordinary annuity is:
1 − (1 + i )− n
PVAn = PMT
i
There are tables available to speed the calculation of the present value of an annuity. A
portion for an annuity due is shown below:
Period 8% 10% 12%
1 0.9259 0.9091 0.8929
2 1.7833 1.7355 1.6901
3 2.5771 2.4869 2.4018
4 3.3121 3.1699 3.0373
5 3.9927 3.7908 3.6048
Illustration: What is the present value of a $1,000 annuity received at the end of the
year for three years when interest is 10% and payments are made at the end of the
period?
Formula solution: PVAn = $1,000 ( 1 - (1 + i)n )
i
= $1,000 ( 1 - (1 + .10)3 )
.10
= $1,000 (2.4869)
= $2,489
Table solution: PVAn = PMT (PVIFAin) where i = 10% and n = 3
= $1,000 (2.4869)
= $2,489
The present value of an annuity due has a similar formula to the one above.
Remember, an annuity due has the cash flow occurring at the beginning of the period.
1 − (1 + i )− n
PVAdue = $1,000 (1 + i )
i
There are also annuities due tables available in order to make calculations easy.
g. Compound interest: When dealing with capital budgeting problems, assume that
interest is compounded annually. However, it is not unusual to have interest stated at
an annual rate but paid semi-annually (bonds), quarterly, or even monthly. In these
cases, the interest rate and number of periods must be adjusted when making the
calculation by using the following steps:
(1) Adjust the interest rate according to the compounding period. For example, 16%
interest compounded semi-annually is 8% each period.
(2) Adjust n from the number of years to the number of compounding periods. For
example, five years compounded semi-annually is 10 periods.
Problem: Compute the payback period for an investment opportunity that costs
$20,000 and provides equal annual cash flows of $4,000 per year for eight years.
Solution: $20,000
= 5 years
$4,000
Notice how the payback method does not even consider the cash flows between the
end of the payback period and the end of the investment’s useful life—in this case,
between years 5 through 8.
b. The average annual profits (or increased income) is the numerator. This method does
not differentiate between projects that have high early cash inflows and projects that
have high late cash inflows. Thus, the time value of money is not taken into account.
c. There is not unanimous agreement as to how the investment should be defined. Some
alternatives include (a) initial cost, (b) average book value, or (c) annual book value.
Either (a) or (b) is preferable to (c). The use of annual book value results in an increasing
rate of return over the life of the investment, providing the accounting income was
relatively stable over the life of the project, since the book value would decrease as
accumulated depreciation increases.
d. When using this method as a screening tool, the accounting rate of return must be
equal to or greater than the company’s hurdle rate for the project to be acceptable.
e. This method is popular due to consistency with financial reporting techniques
commonly used to evaluate company and divisional performance; however, it does
bring all of the accrual income problems into the formula. It is also simple and easy to
understand.
f. Illustration: Accounting rate of return method (simple rate of return)
Problem: Compute the accounting rate of return for an investment opportunity that
costs $100,000 and provides equal net cash inflows of $20,000 per year for 10 years.
Straight-line depreciation is used in calculating net income; therefore, annual
depreciation will be $10,000 ($100,000 ÷ 10 years). The average investment will be the
initial investment divided by 2, or $50,000.
Solution:
Net cash inflow - Depreciation
Accounting rate of return =
Investment
$20,000 - $10,000
=
$50,000
= 20%
b. When several investment proposals with reasonably similar investment sizes are being
considered, they can be ranked by net present values. The proposal with the highest
NPV should be chosen. An implicit assumption of this method is that all net cash inflows
can be reinvested at the discount rate used in computing NPV. If the proposals have
unequal investments, they should be ranked according to their profitability index. The
profitability index is computed by dividing the present value of the cash flows by the
initial investment.
The net present value method can also be used when examining cost-saving projects.
For example, when considering projects that would replace equipment with newer,
more efficient models, the costs with the lowest NPV would be the most desirable
alternative.
c. Illustration: Net present value
Problem: An investment costing $20,000 will reduce production costs by $5,000 per
year. The useful life will be eight years, and there will be a zero salvage value. The
desired minimum rate of return of 12% is used for capital budgeting purposes. Should
this investment be considered?
Solution: The NPV is calculated by determining the PV of the $5,000 annual cash savings
for eight periods using a 12% rate. The present value factor for eight periods at 12% is
4.968.
Net present value:
Cash flows ($5,000 × 4.968) $ 24,840
Initial investment (20,000)
Net present value $ 4,840
Since the NPV is negative, this is not a desirable investment at a 14% return.
e. The discount rate (or hurdle rate) is the required internal rate of return for projects
considered by a company or investor. This rate is often based upon a company’s
weighted-average cost of capital, incorporating a risk premium related to the riskiness
of the project in question. In other words, this rate must reward the investor for the risk
that is being assumed when undertaking the investment.
5335.07 Internal rate of return (time-adjusted rate of return)—equal annual cash flows
a. The internal rate of return (IRR) is the rate of discount (interest) that equates the
present value of the net cash flows including the initial cash outlay to zero. This method
can be used when the cost of the investment and annual cash flows are known, and the
calculated rate of return is then compared to the firm’s desired rate of return (hurdle
rate). If the return is less than the desired rate of return, the project should be rejected.
This method is also known as the time-adjusted rate of return.
Investment required
= Factor of the internal rate of return
Net annual cash inflow
b. After the factor of the internal rate of return is calculated, the present value tables of an
annuity are used in order to determine the rate of return that equates the present value
of the cash inflows and outflows to zero. Since the discount factors frequently do not
coincide with the annuity table, extrapolation is often necessary.
c. The IRR can be interpreted as the highest rate of interest an investor could pay for
borrowed funds to finance the investment being considered and be no worse off than if
the investment were not undertaken. In other words, the investment equals the present
value of the payoff.
d. When more than one project is being considered, they are ranked according to their
projected rate of return. Those with the highest rate of return are the most desirable.
An implicit assumption of this method is that all net cash inflows from the project under
consideration can be reinvested at the computed rate of return, and the method is,
therefore, a less reliable method than the NPV method for ranking investments. This
method is a good tool for making an accept/reject decision for potential projects that
have met the initial requirements for consideration. Using this method, all projects that
have a return greater than the company’s hurdle rate should be given further
consideration.
e. The NPV and IRR methods will potentially rank projects differently if the initial
investment differs and/or the timing of the cash flows differs between the projects. The
NPV method is the better method since the reinvestment assumption is the cost of
capital as opposed to the IRR assumption of reinvestment at the computed rate of
return on the cash flows for the particular project.
f. The payback reciprocal can be used to approximate a project’s internal rate of return if
the cash flow pattern is relatively stable.
g. Illustration: Internal rate of return
Problem: The company is considering the acquisition of a new machine that costs
$20,000. It will provide a savings of $5,000 per year over its useful life of eight years.
The salvage value is expected to be zero. What is the internal rate of return, and should
the machine be acquired?
$20,000
=
$5,000
5335.08 Internal rate of return (time-adjusted rate of return)—unequal annual cash flows
a. The internal rate of return (IRR) is not easily applied when the cash flows are not equal
annuity payments. A trial-and-error method must be used to obtain the rate of return
when hand-calculated; however, computer software is available to compute the IRR
that frequently uses an iterative trial-and-error method. Required inputs to such a
program are cash flows (both inflows and outflows) by period and a seed internal rate of
return. The computer program uses the seed (rough approximation) IRR on the initial
calculation and refines the rate on subsequent calculations until it has a rate that
produces a net present value of zero (or very close to zero).
b. Illustration: An investment opportunity costs $10,000 and returns $5,000, $4,000, and
$3,000, respectively, the first three years. What is the internal rate of return (IRR)?
(1) First attempt: Using an interest rate of 10%, the net present value is $105,
computed as follows:
Cash flows:
$5,000 × .9091 $ 4,545.50
4,000 × .8264 3,305.60
3,000 × .7513 2,253.90
Present value cash inflows 10,105.00
Cash investment (10,000.00)
Net present value $ 105.00
(2) Second attempt: Using an interest rate of 11%, the net present value is ($55.50),
computed as follows:
Cash flows:
$5,000 × .9009 $4,504.50
4,000 × .8116 3,246.40
3,000 × .7312 2,193.60
Present value cash inflows 9,944.50
Cash investment (10,000.00)
Net present value ($ 55.50)
Since the NPV is a negative amount, the interest rate is somewhat less than 11%. By
extrapolation, a rate of approximately 10.7% is obtained. If the desired minimum
rate of return is greater than 10.7%, the project should be rejected.
Using this alternative formula and the above data for Project 1, the profitability index
would be calculated as:
$20,250
+1 = 0.41 + 1 = 1.41
$50,000
f. The project profitability index is an additional variation of the profitability index. When
an organization with limited funds is considering projects that require different initial
investments, the NPV of one project cannot be directly compared to the NPV of another
potential project. If the NPV of each project is divided by the required investment of
that project, the project profitabilities can then be compared. The project with the
highest project profitability index would be the most desirable project.
Net present value of the project
Project profitability index =
Investment required
If machine hours were the constraint (bottleneck) and multiple segments of the
organization could use the limited machine hours to produce incremental profits, then
by determining the profitability of one unit of the constraint, management can
determine the most profitable way to use the constraint.
This formula is really the basis for the formula shown in section (a) to be used for
projects of a more long-term nature. For that formula, the incremental profits from the
segment are defined as the positive net present values of the cash flows after the initial
investment, and the amount of the constrained resource is defined as the investment
required by the project.
i. Example: The following information is available regarding incremental profits that could
be produced by two different segments using a constrained resource (machine hours).
Product 1 Product 2
Incremental profit possible $250,000 $400,000
Amount of machine hours necessary to produce
the incremental profit 100 hours 200 hours
c. Margin of safety is the excess sales over the breakeven sales point. This can be
important when looking at projects with projected sales. The methods used to evaluate
capital budgeting decisions use projected sales (savings). If, however, the projections
are incorrect, then there could be significant changes in the calculations of the payback
period, NPV, and IRR. When evaluating a project, the likelihood that the
sales/production projected will be achieved needs to assessed.
Illustration: Using the above illustration, calculate the margin of safety:
Margin of safety = 100,000 projected unit - 25,000 breakeven units
= 75,000 units
This could also be expressed in term of dollars. The margin of safety for sales is
$375,000 ($5 per unit × 75,000 units). In this case, sales could fall significantly (by 75%)
and there would still be a profit.
(4) The salvage value of the new machinery has no tax consequences in Year 6 since
the salvage value will equal book value. The cash flows for the six years will look
like:
Period Cash Flow
0 ($25,200)
1 7,600
2 7,600
3 7,600
4 7,600
5 7,600
6 7,600 + 6,000
The calculation of net present value at 20% for six years:
Cash savings
Annual: $7,600 × 3.326 = $25,278
Salvage, Year 6: 6,000 × 0.335 = 2,010
Present value of cash flows 27,288
Less: Initial investment 25,200
Net present value $ 2,088
Since the NPV is positive, replacing the old machine would be a desirable
investment.
5411.03 IT advancements make it possible for an organization to strengthen its internal control
structure.
a. New controls were developed to eliminate or minimize the risks introduced when
companies moved to a mainframe-based computer system from a paper-based system.
b. New controls were also developed when businesses moved to a distributed,
client/server system from a centralized, mainframe-based system.
c. Now, new controls are needed to control the risks that arise as we move into an
Internet-based, e-commerce environment.
5411.04 The Assurance Services Executive Committee of the AICPA has introduced Trust Services,
including SysTrust and WebTrust, which are defined as a set of attestation and advisory
services based on a core set of principles and criteria that addresses the risks and
opportunities of IT-enabled systems and privacy programs. The following five principles have
been developed by the AICPA and CICA (Canadian Institute of Chartered Accountants) for
use by practitioners in the performance of Trust Services engagements:
1. Security: The system is protected against unauthorized physical and logical access to
prevent or minimize the theft, improper use, alteration, destruction, or disclosure of
data and software.
2. Availability: The system is available for operation and use as committed or agreed.
3. Processing integrity: System processing is complete, accurate, timely, and authorized.
4. Confidentiality: Information designated as confidential is protected as committed or
agreed.
5. Privacy: Personal information is collected, used, retained, disclosed, and destroyed in
conformity with the commitments in the entity’s privacy notice and with criteria set
forth in generally accepted privacy principles issued by the AICPA and CICA.
5411.05 Ensuring system reliability is a top management issue, not an information technology issue.
Management is responsible for the reliability and accuracy of the organization’s information.
The Sarbanes-Oxley Act requires:
a. the CEO and CFO to certify that financial statements accurately represent company
activities.
b. an annual report on the company’s internal controls. Included in the report is:
(1) a management acknowledgement that it is responsible for designing and
maintaining internal controls.
(2) management’s assessment of the effectiveness of those controls.
5411.07 To ensure system reliability, companies must implement a set of preventive controls and
supplement them with methods for detecting incidents and procedures for taking corrective
remedial action. A company must also employ multiple layers of controls so that if one
control fails or is circumvented, another control will prevent, detect, or correct the reliability
breakdown.
5411.09 Strategic planning is an organization’s formal process of defining its future course or
direction. By devising appropriate strategies and making decisions on allocating its resources
(e.g., staff and capital) to pursue those strategies, an organization can determine its
direction over a span of time, usually three to five years. To determine that direction, an
organization needs to know its current status or position, its future goals, and the strategy or
strategies needed to achieve those goals. The results of this process are gathered in a
document called the strategic plan.
5411.10 While strategic planning is an effective tool for determining the direction of an organization,
the strategic plan cannot predict how a market will evolve and what other issues may
surface that affect an organization’s strategy. A turbulent business climate requires
innovation and modification of the strategic plan for an organization to survive.
5411.12 Strategic planning uses business analysis techniques such as the following:
a. EPISTELS (environment, political, informatic, social, technological, economic, legal, and
spiritual)
b. PEST (political, economic, social, and technological analysis)
c. STEER (sociocultural, technological, economic, ecological, and regulatory factors)
d. SWOT (strengths, weaknesses, opportunities, and threats)
5411.13 An operational plan is part of the strategic plan and describes short-term methods of
achieving milestones. It explains how the strategic plan (or part of the plan) will be put into
operation during a given operational period (e.g., a fiscal year). Operational plans establish
the activities and budgets for each part of the organization for the next one to three years.
They link the strategic plan with the activities the organization will deliver and the resources
needed to deliver them. An operational plan describes agency and program missions and
goals, program objectives, and program activities, and draws directly from the agency and
program strategic plan.
5411.14 Operational plans should contain clear objectives, desired outcomes, activities to be
delivered, staffing and resource requirements, quality standards, implementation
timetables, and a process for monitoring progress. The operational plan justifies and
provides the basis for annual operating budget requests. Therefore, a separate operational
plan is needed for each operating budget.
5411.16 An operational plan should be prepared by the same staff responsible for its implementation
and there is a need for cross-departmental cooperation, as plans created by one part of the
organization inevitably have implications for the other parts.
5411.17 When preparing an operating budget request, an operational plan provides a plan for
resource allocation. Later in the process, the plan may be modified to reflect policy decisions
or financial changes made during the budget development process.
5411.19 Documentation
It is hard to ensure systems reliability without proper documentation, as this can result in
ineffective system design, operation, review, audit, and modification. Documentation can
also be used in training system users. It is important to develop and maintain three different
types of documentation:
a. Administrative (standards and procedures for analysis, design, programming, processing
file handling, and storage)
b. Systems (input, processing, output, and error handling procedures for each application)
c. Operating (equipment configurations, programs, files, setup and execution procedures,
and corrective actions for each program)
5411.20 Training
Organizations should spend the time and money necessary to make sure the people who
design, develop, implement, and operate the company’s information system are qualified
and well trained. There should be an adequately staffed information center help desk in
computer operations to deal with errors in user interactions with the information system.
The same holds true for those in charge of any aspect of system security.
5421.02 The Institute of Management Accountants (IMA) conducted an intensive analysis of the job
duties of corporate accountants. Respondents clearly indicated that work relating to
accounting systems was the single most important activity performed by corporate
accountants.
5421.03 A system is a set of two or more interrelated components that interact to achieve a goal.
Systems are almost always composed of smaller subsystems, each performing a specific
function important to and supportive of the larger system of which it is a part. Accounting
systems generally consist of several accounting subsystems, each designed to process
transactions of a particular type. Although they differ with respect to the type of
transactions processed, all accounting subsystems follow the same sequence of procedures.
These procedures are referred to as accounting cycles.
5421.04 An effective information system is essential to any organization’s long-run success. Without
a means of monitoring the events that occur, there is no way to determine how well the
organization is performing or to track the effects of various business activities on company
resources.
5421.06 The accounting information system (AIS) is a subset of the management information system
(MIS). The AIS is composed of both the human and capital resources within an organization
that are responsible for:
a. the preparation of financial information and
b. the information obtained from collecting and processing company transactions.
The AIS performs several important functions in an organization:
a. It collects and stores data about activities and business transactions.
b. It processes data into information and provides reports that are useful for making
decisions.
c. It provides adequate controls to safeguard the organization’s assets.
5. Properly present transactions and related disclosures in the financial statements. Failing
to disclose a lawsuit or a contingent liability could mislead the reader of a financial
statement.
h. Improve decision making. An information system should provide feedback on the results
of actions so it can be used to improve future decisions. For example:
(1) Reports can help to identify potential problems that require management action.
(2) Query languages can facilitate the gathering of relevant data upon which to make
decisions.
(3) Tools (such as graphical interfaces, decision models, and analytical tools) can help
the decision maker interpret information and evaluate and choose among
alternative courses of action.
5421.10 Accounting information system (AIS): The AIS includes the human and capital resources
within an organization that are responsible for (a) the preparation of financial information
and (b) the information obtained from collecting and processing company transactions. The
AIS is a subset of the management information system.
5421.11 Decision support system (DSS): A DSS is an extension of a management information system.
It is designed to aid in semi-structured decision making. Some processes and functions are
preprogrammed in a DSS; the decision is not made by the system. Instead, the decision
maker must apply insight and exercise judgment while interacting with the system.
5421.12 Executive information system (EIS): An EIS is an information system designed to provide
executives with the needed information to make strategic plans, to control and operate the
company, to monitor business conditions in general, and to identify business problems and
opportunities.
5421.13 Expert system (ES): An ES is a computerized information system that allows non-experts to
make decisions about a particular problem that are comparable with decisions of experts in
the area. These systems include knowledge bases that represent facts and inferences they
know. They use heuristics (rules of thumb) in processing information.
5421.14 Strategic information system (SIS): An SIS provides information that may allow an
organization to make strategic, competitive decisions.
5421.15 Transaction processing system (TPS): A TPS is a completely programmed and automated
system, treating every problem in exactly the same way. This type of system is used in AISs
so that each transaction is processed in an identical manner.
5421.16 Big data is a term that describes the large volume of data—both structured and
unstructured—available to business on a day-to-day basis. The data sets are so voluminous
and diverse that traditional data processing application software is inadequate to deal with
them. Big data can be analyzed for insights that lead to better decisions and strategic
business moves. Big data is often defined by the three “V’s”: volume, velocity, and variety.
5421.17 Data analytics rapidly examine large amounts of data to identify hidden patterns and
correlations. In his report Big Data in Big Companies, IIA (Institute of Internal Auditors)
Director of Research Tom Davenport interviewed more than 50 businesses to understand
how they used big data. He found they got value in the following ways:
a. Cost reduction: Big data technologies such as Hadoop and cloud-based analytics bring
significant cost advantages when it comes to storing large amounts of data—plus they
can identify more efficient ways of doing business.
b. Faster, better decision making: With the speed of Hadoop and in-memory analytics,
combined with the ability to analyze new sources of data, businesses are able to analyze
information immediately—and make decisions based on what they have learned.
c. New products and services: With the ability to gauge customer needs and satisfaction
through analytics comes the power to give customers what they want. Davenport points
out that with big data analytics, more companies are creating new products to meet
customers’ needs.
5421.18 Data visualization is a general term that describes any effort to help people understand the
significance of data by placing it in a visual (pictorial or graphical) context, helping to
understand and communicate complex concepts and ideas.
a. Data visualization software plays an important role in big data by detecting patterns,
relationships, emerging trends, and correlations that might have gone undetected in
analyzing text-based data.
b. Visualization is central to advanced analytics as well. When a data scientist is writing
advanced predictive analytics or machine-learning algorithms, it becomes important to
visualize the outputs to monitor results and ensure that models are performing as
intended; visualizations of complex algorithms are generally easier to understand and
interpret than numerical outputs.
5421.19 Most data visualization tools come with connectors to popular data sources, including many
of the more common relational databases, Hadoop, and cloud storage platforms. The
visualization software pulls in data from these sources and applies a graphic to the data.
a. Data visualization software allows the user to select the best way of presenting the data,
although sometimes software automates this step. Some tools can automatically
interpret the shape of the data and detect correlations between certain variables and
then place these discoveries into the chart type that the software determines is optimal.
b. Interactive visualization allows the user to drill down into the charts and graphics,
changing the data and observing the revised output.
5421.20 Sources for big data generally fall into one of three categories:
1. Streaming data: Data that reaches IT systems from a web of connected devices. It can
be analyzed as it arrives and used to make decisions on what data to keep, what not to
keep, and what requires further analysis.
2. Social media data: Data extracted from social interaction platforms is used for
marketing, sales, and support functions. It is often in unstructured or semi-structured
forms, creating a unique challenge when it comes to consumption and analysis.
3. Publicly available sources: Massive amounts of data are available through open data
sources like the U.S. government’s data.gov, the CIA World Factbook, or the European
Union Open Data Portal.
5421.21 Several areas of technology are generally used for big data analytics, including the following:
a. Data management: Data needs to be high-quality and well-governed before it can be
reliably analyzed. Once data is reliable, organizations should establish a master data
management program that gets the entire enterprise on the same page.
b. Data mining: Data-mining technology helps examine large amounts of data to discover
patterns. With data-mining software, companies can sift through all the chaotic and
repetitive noise in data, pinpoint what is relevant, use that information to assess likely
outcomes, and then accelerate the pace of making informed decisions.
c. Hadoop: Hadoop is a free, open-source software framework that stores large amounts
of data and rapidly runs applications on clusters of commodity hardware.
d. In-memory analytics: By analyzing data from systems, companies can derive immediate
insights from data analytics and respond accordingly. This technology removes data
preparation and analytical processing latencies to test new scenarios and create models,
enabling organizations to stay agile and make better business decisions, as well as
allowing them to run iterative and interactive analytics scenarios.
e. Predictive analytics: Predictive analytics technology uses data, statistical algorithms,
and machine-learning techniques to identify the likelihood of future outcomes based on
historical data. Some of the most common applications of predictive analytics include
fraud detection, risk, operations, and marketing.
f. Text mining: With text-mining technology, entities can analyze text data from the web,
comment fields, books, and other text-based sources to uncover insights not previously
identified. Text mining uses machine learning or natural language processing technology
to comb through documents—emails, blogs, Twitter feeds, surveys, competitive
intelligence, and more—to help analyze large amounts of information and discover new
topics and term relationships.
5421.22 Customer relationship management (CRM) refers to practices, strategies, and technologies
that businesses use to interact, analyze, forecast, and manage customer relationships,
trends, and behaviors. CRM is preferably a cloud-based system that stores customer and
prospect contact information, accounts, leads, and sales opportunities in one central
database, available to all departments in a business, such as sales, customer service,
accounting, marketing, and business development. Major vendors of CRMs include
Salesforce, Microsoft, SAP, and Oracle.
responsibility for the administration, control, security, and maintenance of the database
and information.
b. Cloud-based CRM (also known as SaaS [software as a service] or on-demand CRM):
Data is stored on an external, remote network that employees can access anytime,
anywhere via an Internet connection. Cloud-based systems have several concerns that
companies must cope with:
(1) Data security: Given that the company does not physically control the storage and
maintenance of its data, an entity’s data can be compromised or lost if the cloud
provider goes out of business or is acquired by another company. Compatibility
issues can also arise when data is initially migrated from a company's internal
system to the cloud.
(2) Cost: The cost of subscription fees for cloud-based CRM software can be more
costly over time than for on-premises models.
c. Open-source CRM: An open-source CRM system, such as OroCRM, SuiteCRM, or
SugarCRM, offers alternatives to the proprietary platforms from Salesforce, Microsoft,
and other vendors. Source code is made available to the public, permitting businesses to
make alterations to the software at no cost, thus enabling the customization of data
links on social media channels.
5421.24 A digital business model defines how a business makes money digitally, based on a digital
strategy. A digital strategy is a plan for maximizing the benefits of data assets and
technology-focused initiatives.
5421.25 A digital business blurs the lines between digital and physical worlds. For example, a
customer could visit a retail clothing store and try on a dress. A computer image would show
the customer with the dress on, offer different colors and accessories, and customize the fit
to the customer. The dress would be tailored to a perfect fit, computer-assembled, and
shipped directly to the customer’s home.
5421.26 The findings of a research project at MIT's Center for Information Systems Research (CISR)
concluded that a great digital business model needs to challenge the traditional physical
business model that relies on places (e.g., bank branches) and people (e.g., sales teams) to
satisfy customers. They see this challenge along three key axes: who owns the customer
interface, how to reconfigure business processes, and how to make customer data an
enterprise-wide resource.
5421.27 It is estimated that in the near future all business leaders and chief information officers
(CIOs) will need to develop digital leadership and technology capabilities in order to create
and implement their digital strategy. New positions will be created, such as chief digital
officer (CDO).
5422.03 Many business activities are repetitive and engaged in frequently. For example, a retail
company sells merchandise to customers in exchange for cash on a daily basis. It frequently
buys inventory from suppliers to replenish the merchandise that is sold.
5422.04 These basic exchanges can be grouped into several transaction cycles or business processes.
The most common are:
a. the revenue cycle, where goods or services are sold in exchange for cash or a promise to
pay cash.
b. the expenditure cycle, where goods or services are purchased in exchange for cash or a
promise to pay cash.
c. the production cycle, where raw materials are turned into products that can be sold.
d. the human resources (payroll) cycle, where people are hired, trained, compensated for
time and effort they expend, and dismissed.
e. the financing cycle, where companies obtain funds from investors and creditors and pay
dividends and repay borrowings.
5422.05 While the basic cycles do not change, there are differences in cycles among companies. For
example:
a. Some organizations do not use every cycle. For example, retail stores do not have a
production cycle.
b. Some organizations have unique requirements. For example, financial institutions have
demand deposit and customer loan cycles.
c. Transaction cycles differ depending upon the type of organization. For example, the
expenditure cycles of a service company (CPA or law firm) do not purchase, receive, or
pay for merchandise to resell to customers.
5422.06 The transaction cycles are implemented as separate modules in most accounting software
packages (also known as the general ledger system).
Figure 4-2 uses a merchandising company to show how these transaction cycles relate to
one another and their relationship with the general ledger and reporting system, which is
used to generate information for internal and external parties.
Figure 4-2:
Investors
Investment Merchandising
Human Resources
Company
Financial Cycle
statements Labor services
Loans
Taxes
Managerial reports Government
Management and financial statements Agencies
Tax forms and regulations
5422.07 Objectives
Each cycle has specific objectives that it needs to accomplish. Each cycle also faces a number
of threats that put company assets at risk. Controls are put in place to eliminate or minimize
these risks. There are several objectives, threats, and controls that apply to all the cycles.
All cycles have the following objectives:
a. Data must be available when it is needed.
b. All activities must be performed efficiently and effectively.
5422.08 Threat
Multi-cycle threat: Loss or unauthorized disclosure of data. Data can be lost, resulting in:
a. inaccurate external and internal reporting,
b. inaccurate responses to customer and vendor inquiries,
c. customer or vendor information leaked to competitors,
d. inventory and fixed assets not being monitored effectively,
e. manufacturing activities not being performed efficiently and effectively,
f. morale problems,
g. lawsuits, or
h. the potential demise of the entire organization.
5422.09 Controls
a. Back up all master and transaction files regularly, including the general and subsidiary
ledgers.
(1) Keep one backup copy on-site so it can be immediately accessed if the original is
damaged or destroyed.
(2) Store another backup copy off-site as insurance against a major disaster such as a
fire or earthquake.
b. Develop, implement, and periodically test a disaster recovery plan to handle major
disasters that can destroy a computer system.
c. Use external and internal file labels on all disks, tapes, and other storage media:
(1) to reduce the possibility of accidentally erasing important files and
(2) to make sure the correct version of the master file is updated.
d. Use physical security controls to restrict physical access to data and computer
equipment.
e. Use passwords, user IDs, and an access control matrix to limit logical access to sensitive
files, data items, and terminals to authorized employees. For example:
(1) Only sales staff should be allowed to create sales orders.
(2) Only the cashier should be able to indicate that an invoice has been paid.
(3) Sales staff should have read-only access to customer credit limits and current
account balances.
(4) Access controls should exist for individual terminals so the terminal can only
process transactions its department is authorized to perform.
f. Encrypt data to make it unintelligible to anyone able to get unauthorized access to files
or data.
g. Use simple, easy-to-complete documents with clear instructions so transaction data can
be recorded accurately and efficiently.
h. Record who completed and who reviewed a document as evidence of proper
transaction authorization.
i. Prenumber documents to make sure all transactions are recorded.
j. Restrict access to blank documents and to programs that create documents to reduce
the risk of unauthorized transactions.
k. Log all activities, especially actions involving managerial approval (e.g., extending credit
limits), so they can be reviewed later as part of the audit trail.
5422.10 Threat
Multi-cycle threat: Poor performance. It is important that all business activities be
performed efficiently and effectively.
5422.11 Controls
a. Prepare and review performance reports to assess the efficiency and effectiveness of
cycle activities and to look for inefficient or ineffective performance. For example:
(1) Assess sales force effectiveness by breaking sales down by salesperson, region, or
product.
(2) Assess marketing performance by breaking down the marginal profit contribution of
each territory, customer, distribution channel, salesperson, or product.
(3) Evaluate the frequency and size of back orders to determine how well inventory
management policies satisfy customer needs.
(4) Identify slow-moving products to avoid excessive inventory levels.
(5) Prepare an accounts receivable aging schedule to monitor accounts receivables
collections, estimate bad debts, and evaluate credit policies.
(6) Prepare cash budgets to know when to borrow funds to meet short-term cash
shortages and when to invest excess funds.
(7) Monitor vendor performance using vendor performance reports.
b. Track employee knowledge and skills so they can be shared and used to provide a
sustainable competitive advantage.
c. Monitor employees who telecommute to make sure they are:
(1) working the hours for which they are being paid.
(2) not using company-provided computing resources inappropriately:
(a) visiting pornographic websites, sending or forwarding offensive e-mail, or
storing and using pirated copies of software.
(b) operating a personal business on the side, using company-provided assets for
personal use.
d. Train employees to maximize the efficiency and effectiveness of business processes and
keep abreast of changes in information technology and legal requirements.
Each of the major business cycles is discussed in the following sections.
5422.12 The revenue cycle is a recurring set of business and data processing activities associated with
selling goods and services to customers in exchange for cash.
The revenue cycle produces information that is used by other accounting cycles:
a. Both the expenditure and the production cycles use sales information to purchase or
produce additional inventory.
b. The payroll cycle uses sales information to compensate employees (calculate sales
commissions and bonuses).
c. The general ledger and reporting function uses revenue cycle information to prepare
financial statements and performance reports.
As shown in Figure 4-3, there are four basic business activities in the revenue cycle. Each of
these activities, the threats the activities face, and the controls used to eliminate or
minimize these threats are now discussed.
Figure 4-3:
1. Sales Order
Entry
3. Billing and
Accounts
Receivable
1. Take 3. Check
Customer 2. Approve
Inventory
Order Credit
Availability
3. Billing and
Accounts
Receivable
5422.14 The three steps in the sales order entry process are take customer orders, approve credit,
and check inventory availability.
5422.15 Take customer orders. Customer orders are received in the store, by mail, by phone, or from
a salesperson in the field. Companies have used information technology advancements to
improve the efficiency and effectiveness of the sales order entry process. For example:
a. Customers can enter sales order data themselves on a company website.
b. Customers can use interactive sales order entry websites called choiceboards to design
their own products. For example, Dell customers can mix and match parts and features
to create a computer to meet their needs at an affordable price.
c. Electronic data interchange (EDI), where suppliers and customers are linked, eliminates
the need for data entry and improves the efficiency and effectiveness of the sales order
process.
d. E-mail can be used to communicate price changes, special needs and interests, and sales
promotions.
5422.17 Approve credit. A credit sale should be approved before goods are released from inventory.
d. Do not allow sales order entry clerks or sales personnel to change a customer’s
credit limit.
e. Maintain accurate and current customer account balances and credit limits.
5422.19 Check inventory availability. There are several steps to this process:
a. Have the system find out if there is enough inventory on hand to fill the order. If so:
(1) notify shipping, inventory control, and billing of the sale,
(2) create an order and enter it into the system, and
(3) let the customer know the expected delivery date.
b. If there is insufficient inventory, back-order the item so the production department can
produce the requested item or purchasing can order it.
1. Sales Order
Entry
2. Ship Goods
3. Billing and
Accounts
Receivable
5422.22 Pick and pack orders. After the sales order is entered, a picking ticket that lists the items and
quantities to be shipped is printed. Warehouse employees:
a. use it to identify the items and quantities to remove from inventory,
b. record quantities picked on the picking ticket (or enter them into the system), and
c. forward the inventory to shipping.
b. There are two steps in the billing and accounts receivable activity (as shown in
Figure 4-6).
Figure 4-6:
1. Sales Order
Entry
1. Bill 2. Update
Customers Accounts
Receivable
(1) Bill customers. After goods are shipped, a sales invoice is sent to customers to let
them know how much they owe and where to send payment. EDI and EFT
technology is often used to reduce billing costs and the time required to send
invoices and receive payment.
(2) Update accounts receivable. Sales invoice information is used to update customer
accounts.
(a) There are different ways to maintain accounts receivable depending on when
customers remit payments, how payments are updated on the accounts
receivable master file, and monthly statement format.
i. With the open-invoice method, customer payments are applied against
specific invoices. Two invoice copies are mailed to the customer. One
serves as a remittance advice and is returned with the payment.
ii. With the balance-forward method, customers pay the balance shown on a
monthly statement that lists all sales and payment transactions that
occurred during the past month.
Cycle billing can be used to spread out billing and cash receipts. Monthly
statements are prepared and mailed at different times for subsets of
customers. For example, the master file could be divided into four parts,
and each week monthly statements sent to one-fourth of the customers.
(b) As part of the update process, adjustments to customer accounts are
sometimes needed due to:
i. items that are returned.
ii. allowances for damaged goods.
iii. customer’s accounts that must be written off.
In such cases, the credit manager fills out a credit memo. When the credit
memo is for damaged or returned goods, the customer is sent a copy.
(1) Have two people open all incoming mail to minimize the risk of cash or checks
being stolen.
(2) Prepare a list of all checks received as the mail is opened.
(3) Restrictively endorse each check after it is removed from the envelope.
d. Segregate the recording and custody functions by sending remittance data to
accounts receivable and customer payments to the cashier. Use the following
control checks:
(1) Compare accounts receivable credit totals to cash deposited totals.
(2) Have internal audit compare remittance lists with validated deposit slips and
bank statements to make sure all checks and cash were deposited.
e. Send monthly statements to customers, as they will complain if there are errors.
f. Use cash registers that record all cash receipts when cash is received directly from
customers.
g. Encourage customers to watch their purchase being rung up and to ask for a receipt
by using signs to let them know that:
(1) their purchase is free if they are not given a receipt or
(2) they are entitled to a discount if their receipt has a red star on it.
h. Deposit all customer receipts daily to:
(1) reduce the amount of cash and checks that can be stolen.
(2) make it easier to reconcile bank statements with sales, accounts receivable,
and cash collection records.
i. Make sure the bank statement is reconciled by someone who has no other cash
handling and recording duties. This:
(1) provides an independent check on the cashier and
(2) prevents bank statement manipulations that conceal cash embezzlements.
j. Make sure accounts receivable (a recording function) does not have physical access
to cash or checks.
5422.29 The expenditure cycle is a recurring set of business and data processing activities associated
with purchasing and paying for goods and services. A primary objective of the cycle is to
minimize the cost of acquiring and maintaining inventories, supplies, and other needed
organizational services.
There are several information flows in the expenditure cycle (as shown in Figure 4-7):
Figure 4-7:
Back Orders
Performance Reports Revenue
Management Cycle
Receipt of Goods
Notification
Purchase Orders
Purchase Order
1. Order 1. Sales Order
Goods Entry
4. Cash
Collections
b. prepares a purchase order and sends it to the chosen supplier. A purchase order:
(1) is a formal request to sell and deliver specified products at designated prices.
(2) is a promise to pay.
(3) becomes a contract once the supplier accepts it.
(4) includes supplier and purchasing agent’s names, order and delivery dates,
delivery location, shipping method, and information about the items ordered.
5422.32 Information technology can be used to improve the purchasing process by:
a. using EDI or the Internet to communicate electronically.
(1) Costs can be reduced by eliminating printed and mailed paper documents.
(2) Stockouts can be minimized or reduced by eliminating the delivery time associated
with paper documents.
b. developing vendor-managed inventory programs to minimize or eliminate stockouts.
Suppliers access inventory data in company databases and automatically ship inventory
when needed.
c. running reverse auctions where suppliers compete to sell goods at the lowest price.
d. using trading exchanges (electronic marketplaces) that link buyers and sellers of
commodity-type goods that are differentiated by price.
e. using procurement cards (corporate credit cards) to improve efficiency and reduce
purchasing costs.
(1) Each card has a spending limit.
(2) All employee purchases are summarized on one monthly bill.
(3) The purchases from each card can be allocated to appropriate general ledger
accounts.
e. Use barcode technology to eliminate human data entry and improve perpetual
inventory record accuracy as barcode scanners capture data quickly and are easy to
use. One example of a barcode type is a universal product code (UPC).
3. Threat: Purchasing errors. This includes:
a. buying unnecessary items and
b. failure to take advantage of available volume discounts.
4. Controls
a. Develop an accurate perpetual inventory control system to make sure inventory
information is always current. Periodically count inventory and reconcile the counts
to the perpetual records.
b. Have department supervisors review and approve all purchase requisitions before
they are sent to the purchasing department.
c. Integrate subunit databases so items do not have different numbers in various
databases and make the database available to all who need it.
d. Prepare reports that link item descriptions to part numbers.
4. Controls
a. Make visual inspections to make sure the service was rendered where possible.
b. Hold department supervisors responsible for all service costs they incur.
c. Charge the services to accounts for which the requesting department is responsible.
d. Compare actual and budgeted expenses and investigate any discrepancies.
5422.38 Information technology advancements can be used to improve the efficiency and
effectiveness of the receiving and inventory storage functions:
a. Companies can minimize data entry errors by:
(1) requiring suppliers to barcode their products and receiving clerks to scan the
barcodes to get product numbers and quantity of items received.
(2) requiring vendors to attach, to each crate of goods, a radio frequency identification
tag that emits a signal and installing a receiving unit at the company that can read
the radio signal.
b. Trucks can be equipped with data terminals linked to satellites that track the exact
location of all incoming shipments.
When an invoice is paid, it is removed from the open-invoice file, marked paid, and
stored in the paid-invoice file.
(2) In a voucher system, a disbursement voucher is created to summarize one or more
vendor invoices and specify which general ledger account to debit.
The voucher identifies the supplier, lists all outstanding invoices, and indicates how
much is to be paid after applicable discounts and allowances are deducted.
c. Information technology advancements can help improve the accounts payable process
by:
(1) requiring invoices to be submitted electronically (by EDI or the Internet) and using
the system to automatically match invoices to appropriate purchase orders and
receiving reports.
(2) using an “invoiceless” approach, called evaluated receipt settlement (ERS), that
replaces the traditional three-way matching process (vendor invoice, receiving
report, purchase order) with a two-way match (purchase order, receiving report).
(3) using procurement cards or corporate credit cards to eliminate most small invoices.
(4) using electronic forms for submitting travel expenses.
5422.42 Pay approved invoices. A voucher package consists of a vendor invoice, a purchase order,
and a receiving report. After voucher packages are approved, funds are disbursed by check
or EFT and sent to suppliers.
5422.43 Threats and controls in paying for goods and services (Part 1)
1. Threat: Errors in vendor invoices. These errors include:
a. discrepancies between quoted and actual prices charged, and
b. errors in calculating the total amount due.
2. Controls
a. Have an independent person, or the system:
(1) verify prices and quantities on vendor invoices and compare them with those
on the purchase order and receiving report.
(2) verify the mathematical accuracy of vendor invoices.
b. Require employees who have procurement cards to keep receipts and verify
monthly statement accuracy.
c. Adopt ERS to eliminate vendor invoices errors (companies pay after matching
receiving reports with purchase orders).
d. Train purchasing and accounts payable staff on transportation practices and
terminology so the company does not pay unnecessary transportation costs.
e. Negotiate significant discounts with designated carriers to reduce costs.
f. Capture data on carriers, so reports identifying suppliers who fail to comply with
shipping instructions can be prepared.
3. Threat: Paying for goods not received
4. Controls
a. Compare vendor invoice quantities with the quantities the inventory control person
entered into the system.
Inventory control should make sure that when goods are transferred from
receiving, their count agrees to receiving’s count.
b. Use budgetary controls and careful reviews of departmental expenses to find
potential problems when purchasing services.
5422.44 Threats and controls in paying for goods and services (Part 2)
1. Threat: Failing to take available purchase discounts
2. Controls
a. Set up procedures to make sure approved invoices are filed by due date.
b. Have the system track invoice due dates and print a list of all outstanding invoices
by due date.
c. Use a cash flow budget that indicates expected cash inflows and outstanding
commitments so the company can plan to take advantage of purchase discounts.
3. Threat: Paying the same invoice twice. An invoice can be paid more than once.
a. A duplicate invoice could be sent after the company’s check is in the mail.
b. An invoice could become separated from the documents in the voucher package.
Duplicate payments are usually detected by the supplier, who may or may not
return the duplicate payment. Duplicate payments affect a company’s cash flow
needs and result in incorrect financial records until the duplicate payment is
detected.
4. Controls
a. Approve invoices for payment only when accompanied by a purchase order and
receiving report.
b. Only pay based on an original copy of an invoice; do not authorize payment based
on a photocopy.
c. When payment is made, invoices and voucher packages should be canceled
(marked “paid”) so they cannot be resubmitted for payment.
d. Automate accounts payable, to eliminate most clerical errors associated with
processing invoices.
e. Use invoiceless accounts payable systems to improve efficiency.
(1) Control access to the supplier master file and monitor all changes made to it.
(2) The supplier master file contains the prices of goods being purchased and
unauthorized changes to those prices can result in overpayments to suppliers.
f. Use internal auditors to detect and recover overpayments to suppliers.
5422.45 Threats and controls in paying for goods and services (Part 3)
1. Threat: Accounts payable recording and posting errors. This can result in:
a. errors in financial and performance reports, and
b. poor decisions.
2. Controls
a. Compare supplier account balances before and after processing checks with the
total amount of invoices processed.
b. Reconcile the accounts payable control account balance in the general ledger to
one of the following:
(1) The total of all supplier account balances
(2) Unpaid vouchers
3. Threat: Misappropriated cash, checks, or EFTs
4. Controls
a. Restrict access to cash and blank checks.
b. Pre-number checks sequentially and periodically account for them.
c. Restrict access to the check-signing machine.
d. Require two signatures for checks in excess of a certain amount, to provide an
independent review of the expenditure.
e. Have the cashier or treasurer mail signed checks, rather than returning them to
accounts payable, to make sure checks are mailed to the intended recipients.
f. Segregate duties.
(1) Have accounts payable assemble voucher packages and authorize payments.
(2) Have the treasurer or cashier sign checks.
g. Have internal auditors periodically review the supplier master file to check for
duplicate supplier entries.
h. Have someone who has no involvement in processing cash collections or
disbursements reconcile all bank accounts promptly.
(1) This provides an independent check on accuracy.
(2) It prevents someone from misappropriating cash and concealing the theft by
manipulating the bank reconciliation.
i. To prevent check alteration and forgery:
(1) use check-protection machines that print check amounts in distinctive colors.
(2) use special inks that change colors if altered.
(3) print checks on special paper that contains watermarks.
(4) send a list of all legitimate checks to the bank, which will clear only those
checks appearing on the list.
j. For minor purchases, use a petty cash fund managed by an employee with no cash-
handling or accounting responsibilities.
(1) The petty cash fund should be an imprest fund for a small, fixed amount.
(2) Require vouchers for every disbursement so that, at all times, the sum of cash
plus vouchers equals the imprest fund balance.
(3) When needed, turn the vouchers in to accounts payable to get the fund
replenished.
(4) Cancel petty cash fund vouchers so they cannot be reused.
(5) Have internal auditors make periodic, unannounced counts of the fund
balances and vouchers.
(6) Hold the person in charge of the petty cash fund responsible for any shortages.
5422.46 The production cycle is a recurring set of business and data processing activities associated
with manufacturing products. The production cycle interacts with the other cycles as
indicated in Figure 4-9:
Figure 4-9:
There are four basic activities in the production cycle (as shown in Figure 4-10). Each of
these activities is briefly explained. The threats a system faces during each activity are also
explained, along with the controls used to prevent the threat.
Figure 4-10:
1. Design
Products
3. Produce
Products
(2) Use barcode technology to eliminate human data entry and improve perpetual
inventory record accuracy.
c. Prepare production performance evaluation reports to evaluate production
quantities, product quality, and delivery schedules.
d. Require approval and authorization of production orders:
(1) Use passwords and an access control matrix to restrict access to the production
scheduling program.
(2) Use closed-loop verification to make sure correct production orders are
released. The production planner enters the product number.
(a) The system retrieves the description, order quantity, and other relevant
data.
(b) The system requests the user to verify that the correct production order is
being released.
e. Select suppliers that meet delivery commitments, monitor supplier performance,
and select new suppliers when performance falls below acceptable levels.
3. Threat: Unauthorized acquisition of fixed assets. This can result in:
a. overinvestment in fixed assets and
b. reduced profitability.
4. Controls
a. Require supervisory personnel to recommend large capital expenditures and do a
feasibility analysis of the proposed expenditure.
b. Require a capital investments committee to review acquisition recommendations
and prioritize them.
c. Require small capital expenditures to be purchased directly out of departmental
budgets.
d. Hold managers accountable for their department’s return on fixed assets.
e. Send potential suppliers a request for proposal (RFP) that specifies the desired asset
properties and invite them to send in a bid.
f. Require the capital investments committee to review RFP responses and select the
best bid.
b. Use access controls and compatibility tests to restrict access to information about
inventories and fixed assets to authorized personnel.
c. Document all internal movements of inventory through the production process.
(1) Use materials requisitions to authorize the release of raw materials to
production.
(2) Require inventory control clerks and production employees receiving raw
materials to sign the materials requisition to acknowledge that goods were
released to production.
(3) Document and authorize requests for materials in excess of the amounts
specified in the bill of materials.
(4) Use move tickets to document the movement of inventory through the
production process.
(5) Document any return of materials not used in production.
d. Segregate duties
(1) Have the inventory stores department be responsible for maintaining physical
custody of raw materials and finished goods inventories.
(2) Have department or factory supervisors be responsible for work-in-process
inventories.
(3) Have production planners and the production information system be
responsible for preparing production orders, materials requisitions, and move
tickets.
e. Use barcode scanners and online terminals to record inventory movement and help
maintain accurate perpetual inventory records.
f. Have an employee without any custodial responsibility count inventory on hand
periodically. Discrepancies between physical counts and recorded amounts should
be investigated and corrected.
g. Identify and record all fixed assets.
(1) Make managers responsible and accountable for fixed assets under their
control.
(2) Authorize and document all disposals of fixed assets.
(3) Prepare a report of all fixed-asset transactions periodically and send it to the
controller, who should verify the proper authorization and execution of each
transaction.
h. Maintain adequate insurance to cover losses and replace assets.
5422.55 The human resources management (HRM)/payroll cycle is a recurring set of business and
data processing activities associated with managing employee information and
compensating employees.
a. The more important tasks include the following:
(1) Recruiting and hiring new employees
(2) Training new and existing employees
(3) Assigning jobs to employees
(4) Compensating employees
(5) Evaluating employee performance
(6) Voluntary and involuntary terminations
The first and last activities are performed once for each employee. The other activities
are performed periodically for as long as the person is employed.
b. Compensating employees is the payroll system’s main function. The HRM system takes
care of the other tasks. These two systems are organizationally separate in larger
organizations. The director of human resources manages the HRM system and the
controller manages the payroll system.
The main inputs to the payroll system and their sources are as follows:
(1) The HRM department provides data about hirings and terminations as well as pay-
rate changes.
(2) Employees make changes to their files (deductions, address changes, dependents).
(3) Departments send data about how many hours each employee works.
(4) Government agencies are the source of tax rate changes and instructions for
meeting regulatory requirements.
3. Validate
1. Update Payroll 2. Update Tax Rates
Employee Time,
Master File and Deductions
Attendance Data
6. Calculate
4. Prepare Payroll Employer-Paid
Benefits and Taxes
7. Disburse Payroll
5. Disburse Payroll Taxes, Misc.
Deductions
5422.61 Threats and controls in validating employee time and attendance data
1. Threat: Inaccurate time data. This can:
a. result in inaccurate labor expenses and labor expense reports,
b. result in poor decisions, and
c. hurt employee morale, if paychecks are incorrect or not prepared.
2. Controls
a. Use badge readers, barcode scanners, and online terminals to:
(1) collect data on employee time and attendance electronically and
(2) collect job-time data and send it to the payroll processing system.
b. Include edit checks in data entry programs:
(1) Field checks, for numeric data in fields such as employee number and hours
worked
(2) Limit checks, on the hours worked field
(3) Validity checks, on the employee number field
c. Restrict payroll master file changes to someone not directly involved in processing
payroll.
d. Require payroll master file changes to be approved by appropriate supervisors.
e. Reconcile time clock data to job-time ticket data.
(1) This should be done by someone who does not generate the data.
(2) All time card hours should be accounted for on the job-time tickets.
2. Controls
a. Calculate batch totals prior to data entry.
(1) Check them against totals calculated during each subsequent processing stage.
(2) If original and subsequent totals agree, all payroll records were processed, data
input was accurate, and bogus time cards were not entered during processing.
b. Cross-foot the payroll register to make sure the total of the net pay column equals
the total of gross pay less the total deductions column.
c. Use a payroll clearing account (a zero-balance account) to verify payroll cost
recording accuracy and completeness, and to allocate costs appropriately.
(1) Debit gross pay to the payroll control account.
(2) Credit net pay to cash and credit the various withholdings.
(3) Cost accounting distributes labor costs to appropriate expense categories and
credits the payroll control account for the allocation total. The payroll control
account credit should equal the previous debit to net pay and the various
withholdings.
(4) The payroll control account should be zero after posting both entries.
d. Properly classify workers as employees or independent contractors.
Misclassification can result in back taxes, interest, and penalties.
e. Have the HRM department review decisions to hire temporary or outside help.
5422.68 Threats and controls in disbursing payroll taxes and miscellaneous deductions
1. Threat: Hiring unqualified or dishonest employees. This can result in:
a. increased production expenses and
b. theft of assets.
2. Controls
a. Explicitly state skill qualifications in the job postings for each open position.
b. Require candidates to sign a statement on the job application form that:
(1) confirms the accuracy of the information on the form.
(2) gives permission to do a thorough background check.
c. Perform thorough background checks to verify:
(1) the applicant does not have a prior criminal record.
(2) job applicant skills and references.
(3) college degrees earned.
(4) actual job titles of people listed as references.
3. Threat: Employment law is violated. This can result in:
a. stiff penalties from the government and
b. civil suits by alleged victims of employment discrimination.
4. Controls
a. Carefully document all actions and decisions relating to advertising, recruiting, and
hiring new employees. This demonstrates compliance with applicable government
regulations.
b. Send employees to training courses that keep them current with employment law.
5422.70 The general ledger and reporting system plays a central role in a company’s information
system.
a. Primary functions of this system are to collect and organize:
(1) data from each of the accounting cycle subsystems,
(2) data from the treasurer about financing and investing activities,
(3) data from the budget department, and
(4) adjusting entries from the controller.
b. In the general ledger and reporting system, data is organized and stored such that the
information needs of internal and external users can be met. Examples of these
information needs are:
(1) managers need information to plan, control, and evaluate their operations.
(2) users need regular reports, as well as information, to meet their real-time inquiry
needs.
(3) investors and creditors need financial statements, as well as more detailed and
more frequent reports, to evaluate organizational performance and make
investment and credit-granting decisions.
(4) government agencies require periodic reports of organization performance to
evaluate compliance with laws and regulations.
c. There are four basic activities in the general ledger and reporting system (as shown in
Figure 4-12). Each of these activities is briefly explained. The threats a system faces
during each activity are also explained, along with the controls used to prevent the
threat.
Figure 4-12:
1. Update General
Ledger
2. Post Adjusting
Entries
3. Prepare 4. Prepare
Financial Managerial Reports
Statements
5422.72 Threats and controls in general ledger and reporting system (Part 1)
1. Threat: An error in updating the general ledger produces incorrect information, which
can result in:
a. misleading reports and
b. poor decisions
2. Controls
a. Input controls, such as the following:
(1) Making sure summary journal entries represent actual reporting period activity
(2) Validity checks to make sure a general ledger account exists for each journal
entry account number
(3) Field checks to make sure amount fields in a journal entry contain only numeric
data
(4) Zero-balance checks in journal entries to make sure total debits journal entries
equal total credits
(5) Completeness tests to make sure all pertinent journal entry data is entered
(6) Closed-loop verification tests to make sure account numbers match account
descriptions, so that the correct general ledger account is accessed
(7) Adjusting entry files for standard recurring adjusting entries, such as
depreciation expense
(a) Because the entries are not keyed in each time, this improves input
accuracy.
(b) Because the entries are not forgotten, this improves input completeness.
(8) Signing checks on general ledger account balances to make sure the balance is
of the appropriate nature (debit or credit)
(9) Run-to-run totals to verify batch processing accuracy
(a) Calculate new general ledger account balances, based on beginning
balances and total debits and credits applied to the account.
(b) Compare that total with the actual account balance in the updated general
ledger.
(c) Investigate any discrepancies, as they indicate a processing error that must
be corrected.
b. Reconciliations and controls that help detect general ledger updating errors
(1) Compare total debit balances to total credit balances in a trial balance to
determine if a posting error has occurred.
(2) Determine if clearing and suspense accounts have end-of-period zero balances.
5422.73 Threats and controls in general ledger and reporting system (Part 2)
1. Threat: Unauthorized access to the general ledger, which can result in:
a. confidential data leaks to competitors,
b. corruption of the general ledger, and
c. concealing asset theft
2. Controls
a. User IDs, passwords, and access control matrix to:
(1) control general ledger access and
(2) enforce proper segregation of duties, by limiting the functions legitimate users
may perform
b. Limit the functions performed at a terminal. For example, entering adjusted entries
can be restricted to terminals in the controller’s office.
c. Check for a valid authorization code for each journal voucher record before posting
that transaction to the general ledger.
(1) This authorization code is part of the audit trail.
(2) Inspecting the audit trail can detect unauthorized general ledger access.
Artificial Intelligence
5422.74 Artificial intelligence (AI) is the science and engineering of simulating human intelligence
processes by machines (i.e., making intelligent machines), especially intelligent computer
programs. These processes include learning, reasoning, and self-correction.
a. AI was coined by John McCarthy, an American computer scientist, in 1956. Today, it is
an umbrella term that encompasses everything from robotic process automation to
actual robotics.
b. AI has gained prominence recently due, in part, to big data, or the increase in speed,
size, and variety of data businesses are now collecting. AI can perform tasks such as
identifying patterns in the data more efficiently than humans, enabling businesses to
gain more insight from their data.
5422.75 AI can be categorized as narrow (or weak) AI, or general (or strong) AI:
a. Narrow/weak AI: Designed to perform a narrow task (e.g., facial recognition, Apple’s
Siri)
b. General/strong AI: Strong AI, also known as artificial general intelligence, is an AI
system with generalized human cognitive abilities so that when presented with an
unfamiliar task, it has enough intelligence to find a solution.
Machine Learning
5422.76 Machine learning (ML) is one application of artificial intelligence (AI) based around the idea
that we should be able to give machines access to data and let them learn for themselves.
Machine-learning applications can read text and work out whether the person who wrote it
is making a complaint or offering congratulations. They can also listen to a piece of music,
decide whether it is likely to make someone happy or sad, and find other pieces of music to
match the mood. In some cases, they can even compose their own music expressing the
same themes or which they know is likely to be appreciated by the admirers of the original
piece.
a. Successful systems often use a training set of data with millions of examples, each of
which has been labeled with the correct answer. The system can then look at new
examples and, if the training was sufficient, the system will predict answers with a high
rate of accuracy.
b. The algorithms that drive this process depend on an approach called deep learning,
which uses neural networks. Deep-learning algorithms have a significant advantage over
earlier generations of machine-learning algorithms because they can make better use of
much larger data sets.
c. A neural network is a computer system designed to recognize images and classify them
according to elements they contain. It basically works on a system of probability—based
on data fed to it, it is able to make statements, decisions, or predictions with a degree of
certainty. The addition of a feedback loop enables “learning”—by sensing or being told
whether its decisions are right or wrong, it modifies the approach it takes in the future.
5422.78 Machine learning (ML) brings with it certain risks. ML systems often have low
“interpretability,” meaning that humans have difficulty figuring out how the systems
reached their decisions. Deep neural networks may have hundreds of millions of
connections, each of which contributes a small amount to the ultimate decision.
a. Machines may have hidden biases, derived from the data provided to train the system.
For instance, if a system “learns” which job applicants to accept for an interview by
using a data set of decisions made by human recruiters in the past, it may inadvertently
learn to perpetuate their racial, gender, ethnic, or other biases. Moreover, these biases
may not appear as an explicit rule but, rather, be embedded in subtle interactions
among the thousands of factors considered.
b. Neural network systems deal with statistical truths rather than literal truths. That makes
it almost impossible to guarantee that the system will work in all cases—especially in
situations that were not represented in the training data. Lack of verifiability can be a
concern in mission-critical applications, such as controlling a nuclear power plant, or
when life-or-death decisions are involved.
c. When the ML system does make an error, diagnosing and correcting exactly what went
wrong can be difficult. The underlying structure that led to the solution can be
unimaginably complex, and the solution may be far from optimal if the conditions under
which the system was trained change.
5422.79 Most researchers agree that even a “superintelligent” AI object would not exhibit human
emotions like love or hate, and that there is no reason to expect AI to become intentionally
benevolent or malevolent. Instead, when considering how AI might become a risk, experts
think two scenarios most likely:
1. The AI is programmed to do something devastating: Autonomous weapons are artificial
intelligence systems that are programmed to kill; these weapon systems could easily
cause mass casualties. This risk is present even with narrow AI, but grows as levels of AI
intelligence and autonomy increase.
2. The AI is programmed to do something beneficial, but it develops a destructive
method for achieving its goal: This can happen whenever humans fail to fully align the
AI’s goals with our own, which is strikingly difficult. If a superintelligent system is tasked
with an ambitious geoengineering project, it might wreak havoc with our ecosystem as a
side effect, and view human attempts to stop it as a threat to be met.
Automation
5422.80 Automation is the technology which allows machines to perform tasks once performed by
humans via programmed commands, combined with automatic feedback control, to ensure
proper execution of the instructions without human intervention. Over time, automated
systems have become increasingly sophisticated and complex, often surpassing human
capabilities.
5422.81 Automation technology has resulted in the creation of a number of other technologies, such
as robotics, a specialized branch of automation in which the automated machine possesses
certain anthropomorphic, or humanlike, characteristics. The most humanlike characteristic
of an industrial robot is its powered mechanical arm, which can be programmed to move
through a sequence of motions to perform useful tasks, such as loading and unloading parts
at a production machine or making a sequence of spot-welds on the sheet-metal parts of an
automobile body during assembly.
5422.82 As automation has grown, so too has the fear that humanoid service robots and machine-
learning algorithms will replace millions of service workers in the future. In a widely noted
study published in 2013, Carl Benedikt Frey and Michael Osborne examined the probability
of computerization for 702 occupations and found that 47% of workers in America had jobs
at high risk of potential automation. They concluded that “recent developments in machine
learning will put a substantial share of employment, across a wide range of occupations, at
risk in the near future.”
5422.83 Figures published by the Federal Reserve Bank of St. Louis (in “Jobs Involving Routine Tasks
Aren’t Growing,” 2016) show that in America, employment in nonroutine cognitive jobs that
require highly paid, skilled workers and nonroutine manual jobs that require low-paid,
unskilled workers has grown steadily since the 1980s, whereas employment in routine jobs
has been broadly flat. As more jobs are automated, this “job polarization” or workforce
bifurcation seems likely to continue.
5423.03 Technological advances have made e-business possible. They have also reduced the costs
and increased the speed and accuracy with which business processes are executed. For
example:
a. Businesses can identify suppliers, compare prices, and negotiate prices more effectively
and efficiently.
b. Reverse auctions, where suppliers bid against each other to provide goods or services,
can be used to produce cost savings.
c. Companies have access to more accurate and timely information, which:
(1) allows organizations to better manage inventory.
(2) allows sellers to reduce transportation costs by combining shipments.
(3) significantly improves planning.
(4) allows organizations to improve the effectiveness of advertising and reduce its
costs.
5423.05 The human resources department can be more efficient and effective by allowing employees
to make changes, such as new withholdings, changes in retirement allocations, and name
changes.
a. This empowers employees, increases morale, and can generate cost savings by reducing
HR staff.
b. The effectiveness of the HR function is improved as staff focuses on value-added
activities rather than clerical tasks.
5423.07 1. The success of e-business initiatives is determined by the degree to which e-business:
a. activities are aligned with and support the organization’s overall business strategy.
Implementing e-business processes is not a basic strategy; it is using networking
and communications technology to more efficiently and effectively carry out the
organization’s business processes.
b. processes embody three key business transaction characteristics:
(1) Validity. Both transaction parties need to know the identity of the other to
make sure a transaction is authentic, valid, and enforceable.
(2) Integrity. Both parties want assurance that transaction information is accurate
and has not been altered during transmission.
(3) Privacy. If either party so desires, transaction information should be kept
private and confidential.
2. In traditional business processes:
a. authorizations and approvals (signatures, notarizations, and signature guarantees)
help ensure validity.
b. manual and automated error detection and correction techniques (signatures
across sealed envelopes, certified or hand-delivered communications) help ensure
accuracy and provide assurance that the contents of a message have not been
altered.
c. privacy is achieved by using techniques such as sealed envelopes and courier
delivery.
3. In e-business processes, concerns and objectives are the same but the ways they are
achieved differ. Control techniques are needed to provide more assurance about
validity, integrity, and privacy than is possible in traditional business processes.
5423.08 In addition to the potential for re-engineering processes through the Internet and other
technologies, the principles of operational effectiveness and continuous improvement
should also be considered. As new technologies emerge, organizations must continually
consider ways to improve the effectiveness and efficiency of their processes and operations.
b. There are several benefits to creating and electronically disseminating financial data
using XBRL:
(1) Organizations can publish information once, using XBRL tags, in a format that
anyone can use as they see fit.
(2) Information systems and analysis tools can interpret the XBRL tags. Recipients do
not need to manually reenter data they obtain electronically. Instead, they can be
fed directly into any information system or analysis tool that understands XBRL.
(3) XBRL tags result in more efficient and accurate Internet searches.
E-Business Communications
5423.13 A value-added network (VAN) offers specialized hardware, software, and long-distance
communications to private networks so they can exchange data.
5423.14 A virtual private network (VPN) uses the Internet to create a privately owned network that:
a. provides a secure, cost-effective method of providing remote access to an organization’s
network.
b. saves companies money, as the cost of the VPN software and higher-capacity lines
linking the organizational network to the Internet is less than the leased lines and toll-
free phones that the Internet replaces.
5423.15 A local area network (LAN) links computers and other devices located close to each other,
such as in the same building. These devices are connected to the network using a network
interface card (NIC). LANs can be configured in one of three basic ways:
1. In a star configuration, each device is connected to a central server that controls all
communications between devices. Devices send messages to the server, which forwards
it to the appropriate device.
2. In a ring configuration, each device is linked to two other devices. To control data flow,
a software token (these LANs are often called token ring networks) is continually passed
around the ring. To send a message, a device grabs the token and attaches a message.
Each device checks for messages as the token travels around the ring.
3. In a bus configuration, each device is connected to a main channel called a bus. A
software algorithm controls communications between devices. To send a message, a
device waits until the bus is free. The other devices check the message to see if it is for
them.
c. A metro-area network (MAN) connects multiple sites with multiple workstations for
shared use of common resources.
5423.18 In a client/server system a desktop computer, referred to as the client, sends a data request
to a server. The server retrieves the relevant subset of data from the database and sends it
to the client. Client/server systems can be configured in two ways:
1. A two-tiered system, where the central database is stored on the server and each client
has its own application software
a. Advantages: better able to meet user needs, and easier to design and control
b. Disadvantages: powerful clients, with abundant disk storage, are needed for local
processing; applications software must be purchased and maintained for each
client.
2. A three-tiered client/server system, where a top-tier server stores the central database
and a second-tier server uses applications programs to do most of the data processing.
All the client has to do is request data from the servers and format data.
a. Advantages: inexpensive “thin clients” or network PCs can be used, less applications
software to purchase and maintain, and lower data transmission costs
b. Disadvantages: high cost, and more complex applications development is required
Cryptocurrency
Blockchain
d. Documents can be digitized, coded, and inserted into the blockchain, creating a
sequential record that cannot be changed; the record’s authenticity can be verified by
the entire community using the blockchain instead of a single centralized authority.
5423.22 There are two types of blockchain networks: anonymous networks (also known as
permissionless), where each user has a copy of the entire blockchain and helps process and
confirm transactions; and non-anonymous (or permissioned) networks, where permission is
required to possess a copy of the blockchain and to help process and confirm transactions.
For both types of networks, the business logic is encoded using smart contracts. Smart
contracts are self-executing code on the blockchain framework that allows for straight-
through processing, which means that no manual intervention is required to execute
transactions. Smart contracts rely on data from outside entities referred to as “oracles” and
can act on data associated with any public address or with another smart contract on the
blockchain.
5431.01 Companies that do not adequately protect their personal computers and client/server
networks are vulnerable to the following threats and risks:
a. Unauthorized access to confidential data
b. Damage to computer files and equipment
c. Theft of information assets
5431.02 Personal computers (PCs) and networks are more vulnerable than mainframes because:
a. physical access is more difficult to control:
(1) PCs are everywhere, and many people are familiar with their use and operation.
(2) Networks have many PCs connected to them that must be controlled, and this
increases the risk that a network can be attacked.
b. PC users are usually not as security and control conscious as mainframe users.
c. it is difficult to segregate duties in a PC and network environment, and one person may
be responsible for both developing and operating a PC system.
d. end-user computing (EUC) allows users to develop their own information systems.
These systems are often difficult to integrate with the organizational information
system.
e. networks can be remotely accessed from almost anywhere using modems and phone
lines, the Internet, and EDI.
f. PCs and laptops are portable and subject to theft.
5431.03 Many mainframe control policies and procedures apply to PCs and networks. Other
important controls include the following:
a. Train PC users:
5431.04 All host devices (PCs and servers where programs reside) and applications (software on
those hosts) should be hardened.
a. The default configurations of host devices and operating systems often turn on many
optional settings or special-purpose services that are never used or are not essential.
This maximizes the likelihood of successful installation without the need for customer
support but creates security weaknesses because they have flaws, called vulnerabilities,
which can be exploited to either crash the system or take control of it.
b. Hardening is the process of modifying the configuration of hosts and application
software and deleting, or turning off, unused and unnecessary programs that represent
potential security threats.
Every host should have regularly updated anti-virus software and a software-based firewall
program. Security software can be used to control access to information system resources
such as data files, software, and program libraries.
5431.05 Companies need to exercise caution when conducting business on the Internet because:
a. many websites have serious security flaws.
b. hackers are attracted to the Internet and frequently try and break into company
systems linked to the Internet.
c. many websites are run by fraudsters that lure people into divulging confidential
information which they use or sell to others. Others offer to sell people goods, take their
money, and never deliver the goods.
d. before it arrives at its destination, an Internet message can easily pass through 6 to 10
computers that can read or electronically copy the message. Even messages sent on a
company’s secure intranet can be read.
5431.06 The following controls can improve Internet and telecommunications security.
5431.07 Encryption is transforming data, called plaintext, into unreadable gibberish, called
ciphertext. Decryption reverses this process, transforming ciphertext back into plaintext.
a. Both a key and an algorithm are used to encrypt plaintext into ciphertext and to decrypt
the ciphertext back into plaintext.
(1) The key is a fixed-length string of binary digits (a 128-bit key has 128 0s and 1s). The
longer the key, the less likely it is that someone can break the encryption code.
(2) The algorithm is a formula that combines the key and the text. A strong algorithm
that has been rigorously tested is difficult, if not impossible, to break by using
brute-force guessing techniques.
b. To encrypt a document:
(1) the data to be encrypted is divided into blocks the same length as the key.
(2) the formula is applied to each block of data, producing a ciphertext version of the
data that is the same size as the original.
c. To decrypt a document, the computer divides the ciphertext into 128-bit blocks and
then applies the decryption key to each block.
d. Data is encrypted using one of the following encryption systems:
(1) A symmetric, or single key, system uses the same key to encrypt and decrypt a
message.
(2) An asymmetric, or public key infrastructure (PKI), system uses two keys. The public
key, as the name implies, is publicly available. The private key is kept secret, and
only the owner of that pair of keys knows what it is. Either key can be used to
encode a message, but only the other key can be used to decode it.
e. A symmetric key is simple, fast, and efficient but has the following disadvantages:
(1) Since the sender provides the recipient with the secret key, the two parties need to
have some method for securely exchanging the key. System effectiveness depends
on controlling who knows the secret key.
(2) It requires a secret key for each different party with whom the company
communicates.
(3) Since both parties have the same secret key, there is no way to prove who created a
specific document and therefore no way to create legally binding agreements.
f. The PKI approach has the following advantages over single key systems:
(1) It eliminates problems with exchanging keys, and companies only need one key.
(2) It is more secure, because two different keys are used to encode and decode a
message.
(3) Anyone can use the organization’s public key to encrypt a message, but only the
organization can use its private key to decode the message.
(4) Data can be encrypted with the private key and decrypted with the public key. Since
this makes it possible to prove who created a document, legally binding electronic
agreements can be created.
(5) The main drawback to asymmetric encryption systems is that it is much slower than
single key systems—too slow to be used to exchange large amounts of data over
the Internet.
5431.08 Hashing is taking plaintext of any length and transforming it into a short code called a hash.
Hashing differs from encryption in two important aspects:
1. Encryption always produces ciphertext similar in length to the original plaintext, but
hashing always produces a hash that is of a fixed short length, regardless of the length
of the original plaintext.
2. Encryption is reversible, but hashing is not. Given the decryption key and the algorithm,
ciphertext can be decrypted back into the original plaintext. In contrast, it is not possible
to transform a hash back into the original plaintext, because hashing throws away
information.
5431.09 A digital signature uniquely identifies the sender of an electronic message, similar to how a
handwritten signature identifies the signer of a paper document.
a. In a public key infrastructure (PKI) system, a digital signature is created when the
sender’s private key encrypts the sender’s message. The message can only be decoded
using the corresponding public key.
b. A valid digital signature does not verify the identity of the private key’s owner. It only
proves that the message was sent by the owner of the private key that corresponds to
the public key used to decode the message.
5431.12 A certificate authority issues public and private keys, records the public key in a digital
certificate, and otherwise manages digital certificates.
a. The certificate authority’s digital signature is included in the digital certificate so a
certificate’s validity can be verified.
b. This functions like a hologram, watermark, or other device that verifies the validity of a
driver’s license or passport.
5431.13 Routers control the flow of information sent over the Internet or an internal local area
network.
a. Data to be sent is divided into packets and transmitted, and the device receiving the
packets reassembles the packets to recreate the original message or data.
b. Two important protocols, referred to as TCP/IP, govern the process for transmitting
information over the Internet.
(1) The Transmission Control Protocol (TCP) specifies the procedures for dividing data
into packets and reassembling them.
(2) The Internet Protocol (IP) specifies the structure of the packets and how to route
them to the proper destination.
c. Every IP packet consists of two parts: a header and a body.
The header contains the packet’s origin and destination addresses, as well as
information about the type of data contained in the body of the packet.
d. A router reads the destination address field in an IP packet header to determine where
it is to be sent.
Rules, referred to as an access control list (ACL), determine which packets are allowed
into a system.
e. A border router connects an organization’s information system to the Internet.
(1) It checks the contents of the destination address field of every packet it receives. If
the address is not that of the organization, the packet is forwarded on to another
router on the Internet.
(2) If the destination address matches that of the organization, the source and
destination fields in the IP packet header undergo a number of tests before being
allowed in.
(3) Packets that fail a test are not allowed into the system. Those that do not fail the
tests are passed on to the firewall, where they will be subjected to more detailed
testing before being allowed to enter the organization’s internal network.
5431.14 A firewall prevents outsiders and employees from gaining unauthorized access to a system.
a. It consists of hardware and software that control communications between:
5431.15 An intrusion prevention system (IPS) identifies and drops packets that are part of an attack.
a. An IPS examines the data in the body of an IP packet and therefore is more effective
than routers and firewalls that merely look at data in an IP header. There are several
ways to check the packet contents:
(1) Checking packet contents against a database of patterns (signatures) of known
attack methods
(2) Developing a profile of normal traffic and using statistical analysis to identify
packets that do not fit that profile
(3) Using rule bases that specify acceptable standards for specific types of traffic and
drop all packets that do not conform to those standards
b. The disadvantages of an ISP are:
(1) it is slow, as it takes more time to examine the body of an IP packet than an IP
packet header, and
(2) it is prone to false alarms, which result in rejecting legitimate traffic.
5431.16 The border router, firewall, and intrusion prevention system are complementary tools to
control which information is allowed to enter and leave the organization’s information
system. Multiple filtering devices are more efficient than any one device.
a. Border routers quickly filter out obviously bad packets.
b. Firewalls only allow packets to enter that appear to contain specific types of data for
specific types of programs.
c. The intrusion prevention system (IPS) verifies that packet contents actually conform to
the organization’s security policies.
5431.17 Tunneling can be used to create a virtual private network without the cost of leased private
lines. Tunneling is also used to safeguard internal networks in organizations. In tunneling:
a. networks are connected—firewall to firewall—via the Internet.
b. data are split into small Internet Protocol (IP) packets.
c. the packets are encrypted, mixed with millions of other IP packets, and sent over the
Internet.
d. at their destinations, packets are decrypted and the original message is reassembled.
(4) not automatically broadcasting the access point’s address, called a service set
identifier (SSID). This forces users to manually enter the SSID.
(5) having computer security or internal audit periodically find and disable rogue access
points and discipline the employees who installed them.
5431.20 Electronic envelopes are used to protect e-mail messages. The envelope is created using
private or public key encryption techniques and sent over the Internet.
a. The envelope containing the e-mail is opened using the keys and the message is
decrypted.
b. Encrypted e-mail message authenticity and integrity is guaranteed if the secrecy of the
keys is maintained.
Cybersecurity
5431.22 Cybersecurity is the practice of protecting systems, networks, devices, programs, and data
from digital attack, damage, or unauthorized access. These attacks are generally designed to
access, change, or destroy sensitive information such as intellectual property, financial data,
personal information, or other types of data for which unauthorized access or exposure
could have negative consequences.
5431.23 Some of the more common types of cybersecurity threats include the following:
a. Ransomware: A type of malicious software designed to extort money by blocking access
to files or the computer system until the ransom is paid
b. Malware: A type of software designed to gain unauthorized access or to cause damage
to a computer
c. Social engineering: A tactic designed to trick an individual or entity into revealing
sensitive information. Social engineering can be combined with any of the threats listed
here to make a person more likely to click on links, download malware, or trust a
malicious source.
d. Phishing: The practice of sending fraudulent emails that resemble emails from reputable
sources. The aim is to steal sensitive data like credit card numbers and login
information.
e. Denial of service (DOS): An attack that bombards the receiving server with so much
information that it shuts down, preventing legitimate users from accessing the service
or resources they need
5431.24 From a business perspective, multiple areas of risk need to be addressed, including:
a. security over networks, applications, data and databases, infrastructure, endpoint
devices (computers, smart devices, and routers), mobile devices, and cloud storage;
b. identity management;
c. disaster recovery/business continuity planning; and
d. end-user education.
Common technology used to protect these risk areas includes next-generation firewalls, DNS
(domain name system) filtering, malware protection, antivirus software, and email security
solutions.
5431.26 The National Cyber Security Alliance (NCSA) recommends a top-down approach to
cybersecurity in which corporate management is charged with prioritizing cybersecurity
measures across all business practices. NCSA advises that companies must be prepared to
“respond to the inevitable cyber incident, restore normal operations, and ensure that
company assets and the company’s reputation are protected.”
a. NCSA’s guidelines for conducting cyber-risk assessments focus on three key areas:
identifying an organization’s most valuable information requiring protection, identifying
the threats and risks facing that information, and outlining the damage an organization
would incur should that data be lost or wrongfully exposed.
b. Cyber-risk assessments should also consider any regulations that impact the way an
entity collects, stores, and secures data, such as the Payment Card Industry Data
Security Standard (PCI DSS), Health Insurance Portability and Accountability Act (HIPAA),
Sarbanes-Oxley Act (SOX), and Federal Information Security and Management Act
(FISMA).
5431.27 Many companies that do not have adequate personnel, expertise, equipment, or technology
for performing certain functions may choose instead to outsource those tasks to service
organizations. Common examples of outsourced functions include cloud computing,
security, retirement and investment plans, health care claims management and processing,
and payroll processing. Even though user management has essentially delegated these
functions to a service organization, they are generally still held accountable for establishing
effective controls over those functions by those charged with governance, customers,
shareholders, and regulators.
Service Organization Control (SOC) reports provide user management with information
about the service organization’s controls to assist in assessing and addressing the risks
associated with an outsourced service. These reports, prepared in accordance with AT-C 320,
Reporting on an Examination of Controls at a Service Organization Relevant to User Entities’
Internal Control Over Financial Reporting, are intended to meet the needs of the user
entities and the CPAs that audit the user entities’ financial statements, in evaluating the
effect of the controls at the service organization on the user entities’ financial statements.
(2) Availability: The system, product, or service is available for operation and use
as committed or agreed to by a contract or service level agreement (SLA). This
principle pertains to security-related criteria that may affect availability,
monitoring such items as network performance and availability, site failover,
and security incident handling.
(3) Processing integrity: Addresses whether a system achieves its purpose (i.e.,
delivers the right data at the right price at the right time). Accordingly, data
processing must be complete, valid, accurate, timely, and authorized.
(4) Confidentiality: Information that is designated “confidential” is protected as
committed or agreed to in the contract. Data is considered confidential if its
access and disclosure is restricted to a specified set of persons or organizations
(e.g., data intended only for company personnel, business plans, intellectual
property, or internal price lists). Encryption is an important control for
protecting confidentiality during transmission; network and application
firewalls and access controls are used to safeguard information being
processed or stored on computer systems.
(5) Privacy: Personal information is collected, used, retained, disclosed, and
disposed of in conformity with the commitments in the entity’s privacy notice
and with the privacy principles put forth by the AICPA. Personal identifiable
information (PII) refers to details that can distinguish an individual (e.g., name,
address, or Social Security number). Some personal data related to health,
race, sexuality, and religion are also considered sensitive and generally require
an extra level of protection.
c. There are two types of SOC 2 reports that service organizations can provide to user
entities. A Type 2 report is required if the user entity intends to use the report for
reliance on internal control or if the user entity’s management or auditor plans to
use the report for the assessment of internal controls.
(1) Type 1: In a Type 1 report, the service auditor provides an opinion as to
whether the service organization’s description “fairly presents” the system that
was designed and implemented, and whether the controls were suitably
designed to meet the criteria as of a specified date. Type 1 reports do not
address the operating effectiveness of controls, nor do they provide an opinion
over a period of time. In other words, the report only provides information on
controls that are in place (designed) at a specific point in time and not whether
the controls are operating on a continuous basis throughout a specified time
period.
(2) Type 2: In a Type 2 report, the service auditor provides an opinion on whether
the service organization’s description “fairly presents” the system that was
designed and implemented; the controls were suitably designed to meet the
criteria; the controls operated effectively during the specified period of time;
and the service organization is in compliance with the commitments in its
statement of privacy practices, if the report covers the privacy principle.
3. SOC 3: Report on Controls at a Service Organization Relevant to Security, Availability,
Processing Integrity, Confidentiality, or Privacy: SOC 3 reports are “public-facing”
documents that give a high-level overview of information in the SOC 2 report. A SOC 2
report contains sensitive information about specific systems and network controls which
should stay confidential; therefore, a SOC 3 report summarizes the non-sensitive
content in a SOC 2 report for users who need assurance about the controls at a service
organization relevant to security, availability, processing integrity, confidentiality, or
privacy but do not have the need for or the knowledge necessary to make effective use
of a SOC 2 report. Because they are general-use reports, SOC 3 reports can be freely
distributed.
5431.29 The table below, provided by the AICPA, compares SOC 1, SOC 2, and SOC 3 reports:
SOC 1 Reports SOC 2 Reports SOC 3 Reports
Under what SSAE 16, Reporting on AT 101, Attestation AT 101, Attestation
professional Controls at a Service Engagements Engagements
standard is the Organization
engagement AICPA Guide, Reporting on AICPA Technical Practice
performed? AICPA Guide, Applying SSAE Controls at a Service Aid, Trust Services
No. 16, Reporting on Controls Organization Relevant to Principles, Criteria, and
at a Service Organization Security, Availability, Illustrations
Processing Integrity,
Confidentiality, or Privacy
What is the To provide information to the To provide management of a To provide interested
purpose of the auditor of a user entity’s service organization, user parties with a CPA’s
report? financial statements about entities, and other specified opinion about controls at
controls at a service parties with information and the service organization
organization that may be a CPA’s opinion about that may affect user
relevant to a user entity’s controls at the service entities’ security,
internal control over financial organization that may affect availability, processing
reporting. It enables the user user entities’ security, integrity, confidentiality,
auditor to perform risk availability, processing or privacy.
assessment procedures, and if integrity, confidentiality, or
a Type 2 report is provided, to privacy.
assess the risk of material
misstatement of financial
statement assertions affected
by the service organization’s
processing.
What are the A description of the service A description of the service A service auditor’s report
components of organization’s system. organization’s system. on whether the entity
the report? maintained effective
A service auditor’s report that A service auditor’s report controls over its system as
contains an opinion on the that contains an opinion on it relates to the principle
fairness of the presentation the fairness of the being reported on, i.e.,
of the description of the presentation of the security, availability,
service organization’s system, description of the service processing integrity,
the suitability of the design of organization’s system, the confidentiality, or privacy,
the controls, and in a Type 2 suitability of the design of based on the applicable
report, the operating the controls, and the Trust Services criteria.
effectiveness of the controls. operating effectiveness of
the controls.
In a Type 2 report, a
description of the service In a Type 2 report, a
auditor’s tests of the controls description of the service
and the results of the tests. auditor’s tests of controls
and the results of the tests.
5431.30 To increase the flexibility of SOC 2 reports, and to better address increased cybersecurity
risks, the AICPA released updated Trust Services criteria. The 2017 Trust Services Criteria for
Security, Availability, Processing Integrity, Confidentiality, and Privacy presents control
criteria established by the Assurance Services Executive Committee (ASEC) of the AICPA for
use in attestation or consulting engagements to evaluate and report on controls over the
security, availability, processing integrity, confidentiality, or privacy over information and
systems (a) across an entire entity; (b) at a subsidiary, division, or operating unit level; (c)
within a function relevant to the entity's operational, reporting, or compliance objectives; or
(d) for a particular type of information used by the entity.
5431.31 To supplement the 17 COSO principles, the TSC were organized into the following
categories:
a. Logical and physical access controls: Relate to how the service organization implements
logical and physical access controls that serve to prevent unauthorized access and
protect data assets
b. System operations: The criteria relevant to how the service organization manages the
operation of system(s) and detects and mitigates processing deviations, including logical
and physical security deviations
c. Change management: How service organizations evaluate and determine necessary
changes in infrastructure, data, software, and procedures, which gives them the ability
to securely make changes and prevent unauthorized changes
d. Risk mitigation: The criteria relevant to how the service organization identifies, selects,
and develops risk mitigation activities arising from potential business disruptions and
the use of vendors and business partners
5431.32 The AICPA has provided a mapping of the previous 2016 Trust Services criteria to the
updated 2017 criteria. Below is an example of the mapping:
2017 Trust Services Criteria (TSC) 2016 Trust Services Principles and
Criteria (TSPC)
TSC Ref Criteria Points of Focus TSC Criteria
# CONTROL
ENVIRONMENT
CC1.1 COSO Principle 1: CC1.4 The entity has established
The entity workforce conduct
demonstrates a standards, implemented
commitment to workforce candidate
integrity and background screening
ethical values. procedures, and conducts
enforcement procedures to
enable it to meet its
commitments and system
requirements as they relate
to [insert the principle(s)
addressed by the
engagement: security,
availability, processing
integrity, confidentiality, or
privacy, or any combination
thereof].
Sets the Tone at the Top—
The board of directors and
management, at all levels,
demonstrate through their
directives, actions, and
behavior the importance of
integrity and ethical values
to support the functioning of
the system of internal
control.
Establishes Standards of
Conduct—The expectations
of the board of directors and
senior management
concerning integrity and
ethical values are defined in
the entity’s standards of
conduct and understood at
all levels of the entity and by
outsourced service providers
and business partners.
Evaluates Adherence to
Standards of Conduct—
Processes are in place to
evaluate the performance of
individuals and teams
against the entity’s expected
standards of conduct.
2017 Trust Services Criteria (TSC) 2016 Trust Services Principles and
Criteria (TSPC)
TSC Ref Criteria Points of Focus TSC Criteria
# CONTROL
ENVIRONMENT
Addresses Deviations in a
Timely Manner—Deviations
from the entity’s expected
standards of conduct are
identified and remedied in a
timely and consistent
manner.
Considers Contractors and
Vendor Employees in
Demonstrating Its
Commitment—
Management and the board
of directors consider the use
of contractors and vendor
employees in its processes
for establishing standards of
conduct, evaluating
adherence to those
standards, and addressing
deviations in a timely
manner.
5431.33 Applying the Trust Services criteria requires judgment; therefore, points of focus are
provided for each criterion. COSO’s Internal Control—Integrated Framework states that
points of focus (POF) represent important characteristics of the criteria; POF are intended to
assist management when designing, implementing, and operating controls over security,
availability, processing integrity, confidentiality, and privacy. In addition, POF can assist both
management and the practitioner when they are evaluating whether the controls were
suitably designed and operated effectively to meet the Trust Services criteria.
5431.34 The Trust Services criteria were designed to provide flexibility in application and use for a
variety of different subject matters. Topics a practitioner may be engaged to report on using
the Trust Services criteria include the following:
a. The effectiveness of controls within an entity's cybersecurity risk management program
to achieve the entity's cybersecurity objectives using Trust Services criteria relevant to
security, availability, and confidentiality as control criteria in the cybersecurity risk
management examination
b. The suitability of design and operating effectiveness of controls included in
management's description of a service organization's system relevant to one or more of
the Trust Services criteria over security, availability, processing integrity, confidentiality,
or privacy throughout a specified period to meet those criteria in a Type 2 SOC 2
engagement.
c. The design and operating effectiveness of a service organization's controls over a system
relevant to one or more of the Trust Services criteria over security, availability,
processing integrity, confidentiality, and privacy in a SOC 3 engagement
d. The suitability of design and operating effectiveness of controls of an entity, other than
a service organization, over one or more systems relevant to one or more of the Trust
5431.35 If the practitioner is engaged to report on compliance with laws, regulations, rules,
contracts, or grant agreements in connection with an examination of the design and
operating effectiveness of an entity's controls, the compliance portion of the engagement
would be performed in accordance with AT-C 105, Concepts Common to All Attestation
Engagements, and AT-C 315, Compliance Attestation.
5432 IT Controls
5432.01 The following five classifications of controls are used to make systems more secure:
1. Segregation of systems duties
2. Physical access controls
3. Logical access controls
4. Personal computers and client/server network protection
5. Internet and telecommunication controls
5432.02 When actual, physical access to computer equipment is not restricted, computers, files, and
data can be damaged or accessed by unauthorized parties. Physical access security can be
achieved using the following controls.
5432.03 Where possible, there should only be one unlocked building entrance during normal
business hours. Emergency exits should not permit outside entry and an alarm should sound
when the exit is opened.
5432.04 A security guard or receptionist should be stationed at the main entrance to:
a. verify the identity of employees,
b. require all visitors to sign in, and
c. have an employee escort visitors to their building destination.
5432.05 Locate sensitive computer equipment, programs, and data in locked rooms.
a. Securely lock all entrances to the computer room.
b. Use security guards and closed-circuit television systems to monitor all entrances.
c. Require proper employee ID to gain access to the computer room. Security badges that
have photos and magnetic, electric, or optical codes can be used to automatically log
each employee’s entry and exit. This log should be reviewed periodically by supervisory
personnel.
d. Multiple failed access attempts should trigger an alarm.
e. Rooms where sensitive hardware, software, or data are stored should supplement locks
with other security technologies such as card readers, numeric keypads, or various
biometric devices such as retina scanners, fingerprint readers, or voice recognition.
5432.07 Install security alarms that can detect unauthorized, off-hours access.
5432.08 Lock PCs, laptops, and other computer devices when they are not in use.
5432.09 Place hardware and critical system components away from hazardous or combustible
materials.
5432.10 Restrict access to private, secured telephone lines and to authorized terminals or PCs.
Restrict access to telephone and LAN wiring to prevent wiretapping. Do not expose cables
and wiring to areas accessible to the public. Disconnect wall jacks not in current use.
Securely lock wiring closets containing telecommunications equipment.
5432.11 Laptops, cell phones, and PDA devices require special attention to prevent their theft and
the loss of the data they contain.
a. Employees should always lock their laptops to an immovable object.
b. Store sensitive data on removable media, rather than the hard drive, in an encrypted
format and lock it up at night.
c. Install software on laptops so that if it is stolen the laptop will automatically dial a toll-
free number or use Wi-Fi positioning to reveal its current location when the thief
attempts to use it to connect to the Internet.
5432.12 Restrict access to network printers, because they often store document images on their hard
drives. There have been cases where intruders have stolen the hard drives from printers,
thereby gaining access to sensitive information.
5432.13 Install fire and smoke detectors and use fire extinguishers that do not damage computer
equipment.
5432.14 People can damage information systems without having physical access to the system. They
only need to be able to reach the systems electronically. Several levels of logical access are
needed:
a. Authentication, which ensures that unauthorized users and devices are not allowed to
access any part of the system
b. Authorization, which makes sure the system can recognize authorized users, but restrict
their access to:
(1) data they are not allowed to use and
(2) those functions (reading, copying, updating, adding, deleting, modifying,
downloading, printing) they are authorized to perform
5432.15 Inadequate logical access controls leave a company vulnerable to the unauthorized access
and use of systems software, application programs, data files, and other system resources.
5432.16 The following controls are used to restrict logical access: user identification numbers and
passwords, physical possession identification, and biometric identification.
5432.17 User identification numbers and passwords. This is also known as knowledge identification.
Identification numbers and passwords are used to access a system:
a. Users enter a unique user ID, such as an employee number, and a password (letters,
numbers, and/or special characters known only to the user and the system).
b. The user ID and password are compared to user ID numbers and passwords stored in
the system to see if they are valid.
c. If they are valid, the system assumes an authorized user.
d. The disadvantage of passwords is they can be guessed, lost, or given away, allowing an
unauthorized person to access the system.
5432.19 Physical possession identification. Employees are often identified by something they
possess, such as an ID badge or a smart card.
a. Computers and security devices (door locks, etc.) read the ID cards.
b. Security is strengthened if an ID card and a password are both required to access a
system.
c. Their disadvantage: They can be lost, stolen, or given away.
5432.20 Biometric identification. Unique physical characteristics (fingerprints, voice patterns, retina
prints, facial patterns and features, body odor, signature dynamics, and keyboarding
patterns) can all be used to identify people. A device reads a user’s biometrics and compares
it against those stored in the computer.
a. Biometric devices should be able to adapt to slight personal changes, such as bloodshot
eyes.
b. Remote users must possess a biometric identification device to use the system.
c. Their disadvantage: They are expensive, can be hard to use, and are not 100% accurate.
(1) They may reject legitimate users (voice recognition systems may not recognize an
employee who has a cold).
(2) They may allow access to unauthorized people.
(3) Some, such as fingerprints, have negative connotations that hinder their
acceptance.
(4) The biometric templates (the digital representation of an individual’s fingerprints or
voice) must be stored. Any compromise of those templates would cause serious
problems for people.
5432.22 Authorization controls are implemented using compatibility tests and access control
matrices.
a. Compatibility tests. Companies should classify data based on how its loss or
unauthorized use would impact it and determine the data and program access privileges
of employees and outsiders. When users request access to data or programs or try to
operate the system, a compatibility test can determine if the user is authorized to
perform the desired action. This prevents unintentional errors and deliberate attempts
to manipulate the system. Several confidentiality levels are defined and used:
(1) Some data does not need to be restricted and is put on a website.
(2) Some data is restricted to employees.
(3) Confidential data is restricted to owners and appropriate top management and
employees.
(4) No one should be able to read, add, delete, and change data without authorization
of their activities.
b. Access control matrix. Compatibility tests use an access control matrix, which is:
(1) a list of authorized user ID numbers and passwords and
(2) a list of all files, data, and programs and the access each user has to them.
5432.23 Authentication and authorization controls are also used for computing devices.
a. Each device (PC, workstation, printer, etc.) requires a network interface card (NIC) to
connect to a company’s internal network.
b. Each NIC has a unique identifier, called a media access control (MAC) address.
c. Access should be restricted to devices with recognized MAC addresses. The MAC
addresses can also be used for authorization. For example, only devices in the payroll
department and belonging to appropriate top management should have access to
payroll data.
5433.03 Systems need to be protected from occurrences such as software and hardware failures,
sabotage, political and natural disasters, and power outages. As management relies more on
information systems for support, system downtime has the potential to interrupt business
operations. Three categories of controls are used to ensure information system availability:
(1) minimizing system downtime, (2) disaster recovery plan, and (3) data and program file
backups.
b. Insurance to:
(1) replace equipment lost in the disaster
(2) compensate for business interruptions
c. Specific assignments. A plan coordinator should:
(1) be responsible for implementing the recovery plan.
(2) assign individuals and teams specific recovery responsibilities such as finding new
physical facilities, operating the system, installing software, setting up data
communications linkages, recovering data, and procuring forms and supplies.
d. Backup computer and telecommunications facilities, which can be arranged by:
(1) establishing reciprocal agreements with companies with compatible facilities so
each company can use the other’s computers if an emergency occurs
(2) signing a contract for a contingent site. A hot site is configured to meet user
requirements. A cold site has everything needed (power, air conditioning, and
support systems) to quickly install a computer. Cold site users rely on their
computer vendors for prompt delivery of equipment and software if an emergency
occurs.
(3) fail-soft distributing processing capacity in a multilocation organization so other
facilities can take over if one location is damaged or destroyed
(4) investing in duplicate hardware, software, or data storage devices for critical
applications
e. Periodic testing and revision. A recovery plan must be:
(1) regularly tested with a simulated disaster, with each disaster recovery team
carrying out its assigned activities.
(2) constantly improved, since most plans fail their initial test and tested plans rarely
anticipate and deal with all real-life disaster problems.
(3) reviewed to make sure it reflects current computer application changes, equipment
configurations, and new personnel assignments.
f. Complete documentation of all aspects of the system.
(1) Copies of the documentation should be stored securely at multiple locations.
(2) One copy should be some distance from the system.
d. To protect data privacy, all data should be encrypted before being transmitted.
5433.08 As shown in Figure 4-13, batch processing files are backed up using the grandfather-father-
son concept.
Figure 4-13:
Tuesday’s
Transaction Update
File (1) Program
Tuesday’s FATHER
Master
File (B)
Tuesday’s Processing:
Wednesday’s
Transaction Update
File (2) Program
Wednesday’s SON
Transaction
File (C)
5433.10 PC hard drives can be backed up on CDs, diskettes, and tape files.
Do not store the on-site backup copy next to the PC, as it can be destroyed by any disaster
that wipes out the PC.
5433.11 The company should periodically practice restoring a system from the backup data so
employees know how to quickly restart the system if a failure occurs.
Data Entry
PROCESSING
Manual or Semi-
Automated Entry (via Data Storage
Barcoding or OCR) of
Original Documents
Automatic Data
Capture by ATMs,
Point of Sale
Machines, etc.
5441.01 Historically, paper source documents have been used to collect data about business
activities. This data was later entered into the computer.
5441.02 Today, as much business activity data as possible is recorded directly into computers, often
through a computer data entry screen that retains the same name as the paper source
document it replaced. Computer input devices allow data to be entered into the computer.
Examples of input devices include the following:
a. Keyboard
b. Image scanner
c. Magnetic ink character recognition (MICR)
d. Pointing device (graphics tablet, joystick, light pen, mouse, touchpad, touchscreen)
Additionally, software can be combined with input devices to capture data. One example is
optical character recognition (OCR) software, which can be used to convert images of paper
documents as read by a scanning device into text document computer files.
5441.03 Data such as the following is collected about each business activity or transaction:
a. Date and time it occurs
b. Employees involved
c. Location of the activity or transaction
d. Item(s), such as inventory, that are involved
e. Quantity and price of each item
f. Total amount of the business activity or transaction
g. Bill-to and ship-to addresses
h. Any special delivery or receiving instructions
i. Customer or vendor name
5441.04 The accuracy, control, and efficiency of data input are improved by:
a. using well-designed source documents and data entry screens.
b. grouping logically related data close together.
c. using appropriate shading and borders to clearly separate data items.
d. using prenumbered source documents or by having the system automatically assign a
sequential number to each new transaction. This makes it easier to verify that all
transactions have been recorded and that none of the documents have been misplaced.
e. providing instructions or prompts about the data collected.
f. using check-off boxes or pull-down menus to present the available options.
g. using machine-readable turnaround documents, which are records of company data
sent to an external party and then returned to the system as input to facilitate data
processing. (An example is a utility bill that is read by a special scanning device when it is
returned with a payment.)
h. using source data automation devices (ATMs, bank magnetic ink character devices,
point-of-sale (POS) scanners, and barcode scanners) that capture transaction data in
machine-readable form at the time and place of their origin.
Data Processing
5441.05 The most common data processing activity is data maintenance, which is the periodic
processing of transactions to update stored data. There are four basic types of data
maintenance:
1. Additions (Insert new records into a master file.)
Example: Add a new customer to the customer master file.
2. Deletions (Remove records from a master file.)
Example: Purge the vendor master file of all inactive vendors.
5441.06 Following are the data maintenance steps required to update an accounts receivable master
file record to reflect a sales transaction (as shown in Figure 4-15):
a. The account number field is used to match each transaction file record with the
appropriate record in the master file. In this example, account 0123 is being updated to
reflect the effects of a sales transaction that occurred on February 19, 20XX, as recorded
on invoice number 9876.
b. The sale amount ($360) is added to the existing account balance ($1,500) to produce the
updated current account balance ($1,860).
c. The current balance on the old master file ($1,500) is the new previous balance
(replacing $1,000) in the new master file record.
Figure 4-15:
FILE UPDATE
PROCESS (repeated for all
transactions)
UPDATED ACCOUNTS
RECEIVABLE RECORD (new master file)
Data Storage
5441.08 Another key component of an information system is the storage of data captured and
processed.
Figure 4-16:
Attributes
3 Entities Data
3 Records 16153 Baker Company 1460 Stocks 40,000 21,976 Values
Individual Fields
As this accounts receivable file stores information about three separate entities (Able, Baker,
and Charlie Companies), there are three records in the file. As each record has five separate
attributes (customer number, customer name, address, credit limit, and balance), there are
five different fields in each record. Each field stores a data value that describes an attribute of
a particular entity (customer). For example, 16154 is the customer number of Charlie
Company.
In a spreadsheet, each row represents a different record and each column represents an
attribute. Each cell represents a field. Thus, each intersecting row and column in a
spreadsheet is a field within a record, the contents of which are called a data value.
These data hierarchies are shown in Figure 4-17.
Figure 4-17:
DATABASE Database
e. Reference files are files that are referred to during processing. These would include
items such as price lists and tax rates.
f. A flat file is a traditional file system characterized by having only fixed-length “flat” files.
All files are of equal length.
Databases
5441.13 Historically, new files and programs were created each time an information need arose. This
resulted in a proliferation of master files and the same data stored in two or more separate
master files. When data was updated on one file and not on the other, data inconsistencies
arose. This led to the creation of databases that are an organizational resource managed for
the entire organization, not just a department or function.
A database is a set of interrelated, centrally coordinated files:
a. Their use minimizes or eliminates the proliferation of master files and data
redundancies.
b. Data integration and data sharing is accomplished by combining master files into larger
databases accessed by many application programs.
c. A database management system (DBMS) is a complex software package that permits
users to access information from the database. In addition to basic data movement
(utility) services, the DBMS provides for access and identification security, concurrent
use of data, and backup and recovery. The DBMS is “application independent” and does
not actually run application programs.
d. A database system consists of the database, the DBMS, and application programs that
access the database through the DBMS.
e. The data dictionary contains a description of all data elements, stores, and flows in a
system. Typically, a master copy of the data dictionary is maintained to ensure
consistency and accuracy throughout the development process. The data dictionary is
created using a data definition language (DDL).
f. The data query language (DQL) is used to interrogate the database.
g. The data manipulation language (DML) provides programmers with a facility to update
the database.
h. Data control language (DCL) is used to specify privileges and security rules.
(2) This allows users to change the way they view data relationships (their logical view
of the task) without changing how the data are stored physically. Database
administrators can also make changes to the way data are physically stored to
improve system performance and not have the change affect users or application
programs.
c. The term “schema” was coined to describe the logical structure of a database. In a
database there are several levels of schemas. Two that are important to accountants
are:
(1) the conceptual-level schema lists all data elements and the relationships between
them. As a result, it is referred to as the organization-wide view of the database.
(2) an external-level schema is the portion of the database that is of importance to an
individual user. Each external-level schema is referred to as a subschema.
d. The difference between the two schema levels is illustrated using the revenue cycle.
(1) The conceptual schema contains information about all revenue cycle data such as
customers, sales, cash receipts, salespeople, cash, and inventory.
(2) Each external-level subschema meets the needs of a particular user and is designed
such that users only have access to the data they need to do their job. Access to
other data items is prohibited.
For example, the sales order entry subschema allows access to customer credit
limits and current balances as well as inventory quantities and prices. It would not
allow access to inventory costs or company bank account balances.
Cloud Computing
5441.16 Cloud computing is the practice of using a network of servers, storage, databases, and
software over the Internet (i.e., ”the cloud”) to process, manage, store/backup/recover, and
analyze data, rather than a local server or a personal computer. Shared resources, software,
and information are provided to user computers as a metered service (like electricity). Cloud
computing services can be private, public, or hybrid.
a. Private cloud: Services are delivered internally (on-premises) from a business's own
data center, maintaining flexibility and convenience while preserving management,
ownership control, and security. A private cloud can be physically located on the
company’s on-site data center, or companies can pay a third-party service provider to
host their private cloud.
b. Public cloud: A third-party provider (e.g., Amazon Web Services, Google Compute
Engine) delivers the cloud service over the Internet. All hardware, software, and other
supporting infrastructure is owned and managed by the cloud provider; these services
are accessed and managed using a web browser. Public cloud services are sold on-
demand, typically by the minute or the hour. Customers only pay for the resources they
consume.
c. Hybrid cloud: A combination of public and on-premises private clouds. Companies can
run critical workloads or sensitive applications on the private cloud while using the
public cloud for on-demand workloads. Hybrid cloud gives businesses greater flexibility
and more deployment options.
5441.17 Most cloud computing services fall into three broad categories: infrastructure as a service,
platform as a service, and software as a service. These are sometimes called the cloud
computing stack, because they build on top of one another.
1. Infrastructure as a service (IaaS): IaaS provides the basic building blocks for cloud IT and
typically provides access to IT infrastructure assets—servers and virtual machines (VMs),
storage, networks, operating systems—from a cloud provider who charges on a pay-as-
you-go basis.
2. Platform as a service (PaaS): PaaS refers to cloud computing services that supply an on-
demand environment for developing, testing, delivering, and managing software
applications, allowing developers to focus on creating and delivering those applications
rather than worrying about resource procurement, capacity planning, software
maintenance, or infrastructure management.
3. Software as a service (SaaS): SaaS is a method for delivering software applications over
the Internet, on demand and typically on a subscription basis. With SaaS, cloud
providers host and manage both the software application and underlying infrastructure,
and handle maintenance, like software upgrades and security patching. Users connect
to the application over the Internet, usually with a web browser on their phone, tablet,
or PC (personal computer).
5441.18 Benefits from cloud computing include, but are not limited to, the following:
a. Cost: Cloud computing eliminates the capital expense of investing in data centers,
servers, and hardware, not to mention operating costs and IT staff expenses. Entities
pay only for computing resources actually consumed. Additionally, businesses can lower
their variable usage cost because traffic from hundreds of thousands of customers is
aggregated in the cloud, allowing providers to achieve higher economies of scale and
lower pay-as-you-go rates.
b. Speed: Most cloud computing services are provided on demand, so enormous amounts
of computing resources can be available in minutes, improving flexibility and reducing
capacity planning issues for businesses. The network can process very high volumes of
data with minimal delay, reducing latency and thereby providing nearly real-time access
to rapidly changing data.
c. Reliability: Cloud computing makes data backup, disaster recovery, and business
continuity easier and less expensive, because data can be mirrored at multiple
redundant sites on the cloud provider’s network.
d. Productivity: On-site data centers typically require a lot of “racking and stacking”—
hardware setup, software patching, and other time-consuming IT management chores.
Cloud computing removes the need for many of these tasks, so IT teams can be more
focused on achieving management’s customer and business objectives.
e. Performance: The biggest cloud computing services run on a worldwide network of
secure data centers, which are regularly upgraded to the latest generation of fast and
efficient computing hardware. This offers several benefits over a single corporate data
center, including reduced network latency for applications and greater economies of
scale.
ERP Systems
5441.19 Traditionally, the accounting information system (IS) was called a transaction processing
system because its only concern was financial data and accounting transactions. Useful
nonfinancial information, such as the time of day it occurred, was collected and processed
outside of the accounting information system. It was difficult to effectively integrate data
from the financial and nonfinancial systems and to eliminate data redundancies.
Enterprise resource planning (ERP) systems are designed to overcome these problems. They
integrate all aspects of a company’s operations with its traditional IS. For example, when an
order is entered, transaction data automatically flows to all affected parts of the company.
Inventory is updated, production schedules are adjusted, and purchase orders are initiated
to acquire any needed raw materials and supplies.
a. Collect and store important nonfinancial data, such as the time the activity occurred.
b. Integrate financial and nonfinancial operating data, as both are required for proper and
complete performance evaluation.
c. Collect and store data from external sources, such as data about customer satisfaction,
to determine if the company is meeting its customer’s requirements and expectations.
5441.20 Once data has been captured, processed, and stored, the data must be accessible for use by
the company. Data output, or reporting, is therefore a key component of an information
system.
5441.21 When displayed on a monitor, output is referred to as a soft copy. When printed on paper, it
is referred to as a hard copy.
5441.23 The information an IS (information system) provides falls into three main categories:
1. Financial statements. Financial statements are used by:
a. external parties to make decisions about extending credit to or investing in the
organization.
b. internal parties to meet stewardship requirements.
Sarbanes-Oxley
5442.04 Historically, internal control has been defined as the plan of organization and the methods a
business uses to:
a. safeguard assets.
b. provide accurate and reliable information.
c. promote and improve operational efficiency.
d. encourage adherence to prescribed managerial policies.
5442.05 An internal control structure is the policies and procedures an organization establishes to
provide reasonable assurance that organizational objectives are achieved. A system that
provides complete assurance is difficult, if not impossible, to design and is prohibitively
expensive.
1. A belief system conveys key company values to employees and motivates them to
adhere to such ideals.
a. A belief system should be broad enough to appeal to employees at all levels of the
organization.
b. The belief system should help employees comprehend the direction management
wants the company to take.
2. A boundary system promotes ethical behavior by setting limits beyond which an
employee must not pass.
a. A boundary system should not tell employees exactly what they must do in every
situation, as this can discourage creativity and initiative.
b. Employees should be allowed to meet customer needs and solve problems using
their own ingenuity, as long as they operate within certain limits.
3. A diagnostic control system compares actual performance to planned performance in
order to evaluate company progress.
a. A diagnostic control system facilitates the monitoring of critical performance
outcomes and the progress of individuals, departments, or locations as they
endeavor to attain strategically important goals.
b. This system provides feedback that enables management to modify and fine-tune
inputs and processes so future outputs will more closely match goals.
4. An interactive control system helps top-level managers with high-level activities that
require frequent attention, such as setting company objectives and developing company
strategy.
a. An interactive control system generates data that should be analyzed and discussed
in meetings with superiors, subordinates, and peers present.
5442.08 Control Objectives for Information and Related Technology (COBIT®) was developed by the
Information Systems Audit and Control Association (ISACA). COBIT consolidates standards
from 36 different sources into a single framework of generally applicable IT security and
control practices that:
a. helps balance risk and controls in information systems.
b. provides assurance that security and IT controls are adequate.
c. guides auditors as they form opinions and advise management on internal controls.
5442.09 COBIT is organized using four domains and 34 processes. Each control activity utilizes IT
resources, applies to a process, and satisfies a business requirement. The COBIT framework
looks at controls in three different ways:
1. Business requirements. Information must conform to seven distinct, but overlapping
business requirements:
a. Availability
b. Compliance with legal requirements
c. Confidentiality
d. Effectiveness (relevant, pertinent, and timely)
e. Efficiency
f. Integrity
g. Reliability
2. IT resources:
a. Application systems
b. Information
c. Infrastructure
d. People
3. IT processes:
a. Domains
b. Processes
c. Activities
5442.11 General controls make sure an organization’s overall control environment is secure and well
managed.
(2) Data entry devices verify the check digit each time the ID number is entered by
using the six digits to recalculate the seventh check digit.
(3) The verification calculation will not match the check digit if an error is made in
entering the six digits or the check digit.
i. Use key verification.
(1) An employee rekeys data entered through a keyboard.
(2) The system compares the two sets of keystrokes.
(3) Discrepancies are highlighted for correction.
(4) Key verification is used for crucial input such as customer numbers and amounts.
(4) Review the log periodically to make sure all errors were corrected.
(5) Summarize errors by record type, error type, cause, date, and disposition in an
error report sent to management.
(2) scheduling data processing tasks, so data is not processed without authorization,
and
(3) maintaining an audit trail, so the system can track people who have access to
confidential data.
b. Require employees to sign confidentiality agreements.
c. Create a data control function that:
(1) logs input data as it is received, entered into the system, processed, and dispersed
as output,
(2) checks to make sure all transactions are authorized,
(3) monitors data processing activities for errors and problems,
(4) develops control totals over input and reconciles them to the totals from each
subsequent processing activity,
(5) notifies users of input errors and problems, and
(6) reenters all errors that have been corrected.
d. Perform the following reconciliation procedures on a periodic and regular basis:
(1) Reconcile all system updates to control reports, file status/update reports, or other
control reports.
(2) Reconcile general ledger accounts to subsidiary account totals.
(3) Reconcile database totals with data maintained outside the system.
e. Test stored data to determine whether it is out of date (employees retire or quit;
suppliers and customers move or fail). Unneeded data should be purged.
f. Use the following file labels to protect data files from inadvertent misuse:
(1) External labels that contain the file name, contents, and date processed
(2) Internal labels, such as (1) volume labels that identify the data recording medium
(hard disks, diskettes, and tapes), (2) header labels that contain the file name,
expiration date, and other identification data, and (3) trailer labels that contain file
control totals. These totals are compared to those calculated during data
processing.
g. Use write protection mechanisms that protect files from being accidentally destroyed or
erased.
(1) A tape file protection ring prevents a tape file from being written upon.
(2) Diskettes, USB flash drives, and most other forms of solid-state storage have on/off
switches that accomplish the same thing.
h. Implement the following database protection mechanisms:
(1) Procedures for accessing and updating the database
(2) A data dictionary to make sure data is defined and used consistently
(3) Concurrent update controls to prevent problems when multiple users
simultaneously update a record (Lock other users out of the system until one has
finished updating the file.)
i. Use data conversion controls to make sure new files and databases are free from errors
when data from old files and databases are converted to the new format.
(1) Process data using both the old and new system at least once.
(2) Compare the results to identify differences and problems.
(3) Have internal auditors carefully supervise and review data conversion activities.
j. Protect stored data.
(1) Use a data library, where data files are logged in and out by a librarian.
(2) Back up data files and store a copy at a secure off-site location.
(3) Protect the data library from conditions that harm stored data such as fire, dust,
and excess heat or humidity.
Control Monitoring
5442.19 With the advent of Section 404 of the Sarbanes-Oxley Act and the implementation of top-
down, risk-based approaches for focusing on internal controls, organizations are increasingly
turning to automated controls and continuous monitoring.
5442.21 Most IT controls—from input validation routines to parity checks and other output
controls—are automated by definition, and can therefore have some form of continuous
monitoring applied to them. Many manual controls processes, such as transaction log
reviews and even some basic reconciliation processes, can also be automated.
Operational Effectiveness
5442.22 When thinking of what it takes to make an organization successful, operational effectiveness
is the counterpoint to an organization’s strategy. An organization without strategy has no
direction to follow; however, an organization without operational effectiveness may not be
around long enough to achieve the objectives laid out in its strategies.
a. What is an “effective operation”? By definition, an organization is operationally effective
when it performs activities, including processes, better than its competitors. Continuous
improvement and operational effectiveness therefore go hand in hand. An organization
must be continually adapting its operations and improving its functions to remain
competitive.
b. The implications of continuous improvement impact the entire organization—from the
most granular of transactional processing to the way management makes and carries
out decisions at the top of the organization. The Japanese word kaizen, which focuses
on continual improvement throughout all aspects of life, has become increasingly
popular as a philosophy for improving business functions.
c. Over the past few years, many methods for improving both manual and automated
processes have been developed. One of the most popular is Six Sigma, which is a
business management strategy originally developed by Motorola USA in 1981. It seeks
to improve the quality of process outputs by identifying and removing the causes of
defects and minimizing variability in manufacturing and business processes. Other
continuous improvement methodologies include lean manufacturing, total quality
management (TQM), and theory of constraints (ToC). While all of these methodologies
differ in the way in which they are applied and measured, the basic cycle is as follows:
Figure 4-18:
5442.23 Understanding the “macro” concept of operational effectiveness—the ability to perform the
core components of an organization’s business processes in an effective and efficient
manner—aids in understanding the importance at a “micro” level of internal controls and
monitoring.
IT controls and automated monitoring aid an organization in gathering data about the
operation of a given control. Identifying problems in real-time through ongoing direct
monitoring of controls can help an organization in root-cause analysis, which can in turn be
used to improve the processes for greater operational effectiveness.
As part of SOX (Sarbanes-Oxley Act) Section 404, management must assess both the design
and the operational effectiveness of their internal control environment.
Operational effectiveness: As stated in PCAOB Auditing Standard (AS) 5, the auditor should
test the operating effectiveness of a control by determining whether the control is operating
as designed and whether the person performing the control possesses the necessary
authority and competence to perform the control effectively. Procedures the auditor
performs to test operating effectiveness include a mix of inquiry of appropriate personnel,
observation of the company's operations, inspection of relevant documentation, and
reperformance of the control.
Timing of tests of controls: As stated in PCAOB Auditing Standard (AS) 5, testing controls
over a greater period of time provides more evidence of the effectiveness of controls than
testing over a shorter period of time. Further, testing performed closer to the date of
management's assessment provides more evidence than testing performed earlier in the
year. The auditor should balance performing the tests of controls closer to the as-of date
with the need to test controls over a sufficient period of time to obtain sufficient evidence of
operating effectiveness.
Segregation of Duties
5442.24 In a highly integrated system, a person with unrestricted access to the computer, its
programs, and live data may be able to perpetrate and conceal fraud. To prevent this,
organizations must effectively segregate the following information system duties.
5442.25 Systems administrators make sure the information system operates efficiently and
effectively. The database administrator responsibilities generally include planning the
database, defining schemas and subschemas, selecting the appropriate database
management system (DBMS) software, creating the database structure, establishing policy
and procedures for database usage, teaching users how to work with the DBMS, and
controlling database activity, i.e., database design, operation, and security.
5442.27 Security management makes sure all system components are secure and protected from any
and all internal as well as external threats.
5442.28 Change management makes sure all organizational changes are made effectively and
efficiently.
5442.29 Users:
a. authorize and record transactions,
b. use system output, and
c. are responsible for correcting errors.
5442.31 Programmers:
a. use the design developed by the analysts to develop an information system and
b. write computer programs.
5442.33 Information system librarians maintain company databases, files, and programs in a separate
and secure storage area. Librarians only allow changes into production once the change has
been tested in the test environment by someone other than the programmer.
5442.35 Application programs that are on the server should be stored as object code to prevent
computer operators from making changes to the programs during execution. Object code is
created when a program written in the source language (such as COBOL) has been compiled
by a compiler program. Object code (also known as machine language) is the format
required for the central processing units (CPU) to execute. Changes cannot be made directly
to the object code. Source code is the language that programmers use to write the
programs. Some source code is interpreted or executed line-by-line by the computer and can
be edited at any time. Source code should only be used when in a secure development
and/or testing environment. Once the code has been thoroughly tested in the test
environment, the librarian can move the code into the production environment.
1. Systems
Analysis
4. Implementation 3. Physical
and Conversion Design
5450.06 Systems analysis. During systems analysis, information to purchase or develop a system is
gathered.
a. Systems development requests are received and prioritized so limited development
resources are maximized.
b. A system survey is conducted to:
(1) define the nature and scope of the project.
(2) identify project strengths and weaknesses.
c. An in-depth feasibility study is conducted.
d. Information needs of users and managers are identified and documented. This is a
critical part of systems analysis, as these needs are used to develop and document the
systems requirements used to select or develop a new system.
e. A report summarizing systems analysis results is prepared for the steering committee.
5450.07 Conceptual design. During conceptual design, the company performs the following steps to
determine how to meet user needs:
a. Identify and evaluate appropriate design alternatives, such as:
(1) buying software.
(2) developing software in-house.
5450.08 Physical design. The company uses the conceptual design requirements to:
a. design input and output documents.
b. write computer programs.
c. create files and databases.
d. develop policies and procedures.
e. build controls into the new system.
After the new system is designed and created, physical system design results are
communicated to the steering committee.
5450.09 Implementation and conversion. During this phase all physical design components are
assembled and the system is put into place.
a. An implementation and conversion plan is created and implemented.
b. Hardware and software are installed and tested.
c. Employees are hired and existing employees are relocated.
d. Employees are trained.
e. Processing procedures are developed and tested.
f. Standards and controls are established and implemented.
g. System documentation is completed.
h. The organization converts to the new system and discontinues the old system.
i. Fine-tuning adjustments are made.
j. A post-implementation review is conducted to detect and correct any design
deficiencies.
k. The new system is complete when the operational system is delivered to the
organization.
l. The steering committee is sent a final report.
5450.10 Operation and maintenance. The new, operational system is used as needed.
a. The system is periodically reviewed.
b. Modifications are made as problems arise or needs change.
c. At some point the company will need to make major modifications or replace the
system and the software development life cycle (SDLC) begins again.
In addition, three activities (planning, assessing project feasibility, and managing change) are
performed throughout the life cycle.
Maintainability
5450.11 Two categories of controls are used to make sure a system can be adequately maintained:
1. Project development and acquisition controls
2. Change management controls
Conversion Strategies
5450.14 The most popular conversion strategies are the parallel run, phased rollout, “big bang”
strategy, and hybrid approach.
a. Parallel: The new system is run alongside the legacy system; users are trained on the
new system while the old system is still operational.
b. Phased: The cutover and adoption of the new application occur in phases over an
extended period of time; all users are migrated to the new system in a predetermined
series of steps. This can be by module, business unit, or location.
c. “Big bang” (also called direct changeover): The complete implementation occurs in one
single instance; all legacy applications (if any) are cut out and all users move to the new
system on an agreed-upon day.
d. Hybrid: This approach combines the above methods to adapt to the entity’s unique
situation. For example, the most important modules can be implemented using the “big
bang” strategy, and then the less significant modules can be implemented using a
phased rollout.
5450.15 Below are some of the pros and cons of each conversion method:
“Big Bang”:
Pros Cons
Much faster and less expensive than the other Higher risk—small details can be overlooked in
options the rush to convert
Efficient—complete cutover happens on a Less time to learn the full scope of the
single day application; “sink or swim” environment
Quicker return on investment Little, if any, fallback plans available
Training needed only on the new system; not a Difficult to conduct end-to-end testing prior to
changeover the go-live date
Phased Rollout:
Pros Cons
Less risk; employees learn skills with each Project duration for full implementation is long
phase
More time for users to adapt to the new Operating in a continuous state of change can be
system; no “sink or swim”; small details and disruptive and/or costly
issues can be fixed as you go
Implementation team can learn from the initial “Workarounds” may need to be developed for
phases and use this knowledge for subsequent critical information only available in later
phases modules
Reduces the burden on the IT team Team can lose focus and the need for urgency
Parallel Conversion:
Pros Cons
This is the least risky approach; slower pace Most expensive—maintaining two systems
than “big bang” but faster than phased rollout
Users learn the new system while keeping their Executive commitment may waver when they are
legacy environment presented with the level of effort to conduct such
an implementation
There is always a fallback available if the new Can lead to employee frustration because they
system is unacceptable have to do everything twice, creating
inefficiencies, overtime, and data entry issues
5450.16 Factors to consider when choosing an implementation strategy include the following:
a. Size and complexity of organization: Complex organizations may find the “big bang”
implementation approach quite risky, yet if the different modules or sites are highly
interdependent, it may be the only solution.
b. Risk appetite: Since “big bang” implementations have an inherently higher level of risk
than other approaches, they may be too risky for certain companies’ defined risk
appetite.
c. Software architecture: It is much easier to implement by phase for systems which are
tiered or cloud-based.
d. Anticipated benefits: Phased rollout and parallel deployment may be less daunting for
users and more beneficial to employee morale than the “big bang” deployment. Users
stay motivated and convinced that changing the system is for the greater good.
e. Impact on personnel: If resources are already at full capacity, the “big bang” approach
may overly tax employees; using either the phased rollout or parallel procedure will
spread out the workload. In addition, staff will have more time to perform crucial tests,
which are essential to the success of any implementation or conversion.
f. Budgetary and time constraints: A phased approach is less risky, but it requires more
time. Additional delays translate into additional consulting fees, downtime, and support
costs.
5511.01 Financial measures (or financial ratios) are used as tools to identify the performance of an
organization. The following are examples of financial measures:
a. Earnings per share
b. Price/earnings ratio
c. Cash asset ratio
d. Return on assets
e. Capital asset ratio
f. Current ratio
g. Quick ratio
h. Return on investment
i. Return on equity
j. Free cash flow
5511.02 Financial ratios help management evaluate the overall financial condition of an organization.
5511.03 Investors use financial measures and ratios to compare the strengths and weaknesses of
organizations relative to other firms.
Nonfinancial Measures
5511.04 A variety of nonfinancial measures—from the accounting system or from other data
sources—can be used to evaluate performance. Examples of nonfinancial measures include:
a. inventory turnover,
b. labor efficiency,
c. on-time deliveries,
d. schedule attainment,
e. units per hour,
f. throughput time,
g. measures of customer satisfaction (e.g., from such indicators as surveys, repairs,
complaints, or number of service calls), and
h. measures of environmental compliance.
5512.02 Congruent decision making: At the division level, when one performance measure (such as
return on investment) is not relied upon solely or given a high emphasis, managers are free
to make decisions that are more in tune with the long-term profitability growth objectives of
the company as a whole.
5512.03 Return on investment (ROI) focuses on optimal use of invested capital. Net income from the
income statement is divided by invested capital from the balance sheet as follows:
Net income
ROI =
Invested capital
a. Return on investment can be divided into two elements:
Net income
Profit margin =
Sales
Sales
Capital employed turnover rate =
Invested capital
5512.04 Residual income is the amount of net income in excess of a minimum desired rate of return
on invested capital. The minimum desired net income is first computed by multiplying the
desired rate of return by invested capital. This amount is then subtracted from reported net
income. A positive residual income is viewed favorably because it indicates earnings in
excess of the required minimum.
Example:
Net income $150
Return on investment at minimum rate of return ($1,000 at 0.12) 120
Residual income $ 30
5512.05 Economic value added (EVA) measures surplus or excess value created by an enterprise’s
investments. EVA is calculated as return on capital minus cost of capital multiplied by
invested capital. It is similar to net present value and, specifically, to residual income.
Shortcomings of EVA include the focus on past rather than future performance, and its
complexity restricts its use to larger businesses. EVA can be depicted in an equation as
follows: EVA = (Return on capital – Cost of capital) × Capital investment in project.
5512.06 By integrating total quality control (TQC), total quality management (TQM), and just-in-time
(JIT), several benefits are achieved: (a) fewer defective units are produced before the
problem is identified, (b) the process causing the problem can be fixed before more bad
units are produced, (c) immediate attention gives rise to easier problem identification and
solution, and (d) higher-quality products improve customer satisfaction.
TQC and JIT are often combined so that the production process is stopped whenever
something goes wrong. Because production schedules are so tightly synchronized, all
employees must be involved to identify when a quality control problem occurs. Production is
halted until the problem is identified and corrected.
a. Total quality control (TQC) involves all employees in the quality control effort. It is the
unending quest for perfect quality—the striving for a defect-free manufacturing
process. Emphasis shifts from an acceptable level of quality to total quality control with
zero defects as the objective.
b. Just-in-time (JIT) is a manufacturing philosophy that promotes the simplest, most
efficient means of production. Under ideal conditions, the company would receive raw
materials just in time to go into production, manufacture parts just in time to be
assembled into products, and complete products just in time to be shipped to
customers. JIT shifts the production philosophy from a push approach to a pull
approach.
c. Total quality management (TQM) is the application of quality principles to all of the
organization’s endeavors to satisfy customers. It is a method of managing the
organization to excel on all dimensions. The customer defines quality. Therefore,
product quality performance measures are likely to include product reliability and
service delivery in addition to product cost.
Cost of Quality
5512.07 Every effort should be made to produce a quality product and to produce it right the first
time. Cost of quality (COQ) is an area of cost analysis that measures and monitors quality
costs in four areas:
1. Prevention: Costs of preventing bad quality (such as worker training and quality circles)
2. Appraisal: Costs of monitoring the level of bad quality (such as product testing and
scrap reporting systems)
3. Internal failure: Costs of fixing bad quality (such as rework labor)
4. External failure: Costs of bad quality that are not discovered before shipment (such as
warranty claims or customer ill-will)
The objective is to minimize the total cost of quality. Total cost of quality is the sum of
prevention and appraisal costs, which increase with quality level, and internal and external
failure costs, which decrease with quality level. This is illustrated as follows:
Continuous Improvement
5512.08 Continuous improvement (sometimes called target-based costing) is a concept that seeks to
reduce costs while maintaining quality. Cost control using traditional engineering standards
maintains the status quo. Continuous improvement seeks to reduce costs that frequently
disturb the established order. One way to achieve this is by using market-driven standards
rather than engineering-driven standards. Engineering-driven standards are based on
existing technology. The engineering-driven standard plus a desired markup equals the
desired market price. Market-driven standards work in reverse order. A competitive market
price that yields the desired market share less the desired markup equals allowable cost. The
allowable cost becomes the target standard. Continuous improvement is sought until the
target standard is achieved. This changes the focus from optimizing within existing
technology to modifying technology until the desired standard is achieved. Thus, actual cost
from a prior period becomes a starting point for further tightening.
5512.09 Life cycle costing generally refers to the summation of a purchaser’s costs from the point of
initial purchase to the point of ultimate disposition of the product by the purchaser several
years in the future. This concept deals explicitly with the relationship between what the
customer pays for a product and the total cost the customer incurs over the product’s entire
life. This type of information can be very valuable in the following areas:
a. Better product design to reduce post-purchase costs to the customer
b. Improved market segmentation and product positioning
Balanced Scorecard
5512.10 The balanced scorecard is a strategic planning and management tool, used extensively in
business and industry, government, and nonprofit organizations to align business activities
with the vision and strategy of the organization, improve internal and external
communications, and monitor organizational performance against strategic goals.
5512.11 The balanced scorecard views an organization from four perspectives; metrics are developed
and data is collected and analyzed for each of the following four perspectives:
1. Learning and growth: This perspective forms the foundation of all knowledge-worker
organizations and includes employee training and corporate cultural attitudes related to
individual and corporate self-improvement. Metrics can be used to guide managers in
focusing training funds where they can be the most effective.
2. Business (internal) processes: Metrics based on this perspective provide managers with
feedback regarding how well their business is operating, and whether its products and
services meet customer requirements (the mission).
3. Customer (satisfaction) perspective: If customers are not satisfied, they will eventually
find another supplier(s) that will meet their needs. Poor performance from this
perspective can be a leading indicator of future troubles, even though the current
financial picture may seem strong. In developing metrics for this perspective, customers
are analyzed in terms of “kinds of customers” and “kinds of processes” for which the
entity provides a product or service to those customer groups.
4. Financial: Even under a scorecard system, there remains a need for traditional financial
data, provided that any undue emphasis on financial-related data does not “unbalance"
the other perspectives.
Benchmarking
5512.13 The objectives of benchmarking are to analyze how other organizations achieve their high
performance levels, determine what improvements are needed for the given entity, and
implement this information to improve performance.
Incentive Compensation
5512.14 Incentive compensation programs should promote and maintain an inspired and productive
work environment. An incentive compensation program should be flexible and be developed
alongside the organization’s changing needs.
a. Incentive compensation should only reward employees for a job well done. If an
employee gets a bonus for simply being at work, he or she might feel entitled to the
compensation. The compensation should be used as a motivational tool.
b. An incentive program that is organized with the organization in mind will help steer
employees to the organization’s goals as well. These programs should be incentives to
push employees to the organization’s objectives.
c. Many employees are fearful of change and do not adapt to change very easily. An
incentive program should be designed to provide incentive for employees to become
more interested in changes and to offer further suggestions for change.
d. The program should take into consideration the impact of the employee on the
organization. For example, a controller may be rewarded on average aged accounts
receivable, while a salesman might be awarded on sales.
5521.01 Cost allocation is the process of assigning indirect costs (also called overhead, common
costs, or joint costs) to cost objects. The focus in financial accounting is primarily on the
allocation of factory overhead to inventories and cost of goods sold. However, indirect costs
may also be allocated to help management plan and control operations. It is a controversial
topic because, by its very nature, allocation is an arbitrary process.
5521.04 Variable costing is sometimes called direct or marginal costing. Absorption costing is
sometimes referred to as full product costing. The difference between variable and
absorption costing is the way fixed overhead costs are handled. Under variable costing, they
are charged to expense as a period cost in the period incurred. No fixed overhead costs are
allocated to the units produced. Absorption costing treats fixed costs as a product cost and
allocates them to the units produced. Fixed costs follow the units through work-in-process
and finished goods inventory, and are expensed through cost of goods sold when the units
are sold.
5521.05 The following summarizes the major arguments in favor of each method.
a. Variable costing:
(1) Variable costing is felt to be more useful for management decision making because
it separates fixed costs that do not change with volume. The impact of management
decisions that change volume is more evident.
(2) Fixed factory overhead is more closely related to the capacity to produce than to
the production of specific units. Since fixed costs would be incurred regardless of
production, and since they relate to the capacity to produce for a period of time,
they should be reported as a period cost.
b. Absorption costing:
(1) Fixed factory overhead is a necessary cost of production and units produced should
bear all such costs. It is consistent with the concept that the cost of an asset should
include all necessary costs to get it into proper form, location, and working
condition.
(2) Absorption costing is required by both generally accepted accounting principles and
the Internal Revenue Service (IRS).
5521.06 Over the long run, the difference between variable costing profit and absorption costing
profit will be zero. The difference is not in the amount of expense but in the timing of
expense. In the short run, profit will be different when production and sales are not the
same. Some fixed overhead will be inventoried or taken from inventory under absorption
costing, which would not be done under variable costing.
5521.07 Fixed overhead must be accounted for differently when computing product costs, cost of
goods sold, or net income under the two methods. Fixed overhead is a product cost for
absorption costing and a period cost for variable costing.
5521.08 Illustration: Variable and absorption costing (adapted from the AICPA).
Problem: JV Company began its operations on January 1, 20X1, and produces a single
product that sells for $7.00 per unit. Standard capacity is 100,000 units per year, and
100,000 units were produced and 80,000 units were sold in 20X1.
Manufacturing costs and selling and administrative expenses were as follows:
Fixed Costs Variable Costs
Raw materials — $1.50 per unit produced
Direct labor — 1.00 per unit produced
Factory overhead $150,000 0.50 per unit produced
Selling and administrative 80,000 0.50 per unit sold
There were no variances from the standard variable costs. Any over- or underapplied
overhead is written off directly at year-end as an adjustment to cost of goods sold.
Required:
a. Compute the unit cost under absorption costing for presenting inventory on the balance
sheet at December 31, 20X1.
b. Compute the net income for 20X1 under variable costing.
Solution:
Part a. Absorption costing values units in ending inventory at full cost of production,
including fixed factory overhead.
Production cost/unit:
Raw materials $ 1.50
Direct labor 1.00
Overhead: Variable 0.50
Fixed 1.50a
Total $ 4.50
a
Cost per unit at standard capacity ($150,000 ÷ 100,000 units)
Part b. Variable costing charges fixed factory overhead to expense as incurred. Income is
computed as follows:
Sales (80,000 × $7.00) $560,000
Less variable costs:
Raw materials (80,000 × $1.50) $120,000
Direct labor (80,000 × $1.00) 80,000
Overhead (80,000 × $0.50) 40,000
Selling (80,000 × $0.50) 40,000 280,000
Contribution margin $280,000
Less fixed costs:
Overhead $150,000
Selling 80,000 230,000
Net income $ 50,000
5521.09 Management accounting is the process of measuring and using information for the following
purposes:
a. Management planning, decision making, and control
b. Reporting assets and cost of goods sold for financial statements and income taxes
c. Reporting of costs under cost-based contracts
Because management accounting information is used for many purposes, the accounting
systems need to be sufficiently flexible to allow measuring and combining costs in different
ways.
5521.10 Assignment of costs to cost objects: Cost assignment is the process of tracing direct costs
and allocating indirect/common costs to cost objects. A cost object (or cost objective) is an
item or activity for which a separate and discrete measurement of costs is desired, such as a
product, function, department, division, contract, or service.
5521.11 Costs for external reporting is the financial accounting purpose of reporting the value of
inventory so that the cost of goods sold and income can be determined. Costs are classified
based on their attachability to inventory.
a. Matching principle: Information for external reporting is classified based on
attachability to inventory and follows GAAP. The basic theory is the matching principle:
“Let the expense follow the revenue.” The three methods of recognition are as follows:
(1) Associating cause and effect: Some costs are recognized as expenses on the basis
of a presumed direct association with specific revenue. Examples of expenses that
are recognized by associating cause and effect are sales commissions and costs of
products sold or services provided.
(2) Systematic and rational allocation: In the absence of a direct means of associating
cause and effect, some costs are associated with specific accounting periods as
expenses on the basis of an attempt to allocate costs in a systematic and rational
manner among the periods in which benefits are provided. Examples of items that
are recognized in a systematic and rational manner are depreciation of fixed assets,
amortization of intangible assets, and allocation of rent and insurance.
(3) Immediate recognition: Some costs are associated with the current accounting
period as expenses because:
(a) costs incurred during the period provide no discernible future benefits.
(b) costs recorded as assets in prior periods no longer provide discernible benefits.
(c) allocating costs either on the basis of association with revenue or among
accounting periods is considered to serve no useful purpose. Examples include
officers’ salaries, most selling costs, amounts paid to settle lawsuits, and cost of
resources used in unsuccessful efforts.
b. Product costs: Direct costs can be attached to inventory based on associating cause and
effect. Indirect costs are attached by systematic and rational allocation.
c. Period costs: These costs are not attached to inventory, but are recognized
immediately.
5521.12 Manufacturing product costs are classified into the following categories:
a. Direct materials: Raw material costs that can be reasonably traced to a product on a
per-unit basis.
b. Direct labor: Wages relating to employees who work directly on the product (i.e.,
hands-on) that can be reasonably traced to a product on a per-unit basis.
c. Indirect materials: These are costs incurred in the manufacturing process, but either
they do not become part of the product (e.g., a machine lubricant) or they are not
reasonably traceable to the product (e.g., paint, varnish, glue). Such costs become part
of factory overhead and are attached to the product through an allocation process.
d. Indirect labor: This incorporates factory labor costs for personnel who do not work
directly on the product and whose costs cannot be reasonably traced on a per-unit
basis. This category includes salaries of peripheral production employees, notably
production supervisors/managers, factory janitors, and factory security personnel. Like
indirect materials, these costs become part of factory overhead and are attached to the
product through an allocation process.
e. Factory overhead: This is also called manufacturing overhead, factory burden, or simply
overhead. It includes all costs except direct labor and direct materials that are part of
making the product, but which are not directly traced to a specific product. Costs are
accumulated in factory overhead accounts. They may either vary with production
(variable costs) or not (fixed costs). The costs are traditionally allocated to each unit
produced on some basis such as $ per direct labor hour, per direct labor cost, or per
machine hour.
f. Prime costs: Direct materials and direct labor together are called prime costs.
g. Conversion costs: All factory costs other than direct materials are called conversion
costs; they are required to convert materials into a finished product. Conversion costs
include direct labor and factory overhead.
Raw Materials
Salaries & Wages Payable Work in Process Finished Goods Cost of Goods Sold
Labor Costs
Incurred
Manufacturing Overhead
5521.15 Schedule of cost of goods manufactured: This schedule summarizes the costs assigned to
the manufacturing process during the period. A condensed example follows:
Schedule of Cost of Goods Manufactured
Beginning direct materials inventory $ 10
+ Purchases of direct materials 80
– Ending direct materials inventory (30)
Raw materials used 60
+ Direct labor 70
+ Overhead (supplies, utilities, indirect labor, etc.) 40
Total manufacturing costs for the period 170
+ Beginning WIP inventory 50
Total manufacturing costs 220
– Ending WIP inventory (140)
Cost of goods manufactured $ 80
5521.16 A determination of unit cost is needed for product pricing, cost control, GAAP reporting
(inventory valuation, cost of goods sold), and income tax reporting. There are several
methods of attaching production costs to each unit produced. Within each method, inputs
can be valued using actual, normal, or standard costs.
The choice of method is influenced by the production process. Production processes range
from making products individually or in identifiable batches, to making a continuous stream
of indistinguishable goods. Therefore, job costing can be viewed as existing at one end and
process costing at the other end of a cost-tracking continuum. There are hybrid methods in
between that have been known by various names, such as operation costing.
5521.17 Job costing, also called job order costing, is a tracking system used for customized products
or services.
a. Job cost records (or sheets): Costs are accumulated by the job, or job lot, and are
recorded on job cost records as work progresses. Entries are made into the WIP ledger
account for manufacturing inputs from the job cost records. The job cost records make
up a subsidiary ledger that supports the control account Work in Process.
b. Direct materials: Most raw materials will be traceable to the units produced through
the requisition forms used to issue materials from the raw materials inventory
storeroom to the production department. Indirect materials are charged to
manufacturing overhead and are attached to the product through the allocation
process.
c. Direct labor: This cost is recorded by the employee on a detailed time record listing the
job number, operation performed, hours used, etc. Indirect labor is charged to
manufacturing overhead and attached to the product through the allocation process.
d. Overhead: This is applied to jobs using the predetermined rate applied on the selected
allocation base or cost driver. If this is a normal costing system, actual driver units
experienced are used as the input quantity. If a standard costing system, the number of
driver units that should have been used to complete the related activity is used.
e. Job completion and sale of goods: As jobs are completed and moved to the Finished
Goods warehouse, entries are made to transfer the completed cost from WIP to
Finished Goods and the job cost records become a subsidiary ledger for the Finished
Goods account. Ultimately, as sales occur, the job cost records become supporting
documents for Cost of Goods Sold.
5521.18 Process costing is used for products whose individual units are indistinguishable from other
units (e.g., sugar, corn flakes, textiles, paint). Such products are continuously produced
through a series of uniform production steps called processes. Costs are accumulated by
process, then applied to the units processed using an average cost for input units. There is a
separate Work-in-Process account for each department or process through which the units
flow. At the end of each period, costs are allocated between the units completed and
transferred to the next department and the unfinished units remaining in ending inventory
based on equivalent finished units.
5521.19 Computation of unit costs and assigning costs to inventories: To compute costs relating to
completed units transferred out of the department and costs of unfinished units that remain
in ending inventory, both physical units and total costs must be accounted for in a five-step
process. This process comprises a process costing report. The steps are listed below. Note
that Steps 2 and 3 are interchangeable.
Step 1: Account for all physical units.
Step 2: Compute the equivalent finished units (EFU or EU) for each element of
manufacturing cost that is incurred at a separate rate. Usually two computations are
needed: one for materials units and one for conversion units.
Step 3: Determine the total dollars to be accounted for.
Step 4: Compute the per-unit cost for each separately identified manufacturing element.
This will usually be materials and conversion.
Step 5: Use the per-unit cost to assign the dollars to the units transferred out and to the
units remaining in the ending inventory. The sum of the dollars transferred out to the next
department or to finished goods and the dollars remaining in ending WIP must be equal to
the total dollars identified in Step 3.
(1) Weighted average: In this method, no attempt is made to separate the work done
to complete the beginning inventory units and costs from the work performed on
units started and finished. It is as though the beginning inventory data were “pulled
into” the current period and a simplifying assumption is made that all units
completed during the current period were started during the current period. Thus,
all the costs incurred, including the beginning inventory costs, are averaged over all
the equivalent units. This method is appropriate when costs and production remain
relatively stable from period to period. It is not appropriate in a standard costing
system.
All costs OR Beginning inventory costs + Started costs
All E. units All units completed + (EI × Completion %)
(2) First in, first out (FIFO): In this method, the beginning inventory units and costs are
kept separate from the current period units and costs. The units transferred out are
made up of two “layers”—the beginning inventory units finished up during the
period plus units started and finished. The ending inventory is valued using the
most recent costs. This method is preferred when costs fluctuate from period to
period and is required if a standard cost system is in use.
Current-period costs OR Total costs − Beg. inventory costs
Current-period E. units Completed units + (EI × %) − (BI × %)
Note that when there is no beginning inventory to finish up during the period, the
equivalent units will be the same for the weighted-average and FIFO methods.
5521.21 Materials costs: Materials are added to the manufacturing process at different points in
time. For example, Material A could be added as processing begins in the department and
Material B added at the 60% conversion phase. In CPA Examination problems, materials are
generally added at either the beginning of the process or at the end; therefore, units will be
either 0% or 100% complete as to materials costs.
5521.22 Conversion costs: It is assumed that conversion occurs uniformly during the process;
therefore, the CPA Examination question fact pattern will indicate the percentage of
completion as to these costs for both the beginning and ending inventories. Note that
conversion includes both labor and overhead costs.
b. FIFO method:
Materials Conversion
Completed Units Transferred Out X X
Add Ending Inventory:
Materials - % added during period X
Conversion - % added during period X
Less Beginning Inventory:
Materials - % completed at beginning of period (X) X
Conversion - % completed at beginning of period (X)
Equivalent Finished Units, FIFO (EFU/FIFO) X X
Step 4: Compute per-unit cost for materials and conversion. (Where EFU is the same for
both materials and conversion, only one computation needs to be done.)
a. Weighted-average method:
Total materials costs Total conversion costs
EFU/WA − Materials EFU/WA − Conversion
b. FIFO method:
Current-period materials costs Current-period conversion costs
EFU/FIFO − Materials EFU/FIFO − Conversion
5521.24 Process costing report example: Levittown Company employs a process cost system for its
manufacturing operations. All direct materials are added at the beginning of the process and
conversion costs are added uniformly. Levittown’s production quantity schedule and costs
for November are reproduced below for the first stage of processing:
Beginning WIP (60% complete as to conversion costs) 1,000
Units started during period 5,000
Ending WIP (20% complete as to conversion costs) 2,000
A process cost report for Levittown Company for this process stage would require the
following steps.
Step 4: Compute unit costs based on EFU denominators computed in Step 3. (Use total
costs.)
Materials $7,500 ÷ 6,000 units = $ 1.25 per unit
Conversion $44,000 ÷ 4,400 units = 10.00 per unit
Total unit cost (weighted average) $11.25
Step 5: Use unit costs computed in Step 4 to allocate costs between units completed and
transferred out and units in ending inventory.
Units Transferred Out:
(4,000 × $11.25) $45,000
Ending WIP Inventory
Materials (2,000 × $1.25) $2,500
Conversion (2,000 × 20% × $10.00) 4,000
Total Costs in Ending WIP 6,500
Total Costs Accounted For $51,500
Step 3: Determine total costs to be accounted for. (Note that this also is the same under
both weighted-average and FIFO methods.)
Materials Conversion Total
Beginning inventory costs $2,000 $ 6,000 $ 8,000
Period costs 5,500 38,000 43,500
Total costs to account for $7,500 $44,000 $51,500
Step 4: Compute unit costs based on EFU denominators computed in Step 3. (Use current-
period costs.)
Materials $5,500 ÷ 5,000 units = $ 1.10 per unit
Conversion $38,000 ÷ 3,800 units = 10.00 per unit
Total unit cost (FIFO) $11.10
Step 5: Use unit costs computed in Step 4 to allocate costs between units completed and
transferred out and units in ending inventory.
Units Transferred Out:
1. From Beginning Inventory
Beginning Costs $ 8,000
Costs to Complete Beginning Inventory
Materials added during period -0-
Conversion added to complete during period
(1,000 units × 40%* added × $10.00) 4,000
2. Started and Completed (3,000 × $11.10) 33,300
Total Costs Transferred Out (4,000 units) $45,300
Ending WIP
Materials (2,000 × $1.10) $ 2,200
Conversion (2,000 × 20% × $10.00) 4,000
Total Costs of 2,000 units in WIP 6,200
Total Costs Accounted For $51,500
* Beginning inventory was 60% complete as of the beginning of the month and all
materials had been added in the previous period at the beginning of the process.
Therefore, 40% conversion costs were added during the period and no additional
materials costs were incurred for these units.
5521.27 Spoilage consists of units that are lost, broken, or otherwise defective and cannot be sold
with regular good units.
a. Normal spoilage: A certain amount of spoilage is expected as part of the manufacturing
process. The amount expected spoilage is often expressed as a percent relative to the
good units produced. It is considered to be a cost of producing the good units and is
classified as a product cost.
b. Abnormal spoilage: This is spoilage in excess of the expected level. It is treated as a
period cost. Cost relating to abnormal spoilage units is transferred out of the work-in-
process (WIP) account into a loss account.
c. Treatment in EFU calculation: Abnormal spoilage units are generally always included in
the computation of equivalent units, because their cost must be calculated and
recorded as a separate loss. However, the inclusion of normal spoiled units depends on
whether a company wishes to separately identify and report the extent to which normal
spoilage costs impact the cost of good units. There are, therefore, two treatments for
normal spoilage:
(1) Include the units and add their cost to only the cost of good units passing the
inspection point (separate identification of costs), or
(2) Exclude the units, thereby increasing the manufacturing cost per equivalent unit,
and spreading normal spoilage cost over not only good units completed, but also
over abnormal spoilage units and equivalent units in ending inventory (hidden cost
method).
5521.28 Separately reported normal spoilage costs: Spoiled units are identified at an inspection
point. The inspection point usually occurs upon completion of departmental processing.
Therefore, normal spoilage costs should be identified and added to the costs of good units
completed. No normal spoilage costs should be attached to the incomplete units in ending
inventory. To accomplish this, all spoilage units are included in the equivalent unit
computation to arrive at a manufacturing cost per equivalent unit. The cost of normal
spoilage is added to the cost of good units produced. The cost of abnormal spoilage is
removed from work in process and is charged as an expense of the period.
Note: (1) The normal spoilage costs are separately identified so that (2) the impact of
those costs is seen on the unit cost of good units transferred out ($47,520 total costs ÷
9,000 good units transferred = $5.28, as opposed to $4.80/unit overall manufacturing
cost), and (3) no normal spoilage cost is included in ending inventory. Any spoilage in
those 4,000 units will be identified upon their completion and allocated to those units at
that time.
5521.30 Hidden normal spoilage costs: (Use this method when no information is given about
spoilage inspection point or when not told to attach cost to only good units produced.)
Some companies choose not to separately identify normal spoilage costs. This is appropriate
when such costs are relatively small. In this procedure, abnormal spoilage is included in the
equivalent unit computation, but normal spoilage is excluded. Thus, the per-unit
manufacturing cost is proportionately higher. The normal spoilage cost is absorbed into the
cost of all remaining equivalent units (including abnormal spoilage and ending inventory).
5521.31 Example of hidden normal spoilage costs: Assume the same facts for Hamilton Company as
described previously, but that normal spoilage costs are not separately identified.
1. Computation of spoilage units would be the same as with separately reported normal
spoilage costs.
Total units completed 10,000
Less total spoilage 1,000
Good units completed 9,000
Total spoilage 1,000
Less normal spoilage (9,000 good × 10%) 900
Abnormal spoilage 100
Note: (1) The normal spoilage costs are not separately identified; rather, (2) they are hidden
in the unit manufacturing cost of $5.18, and (3) both the abnormal spoilage units and ending
inventory include some normal spoilage costs. For this reason, some authorities believe this
method is not theoretically correct and do not advocate its use.
5521.32 Spoilage in job costing: The net cost (cost less any disposal value) related to normal spoilage
that is not attributable to a particular job is charged to factory overhead. It is thus allocated
to all jobs worked on through the normal overhead application procedure. The net cost of
spoilage attributable to the characteristics of a particular job or job lot is charged to that
particular job. The cost of abnormal spoilage not attributable to a particular job is treated as
a period cost and charged to a special loss account.
5521.33 Scrap: Normally, no cost is assigned to scrap. Its disposal value is most commonly credited to
factory overhead. Where feasible, it may be used to reduce the cost of a particular job or
type of job.
5521.34 Rework costs: Costs of reworking defective units are generally charged to factory overhead
and are spread among all units produced through the overhead application process. Special
circumstances may cause such costs to be charged to a particular job.
Costing Methods
5521.35 Costing methods: Inputs into the production process can be accounted for by three different
costing methods: actual, normal, and standard costing.
Input Actual Costing Normal Costing Standard Costing
Direct Materials Actual Actual Standard Input at Budgeted Rate
Direct Labor Actual Actual Standard Input at Budgeted Rate
Factory Overhead Actual Actual Input at Standard Input at Budgeted Rate
Budgeted Rate
5521.36 Actual costing: The inputs to production are valued at their actual cost, the actual number
of dollars paid for direct materials, direct labor, and overhead. All factory overhead costs
incurred (debit to Factory Overhead Control) would be applied to Work-in-Process (credit
Factory Overhead Control). The problem is that the total actual costs of overhead are not
known until the end of the period. It may not be practical to wait until the end of the year to
cost products, particularly if cost is a factor in setting prices. If monthly overhead costs are
used, unit costs could vary widely.
5521.37 Normal costing: The inputs of direct materials and direct labor are valued at their actual
cost, but overhead is applied on a budgeted or “normalized” basis. Estimated overhead costs
are divided by estimated production for the year to determine an application rate. If
overhead is applied on a direct labor hour basis, then the overhead rate would be applied to
actual direct labor hours.
5521.38 Standard costing: In a standard costing system, standard costs are developed for all inputs
per unit of production. For example, if overhead is applied on direct labor hours, it would be
applied according to the number of hours that should have been used to make that unit, not
the actual number of direct labor hours used. Standard costs are designed to control the two
components of the cost of producing something: the price paid and the quantity used. Price
and quantity standards are set for materials, labor, and overhead. Standard cost is a unit
concept. It represents the budget to produce one unit. Standard costs are often used to
develop a master budget and flexible budget, and variances from standard costs are then
analyzed.
a. Ideal standards: These are standards that can be attained only under perfect conditions.
There is no allowance for machinery to break down, or to wait for materials, or for other
delays that can be expected in a normal world. Ideal standards require at all times a
level of effort and skill that would be expected from the very best employees at their
peak performance. Ideal standards are generally not used in practice, because they
discourage the efforts of ordinary employees who know they cannot make the standard,
so “why bother?” Variances are often viewed as useless because part of the variance is
“normal.” Expected differences from the standard obscure the actual variances that
need to be investigated. However, advocates of TQM (total quality management) and
continuous improvement are often in favor of the use of ideal standards because such
standards can encourage greater efficiency over time.
b. Attainable standards: Practical standards are tight, but attainable. They can be achieved
by normal workers working with reasonable, efficient effort, with a reasonable
allowance for breakdowns, etc. Variances provide useful information because they
represent abnormal conditions that need to be investigated. Attainable standards are
also useful in the planning and budget process when calculating the production budget,
etc.
Applying Overhead
5521.39 Applying overhead: In both normal and standard costing systems, a process must be
developed for calculating and applying a budgeted overhead allocation rate.
5521.40 Overhead application/allocation rate: Overhead is allocated to cost objects (e.g., units or
jobs) using an application rate (also called an allocation rate), which may be calculated in
several different ways:
a. Actual overhead application rate: This rate is calculated by dividing the overhead costs
by the actual volume of the allocation base. An actual rate is useful when the goal is to
measure actual costs; however, actual data may not be available on a timely basis.
b. Budgeted (or predetermined) overhead application rate: This rate is calculated by
dividing estimated (or budgeted) overhead cost by the estimated (or budgeted) volume
of the allocation base. A standard overhead application rate is an example of a
predetermined rate.
1. Calculate the overhead application rate using (a) material cost and (b) machine hours as
the allocation base.
2. Calculate the overhead charged to Job A if 900 hours of machine time and $20,000 of
materials were used, assuming that the allocation base is (a) material cost and (b)
machine hours.
Solutions:
1. Calculation of application rate:
a. Overhead application rate based on material cost (overhead cost divided by
material cost):
$195,000 = $0.50 per dollar of material cost (or 50% of material cost)
$390,000
b. Overhead rate based on machine hours (overhead cost divided by machine hours):
$195,000 = $10 per machine hour
19,500 hours
The new accounting standard is likely to increase the number of companies that account
for fixed and variable overhead separately. In addition, more companies are likely to use
normal capacity as the budgeted volume when establishing predetermined overhead
rates.
b. Adjustment for fixed overhead: In general, over/underapplied fixed overhead should be
assigned to cost of goods sold. However, any material amounts of over/underapplied
fixed overhead caused by an inaccurate cost budget should be prorated among ending
units in work in process, finished goods, and cost of goods sold.
c. Adjustment for variable overhead: By definition, variable overhead is expected to vary
with the volume of production. This means that over/underapplied variable overhead is
caused by error in the budgeting process. Accordingly, any material amounts of
over/underapplied variable overhead should be prorated among ending units in work in
process, finished goods, and cost of goods sold. If the over/underapplied amount is not
material, it may be adjusted to cost of goods sold alone.
5521.44 Example of disposition of under/overapplied overhead: Assume that XYZ Company incurred
$180,000 of manufacturing expenses and that actual machine hours were 17,000. Before
any adjustment is made, the amounts of fixed overhead applied units in work-in-process,
finished goods, and cost of goods sold are $17,000, $34,000, and $119,000, respectively.
1. Calculate the actual fixed overhead rate based on (a) actual machine hours and (b)
normal capacity.
2. Calculate the amount of over/underapplied fixed overhead, then calculate the amount
attributable to (a) inaccurate cost budget and (b) difference between actual and normal
capacity.
3. What adjustments should be made if the fixed overhead spending variance is
considered immaterial?
4. What adjustments should be made if the fixed overhead spending variance is
considered material?
Solutions:
1a. Actual fixed overhead rate based on actual machine hours:
$180,000 = $10.59 per machine hour
17,000
Note: The rate of $9.2308 should be used to assign fixed overhead to inventory.
Journal entry:
Cost of goods sold 13,923
Work in process 1,308
Finished goods 2,615
Overhead control 10,000
4. If the $13,077 amount is considered immaterial, then it may be netted against the
$23,077 amount and assigned directly to COGS:
Journal entry:
Cost of goods sold ($23,077 – $13,077) 10,000
Overhead control 10,000
Activity-Based Management
5521.46 ABC/ABM cross: Activity-based costing (ABC) can be viewed from two perspectives, each of
which has activities as the central focus. The differing perspectives are depicted in the
following ABC/ABM cross diagram:
Resources
Products
Viewed vertically, the ABC (activity-based costing) cross depicts a cost accounting allocation
tool. Products result from activities; those activities consume resources. Resource costs are
allocated to cost objects on the basis of the relative amount of the particular activities
consumed to produce the particular product.
Viewed horizontally, however, the cross shows a management process (activity-based
management). This process is a series of activities measurable by their related drivers.
Examination of the relationship between these drivers and activities is useful to
management for performance evaluation, planning and control, and decision making. Thus,
implementation of activity-based costing methods provides information that can lead to
better management decisions, i.e., activity-based management (ABM).
5521.47 Steps in ABM: No consensus exists about the exact steps in ABM (activity-based
management), but it typically involves first applying activity-based costing (ABC). Activities,
costs, and drivers are identified, and costs are assigned to activities. However, the focus of
ABM is on the processes, not the costs. ABM uses cost information to understand, manage,
and improve processes.
5521.49 Common ABM applications: The following are common uses for ABM (activity-based
management):
a. Customer profitability analysis
b. Strategic planning
c. Life cycle management
d. Performance measurement
e. Value analysis
f. Budgeting
g. Product costing
h. Measure and monitor capacity utilization
i. Analyze costs of quality
5521.50 Activity-based costing (ABC) addresses several of the concerns about traditional overhead
costing systems. Instead of using a small number of overhead cost pools and arbitrary
allocation bases, overhead costs are accumulated in cost pools related to separately
identified activities that are allocated based on cost drivers.
5521.51 Activities and cost hierarchy: Activities are the separately identifiable tasks required to
produce a product or provide a service. They can be referred to as the work to be done. In
an ABC system, activities are grouped into cost pools, which could be a machine, a group of
machines, a department, or a function.
The following hierarchy of activity levels is useful in helping accountants and managers
identify activities for their organization:
a. Unit level: These activities are done a certain number of times per unit of output, and
are thus volume-related. For example, assembly of parts into a subassembly unit is a
unit-level activity.
b. Batch level: These activities are performed once for each batch, or group, of units. An
example is machine setup that has to be done for each production run no matter how
many units are in a batch.
c. Customer level: These activities support specific customers or groups of customers.
They cannot be specifically correlated to a particular product, batch, or unit. Examples
include sales calls, service and support, and advertising aimed at particular customers.
d. Product level: These activities support specific product types or services. They cannot be
specifically correlated to a particular unit or batch on a cause-and-effect basis. They are
incurred regardless of the number of units produced, and include such items as product
advertising, research and development, engineering specifications, and design costs.
e. Plant (facility) level: These activities and resources are not traceable to specific
products, customers, or services, but they support an individual plant or facility.
Examples include building (facility) maintenance, security, and plant manager’s salary.
f. Organization level: These activities are not traceable to specific products or services or
to specific facilities; they support the organization as a whole. Examples include
headquarters building maintenance and depreciation, CEO and CFO compensation, and
centralized accounting functions.
5521.52 Cost driver: One of the major differences between activity-based costing (ABC) and
traditional costing systems is that greater effort is made to identify cost allocation bases that
are cost drivers. For example, in an assembly activity, the more parts that have to be put
together, the more the cost of that activity increases. Therefore, the cost driver for an
assembly activity would be the number of parts. Increased activity of the cost driver causes
the total costs related to that activity to increase.
5521.53 Steps in activity-based costing: Activity-based costing applies the accumulated costs of
identified activities to cost objects using a separate driver for each activity, using the
following series of steps:
a. Identify causal relationship: Identify the causal relationship between cost incurrence
and activities.
b. Determine driver: Determine the underlying driver of overhead costs for activities.
c. Establish cost pools: Establish homogeneous cost pools related to individual drivers.
d. Determine cost ratio: Develop a predetermined cost ratio for each activity cost pool.
e. Apply costs: Apply the accumulated costs (in the cost pools) to the cost objects on the
basis of resources consumed (as determined by the amount of cost drivers used). Costs
may be assigned to any type of cost object depending on the information management
wants. For example, costs may be assigned to units of product, groups of customers, or
product lines.
5521.54 Example of activity-based costing: Glide-Free, Inc., manufactures two models of patio
swings: Deluxe (upholstered with a woven-vinyl material) and Standard (with stained-wood
finish). The following data relate to the company’s production activities and costs:
Budgeted
Manufacturing Activity Costs Cost Driver Used Cost per Unit of
Activity for Year as Allocation Base Allocation Base
Materials handling $ 300,000 Number of parts $ 0.40
Cutting 1,500,000 Number of parts 2.00
Assembly 4,500,000 Number of parts 6.00
Upholstery 1,200,000 Direct labor hours 20.00
Wood refinishing 3,000,000 Direct labor hours 15.00
1. Compute the total manufacturing costs and manufacturing cost per unit for each
product.
2. In addition, various product design and research costs (also called “upstream” costs)
have been identified and applied to the Standard model at $15 per unit and to the
Deluxe model at $25 per unit. Also, marketing, distribution, and customer service costs
(also called “downstream” costs) of $120 and $180 have been applied to each unit of
the Standard and Deluxe models, respectively. Compute the full product cost per unit of
each model.
Solutions:
1. Standard model:
Direct Materials $100,000
Conversion Costs:
Materials Handling: 64,000 × $0.40 $ 25,600
Cutting: 64,000 × $2.00 128,000
Assembly: 64,000 × $6.00 384,000
Wood Refinishing: 16,200 × $15.00 243,000 780,600
Total Manufacturing Costs $880,600
Manufacturing Cost per Unit: $880,600 ÷ 8,000 units $110.075
Deluxe model:
Direct Materials $ 62,500
Conversion Costs:
Materials Handling: 30,000 × $0.40 $ 12,000
Cutting: 30,000 × $2.00 60,000
Assembly: 30,000 × $6.00 180,000
Upholstery: 5,100 × $20.00 102,000 354,000
Total Manufacturing Costs $416,500
Manufacturing Cost per Unit: $416,500 ÷ 2,500 units $166.60
2.
Standard Deluxe
Manufacturing cost per unit $110.075 $166.60
Upstream costs 15.000 25.00
Downstream costs 120.000 180.00
Total unit cost $245.075 $371.60
5521.56 Applicability and benefits of ABC: Activity-based costing (ABC) procedures are most
applicable in settings where overhead costs other than organization-level are a relatively
large component of costs. Benefits of implementation include the following:
a. Provides an understanding of complex product costs and product profitability for
improved resource management
b. Allows management to focus on the nature of activities performed so that nonvalue-
added activities can be controlled or eliminated, thus streamlining the production
process
c. Highlights the interrelationship of costs and activities and thus provides opportunities to
reduce costs
d. Provides a more appropriate means of charging overhead costs to products
5521.57 Allocation of support department costs: Costs incurred by support (or service) departments,
such as Human Resources, Purchasing, Building Maintenance, and Power Plant, that directly
assist the production process can be a large part of the overhead to be allocated to user
departments, which use the support services.
5521.58 Allocation bases: Support costs are allocated to user departments using an allocation base,
preferably a cost driver. Typical allocation bases are as follows:
a. Human Resources: Number of employees served
b. Purchasing: Number of purchase requisitions processed
5521.59 Cost allocation methods: Three major methods are used to assign support costs to user
departments:
1. Direct method: Service department costs are allocated directly to the production
departments without any allocation to other service departments for which work might
have been performed.
2. Step-down (or sequential) method: Since service departments often provide services to
the other service departments as well as to the production departments, service
department costs are allocated to other service departments and to production
departments in a predetermined sequence. Generally, the service department allocated
first is the one that provides the most service to other service departments. Where
there is no determinable difference in service levels, total costs in each service
department are used, starting with the department that has the most costs, and then
moving down. Once a service department’s costs have been allocated, no further costs
are allocated to that service department.
3. Reciprocal method: Service department costs may also be allocated using a method
that fully incorporates the interdepartmental support provided among the service
departments. This method requires simultaneous equations and, where there are more
than two service departments involved, a computer program is generally used. It is
theoretically the most accurate method; however, the added precision resulting from its
use may not exceed the cost of using it.
5521.60 Single- versus dual-rate allocation systems: Regardless of which of the allocation methods is
used, support costs may be allocated from a single pool (single-rate method) or from
separate pools for fixed and variable costs (dual-rate method). A dual-rate system for fixed
and variable costs is a more correct procedure, especially if the actual volume of support
activity can vary significantly from the budgeted volume.
5521.61 Example of allocation of support department costs: Spalding Enterprises has two support
departments, Maintenance and Receiving, whose costs are allocated to two production
functions, Machining and Assembly. During the current year, Maintenance and Receiving
incurred the following costs and budgeted volume of activity:
Costs/Departments Maintenance Receiving Totals
Fixed costs $10,000 $15,000 $25,000
Variable costs 30,000 20,000 50,000
Total costs to be allocated $40,000 $35,000 $75,000
Using a single-rate system, the calculations shown below demonstrate allocation of support
costs to the Machining and Assembly departments using the (a) direct method and (b) step-
down method.
a. Direct method:
First, determine the ratio of support activity supplied among the user departments.
Disregard any support provided by one service department to another (ratios rounded
to whole percents).
Maintenance Support:
Machining function 12,000/27,000 hours 44%
Assembly function 15,000/27,000 hours 56%
Receiving Support:
Machining function 1,500/5,500 deliveries 27%
Assembly function 4,000/5,500 deliveries 73%
Next, use ratios to allocate support department costs directly to user departments.
Maintenance Receiving Machining Assembly Totals
Costs to be allocated: $40,000 $35,000
Maintenance
(44%, 56%) (40,000) $17,600 $22,400 $40,000
$ 0
Receiving
(27%, 73%) (35,000) 9,450 25,550 35,000
Totals $ 0 $27,050 $47,950 $75,000
5521.62 Joint products are two or more products whose total resale value is significant and which
are produced simultaneously in the same processing operation. The joint (or common)
processing costs must be allocated to the joint products using estimates. The allocations are
typically based on the relative sales value or relative measures (i.e., units or pounds) of the
products.
5521.63 Byproducts are a product of a process having a relatively small total resale value in relation
to the sales value(s) of the main or joint products. The recovery value may be treated either
as other income or a reduction of main product joint cost.
5521.64 Allocation methods: Four methods are commonly used for allocating joint costs:
1. Relative sales value at split-off
2. Physical output
3. Net realizable value (NRV)
4. Constant gross margin NRV
5521.65 Relative sales value at split-off: Split-off is the point at which separately identifiable main
products emerge from the joint processing phase. Where sales values at split-off are known,
they are usually used to determine the allocation percentage of joint costs to each product.
This method cannot be used in cases where no market exists for one of more products at the
split-off point (i.e., some products require further processing before they are marketable).
5521.66 Physical output: The joint product cost is allocated based on each product’s percentage of
the total physical quantity of all the products. This method is the least preferred because it
does not consider a product’s ability to generate enough revenue to cover the costs
allocated (ability-to-bear concept). In addition, this method requires all joint products to be
measured in the same physical output, which is not always possible.
5521.67 Example of joint product cost: Gyro Company produces 15,000 pounds of Product A and
10,000 pounds of Product B in a joint production process at a total cost of $150,000. The
aggregate sales values of the Products A and B at split-off are $180,000 and $90,000,
respectively. Calculate the joint cost allocable to each product using both the relative sales
value and the unit of measure approaches.
Relative sales value approach: This method will result in the same gross profit percentage
for each joint product at the point of split-off. Each product receives part of the joint costs in
proportion to its sales value in relation to the total sales value of all the products. In this
case, the total sales value is $180,000 + $90,000 = $270,000.
Product A $180,000 × $150,000 joint cost = $100,000
$270,000
Physical output method: This approach will result in the same cost per unit of physical
output (in this example, pounds) for each joint product.
Product A 15,000 (15,000/25,000 = 60%)
Product B + 10,000 (10,000/25,000 = 40%)
25,000 pounds
5521.68 Net realizable value (NRV): This is a variation of the relative sales value method and can be
used in situations where no market value exists for one or more joint products at the split-
off point. The ultimate sales value of each product is reduced dollar for dollar by any
additional processing costs after split-off (i.e., its separable costs) to arrive at its net
realizable value (NRV) at the time of split-off. The NRV is then used to allocate joint costs.
5521.69 Example of net realizable value: Products A and B are produced in Department 1 at a joint
cost of $45,000. Product A is further processed in Department 2 at a cost of $15,000, and
Product B is further processed in Department 3 at a cost of $10,000. Products A and B are
not in salable form after Department 1 processing. Their sales values, after all production
has been completed, are $60,000 and $40,000, respectively. Using the net realizable values
at split-off, calculate the Department 1 joint cost allocable to each product.
Product A Product B Total
Sales value $60,000 $40,000
Less separable costs 15,000 10,000
NRV at split-off $45,000 $30,000 $75,000
5521.70 Constant gross margin net realizable value (NRV): Under this method, joint costs are
allocated so that the gross margin percentage (after further processing, if any) for each main
product is identical. Because this method requires simultaneous equations, calculations are
not likely to be tested on the CPA Examination. However, this method is theoretically
superior to the other methods because it avoids making some joint products appear to be
more profitable than other joint products. In other words, it preserves the idea that one
joint product cannot be produced without also producing the other joint product(s). By
ensuring that the gross margin for all joint products is identical, this method reduces the
likelihood that managers will make inappropriate joint product decisions; managers will be
encouraged to consider the entire group of joint products when making decisions to keep or
drop the joint products, process a joint product further, or emphasize one joint product over
another.
Cost Reporting
5521.72 Cost reporting: Operating statements are prepared for both external and internal purposes.
For external purposes, GAAP must be followed. The matching principle requires that an
absorption (full costing) statement be presented so that product costs (both variable and
fixed) are deducted when sold, and period costs are deducted when incurred. For internal
purposes, management is more interested in the progress it is making toward making a
profit for the period. A profit cannot be made until and unless all fixed costs for the period
are covered. Therefore, an internal report is focused on cost behavior, not on the distinction
between product and period costs.
5521.73 Absorption (full) costing: This is the traditionally prepared income statement. Cost of goods
sold is deducted from sales to arrive at gross margin. The cost of goods sold is made up of all
manufacturing costs attached to the product, including fixed manufacturing overhead per
unit. Fixed costs attached to units not sold are on the balance sheet in the ending inventory.
A major problem with absorption costing is that managers have an incentive to produce
more units, which allows them to defer more fixed costs in ending inventory. This method
discourages managers from rapidly reducing production volumes if sales drop.
5521.75 Operating income: Absorption versus variable costing: Because fixed manufacturing costs
are treated differently between these two reporting approaches, operating income will be
different if sales and production are not equal during the period. The amount of the
difference is the change in inventory units times fixed overhead per unit.
a. Production greater than sales: In this scenario, inventory increases. Absorption costing
operating income will be greater than variable costing net income because some of the
current period’s fixed costs are held in inventory until the next period. On the variable
statement, all fixed costs are deducted regardless of unit sales.
b. Production less than sales: If sales are greater than production, inventory decreases.
Absorption costing operating income will be less than variable costing because, in
addition to deducting all fixed costs relating to this period’s production, some fixed costs
from the previous period will also be in cost of goods sold (beginning inventory), thus
reducing the current period’s income. However, the variable costing statement will still
include only this period’s fixed costs.
5521.76 Example of variable and absorption costing: Smith Company manufactures and sells one
product. The following is information about its operations for the first two months of the
current year.
Selling price per unit $ 11
Variable manufacturing cost per unit 4
Variable selling cost per unit sold 2
Monthly fixed manufacturing costs (normal production
capacity of 10,000 units) 30,000
Fixed manufacturing cost per unit (based on normal
production capacity of 10,000 units) 3
Monthly fixed selling costs 10,000
January February
Beginning inventory -0- $ 1,000
Production $10,000 10,000
Sales 9,000 10,500
Ending inventory 1,000 500
Variable and absorption costing reports for the first two months of the year are presented below:
JANUARY – Variable Costing Report JANUARY – Absorption Costing Report
Sales – 9,000 units × $11 $99,000 Sales – 9,000 units at $11 $99,000
Variable Cost of Goods Sold Full Cost of Goods Sold
Beginning Inventory $ 0 Beginning Inventory $ 0
Plus Production (10,000 units × Plus Production (10,000
$4 per unit) 40,000 units × $7 per unit) 70,000
Less Ending Inventory Less Ending Inventory
(1,000 units × $4) 4,000 36,000 (1,000 units × $7) 7,000 63,000
Manufacturing Margin $63,000 Gross Profit $36,000
Less Variable Selling Expenses
(9,000 units × $2 per unit) 18,000
Contribution Margin $45,000
Less Fixed Costs: Less Period Costs:
Variable Selling Costs
Fixed Manufacturing Costs $30,000 (9,000 units × $2 per unit) $18,000
Fixed Selling Expenses 10,000 40,000 Fixed Selling Costs 10,000 28,000
Operating Income $ 5,000 Operating Income $ 8,000
Inventory increased from 0 to 1,000 units. Therefore, absorption income as reported is higher than variable income. The
difference in dollars is the Change in inventory units × Fixed manufacturing costs per unit (1,000 units × $3 per unit), or $3,000.
Note that the differences in operating incomes shown above represent only a timing difference in the recognition
of fixed manufacturing costs as expenses. Over the long run, total operating income reported under both methods
will be the same.
5522.01 The objective of a standard cost system is to help a business operate more effectively and
more efficiently. It helps accomplish organizational goals by obtaining the optimum output
from the inputs used. Standards and budgets are related. When the budget is broken down
on a unit basis, it is called a standard. Standard costs are budgeted manufacturing costs on a
unit basis.
5522.02 Standard costs have several benefits. They save on bookkeeping costs, provide a basis for
performance measurements, provide useful data in product costing, and are used in the
budgetary process. There are three basic types of standards that may be employed: (1) fixed
or unchanging, (2) ideal or theoretical, and (3) currently attainable. Each type of standard
has some advantages, but currently attainable standards are most commonly used.
a. Fixed or basic standards: Once established, are not changed. Initially, a fixed standard
may be ideal and/or attainable, but because it is never changed, it is of limited use to
management.
b. Ideal or theoretical standards: Based on utopian conditions for a given manufacturing
process. Ideal standards presume that manufacturing costs will always be purchased at
minimum prices, used at optimum efficiency, and produced at 100% manufacturing
capacity. Ideal standards cannot be met and will consistently result in unfavorable
variances. For this reason, they have undesirable behavioral consequences.
c. Attainable standards: Based on a high degree of efficiency. They can be met or even
surpassed by employing excellent managers and employees. Attainable standards
consider that manufacturing components can be purchased at good overall prices.
Attainable standards also consider that (a) labor is not 100% efficient, (b) material usage
will have some normal spoilage, and (c) manufacturing capacity cannot produce at 100%
of theoretical capacity. Attainable standards are set above average levels of efficiency
but may be met or surpassed by efficient production.
5522.03 Many people are generally involved in developing the standard costs. Those commonly
involved include line personnel directly responsible for meeting the standard, engineers,
accountants, and time and motion analysts. Accountants provide valuable cost data and
variance analyses on existing standards that are very helpful in developing good standards.
5522.04 The approach generally followed in developing standards is to estimate the amount and cost
of material, labor, and overhead required to efficiently produce one unit of product in the
current period. The standards are used by accountants to value inventories of raw materials,
work-in-process, and finished goods, and to value the cost of goods sold.
5522.05 When actual amounts differ from standard amounts, a variance is recorded. The variance
should be assigned to the department or division that has the ability to control the activity
that causes the variance. As a general rule, all variances should be recorded as soon as
possible so corrective action may be taken. By doing so, variance analysis is useful to
management because it points to the areas of operations most in need of management’s
attention.
5522.06 Manufacturing accounts used in a standard cost system are similar to the accounts used for
an actual cost system. Several variance accounts are added. Direct material has two variance
accounts—material price variance and material usage variance. Direct labor also has two
variance accounts—labor price (rate) variance and labor usage (efficiency) variance. Over- or
underapplied overhead is analyzed to determine the cause. Separate overhead variance
accounts may also be added.
5522.07 Exhibit A illustrates the flow of costs through the accounts. Lines and arrows connect the
parts of each journal entry. Where the amounts at the ends of the lines are not equal, it is
because a variance is being recorded. The letters and numbers in parentheses identify the
variance account.
5522.08 Direct materials are usually recorded in a raw material inventory account at the standard
cost. The difference between the actual cost and the standard cost for the direct materials
purchased is a price variance. Its computation is:
Direct material price variance = Quantity purchased × (Standard price – Actual price)
5522.09 If the actual price exceeds the standard price, it is an unfavorable variance. By subtracting
the actual price from the standard price, a negative amount will indicate that it is
unfavorable:
Standard price – Actual price
$10.00 – $10.20 = –$0.20
5522.10 Some have proposed that both price and usage variances be recorded at the time the
material is transferred into work-in-process. However, during a given accounting period, the
quantity purchased may not be the same as the quantity used, and the price variance would
be reported in the wrong period. The variance is best associated with the activity when the
price variance is recorded at the time of purchase and the usage variance is recorded at the
time of use.
5522.11 Under manufacturing processes where a specified number of units are being produced, the
standard quantity of raw materials needed for production is known and the usage variance
can be recorded at the beginning of the production process. The standard quantity is
recorded in work-in-process at the standard price. Any difference between the standard
quantity needed and the amount requisitioned is a usage variance. The computation is:
Direct material usage variance = Standard price × (Standard quantity – Actual quantity)
5522.12 Under a manufacturing process where a given amount of raw material is processed to obtain
as much finished product as possible, the usage variance is recorded at the completion of
the production process. The actual raw materials used in production are transferred into
work-in-process at the standard price. When production is complete and the number of
units actually produced is known, the usage variance is computed as the difference between
the raw materials that should have been used (standard quantity) and the amount actually
used (actual quantity) to produce the units of finished product. The formula is the same, but
the time of recording is different.
5522.13 The following summarizes the direct material variances when there is no change in the raw
material inventory balance. This assumes that the quantity purchased is used immediately in
work-in-process as would occur in a just-in-time (JIT) manufacturing process.
5522.14 Workers classified as direct labor are paid for the time worked at the actual or agreed price.
The objective, however, is to record in work-in-process the standard quantity of labor
needed to produce the units at the standard labor rate. The difference between the amount
paid and the amount recorded in work-in-process can be separated into a price (rate) and a
usage (efficiency) variance as follows:
Direct labor rate variance = Hours worked × (Standard price – Actual price)
Direct labor efficiency variance = Standard price × (Standard hours – Actual hours)
5522.15 In some process cost systems, the standard quantity of labor cannot be determined until the
end of the period when the number of units actually produced is known. Under these
systems, the actual direct labor used is recorded in work-in-process, and any usage variance
is recognized when the units are completed and transferred to the next process or into
finished goods inventory.
5522.17 The over- or underapplied overhead can be analyzed in a variety of ways. The most common
are called two-way, three-way, or four-way analysis. The following example will be used to
illustrate each of these.
5522.18 Example: Standard cost data per unit includes the following:
Direct labor: 3 hours at $5.00 per hour
Direct material: 4 pounds at $4.15 per pound
Overhead: $1.10 per direct labor hour × 3 hours = $3.30 per unit
The flexible budget formula is based on fixed costs of $18,000 and variable costs of $0.50 per
direct labor hour. A normal volume of 30,000 direct labor hours is projected for the period.
The overhead rate is $1.10, computed as follows:
$18,000 + ($0.50 × 30,000 hours)
$1.10 =
30,000 hours
Actual cost data for the period:
Overhead Costs
Fixed $18,500
Variable 16,000
Total $34,500
Units produced: 9,000 units
Direct labor hours worked: 28,000 hours
Summary of over- or underapplied overhead:
a
The earned or standard volume was 27,000 direct labor hours. This is the
number of hours that should have been worked to produce 9,000 units
(9,000 units x 3 hours per unit = 27,000 DL hours).
5522.19 A two-way analysis separates over- or underapplied overhead into two variances—
controllable and uncontrollable or volume. The difference between the actual overhead cost
and the amount of cost that should be incurred at the earned or standard volume is the
controllable or budget variance. The difference between the cost that should have been
incurred at the earned volume and the cost allocated to work-in-process is uncontrollable
because it results from a difference between the normal volume and the earned or standard
volume. This uncontrollable variance is often called a volume variance. The over- or
underapplied overhead is split into these two variances by computing the overhead cost that
should have been incurred at the earned or standard volume. The flexible budget formula is
used. Based on the data shown in the example, the following variances are computed under
a two-way analysis:
Notice that the total of the two variances equals the underapplied overhead. Notice also
that the volume variance results from inefficient utilization of fixed overhead. At the normal
volume of 30,000 direct labor hours, fixed costs are $0.60 per direct labor hour. Earned
volume was 3,000 direct labor hours (30,000 – 27,000) below normal, which accounts for the
$1,800 ($0.60 × 3,000) unfavorable volume variance.
5522.20 A three-way analysis separates over- or underapplied overhead into three variances: (1)
spending, (2) efficiency, and (3) volume. The volume variance is the same as computed in
the two-way analysis. However, what was a controllable variance in a two-way analysis is
split into two parts. The difference between the actual costs and the overhead costs that
should have been incurred at the actual volume of production is a spending variance. This
results from prices that are different from those expected at the time the flexible budget
formula was developed. The difference between the overhead costs that should have been
incurred at the actual volume and the costs that should have been incurred at the earned
volume is an efficiency variance. Based on the data shown in the example, the following
variances are computed under a three-way analysis:
Notice that the efficiency variance results from the direct laborers spending 28,000 hours to
do what they should have been able to do in 27,000 hours. Since variable overhead costs
increase $0.50 with each direct labor hour worked, $500 ($0.50 × 1,000 hours) of additional
overhead cost resulted from this inefficiency.
5522.21 A four-way analysis of variance separates over- or underapplied overhead into four
variances: (1) fixed spending, (2) variable spending, (3) efficiency, and (4) volume. The
efficiency and volume variances are the same as computed in a three-way analysis. The
spending variance is split between the fixed and variable expenses. Actual costs are
compared with budgeted costs at actual volume for both fixed and variable expenses to
compute the variances. Based on the data shown in the example, the following variances are
computed under a four-way analysis:
Notice that both the $500 fixed spending variance ($18,000 – $18,500) and the $2,000
variable spending variance ($14,000 – $16,000) are unfavorable and that their combined
total is equal to the total spending variance computed under the three-way approach.
5522.22 A decision must be made as to the disposition of the balance in the variance accounts. The
account balances in question include the material price variance, material usage variance,
labor price (rate) variance, labor usage (efficiency) variance, and manufacturing overhead
variances.
5522.24 Some production processes require a mix of materials and labor to produce a quality
finished product. The process is complicated by the ability to substitute one type of material
or one type of labor for another type of material or labor. However, the substitute material
or labor is often more expensive and the substitution can increase the production cost.
Evaluating the impact on production cost for altering the mix of materials and labor is called
“mix and yield variance analysis.”
5522.25 Mix refers to the relative proportion of components in a mixture. Production mix refers to
the relative proportion of different types of direct materials or direct labor used to
manufacture a given product. Mix is measured relative to a standard proportion of inputs
and has two components—materials mix, which is analyzed by calculating a materials mix
variance, and labor mix, which is analyzed by calculating a labor mix variance.
5522.26 Yield refers to the quantity of finished output obtained from a given amount of input. Yield is
measured relative to a standard amount of good output for a given amount of input. Yield
also has two components—materials yield, which is analyzed by calculating a materials yield
variance, and labor yield, which is analyzed by calculating a labor yield variance.
5522.27 In calculating mix and yield variances, materials and labor should be analyzed separately.
The materials usage variance is divided into two variances—materials mix variance and
materials yield variance. The sum of the materials mix variance and the materials yield
variance should equal the material usage variance.
The labor efficiency variance is divided into two variances—labor mix variance and labor
yield variance. The sum of the labor mix variance and the labor yield variance should equal
the labor efficiency variance.
The usage (efficiency) variance when several inputs are involved is equal to the difference
between (a) Standard price × Actual mix × Actual total quantity, and (b) Standard price ×
Standard mix × Standard total quantity. The mix and yield portions are identified with an
intermediate calculation: Standard price × Standard mix × Actual total quantity. This is
illustrated as follows:
5531.02 Managing a business from a process perspective requires focusing on three crucial factors:
vision, mission, and core values.
1. A vision is a long-term aspiration: a source of motivation to build a business structure,
determine strategic plans, and guide its long-term decision making. A vision is a set of
ideas that describe a future state; the company carries out “vision-achieving operations”
as one of its primary goals.
5531.03 There are multiple stages, or steps, that a PBM organization would follow.
a. The first step is to identify, define, and document the processes that support the
business. These processes may be strategic, operational, or support-based. The
processes should then be prioritized, based upon the impact they have on the business.
The sequence of the processes should be documented, including the input needed for
the process to be operational, the expected output of the process, and the people or
departments responsible for each part of the process. Documentation might consist of
such items as a process map or flowchart, departmental procedure manuals, customer-
based agreements, and purchasing instructions.
b. Once the process is defined and documented, the next step is to establish indicators and
measures that will be used to evaluate how well the process performed, in terms of
efficiency, quality, and timeliness. Measurement metrics should preferably be
quantifiable (numerical). If that is not possible, the next best measurement would be
comparative data and relevant benchmarks such as pie charts, bar charts, cause-and-
effect analysis, and gap analysis.
c. The third step is to analyze process performance to determine if the process itself has
indeed achieved its purpose; additionally, identification of opportunities, threats, and
areas in need of improvement would occur during this phase. Compliance audits are
often used in analyzing process stability. Technological improvements should be
reviewed. If necessary, new goals and process improvements should be planned and
aligned with the vision, mission, and core values of the organization.
d. The final step is to implement the improvement quickly and in a nondisruptive manner,
including providing sufficient resources for the implementation and ongoing operation
of the new procedure. Employee training may be conducted during this phase, as well as
staff education to ensure their support and commitment to the improvements.
5531.04 Business process modeling (BPM) is used in PBM and systems engineering to diagram (or
model) the activities of an organization so that the current process can be identified,
documented, analyzed, and improved. BPM is typically performed by external business
analysts and/or subject matter experts, using one or more of the following business process
modeling tools:
a. Event-driven process chain (EPC): A type of flowchart used for business process
modeling used for modeling, analyzing, and redesigning business processes. An EPC is an
ordered graph of events and functions, providing connectors (e.g., AND, OR) that allow
alternative and parallel execution of processes.
b. Cognition-enhanced natural language information analysis method (CogNIAM): A
conceptual fact-based modeling method that aims to integrate the different dimensions
of knowledge: data, rules, processes, and semantics. CogNIAM structures knowledge,
5531.05 Shared services means that the service is internal; it is the exact opposite of outsourcing. For
example, two hospitals might decide to merge their HR functions.
5531.06 Manufacturers are increasingly using outsourcing to reduce costs and improve their
flexibility to modify strategies and operating plans with changing circumstances. This is the
process of subcontracting services to a third party in exchange for payment. One concern is
the outsourcing of jobs to suppliers outside the country (off-shore locations). Outsourcing
research and development to countries such as India and China has become popular because
of the availability of engineers and the lower cost. Some of the reasons for outsourcing
include the following:
a. Cost savings: the lowering of overall costs of the services to the business
b. Focus on core business: A company may outsource a service such as their information
technology so that they can concentrate on developing their core (profit center)
business and have a specialized company oversee the service.
c. Improve quality: achieve increased quality by contracting out the service to a specialist
d. Capacity management: The risk in providing excess capacity is borne by the service
supplier, not the firm, which improves the methods of capacity management of the
services and technology provided.
5531.07 Off-shore operations means that the organization is incorporated outside the jurisdiction of
the primary operations. For example, a U.S. citizen may use a Swiss company for a certain
portion of his or her business. Competitive pressure and cost reduction are the main goals of
off-shoring. Call centers and IT seem to be the best off-shoring cost-cutters.
5531.08 Business process reengineering (BPR) involves creative, breakthrough thinking and results in
the implementation of new ways of doing business that produce radical improvements in
areas of cost, quality, service, and speed. It requires a horizontal, cross-functional view of
what needs to be done to serve the customer’s needs. However, the focus is on business
processes rather than on the actual products and services. Most change efforts start with
what exists and fixes it. Reengineering starts with the future and works backwards,
unconstrained by existing methods, people, or departments.
5531.09 A combination of factors induces companies to pursue business process reengineering (BPR),
including reducing costs and increasing profits, keeping up with the competition, or
achieving breakthrough results that will set the company apart from the competition.
Reengineering is more than continuous quality improvement, which calls for small,
continuous, incremental steps forward, and which focuses on improving what people are
doing. It involves more radical changes and attempts to change or realign activities so that
the customer is better served.
5531.10 Manufacturing enterprises have been among the first to embrace business process
reengineering (BPR) because of the need to be more competitive internationally and reduce
product cycle time (the time required to convert raw materials into finished products).
Improvements in products can be copied by competitors (reverse engineered) without their
incurring the research and development costs. Improvements in business processes, on the
other hand, provide more leverage and are more difficult to copy.
5531.11 Three tests can be used to determine if a company is doing business process reengineering
(BPR):
1. Is the effort focused on critical business processes that, if they are changed, can have a
major impact on performance? A company’s processes can be classified into critical
processes (those that are central to providing goods and services to customers) and
support processes (those that support critical business processes). BPR focuses on
critical business processes.
2. How ambitious is the desired improvement? A company looking for a 10%–20%
improvement would not be considered to be using BPR, which generally looks to
improve the process by 100%–150%.
3. How receptive is senior management to change? Because BPR requires fundamental
changes in critical business processes, senior management must be committed to
change.
5531.12 Information technology (IT) is usually at the heart of business process reengineering (BPR).
IT is the engine that allows BPR to take place. Intelligent workstations allow the company to
decentralize control throughout the organization. Groupware software helps coordinate a
variety of activities performed by many individuals, and all of this is tied together with
communication networks in a client-server environment.
obtain (or maintain) a competitive advantage, a company must examine the ways it can
add customer value as part of its strategic planning process.
b. Contextual approach: Adding value is contextual. It is different for each industry, for
each firm, for each product or service, and for each process and function within a firm.
For example, adding value to a product that has only just been introduced to the market
requires a different approach (i.e., strategy) than for a mature product. Furthermore,
activities and processes that are initially determined to be value-adding may become
obsolete and nonvalue-adding.
5531.14 In value chain analysis, business processes are viewed as a sequence of activities—a chain of
events. The chain includes both internal and external processes.
5531.15 Value chain activities: Michael Porter, the originator of the idea of the value chain, grouped
the activities into two categories: primary and support activities.
1. Primary activities: These transform inputs into outputs, and usually include the
following:
a. Inbound logistics (material handling and warehousing)
b. Operations (manufacture of the final product)
c. Outbound logistics (processing orders and distribution)
d. Marketing and sales (communication, pricing, and channel management)
e. Service (installation, repair, and parts)
2. Support activities: These are handled by staff functions and include the following:
a. Procurement (purchase of raw materials, supplies, etc.)
b. Technology development (acquisition of know-how and technological inputs)
c. Human resource management (selection, placement, promotion, development,
etc.)
d. infrastructure (general management, planning, finance, accounting, etc.)
5531.16 Steps in value chain analysis: The value chain approach is a vital part of the strategic
planning process. It is a continuous process of gathering, evaluating, and communicating
information for business decision making. It allows a firm to understand its segments,
distribution channels, price points, product differentiation, selling options, and the value
chain configurations that will maximize competitive advantage. Three types of analyses are
conducted: internal cost analysis, internal differentiation analysis, and vertical linkage
analysis.
1. Internal cost analysis: An examination of the internal processes (activities) of the firm is
the starting point of value chain analysis. Its purpose is to identify the sources of
profitability and to determine the relative costs of internal activities and processes. The
principal steps include the following (note the similarities to activity-based costing):
a. Identify value-creating processes and activities.
b. Determine portion of total cost of the product or service attributable to each value-
creating process.
c. Identify cost drivers for each process.
d. Identify links among processes (i.e., could a change in one process cause a change
in a related process; e.g., change the design of a product to increase its
manufacturability, reduce costs, and improve quality).
e. Evaluate opportunities for achieving relative cost advantage (e.g., consideration of
outsourcing a function or activity, development of special relationships with
suppliers, etc.).
2. Internal differentiation analysis: The purpose of this analysis is to identify the sources
of differentiation within the internal activities that can provide opportunities for
creating or maintaining superior differentiation advantage. The focus here is on the
customer’s perceived value of the products/services. It includes the following:
a. Identify customer’s value-creating processes (i.e., how can the firm enhance a
customer’s product so that it will, in turn, add value to something the customer
does?; e.g., how can Boeing enhance an airplane so United Airlines can itself
provide more valuable service to its customers?).
b. Evaluate differentiation strategies for enhancing customer value (e.g., product
features that are aesthetically pleasing or functionally superior; marketing channels
that provide superior responsiveness, convenience, variety, and information;
service and support appropriate to the customer’s specific needs; brand or image
positioning indicative of quality; and price issues relative to customer’s use of the
product or service).
c. Determine the best sustainable differentiation strategies (i.e., use the best selection
of resources available to create value for the customer).
3. Vertical linkage analysis: This has an external focus. It examines all potential value-
adding opportunities in upstream and downstream processes throughout the industry.
It considers the competitive advantage that can come from strategic links with value-
adding processes of suppliers, distribution channels, or users. The analysis includes the
following steps:
a. Identify industry value chain and assignment of costs, revenues, and assets to
value-creating processes. An industry’s value chain includes the value-creating
activities of suppliers of materials and components, as well as varied buyers (end
users), and ends with the disposal and recycling of materials. Included in this
analysis are competitive benchmarking activities and strategic decisions about what
processes of the industry value chain a firm will engage in (i.e., where does the firm
want to participate within the industry value chain?).
b. Determine cost drivers of each value-creating process throughout the industry
value chain.
c. Evaluate opportunities for creating or maintaining competitive advantage. This
requires an examination of how each company within the industry competes so as
to determine where competitive niches are available.
5531.17 Limitations of value chain analysis: The following three limitations need to be taken into
account:
1. Periodicity of financial data: The figures used for value chain analysis are generally
derived from one period’s financial data. Long-term, strategic decision-making requires
information about changes in cost structures, market prices, and capital investments
from one period to another. Where figures are not valid from one period to another, the
analysis needs to be redone under the new conditions.
2. Accessibility of data: Much of the analysis requires collecting data from hard-to-reach
sources.
3. Isolation of information: The determination of cost drivers for each value-creating
activity, the identification of value-chain linkages, and the computation of supplier and
customer profit margins are all difficult. However, the information obtained from value
chain analysis is crucial to the development of an understanding of a firm’s competitive
situation, cost structure, and linkages with suppliers and customers.
Improvement Initiatives
5532.03 The process of selecting improvement initiatives depends on what the organization is trying
to accomplish. One of the factors in choosing an initiative is to assess the degree of fit
between the company’s priorities and the initiatives’ attributes. A report by management
consultant firm Bain & Company suggested four principles for the usage of tools:
1. Get the facts, e.g., strengths, weaknesses, full effects, and side effects of each tool.
2. Champion realistic and strategic directions, not trendy concepts.
3. Choose the best tools for the job.
4. Adapt tools to the business system, but not vice versa.
Source: Management Tools and Trends 2005; Darell Rigby and Barbara Bilodeau; Bain &
Company, 2005.
5532.04 Other factors that influence the choice of an initiative are (1) strategic priorities, (2) past
experiences on improvement programs, and (3) internal and external environment. (Source:
Cagliano, R. and Spina, G. (2000), International Journal of Operations and Production
Management, Vol. 20, No. 7, pp. 772–791.)
5532.05 The decision path of selecting an improvement initiative starts from a company’s key areas
of concern and links it to the improvement initiative’s primary objective. For example, if a
company is feeling pressure to be more competitive in the market, it may look at ISO 900. If
the company has significant waste in the production of its product(s), then it should examine
lean production. If the quality of its products varies greatly, it should implement Six Sigma.
(Source: Bendell, T. (2005), Total Quality Management, Vol. 16, No. 8–9, pp. 969–978.)
5532.06 Management is tasked with choosing the best approach to suit its organization. Is the
chosen initiative appropriate for the organization’s culture? Will it deliver the desired results
on an ongoing basis? Are the organization’s employees able to handle the required
processes?
5532.07 The process for choosing an improvement initiative begins with identifying company needs,
which can be done through a SWOT analysis (strengths, weaknesses, opportunities, threats),
gap analysis, or even using a business excellence framework to identify organizational needs.
Using one of these techniques, the organization can establish a baseline for where it
currently stands, establish its objectives for moving forward, and then establish if it has the
capabilities and resources to begin a specific initiative.
5532.08 Management must not only support the change, but be able to get buy-in from the entire
organization as to the need for the change. The employee’s experience can provide some of
the best data relating to workflow changes, customer satisfaction, and waste, as employees
are closest to the process.
5532.09 Employees at every level of the organization must be active in the implementation as well as
in two-way communication. Once an initiative is chosen, effective communication regarding
the implementation of the initiative must be ongoing. All employees should be provided a
timeline for the implementation, including resource factors such as personnel and
equipment.
5532.10 Implementation of new initiatives will require data collection during and after the process to
determine if the implementation is delivering the desired results. Management should be
prepared to modify the initiative when measured results do not meet targeted expectations.
These modifications can include changing employees, changing equipment, or adding
additional training.
Just-in-Time (JIT)
5532.11 Just-in-time (JIT) refers to systems in which items are delivered where and when needed.
5532.12 Just-in-time (JIT) is a manufacturing philosophy that promotes the simplest, most efficient
means of production. Under ideal conditions, the company would receive raw materials just
in time to go into production, manufacture parts just in time to be assembled into products,
and complete products just in time to be shipped to customers. In other words, goods are
produced to meet demand and are delivered as they are produced, essentially eliminating
finished goods inventory.
a. JIT shifts the production philosophy from a push approach to a pull approach.
(1) The push approach begins with the acquisition of raw materials, which are pushed
into production, through the various stages of production, and finally into finished
goods inventory. It is hoped that finished goods can ultimately be sold to
customers.
(2) The pull approach starts with the demand for finished products from a customer.
The customer order triggers the production of the goods. Inventories are acquired
only to meet production needs.
b. The rate of production for individual departments is determined by the needs of each
succeeding department. Inventory buffers with predetermined maximum sizes are used
to signal when a department should be working. The department works as long as the
buffer is less than the maximum. This eliminates most work-in-process inventory.
c. JIT requires a detailed system for integrating the sequential operations of the process.
Accurate sales forecasts are needed for effective finished goods planning and
production scheduling. If possible, products should be designed to use standardized
parts to reduce manufacturing time and per-unit input costs.
d. Quality control: Little or no inventory is kept on hand, so production is interrupted if
parts do not arrive as needed or are defective. Quality control is critical, as is a good
relationship with vendors to obtain high-quality materials with frequent deliveries. If
quality control procedures are not in place, the entire process breaks down.
e. Kanban: The basic principle of just-in-time production is not to produce anything until
the next stage in the production process is ready to work on it. The procedure
originated in Japan and is based on the “kanban” concept. A kanban is a communication
device, which may be a card, sign, or production bin. When work is completed in one
area, the card, sign, or empty bin is returned to the preceding work area to signal that
additional materials/products are demanded. This demand triggers production at the
previous station, hence the name “demand-pull.” The ultimate demand is the
customer’s order or anticipated orders.
5532.13 Changes are required in several areas of a manufacturing process to make JIT successful.
a. Eliminate activities that do not add value to a product or a service. Activities are
identified as value-adding activities and nonvalue-adding activities. A nonvalue-adding
activity is any activity that could be eliminated without detracting in any way from
customers’ satisfaction with the final product. Business process reengineering is often
used to identify and eliminate nonvalue-adding activities. Machining, milling, and
polishing of products are all value-adding activities, as customer satisfaction would be
reduced without them. Moving time, inspection time, and time in queues are all
nonvalue-adding activities and should be eliminated as much as possible.
b. Obtain agreements with suppliers. Without a supply of raw materials in inventory, the
company becomes very dependent on its suppliers. To minimize this problem, the
number of suppliers is limited to a select few who guarantee both the quality and
timeliness of raw materials. Raw materials are usually delivered in small quantities as
needed for production. Responsibility for raw material quality is also shifted to the
supplier.
c. Develop manufacturing cells. Traditional departments and processes are replaced with
manufacturing cells that contain families of machines typically spread through several
departments. A component or a complete product can be produced within one cell
rather than being transferred between several departments.
d. Obtain a flexible, highly trained workforce. The cellular manufacturing environment
requires a labor force that is capable of operating several different machines. The
workforce must be capable of servicing the machines and performing inspections on
finished products.
e. Implement total quality control. The streamlined flow of raw materials and
components through the production process does not allow for product defects. No
defects can be allowed in parts or raw materials received from suppliers, in work-in-
process, or in finished goods. Emphasis is placed on doing things right the first time and
avoiding rework or waste of any type. Products are inspected for quality at each stage of
the process. Any defect halts the process until it is corrected. Responsibility for
inspecting products is shifted from inspectors to production workers. There is a
commitment for continuous improvement in all aspects of the company.
f. Reduced setup time. Setup time is the time required to change equipment, move
materials, and obtain forms needed to shift production from one product to another.
Meeting customer demand on short notice without any finished goods inventory
requires the ability to shift production from one product to another with a short setup
time.
Lean Manufacturing
5532.14 Lean is a concept closely related to the notion of efficiency that focuses on maximizing
customer value and minimizing costs. It fundamentally means to operate with little waste.
5532.15 To cut back on costs or waste, the organization must consider overproduction, material
handling, and unnecessary inventory. These are only a few of the items that may affect the
waste of the organization.
5532.16 To create a more efficient organization (i.e., one with a lean culture), firms might employ a
few best practices:
a. Teamwork: The more employees within the organization work together, the more
efficient the company becomes. This can help reduce costs and also provide more value
to the customer.
b. 5S: Keep only materials that are needed, arrange material to promote flow, clean
workplace regularly, keep policies consistent, and maintain standards.
c. Mura, muda, and muri: Mura is a Japanese term that describes unevenness in sales and
production. Muda is a Japanese term that describes overproduction, unnecessary
waiting, conveyance, processing, inventory, motion, and correction. Muri is a Japanese
term that relates to overburdening people and equipment. Firms should focus on
identifying these problems and then implementing processes to minimize their impact.
d. Total productive maintenance is the process of identifying issues as soon as possible
and preventing problems before they happen.
5532.17 Demand flow is a strategy driven by customer demand that defines business processes in a
specific workflow. Since demand management helps the organization understand,
coordinate, and control the source of demand to deliver the product, the demand flow has
to be in place to keep the processes running smoothly.
5532.18 The theory of constraints (TOC) is based on the idea that an organization has a single goal
that is achieved through a complex system of linked activities. Improvement in individual
activities often has little effect on achievement of the overall goal. Therefore, TOC focuses
on identifying and addressing variables—called constraints—that have a significant effect on
global performance.
5532.19 Steps in TOC analysis: E. M. Goldratt proposed the following five steps for implementing the
theory of constraints (TOC) as a continuous improvement process:
1. Identify: The first step is to identify the system’s constraints; it asks the question: What
to change? In TOC, a constraint is a limitation on a system’s performance. Goldratt
argued that there are very few (perhaps only one) true constraints in most systems,
although there may be many bottlenecks. Fixing a bottleneck (such as a limited physical
resource) improves performance in only one subpart of the entire system, whereas
fixing a constraint improves total system performance.
2. Exploit: The second step is to decide how to exploit the system’s constraints; it asks the
question: What to change to? This step involves assumptions underlying the current
system and identifying methods for relieving the constraint, leading to improved system
performance. The focus in this step is on identifying simple, practical solutions rather
than complex fixes. Attention is given to breaking through conflicts, avoiding
unintended negative consequences, and ensuring success.
3. Subordinate: The third step is to subordinate everything else; it asks the question: How
to change? This step involves gaining commitment and developing a successful action
plan. Sometimes organizations must change long-standing patterns of behavior in parts
of the system that are not constrained, and there can be emotional resistance to
change. It is essential to actively involve employees in developing the action plan and to
subordinate other activities to relieve the constraint.
4. Elevate: The fourth step is to determine whether desired results have been achieved—
i.e., whether output meets market demand—and, if not, to elevate the process by
obtaining more capacity. Elevation might occur by further modifying capacity use,
outsourcing, or expanding capacity.
5. Go back to step 1: The fifth step is to engage in continuous improvement by repeating
the TOC analysis. Once the system’s constraint has been relieved, the attention turns to
the next constraint. (Theoretically, there is always a constraint.)
5532.20 Throughput is the output relative to input of a system. In TOC (theory of constraints),
throughput is defined as the rate at which the system generates money based on sales.
Thus, the focus is on producing goods that are sold, rather than simply producing goods.
5532.21 The drum-buffer-rope system is a tool for managing product flow, inspired by the theory of
constraints (TOC). The system focuses on maximizing throughput around the constraint as
follows:
a. Drum: The constraint in the system sets the pace (i.e., the “drum”) for production.
Production cannot exceed the slowest path through the system.
b. Buffer: Protective inventory is built into the system to provide slack around the
constraint. Unlike JIT (just-in-time production), in which desirable inventory is zero, TOC
builds an allowance for problems into the system. Such problems might include last-
minute customer orders, delays in vendor shipments, employee absenteeism, and
equipment failure. An inventory buffer is needed to reduce risks to throughput.
c. Rope: a “pull” system such as kanban to trigger product flow
Six Sigma
5532.24 Lean cannot bring a process under statistical control by itself, and Six Sigma cannot improve
processes or reduce capital by itself.
5532.25 Material resource planning (MRP): The acronym MRP refers to “material resource planning”
or to “material requisition planning.” MRP refers to the use of a computerized information
system to schedule and order inventories of raw materials and other components used in
manufacturing. A sales forecast or specific customer’s order for completed product units
specifies a given requirement date. The system then uses a bill of material and master
production schedule to identify the type and quantity of components needed and
determines the amount and timing of orders to be placed for the needed components.
Orders may be placed through electronic data interchange (EDI) or a more traditional
ordering system. In recent years, MRP has evolved to include both materials and financing—
the planning of cash requirements to pay for inventory purchases.
5532.26 Value engineering: This is the system of examining the cost of each component relative to
its function to determine if the cost is equal to the value added by the function. Where the
function is not justified by the cost expended, the component should be eliminated or
5532.28 Pareto principle: Known as the 80/20 rule, the Pareto principle states that a small number of
causes (20%) are responsible for a large number of effects (80%). In business process
performance, this principle suggests that correcting a small number of design,
implementation, or other types of problems can have a large positive effect on overall
performance.
5540.01 In the process of strategic planning, the organization must consider integrated planning.
Integrated planning ensures communication between all stakeholders and affected
departments (work groups) participating in the project.
5540.02 The main goal of integrated planning is to evaluate all costs and benefits affiliated with the
project. The evaluation will help the people involved determine the most appropriate course
of action.
5540.03 The integrated planning process helps link strategic planning and operations management
with financial planning.
5540.04 Interdependence is the level to which work groups are organized. There are three levels of
interdependence to take into consideration:
1. Pooled interdependence: There is a common source of resource, but no
interrelationship between the work groups.
2. Sequential interdependence: The work groups coordinate the flow of information,
tasks, or resources from one group to another.
3. Reciprocal interdependence: Information, tasks, and resources are passed back and
forth between the groups.
5540.05 The information, tasks, and resources must reach the different work groups or organization.
In order for this to happen, an organization must have integrating mechanisms.
5540.06 There are three major integrating mechanisms. These are used to process information and
coordinate the resources of interdependent work groups:
1. General management systems include managerial hierarchy, rules and procedures, and
plans and goals for the organization.
2. Increasing coordination potential: Information systems and lateral relationships are the
most common ways to increase coordinating potential.
3. Reducing the need for coordination: To accomplish this, the organization must create
only pooled interdependence.
5540.07 There are costs associated to the employees related with the work groups, information
systems, and other positions within the organization. There are also costs associated with
the hardware and software for the work groups. An organization must develop cost-effective
integration systems that still meet the needs of the integration and also minimize the
financial and managerial resources necessary to maintain the integrated system.
5540.08 Planning is the selection of goals, predicting potential results under alternative means of
achieving the goals, and a decision on which alternative to implement. Planning establishes
the goals and objectives of the firm. Planning is oriented toward the future. It is the first part
of the budget cycle to assign resources to operating units.
5540.09 Control is the performance evaluation that provides feedback of the results. Control tries to
ensure that the objectives are met, and thus it is oriented toward both the present and the
past. Planning and control are continuous and interdependent.
5540.10 There are many terms describing the hierarchy of planning, and few companies would have
plans labeled with all of the levels given below. However, it is important to remember that
planning is a continuum from the long-term and large-scale to the immediate and detailed.
All the parts of an organization’s planning must fit together regardless of the terms used.
a. Mission and strategic goals: The mission of an organization is its central purpose, the
reason for its existence. It answers the questions “What business are we in?” and “What
business do we want to be in?” Although these may seem like obvious questions, they
are not always easy to answer. A company can meander from one mistake to another
on its way to failure if it does not know what business it is really in. Strategic goals are
the major goals that relate to the company’s mission. The mission defines the broad
goals that are compatible with the company’s mission. There are generally goals related
to such factors as profitability, market share, service and product quality, employee
development, and community service.
b. Strategic objectives and plans: Goals, objectives, plans, etc. are related terms that vary
from organization to organization. The process of translating long-range plans into
short-term action is important here. Strategic plans are company-wide plans with a long
planning horizon, sometimes as long as 10 to 30 years. Many companies have a “5-year
plan” which is revised annually, with a new year added each time. In for-profit
companies, a major part of strategic planning would be profit-related. Capital budgeting
has a strategic component in that the decision to, for example, build a new factory will
affect the company’s profits for many years.
c. Tactical objectives and plans translate overall strategy into shorter-range specifics for
each division or department. These plans contain more detail and are more action-
oriented.
d. Operational objectives and plans are detailed implementation of the company’s
strategy. They are annual plans with specific objectives for individual operating units.
These plans include what are thought of separately as budgets and financing plans. For
some companies, these plans may be the only parts of overall planning that are formal
written documents.
e. Contingency plans are sometimes called disaster recovery plans, to be implemented in
case of fire, flood, power outages, etc. Planning before a catastrophe can result in a
quicker response in the midst of chaos.
5540.11 An effective control system is necessary for an organization to realize its strategic, tactical,
and operational goals. Control points placed so that potential problems are identified before
they occur have much more strategic value than “after the fact” problem identification. An
effective control system:
a. monitors activities and processes so that progress can be measured against established
strategic goals and objectives (measurement of progress),
b. provides for identification of deviations from desired performance so that
implementation of timely corrective action can occur (corrective action), and
c. has the flexibility to be updated to allow for changing conditions (adaptation to
change).
5540.12 A benchmark is a point of reference from which measurements can be made. It serves as a
standard of comparison. Benchmarking, therefore, is the process of identifying products and
services, organizational activities, functions, and processes that represent “best practice”
(i.e., the benchmark) to be strived for and attained. It is a continuous evaluation process and
an essential element of quality audits and gap analysis. Benchmarks can be used in
establishing reasonable goals when creating budgets.
5540.13 Not all capital budgeting projects will have the same level of risk to the company. For
example, replacing a machine is not the same as entering a new market. Various methods
are available to adjust for the levels of risk in different types of projects. The procedures
include the following:
a. Different discount rates: Establish different discount rates.
b. Shorten payback period: Shorten the required payback period.
c. Reduce cash flows estimate: Reduce the estimated future cash flows.
d. Sensitivity analysis: Perform sensitivity analysis, which evaluates how outcomes are
expected to vary with changes in assumptions. Sensitivity analysis should focus on areas
having the greatest uncertainty and/or largest expected impact.
e. Probability distributions: Develop probability distributions for the future cash flows.
f. Certainty equivalent: The certain amount that would be acceptable in lieu of the
uncertain payoffs from the capital project. To compensate for risk, the certainty
equivalent should be greater than the expected value of the capital budget payoffs.
5540.14 Strategic planning determines the path an organization chooses for attaining its mission and
long-term goals. Goals, objectives, plans, etc. are related terms that may vary from
organization to organization.
5540.15 The mission of an organization is its central purpose, the reason for its existence. It answers
the questions “What business are we in?” and “What business do we want to be in?”
Although these may seem like obvious questions, they are not always easy to answer. A
company can meander from one mistake to another on its way to failure if it does not know
what business it is really in.
5540.16 Strategic goals are the major goals that relate to the company’s mission. The mission defines
the broad goals that are compatible with the company’s mission. The mission is translated
into strategic goals that an organization works toward over the long term. There are
generally goals related to profitability, market share, service and product quality, employee
development, community service, and so on.
5540.17 Strategic plans are company-wide plans with a long planning horizon, sometimes as long as
10 to 30 years. Many companies have a “5-year plan” that is revised annually, with a new
year added each time. In for-profit companies, a major part of strategic planning would be
profit related. Capital budgeting has a strategic component in that the decision to, for
example, build a new factory will affect the company’s profits for many years.
5540.18 Strengths, weaknesses, opportunities, and threats (SWOT) analysis: Strategic plan
formation involves an analysis of the enterprise SWOT qualities: internal strengths and
weaknesses (resources) and external opportunities and threats (environment). SWOT
analysis helps managers clarify the organization’s mission, which in turn leads to long-term
objectives such as business diversification, addition or deletion of product lines, or
penetration of new markets.
a. Strengths: These internal factors focus on areas where the company has a comparative
advantage because of internally created resources. Examples are patents, copyrights, or
highly trained people.
b. Weaknesses: These internal factors are the opposite of strengths; they represent areas
where the company is weak or its competitors have comparative advantages over the
company’s internally created resources.
c. Opportunities: These are external factors that involve matching the environment to the
organization. The focus and attention is on critical issues and choices facing the
enterprise. Opportunities can, and should, be prioritized.
d. Threats: External in nature, these factors can arise from a competitor or from changes in
market demographics. Threats may also be the consequence of an adverse change in
one of the company’s necessary inputs.
5541.01 A budget is a formal quantitative expression of an enterprise’s plans. The budget is prepared
for the entire business or a major segment of the business. When the budget is broken down
on a unit basis, it is called a standard.
The traditional or incremental approach to budgeting starts with the activities and costs of
the prior year and adds an incremental layer for growth and/or inflation. Zero base
budgeting is intended to challenge the incremental approach by considering nothing
sacrosanct about the current year as a base for next year’s estimates. Each year you must
start at zero and justify the appropriateness of the item and its cost.
5541.02 The purpose of a budget is to help plan, coordinate, implement, and control enterprise
activities. The budget should be set at an efficient level of operations, but it should not be
viewed as a cost reduction tool. Cost reductions, however, are often achieved when a
budgeting system is adopted because of the inefficiencies due to poor planning,
coordination, and control prior to its adoption.
5541.03 Problems commonly associated with budgeting include inaccurate estimates and adverse
reactions from employees. The budget is an estimate of what should be considered efficient
operations in the future period. Estimates may be inaccurate and events may transpire that
reduce the usefulness of the budget. Employees may react negatively to budgeting
procedures. There is a tremendous amount of work involved in developing a good budget.
Once it is developed, it tends to limit flexibility.
Master Budget
5541.05 A complete budget for a company is called a master budget. A master budget may be
divided into several subparts as follows:
Master Budget
A. Operating Budget B. Financial Budget C. Special Budgets
1. Sales forecast 1. Cash forecast 1. Performance budget
2. Production budget Cash receipts 2. Capital budgets
Material purchases Cash disbursements
Labor costs 2. Budgeted balance sheet
Overhead costs 3. Budgeted cash flows statement
Inventory levels
3. Cost of goods sold budget
4. Selling and administrative expense budget
5541.06 The form of each budget is similar to the form used in developing normal financial
statements. They are called pro forma statements rather than actual statements of results
because they are statements that would be prepared if certain conditions and assumed
events occurred.
5541.08 Problems on the CPA Examination focus on one or more subparts of the master budget. The
most common areas are production or purchases budgets and cash forecasts.
5541.09 Segment reporting is a process of dividing an enterprise into reportable segments and
preparing financial information by segment. We are concerned in this section only with
segmented reporting for internal management use, not for external reporting.
5541.10 Responsibility accounting: Segmented reporting is useful in the management and control of
an enterprise. It is consistent with the concept of responsibility accounting. Responsibility
accounting breaks the company into responsibility centers according to the ability of
individuals to control various activities. Accounting reports are then prepared for each
responsibility center consistent with the assignment of responsibility. There are three types
of responsibility centers:
1. Cost centers: Subdivisions of a business that are assigned responsibility for the
incurrence and proper utilization of costs
2. Profit centers: Subdivisions responsible for both costs and revenues
3. Investment centers: Subdivisions responsible for costs, revenues, and profitable
utilization of invested capital
5541.11 The development of internal reports must be consistent with the type of responsibility
center. Cost center reports should list controllable costs and may also list allocated portions
of fixed costs. There are two general approaches to profit and investment center reporting—
the contribution margin approach and the net income approach.
5541.12 The contribution margin approach subtracts variable costs and fixed costs that can be
identified to a specific segment from segment revenue to give an amount called segment
contribution margin. It is the amount provided by the segment to help cover unallocated
fixed costs and yield income. The contribution margin approach is most useful when making
a decision to add or delete a segment of the business. The following form illustrates this
approach:
Division
Entire Business A B
Sales $ $ $
Less: Variable costs $ $ $
Contribution margin $ $ $
Less: Separable fixed costs $ $ $
Segment contribution margin $ $ $
Less: Unallocated costs $
Net income $
5541.13 The net income approach computes net income for the revenue-generating segments of the
business. In computing net income for these segments, cost allocation is required for fixed
costs that are used jointly by two segments and service department costs that generate no
revenue but provide services to the revenue-producing segments.
5541.15 When service departments render services to other service departments as well as
production departments, you may be required to use direct allocation or step allocation.
a. Direct allocation: No costs of one service department are allocated to another service
department. Costs are allocated directly to producing departments.
b. Step allocation: One service department’s cost may be allocated to another service
department, beginning with the service department that renders the most services to
other departments.
5541.16 Different methods are needed to evaluate the performance of the different types of
responsibility centers.
a. Cost centers: Evaluated on costs incurred relative to budgeted and standard costs
b. Profit centers: Evaluated on sales and costs relative to budgeted amounts. Contribution
margin and vertical ratio analysis can also be applied.
c. Investment centers: Evaluated on the relationship of profits to invested capital.
Alternative methods of evaluation are return on investment, residual income, and
segment margin analysis.
5541.17 Segment margin analysis: Segment margin is the contribution margin of a segment less fixed
costs that can be traced to the segment. Segment fixed costs are those that would be
avoided if the segment is discontinued. Segment margin represents the amount the segment
contributes toward covering common fixed costs and toward net income.
5541.18 Gross margin analysis (or gross profit analysis) is a way of analyzing unexpected changes in
gross margin. It is a technique to isolate changes in each of the sales and cost elements that
comprise the total change in gross margin.
5541.19 The first step in gross margin analysis is to compute the difference between actual gross
margin and budgeted gross margin. This difference is called the gross margin variance. The
gross margin variance is then divided into component parts.
a. Sales price variance: The effect on gross margin resulting from differences in sales price
between planned and actual results
b. Sales mix variance: The effect on gross margin resulting from differences in sales mix
between planned and actual results
c. Sales volume variance: The effect on gross margin resulting from differences in the
quantity sold between planned and actual results
d. Cost price variance: The effect on gross margin resulting from differences in
manufacturing costs between planned and actual results
5541.20 The following example illustrates the computations required in gross margin analysis
(adapted from the AICPA).
Problem: Garfield Company, which sells a single product, provided the following data from
its income statements for the calendar years 20X2 and 20X1:
20X2
Sales (150,000 units) $750,000
Cost of goods sold 525,000
Gross profit $225,000
20X1
(Base Year)
Sales (180,000 units) $720,000
Cost of goods sold 575,000
Gross profits $145,000
In an analysis of variation in gross profit between the two years, what would be the effects
of changes in sales price, sales volume, and cost?
Solution:
The change in gross profit to be accounted for is computed as follows:
Gross profit: 20X2 $225,000
20X1 145,000
Total increase $ 80,000 Favorable
Total cost variance:
($575,000 - $525,000) $ 50,000 Favorable
Sales price variance:
20X1 sales price
($720,000 ÷ 180,000 units) $ 4.00
20X2 sales price
($750,000 ÷ 150,000 units) $ 5.00
Increase $ 1.00
Times units sold 150,000 $150,000 Favorable
5541.21 Product pricing: Product and service pricing refers to assigning monetary values to goods
and services exchanged with external entities.
5541.22 Pricing decisions differ substantially depending on the market structure within which a
company operates. Market structure is determined by the number of firms in the market,
barriers to entry, and communication among firms. Pure competition and monopoly are at
the two extremes of market structure, with monopolistic competition and oligopoly in
between. Very few markets are classified as purely competitive or purely monopolistic; most
markets are monopolistic competition or oligopoly. Pricing strategies will differ depending
on the type of market within which a company operates. Generally, a company will
maximize profits when it prices its products and operates at a volume where marginal
revenues equal marginal costs. These economic concepts are conceptually sound but
difficult to implement because they are based on demand curves, cost functions, and
competitor strategies that are difficult to measure and quantify. Therefore, most companies
use a cost-based pricing strategy, but monitor and adjust the price based on competitor’s
prices, alternative products, customer’s attitude, and the actual profit a given price provides.
5541.23 A long-term pricing policy requires the product’s price to be high enough to cover all costs
and provide an adequate return on invested capital. Methods for computing the product’s
price include the following:
a. Rate of markup: A markup is added to the cost of the product to cover other operating
costs and provide net income:
Selling price = Product cost + (Product cost × Markup percentage)
The markup percentage is based on a desired return on assets employed, the expected
amount of other costs, anticipated sales volume, and the unit’s cost. If product cost is
based on total manufacturing costs (using absorption costing), the following formula
would be used to calculate the markup percentage:
(Desired R of R × Assets employed) + S & A expenses
Markup % =
Unit volume × Unit manufacturing costs
5541.24 Service pricing: Many organizations (like repair shops, doctor offices, consulting
organizations, and CPA firms) have labor as a major portion of the cost of the service they
provide to their customers.
5541.25 Pricing strategy: Organizations like these typically use a time and materials pricing strategy.
Two rates are used—one for time and a second for materials.
1. Time: A rate for time can be developed by estimating all the costs associated with the
time component of the business, adding a desired amount for profit, and dividing by the
estimated number of hours expected to be worked during the time period. An
alternative to this (used by most CPA and consulting firms) is to calculate the desired
gross pay per hour for each employee and multiply it by a standard multiple. When the
employee works on a particular job, the amount charged to the client is equal to the
number of hours worked times the rate per hour.
Example: If a new accountant is to be paid $50,000 per year and is expected to work
2,000 hours for the year, the gross pay per hour is $25 ($50,000 ÷ 2,000 hours). Using a
standard multiple of four, the new accountant’s billing rate would be $100 per hour.
Supposing the new accountant spends 20 hours on a particular job, the client would be
charged $2,000 ($100 × 20 hours).
2. Material loading charge: A material loading charge is developed by estimating the costs
associated with purchasing, handling, and storing the materials for the time period,
adding a desired profit on the materials provided, and dividing by the total cost of the
materials for the time period. When materials are used on a particular job, the amount
charged to the customer is equal to the actual cost of the materials plus the materials
loading charge.
Example: Assume that a company expects to use $100,000 of materials during the year,
and they expect to spend $40,000 on purchasing and handling the materials, plus they
would like to earn $20,000 on the materials handled. Their materials loading rate would
Transfer Pricing
5541.26 Transfer price is the internal price charged by a selling department, division, or subsidiary of
a company for a raw material, component, finished good, or service to a buying department,
division, or subsidiary of the same company.
5541.27 Transfer pricing methods: There are three basic methods for establishing a transfer price:
a. Market-based price: The transferring entity prices the goods or services to the buying
entity at a price equal to that prevailing for those goods or services on the open market.
Theoretically, this should be the same price at which the selling entity would sell the
same goods or services to an external entity.
Advantages of this method include the following:
(1) Business units can operate as independent profit centers, with the manager being
held responsible for the center’s performance. This increases motivation for
managers and aids performance evaluation by top management.
(2) It fairly distributes profits among the various internal entities. This results in fair
taxation by various countries for multinational organizations.
Disadvantages of this method include the following:
(1) A market price may not be available for intermediate products because no external
market exists.
(2) Using a market price and allowing a subdivision of a company to purchase
intermediate parts from outside organizations may suboptimize net income for the
company as a whole when idle capacity exists within the providing subdivision of
the company.
b. Cost-based prices: The transferring entity prices the goods or services to the buying
entity at its cost of providing the good or service. The cost may be full cost, full standard
cost, actual variable cost, or marginal cost. A decision must also be made about charging
part of research and development to the cost of the product.
Advantages for using cost include the following:
(1) Cost data is generally available within the company.
(2) This method is generally accepted by taxing authorities in a multinational
environment.
Disadvantages associated with this method include the following:
(1) Identifying the cost of the good or service may be manipulated by the costs
included in the calculation.
(2) There is a loss of motivation to control costs within the providing division of the
organization because all costs are passed on to the purchasing division. This erodes
the competitiveness of the firm.
(3) It does not provide any meaningful measure to evaluate the performance of the
selling division of the company.
c. Negotiated price: The transferring entity prices the goods or services to the buying
entity at a price that is mutually acceptable to both the selling and buying divisions.
Advantages with a negotiated price include the following:
(1) A negotiated price can always be developed.
(2) It makes managers responsible for their performance, provides reasonable
measures to evaluate performance, and encourages cost minimization.
Disadvantages with using this method include the following:
(1) There may be significant time associated with negotiating a fair price.
(2) It can result in suboptimization of enterprise net income.
5541.28 There is a range within which the transfer price should be established to optimize net
income for the enterprise as a whole. The minimum transfer price should be equal to the
variable costs of the selling division plus any lost contribution margin per unit on sales it
could have made to outside entities that are lost by providing the goods or services to the
internal division. The maximum transfer price should be equal to the price at which the
buying division could purchase the good or service from an external entity.
5541.29 Objectives in transfer pricing: There are several objectives that must be balanced in
selecting a transfer price, especially when the company operates in several different
countries.
a. Evaluate performance: One key objective of a transfer price is to fairly and accurately
measure the performance of a segment and its management.
b. Minimize taxation: Global taxes can be minimized by using transfer prices to move
products at cost out of countries with high corporate taxes and generate profits in
countries with low corporate taxes.
c. Minimize tariffs: Tariffs can be minimized by using transfer prices to move products at
cost to or from countries with high duties or customs and at market value for countries
with low tariffs.
d. Hedge exchange rates: Transfer prices can be used to reduce exposure to exchange rate
risks by moving funds out of weak currencies into strong currencies.
e. Move cash: Transfer prices can be used to move cash from countries with exchange
controls that restrict the repatriation of dividends and capital.
f. Improve competitive position: Lowering the transfer price to a subsidiary can enable it
to lower its prices to match or undercut local competition.
5542.01 Introduction
a. Projected financial statements are used by managers to:
(1) plan future performance and compare the projections to goals.
(2) plan future financing needs.
(3) look at the effects of proposed changes in operations.
b. Mission statements define the general purpose of the company, including the purpose
of the firm’s existence (often maximizing stockholder wealth); the scope of the company
(defining the industry and location of operations); and the objectives of the organization
(general philosophy).
c. Corporate strategies are developed after the mission statement. They are broad, long-
range plans such as developing new technologies in a particular field. These strategies
need to be realistically obtainable in the long term and attuned with the organization’s
purpose, scope, and objectives.
d. Operating plans are then developed and used to implement the corporate strategy, and
are usually 1–5 years in length with the first year of the plan being the most detailed.
The plan will identify responsibility for implementation, specific tasks to be
accomplished, and revenue and cost targets, among other items.
would need to modify the original financial plan to compensate for actual conditions.
Since forecasts are based upon anticipated economic conditions and expected customer
needs and wants, adjustments are often necessary as circumstances change (strong vs.
weak economy, response by the competition, natural events such as floods, disasters,
etc.).
g. A performance-based compensation system may need to be developed in order to
ensure that management is rewarded for the accomplishment of goals related to the
attainment of strategic goals.
Forecasting
exponential tread would be sales that have increased by 20% for each of the past four
years and are expected to continue increasing at that 20% rate into the future. The
formula for exponential growth is:
Pt = Po × ekt
Where: Pt = Forecasted amount after time t
Po = Original amount
e = 2.7183
k = Growth or decline rate
t = Time
f. Constant, linear, and exponential trends can be determined through the use of
graphing, algebraic approaches, spreadsheet functions. The analyst, however, must take
care not to identify a trend that does not exist. Not all sets of historic data points will
form a trend that can be used to predict the future.
g. POS forecasting incorporates point of sale (POS) data into the forecast in an attempt to
improve the accuracy of various forecasts. The hope is to improve the operations of a
company and increase profits by having relatively accurate predictions. Such data as
consumer demand, retail channel demand, and wholesaler channel demand are
incorporated into a sophisticated model in an attempt to improve the accuracy of a
particular forecast. Due to the fact that there are many forces at play in the real-world
markets, a particular set of circumstances is likely to occur twice, thus complicating the
entire forecasting process. By developing a model that considers how sales respond to
various events and combination of events, a greater degree of accuracy can potentially
be achieved. When using POS forecasting, such items as the following are taken into
consideration:
(1) Effect of advertising, promotions, and special packaging on sales
(2) Historical data related to stock outages
(3) Distribution channels and anticipated changes in those channels
(4) Actions of competitors
(5) Seasonality of sales
(6) Current market trends
d. Surveys prepared by current customers or test markets can be used to predict sale
forecasts.
e. Forecasts by analogy can be useful when introducing a new product. Most products go
through the product life cycle stages of introduction, growth, maturity, and decline.
When introducing a new product, the life cycle trends of similar products can be used to
predict the potential sales trends for the new product.
Regression Analysis
5542.05 Regression analysis, often utilizing the least-squares method, is used to estimate the
relationship between a dependent variable and one or more independent variables. It is
often used as a method to estimate a cost function, i.e., to use historical relationships to
predict future costs.
5542.06 Simple and multiple regression analysis: Simple regression analysis indicates that the
dependent variable is predicted on the basis of one independent variable. Multiple
regression analysis is concerned with predicting the dependent variable on the basis of two
or more independent variables.
5542.07 Caution about predicting the future: Regression analysis typically relies on historical data
from events that have already occurred. One hazard regarding regression analysis is that it
requires an assumption that the relationships will continue into the future. If the
environment changes in the future, then historical data may not be helpful for predicting the
future. For example, if there is a change toward automated manufacturing, then prior data
regarding the relationship between direct labor and unit costs may no longer be valid.
5542.08 Assumptions: There are four assumptions that need to be met to produce valid results with
regression analysis using the least squares method:
1. Linearity: The relationship between dependent and independent variables must be
linear within the relevant range of activity. This can be seen visually in a scatter diagram.
If the actual relationship is not linear, a linear regression line will not fit very well.
2. Constant variance of the residuals: Residuals represent the differences between
estimated and actual values. The dispersion of the data points around the regression
line should be the same along the regression line. Violation of this assumption makes
the standard error of the residuals less reliable.
3. Independence of the residuals: The deviation of one data point around the regression
line should be unrelated to the deviation of any other point. If the deviations are
related, it complicates the analysis of the standard errors.
4. Normal distribution of residuals: The normal distribution is the “bell curve.” This
assumption is necessary to be able to make confidence-level statements using the
standard error of the residuals.
Where:
y′ = the estimated value of the dependent variable (sometimes written as ^y; when
estimating a cost function, y′ is the estimated cost for a given level of the cost
driver)
x = the independent variable (when estimating a cost function, x is a cost driver)
a = the point of intersection of the linear regression line with the Y axis (also called
the intercept or the constant; when estimating a cost function, a is the fixed cost)
b = the slope of the regression line (when estimating a cost function, b is the
variable cost per unit of the cost driver)
b. Best-fitting regression line: Under the least squares method, the best-fitting regression
line (and equation) is that for which the sum of the squared deviations between the
estimated and actual values of the dependent variable for the sample data is minimized.
c. Estimate dependent variable: Once the regression equation has been calculated, it can
then be used to estimate the value of the dependent variable given the value of the
independent variable. However, such estimation would only be done within the range of
the independent values originally sampled, since there is no statistical basis to assume
that the regression line is appropriate outside these limits.
5542.10 Standard error of the estimate (Se): The range within which the true y is located. Under the
regression equation assumptions, there is a 68.3% probability that the true y is within one Se
of y′, and there is a 95.5% probability that the true y is within two Se of y′. Se is also called the
standard error of the residuals, as it is the square root of the variances of the residuals
(difference between the predicted and the actual values of the data points).
∑(y-y ' )2
Se = �
n-2
5542.11 Confidence interval for estimated values of y: A confidence interval for an estimated y value
can be constructed in a similar manner.
5542.12 Coefficient of determination (R2): A ratio that indicates the proportion of variance in the
dependent variable that is statistically explained (determined) by the independent variable
using the regression equation. The total variation of the dependent variable y from its mean
can be divided into the part explained by the regression equation and the part that is not
explained, but is produced by chance. The values range from 0 (no explanation) to 1 (all
explained). It can also be expressed as a percentage, as in “x explains 70% of the variation in
y.”
5542.13 T-statistic (t-value): The population standard deviation is assumed to be normal, but it is
unknown. The t-value of the slope coefficient (b) is used to evaluate the significance of the
relationship between changes in the dependent variable (y) and the independent variable
(x). The higher the t statistic, the less likely the relationship between the variables occurred
by chance.
After calculating the standard error of the coefficient b, the standard error of the coefficient
is divided into the coefficient. This gives a measure of the relative size of the standard error,
or precision.
Example: If the b coefficient had a value of 100 and its standard error was also 100, the
range estimate of the b coefficient would be 100+/–100, and the t-statistic would be
100/100, or 1. This is not a very accurate (precise) b. In fact, the higher the t-statistic is, the
more useful it is. If the standard error of b were 10, rather than 100, the relationship would
be 100+/–10 with a t-statistic of 100/10, or 10. The t-value should be at least 2 to be
considered statistically meaningful.
Coefficient (b)
t-value =
Standard error of the b coefficient
𝑆𝑆
𝑒𝑒
Standard error of the b coefficient (Sb) = �∑(𝑥𝑥−𝑥𝑥)2
5542.14 Regression example for a cost function: To use regression to estimate a cost function, i.e.,
to separate a mixed cost into fixed and variable components, we assume that number of
units is the cost driver. Data for total cost (TC) and units (x) are as follows:
Units (x) TC $ (y)
80 350
90 390
85 365
95 405
100 410
110 425
Output for these data using the Excel regression function is as follows:
SUMMARY OUTPUT
Regression Statistics
Multiple R 0.962584132
R Square 0.926568212
Adjusted R Square 0.908210265
Standard Error 8.644775466
Observations 6
ANOVA
df SS MS F Significance F
Regression 1 3771.904762 3771.9048 50.472322 0.002073731
Residual 4 298.9285714 74.732143
Total 5 4070.833333
5542.15 Learning curve analysis: This method involves applying a mathematical formula to estimate
future time to produce units. Learning curves were first developed in the aircraft industry
before World War II. Analysts discovered that the direct labor input per airplane declined as
more and more airplanes were produced. Surveys of manufacturers showed that the direct
labor for the 8th unit was only 80% of the direct labor required for the 4th unit, and the
direct labor for the 12th unit was only 80% of the labor for the 6th unit.
5542.16 Competitive strategy: Learning curves enable the manager to project the manufacturing
cost per unit.
a. Application: Firms that emphasize low price and large volumes must move down the
learning curve to reduce costs per unit. In the electronics industry, the cost of
development can be quite high, so the first units must be priced relatively high. As costs
and prices fall, the first companies in the market have a big advantage. Newcomers
must have large start-up costs and suffer large initial losses.
b. Product changes: Product changes can disrupt the benefits of the learning curve model
and cost reductions may not be realized.
5542.17 Assumptions
a. Increasing efficiency: The direct labor required for the n + 1st unit will always be less
than the labor required for the n unit.
b. Decrease will decline: Direct labor will decrease at a declining rate as cumulative
production increases.
c. Exponential curve: The reduction in time will follow an exponential curve. In other
words, the production time per unit is reduced by a fixed percentage each time
production is doubled.
production, the direct labor costs will exceed the average and in the later stages the
reverse will be true. This can be used to provide a basis for comparing prices and costs.
c. Projection of labor requirements: For a given production schedule, the curve can be
used to project the labor requirements. This can be used to estimate training and to
develop hiring plans.
d. Managerial considerations: The following are issues to keep in mind when using
learning curves:
(1) Curve estimate: An estimate of the curve is necessary. Using industry averages may
be risky.
(2) Unit estimate: The estimate of the first unit is an important element.
(3) Influencing factors: The learning rate depends on factors such as product
complexity and the rate of capital addition. The simpler the product, the less is the
learning rate. A complex product offers more opportunity to improve work
methods and processes. TQM (total quality management) and continuous
improvement will also affect learning curves. Automation alters the learning rate as
there is less opportunity to make reductions in the required hours per unit.
(4) Benefit: The greatest advantages from the learning curve come in the early stages
of a new product.
Example 1: A 90% learning curve of hours is developed as follows. Assume that the first unit
requires 300 hours:
Cumulative Average Cumulative
Units Hours per Unit Total Hours
1 300 300
2 270 540 (2 × 270)
4 243 972 (4 × 243)
8 218.7 1,749.6 (8 × 218.7)
Calculate the number of hours needed for a particular number of units by subtracting their
cumulative hours from the previous units.
Answer: Number of hours needed for the last 4 units = 1749.6 – 972 = 777.6. Average hours
per unit for the last 4 units = 777.6/4 = 194.4 hours per unit.
Example 2: A company has a learning curve of 80%, and the first 50 units took 1,000 hours to
make. How long will the next order for 150 units take?
Answer: 50 units @ 1,000 hours = 20 hours per unit; therefore, on the first doubling to 100
units the cumulative average is 16 hours (20 hours × 0.80) per unit. The second doubling to
200 units produces a cumulative average of 12.8 hours (16 hours × 0.80) per unit.
200 units × 12.8 hours per unit 2,560 total hours
Subtract the hours for the first 50 units 1,000 hours
Hours to make another 150 units 1,560 hours
5542.19 Time series analysis: These methods examine historical data and incorporate the
assumption that the future can be predicted from the past. With time series analysis, the
independent variable is time. The object is to identify consistent patterns over time that can
be used to forecast the future behavior of the dependent variable.
5542.20 Components: There are four components that contribute to patterns over time. Each time
series of data may be influenced by any or all of the components. The problem is to
determine how much of each component influences the series. A moving average can be
used to smooth the data, so the trend is easier to see by filtering out seasonal or cyclical
variation.
1. Trend: This is the long-term movement of the dependent variable over years or decades
(e.g., long-term inflation).
2. Seasonal variation: This is movement within the year, such as peak sales periods from
selling ski equipment during the winter and Christmas or back-to-school sales.
3. Cyclical variation: This is movement upward and downward over periods longer than a
year, such as the business cycle of prosperity, recession, and recovery.
4. Irregular: This is unpredictable or random movement of the dependent variable. It can
also be valuable information to know that this variable cannot be used in forecasting.
5542.21 Classical decomposition of a time series: Time series analysis is applicable when data shows
trend, seasonality, and/or cyclic variations. This method isolates these components to
analyze their effects individually. It does not isolate random and irregular variations.
a. Naive (or random walk) approach: Using a sequence of observations, the latest is used
to predict the next. While simple and involving little cost, this method is inaccurate
unless there is little change in values from one period to another. However, this
approach may be appropriate for some types of time series, such as prices of firms in
the stock market, that stagger up and down over time (i.e., where the change from one
time period to the next is random).
b. Moving average: Fluctuations in data are smoothed by updating an average each time a
new observation is added. Moving average information reflects trends that occur in the
data over time, but changes in patterns of data are not immediately apparent; moving
averages typically lag trends. Also, seasonal variations are not considered. This method
assumes that the resulting average occurred in the middle of the time sequence used.
c. Non-weighted moving average: In this procedure, values are forecast based on the
mean value over a specific number of prior periods. Each of the prior periods is given
equal weight.
Example: The forecasts using the naive (or random walk) method and a three-month
moving average for orders delivered are as follows:
3-Month
Orders Naïve Moving
Month Delivered Forecast Average
January 130 300
February 90 130
March 100 90 106.7 (change assumed to have occurred in
February)
April 75 100 88.3 (change assumed to have occurred in
March)
May 110 75 95.0 (etc.)
June 50 110 78.3
July 75 50 78.3
August 130 75 85.0
September 110 130 105.0
October 90 110 110.0
Solution: The March and April moving averages are calculated as follows:
March: (120 + 90 + 100) ÷ 3 = 103.3
April: (90 + 100 + 75) ÷ 3 = 88.3
If it is the end of October and you wish to forecast November deliveries, simply use the
last moving average you are able to calculate as the forecast for the next period. In this
case:
(130 + 110 + 90) ÷ 3 = 110, the forecast for November
d. Weighted moving average: Based on experience, assign weights to each of the prior
periods.
Example: For the preceding example, suppose past experience indicates that the
number of orders in the most recent month is three times as important as the number
of orders three months ago, and that the number of orders two months ago is twice as
important as the number of orders three months ago. In other words, calculate a three-
month weighted moving average for November in which October orders are weighted
times 3, September orders are weighted times 2, and August only 1.
Solution: [3(90) + 2(110) + 1(130)] ÷ (3 + 2 + 1) = 103.3, the forecast for November
e. Exponential smoothing: A method that iteratively calculates a weighted average by
applying weights to the most recent observation and to prior-period smoothed
weighted averages. The weighting factor is based on experience and judgment.
Exponential smoothing techniques can be made more sophisticated by making
adjustments for trend and seasonality.
The formula for simple (also called single) exponential smoothing is:
f(t) = wy(t-1) + (1-w)f(t-1)
Where:
f(t) is the exponential weighted moving average for time t
y(t-1) is the observation for the previous period
f(t-1) is the exponential weighted moving average for the previous period
w is the weighting factor (also called the smoothing constant)
Example: Use the previous data and a weight of 0.10 to calculate the simple exponential
weighted moving average for March through November. Assume that the initial
exponential weighted moving average (i.e., for February) is equal to the actual value for
the first month (January).
Solution:
Exponential Weighted
Orders Weighted (Smoothed)
Month Delivered Moving Average
January 130
February 90 130
March 100 126 [0.10(90) + 0.90(130)]
April 75 123.4 [0.10(100) + 0.90(126)]
May 110 118.6 [0.10(75) + 0.90(123.4)]
June 50 117.7 [0.10(110) + 0.90(118.6)]
July 75 110.9 [0.10(50) + 0.90(117.7)]
August 130 107.3 [0.10(75) + 0.90(110.9)]
September 110 110 [0.10(130) + 0.90(107.3)]
October 90 109.6 [0.10(110) + 0.90(109.6)]
Expected Value
5542.22 Expected value: Under conditions of uncertainty (or risk) when it is possible to assign
probabilities, expected value is the sum of the outcomes (payoff) of each event multiplied by
the probability of each event occurring. The payoff table includes probabilities relating to the
states of nature:
States of Nature
Decision
Options Good Economic Poor Economic
(Purchases) Conditions Conditions
60% 40%
Apartments $ 50,000 $30,000
Office Building 100,000 (40,000)
Warehouse 30,000 10,000
Now, for each decision option, calculate the expected value as follows:
Expected value of apartment: 0.6(50,000) + 0.4(30,000) = $42,000
Expected value of office: 0.6(100,000) + 0.4(–40,000) = $44,000
Expected value of warehouse: 0.6(30,000) + 0.4(10,000) = $22,000
Another, slightly more complicated example follows:
Example: The College Honor Society sells hot pretzels at the home football games. The
frequency distribution of the demand for pretzels per game is presented below:
Unit Sales Volume Probability
2,000 pretzels 0.10
3,000 pretzels 0.15
4,000 pretzels 0.20
5,000 pretzels 0.35
6,000 pretzels 0.20
The pretzels are sold for $1.00 each, and the cost per pretzel is $0.30. Any unsold pretzels
are discarded because they will be stale before the next home game.
Compute the payoff table amounts for each action and state of nature as follows:
If demand is 2,000 and 2,000 pretzels are stocked, the profit is $1,400 ($2,000 – $600 = $1,400).
If demand is 2,000 and 3,000 pretzels are stocked, the profit is $1,100 ($2,000 – $900 = $1,100).
If demand is 2,000 and 4,000 pretzels are stocked, the profit is $800 ($2,000 – $1,200 = $800), etc.
Notice that in creating the profit for each action, the probabilities are ignored until
calculation of the expected value of an action column (i.e., the last row).
5542.23 Expected value of perfect information: If the decision maker had perfect knowledge as to
which state of nature would occur, the decision offering the highest payoff could be
selected. The difference between the expected value with the highest payoff (complete
information) and the expected value calculated in the payoff table (condition of risk) is the
amount one would be willing to pay to acquire perfect information. This can be useful when
considering how much to spend to gather more data—cost versus benefits. To calculate the
expected value with complete knowledge, select the highest payoff for each state of nature:
State of Nature Highest Payoff Probability
2,000 $1,400 × 0.10 = $ 140
3,000 $2,100 × 0.15 = 315
4,000 $2,800 × 0.20 = 560
5,000 $3,500 × 0.35 = 1,225
6,000 $4,200 × 0.20 = 840
Expected value with highest payoff $3,080
Best expected value from the payoff table 2,700
Amount willing to pay for perfect information 380
5542.25 Expected value and cash flows with probabilities: If there is a series of future cash flows,
and each has a probability of occurring, the expected value (the same as the mean) of these
cash flows can be determined. Then, the variance and the standard deviation of these future
cash flows can be determined. The calculations are slightly different from the variance
without probabilities.
Probability of
Cash Flow Occurrence
10,000 10%
12,000 40%
15,000 30%
18,000 20%
The expected value or mean of this estimate of cash flows is:
0.1(10,000) + 0.4(12,000) + 0.3(15000) + 0.2(18,000) = 13,900
(5) With the introduction of a new product, a firm can often have a competitive
advantage with increased market share until the competition can develop similar
products.
Actual Forecasted
Current Relationship Projected
Sales $110,000 110% of current $121,000
Costs except depr. 80,000 73% of sales 88,300
Depreciation 12,000 15% of P&E 15,200
EBIT 18,000 17,500
Interest 5,000 Same as current 5,000
EBT 13,000 12,500
Taxes (40%) 5,200 5,000
NI before dividends 7,800 7,500
Dividends 1,000 Same as current 1,000
Addt’s to RE $ 6,800 $ 6,500
c. Spontaneous funds are generated through the use of credit. As inventory purchases and
production increase (driven by increased sales), current liabilities such as accounts
payable and accrued labor also increase. There is a direct relationship between many
current liabilities and projected sales.
d. The relationship between sales and property plant and equipment (PP&E) is often not as
clear as the relationships between working capital items and sales. For example,
companies often build facilities with capacities greater than current needs; thus, sales
increases for a limited number of years can often be handled by current PP&E. In the
long run, however, no organization can increase sales indefinitely without increasing
capacity. When projecting necessary increases in capital items, management must look
closely at current facilities and the efficiency of operations in relation to projected sales.
Increases in PP&E are likely to be “lumpy” and occur as the firm reaches capacity
constraints.
e. Increases in long-term debt are closely tied to permanent increases in working capital
and PP&E. For example, higher sales must be supported by a higher level of assets
(current and long term). Some of this support comes from the spontaneous credit just
discussed, and any remaining shortfalls must come from additional short-term debt,
long-term debt, or stock issues. Management must decide whether to use debt or
equity for additional financing needs. Note that additional interest-bearing financing will
change the interest expense on the income statement.
f. Projected retained earnings (RE) is current RE plus forecasted net income less projected
dividends to be paid.
b. In the example that will be used, management has made the decision that no new stock
will be sold, and the only increase in equity will be the result of additional retained
earnings. In addition, no dividends will be paid in the coming year due to large planned
fixed-asset expansion. It has been determined that additional debt can be borrowed at
the same rate as the current rate (10%). In order to determine the additional
borrowings necessary, an iteration approach (a guess) can be employed. Once an
estimate of the possible ending LT debt amount is made, the interest expense, EBT, tax,
EAT, and ending equity can be calculated. Obviously, the goal is to have the ending total
assets equal the total liabilities and equity. It is assumed that the interest expense for
the year will be calculated upon the average debt held for the year. A “guess” of the
ending LT debt of $13,000 provides the following results:
(1) Interest = 10% average debt held during year = 10% (($11,000 + $13,000) ÷ 2)
= $1,200
(2) EBT = EBIT – Interest expense = $8,000 – $1,200 = $6,800
(3) Tax = 40% of EBT = 40% × $6,800 = $2,720
(4) EAT = EBT – Tax = $6,800 – $2,720 = $4,080
(5) Equity = Beginning equity + EAT = $12,000 + $4,080 = $16,080
Income Statement Balance Sheet
Projected Beginning Ending
Revenue $100,000 Assets
COGS 80,000 Current $ 8,000 $10,000
Gross Margin 20,000 Fixed 20,000 25,000
Expenses 12,000 Total Assets $28,000 $35,000
EBIT 8,000
Interest 1,200 Liabilities & Equity
EBT 6,800 Current $ 5,000 $ 6,000
Tax 2,720 LT Debt 11,000 13,000
EAT $ 4,080 Equity 12,000 16,080
Total L&E $28,000 $35.080
Since total asset do not equal liabilities and equity, the iteration using a LT debt amount
of $13,000 is incorrect.
c. A series of iterations will potentially be necessary in order to find the ending LT debt
amount that will result in a “balanced” balance sheet. Generally, only a few iterations
are necessary in order to achieve the necessary results. In fact, the use of a spreadsheet
can facilitate the entire iteration process. If the total liabilities and equity are more than
total assets, then a lower LT debt amount can be tried. If the total liabilities and equity
are more than total assets, then a higher LT debt amount can be tried. As part of the
planning process, management will then seriously consider whether the planned
expansion is feasible in one year or whether plans needs to be scaled back based upon
the ability to obtain the necessary debt financing. In the example, the use of an ending
LT debt level of $12,918 brings the projected balance sheet into balance.
(1) Interest = 10% average debt held during year = 10% (($11,000 + $12,918) ÷ 2)
= $1,196
(2) EBT = EBIT – Interest expense = $8,000 – $1,196 = $6,804
(3) Tax = 40% of EBT = 40% × $6,804 = $2,722
(4) EAT = EBT – Tax = $6,804 – $2,722 = $4,082
(5) Equity = Beginning equity + EAT = $12,000 + $4,082 = $16,082
Projections
annual depreciation is also included in the change in the beginning and ending net PP&E
figures. Gross investment in operating for a period can be summarized as follows:
Gross investment in operating capital = Net investment for the period +
depreciation and amortization
d. Free cash flow can be expressed in an equation form as follows:
Free cash flow = Operating cash flow - Gross investment in operating capital
If operating cash flow and gross investment in operating capital equations are
substituted in the above formula, the result is:
Free cash flow = NOPAT - Net investment in operating capital
e. When planning, the calculation of free cash flow can be an important analytical tool.
With the planned level of capital expenditures for expansion, will there be enough cash
available for debt and equity investors? The results for a single year are not important in
isolation. This is an equation that draws importance over time and allows a firm to
answer the question as to whether an expansion today can be expected to provide the
cash flow necessary to satisfy investors in the future.
This ratio can be very helpful when determining estimates of funds that will be necessary to
support increased sales volumes.
Where:
Assets0
Required increase in assets = × Change in sales
Sales0
Liabilities0
Spontaneous increase in assets = × Change in sales
Sales0
Actual sales
Full capacity sales =
Percentage of capacity used to create sales
For example, if sales were at $500,000 and 80% of capacity was used to produce those
sales, then the amount that sales could increase before expansion could be estimated as
follows.
$500,000
Full capacity sales = = $625,000
80%
Thus, sales could increase by $125,000 ($625,000 at full capacity - $500,000 at 80%
capacity) before plant expansion would be necessary.
b. This information can also be used to project the necessary increase in production
capacity for a given sales level given the following relationship:
Actual fixed assets Targeted fixed assets
=
Full capacity sales Projected sales
Using the above example, if $200,000 of production assets would be needed to support
$625,000 of full capacity sales, then what would the target fixed assets be for $750,000
of projected sales?
In reality, indicating a need to increase assets by $40,000 may not be absolutely correct
since additions to production assets might not show a linear relationship due to
increased efficiency of new equipment. Management’s judgment may be essential when
estimating the actual additional assets that would be required and determining the
timing of those additions.
It is expected that 10% of the sales for each month will be cash sales. Of the remaining
sales, 15% will be collected the month of the sale, 50% the month following the sale,
25% the second month following the sale, and the remaining 10% in the third month.
Projected total sales for January are $500,000. What are the projected cash inflows for
the month of January?
Solution:
Cash Inflows
10% of January sales of $500,000 as cash sales $ 50,000
15% of January credit sales of $450,000 ($500,000 × 90%) 67,500
50% of December credit sales of $630,000 ($700,000 × 90%) 315,000
25% of November credit sales of $495,000 ($550,000 × 90%) 123,750
10% of October credit sales of $405,000 ($450,000 × 90%) 40,500
Total projected cash inflows in January $596,750
d. Other anticipated cash receipts from interest, dividends, sale of marketable securities or
other investments, proceeds from the sale of property, plant, and equipment, and
proceeds from the company’s own stock issues are also listed as cash receipts. Many of
these inflows are discretionary and result from decisions made by management.
e. The cash disbursements section itemizes all of the cash outflows except those due to
financing that are expected during the period. Major sources for outflows are expected
payments for materials (merchandise), labor, overhead, selling, and administrative
expenses. Again, the various operating expense budgets do not represent the flow of
cash due to the use of the accrual method of accounting. Each listed expense category
needs to be converted from an accrual to a cash basis in order to project the related
expected cash outflows. The disbursement section of the cash budget is a bridge
between the beginning balance sheet and the projected ending balance sheet. This can
be summarized by:
Beginning Accounts Payable Balance
+ Projected Purchases on Account
– Projected Cash Payments on Account
= Projected Ending Accounts Payable Balance
In order to determine the projected cash payments, it is necessary to know the credit
terms as well as the payment philosophy of the organization. Are obligations paid within
the credit terms? Are discounts generally taken?
Illustration: The Johnson Company adheres to credit terms in the payment of accounts
payable of either 2/10, net 30 (when offered) or simply net 30. Discounts of 2% are
offered on 20% of all purchases. The remaining 80% have terms net 30. Assume that
purchases are made evenly throughout the month. Purchases of $310,000 are projected
to be made in March and $360,000 during the month of April. What are the projected
cash disbursements for April?
Solution:
Average daily purchases for March = $310,000 ÷ 31 = $10,000
Average daily purchase for April = $360,000 ÷ 30 = $12,000
Projected
April Cash Disbursement Items Disbursement
80% of March purchases with net 30 terms ($310,000 × 80%) $248,000
Since payments taking discounts are paid on the 10th day, 9 days of
March purchases qualifying for discounts are outstanding on April 1.
[($10,000 × 20%) × 9 days] × 98% 17,640
Since payments taking discount are paid on the 10th day, 21 days of April
purchases qualifying for discounts are paid in April. [($12,000 × 20%) ×
21 days] × 98% 49,392
Total projected April disbursement for Accounts Payable $315,032
The outstanding accounts payable at the end of April could also be calculated in the
following manner:
80% of the April purchases with the terms net 30 will be
outstanding on April 30 ($360,000 × 80%). $288,000
The remaining portion of the $72,000 of 2/10, net 30
purchases will be outstanding at the end of the month
($72,000 – $50,400). 21,600
Total Accounts Payable at the end of April $309,600
f. Other disbursements such as capital purchases, dividend payments, stock repurchases,
purchases of investments, or other various disbursements other than for financing are
also listed as cash outflows. Many of these items are discretionary payments and can
vary from period to period.
g. The cash inflows before financing are added to the beginning cash balance for the
period from which the cash outflows before financing are deducted to arrive at the
excess cash (or deficit) available.
Beginning Cash Balance
+ Projected Cash Receipts Other Than Financing
– Projected Cash Disbursement Other Than Financing
= Excess (Deficit) Cash Available Over Disbursements
h. The financing section summarizes the borrowing, repayments, and cash interest
payments deemed to be necessary during the period. Most organizations have a target
cash balance that is the desired cash assumed to be necessary to safely conduct
business. If the excess (deficit) cash is less than the target cash balance and required
interest payments, then it is necessary to finance the cash needs of the organization. If
there is an amount in excess of the target cash balance after interest payments at the
end of the projected period, previous financing can be repaid or the funds can be
invested.
The cash inflows before financing are added to the beginning cash balance for the
period from which the cash outflows before financing are deducted to arrive at the
excess cash (or deficit) available:
Excess (Deficit) Cash Available Over Disbursements
+ New Borrowing
– Repayments
– Interest Payments
= Target Cash Balance
Illustration: A firm has an opening cash balance of $25,000 and outstanding long-term
financing (due in three years) of $60,000 at a 10% interest rate at the beginning of the
year. It is projected that cash receipts will equal $750,000 and that cash disbursements
will equal $739,250. The desired target cash balance at the end of the year is $32,000.
Will the company have to borrow funds in order to meet the desired target cash
balance? Assume that all additional financing will be acquired at the beginning of the
period at a 10% rate and that all repayments will be at the end of the period.
Solution:
Beginning Cash Balance $ 25,000
+ Projected Cash Receipts 750,000
Projected Cash Disbursements (739,250)
= Excess (Deficit) Cash Available 35,750
Projected Additional Financing ?
Projected Repayments ?
Projected Interest Payments ?
Projected Ending Cash Balance $ 32,000
If there is no additional financing, then $6,000 of interest payments ($60,000 × 10%) will
be paid out, leaving a balance less than the desired target cash balance of $32,000.
(Projected excess cash available of $35,750 less projected interest payments of $6,000 is
$29,750.) Therefore, additional financing will be necessary. The equation needed to
determine the additional cash needed is as follows:
Projected excess funds + Projected additional financing – 10% (Original financing +
Projected additional financing) = Target cash balance
$35,750 + Pro. add’l financing – 10% ($60,000 + Pro. add’l financing) = $32,000
Solving for the project additional financing, it can be determined that an additional
$2,500 will need to be borrowed. The projected interest payments will be $6,250 [10%
($60,000 + $2,500)].
Beginning Cash Balance $ 25,000
+ Projected Cash Receipts 750,000
Projected Cash Disbursements (739,250)
= Excess (Deficit) Cash Available 35,750
Projected Additional Financing 2,500
Projected Repayments 0
Projected Interest Payments (6,250)
Projected Ending Cash Balance $ 32,000
i. Most organizations have a master cash budget that is then broken down into monthly or
even weekly mini-budgets due to the fact that cash flows and disbursements do not
occur evenly throughout the year. The need for financing and/or the repayment of prior
financing will often vary from month to month depending upon the business cycle. The
target cash balance may also be varied during the year.
j. Certain items such as depreciation add complexity to the cash budget. Even though
depreciation does not require the use of cash, it does have an effect on taxable income
and ultimately the taxes paid.
Using the above example, the cash flow of the proposed project can be calculated:
Cash flow = [$100,000 - ($65,000 + $15,000)] × (1 - 0.40) + ($15,000 × 0.40)
= $18,000
The use of this formula does not include any financing expenses since the manner of the
financing is not relevant when evaluating/comparing projects for acceptance. When
using a discounted cash flow method of evaluation, the cost of capital is taken into
account; therefore, if financing costs were included in the projected cash flows that are
then discounted, the financing costs would be double counted. After a project is
potentially accepted, the best financing method can then be chosen.
5542.39 Probability analysis: Most management decisions involve uncertainty. Uncertainty means
that decisions must be made before knowing which of a set of possible future events will
occur. Possible events are called states of nature. In decision making under uncertainty, the
decision maker must choose a course of action before knowing which state of nature will
occur. A prerequisite to making rational decisions under uncertainty is having some measure
of the likelihood of occurrence of each possible state of nature. Such a measure is called
probability and the development and use of probabilities is called probability analysis.
5542.40 Developing probabilities: The probability of an event P(m) is computed by dividing the
number of elements in the subset n(m) by the number of elements in the universal set n(u).
n(m)
P(m) =
n(u)
Example: Historical data shows that the last 10 units sold consisted of 8 units of model t and
2 units of model s. What is the probability that model t will be the next unit sold?
8
P(t) = = 80%
10
The development of probabilities is complicated when dealing with joint events (events that
occur together), additive events (a collection of events such that any or all events can occur
concurrently), and conditional probability (the probability of a particular event being
contingent on another event).
Example: Suppose we are considering a capital budgeting decision to buy new equipment,
but we need to know the expected payoff to the company if it is purchased. Our random
variable is the expected cash flow, and the probability associated with each outcome is
shown in the following probability distribution:
Expected Cash Flow Probability Distribution
$10,000 .10
15,000 .20
20,000 .50
30,000 .20
What is the expected value?
Solution:
E(x) = ($10,000 × 0.10) + ($15,000 × 0.20) + ($20,000 × 0.50) + ($30,000 × 0.20)
= $20,000
Keep in mind that, if the machine is acquired, the actual cash flow will probably not be
$20,000. However, if this decision is repeated many times and if the probability distribution
is accurate, on the average, the payoff will be $20,000 per year for each machine.
5542.41 Variance: The variance of a random variable (x) shows the dispersion of the observations. It
is defined as the sum of the squared deviations of x from the expected value, weighted by
their probability.
A small variance indicates that the outcomes are closely clustered about the expected value.
5542.42 Standard deviation: The standard deviation is frequently reported in place of the variance.
The standard deviation of x is the square root of the variance of x.
5542.43 Payoff table: A decision maker is frequently faced with several decision alternatives and
several possible states of nature. A state of nature is an uncontrollable event that will impact
a decision. Each state of nature has a probability associated with its occurrence. The
probabilities associated with the various states of nature must total 1.0.
A payoff table is developed by showing the states of nature in the column headings along
with the probability associated with each. The decision alternatives are shown as row
headings. Within each square of the matrix, the payoff associated with that decision
alternative and state of nature is inserted.
Example: Management is deciding whether or not to build a new wing on the factory. The
decision alternatives are “build” and “not build.” There are three possible states of nature:
improvement in the economy, economy stays the same, and economy deteriorates. The
probability of each state is 0.20, 0.50, and 0.30, respectively. The various payoffs are shown
in the matrix. If, for example, the wing is built and the economy improves, the payoff will be
a $10,000 increase in profits per month. If it is built and the economy gets worse, profits will
decrease by $5,000 per month.
There are several decision rules that can be used; the most common is expected value. The
expected value for each decision alternative is computed (by multiplying the probability of
each state of nature by the payoff for that state of nature and summing for each decision
alternative), and the alternative with the highest expected value is selected.
Breakeven Analysis
5542.44 Breakeven (also called cost-volume-profit (CVP) analysis is a technique to evaluate the
relationship between costs, volume of activity, and profit. It is useful in providing answers to
the following types of questions:
a. At what sales volume will we break even?
b. What must the sales volume be to earn $X profit?
c. How will profits be affected if the sales price is changed by $Y per unit?
d. What is the effect on the breakeven point if variable costs are converted to fixed costs
by increased mechanization?
5542.45 The key factors in cost-volume-profit analysis are revenues, fixed costs, and variable costs.
Accountants generally assume that both revenues and costs are linear. As long as revenue
per unit exceeds variable cost per unit, there will be a point at which the firm will break
even. The breakeven point is where total revenues equal total costs. Above (below) this
point there will be a profit (loss). The following diagram illustrates these relationships:
5542.46 Contribution margin per unit is the sales price less variable cost per unit. If the sales price
exceeds the variable cost, the contribution margin is positive and each unit is providing
something to cover fixed cost and generate a profit. A breakeven point is reached as soon as
the cumulative contribution margin from all units equals the fixed cost. Expressed in
equation form:
FC
Breakeven units =
(P - VC)
where: FC = Total fixed costs
P= Price per unit
VC = Variable cost per unit
5542.47 Questions on the CPA Examination often ask for breakeven stated in sales dollars. This can
be obtained by computing breakeven in units and multiplying it by the sales price, or it can
be computed directly as follows:
FC
Breakeven sales dollars =
P - VC
P
The computation (P - VC) ÷ P is called the contribution margin ratio.
5542.48 If the problem is to determine the number of units or sales volume required to generate a
profit, the amount of desired profit is added to the fixed cost in the equations shown.
Additional units must be sold to accumulate enough contribution margin to cover both the
fixed cost and the desired profit. Note that this is profit before income tax. Variable cost per
unit does not include any income tax.
FC + NI FC + NI
Units = Sales dollars =
(P - VC) (P - VC) ÷ P
Where: NI = Net income before income taxes
5542.49 If the desired profit is specified as an after-tax amount, the after-tax amount must be
adjusted to a before-tax amount for inclusion in the equations. To get before-tax profit,
divide the after-tax profit by one minus the tax rate:
After-tax NI
Before-tax NI =
1 – Tax rate
5542.50 Cost-volume-profit analysis, as described, relies on several assumptions built into it. If these
are not valid, the results will be inaccurate. Assumptions that underlie cost-volume-profit
analysis include the following:
a. Costs can be classified as either fixed or variable.
b. Total variable costs change at a linear rate. (Variable costs per unit remain unchanged.)
c. Total fixed costs remain unchanged over the relevant range of the breakeven chart.
(Fixed costs per unit change.)
d. Sales price does not change with changes in volume.
e. There is only one product and, if there are multiple products, the sales mix remains
constant.
f. Productive efficiency does not change.
g. Inventories are kept constant or at zero.
h. Volume is the only relevant factor affecting costs.
5542.51 The results of cost-volume-profit analysis are frequently illustrated in a profit-volume graph.
As the name implies, a profit-volume (P/V) graph focuses entirely on the relationship
between volume and profit. It shows profit directly rather than as the difference between
the revenue and cost curves as on a breakeven chart (see section 5542.45).
The vertical axis represents net income (net loss), and the horizontal axis represents volume.
The vertical axis is extended below the origin to show a net loss. The measure of volume can
be either units of production or sales dollars. The profit (loss) line is plotted by computing
the profit (loss) at two levels of output. The two points most commonly used are zero
production level and breakeven point. At an output of zero, the loss will be the fixed costs.
The slope of the profit line is equal to the contribution margin. It slopes upward to the right
and crosses the horizontal axis at the breakeven point.
5542.52 Illustration: Two examples follow to illustrate the computations involved in breakeven
analysis (adapted from the AICPA).
Example: The Oliver Company plans to market a new product. Based on its market studies,
Oliver estimates that it can sell 5,500 units in 20X1. The selling price will be $2.00 per unit.
Variable costs are estimated to be 40% of the selling price. Fixed costs are estimated to be
$6,000. What is the breakeven point?
Solution:
Variable cost per unit = 40% × $2.00 = $0.80
$6,000
Breakeven units = = 5,000 units
$2.00 - $0.80
$6,000
Breakeven sales dollars = = $10,000
($2.00 - $0.80) ÷ $2.00
Example: The Brieden Company sells rodaks for $6.00 per unit. Variable costs are $2.00 per
unit. Fixed costs are $37,500. How many rodaks must be sold to realize a profit before
income taxes of 15% of sales?
Solution:
Breakeven in sales ($) dollars:
$37,500 + $0.15(S)
S= = $72,580
($6.00 - $2.00) ÷ $6.00
$72,580
Breakeven in units: = 12,097 units
$6
5542.53 Inventory decision models are designed to answer two basic questions.
1. What quantity of an item should be ordered or produced?
2. How often should an item be ordered or produced?
Information on demand, cost, lead times, and variability are required by the models.
5542.54 Information on demand, cost, lead times, and variability are required by the models.
a. Demand: Most inventory models assume that there is a stable and uniform demand
over the period. Demand in the economic order quantity (EOQ) model is specified in
number of units per year.
b. Costs: The relevant costs in the EOQ model are (a) annual costs of carrying one unit in
stock, (b) ordering or setup cost to purchase or produce the item, and (c) shortage costs.
The purchase or production cost is not relevant because it will not change with the
quantity ordered or produced. The annual carrying cost includes imputed interest on the
investment in inventory, storage, insurance, obsolescence, spoilage, and theft. Ordering
or setup costs include preparing for a production run, preparing an order, evaluating
bids, and receiving and paying vendor’s invoices. Shortage costs include cost of lost
sales, customer dissatisfaction, and lost customers.
5542.55 Safety stock: An additional quantity of inventory that is held to cover periods in which
demand is greater than normal. The size of the safety stock is based on the degree of
variability in demand (or standard deviation) and the acceptable risk of being out of stock.
5542.56 Cost relationships: Inventory decision models work with opposing costs. Carrying costs
increase as the size of the order increases. Setup or ordering costs, however, decrease as the
size of the production run or order increases. The EOQ model seeks to minimize the
combined total of these two costs, as illustrated in the following graph:
5542.57 Fixed quantity system: The fixed quantity system calculates a fixed order quantity (Q) that is
used whenever an order is placed. The order is placed at the reorder point (RP), which is
dependent on the lead time (LT) and safety stock (SS). The amount of the order is fixed, but
the time of order placement varies from period to period, depending on demand. This
system has two fixed parameters—order quantity and reorder point. Whenever the
inventory level drops below the reorder point, an order is placed for the order quantity (Q).
These relationships are illustrated as follows:
5542.58 Order quantity computation: The formula for computing the economic order quantity (Q) is:
5542.59 Reorder point computation: Assuming that there is constant demand and lead time, the
reorder point is the expected units of demand during lead time plus any required safety
stock. This relationship can be illustrated as follows:
5542.60 Example: The following data will be used to illustrate the computation of order quantity and
order point.
Demand (D) = 40,000 units per year
Unit cost (C) = $50
Carrying cost rate (i) = 20% per year
Order cost (S) = $115 per order
Lead time (LT) = 3 days (assume 300 business days per year)
Safety stock (SS) = 500 units
Solution:
Reorder point:
RP = ((3 ÷ 300) × 40,000) + 500
= 900 units
In this example, when the stock of inventory is reduced to 900 units, an order will be placed
for 960 units.
Relevant Costs
individual project approach, the total costs or revenues of each project are first computed,
and the project with the lowest cost or highest revenue is selected. When the incremental
approach is used, the incremental or added costs to move to the alternative proposal are
subtracted from the incremental revenue provided by it. If the result is positive, it is
considered a desirable alternative.
5542.62 Sunk costs
Critical in management decision making is the understanding that sunk (past) costs are
irrelevant. Relevant costs for decision making are expected future costs that will differ
among alternatives available to the firm. Historical costs are useful only as they help predict
the future.
a. The book value of old equipment is not relevant since it is an historic cost. Nothing can
change what has already been spent.
b. The current disposal price of old equipment is relevant since expected future cash flows
will differ among alternatives (replacing vs. keeping).
c. The gain or loss on sale (or disposal) of equipment is irrelevant to decision making since
this is merely an algebraic difference between book value and disposal price. Note,
however, that any tax consequences resulting from the gain or loss from the sale would
be relevant.
d. The cost of new equipment is relevant since expected cash outflows will differ between
alternatives.
e. Fixed costs that will not change among alternatives are irrelevant. Fixed costs that will
be eliminated or added due to the project under consideration are relevant.
f. Variable costs are relevant if they differ among alternatives (per unit cost) or if
production volume is changed (change in total cost).
g. The timing of expected future costs and revenues is relevant to decision making due to
the time value of money. The use of discounted cash flow methods will take timing
issues into consideration.
h. Illustration: An uninsured machine costing $50,000 is destroyed the first day. There are
two alternatives: (a) dispose of it for $5,000 and replace it with another machine for
$51,000 or (b) rebuild it to original state for $45,000. Which alternative should be
selected?
Solution: Rebuild, based on the following analysis. The $50,000 original cost is a sunk
cost and irrelevant.
(a) New $51,000 (b) Rebuild $45,000
Less Salvage 5,000
Net Cost $46,000
In this situation, it would cost less to rebuild the equipment than to replace it.
Decision-Making Process
5542.63 a. Steps to use in the decision-making process:
(1) Define the problem. Replace equipment, make or buy a component part, expand or
contract, accept a special order, etc.
(2) Determine possible alternatives and compare each to the firm’s goals for
congruence.
(3) Prepare estimates of costs/revenues, benefits, and risks for each alternative that is
complementary to the firm’s goals.
(4) Identify and estimate possible constraints such as availability of funds, capacity,
availability of raw materials, and skilled labor.
(5) Select the alternative that will most likely maximize the present value of the
company’s future cash flows without violating any of the constraints.
b. Since cash is the most important aspect of running a business, decision making leans
toward defining cost and benefits in terms of cash flows—positive, negative, or both.
Since the choice of one action will potentially prevent another from being taken, the
decision should be based on perceived differences among the alternatives being
considered—the incremental approach. When future cash is affected, the PV of the
incremental cash flows should be used.
c. An opportunity cost is a net cash inflow that will be lost if the course of action under
consideration is pursued. An action that eliminates a cash inflow is equivalent to an
action that produces a cash outflow.
Illustration: An overstock of 5,000 widgets could be sold as a special promotion at $100
each. If the action is rejected, the incremental cash inflow of ($500,000) represents the
opportunity cost of not selling the widgets.
d. Any past cash outflow (or committed future cash outflow) that is not affected by the
decision to pursue the course of action under consideration is a sunk cost to the
decision at hand. In the example used above, the cost of the widgets would be a sunk
cost.
Some managers would choose not to replace an old machine with current depreciable
book value greater than zero due to the fact that they do not want to recognize a loss
on disposal. “Depreciation expense” over the next few years is often considered to be
more appealing than a “loss on disposal.” For many, there appears to be a psychological
issue regarding the appearance of “loss on disposal” on the income statement even
though it would be in the best interest of the organization to replace current production
assets with new ones.
e. Pitfalls of relevant costing:
(1) Assuming that all variable costs are relevant. For example, normal variable
marketing costs may not be relevant for a special order, or there might be
incremental marketing costs associated with the proposed alternative.
(2) Assuming that all fixed costs are irrelevant. If a proposal would require operating
outside the relevant range, fixed costs could increase or decrease. When purchasing
new equipment, fixed costs are likely to change.
(3) The use of unit costs can be misleading. Original unit costs could contain a
combination of both relevant and irrelevant costs.
point. If incremental revenue exceeds incremental cost, the additional processing will always
be profitable.
5542.65 Illustration
An organization has three joint products (copper, silver, lead) made from copper ore that
can each be processed further after the split-off point or sold without further processing.
The question is which sales point—with or without further processing—is most beneficial to
the company. Data regarding overall profitability is shown below if all products are
processed further after the split-off point.
Final Sales Revenue $500,000
Less Cost of Producing End Products
Cost of Copper Ore $200,000
Cost of Separating Ore 60,000
Cost of Further Processing 150,000 410,000
Profit $ 90,000
As can be seen, the company is making a profit; however, the question is whether or not the
organization is maximizing profit by the further processing of all three joint products or
could increase profits by further processing only one or two of the joint products.
Solution: In order to make the best choice, an incremental approach needs to be used. Are
the incremental revenues of further processing greater than the incremental costs of that
processing? Each product needs to be examined individually in order to determine whether
further processing of that particular product is maximizing revenues. This could be done as
follows:
Copper Silver Lead
Final Sales Revenue—After Processing $150,000 $260,000 $90,000
Less Sales Value at Split-Off Point 100,000 180,000 70,000
Incremental Revenue Provided by Further Processing 50,000 80,000 20,000
Less Cost of Further Processing 40,000 95,000 15,000
Profit (Loss) from Further Processing $ 10,000 ($ 15,000) $ 5,000
As can be seen in this instance, the company would increase profits by $15,000 if the silver
ore is sold at the split-off point and not processed further; however, the other two products,
copper and lead, should continue with further processing.
Make or Buy
5542.67 Illustration
a. A company currently makes a component part that another firm has offered to provide
for $250 per unit. Currently 1,000 units are produced at the following costs:
Direct Materials $ 50,000
Direct Labor 75,000
Manufacturing Overhead
Variable 30,000
Fixed 80,000
Total Costs for 1,000 units $235,000 or $235 per unit
Additionally, if the part was purchased, fixed costs of $25,000 would be eliminated and
the current production space could be rented for $30,000 a year. Should the company
make or buy the part?
Solution:
The opportunity costs associated with not renting the space are additional costs added
to the cost of making the part due to the “lost opportunity.” The relevant fixed costs are
those that will differ among the alternatives; therefore, the fixed costs that would
continue under either circumstance need to be removed from the cost of producing the
product in order to compare the cost of making the product to that of buying the
component part. In this scenario, it would be $40,000 more expensive for the company
to purchase the part than to make it.
b. Solution using relevant cost approach:
Total
Costs Per Unit Cost
Original cost to produce 1,000 units (from the illustration) $235,000 $235
Lost opportunity cost for not renting space 30,000 30
Nonrelevant fixed costs (55,000) (55)
d. Outsourcing is the purchasing of goods and services as opposed to producing the goods
and services internally (insourcing). The make or buy decision shown above would be
equally valid for services—human resources, accounting, maintenance, etc. Often the
most important elements of a make/buy decision are nonquantifiable variables such as
quality or dependability.
Keep or Replace
5542.69 Illustration
A company is considering replacing an older piece of equipment with a more efficient
machine. Cost of Capital is 10%. Available data is as follows:
Old Equipment New Equipment
Original Cost $500,000 $600,000
Useful Life 10 years 5 years
Remaining Life 5 years 5 years
Accumulated Depreciation 250,000 N/A
Current Disposal Value 175,000 N/A
Disposal Value in 5 Years 0 0
Annual Revenues 400,000 400,000
Annual Operating Costs 150,000 50,000
Average Tax Rate 40% 40%
(3) Analysis of the cash (outflows) and inflows for the five-year period is:
Keep Replace Difference
Total Operating Costs Outflow
($150,000 × 5) ($750,000) ($750,000)
($50,000 × 5) ($250,000) (250,000)
(500,000)
Disposal Value 0 175,000 (175,000)
Cost of New Machine 0 (600,000) 600,000
Total Relevant Costs ($750,000) ($675,000) ($ 75,000)
Given this analysis, it would appear that replacing the equipment would save the
company $75,000 before taxes due to lower annual operational cost outflows. For this
case, the alternative with the least amount of cash outflows is the most desirable.
Solution with tax and PV considerations:
(1) In this analysis, the assumption is made that all tax payments occur at the end of
the period as well as revenues and related costs.
(2) The gain/loss is not relevant in itself; however, the tax consequences of that event
are relevant.
(3) Annual cash flows:
Keep Replace
Revenues $400,000 $400,000
Cash Operating Expenses (150,000) (50,000)
Depreciation (50,000) (120,000)
Taxable Income 200,000 230,000
Tax (40%) 80,000 92,000
Net Income 120,000 138,000
Add Back Depreciation 50,000 120,000
Annual Cash Flow 170,000 258,000
PV of Annual Cash Flow * $644,436 $978,026
5542.70 Pricing of special orders: Occasionally, a request is received to manufacture a product and
sell it below the normal sales price. It is generally assumed that the special order will not
encroach on existing sales. If it does, the order is generally rejected because the regular price
is established to provide long-run profitability based on normal demand for the product. If
the assumption is valid, profitability can be increased by accepting the special order if the
price per unit exceeds the variable costs. Fixed costs normally allocated to all units produced
should be excluded from the decision because they will be covered by normal sales at the
normal price.
be examined item by item in order to determine what costs could be eliminated and
what costs will continue.
c. Illustration: The Gentry Company currently has two divisions that had operating results
as follows for last year.
Since Division 2 sustained an operating loss during the year, the president is considering
the elimination of this division. Assume that the direct fixed costs of Division 2 could be
avoided if the division is eliminated. Should the division be eliminated?
Solution: If Division 2 is eliminated, all of the variable costs for the division as well as the
direct fixed costs for the division would also be eliminated. However, the allocated
corporate costs would not be eliminated; thus, total corporate income would decrease
by $20,000 ($30,000 of current net income - $20,000 of lost division margin). Currently
Division 2 contributes $20,000 toward covering the corporate costs. Division 2 should be
continued until some other more profitable manner to absorb the $20,000 in allocated
corporate costs can be found.
5542.73 Expansion
The decision-making process when considering adding a customer, division, or business
segment is similar to the disposal decision. As long as the revenues for the new segment are
larger than the directly related costs of that segment, a positive contribution will be made to
the organization as a whole, thus increasing profits. The fact that a portion of the current
corporate-office costs will be allocated to the new segment is irrelevant since these costs
existed prior to the consideration of the new segment. Only revenues and costs that will
change as a result of the expansion need to be considered.
Constrained Resources
Solution: Analyzing the contribution margin per machine hours yields the following
results:
Regular Deluxe
CM per Machine Hour $125 $110
(Calculation) ($250 ÷ 2) ($330 ÷ 3)
Contribution Available for
2,000 Assembly Hours $250,000 $220,000
(Calculation) ($125 × 2,000) ($110 × 2,000)
Due to the assembly constraint (and assuming sufficient consumer demand), the regular
racing bike should be produced in order to maximize profits. The contribution margin
per machine hour for each model is the relevant item in this case—not the contribution
margins of the separate models.
c. Capacity can possibly be increased within the constraint in the short run by:
(1) working overtime at the constraint, providing the CM minus the increased wages
due to overtime is still greater than zero.
(2) subcontracting some of the constraint area output. Again, the cost must be less
than the benefit.
d. Capacity can possibly be increased within the constraint in the long run by:
(1) reengineering the process in the constraint area.
(2) investing in additional capacity.