Tag Archives: measures of regulation

The Economics of Regulation Part 3: How to Estimate the Effect of Regulatory Accumulation on the Economy? Exploring Endogenous Growth and Other Models

This post is the third part in a three part series spurred by a recent study by economists John Dawson and John Seater that estimates that the accumulation of federal regulation has slowed economic growth in the US by about 2% annually.  The first part discussed generally how Dawson and Seater’s study and other investigations into the consequences of regulation are important because they highlight the cumulative drag of our regulatory system. The second part went into detail on some of the ways that economists measure regulation, highlighting the strengths and weaknesses of each.  This post – the final one in the series – looks at how those measures of regulation are used to estimate the consequences of regulatory policy.  As always, economists do it with models.  In the case of Dawson and Seater, they appeal to a well-established family of endogenous growth models built upon the foundational principle of creative destruction, in the tradition of Joseph Schumpeter.

So, what is an endogenous growth model?

First, a brief discussion of models:  In a social or hard science, the ideal model is one that is useful (applicable to the real world using observable inputs to predict outcomes of interest), testable (predictions can be tested with observed outcomes), flexible (able to adapt to a wide variety of input data), and tractable (not too cumbersome to work with).  Suppose a map predicts that following a certain route will lead to a certain location.  When you follow that route in the real world, if you do not actually end up at the predicted location, you will probably stop using that map.  Same thing with models: if a model does a good job at predicting real world outcomes, then it sticks around until someone invents one that does an even better job.  If it doesn’t predict things well, then it usually gets abandoned quickly.

Economists have been obsessed with modeling the growth of national economies at least since Nobel prize winner Simon Kuznets began exploring how to measure GDP in the 1930s.  Growth models generally refer to models that try to represent how the scale of an economy, using metrics such as GDP, grows over time.  For a long time, economists relied on neoclassical growth models, which primarily use capital accumulation, population growth, technology, and productivity as the main explanatory factors in predicting the economic growth of a country. One of the first and most famous of such economic growth models is the Solow model, which has a one-to-one (simple) mapping from increasing levels of the accumulated stock of capital to increasing levels of GDP.  In the Solow model, GDP does not increase at the same rate as capital accumulation due to the diminishing marginal returns to capital.  Even though the Solow model was a breakthrough in describing the growth of GDP from capital stock accumulation, most factors in this growth process (and, generally speaking, in the growth processes of other models in the neoclassical family of growth models) are generated by economic decisions that are outside of the model. As a result, these factors are dubbed exogenous, as opposed to endogenous factors which are generated inside of the model as a result of the economic decisions made by the actors being modeled.

Much of the research into growth modeling over the subsequent decades following Solow’s breakthrough has been dedicated to trying to “endogenize” those exogenous forces (i.e. move them inside the model). For instance, a major accomplishment was endogenizing the savings rate – how much of household income was saved and invested in expanding firms’ capital stocks. Even with this endogenous savings rate, as well as exogenous growth in the population providing labor for production, the accumulating capital stocks in these neoclassical growth models could not explain all of the growth in GDP. The difference, called the Solow Residual, was interpreted as the growth in productivity due to technological development and was like manna from heaven for the actors in the economy – exogenously growing over time regardless of the decisions made by the actors in the model.

But it should be fairly obvious that decisions we make today can affect our future productivity through technological development, and not just through the accumulation of capital stocks or population growth. Technological development is not free. It is the result of someone’s decision to invest in developing technologies. Because technological development is the endogenous result of an economic decision, it can be affected by any factors that distort the incentives involved in such investment decisions (e.g., taxes and regulations). 

This is the primary improvement of endogenous growth theory over neoclassical growth models.  Endogenous growth models take into account the idea that innovative firms invest in both capital and technology, which has the aggregate effect of moving out the entire production possibilities curve.  Further, policies such as increasing regulatory restrictions or changing tax rates will affect the incentives and abilities of people in the economy to innovate and produce.  The Dawson and Seater study relies on a model originally developed by Pietro Peretto to examine the effects of taxes on economic growth.  Dawson and Seater adapt the model to include regulation as another endogenous variable, although they do not formally model the exact mechanism by which regulation affects investment choices in the same way as taxes.  Nonetheless, it’s perfectly feasible that regulation does affect investment, and, to a degree, it is simply an empirical question of how much.

So, now that you at least know that Dawson and Seater selected an accepted and feasible model—a model that, like a good map, makes reliable predictions about real world outcomes—you’re surely asking how that model provided empirical evidence of regulation’s effect on economic growth.  The answer depends on what empirical means.  Consider a much better established model: gravity.  A simple model of gravity states that an object in a vacuum near the Earth’s surface will accelerate towards the Earth at 9.81 meters per second squared. On other planets, that number may be higher or lower, depending on the planet’s massiveness and the object’s distance from the center of the planet.  In this analogy, consider taxes the equivalent of mass – we know from previous endogenous growth models that taxes have a fairly known effect on the economy, just like we know that mass has a known effect on the rate of acceleration from gravitational forces.  Dawson and Seater have effectively said that regulations must have a similar effect on the economy as taxes.  Maybe the coefficient isn’t 9.81, but the generalized model will allow them to estimate what that coefficient is – so long as they can measure the “mass” equivalent of regulation and control for “distance.”  They had to rely on the model, in fact, to produce the counterfactual, or to use a term from science experiments, a control group.  If you know that mass affects acceleration at some given constant, then you can figure out what acceleration is for a different level of mass without actually observing it.  Similarly, if you know that regulations affect economic growth in some established pattern, then you can deduce what economic growth would be without regulations.  Dawson and Seater appealed to an endogenous growth model (courtesy of Perreto) to simulate a counterfactual economy that maintained regulation levels seen in the year 1949.  By the year 2005, that counterfactual economy had become considerably larger than the actual economy – the one in which we’ve seen regulation increase to include over 1,000,000 restrictions.

The Economics of Regulation Part 2: Quantifying Regulation

I recently wrote about a new study from economists John Dawson and John Seater that shows that federal regulations have slowed economic growth in the US by an average of 2% per year.  The study was novel and important enough from my perspective that it deserved some detailed coverage.  In this post, which is part two of a three part series (part one here), I go into some detail on the various ways that economists measure regulation.  This will help put into context the measure that Dawson and Seater used, which is the main innovation of their study.  The third part of the series will discuss the endogenous growth model in which they used their new measure of regulation to estimate its effect on economic growth.

From the macroeconomic perspective, the main policy interventions—that is, instruments wielded in a way to change individual or firm behavior—used by governments are taxes and regulations.  Others might include spending/deficit spending and monetary policy in that list, but a large percentage of economics studies on interventions intended to change behavior have focused on taxes, for one simple reason: taxes are relatively easy to quantify.  As a result, we know a lot more about taxes than we do about regulations, even if much of that knowledge is not well implemented.  Economists can calculate changes to marginal tax rates caused by specific policies, and by simultaneously tracking outcomes such as changes in tax revenue and the behavior of taxed and untaxed groups, deduce specific numbers with which to characterize the consequences of those taxation policies.  In short, with taxes, you have specific dollar values or percentages to work with. With regulations, not so much.

In fact, the actual burden of regulation is notoriously hidden, especially when directly compared to taxes that attempt to achieve the same policy objective.  For example, since fuel economy regulations (called Corporate Average Fuel Economy, or CAFE, standards) were first implemented in the 1970s, it has been broadly recognized that the goal of reducing gasoline consumption could be more efficiently achieved through a gasoline tax rather than vehicle design or performance standards.  However, it is much easier for a politician to tell her constituents that she will make auto manufacturers build more fuel-efficient cars than to tell constituents that they now face higher gasoline prices because of a fuel tax.  In econospeak, taxes are salient to voters—remembered as important and costly—whereas regulations are not. Even when comparing taxes to taxes, some, such as property taxes, are apparently more salient than others, such as payroll taxes, as this recent study shows.  If some taxes that workers pay on a regular basis are relatively unnoticed, how much easier is it to hide a tax in the form of a regulation?  Indeed, it is arguably because regulations are uniquely opaque as policy instruments that all presidents since Jimmy Carter have required some form of benefit-cost analysis on new regulations prior to their enactment (note, however, that the average quality of those analyses is astonishingly low).  Of course, it is for these same obfuscatory qualities that politicians seem to prefer regulations to taxes.

Despite the inherent difficulty, scholars have been analyzing the consequences of regulation for decades, leading to a fairly large literature. Studies typically examine the causal effect of a unique regulation or a small collection of related regulations, such as air quality standards stemming from the Clean Air Act.  Compared to the thousands of actual regulations that are in effect, the regulation typically studied is relatively limited in scope, even if its effects can be far-reaching.  Because most studies on regulation focus only on one or perhaps a few specific regulations, there is a lot of room for more research to be done.  Specifically, improved metrics of regulation, especially metrics that can be used either in multi-industry microeconomic studies or in macroeconomic contexts, could help advance our understanding of the overall effect of all regulations.

With that goal in mind, some attempts have been made to more comprehensively measure regulation through the use of surveys and legal studies.  The most famous example is probably the Doing Business index from the World Bank, while perhaps the most widely used in academic studies is the Indicators of Product Market Regulation from the OECD.  Since 2003, the World Bank has produced the Doing Business Index, which combines survey data with observational data into a single number designed to tell how much it would cost to “do business,” e.g. set up a company, get construction permits, get electricity, register property, etc., in set of 185 countries.  The Doing Business index is perhaps most useful for identifying good practices to follow in early to middle stages of economic development, when property rights and other beneficial institutions can be created and strengthened.

The OECD’s Indicators of Product Market Regulation database focuses more narrowly on types of regulation that are more relevant to developed economies.  Specifically, the original OECD data considered only product market and employment protection regulations, both of which are measured at “economy-wide” level—meaning the OECD measured whether those types of regulations existed in a given country, regardless of whether they were applicable to only certain individuals or particular industries.  The OECD later extended the data by adding barriers to entry, public ownership, vertical integration, market structure, and price controls for a small subset of broadly defined industries (gas, electricity, post, telecommunications, passenger air transport, railways, and road freight).  The OECD develops its database by surveying government officials in several countries and aggregating their responses, with weightings, into several indexes.

By design, the OECD and Doing Business approaches do a good job of relating obscure macroeconomic data to actual people and businesses.  Consider the chart below, taken from the OECD description of how the Product Market Regulation database is created.  As I wrote last week and as the chart shows, the rather sanitized term “product market regulation” actually consists of several components that are directly relevant to a would-be entrepreneur (such as the opacity of a country’s licenses and permits system and administrative burdens for sole proprietorships) and to a consumer (such as price controls and barriers to foreign direct investment).  You can click on the chart below to see some of the other components that are considered in OECD’s product market regulation indicator.

oecd product regulation tree structure

Still, there are two major shortcomings of the OECD data (shortcomings that are equally applicable to similar indexes produced by the World Bank and others).  First, they cover relatively short time spans.  Changes in regulatory policy often require several years, if not decades, to implement, so the results of these changes may not be reflected in short time frames (to a degree, this can be overcome by measuring regulation for several different countries or different industries, so that results of different policies can be compared across countries or industries).

Second, and in my mind, more importantly, the Doing Business Index is not comprehensive.  Instead, it is focused on a few areas of regulation, and then only on whether regulations exist—not how complex or burdensome they are.  As Dawson and Seater explain:

[M]easures of regulation [such as the Doing Business Index and the OECD Indicators] generally proceed by constructing indices based on binary indicators of whether or not various kinds of regulation exist, assigning a value of 1 to each type of regulation that exists and a 0 to those that do not exist.  The index then is constructed as a weighted sum of all the binary indicators.  Such measures capture the existence of given types of regulation but cannot capture their extent or complexity.

Dawson and Seater go out of their way to mention at least twice that the OECD dataset ignores environmental and occupational health and safety regulations.  Theirs is a good point – in the US, at least, environmental regulations from the EPA alone accounted for about 15% of all restrictions published in federal regulations in 2010, and that percentage has consistently grown for the past decade, as can be seen in the graph below (created using data from RegData).  Occupational health and safety regulations take up a significant portion of the regulatory code as well.

env regs as percentage of total

In contrast, one could measure all federal regulations, not just a few select types.  But then the process requires some usage of the actual legal texts containing regulations.  There have been a few attempts to create all-inclusive time series measures of regulation based on the voluminous legal documents detailing regulatory activity at the federal level.   For the most part, studies have relied on the Federal Register, the government’s daily journal of newly proposed and final regulations.  For example, many scholars have counted pages in the Federal Register to test for the existence of the midnight regulations phenomenon—the observation that the administrations of outgoing presidents seem to produce abnormally large numbers of regulations during the lame-duck period

There are problems with using the Federal Register to measure regulation (I say this despite having used it in some of my own papers).  First and foremost, the Federal Register includes deregulatory activity.  When a regulatory agency eliminates words, paragraphs, or even entire chapters from the CFR, the agency has to notify the public of the changes.  The agency does this by printing a notice of proposed rulemaking in the Federal Register that explains the agencies intentions.  Then, once the public has had adequate time to comment on the agencies proposed actions, the agency has to publish a final rule in the Federal Register—another set of pages that detail the final actions the agency is taking.  Obviously, if one is counting pages published in the Federal Register and using that as a proxy for the growth of regulation, deregulatory activity that produces positive page counts would lead to incorrect measurements.  

Furthermore, pages published in the Federal Register may be a biased measure because the number of pages associated with individual rulemakings has increased over time as acts of Congress or executive orders have required more analyses. In his Ten-Thousand Commandments series, Wayne Crews mitigates this drawback to some degree by focusing only on pages devoted to final rules.  The Ten-Thousand Commandments series keeps track of both the annual number of final regulations published in the Federal Register and the annual number of Federal Register pages devoted to final regulations.

Dawson and Seater instead rely on the Code of Federal Regulations, another set of legal documents related to federal regulationsActually, the CFR would be better described as the books that contain the actual text of regulations in effect each year.  When a regulatory agency creates new regulations, or alters existing regulations, those changes are reflected in the next publication of the CFR.  Dawson and Seater collected data on the total number of pages in the CFR in each year from 1949 to 2005. I’ve graphed their data below.

dawson and seater cfr pages

*Dawson and Seater exclude Titles 1 – 3 and 32 from their total page counts because they argue that those Titles do not contain regulation, so comparing this graph with page count graphs produced elsewhere will show some discrepancies.

Perhaps the most significant advantage of the CFR over counting pages in the Federal Register is that it allows for decreases in regulations. However, using the CFR arguably has several advantages over indexes like the OECD product market regulation index and the World Bank Doing Business index.  First, using the CFR captures all federal regulation, not just a select few types.  Dawson and Seater point out:

Incomplete coverage leads to two problems: (1) omitted variables bias, and, in any time series study, (2) divergence between the time series behavior of subsets of regulation on the one hand and of total regulation on the other.

In other words, ignoring potentially important variables (such as environmental regulations) can cause estimates of the effect of regulation to be wrong.

Second, the number of pages in the CFR may reflect the complexity of regulations to some degree.  In contrast, the index metrics of regulation typically only consider whether a regulation exists—a binary variable equal to 1 or 0, with nothing in between.  Third, the CFR offers a long time series – almost three times as long as the OECD index, although it is shorter than the Federal Register time series.

Of course, there are downsides to using the CFR.  For one, it is possible that legal drafting standards and language norms have changed over the 57 years, which could introduce bias to their measure (Dawson and Seater brush this concern aside, but not convincingly in my opinion).  Second, the CFR is limited to only one country—the United States—whereas the OECD and World Bank products cover many countries.  Data on multiple countries (or multiple industries within a country, like RegData offers) allow comparisons of real-world outcomes and how they respond to different regulatory treatments.  In contrast, Dawson and Seater are limited to constructing a “counterfactual” economy – one that their model predicts would exist had regulations stayed at the level they were in 1949.  In my next post, I’ll go into more detail on the model they use to do this.

The Economics of Regulation Part 1: A New Study Shows That Regulatory Accumulation Hurts the Economy

In June, John Dawson and John Seater, economists at Appalachian State University and North Carolina State University, respectively, published a potentially important study (ungated version here) in the Journal of Economic Growth that shows the effects of regulatory accumulation on the US economy.  Several others have already summarized the study’s results (two examples here and here) with respect to how the accumulation of federal regulation caused substantial reductions in the growth rate of GDP.  So, while the results are important, I won’t dwell on them here.  The short summary is this: using a new measure of federal regulation in an endogenous growth model, Dawson and Seater find that, on average, federal regulation reduced economic growth in the US by about 2% annually in the period from 1949 to 2005.  Considering that economic growth is an exponential process, an average reduction of 2% over 57 years makes a big difference.  A relevant excerpt tells just how big of a difference:

 We can convert the reduction in output caused by regulation to more tangible terms by computing the dollar value of the loss involved.  […] In 2011, nominal GDP was $15.1 trillion.  Had regulation remained at its 1949 level, current GDP would have been about $53.9 trillion, an increase of $38.8 trillion.  With about 140 million households and 300 million people, an annual loss of $38.8 trillion converts to about $277,100 per household and $129,300 per person.

These are large numbers, but in fact they aren’t much different from what a bevy of previous studies have found about the effects of regulation.  The key differences between this study and most previous studies are the method of measuring regulation and the model used to estimate regulation’s effect on economic growth and total factor productivity.

In a multi-part series, I will focus on the tools that allowed Dawson and Seater to produce this study: 1. A new time series measure of total federal regulation, and 2. Models of endogenous growth.  My next post will go into detail on Dawson and Seater’s new time series measure of regulation, and compares it to other metrics that have been used.  Then I’ll follow up with a post discussing endogenous growth models, which consider that policy decisions can affect the accumulation of knowledge and the rates of innovation and entrepreneurship in an economy, and through these mechanisms affect economic growth.

Why should you care about something as obscure as a “time series measure of regulation” and “endogenous growth theory?”  Regulations—a form of law that lawyers call administrative law—create a hidden tax.  When the Department of Transportation creates new regulations that mandate that cars must become more fuel efficient, all cars become more expensive, in the same way that a tax on cars would make them more expensive.  Even worse, the accumulation of regulations over time stifle innovation, hinder entrepreneurship, and create unintended consequences by altering the prices of everyday purchases and activities.  For an example of hindering entrepreneurship, occupational licensing requirements in 17 states make it illegal for someone to braid hair for a living without first being licensed, a process which, in Pennsylvania at least, requires 300 hours of training, at least a 10th grade education, and passing a practical and a theory exam. Oh, and after you’ve paid for all that training, you still have to pay for a license.

And for an example of unintended consequences: Transportation Security Administration procedures in airports obviously slow down travel.  So now you have to leave work or home 30 minutes or even an hour earlier than you would have otherwise, and you lose the chance to spend another hour with your family or finishing some important project.  Furthermore, because of increased travel times when flying, some people choose to drive instead of fly.  Because driving involves a higher risk of accident and death than does flying, this shift, caused by regulation, of travelers from plane to car actually causes people to die (statistically speaking), as this paper showed.

Economists have realized the accumulation of regulation must be causing serious problems in the economy.  As a result, they have been trying to measure regulation in different ways, in order to include regulation in their models and better study its impact.  One famous measure of regulation, which I’ll discuss in more detail in my next post, is the OECD’s index of Product Market Regulation.  That rather sanitized term, “product market regulation,” actually consists of several components that are directly relevant to a would-be entrepreneur (such as the opacity of a country’s licenses and permits system and administrative burdens for sole proprietorships) and to a consumer (such as price controls, which can lead to shortages like we often see after hurricanes where anti-price gouging laws exist, and barriers to foreign direct investment, which could prevent multinational firms like Toyota from building a new facility and creating new jobs in a country).  But as you’ll see in the next post, that OECD measure (and many other measures) of regulation miss a lot of regulations that also directly affect every individual and business.  In any science, correct measurement is a necessary first step to empirical hypothesis testing.

Dawson and Seater have contributed a new measure of regulation that improves upon previously existing ones in many ways, although it also has its drawbacks.  And because their new measure of regulation offers many more years of observations than most other measures, it can be used in an endogenous growth model to estimate how regulation has affected the growth of the US economy.  Again, in endogenous growth models, policy decisions (such as how much regulation to create) affect economic growth if they affect the rates of accumulation of knowledge, innovation, and entrepreneurship. It’s by using their measure in an endogenous growth model that Dawson and Seater were able to estimate that individuals in the US would have been $129,300 richer if regulations had stayed at their 1949 level.  I’ll explain a bit more about endogenous growth theory in a second follow-up post.  But first things first—my next post will go into detail on measures of regulation and Dawson and Seater’s innovation.