John Locke: Law Is Only Legitimate When It Is Founded on the Law of Nature

Link to the video above on YouTube of Gilbert and Sullivan's Mikado singing about making the punishment fit the crime

In section 12 of his 2d Treatise on Government: On Civil Government," John Locke makes two key points. The first is that a punishment should be just severe enough "as will suffice to make it an ill bargain to the offender, give him cause to repent, and terrify others from doing the like":

By the same reason may a man in the state of nature punish the lesser breaches of that law. It will perhaps be demanded, with death? I answer, each transgression may be punished to that degree, and with so much severity, as will suffice to make it an ill bargain to the offender, give him cause to repent, and terrify others from doing the like.

I wrote about how this principle of a punishment just severe enough to deter applies could apply to civil law in "Reparation and Deterrence."

The second key point is that to be legitimate, laws must be founded on the underlying morality of the law of nature.

Every offence, that can be committed in the state of nature, may in the state of nature be also punished equally, and as far forth as it may, in a commonwealth: for though it would be besides my present purpose, to enter here into the particulars of the law of nature, or its measures of punishment; yet, it is certain there is such a law, and that too, as intelligible and plain to a rational creature, and a studier of that law, as the positive laws of commonwealths; nay, possibly plainer; as much as reason is easier to be understood, than the fancies and intricate contrivances of men, following contrary and hidden interests put into words; for so truly are a great part of the municipal laws of countries, which are only so far right, as they are founded on the law of nature, by which they are to be regulated and interpreted.

I marvel at the number of people I run across who think there is some magic in democracy that makes everything that 51% of people vote for good and right—or in any case legitimate to impose on everyone else. Some people are more circumspect and argue that the large supermajorities needed to pass a constitutional amendment are necessary to make something good and right—or at least legitimate to impose on everyone else. But it is not so. Something can be wrong to impose on people even if a large supermajority wants to.

I do not mean to claim that it is always clear what the law of nature allows and what it doesn't. Where some freedom is compromised to raise overall social welfare or where overall social welfare is compromised for the sake of freedom, I don't always know what is legitimate and what isn't, let alone what is the right thing to do. But there is one case where I know something is illegitimate and contrary to Locke's law of nature: when it both compromises freedom and lowers overall social welfare in order to benefit some at the expense of others. 

There are many such laws that both compromise freedom and lower overall social welfare in order to benefit some people at the expense of others. I mention a few in 'Keep the Riffraff Out!'" but attentive readers will notice how this is a theme throughout this entire blog. 

On the underlying principle that democracy doesn't necessarily make things right, see "Democracy is Not Freedom" and "John Locke: The Right to Enforce the Law of Nature Does Not Depend on Any Social Contract." Also, I remember being inspired by the Milton Friedman video below on "The Proper Role of Government" to write the gloss "Legally formalizing ethical rules is different from other, more arbitrary legislation" in my post "Milton Friedman: Celebrating His 100th Birthday with Videos of Milton."

Economics Needs to Tackle All of the Big Questions in the Social Sciences

The Economist's Free Exchange column on January 14, 2017, reporting on the American Social Science Association meetings that month argues "To be relevant, economists need to take politics into account." For example:

Their theories had always shown that globalisation would produce losers as well as winners. But too many economists worried that emphasising these costs might undermine support for liberal policies. A “circle the wagons” approach to criticism of globalisation weakened the case for mitigating policies that might have protected it from a Trumpian backlash. Perhaps the greatest omissions were the questions not asked at all. Most dismal scientists exclude politics from their models altogether. As Joseph Stiglitz, a Nobel laureate, put it on one star-studded AEA panel, economists need to pay attention not just to what is theoretically feasible but also to “what is likely to happen given how the political system works”.

But I worry that people reading this will think a bandaid is enough. Indeed, the author advises halfway measures:

Political and social institutions are much harder to model and quantify than commodity or labour markets. But a qualitative approach might actually be far more scientific than equations offering little guide to how the future will unfold.

What is needed is a full-fledged research program that does the hard work of modeling and quantifying. In order for economics to influence policy in the way I believe it should, we need to take all of social science as our area of concern, not a limited slice of the social sciences.

What is economics? In my view, economics should encompass every important social science question that those with training in economics are well suited to study. If economists have a comparative advantage in investigating an issue, that issue should be part of economics. 

That doesn't mean the economists should ignore the work done by other social scientists, but neither should they be overly deferential to that work. Social scientists outside of economics have turned over many stones and know many things. Economists need to absorb the key bits of that knowledge. And the best scholars in any social science field are smarter than mediocre economists. But in many cases, economists who are not dogmatic can learn about social science questions outside their normal purview and see theoretical and statistical angles to studying those questions that others have not. 

Because of the great practical need for a deep understanding of the social sciences, I consider it a cardinal sin for any economist to discourage another economist from trying to tackle an important social science questions because it is "not economics." No! If it is an important social science question, and the economist who wants to study it can make a real contribution, it is economics.  

That is, for practical reasons of saving the world, social science understanding outside the traditional domain of economics is so desperately needed that we should not hold fellow economists back. In addition, concerns in the traditional domains of different social sciences interact with one another, so it makes sense for economists to go there.

Linguistics is the social science that seems the furthest from traditional economist to me. But even the border between Linguistics and the humanities is of great relevance to core economic questions, as the description in the column of Robert Shiller's talk indicates: 

In a keynote address, Robert Shiller—a Nobel prizewinner, habitual freethinker and outgoing AEA president—suggested that economists should think more broadly about the factors that affect human behaviour. Narratives matter, he argued. Powerful ideas, captured in memorable stories, can spread like epidemics, wreaking economic havoc as they go.

If we understood what makes narratives striking and memorable, and what makes people accept them as a basis for making decisions, we would understand both politics and movements in the stock market better.

George Lakoff in particular has done a good job of applying linguistics to understanding politics. The psychologist Jonathan Haidt is operating in similar territory in his book The Righteous Mind. Economists who want to be more politically relevant would do well to read their work. 

The model I am putting forward is one in which the different social science disciplines act as different schools of thought for addressing social science questions rather than any social science claiming a monopoly on a particular question. This goes both ways. On questions within the traditional domain of economics, the group that with a different perspective that must be taken quite seriously are business school professors and business practicioners. For example, economists should know the work of Clay Christensen. I have written these posts about his work:

  1. Clay Christensen, Jeffrey Flier and Vineeta Vijayaraghavan on How to Make Health Care More Cost Effective
  2. Saint Clay
  3. Clay Christensen, Jerome Grossman and Jason Hwang on the Agenda for the Transformation of Health Care
  4. Clay Christensen, Jerome Grossman and Jason Hwang on How the History of Other Industries Gives Hope for Health Care
  5. Clay Christensen, Jerome Grossman and Jason Hwang on Intuitive Medicine vs. Precision Medicine
  6. Clay Christensen, Jerome Grossman and Jason Hwang on the Personal Computer Revolution
  7. Clay Christensen, Jerome Grossman and Jason Hwang on the Three Basic Types of Business Models
  8. Clay Christensen, Jerome Grossman and Jason Hwang: How to Divide and Conquer Our Health Care Problems
  9. In Defense of Clay Christensen: Even the ‘Nicest Man Ever to Lecture’ at Harvard Can’t Innovate without Upsetting a Few People

Despite the efforts of all the human social scientists that have ever lived put together, we don't understand politics all that well. But we will. And free-range minds will ultimately bring us to a better understanding of many other social-science questions as well. 

Update: This New York Times essay by Neil Irwin (and flagged by Justin Wolfers) talks about some of the questions and concerns of sociologists that economists need to bring into their work:

What if Sociologists Had as Much Influence as Economists?

Neil Irwin talks about sociologists becoming more influential. But it is much easier for economists to incorporate insights from sociology into their work than for sociologists to become as influential as economists. For one thing, there are probably about 10 times as many economists in the world as sociologists, judging from association membership and so on that I had to look up once for a Templeton Foundation grant proposal I did with three sociologists. (I have two unpublished papers coauthored with sociologists.) 

Don't miss the follow-up post "Defining Economics."

 

The Volcker Shock

"October 6, 1979, was a chilly Saturday in Washington. The coming Monday was a government holiday, Columbus Day, and much of official Washington had scattered for the long weekend. Many of those who remained, along with the news media, were keeping an eye on Pope John Paul II, who was paying the first-ever papal visit to the White House to meet with President Carter; he would lead an open-air mass at the foot of the US Capitol the following day. With almost everyone’s attention elsewhere, it was a good day for a secret meeting at the Federal Reserve. ...

After a full day of discussion, the central bankers agreed on a plan. As Chairman Paul Volcker told reporters at an unusual press conference that night, the Federal Reserve would stop trying to stabilize prices by adjusting short-term interest rates. Instead, it would target the total amount of reserves held by the thousands of banks in the Federal Reserve system. 'By emphasizing the supply of reserves and constraining the growth of the money supply through the reserve mechanism, we think we can get firmer control over the growth in money supply in a shorter period of time,' Volcker intoned. Not one in a thousand Americans could explain what that meant. But the message got through to Wall Street, where traders dissect every word of every utterance by every Fed official. Adding up banks’ nonborrowed reserves was one of many ways to measure the nation’s money supply. By making reserves its main gauge, the Federal Reserve, like the Bank of England four months earlier, was embracing monetarism.

Neither Volcker nor any other policymaker at the US central bank was a committed believer in mechanically regulating the money supply as the monetarists counseled. Their responsibility, as all of them saw it, was to receive a stream of data and anecdotal reports, evaluate them to assess the state of the economy, and then adjust monetary policy accordingly. The October 6 announcement, known forever after as the 'Volcker shock,' seemed to eliminate the Fed’s discretion to make those month-to-month adjustments. Henceforth the central bank would be bound by an ironclad rule governing how fast the money supply should grow.

But that was not really Volcker’s intention. By appearing to put monetary policy on autopilot, the Fed was trying to sweep away two political obstacles to its goal of lowering inflation. It hoped to blunt the ceaseless attacks of its most vociferous critics, the influential monetarist economists and their allies at places like The Wall Street Journal, who harped constantly on the Fed’s erratic policies. If perchance their Fed-bashing turned to praise, perhaps the financial markets would believe that inflation would be coming down. If that occurred, interest rates would fall, and lower interest rates on mortgages and business loans might in fact help bring inflation down. The Fed also hoped its new stance would shield it from the political assaults that were sure to come. Quelling inflation, which was running at a 12 percent rate, previously experienced only when wartime price controls were removed, seemed likely to require much higher interest rates than the United States had ever known.

If the Fed openly made interest rates the target of its policy, announcing that it was raising short-term rates to 15 or 20 percent, then auto dealers, construction workers, and corporate executives would cry foul and enraged members of Congress might strip the central bank of its independence. If, however, high interest rates were merely the byproduct of its much-praised shift to the monetary policy rules the monetarists were demanding, the Fed would have some protection from its critics. As Volcker put it to his colleagues at that Saturday meeting, 'It’s an easier political sale.'”

—Marc Levinson, An Extraordinary Time: The End of the Postwar Boom and the Return of the Ordinary Economy, chapter 13. 

One for All: John Woodland "Jack" Welch on Edward Lawrence Kimball

Jack Welch was one of my Dad's colleagues at Brigham Young University's J. Reuben Clark Law School. Jack combined his work as a law professor with very interesting apologetics of Mormonism's historical claims--especially Mormonism's claims about ancient history. Indeed, Jack Welch's Foundation for Ancient Research and Mormon Studies brought about one of the most notable cases of the views of Mormon officialdom being changed by intellectual inquiry: a shift from and emphasis on the Book of Mormon as an ancient history of the Americas as a whole to an acceptance of the idea that the action in the Book of Mormon might be primarily confined to ancient Mesoamerica.

My Dad always greatly respected Jack. Recognizing that, my siblings and I chose Jack to talk about my Dad's professional activities at the memorial service on December 3, 2016. Jack was good enough to give me permission to post a revised version of his tribute to my Dad. (Also see my own tribute and those of my brothers Chris and Joseph and of my sisters Paula, Mary and Sarah.) 

From more than one discussion with my Dad, I know that an important part of my Dad's belief that the Book of Mormon is what it claims to be was based on Jack's work on Chiasmus in the Book of Mormon, which you can find here. Jack's work on "The Legal Cases in the Book of Mormon" is also very interesting.

Here is Jack's tribute to my Dad:  


I am humbled to add my sincere words of love and commendation at this service for Ed. I am so glad I knew Edward L. Kimball, a valued colleague and a happy friend. 

All of us old-timers at the Law School remember Ed with great respect and appreciation as a voice of profound wisdom in dealing with hard rules and difficult cases, as a powerful minute-taker in faculty meetings, as a joyous singer at law school Christmas parties, and a meticulous legal scholar, a very effective teacher, and a pioneer on the new Multi-State Bar Examination Committee of the National Conference of Bar Examiners. In 1973, Ed was the first faculty member to commit to Rex Lee to join the newly emerging J. Reuben Clark Law School. Gordon Smith, now Dean of the BYU Law School, who came to BYU from the same University of Wisconsin Law School where Ed had taught for ten years, remembers how Ed’s great legacy was still felt there 30 years after Ed had left for Provo.

I, like so many other people, always found much to admire in Ed. I wish I had written a monthly letter to each of my ward missionaries, as Ed did as a bishop. One of his students described Ed well: “Professor Kimball is a scholarly man and an extremely good man . . . motivated by a desire to serve others through the Church and the law.” Who else would have—or could have—served for over a decade on the Utah Board of Pardons at Point of the Mountain?  

As the editor of BYU Studies, where Ed had served for five years on the editorial board, I worked closely with him on three of his major publications. He was always sensitive, well informed, careful, precise, and punctual. Ed will be best known for the well-documented 2-volume biography of his father and also the one volume on Camilla, his mother. But let me mention another important but overlooked contribution, namely his lengthy article on confession (published in BYU Studies, vol. 36, no. 2), an essential step for all of us in repentance. I recommend this article to everyone in the Church. Typical of Ed’s work, it is scripturally grounded and enriched by comparing LDS confessional teachings with the confessional practices of those of other faiths. It is seasoned by Ed's own growth as one who heard confessions and took those sincere opportunities to aid, encourage, and restore the lost confidence of the penitent. For Ed, scholarship was never just academic. Indeed, his article on confession grew out of his legal interest in the laws of evidence relating to the priest-penitent privilege, especially as that privilege of non-disclosure might relate to confessions made to a bishop or cleric in cases involving child abuse.

While Ed always laid a rigorous foundation for his wise and truth-loving statements, he also sensitively sought to make his academic wisdom useful in strengthening the lives of his family, friends and fellow saints. This desire came naturally for Ed. For him, the reconciliation of faith and intellect traveled across the bridge of usefulness. Ed never allowed faith to diminish his rigor. Rather, his faith drove him to be sure that that knowledge was thoroughly developed and flawlessly well-grounded, as rigorously reliable as possible.

Now I wish to say something to the children here and also to the young in heart. You probably know the thirteen Articles of Faith of the Church of Jesus Christ of Latter-Day Saints:

  1. We believe in God, the Eternal Father, and in His Son, Jesus Christ, and in the Holy Ghost.
  2. We believe that men will be punished for their own sins, and not for Adam's transgression.
  3. We believe that through the Atonement of Christ, all mankind may be saved, by obedience to the laws and ordinances of the Gospel.
  4. We believe that the first principles and ordinances of the Gospel are: first, Faith in the Lord Jesus Christ; second, Repentance; third, Baptism by immersion for the remission of sins; fourth, Laying on of hands for the gift of the Holy Ghost.
  5. We believe that a man must be called of God, by "prophecy, and by the laying on of hands" by those who are in authority, to preach the Gospel and administer in the ordinances thereof.
  6. We believe in the same organization that existed in the Primitive Church, namely, apostles, prophets, pastors, teachers, evangelists,  and so forth.
  7. We believe in the gift of tongues, prophecy, revelation, visions, healing, interpretation of tongues, and so forth.
  8. We believe the Bible to be the word of God as far as it is translated correctly; we also believe the Book of Mormon to be the word of God.
  9. We believe all that God has revealed, all that He does now reveal, and we believe that He will yet reveal many great and important things pertaining to the Kingdom of God.
  10. We believe in the literal gathering of Israel and in the restoration of the Ten Tribes; that Zion (the New Jerusalem) will be built upon this, the American continent; that Christ will reign personally upon the earth; and, that the earth will be renewed and receive its paradisiacal glory.
  11. We claim the privilege of worshiping Almighty God according to the dictates of our own conscience, and allow all men the same privilege, let them worship how, where, or what they may.
  12. We believe in being subject to kings, presidents, rulers, and magistrates, in obeying, honoring, and sustaining the law.
  13. We believe in being honest, true, chaste, benevolent, virtuous, and in doing good to all men; indeed, we may say that we follow the admonition of Paul—"We believe all things, we hope all things," we have endured many things, and hope to be able to endure all things. If there is anything virtuous, lovely, or of good report or praise worthy praiseworthy, we seek after these things.

But did you know that the word “all” appears eight times in the Articles of Faith? Let’s look at these and see how they describe Ed’s faith, as well as your own.

1. Ed believed that “through the Atonement of Jesus Christ all mankind may be saved by obedience to the laws and ordinances of the Gospel” (Articles of Faith 3). And thus Ed embraced all mankind, from every nation, tongue and people. He empathized with anyone who felt awkward or inadequate in any way. He served very effectively at the Law School on the Faculty-Student Diversity committee. 

2. Next, “We believe all that God has revealed” (A of F 9), and Ed did too. His testimony about the Book of Mormon is printed at the end of today’s program. The Book of Mormon itself stood for him as a persuasive artifact. He and I talked often about the evidences of the miraculously short time in which Joseph Smith translated the Book of Mormon and how it contains such elegant chiastic literary patterns. The power this book spoke to his soul led him to believe the Book of Mormon as something God has revealed.

3. Ed also believed “all that God does now reveal” (A of F 9), through continuing revelation. His monumental chapters and article on his father’s 1978 priesthood revelation (BYU Studies, vol. 47, no. 2) takes readers into the struggle and joy of real revelation. Ed believed that revelation happens, that it needs to be sought diligently, that it does not always come easily, and thus should never be taken lightly. 

4. Ed claimed “the privilege of worshiping God according to the dictates of his own conscience and allowed all men the same” (A of F 11). Ed was passionate about freedom of religion, free agency, consequences, and accountability, knowing that God will be just and merciful to all. 

5. He believed, as we believe, “in being honest and in doing good to all men” (A of F 13). Ed mastered the art of doing acts of kindness to each life he touched.

The 6th and 7th “alls” state: “We believe all things. We hope all things” (A of F 13). With his open, inquisitive mind, Ed believed all things. Filled with the love of Christ, one can hope that all things are possible with God. As Ed often said, the words in Ether 12:6, “dispute not because ye see not, for ye receive no witness until after the trial of your faith,” gave him the brightness of hope that he and we would someday hear and see the unimaginable things which God has prepared for all who love him (1 Cor. 2:9). It was that belief and hope that allowed Ed to keep putting one foot ahead of the other.

And finally, number 8, Ed successfully endured many things, and because of that he hoped “to endure all things” (A of F 13). He sought after everything that was of good report or praiseworthy. He wanted all: all the truth, and nothing but the truth, and he endured faithfully to the end.

So, as you repeat these and all the Articles of Faith, I hope they will always remind all of you of your father, your grandfather, and your friend Ed.

And now to you seven family members who have spoken today, thank you. Your words stand as seven seals on Ed’s book of life. And because Ed and Bee were sealed by virtue of the holy priesthood in the Temple, you are all sealed to them and with each other. What a blessing. I hope you all appreciate all that it means to be born in the covenant.  

I have the privilege of sealing children to parents. As I was performing sealings a month ago, I was moved to tell the following to one of the couples there, who was expecting to welcome a baby into their family soon: Be sure to teach that child what it means to be born in the covenant, to come into this life with a pure gift of self-worth and purpose, given by heavenly parents at the instant of mortal birth. It entitles you to possess and receive as equal heirs with all the children every blessing afforded by God’s everlasting covenant.

Because you are gratefully and forever the posterity of Ed and Bee Kimball, please know, and I testify, that this generously given privilege is eternally yours, through God’s unfailing plan and the effulgent grace of our Lord Jesus Christ. All this I witness in his holy name, and in the memory and honor of Ed Kimball, amen.

 

The Economist: The Coexistence of Pinyin and Chinese Characters Highlights the Role of Emotion in Language Decisions

I found this very interesting, even though I knew some of it already. Pinyin the system for using Roman letters to write Chinese, which is also used to telling computers what Chinese character to put down. The key passage is this:

Why don’t the Chinese just adopt pinyin? One is the many homophones (though these are not usually a problem in context). Another is that Chinese characters are used throughout the Chinese-speaking world, not just by Mandarin-speakers but also speakers of Cantonese, Shanghainese and other varieties. These are as different from each other as the big Romance languages are, but the writing system unifies the Chinese world. In fact, character-based writing is, in effect, written Mandarin. This is not obvious from looking at the characters, but it is obvious if you look at pinyin. If China adopted it wholesale, the linguistic divisions in China would be far more apparent.

But there is another reason for attachment to the characters. They represent tradition, history, literature, scholarship and even art on an emotional level ...

Next Generation Monetary Policy

Link to Wikipedia article on Star Trek: The Next Generation

Link to Wikipedia article on Star Trek: The Next Generation

On September 7, 2016, I gave a talk at the Mercatus Center conference on “Monetary Rules in a Post-Crisis World.” You can see it here:  Next Generation Monetary Policy: The Video. (And here is a recently revised version of the Powerpoint file.) I have written those ideas out more carefully in a paper I have submitted to the Journal of Macroeconomics (Update: now published in the Journal of Macroeconomics), which co-sponsored the conference. Here is a link to the paper as published that you can download. Below is an earlier version of the text of that working paper as a blog post. 

Abstract: This paper argues there is still a great deal of room for improvement in monetary policy. Sticking to interest rate rules, potential improvements include (1) eliminating any effective lower bound on interest rates, (2) tripling the coefficients in the Taylor rule, (3) reducing the penalty for changing directions, (4) reducing interest rate smoothing, (5) more attention to the output gap relative to the inflation gap, (6) more attention to durables prices, (7) mechanically adjusting for risk premia, (8) strengthening macroprudential measures to reduce the financial stability burden on interest rate policy, (9) providing more of a nominal anchor.


Monetary policy is far inside the production possibility frontier. We can do better. That is what I want to convince you of with this paper.

The worst danger we face in monetary policy during the next five years is complacency. I worry that central banks are patting themselves on the back because the world is finally crawling out of its business cycle hole. It is right to be grateful that things were not even worse in the aftermath of the Financial Crisis of 2008. But dangers abound, including (a) the likelihood that lesson that higher capital requirements are needed to avoid financial crises will be forgotten by key policy-makers, (b) the possibility of a dramatic collapse of Chinese real estate prices, and (c) the not-so-remote chance of a serious trade war.

A longer-term danger is that the rate of improvement in monetary policy will slow down as monetary policy researchers are drawn into simply justifying what central banks already happen to be doing. The assumption that people know what they are doing and are already following the best possible strategy may have an appropriate place in some areas of economics, but is especially inapt when applied to central banks. For one thing, the field of optimal monetary policy is very young, and its influence on central banks even more younger. For another, it is not easy to optimize in the face of complex political pressures and many self-interested actors eager to cloud one’s understanding with false worldviews. 

Advances in monetary policy cannot come from armchair theorizing alone. trying ideas out is the only way to make a fully convincing case that they work—or don’t. But in general, the value of experimentation in public policy is underrated. Because central banks are usually in a good position to reverse something they try that doesn’t work, they are in an especially good position to do policy experiments for which a good argument can be made even if it is not 100 percent certain the experiment will be successful. 

The Value of Interest Rate Rules

Because optimal monetary policy is still a work in progress, legislation that tied monetary policy to a specific rule would be a bad idea. But legislation requiring a central bank to choose some rule and to explain actions that deviate from that rule could be useful. To be precise, being required to choose a rule and explain deviations from it would be very helpful if the central bank did not hesitate to depart from the rule. Under this type of a rule, the emphasis is on the central bank explaining its actions. The point is not to directly constrain policy, but to force the central bank to approach monetary policy scientifically by noticing when it is departing from the rule it set itself and why.

So far, even without legislation, the Taylor Rule—essentially a description of the policy of the Fed under Volcker and Greenspan—has served this kind of useful function. Economists notice when the Fed departs from the Taylor Rule and talk about why.

The Taylor Rule has served another function: as a scaffolding for describing other words. Many important alternatives to the Taylor rule can readily be described as a modification of one of the Taylor Rule’s parameters or as the addition of an additional variable to the Taylor Rule.

In recent years, a shift toward purchasing assets other than short-term Treasury bills has complicated the specification of monetary policy. But Ricardo Reis, in his 2016 Jackson Hole presentation, argues that once the balance sheet is sufficiently large (about $1 trillion), it is again interest rates that matter after that. That is the perspective I take here.

What follows is, in effect, a research agenda for interest rate rules—a research agenda that will take many people to complete. The following sections discuss the following:

  1. eliminating the zero lower bound or any effective lower bound on interest rates
  2. tripling the coefficients in the Taylor rule
  3. reducing the penalty for changing directions
  4. reducing the presumption against moving more than 25 basis points at any given meeting
  5. a more equal balance between worrying about the output gap and worrying about fluctuations in inflation
  6. focusing on a price index that gives a greater weight to durables
  7. adjusting for risk premia
  8. pushing for strict enough leverage limits for financial firms that interest rate policy is freed up to focus on issues other than financial stability.
  9. having a nominal anchor.

Considering these possible developments in monetary policy is indeed only a research program, not a done deal. For each item individually, I have enough confidence research will ultimately vindicate it to be willing to state my views assertively. But though I don’t know which, it is likely that I have made some significant error in my thinking about at least one of the nine.

1.     Eliminating the Zero Lower Bound and Any Effective Lower Bound for Nominal Rates

Eliminating any lower bound on interest rates is important for making sure that interest rate rules can get the job done in monetary policy. In “Negative Interest Rate Policy as Conventional Monetary Policy,” in the posts I flag in “How and Why to Eliminate the Zero Lower Bound: A Reader’s Guide” and with Ruchir Agarwal, in “Breaking Through the Zero Lower Bound,” I arguethat eliminating any lower bound on interest rates is quite feasible, both technically and politically. Although this has not happened yet, the attitude of central bankers and economists toward negative interest rates and towards the paper currency policies needed to eliminate any lower bound on interest rates have shifted remarkably in the last five years. “Negative Interest Rate Policy as Conventional Monetary Policy” gave this list of milestones for negative interest rate policy up to November 2015:

1.     …. mild negative interest rates in the euro zone, Switzerland, Denmark and Sweden (see, for example, Randow, 2015).

2.     The boldly titled 18 May, 2015 London conference on ‘Removing the Zero Lower Bound on Interest Rates’, cosponsored by Imperial College Business School, the Brevan Howard Centre for Financial Analysis, the Centre for Economic Policy Research and the Swiss National Bank.

3.     The 19 May, 2015 Chief Economist’s Workshop at the Bank of England, which included keynote speeches by Ken Rogoff, presenting ‘Costs and Benefits to Phasing Out Paper Currency’ (Rogoff, 2014) and my presentation ‘18 Misconceptions about Eliminating the Zero Lower Bound’ (Kimball, 2015). …

4.     Bank of England Chief Economist Andrew Haldane’s 18 September, 2015 speech, ‘How low can you go?’ (Haldane, 2015).

5.     Ben Bernanke’s discussion of negative interest rate policy on his book tour (for Bernanke, 2015), ably reported by journalist Greg Robb in his Market Watch article ‘Fed officials seem ready to deploy negative rates in next crisis’ (Robb, 2015).

There have been at least four important milestones since then:

6.     Narayana Kocherlakota’s advocacy of a robust negative interest rate policy. (See Kocherlakota, February 1, February 9, June 9 and September 1, 2016.)

7.     The Brookings Institution’s June 6, 2016 conference “Negative interest rates: Lessons learned…so far(video available at https://www.brookings.edu/events/negative-interest-rates-lessons-learned-so-far/ )

8.     The publication of Ken Rogoff’s book, The Curse of Cash.

9.     Marvin Goodfriend’s 2016 Jackson Hole talk “The Case for Unencumbering Interest Rate Policy at the Zero Bound” and other discussion of negative interest rate policy at that conference.

10.  Ben Bernanke’s embrace of negative interest rates as a better alternative than raising the inflation target in his blog post “Modifying the Fed’s policy framework: Does a higher inflation target beat negative interest rates?”

In addition, I can testify to important shifts in attitudes based on the feedback from central bankers in presenting these ideas at central banks around the world since May 2013 and as recently as December 2016.[1] One objective indicator of this has been increased access by proponents of a robust negative interest rate policy to high-ranking central bank officials.

2.     The Value of Large Movements in Interest Rates

Intuitively, the speed with which output gaps are closed and inflation fluctuations are stabilized depends is likely to depend on how dramatically a central bank is willing to move interest rates. It is notable that in explaining their monetary policy objectives, most central banks emphasize stabilizing inflation and unemployment, or stabilizing inflation alone.[2] But implicitly, many central banks act as if interest rate stabilization was also part of their mandates. This may be a mistake. The paper in this volume “The Yellen Rules” by Alex Nikolsko-Rzhevskyy, David H. Papell and Ruxandra Prodan provides suggestive evidence that increasing the size of the coefficients in the Taylor rule yields good outcomes; increasing the size of all the coefficients in the Taylor rule corresponds to increasing the size of interest rate movements. This is an idea that deserves much additional research.

For concreteness, let me talk about this as tripling the coefficients in the Taylor rule. Note that tripling the size of the coefficients in the Taylor rule only works on the downside if the lower bound on interest rates has been eliminated by changes in paper currency policy. Once the lower bound on interest rates has been eliminated, tripling the coefficients in the Taylor rule is likely to be a relatively safe experiment for a central bank to try. 

3.     Data Dependence in Both Directions

“Data dependence” has deservedly gained prominence as a central banking phrase. But in its current usage, “data dependence” of the action of a central bank at a given meeting almost always refers to incoming data determining either (a) whether the target interest rate will stay the same or go up in one part of the cycle, or (b) whether the target interest rate will stay the same or go down in another part of the cycle. Seldom does it refer to incoming data determining (c) whether the target interest rate stays the same, goes up or goes down. This is not the way formal models of optimal monetary policy work. If the target rate is where it should be at one meeting, there will be some optimal expected drift of the rate between that meeting and the next, but if meetings are relatively close together it would be strange if incoming news weren’t sometimes large enough to outweigh any such expected drift in the optimal policy. Indeed, if relevant news drove a diffusion process, the standard deviation of news would be proportional to the square root of the time gap between meetings, while the drift would be directly proportional to the time gap between meetings. Thus, when meetings are as close together as they are for most central banks (often as little as six weeks or one month), it would take quite large drift terms in the optimal policy to overcome the implications of news with such regularity that graphs of the target rate would take on the familiar stair-step pattern (going up the stairs, plateauing, then going down the stairs, and so on).

Indeed, I remember well a discussion with an economist at a foreign central bank in charge of calculating the implications of a formal optimal monetary policy rule. The rule had exactly the two-way character I am talking about, requiring the economist to add a fudge factor in order to get a one-way recommendation acceptable to the higher-ups.

One of the most important benefits of data dependence in both directions is that it allows more decisive movements in rates since it is considered easy to reverse course if later data suggests that move was too large. Indeed, the benchmark of optimal instrument theory with a quadratic loss function indicates that additive uncertainty does not affect the optimal level of the instrument. Thus, there is no need for a central bank that embraces “data dependence in both directions” to water down its reaction to a shock simply because it does not know what the future will bring. Staying put is more likely to be a big mistake than ignoring the uncertainty and going with the central bank’s best guess of the situation.

It is only when the effect size of the instrument itself is uncertain that movements in the instrument should be damped down. But one of the big advantages of interest rate policy—even negative interest rate policy—as compared to quantitative easing is that the size of the effect of a given movement in interest rates is much easier to judge given theory and experience than the effect of a given amount of long-term or risky asset purchases. Simple formal models predict that selling Treasury bills to buy long-term assets will have no effect, so any effect of quantitative easing is due to a nonstandard effect. Thus, any effect of such actions is due to one or more of many, many candidate nonstandard mechanisms. Theory then gives less guidance about the effects of quantitative easing than one might wish. And since the quantity of large scale asset purchases since 2008 was not enough to get economies quickly back on track, the relevant size of large scale asset purchases if one were to rely primarily on them in the next big recession is larger than anything for which we have experience. By contrast, in simple formal models, it is the real interest rate that matters for monetary policy. Monetary history provides a surprisingly large amount of experience with deep negative real rates. Other than paper currency problems that can be neutralized, only nonstandard mechanisms—such as institutional rigidities—would make low real rates coming from low nominal rates significantly different from low real rates coming from higher inflation.

One objection often made to “data dependence in both directions” is that reversing directions will make the central bank look bad—as if it doesn’t know what it is doing. A good way to deal with this communications problem is to spell out a formula in advance for how incoming data will determine the starting point for the monetary policy committee’s discussion of interest rates. Then it will be clear that it is the data itself that is reversing directions, not some central banker whim.

4. Stepping Away from Interest Rate Smoothing. The stair-step pattern of the target rate over time reflects not only an aversion to changing directions, but also an aversion to changing the interest rate too many basis points at any one meeting, where in the US the definition of “too many” is that traditionally, there is a substantial presumption against a movement of more than 25 basis points. The theoretical analogue of this behavior is often called “interest rate smoothing.” 

There is a myth among many macroeconomists that commitment issues make it sensible to tie the current level of the target rate to the past level of the target rate with interest rate smoothing. Commitment issues can make the drift term in the optimal monetary policy rule stronger and thereby tilt things somewhat toward a predominant direction of movement, but it is hard to see how they would penalize large movements in the target rate at a given meeting of the monetary policy committee. 

For “data dependence in both directions” and “stepping away from interest rate smoothing,” one need not embrace perfectly optimal monetary policy to get substantial improvements in monetary policy. Simply reducing the implicit penalty on changing directions or making a large movement in the target rate at a given meeting would bring benefits. 

5.     A More Equal Balance Between Output Stabilization and Inflation Stabilization

The Classical Dichotomy between real and nominal is akin to Descartes’ dichotomy between matter and mind. To break the mind-matter dichotomy, Descartes imagined that mind affected the body through the pineal gland (because it is one of the few parts of brain that does not come in a left-right pair). Where is the location analogous to the pineal gland where the Classical Dichotomy is broken in sticky price models? The actual markup of price over marginal cost (P/MC). For many sticky-price models, all the real part of the model needs to know is communicated by the actual markup (P/MC) acting as a sufficient statistic.[3] That is, if one takes as given the path of P/MC, one can often determine the behavior of all the relative prices and real quantities without knowing anything else about the monetary side of the model.

In cases where only one aggregate value for P/MC matters, the monetary model will differ from the real model it is built on top only in one dimension corresponding to the consequences of different values of P/MC. Thus, when duplicating the behavior of the underlying real model is attractive from the standpoint of welfare, all that is needed for excellent monetary policy is to choose the level of stimulus or reining in that yields the value of P/MC the underlying real model would have. Even when firm-level heterogeneity also matters for welfare, duplicating the aggregate behavior of the underlying real model is often an excellent monetary policy.

This logic is a big part of what lies behind the “divine coincidence” (Olivier Blanchard and Jordi Gali, 2007). That is, having, for many sticky-price models, only one dimension in which the vector of aggregate variables of the model can depart the vector of aggregate variables that would prevail in the underlying real model is a big part of the reason that the monetary policy that makes the output gap zero in those models also stabilizes inflation.

There are two main reasons the divine coincidence might fail. First, staying at the natural level of output generated by the underlying real model may not be the best feasible policy if there are fluctuations in the gap between the natural level of output and the ideal level of output (that would equate the social marginal benefit and social marginal cost of additional output).

Second, there may be more than one sectoral actual markup ratio P/MC that matters. Labor can be considered an intermediate good produced by its own sector far upstream, so sticky wages count as an additional sectoral price. But having sticky prices in the durables sector that follow a different path from sticky prices in the nondurables sector will also give the model more than one dimension in which it can depart from the underlying real model. The key to this second type of failure of the divine coincidence is fluctuations in the ratio between two different sticky-price aggregates.

When two aggregate markup ratios matter, it can be difficult for monetary policy to duplicate the behavior of the underlying real model. For monetary policy to keep the actual markups in two sectors simultaneously at the levels that would duplicate the behavior of the underlying real model even in the face of frequent shocks, monetary policy would have to be based on two instruments different enough in their effects that they could have a two-dimensional effect on the economy.

When the divine coincidence fails, there is a tradeoff between stabilizing output and stabilizing inflation. Then it becomes important to know how bad output gaps are compared to aggregate price fluctuations. Much of the cost of output gaps is quite intuitive: less smoothing of labor hours. The cost of departures from steady inflation are a little less intuitive: in formal models the main cost of price fluctuations is intensified leapfrogging of various firms’ prices over each other that can lead people to go to stores or to products that make no sense from an efficiency point of view. This cost of leapfrogging prices causing inefficient purchasing patterns is proportional to how responsive people are to prices. The higher the price elasticity of demand, the bigger the misallocations from the microeconomic price disturbances caused by leapfrogging prices when macroeconomic forces make aggregate prices to fluctuate.

It is common in optimal monetary policy models to calibrate the price elasticity of demand at the firm level to a quite high value—often as high as 11. Where does this relatively high number come from? I have a suspicion. Susanto Basu and John Fernald (1996, 2002) and Basu (1997) estimate average returns to scale in the US economy of 1.1. Given free entry and exit of monopolistically competitive firms, in steady state average cost (AC) should equal price (P): if P > AC, there should be entry, while if P < AC there should be exit, leading to P=AC in steady state. Price adjustment makes marginal cost equal to marginal revenue in steady state. Thus, with free entry, in steady state,

where by a useful identity[4] γ is equal to the degree of returns to scale and μ is a bit of notation for the desired markup ratio P/MR. The desired markup ratio is equal in turn to the price elasticity of demand epsilon divided by epsilon minus one: 

Thus, a returns-to-scale estimate of γ=1.1 implies a price elasticity of demand ε11.

There are several holes in this logic. First, the average returns-to-scale estimate of 1.1 is suspect. Disaggregating by sectors in Basu, Fernald and Kimball (2006), the returns to scale estimate for nondurable manufactures was .87, and within durables production, lumber was .51. This is implausible. Tjalling Koopmans argued persuasively that if all factors are accounted for, and plenty of each factor is available, a production process can always be replicated, guaranteeing that constant returns to scale are possible. That is, if all factor prices are held constant, twice as much output can always be produced at twice the cost by doubling all inputs and doing the same thing all over again. It may be possible to do better—increasing returns to scale—but the firm will never have to do worse than replicating its production process. Therefore, returns to scale are at least constant when defined by what happens to costs with increases in scale when factor prices are held fixed.

Why might the estimate of returns to scale for nondurables come in at .87, and for lumber .51? There is more than one possible answer, but an obvious possibility is that in the boom, firms need to employ worse quality inputs, while in the bust, firms only hang onto the best quality inputs. It is not possibly to fully adjust for input quality using the available indicators of quality. So what looks like 1% increase in inputs might be only a .87% increase in quality-adjusted inputs or less if one could only adjust for unmeasured quality. If the quality profile of hiring and layoffs remains similar in many different circumstances, the false returns-to-scale estimate of .87 for nondurables might still be useful for isolating movements in productivity from these regular patterns of expansion and contraction of scale with accompanying changes in quality, but the false returns-to-scale estimate is not appropriate as part of the average for estimating the  that should be equal to the desired markup ratio  in the long run if there is free entry and exit.

Moreover, given the logic of using only the very best inputs in the bust, but worse inputs in the boom, the bias that affects nondurables manufacturing is likely to affect other sectors as well. Thus, true returns to scale and therefore the markup ratio could be much bigger than 1.1. The bigger the markup ratio, the lower the implied price elasticity of demand and the lower the costs from leapfrogging prices causing inefficient purchases.

Another reason the desired markup ratio may be underestimated—and therefore the price elasticity of demand therefore overestimated—is that firms face substantial sunk costs in entering a market. For example, John Laitner and Dmitriy Stolyarov (2011) emphasize the high failure rate in the first few years of a new business. This high failure rate is likely to deter entry in a way that will not show up in any observed flow fixed cost once the business is a solidly going concern. Therefore, one can expect that price will be above the observed average cost. Yet the observed average cost is what is relevant for estimates of returns to scale from looking at what happens to going concerns in booms and busts. Thus, P > AC implies that mu>gamma even for the true gamma when the true gamma is interpreted in this way as what happens when going concerns vary their scale.

Finally, the price elasticity of demand that bears a long-run relationship to the degree of returns to scale is the long-run price elasticity of demand. But the price elasticity of demand that matters for leapfrogging prices causing inefficient purchases is the short-run price elasticity of demand. The short-run price elasticity of demand is likely to be much smaller than the long-run price elasticity of demand. This gives one more reason why optimal monetary policy models with common calibrations might overestimate the welfare cost of inefficient purchases caused by the leapfrogging prices accompanying aggregate price fluctuations. Thus, it is possible that the cost of output gaps is of roughly the same size as the cost of price fluctuations rather than being significantly smaller as common calibrations imply.

6.     The Importance of Durables

In “Sticky Price Models and Durable Goods,” Robert Barky, Chris House and I show that as soon as a sticky price model has a durable goods sector (which many do not) it is sticky prices for durable goods that give monetary policy its power. Sticky prices in nondurables only reduce the power of monetary policy greatly, and lead to negative comovement durables and aggregate output. This importance of durables prices for sticky-price models suggests that durables prices ought to also be important for monetary policy. Indeed, Barsky, Christoph Boehm, House and Kimball (2016) argue that the price index a central bank tries to stabilize should have a weight on durables larger than the weight of durables in GDP. By contrast, some central banks pay most attention to the consumption deflator, which has a much smaller weight on durable goods prices than the weight of durables in GDP.

The importance of this for monetary policy can be illustrated by the impulse responses in Basu, Fernald, Jonas Fisher and Kimball (2013). There, technology improvements in the nondurables and services sector are estimated to expand employment, while technology improvement in the durable goods sector lead to lower employment. Why might this be? A simple story is that the Fed is staring at the consumption deflator. When there is an improvement in nondurables and services technology, this shows up in a lower consumption deflator, and the Fed provides some stimulus to restore the consumption deflator to it target growth rate. When there is an improvement in durables technology, particularly for business equipment, the consumption deflator is affected much less. Therefore, the Fed does not accommodate as much, and the ability to produce more output with the same inputs instead leads to something closer to producing the same output with fewer inputs. If this is true, it points to a serious problem with focusing primarily on a price index that has a low weight on durables.

There is another message for research on monetary policy. Including a durable goods sector in monetary policy models is crucial. Too much of the literature does not do this. Adding durables goods sectors to monetary policy models should be a big part of the research agenda in the coming decade. Moreover, models with a durable goods sector should explore a variety of parameter values: parameter values that make the durable goods sector more important for the results as well as parameter values that keep the durable goods sector relatively tame so that it does not affect results much.

7.     Adjusting for the Risk Premium

Vasco Curdia and Michael Woodford (2010) argue that the usual target rate—assumed to be a safe short-term interest rate—should drop by about 80% of the rise in the risk premium—say as measured by the spread between the T-bill repo rate and the commercial paper rate. If borrowers have a much higher elasticity of intertemporal substitution than savers, the recommended percentage of compensation for a rise in the risk premium could get even higher, getting close to what is effectively a commercial paper rate target. It would be very helpful for central banks to bring this kind of thinking explicitly into policy-making.

There are two other points to make here. First, note that having something close to an effective commercial paper rate target would do much of what people saying interest rate policy should respond to financial stability concerns want in practice, but for a different reason that does not depend on any fear of disaster. Second, when the risk premium rises dramatically, following such a policy requires having eliminated the lower bound on interest rates—one more way in which eliminating the lower bound on interest rates is foundational for next generation monetary policy.

8.     Coordination with Financial Stability

Beyond the compensation for movements in the risk premium discussed above, let me argue that if financial stability concerns are significantly affecting interest rate policy, other tools to enhance financial stability are being underused. Here, my perspective is strongly influenced by Anat Admati and Martin Hellwig’s The Banker’s New Clothes. They make the case that there is very little social cost to strict leverage limits on financial firms. John Cochrane (2013), in his review of The Banker’s New Clothes, writes memorably:

Capital is not an inherently more expensive source of funds than debt. Banks have to promise stockholders high returns only because bank stock is risky. If banks issued much more stock, the authors patiently explain, banks' stock would be much less risky and their cost of capital lower. "Stocks" with bond-like risk need pay only bond-like returns. Investors who desire higher risk and returns can do their own leveraging—without government guarantees, thank you very much—to buy such stocks. …

Why do banks and protective regulators howl so loudly at these simple suggestions? As Ms. Admati and Mr. Hellwig detail in their chapter "Sweet Subsidies," it's because bank debt is highly subsidized, and leverage increases the value of the subsidies to management and shareholders. To borrow without the government guarantees and expected bailouts, a bank with 3% capital would have to offer very high interest rates—rates that would make equity look cheap. Equity is expensive to banks only because it dilutes the subsidies they get from the government. That's exactly why increasing bank equity would be cheap for taxpayers and the economy, to say nothing of removing the costs of occasional crises. …

How much capital should banks issue? Enough so that it doesn't matter! Enough so that we never, ever hear again the cry that "banks need to be recapitalized" (at taxpayer expense)!

I have yet to hear a persuasive argument that the social as opposed to the private cost of strict leverage limits are anything to be concerned about. There are only two concerns that come close. First, it may be important to cut capital taxation in other ways to compensate for the higher effective tax rates when firms have more equity. But this is a plausible political tradeoff to make as part of a legislative package that dramatically raises capital requirements. Second, the reduced subsidy to financial firms could reduce aggregate demand. But eliminating the zero lower bound guarantees that there will longer be any shortage of aggregate demand.

But what if the interest rate cuts that bring aggregate demand back up themselves hurt financial stability? The answer is that if leverage limits are strict enough, this is unlikely to be a problem. One hears arguments (a) that interest rates can’t be cut because it will hurt financial stability, and (b) that capital (equity-finance) requirements can’t be raised because they will hurt aggregate demand. However, as long as, in absolute values, the ratio of aggregate demand to financial stability effects of interest rate cuts is larger than the ratio of aggregate demand to financial stability effects of higher capital requirements. This can be seen from the following diagram, taken from my blog post “Why Financial Stability Concerns Are Not a Reason to Shy Away from a Robust Negative Interest Rate Policy.”

The diagram above shows that, as long as interest rate cuts—which are primarily intended to effect aggregate demand—and higher equity requirements—which are primarily intended to affect financial stability—each have intended effects that are bigger than their side effects, then a combination of interest rate cuts with increases in equity requirements will both increase aggregate demand and improve financial stability. Careful study of the diagram indicates that what is crucial is that the vector of effects from interest rate cuts have a smaller negative slope than the vector of effects from higher equity requirements; if that condition is satisfied, getting the relative dosages of lower rates and higher equity requirements will both increase aggregate demand and improve financial stability.

Though they do not act alone, the Fed and other central banks have significant influence over the level of capital requirements. They should use their influence in this area vigorously to push for higher capital requirements, both for the direct benefits of higher financial stability and to free up interest rate policy to close output gaps and stabilize inflation.

9.     A Nominal Anchor

My understanding of the importance of nominal anchors is more second-hand and impressionistic than I would like. But it seems that many who study monetary policy have argued for targeting a path for prices as opposed to a level for inflation. Others have argued for targeting the velocity-adjusted money supply, also known as nominal GDP. Both ideas involve having a nominal anchor for monetary policy. During the Great Recession, either type of nominal anchor would have led to a rule that recommended more monetary stimulus than major economies in fact received—a recommendation that looks good in hindsight.

It is also my impression that nominal anchors make monetary policy rules “stable” in the theoretical sense to a wider set of shocks and under a wider range of solution concepts than monetary policy rules that do not have nominal anchors. Our ignorance about the future shocks economies and which solution concepts are appropriate make that kind of robust theoretical stability valuable.

If price-level targets are explained in terms of their implications for inflation, they seem unintuitive and even dangerous to many people, since viewed through that lens they call for “catch-up” inflation if inflation has been below the growth rate of the price-level target. Someday, after the lower bound on interest rates has been eliminated, it may be possible to deal with this communications problem by having a target of absolute price stability, in which the price level target does not change at all (within the limits of how well aggregate prices can be measured). An unchanging price-level target is much easier to explain to the public than a moving price-level target.

Advocates of a nominal GDP target have demonstrated that a nominal GDP target can be explained reasonably well. I do not think a nominal GDP target faces any insuperable communications difficulties.

An Example of an Interest Rate Rule Along the Lines of These Principles

Primarily as a way of illustrating the ideas above, let me set down the following interest-rate rule:

where

and g is a periodically revised estimate of the current trend growth rate of potential (natural) real GDP, determined as the median of all the values put forward by each member of the monetary policy committee.

Note that the coefficient of 1.5 on log nominal GDP is equivalent to a coefficient of 1.5 on log real GDP and a coefficient of 1.5 on the log GDP deflator. Thus, it includes both a tripling of the regular Taylor rule coefficient on log real GDP and a coefficient of 1.5 on the log price level in order to provide a nominal anchor. One way in which the nominal anchor here provides additional stability is that if the estimate of the natural level of output is off, the movement of the price level will still eventually restore a reasonable interest rate rule. If g is wrong as an estimate of the growth rate of natural output, it should be possible to eventually discover and correct that. If not, a persistent error in g could lead to inflation persistently away from the target of zero. But things still would not spiral out of control: the inflation rate would simply be away from target by the amount of the persistent error in g.

The Potential of Interest Rate Policy

The picture I want to paint is of the possibility of a monetary policy rule that would be much better than the monetary policy rules we have become used to. Much is uncertain. Research that hasn’t been done yet must be followed by trial runs yet to be conducted on the most promising ideas. But it seems possible interest rate policy could close output gaps and stabilize inflation within about a year—that is, with the speed of getting the economy back on track limited only by the well-known 9- to 18-month lag in the effect of monetary policy. And for shocks that have some forewarning, that 9- to 18-month clock can start from the moment of that forewarning.

With no effective lower bound on interest rates, there will be no need to conduct open market operations in anything but T-bills. There will be no need for fiscal stabilization beyond automatic stabilizers and possibly “national lines of credit” to households a la Kimball (2012) to stimulate the economy in the period before monetary policy stimulus kicks in given its lag.

One level of success would be to return to the Great Moderation. A higher level of success would be stay at the natural level of output all the time, except for the period between the arrival of a shock and when the effects of monetary policy kick in. A bit higher level of success would be to essentially solve the monetary policy problem.

Of course, solving the monetary policy problem does not mean all our problems are solved! The long-run Classical Dichotomy means that long-run economic growth depends on other policies. And although it might help, economic growth alone cannot solve the problems of war and peace, nor answer the question of what our highest human objectives should be.

However, for those of us who have spent or are spending an important part of our individual careers in trying to improve monetary policy, helping to understand and implement the principles of next generation monetary policy is a worthy endeavor indeed.

[1] See https://blog.supplysideliberal.com/post/53171609818/electronic-money-the-powerpoint-file for a list of these seminars.

[2] The Fed’s mandate is “to promote effectively the goals of maximum employment, stable prices, and moderate long-term interest rates.” Moderate long-term interest rates come from low inflation. And term-structure relationships make stable long-term interest rates consistent with sharp movements in short-term interest rates for short periods of time.

[3] The actual markup P/MC should be distinguished from the desired markup: price over marginal revenue P/MR. The desired markup mainly matters for price-setting, and is one of the ingredients that determines the actual markup P/MC.

References

Agarwal, Ruchir, and Miles Kimball, 2015. “Breaking Through the Zero Lower Bound,” IMF Working Paper 15/224.

Admati, Anat, and Martin Hellwig, 2013. The Banker’s New Clothes: What’s Wrong with Banking and What to Do about It, Princeton University Press.

Barsky, Robert, Christoph Boehm, Chris House and Miles Kimball, 2016. “Monetary Policy and Durable Goods,” Federal Reserve Bank of Chicago Working Paper No. WP-2016-18.

Barsky, Robert, Chris House, and Miles Kimball, 2007: “Sticky-Price Models and Durable Goods,” American Economic Review, 97 (June), 984-998.

Basu, Susanto, 1997. “Returns to Scale in U.S. Production: Estimates and Implications.” Journal of Political Economy 105 (April), pp. 249-83.

Basu, Susanto, and John Fernald, 1996. “Procyclical Productivity: Increasing Returns or Cyclical Utilization?” Quarterly Journal of Economics 111 (August), pp. 719-51.

Basu, Susanto, and John Fernald, 2002. “Aggregate Productivity and Aggregate Technology.” European Economic Review 46 (June), pp. 963-991.

Basu, Susanto, John Fernald and Miles Kimball, 2006. “Are Technology Improvements Contractionary?” American Economic Review 96 (December), pp. 1418-1448.

Basu, Susanto, John Fernald, Jonas Fisher, and Miles Kimball, 2013. “Sector-Specific Technical Change.” Federal Reserve Bank of San Francisco mimeo.

Bernanke, Ben, September 2016. “Modifying the Fed’s policy framework: Does a higher inflation target beat negative interest rates?” Ben Bernanke’s blog https://www.brookings.edu/blog/ben-bernanke/2016/09/13/modifying-the-feds-policy-framework-does-a-higher-inflation-target-beat-negative-interest-rates/#cancel

Blanchard, Olivier and Jordi Gali, 2007. “Real Wage Rigidities and the New Keynesian Model”
Journal of Money, Credit, and Banking, supplement to vol. 39 (1), 2007, 35-66

Cochrane, John, March 1, 2013. “Running on Empty: Banks should raise more capital, carry less debt—and never need a bailout again,” Wall Street Journal op-ed.

Curdia, Vasco, and Michael Woodford, 2010. "Credit Spreads and Monetary Policy," Journal of Money, Credit and Banking, Blackwell Publishing, vol. 42 (1), 3-35

Kimball, Miles, (2012). “Getting the Biggest Bang for the Buck in Fiscal Policy,” pdf available at https://blog.supplysideliberal.com/post/24014550541/getting-the-biggest-bang-for-the-buck-in-fiscal recommended over the NBER Working Paper 18142 version.

Kimball, Miles, originally 2013, regularly updated. “How and Why to Eliminate the Zero Lower Bound,” https://blog.supplysideliberal.com/post/62693219358/how-and-why-to-eliminate-the-zero-lower-bound-a

Kimball, Miles, 2015. “Negative Interest Rate Policy as Conventional Monetary Policy,” National Institute Economic Review, 234:1, R5-R14.

Kimball, Miles, 2016. “Why Financial Stability Concerns Are Not a Reason to Shy Away from a Robust Negative Interest Rate Policy,” Confessions of a Supply-Side Liberal blog, https://blog.supplysideliberal.com/post/144132627184/why-financial-stability-concerns-are-not-a-reason

Kocherlakota, Narayana, February 1, 2016. “The Potential Power of Negative Nominal Interest Rates” https://sites.google.com/site/kocherlakota009/home/policy/thoughts-on-policy/2-1-16

Kocherlakota, Narayana, February 9, 2016. “Negative Rates: A Gigantic Fiscal Policy Failure” https://sites.google.com/site/kocherlakota009/home/policy/thoughts-on-policy/2-9-16

Kocherlakota, Narayana, June 9, 2016. “Negative Interest Rates Are Nothing to Fear,” Bloomberg View https://www.bloomberg.com/view/articles/2016-06-09/negative-interest-rates-are-nothing-to-fear

Kocherlakota, Narayana, September 1, 2016. “Want a Free Market? Abolish Cash,” Bloomberg View https://www.bloomberg.com/view/articles/2016-09-01/want-a-free-market-abolish-cash

Laitner, John and Dmitriy Stolyarov, 2011. “Entry Conditions and the Market Value of Businesses,” University of Michigan mimeo.

Reis, Ricardo (2017) “Funding Quantitative Easing to Target Inflation” in Designing Resilient Monetary Policy Frameworks for the Future, Federal Reserve Bank of Kansas City.

Rogoff, Kenneth, August 2016. The Curse of Cash, Princeton University Press. 

 

Building Up With Grace

High rises don't have to be soulless. In "Density is Destiny" I wrote about the importance of density for economic growth. The way to have density and still have space for people in their homes is to build up. I also tried my hand in "Density is Destiny" at sketching a possible design for pleasant high-rises that I think can ultimately be built at non-luxury prices. Anna Baddeley's article linked above talks about some innovative architectural designs for high rises that I suspect might be at the higher end, but might become more affordable with technological progress. 

Urban density contributes to creativity and therefore to economic growth. Moreover, greater density through relaxing height restrictions is important for helping those with lower incomes and therefore for social justice. Experience has shown that it is very difficult to move good jobs to people. But it is easy for people to move to where good jobs are if affordable housing is available. The phrase "affordable housing" has often been used for a token amount of subsidized housing. But the principles of supply and demand mean that the only way to have affordable housing for the masses is to allow construction of many more units.

High density does not mean that height restrictions need to be removed everywhere within an urban area. That can be some low-density subdivisions for the rich folks. But in my view social justice requires that every metropolitan area has some large section with essentially no height restrictions for residential buildings (other than those genuinely required by earthquake concerns) and very frequent and convenient bus service from the high section of the city to jobs in the rest of the city.

It is far from certain that we will realize the social justice requirement of a high section to every city. But I want to encourage social-justice-minded architects to prepare for that hoped-for day by thinking about graceful and pleasant designs that can give good homes to people at a wide range of different income levels. Showing in principle how good high rises can be even for those of modest means could do a lot to make it politically feasible to shift policies toward letting more people into our most vibrant cities. 

John Locke: Theft as the Little Murder

In section 11 of his 2d Treatise on Government: On Civil Government," John Locke sets down a very different moral basis and very different location of the right of enforcement for civil as opposed to criminal law:

From these two distinct rights, the one of punishing the crime for restraint, and preventing the like offence, which right of punishing is in every body; the other of taking reparation, which belongs only to the injured party, comes it to pass that the magistrate, who by being magistrate hath the common right of punishing put into his hands, can often, where the public good demands not the execution of the law, remit the punishment of criminal offences by his own authority, but yet cannot remit the satisfaction due to any private man for the damage he has received. That, he who has suffered the damage has a right to demand in his own name, and he alone can remit: the damnified person has this power of appropriating to himself the goods or service of the offender, by right of self-preservation, as every man has a power to punish the crime, to prevent its being committed again, by the right he has of preserving all mankind, and doing all reasonable things he can in order to that end: and thus it is, that every man, in the state of nature, has a power to kill a murderer, both to deter others from doing the like injury, which no reparation can compensate, by the example of the punishment that attends it from every body, and also to secure men from the attempts of a criminal, who having renounced reason, the common rule and measure God hath given to mankind, hath, by the unjust violence and slaughter he hath committed upon one, declared war against all mankind, and therefore may be destroyed as a lion or a tyger, one of those wild savage beasts, with whom men can have no society nor security: and upon this is grounded that great law of nature, “Whoso sheddeth man’s blood, by man shall his blood be shed.” And Cain was so fully convinced, that every one had a right to destroy such a criminal, that after the murder of his brother, he cries out, Every one that findeth me, shall slay me; so plain was it writ in the hearts of all mankind.

That is, in the state of nature, everyone has the right to enforce natural criminal law, but the right to enforce natural civil law (or have an agent enforce it) belongs to the one wronged. 

In his discussion of natural criminal law, I am struck by John Locke's repeated recourse to murder as a metaphor for other crimes as well. In section 6, he makes the connection explicit: 

[One] may not, unless it be to do justice on an offender, take away, or impair the life, or what tends to the preservation of the life, the liberty, health, limb, or goods of another.

A possible reaction to this might be that in 1689, when John Locke published the Two Treatises of Government, even many people in upper percentiles of the population in income were living close enough to the margin that the typical effect of having goods stolen on mortality was much greater than it would be for the rich today. But that is much too partial an analysis. If theft is not deterred in a nation, the whole nation will remain poor and many, many will die. Blocking theft, deception and threats of violence are the most basic tasks of government in order to give a nation a fighting chance to become rich. (See my column "The Government and the Mob" for a related discussion.) Most of the poorest nations on earth are poor either because of some failure to block theft, deception and threats of violence on the part of private individuals or because the government itself has resorted too readily to theft, deception and threats of violence.  

Nevertheless, when carefully regulated so as not to interfere with economic growth, "theft" by the government of resources from the very wealthy probably cannot be appropriately called a "little murder." But other things equal, actions that do inhibit economic growth will have a cost in lives, and are hard to justify unless they contribute to the preservation of life enough in other ways to counterbalance the lost life from slower economic growth.  

The Supply and Demand for Paper Currency When Interest Rates Are Negative

Note: The acronym "ROERC" is pronounced in the same way as Howard Roark's last name. It stands for "Rate of Effective Return on/for/of Cash.

Note: The acronym "ROERC" is pronounced in the same way as Howard Roark's last name. It stands for "Rate of Effective Return on/for/of Cash.

Why the Rate of Effective Return on Cash Curve Slopes Down

Ruchir Agarwal and I argue in our IMF working paper "Breaking Through the Zero Lower Bound" that when deep negative interest rates are needed for prompt macroeconomic stabilization, central banks should take paper currency off par. (On this point, also see "How and Why to Eliminate the Zero Lower Bound: A Reader's Guide.") But what if a central bank is legally or politically unable—or unwilling—to take paper currency off par? In our nascent paper "Implementing Deep Negative Rates"—and the Powerpoint presentation of the same name—Ruchir and I argue that a variety of measures can be taken to make an arbitrage using massive paper currency storage more difficult. We call this "the dirty approach" to dealing with the paper currency problem. Moreover, we argue that political economy forces are favorable for many of the measures in the dirty approach, whether taken by a central bank or by other arms of its government. First, versions of many of these measures are already being taken. Second, how many governments would really stand by and do nothing while private agents piled up in storage a substantial fraction of GDP in paper currency? That is, for the US, is it plausible that the government would stand by and do nothing as private agents pile up trillions and trillions of dollars worth of paper currency? 

Six measures in the dirty approach to the paper currency problem (which you can see discussed on video in "Enabling Deeper Negative Rates by Managing the Side Effects of a Zero Paper Currency Interest Rate: The Video") are

  1. Ban Electrification of Paper Currency. That is, prohibit any money market mutual funds or similar securities that have on their asset side more than a small percentage of paper currency. Though only a partial implementation of this measure, it is an open secret I heard from multiple sources in various places around Europe and elsewhere in the world that the Swiss National Bank has been using moral suasion to discourage financial firms from offering easily traded assets backed by paper currency.
  2. Use the Interest on Reserves Formula to Subsidize Zero Rates for Small Household Accounts. This would reduce the motivation for households to withdraw money from the bank and store it as paper currency, helping to avoid a "death of a thousand cuts" of small-scale paper currency storage by many households that could add up to a large total quantity of storage. See "How to Handle Worries about the Effect of Negative Interest Rates on Bank Profits with Two-Tiered Interest on Reserves Policies," "Ben Bernanke: Negative Interest Rates are Better than a Higher Inflation Target," and "The Bank of Japan Renews Its Commitment to Do Whatever it Takes." No central bank has done this yet, but the reception I have had for this proposal is quite favorable. And several central banks already have quite complex tiered interest-on-reserves formulas.  
  3. Charge Banks for Excess Paper Currency Withdrawals from the Cash Window, Allowing Them to Impose Restrictions in Turn. The Swiss National Bank and the Bank of Japan already have provisions in their tiered interest-on-reserves formulas that penalize banks for cumulative withdrawals of paper currency. This is a way of directly having negative interest rates on paper currency in the relationship between the central bank and private banks. However pass-through of paper currency to households may not currently be counted as part of those cumulative withdrawals. If it were, that would inhibit banks from giving out paper currency so freely. 
  4. Retire Large Denomination Notes of Paper Currency. This is under serious discussion in many countries. In action, the ECB has decided to eliminate the 500-euro note
  5. Ban Storage of Paper Currency as a Business. Note that, given the substantial economies of scale and what would be sunk costs in setting up a paper currency storage business, even the threat of future action banning this business can do a lot to stunt its growth.
  6. Put Tight Restrictions on Flows of Paper Currency Out of the Country. This is a small step from restrictions and hindrances to flows of paper currencies in and out of a country that are in place for other reasons. 

I argue that these kinds of measures (and perhaps inherent obstacles in place even without these kinds of measures) make the marginal rate of effective return on paper currency decreasing in the currency-region-wide amount of paper currency being stored, as shown in the graph of the rate of effective return on cash (ROERC) curve at the top of this post. Such difficulties tend to push the rate of effective return on cash below its superficial return of zero. If any dirty hindrances to redemption of paper currency into reserves (electronic money) are added, that would add to the downward slope of the ROERC curve on the right. 

On the left of the curve, the rate of effective return on cash is far above its superficial return of zero for the amount of paper currency needed to carry on high-demand illegal activities—including, notably, tax evasion. (The secrets people want most desperately to keep are typically secrets from the government, but there is some demand for cash in order to keep secrets from others.)

The Opportunity Cost of Cash Curve

The ROERC curve is the demand curve for paper currency. The supply curve for paper currency is the Opportunity Cost of Cash (OCC) curve.

To the private sector, the maximum possible potential supply quantity of paper currency is the entire monetary base. Without reducing the amount of loans, the potential supply of paper currency is the monetary base minus required reserves. But even beyond required reserves, there may be some amount of excess reserves that has value as a buffer stock and so has an implicit return beyond the interest on reserves (shown as a negative rate above, as in several countries now). If the monetary base is large, as it is now, the marginal opportunity cost for quite a bit of paper currency would be very close to the interest on reserves (IOR). But beyond some level, additional paper currency would require reducing excess reserves that have some significant value as a buffer stock, so above that level of paper currency, the opportunity cost of cash would rise above IOR. 

In drawing the opportunity cost of cash curve, I am heavily influenced by having seen Ricardo Reis's Jackson Hole Presentation "Funding Quantitative Easing to Target Inflation."  Ricardo argues that in the US, the value of reserves—and therefore the opportunity cost of cash—only drops to IOR after excess reserves reach about $1 trillion. 

Putting the ROERC and OCC Curves Together

The Classic Zero Lower Bound Argument

The classic zero lower bound argument corresponds to a ROERC curve flat at zero once certain high-priority demands for paper currency have been satisfied:

With a ROERC curve flat at zero beyond a certain point, the interest rate won't drop below zero even if the monetary base is large and IOR is negative. Moreover, starting from that situation, increasing the monetary base will do nothing to the interest rate, which will stay at zero. 

Similarly, starting from that situation, reducing IOR even further will do nothing to the interest rate: 

An Effective Lower Bound Below IOR

Now, what if there is a constant storage cost of paper currency. This make the ROERC curve flat at a level below zero after a certain point. This flat ROERC curve creates an effective lower bound below zero. If the IOR is still above that effective lower bound, a cut in IOR will lower the interest rate: 

As long as the effect rate of return on cash is below IOR, it is IOR, not the effective return on cash that is crucial for determining the interest rate. Indeed, assuming the monetary base is large enough that the function of marginal reserve is simply to earn IOR on them, increasing the monetary base further has no effect on the interest rate, it just increases excess reserves.

Again, because the effect return on cash is below IOR, paper currency isn't affecting things very much. 

An Effective Lower Bound Above IOR

By contrast, if the effective rate of return on cash is above IOR, constant beyond a certain amount of paper currency that has already been exceeded, then both reductions in IOR and open market operations do nothing to the interest rate:

The Clean Approach: Cutting the Effective Rate of Return on Paper Currency Using a Depreciation Mechanism

Sticking for simplicity with a flat ROERC curve beyond a certain point, one can illustrate the effect of having the exchange rate for paper currency in terms of reserves gradually decline over time with a downward shift in the ROERC curve. Here the idea is to cut the paper currency interest rate (PCIR) in tandem with cutting IOR: 

The PCIR or effective rate of return on cash is cut by making the exchange rate at which paper currency can be exchanged for reserves at the cash window drift downward more strongly. Obviously, doing this requires being willing to contemplate an exchange rate between paper currency and reserves that can become different from par. Being willing to take paper currency off par opens the possibility of making the effective return on paper currency a policy variable. The bulk of the links in "How and Why to Eliminate the Zero Lower Bound: A Reader's Guide" point to discussions of this approach. (On the cash window itself, see "An Underappreciated Power of a Central Bank: Determining the Relative Prices between the Various Forms of Money Under Its Jurisdiction.")

A Downward Sloping ROERC

If the ROERC curve is downward sloping, for whatever reason—and the monetary base is large enough to make the opportunity cost of cash equal to IOR—reductions in IOR lower the interest rate even if the exchange rate at the central bank's cash window between paper currency and reserves is left at par:

 

With the monetary base already large enough to make the opportunity cost of cash equal to IOR, open market purchases without any reduction in IOR do not affect the interest rate:

On the other hand, if the monetary base isn't large enough to make the opportunity cost of cash equal to IOR, reductions in IOR don't lower the interest rate, but open market purchases do: 

The Lump-of-Labor Model

"In France, as in other parts of Western Europe, the reigning explanation for unemployment was what economists call the “lump of labor” theory. The theory holds that a society has only a fixed amount of work that needs to be done, and therefore the only way to reduce unemployment is to share the available work. This was reflected in Mitterrand’s initial program. The government lowered the retirement age to sixty to push older people out of the workforce; this was expected to create openings for youngsters, on the assumption that each employer needed only a certain amount of labor and would replace departing workers one for one. Workers who reached age fifty-five could collect pensions equal to 80 percent of their wages if their employers agreed to replace each retiree with a worker under twenty-five. The regular workweek was cut from forty hours to thirty-nine, and the maximum workweek was reduced as well, in the expectation that employers might cover those hours by adding workers. The possibility that less work for the same pay might deter hiring, or that young workers might lack the skills of the experienced workers they were replacing and therefore have lower productivity, was not widely discussed in the France of 1981."

--Marc Levinson, An Extraordinary Time: The End of the Postwar Boom and the Return of the Ordinary Economy