SI
SI
discoversearch

We've detected that you're using an ad content blocking browser plug-in or feature. Ads provide a critical source of revenue to the continued operation of Silicon Investor.  We ask that you disable ad blocking while on Silicon Investor in the best interests of our community.  If you are not using an ad blocker but are still receiving this message, make sure your browser's tracking protection is set to the 'standard' level.
Strategies & Market Trends : The Residential Real Estate Crash Index -- Ignore unavailable to you. Want to Upgrade?


To: Poet who wrote (188131)3/3/2009 11:52:19 PM
From: Broken_ClockRead Replies (1) | Respond to of 306849
 
nor Bush, Clinton, Bush, Reagan etc.



To: Poet who wrote (188131)3/4/2009 12:39:34 AM
From: stockman_scottRespond to of 306849
 
Behind the debate over remaking U.S. financial policy will be a debate
over who’s to blame. It’s crucial to get the history right, writes a
Nobel-laureate economist, identifying five key mistakes—under Reagan,
Clinton, and Bush II—and one national delusion.

Capitalist Fools
By Joseph E. Stiglitz
Vanity Fair Magazine
January 2009
vanityfair.com

There will come a moment when the most urgent threats posed by the
credit crisis have eased and the larger task before us will be to
chart a direction for the economic steps ahead. This will be a
dangerous moment. Behind the debates over future policy is a debate
over history—a debate over the causes of our current situation. The
battle for the past will determine the battle for the present. So it’s
crucial to get the history straight.

What were the critical decisions that led to the crisis? Mistakes were
made at every fork in the road—we had what engineers call a “system
failure,” when not a single decision but a cascade of decisions
produce a tragic result. Let’s look at five key moments.

No. 1: Firing the Chairman

In 1987 the Reagan administration decided to remove Paul Volcker as
chairman of the Federal Reserve Board and appoint Alan Greenspan in
his place. Volcker had done what central bankers are supposed to do.
On his watch, inflation had been brought down from more than 11
percent to under 4 percent. In the world of central banking, that
should have earned him a grade of A+++ and assured his re-appointment.
But Volcker also understood that financial markets need to be
regulated. Reagan wanted someone who did not believe any such thing,
and he found him in a devotee of the objectivist philosopher and
free-market zealot Ayn Rand.

Greenspan played a double role. The Fed controls the money spigot, and
in the early years of this decade, he turned it on full force. But the
Fed is also a regulator. If you appoint an anti-regulator as your
enforcer, you know what kind of enforcement you’ll get. A flood of
liquidity combined with the failed levees of regulation proved
disastrous.

Greenspan presided over not one but two financial bubbles. After the
high-tech bubble popped, in 2000–2001, he helped inflate the housing
bubble. The first responsibility of a central bank should be to
maintain the stability of the financial system. If banks lend on the
basis of artificially high asset prices, the result can be a
meltdown—as we are seeing now, and as Greenspan should have known. He
had many of the tools he needed to cope with the situation. To deal
with the high-tech bubble, he could have increased margin requirements
(the amount of cash people need to put down to buy stock). To deflate
the housing bubble, he could have curbed predatory lending to
low-income households and prohibited other insidious practices (the
no-documentation—or “liar”—loans, the interest-only loans, and so on).
This would have gone a long way toward protecting us. If he didn’t
have the tools, he could have gone to Congress and asked for them.

Of course, the current problems with our financial system are not
solely the result of bad lending. The banks have made mega-bets with
one another through complicated instruments such as derivatives,
credit-default swaps, and so forth. With these, one party pays another
if certain events happen—for instance, if Bear Stearns goes bankrupt,
or if the dollar soars. These instruments were originally created to
help manage risk—but they can also be used to gamble. Thus, if you
felt confident that the dollar was going to fall, you could make a big
bet accordingly, and if the dollar indeed fell, your profits would
soar. The problem is that, with this complicated intertwining of bets
of great magnitude, no one could be sure of the financial position of
anyone else—or even of one’s own position. Not surprisingly, the
credit markets froze.

Here too Greenspan played a role. When I was chairman of the Council
of Economic Advisers, during the Clinton administration, I served on a
committee of all the major federal financial regulators, a group that
included Greenspan and Treasury Secretary Robert Rubin. Even then, it
was clear that derivatives posed a danger. We didn’t put it as
memorably as Warren Buffett—who saw derivatives as “financial weapons
of mass destruction”—but we took his point. And yet, for all the risk,
the deregulators in charge of the financial system—at the Fed, at the
Securities and Exchange Commission, and elsewhere—decided to do
nothing, worried that any action might interfere with “innovation” in
the financial system. But innovation, like “change,” has no inherent
value. It can be bad (the “liar” loans are a good example) as well as
good.

No. 2: Tearing Down the Walls

The deregulation philosophy would pay unwelcome dividends for years to
come. In November 1999, Congress repealed the Glass-Steagall Act—the
culmination of a $300 million lobbying effort by the banking and
financial-services industries, and spearheaded in Congress by Senator
Phil Gramm. Glass-Steagall had long separated commercial banks (which
lend money) and investment banks (which organize the sale of bonds and
equities); it had been enacted in the aftermath of the Great
Depression and was meant to curb the excesses of that era, including
grave conflicts of interest. For instance, without separation, if a
company whose shares had been issued by an investment bank, with its
strong endorsement, got into trouble, wouldn’t its commercial arm, if
it had one, feel pressure to lend it money, perhaps unwisely? An
ensuing spiral of bad judgment is not hard to foresee. I had opposed
repeal of Glass-Steagall. The proponents said, in effect, Trust us: we
will create Chinese walls to make sure that the problems of the past
do not recur. As an economist, I certainly possessed a healthy degree
of trust, trust in the power of economic incentives to bend human
behavior toward self-interest—toward short-term self-interest, at any
rate, rather than Tocqueville’s “self interest rightly understood.”

The most important consequence of the repeal of Glass-Steagall was
indirect—it lay in the way repeal changed an entire culture.
Commercial banks are not supposed to be high-risk ventures; they are
supposed to manage other people’s money very conservatively. It is
with this understanding that the government agrees to pick up the tab
should they fail. Investment banks, on the other hand, have
traditionally managed rich people’s money—people who can take bigger
risks in order to get bigger returns. When repeal of Glass-Steagall
brought investment and commercial banks together, the investment-bank
culture came out on top. There was a demand for the kind of high
returns that could be obtained only through high leverage and big
risktaking.

There were other important steps down the deregulatory path. One was
the decision in April 2004 by the Securities and Exchange Commission,
at a meeting attended by virtually no one and largely overlooked at
the time, to allow big investment banks to increase their
debt-to-capital ratio (from 12:1 to 30:1, or higher) so that they
could buy more mortgage-backed securities, inflating the housing
bubble in the process. In agreeing to this measure, the S.E.C. argued
for the virtues of self-regulation: the peculiar notion that banks can
effectively police themselves. Self-regulation is preposterous, as
even Alan Greenspan now concedes, and as a practical matter it can’t,
in any case, identify systemic risks—the kinds of risks that arise
when, for instance, the models used by each of the banks to manage
their portfolios tell all the banks to sell some security all at once.

As we stripped back the old regulations, we did nothing to address the
new challenges posed by 21st-century markets. The most important
challenge was that posed by derivatives. In 1998 the head of the
Commodity Futures Trading Commission, Brooksley Born, had called for
such regulation—a concern that took on urgency after the Fed, in that
same year, engineered the bailout of Long-Term Capital Management, a
hedge fund whose trillion-dollar-plus failure threatened global
financial markets. But Secretary of the Treasury Robert Rubin, his
deputy, Larry Summers, and Greenspan were adamant—and successful—in
their opposition. Nothing was done.

No. 3: Applying the Leeches

Then along came the Bush tax cuts, enacted first on June 7, 2001, with
a follow-on installment two years later. The president and his
advisers seemed to believe that tax cuts, especially for upper-income
Americans and corporations, were a cure-all for any economic
disease—the modern-day equivalent of leeches. The tax cuts played a
pivotal role in shaping the background conditions of the current
crisis. Because they did very little to stimulate the economy, real
stimulation was left to the Fed, which took up the task with
unprecedented low-interest rates and liquidity. The war in Iraq made
matters worse, because it led to soaring oil prices. With America so
dependent on oil imports, we had to spend several hundred billion more
to purchase oil—money that otherwise would have been spent on American
goods. Normally this would have led to an economic slowdown, as it had
in the 1970s. But the Fed met the challenge in the most myopic way
imaginable. The flood of liquidity made money readily available in
mortgage markets, even to those who would normally not be able to
borrow. And, yes, this succeeded in forestalling an economic downturn;
America’s household saving rate plummeted to zero. But it should have
been clear that we were living on borrowed money and borrowed time.

The cut in the tax rate on capital gains contributed to the crisis in
another way. It was a decision that turned on values: those who
speculated (read: gambled) and won were taxed more lightly than wage
earners who simply worked hard. But more than that, the decision
encouraged leveraging, because interest was tax-deductible. If, for
instance, you borrowed a million to buy a home or took a $100,000
home-equity loan to buy stock, the interest would be fully deductible
every year. Any capital gains you made were taxed lightly—and at some
possibly remote day in the future. The Bush administration was
providing an open invitation to excessive borrowing and lending—not
that American consumers needed any more encouragement.

No. 4: Faking the Numbers

Meanwhile, on July 30, 2002, in the wake of a series of major
scandals—notably the collapse of WorldCom and Enron—Congress passed
the Sarbanes-Oxley Act. The scandals had involved every major American
accounting firm, most of our banks, and some of our premier companies,
and made it clear that we had serious problems with our accounting
system. Accounting is a sleep-inducing topic for most people, but if
you can’t have faith in a company’s numbers, then you can’t have faith
in anything about a company at all. Unfortunately, in the negotiations
over what became Sarbanes-Oxley a decision was made not to deal with
what many, including the respected former head of the S.E.C. Arthur
Levitt, believed to be a fundamental underlying problem: stock
options. Stock options have been defended as providing healthy
incentives toward good management, but in fact they are “incentive
pay” in name only. If a company does well, the C.E.O. gets great
rewards in the form of stock options; if a company does poorly, the
compensation is almost as substantial but is bestowed in other ways.
This is bad enough. But a collateral problem with stock options is
that they provide incentives for bad accounting: top management has
every incentive to provide distorted information in order to pump up
share prices.

The incentive structure of the rating agencies also proved perverse.
Agencies such as Moody’s and Standard & Poor’s are paid by the very
people they are supposed to grade. As a result, they’ve had every
reason to give companies high ratings, in a financial version of what
college professors know as grade inflation. The rating agencies, like
the investment banks that were paying them, believed in financial
alchemy—that F-rated toxic mortgages could be converted into products
that were safe enough to be held by commercial banks and pension
funds. We had seen this same failure of the rating agencies during the
East Asia crisis of the 1990s: high ratings facilitated a rush of
money into the region, and then a sudden reversal in the ratings
brought devastation. But the financial overseers paid no attention.

No. 5: Letting It Bleed

The final turning point came with the passage of a bailout package on
October 3, 2008—that is, with the administration’s response to the
crisis itself. We will be feeling the consequences for years to come.
Both the administration and the Fed had long been driven by wishful
thinking, hoping that the bad news was just a blip, and that a return
to growth was just around the corner. As America’s banks faced
collapse, the administration veered from one course of action to
another. Some institutions (Bear Stearns, A.I.G., Fannie Mae, Freddie
Mac) were bailed out. Lehman Brothers was not. Some shareholders got
something back. Others did not.

The original proposal by Treasury Secretary Henry Paulson, a
three-page document that would have provided $700 billion for the
secretary to spend at his sole discretion, without oversight or
judicial review, was an act of extraordinary arrogance. He sold the
program as necessary to restore confidence. But it didn’t address the
underlying reasons for the loss of confidence. The banks had made too
many bad loans. There were big holes in their balance sheets. No one
knew what was truth and what was fiction. The bailout package was like
a massive transfusion to a patient suffering from internal
bleeding—and nothing was being done about the source of the problem,
namely all those foreclosures. Valuable time was wasted as Paulson
pushed his own plan, “cash for trash,” buying up the bad assets and
putting the risk onto American taxpayers. When he finally abandoned
it, providing banks with money they needed, he did it in a way that
not only cheated America’s taxpayers but failed to ensure that the
banks would use the money to re-start lending. He even allowed the
banks to pour out money to their shareholders as taxpayers were
pouring money into the banks.

The other problem not addressed involved the looming weaknesses in the
economy. The economy had been sustained by excessive borrowing. That
game was up. As consumption contracted, exports kept the economy
going, but with the dollar strengthening and Europe and the rest of
the world declining, it was hard to see how that could continue.
Meanwhile, states faced massive drop-offs in revenues—they would have
to cut back on expenditures. Without quick action by government, the
economy faced a downturn. And even if banks had lent wisely—which they
hadn’t—the downturn was sure to mean an increase in bad debts, further
weakening the struggling financial sector.

The administration talked about confidence building, but what it
delivered was actually a confidence trick. If the administration had
really wanted to restore confidence in the financial system, it would
have begun by addressing the underlying problems—the flawed incentive
structures and the inadequate regulatory system.

Was there any single decision which, had it been reversed, would have
changed the course of history? Every decision—including decisions not
to do something, as many of our bad economic decisions have been—is a
consequence of prior decisions, an interlinked web stretching from the
distant past into the future. You’ll hear some on the right point to
certain actions by the government itself—such as the Community
Reinvestment Act, which requires banks to make mortgage money
available in low-income neighborhoods. (Defaults on C.R.A. lending
were actually much lower than on other lending.) There has been much
finger-pointing at Fannie Mae and Freddie Mac, the two huge mortgage
lenders, which were originally government-owned. But in fact they came
late to the subprime game, and their problem was similar to that of
the private sector: their C.E.O.’s had the same perverse incentive to
indulge in gambling.

The truth is most of the individual mistakes boil down to just one: a
belief that markets are self-adjusting and that the role of government
should be minimal. Looking back at that belief during hearings this
fall on Capitol Hill, Alan Greenspan said out loud, “I have found a
flaw.” Congressman Henry Waxman pushed him, responding, “In other
words, you found that your view of the world, your ideology, was not
right; it was not working.” “Absolutely, precisely,” Greenspan said.
The embrace by America—and much of the rest of the world—of this
flawed economic philosophy made it inevitable that we would eventually
arrive at the place we are today.

*Joseph E. Stiglitz, a Nobel Prize winning economist, is a professor
at Columbia University.



To: Poet who wrote (188131)3/4/2009 6:03:33 AM
From: stockman_scottRead Replies (3) | Respond to of 306849
 
An interesting comment on a New York Times blog:

dealbook.blogs.nytimes.com

Yesterday my wife and I got a lengthy letter from the President of Duke University where our son is an undergraduate. The letter essentially gave parents an update on the impact of the current economic environment on the University. In short, salaries are being frozen, job cuts are on the way and new capital projects are being suspended. While the letter explicitly linked declining asset values in the University’s endowment to the coming cut-backs, it implied that this was simply the outcome of exposure to systemic risk. In fact, like Harvard, Duke’s investment managers exposed the endowment to a hefty allocation in high-risk, leveraged and illiquid investments. For a good few years now the entire foundations and endowment world has been in a swoon over Harvard’s outstanding performance and has taken an increasingly dim view of ‘marketable securities’ - i.e. stocks and bonds. Going forward I hope the best for these institutions’ fortunes, and I hope that they will be more risk-averse and conservative in their asset allocation strategies.

— Posted by John

March 3rd, 2009 3:01 pm