“Reagan doesn’t have that presidential look” — United Artists Executive Rejecting Reagan for Lead Role in The Best Man (1964)
The liberal wave started during the Progressive Era, gained momentum with FDR’s New Deal in the 1930s, crested with LBJ’s Great Society in the 1960s, and ran out of steam by the mid-1970s. Conservative Ronald Reagan won the 1980 presidential election by arguing that “In this present crisis, government is not the solution to our problem; government is the problem.” The Departments of Energy (1977-) and Education (1979-) expanded federal bureaucracy some in the ’70s, but the public mood was shifting back toward smaller government at all levels. Ralph Nader’s idea of creating a bureaucracy for consumer protection went nowhere. The Civil Aeronautics Board went away in 1985 and the Interstate Commerce Commission (railroads and trucking) in 1995, their safety enforcements transferred to other agencies. America, of course, never swings all the way in one direction or the other — right or left — but the momentum and rhetoric shifted right during the conservative resurgence of the 1970s and after. To repurpose for liberalism what Winston Churchill said about the meaning of World War II’s Battle of El-Alamein for Nazi Germany: “this was not the end. It was not even the beginning of the end. But it was, perhaps, the end of the beginning.”
Does America really have a “big government?” Proportional to the size of its country, America’s government as of 2015 was fairly small by international standards, spending around 14.5% of national GDP output compared to Australia (18%), Germany (19%), Russia (19%), the United Kingdom (19%), and Canada (21%). These World Bank figures are lower than the Office of Management & Budget figures (right) that usually hover ~20%. The U.S. leads the world in total expenditures, though, spending ~$3.8 trillion in 2016 while collecting $3.3 trillion in revenue. America spends more on its military than its next eight competitors and twice that of China and Russia combined (SIPRI), but doesn’t provide health insurance for those under 65, except for Medicaid at the state level. Of course, spending and military power aren’t the only measures of a government’s size or reach; there’s also its legal/regulatory system and the power of law enforcement.
Domestically, there’s a lot of noise about “tyranny” but, collectively, U.S. laws don’t stand out as being overly oppressive in comparison to other nations. Look at the chart on the right and think of how steadily the drumbeat has grown over the last decades about the government getting bigger and bigger. Some of that noise originates among those profiting from beating the drum, selling airtime on radio and cable or ads on blogs. Some of the noise originates from quarters unfamiliar with real tyranny or suspicious of a deep state within intelligence agencies operating outside the public or even regular government’s purview. Conspiracy theories are more entertaining and profitable than real knowledge or perspective. Still, many laws originate in agencies run by non-elected bureaucrats and housed under the executive branch such as the IRS, FDA, OSHA, FCC, EPA, etc., and it’s not surprising that citizens resist or resent these laws when they’re onerous or that bureaucrats don’t communicate well with citizens as to why they’re enacted or they’re administered with an inflexible or heavy hand. Governments grow because voting citizens across the political spectrum demand things, because agencies grow to police other agencies, because big companies lobby (i.e. bribe) politicians to pass regulations that small companies can’t afford to comply with, and because bureaucracies have a natural tendency to grow like fungi regardless of the first three reasons.
By the 1970s and ’80s, the Great Society era launched by Lyndon Johnson was waning, and the public was ready to send the pendulum back in the other direction, despite still wanting the services and protection government provided. You could trace one turning point in liberalism’s demise to New York City’s near-bankruptcy, when they ran out of money to pay public employees and had to look overseas to sell municipal bonds. While New York was the first city to tighten its budgetary belt, California was the first state, though in their case they cut taxes more than they cut spending. California passed Proposition 13 in 1978, a law banning increases in property taxes (beyond inflation) unless authorized by a two-thirds majority. The brewing conservative resurgence wasn’t just about cutting taxes, but also reducing government intervention in the economy and reinjecting religion into politics. With the election of Ronald Reagan in 1980, the conservative revolution launched under Barry Goldwater in 1964 led to a fundamental changing of the guard in Washington and in many states. Conservatives didn’t take over, but they stole the momentum and have more or less maintained it ever since.
Stagflation & Energy Crisis
Yet, the first president we’ll discuss as part of this conservative resurgence was a Democrat. By the 1976 election, the public wanted the most anti-Nixon, anti-Vietnam, anti-Watergate type candidate they could find. They found it in Jimmy Carter (D), a born-again, peanut-farming governor from Georgia untarnished by Washington politics. Watergate began a trend toward outsider candidates, resulting in state governors Carter, Reagan, Clinton and Bush the Younger all winning the presidency. After defeating Gerald Ford in a tight 1976 election, Carter came to Washington with what some congressmen perceived to be a holier-than-thou attitude and didn’t work well with what he perceived to be corrupt Washington insiders. Like Richard Nixon, he had a fortress mentality in the White House, not initiating relations with congressmen on Capitol Hill. He alienated conservatives by creating the Department of Energy to try to wean the country off Arab oil (the GOP didn’t want more bureaucracy, and oil companies feared breakthroughs on alternative energy), and he alienated Great Society Democrats by overseeing economic deregulation (about which more later) and insisting on a balanced budget. In that way, Carter was more of an independent and fiscal (budgetary) conservative than a party-line Democrat. As a fiscal conservative, Carter alienated Democrats by refusing to go further into debt and Republicans by refusing to cut taxes. But his main, seemingly intractable problem was stagflation.
Normally prices don’t rise during a recession but, by the mid-1970s, the U.S. was mired in stagflation: the unlikely combination of inflation and high unemployment. LBJ’s Great Society and the lengthy Vietnam War raised the federal deficit, contributing to inflation. So, too, did President Nixon decoupling American currency from the Gold Standard in 1971, due to too many trade surplus nations swapping greenbacks for a dwindling gold supply (American dollars were convertible to gold dating back to the Bretton Woods Conference in 1944). Going off the Gold Standard created confidence, or fiat, currency instead. Greenbacks from then on were worth whatever people thought they were worth, based on their confidence in America’s solvency and survival. As for coins, they cost more to mint than the amount of copper or nickel on the coin is worth. Caught in between its two mandates — curbing inflation and supporting job growth — the Federal Reserve initially didn’t raise interest rates to stem inflation because they feared that would further weaken the job market (also, inflation helps an indebted government because, in real dollars, it reduces the money it owes).
As we saw in the previous chapter, OPEC (Organization of Petroleum Exporting Countries) embargoed oil in 1973 to show the West how dependent they’d become while punishing them for supporting Israel. Then they raised prices and within a few short years, oil climbed from $3 to $12/barrel. The Iranian Revolution that we’ll cover later in this chapter caused another big spike in oil prices in 1978-79. This coincided with increased dependence on foreign oil. Oil hit $1 gallon for the first time in American history, even adjusted for inflation far higher than the 5¢ it cost in the 1950s.
To offset high prices and Peak Oil, the U.S. built the Trans-Alaska Pipeline from the North Slope, which was iced in much of the year, to the Port of Valdez, where tankers could ferry it to the Lower 48. That still wasn’t enough to offset the price hike. Many Americans overreacted by trying to hoard oil, not realizing that global price and local supply aren’t always linked. There was often no actual shortage in the pipeline. Either way, prices remained high, and oil is so important to industrialized societies that it can drive up inflation even during a recession, thus the stagflation. European countries tax oil enough to deliberately make it expensive and then use the taxes to pay for mass transit, encouraging people to conserve. The United States has historically devoted more to its military budget instead, partly to ensure the flow of cheap oil.
Detroit wasn’t well prepared to manufacture fuel-efficient cars in the 1970s. For years leading up to then, bigger was better. The oil crisis coincided with the overall rise of European and Japanese industry from the ashes of WWII, and they were better at making smaller cars that got good mileage. The U.S. rebuilt those countries as industrial powerhouses after the war and succeeded beyond their expectations. They were now fully rebuilt, which was good for the world economy, but also meant that the U.S. wasn’t the only kid on the block. Among other things, that meant American unions would steadily weaken from that point forward, and most middle and working-class families would need both parents working to keep up.
American manufacturing was entering a long, slow period of decline, as one factory after another shut down in the Rust Belt of the industrial Northeast and Midwest. That trend continued into the early 21st century if you look at manufacturing jobs. But if you measure by output, American manufacturing has done well in recent decades. In fact, the U.S. is producing far more than ever; it’s just producing more with automation and lowered-paid staff rather than union workers. And there is even an unmet need for semi-skilled workers in American manufacturing that can only be offset by immigration or more young citizens seeking work in manufacturing. As factory unions weakened, public unions suffered setbacks as well. Major cities like New York struggled with a combination of low taxes and well-pensioned public unions (policemen, firemen, sanitation workers, teachers, etc.).
President Carter described the need to transition to alternative energy as a moral imperative, or the “moral equivalent of war.” But his pleas for Americans to conserve energy out of a sense of patriotism mainly fell on deaf ears. Carter learned that, while most Americans are patriotic when it comes to wars, they’re less enthusiastic about turning down the AC in the summer or heat in the winter. The moral argument ran headlong into strands of Sunbelt Christianity that associated with the oil industry (what historian Darren Dochuk called “wildcat Christianity” or “oil patch evangelicalism”). Then a near meltdown at the Three Mile Island nuclear plant outside Harrisburg, Pennsylvania in 1979 dampened the public’s enthusiasm to pursue atomic energy as a viable alternative to fossil fuels. Carter had been a nuclear engineer himself, commanding a nuclear sub in the early Cold War, and visited the plant personally during the height of the crisis. Engineers staved off a meltdown of the inner core reactor, but the near-miss spooked the public and few new reactors went into construction afterward. Had they not been able to cool the reactor it would’ve sunk into the Earth, radiating the soil and water around it before exploding when it contacted groundwater.
A popular movie called The China Syndrome, so-named because one scenario is that a reactor might melt through the Earth’s core “all the way to China,” prophesied the near-meltdown just prior to the actual emergency. Problems with waste disposal and much worse crises in the USSR in 1986 (Chernobyl) and Japan in 2011 (Fukushima Dai-Chi) dampened the industry’s prospects despite the fact that nuclear power is mostly carbon-free. President Eisenhower’s dream of always having one of the giant reactor chimneys within site as one drove across the country didn’t materialize, and neither did mini-reactors in each home or smaller devices to propel vacuum cleaners and other appliances. Nuclear-powered coils that could melt driveway and sidewalk snow didn’t happen in the 1950s and they weren’t about to after Three Mile Island. Today, nuclear reactors supply ~ 20% of America’s electrical power.
When Carter talked about the economic malaise the country was in, he came across as ineffectual in fixing the situation or as telling Americans something they didn’t want to hear. His administration was struggling to keep the basic cost of living and energy from rising faster than wages. For many Americans, especially white-collar workers or union workers with automatic cost-of-living adjustments (COLAs), the price-wage spiral allowed them to keep up (people are skeptical about this concept but inflation requires wage spirals; otherwise no one would be able to afford the higher prices). However, the wage hikes were uneven across occupations. Some blue-collar workers lost ground and retirees on fixed incomes saw their savings shrink at the rate of inflation — probably the cruelest effect of high inflation. Investors weren’t happy either; while stock prices rose in terms of nominal dollars, there was no real increase (adjusted for inflation) from 1965 to 1982. Still, stocks were a better investment than a low-yielding savings account which, in effect, lost money as it grew gradually at a lower rate than inflation.
Just as the capacity of wages to rise along with inflation varies among occupations, so too, inflation is uneven from sector to sector. For instance, in the 21st century, despite low inflation in retail and deflation in electronics and air travel, prices in housing, healthcare, and education have risen relative to income. In an inflationary environment, borrowing makes sense because the amount you’ll owe later is, in effect, less, so people that hang on to their jobs keep borrowing and spending, which drives inflation even more. In a desperate move to stop double-digit inflation, which stood at a staggering 13% by 1980, Federal Reserve Chair Paul Volcker raised interest rates dramatically, slowing the economy because fewer people could borrow, but at least reversing inflation’s rise. As the short-term borrowing rate between banks soared to 20%, the nation dipped into its worst recession since the 1930s, this one deliberately caused.
Some economists argue that Volcker’s drastic actions were unnecessary because those prices would’ve subsided on their own, without government action. There were many causes of inflation in the 1970s besides low-interest rates, and those would’ve naturally taken care of themselves, according to this argument.
Slamming the brakes on the economy by raising rates also raised unemployment, which Volcker supporters like Milton Friedman claimed hovered naturally around 5% anyway. Unemployment shot up to nearly 11%, above. In truth, the country was in a bind that didn’t offer any easy solutions, and Carter sided with the conservative approach of Volcker and Friedman. America took its Volcker chemotherapy, killing inflation cells along with growth and employment cells. Things got worse before they got better and the economy didn’t bottom out until 1982. Farmers with variable-rate land mortgages also felt the pinch of higher rates, especially those that lost their wheat and corn export trade to the USSR with the end of détente (more below).
Jimmy Carter made some changes that helped the economy long-term besides initiating the painful process to slow inflation. To deregulate is to lessen the government’s role in a sector. One theme you’ll see in this chapter is that, much like its inverse regulation, deregulation can have unpredictable and unforeseen consequences, good and bad. If it was as simple as the-more-laws-the-better or the-fewer-laws-the-better we would’ve figured that out already. Spurred by Republicans and his intraparty rival, Senator Ted Kennedy (D-MA), Carter deregulated some industries that had been under the government’s control, including transportation (airlines, trucking, rail) and natural gas lines. In 1978, regional startups such as Southwest began to undersell big national airlines on an open market, challenging the original five-headed government-sanctioned oligopoly of United, Eastern, Braniff, American, and Delta. The impact of airline deregulation has been to dramatically cheapen the cost of flying, adjusted for inflation, so that by the late 20th century it was common for middle-class Americans to fly. However, in comparison to the mid-20th century, airlines now put less emphasis on customer service because they distinguish themselves by selling cheaper coach tickets rather than serving customers that are all paying similar, high rates.
Communications followed the same deregulatory trend, triggered by a 1974 anti-trust lawsuit breaking up Ma Bell, by 1982-84, into the new “Baby Bells” that reemerged into Verizon, AT&T, CenturyLink, etc. That opened up telecom for competitive pricing just before the advent of cell phones. Also, credit card companies — to the detriment of the working poor and unfrugal middle class, but in keeping with free-market principles — won the right to charge unlimited interest rates.
Carter also signed off on legislation allowing for the creation of Business Development Companies (BDC’s) that gave small investors a tax-friendly way to invest in private businesses and riskier start-ups than those otherwise allowed in the SEC-regulated public stock markets. All this helped lay the foundation for economic recovery in the 1980s but wasn’t enough to help Carter at the time. Finally, on a lighter note, Carter signed regulation legalizing home-brewing in 1978 that advanced the rise of craft beers as independent, artisanal brewers initially honed their craft in garages.
Meanwhile, Jimmy Carter had plenty of problems overseas to deal with. Building on Richard Nixon’s foundation, the U.S. officially normalized relations with China in 1979 but Nixon’s détente with the Soviets came unraveled under Carter. The Soviets gained influence in Ethiopia (East Africa) as a Marxist state killed hundreds of thousands in a Red Terror and various relocation schemes. Carter’s emphasis on human rights threatened the Soviets and the arms race spiked dramatically because of better ICBMs (inter-continental ballistic missiles) and multi-warhead MIRVs (multiple independently targetable reentry vehicles) that, according to rumor at least, could be dosed with biological weapons. This LGM-118 “Peacekeeper” MIRV the U.S. tested over the Kwajalein Atoll divides into eight 300 kiloton warheads, each ~20x more powerful than the Hiroshima bomb if detonated.
In the 1970s, each side was testing submarine-based ballistic missiles (SLBM’s) and air-launched cruise missiles that could be loaded onto traditional bombers like America’s B-52’s. The U.S. placed medium-range cruise missiles similar in design to the Nazis’ old V-1 “flying bombs” in southern England, loaded onto mobile launch pads in the payloads of trucks parked in underground bunkers. These relatively cheap warheads, each costing only ~ ¼ of a jet fighter, had a 2k-mile range and could incinerate a small city and burn everyone to death in a ten-mile radius. In general, Soviets focused more on size whereas the U.S. focused on accuracy. Each side additionally worked on neutron bombs that could wipe out life without destroying property, though the U.S. shelved plans to arm NATO with the new weapons due to public pressure. Near the end of his presidency, Carter announced that both sides had more than 5x as many warheads as they had in the early 1970s. A round of SALT (strategic arms limitation talks) between Carter and the Soviets slowed the madness some, limiting each side to 2400 warheads and convincing the Soviets to halt production on new MIRVs that carried up to 38 separate warheads.
Unfortunately, six months after the SALT II talks in Vienna, in December 1979, the Soviets invaded Afghanistan to support communist forces in a civil war there against the jihadist Mujahideen. In response, Carter embargoed agricultural trade, boycotted the 1980 Moscow Olympics, and issued a doctrine stating America’s intention to protect its oil interests in the Middle East. However, nothing could compel the Soviets to relent in their misguided quest to conquer Afghanistan and arms reduction talks stalled.
West of Afghanistan, resentment had been building in Iran against the U.S. ever since 1953 when the CIA helped overthrow their socialist democracy and replaced it with a dictatorship (the Shah) that sold the West cheap oil. America’s role in that overthrow grew in the Iranian imagination as the years passed. Unfortunately for Carter, he reaped what President Eisenhower and his successors sowed. The only accommodation the Shah had made to free speech was within mosques, so anti-Western sentiment fused with fundamentalist Islam there over the decades. When Islamic revolutionaries took over the country in 1978, the Shah escaped to Egypt and Mexico, then sought cancer treatment in the U.S. after Henry Kissinger insisted on it in exchange for endorsing Carter’s SALT II treaty. Granting the Shah exile was the straw that broke the camel’s back for the new leaders who seized the American embassy in Tehran, capturing diplomats and Marines in the process.
At first, Carter hoped the Iranian Hostage Crisis could divert Americans’ attention away from the domestic economy, but that backfired as the crisis wore on and ABC’s Nightline covered angry Iranians burning Uncle Sam in effigy on a nightly basis. For Americans still in a post-Vietnam funk, it was like getting salt poured in their wound. Finally, Carter ordered a military rescue, but a sandstorm compromised the mostly helicopter-based operation and a C-130 tanker aircraft crashed, killing eight. The fiasco only raised the prestige of revolutionary leader Ayatollah Khomeini, exiled by the Shah in 1963.
The one big feather in Carter’s foreign policy cap was negotiating peace between Israel and its most formidable adversary, Egypt. Carter built on the Shuttle Diplomacy initiated by Henry Kissinger in the early 1970s, whereby the U.S. no longer supported Israel unconditionally but rather tried to broker peace between Israel and its neighbors. Building on an idea raised by CBS News’ Walter Cronkite in a split-screen interview with the leaders of Israel and Egypt, Menachem Begin and Anwar el-Sadat, Carter invited both to Camp David, Maryland for a retreat. At first, he had to walk from one end of the compound to the other to relay messages, but he eventually got both men in the same room to talk through interpreters. In the Camp David Accords, Israel agreed to swap the Sinai Peninsula in exchange for Egypt’s recognition of Israel’s right to exist. The two have been at peace ever since, though the democratic revolution in Egypt in 2011 threatened the relationship because, potentially, a populist Muslim Brotherhood in Egypt might favor war with Israel. As for Sadat, his own army assassinated him during a parade for negotiating with Israel, underscoring the resistance to peaceful compromise that Middle Eastern leaders face among their own populations. A similar fate awaited Israeli leader Yitzhak Rabin after he laid out a framework for peace with Palestinians within Israel in the 1993 Oslo Accords. He was killed by a right-wing Israeli opposed to peace.
For Carter, his success with Israel and Egypt wasn’t enough to offset setbacks in Iran and Afghanistan. In retrospect, Afghanistan was causing the Soviets more harm than the Soviets were causing the U.S., but Iran plagued Carter as he approached reelection in 1980. After months of negotiations to get their assets unfrozen in American banks, Iranians released American hostages within minutes of when Carter left office. As new president Ronald Reagan took the inaugural oath in January 1981, freed hostages hit the airport tarmac.
Morning In America
It’s hard to say whether the Iranian Crisis cost Carter re-election or not. By 1980, the time was right for the Reagan Revolution as Americans were ready for a conservative change of pace. The Misery Index (stagflation) as economists came to call it, set the stage for Republican victory by Californian Ronald Reagan over Carter in 1980. The actor and former Democrat had veered right in the early Cold War and become governor of California in 1966 after campaigning for Barry Goldwater’s presidential run in 1964, saying of his defection that “I didn’t leave the Democratic Party; it left me.” At the 1976 GOP convention — the last contested convention to date — Reagan had nearly stolen the nomination from Gerald Ford. With his telegenic charisma and jocular charm, the “Gipper” gave a rousing speech even as the party crowned Ford as their candidate. Reagan had played dying Notre Dame football player George Gipper in Knute Rockne: All-American (1940).1980 was a watershed election, on par with 1932 in terms of swinging the American political pendulum back toward the right, just as ’32 had swung it to the left. For the first time since the 1930s, Republicans managed to pry away a significant chunk of blue-collar workers. Many of these Reagan Democrats wanted to restore military pride or, in the case of some Christians, opposed the Democrats’ pro-choice abortion platform. The Supreme Court legalized abortion in Roe v. Wade (1973). Reagan tapped into public skepticism about government agencies and programs launched under Johnson in the 1960s like welfare, public housing, and food stamps, whereas Goldwater had taken on a Social Security program that was popular among working-class Whites. As we read in Chapter 16, one commentator said that “Goldwater lost against the New Deal, but Reagan won against the Great Society.” Reagan’s motto in 1980 was “Let’s Make America Great Again” but there weren’t red LMAGA hats.
One of Reagan’s first speeches after his nomination was outside Philadelphia, Mississippi, where three civil rights workers were killed in 1963. Why Philadelphia? Obviously, Reagan and his advisers weren’t just throwing darts at a map and picking random towns instead of larger cities. While not endorsing segregation or violence, he used the occasion to remind the townspeople of how much he’d always appreciated their commitment to states’ rights in the context of a talk on the innocuous subject of education. Like many who went before him, he’d learn to couch racism in an anti-federal context, especially appealing to suburban racists who, like Reagan, obsessed over not being labeled racist. This time-honored political trick is known as a dog-whistle (for whistles that dogs can hear but not humans, like coded messages that racists hear but not mainstream voters).
In the 1960s, Reagan likewise opposed the national government’s role in the 1964 Civil Rights Act and ’65 Voting Rights, which he denounced as “humiliating” the South and later tried to weaken as president. Other times he was franker: as California’s governor, he appealed to Whites in areas like Orange County by denouncing the 1968 Fair Housing Act, arguing that “If an individual wants to discriminate against Negroes or others in selling or renting his house, it is his right to do so.” In a 1971 phone call to Richard Nixon that he didn’t realize was being recorded, he referred to African leaders as “monkeys…still uncomfortable wearing shoes,” in keeping with his characterization of American cities as “jungles” and frequent jokes about African cannibalism. Nixon admired how Reagan played on the “emotional distress of those who fear or resent the Negro, and who expect Reagan somehow to keep him ‘in his place.’ ”
But Reagan was craftier on-the-record by the late 1970s, exploiting the public’s resentment toward people who were taking unfair advantage of the welfare system. He asked working-class white audiences if they were tired of working hard for their paycheck then going to the grocery store and seeing a “strapping young buck” ahead of them in line with food stamps. Prior to the Civil War, “bucks” or “studs” referred to male slaves that masters encouraged to reproduce with “wenches.” During his 1976 campaign, Reagan popularized the term welfare queen for women who took advantage of the government’s well-intentioned idea of paying single unemployed moms more than married couples. The term derived from an African-American Chicago woman named Linda Taylor who, in 1974, was caught defrauding the government with multiple identities and sentenced to prison. Reagan’s audience understood who the strapping young buck and welfare queen were. He wasn’t going to beat people over the head with explicit racism (he wasn’t stupid), but neither was he going to leave the old southern Democratic voters and George Wallace supporters on the table (because, again, he wasn’t stupid). In an infamous 1981 interview (YouTube), South Carolina GOP strategist Lee Atwater explained how his party won over racists without sounding overtly racist.
Reagan was continuing with a variation on the GOP’s Southern Strategy, begun under Nixon to siphon off the old Democratic “negrophobes.” The concern over welfare abuse transcended race, though. In Hillbilly Elegy: A Memoir of a Family & Culture In Crisis (2016), J.D. Vance recalled how working-class Whites in the Appalachia of his youth resented other lazy Whites that ate (and drank) better than they did by staying on the public dole permanently, spurring the workers to abandon the Democrat Party.
Republicans gained control of the South, fulfilling LBJ’s prophecy about the Civil Rights movement, partly through various Southern Strategies on race, partly through their general limited government philosophy (including cracking down on welfare abuse among all races), and partly through their new alliance with Christian Fundamentalists. Fundamentalism had been growing since the 1970s and abortion, legal since Roe v. Wade in 1973, gave Republicans an excellent wedge issue to galvanize their alliance around, along with Christian nationalism. Consequently, many Reagan Democrats, North and South, Protestant and Catholic, crossed the aisle and voted GOP for the first time, regardless of their economic class and Republican hostility toward unions that would lack the right to collective bargaining in a free market. Economically, they were either willing to sacrifice their interests on behalf of outlawing abortion and strengthening the military or were convinced that helping the wealthy would ultimately create working-class jobs through trickle-down economics (more below on Reaganomics).
Evangelicals didn’t agree on whether this political alliance was a wise deal, given the amorality of politics. Jerry Fallwell bridged Fundamentalism and the GOP with his Moral Majority, but the most famous evangelist, Billy Graham, who counseled every president from Harry Truman to Barack Obama, urged Christians to avoid partisanship (right). Either way, religion took on a revived role in American politics during the Carter-Reagan era and after. Carter was born-again and wanted to carry Christ’s message of peace into the real world. Reagan, too, was interested in the New Testament, especially its last chapter, the Book of Revelations, that he suspected might foreshadow an apocalyptic showdown between America and the USSR. He secured a lasting alliance between Christian fundamentalists and the Republican Party. Today no candidate could run for president without fully explaining his or her faith. Mormons (e.g. Mitt Romney 2012) and Jews are more or less welcome to join the sweepstakes along with Christians, while this 2019 Gallup poll suggests that other groups’ prospects are at least improving.
Economics, though, was where the right-wing Reagan set himself apart the most from Democrats. Just as FDR wanted to jump-start the economy through government spending, Reagan’s supply-side economics reversed the concept of Keynesian demand-side stimuli, focusing not on government spending but on tax cuts, especially for the wealthy and corporations. Several major corporations were essentially on welfare throughout Reagan’s presidency because their tax rebates exceeded their tax bills. It’s difficult to tell how much the wealthy were actually paying on income taxes prior to 1980, but the top rates went from 70% when Reagan came into office down to 28% by 1988, so they were the biggest and most obvious beneficiaries of his election.
Was all this “Reagan-Hood” as his critics charged? In other words, did Reaganomics really steal from the poor and give to the rich, the opposite of the legendary Robin Hood? Yes and no. He helped the rich plenty, but his record was mixed on the poor. He reduced food stamps, most forms of student aid (e.g. Pell Grants), and painkillers from disability coverage, leading to a black market in drugs like oxycodone. However, spending continued through Reagan’s presidency on most of the core New Deal programs and even much of the welfare from the Great Society. Some tax burdens were shifted to the states but still came out of paychecks just the same. Really, Reaganomics kicked off an era when Americans continued to spend on core entitlements (Social Security and Medicare) while voting themselves tax cuts.
Reagan’s budget director, David Stockman (right), a follower of Austrian free-market economist Friedrich Hayek (Chapter 9), quickly learned the limits of what a full-blown Reagan economic revolution would entail. Stockman, the “father of Reaganomics,” realized that cutting most non-military spending would decimate “Social Security recipients, veterans, farmers, educators, state and local officials, [and] the housing industry…democracy had defeated the [free market] doctrine.” Senate Majority Leader Mitch McConnell reiterated this point in 2018 when Republicans had control of both chambers and the White House, meaning that, unless they’d been bluffing all these years, they could finally cut Social Security and Medicare all they wanted to balance the budget and offset tax cuts. That would’ve required legal changes rather than just lowering the amounts (apportionments) because those entitlements come under mandatory spending rather than discretionary, but they could’ve changed the law. The problem is that most GOP voters who favor lower spending and smaller government don’t actually want cuts to include Social Security or Medicare, just Medicaid for the poor. In an unprecedented burst of honesty, McConnell said the pain that comes with meaningful cuts makes it “difficult if not impossible to achieve when you have unified government” (translation: we need the Democrats to win back at least one chamber so that they can share the blame among a voting public that, collectively, wants higher spending for lower taxes). The truth is that democracies and budgets don’t dovetail well. To be precise, in America’s case they’re off by ~ 15% as measured by 2016’s $3.8 trillion in expenditures versus $3.3 trillion in revenue. Even consistently running huge deficits and burdening their grandchildren with the tab leaves the public complaining about insufficient services and high taxes. In 2020, Donald Trump asked a private audience at Mar-a-Lago “Who the hell cares about the budget?” The current answer is Republicans when a Democrat is in office and nobody when Republicans are in office.
The result of Reagan’s concession to core New Deal programs, when combined with increased military spending and tax cuts, was ballooning debt. Rather than always trying to sell the multiplier effect that tax cuts would increase revenue, Reagan often candidly told American voters that he was willing to plunge the country into debt to win the Cold War if that’s what it took. And, Reagan likely knew (wisely if cynically) that overspending helps a sitting president and punishes his successors — a lesson his successors took to heart. George W. Bush’s VP Dick Cheney said Reagan proved that, in politics, “deficits don’t matter.” Adjusted for inflation, Abraham Lincoln and Franklin Roosevelt are the runaway leaders in growing the size and cost of the federal government because of the Civil War and World War II. But aside from them, what presidents oversaw the most growth in the size of the national government? Surprisingly, George W. Bush (87%) and Reagan (82%) in non-inflation-adjusted numbers (Source: USGovernmentSpending.com). This is as good a time as any to remind ourselves that, while presidents submit budget proposals under the Constitution, Congress is in charge of the nation’s purse strings as far as setting budgets, though presidents sign off. Given our current partisanship, it’s heartwarming that Reagan got along well with House Speaker Tip O’Neill (D-MA), but really their bipartisan budget compromises just meant mounting debt for future Americans.
Reagan’s supporters often claim that the debt-to-GDP ratio actually shrank under Reagan, meaning that the economy grew more than the debt, and the ratio of federal spending to the overall economy fell, but that’s not the case. The Debt-to-GDP ratio grew from ~30% to 40% in the 1980s (at the top of the chapter we looked at annual spending vs. GDP rather than debt). According to the multiplier theory of supply-side economics, lower tax rates were supposed to stimulate growth enough to increase overall tax revenues even though tax rates dropped but that didn’t happen. Nonetheless, the economy took off on a long bull run, lasting through the late 1990s. By that measure, and a booming stock market, Reaganomics was a success.
But did increased wealth “trickle-down” to workers as supply-side advocates promised? Again, yes and no. There’s no doubt the economy grew over the next twenty years, and the booming stock market of the 1980s and ’90s helped all workers tied to defined-benefit pensions and directed-benefit 401(k) or IRA retirement funds, along with stimulating overall growth. As the graph to the left shows, working classes didn’t suffer significant wage reductions on average between 1980 and 2007, just before the Great Recession. The overall earning power of most workers stagnated, though, except for people in the upper 20%. Also, most workers didn’t stay at one job long enough to take full advantage of either type of retirement fund, and most weren’t investment-savvy enough to manage their own directed funds to the fullest advantage.
Many economists claim that it’s wrong to look at the economy like a pie and ask who is getting the biggest piece because the pie itself is growing. That’s true to a certain extent and, yes, it’s true that America’s poor suffer more from diseases of abundance (diabetes, obesity, etc.) than hunger. But regardless of the size of the pie, it is finite at any given moment, and the gap between rich and poor has widened between 1980 and the present, with most of the money trickling up. Really, it was more like a tidal wave than a trickle. While the rich have gained more proportionally than the working and middle-classes, the ultra-rich (top 1%) have gained far more than the poor, middle or even regular rich. The top 1% lost ground in the Great Recession of 2008-09, but when the slow recovery kicked in around 2010, they increased their lead over the bottom 99%. Much of that wealth is in the hands of entrepreneurs who’ve created jobs and products for the rest, but much of it has gone to investment bankers and hedge-fund managers who stash their earnings in offshore accounts to avoid paying taxes. For them, deregulating Wall Street was a welcome part of Reaganomics — the trickle-down part not so much. For pennies-on-the-dollar, they lobbied politicians to legislate tax loopholes (to wit: the amount needed to bribe/lobby the politicians was far less than the amount the saved on taxes). Tax shelters, combined with the fact that taxes are lower on investments than earned income, mean that most wealthy now pay a lower effective tax rate than the middle and upper-middle classes and, if you don’t like it, you’re guilty of fomenting “class warfare.”
By 2012, the richest 400 people in the U.S. had more money than the bottom 50% of the population and wages stood at an all-time low as a percentage of GDP. As we saw in Chapter 5, one of the wealthiest, Warren Buffet, suggested that millionaires pay a 30% minimum tax rate regardless of whether their earnings are from work or investments (some already do). The majority of Americans favor the idea, but it’s difficult to see the Buffet Rule going anywhere given politicians’ need to win campaign donations from the very people (0.06% of the population) who would suffer under the heavy hand of “totalitarian government” in that scenario. Opponents argue that keeping taxes lower on investments than earned income spurs growth and that the revenue increase would be minor anyway.
Another key to the Reagan Revolution was deregulation. Reagan was adamant in his aforementioned philosophy that government is not the solution to our problem(s); government is the problem. The deregulatory trend started under Carter in the 1970s but gained momentum under Reagan. He rolled back environmental and workplace safety regulations and got rid of many rules governing banking and accounting. The financial changes, especially the evolution of Special Purpose Entities (SPE’s), contributed to problems like the Savings & Loan Crisis, Michael Milken’s junk bond-related fraud, and Enron in the late 1990s. The taxpayer bailout of corrupt Savings & Loans cost Americans 3-4% of GDP between 1986 and ’96. On the other hand, allowing accountants to “cook their books” may have stimulated the economy and Milken used some of the money he stole to help fund medical research. But, ideally, one purpose of accounting, other than to run a business responsibly for your own sake, is so that other people (employees, investors, IRS) can get a feel for what’s going on, not what’s not going on. Reagan also cut funding for the Small Business Administration (1953-) the only government agency aimed at helping small entrepreneurs, but one that could be spun as yet more bureaucracy and costing taxpayers because of some failed loans.
Mergers were another hallmark of Reaganomics, as courts were hesitant to prosecute monopoly cases. Remember all the hullabaloo about trust-busting in the Progressive Era? In 1986, an upstart football league called the USFL led by, among others, Donald Trump, sued the NFL in an antitrust case. A lower court determined that the USFL was right — the NFL did have a monopoly on pro football — and awarded the USFL a grand total of $1. While that particular case wasn’t influential, it symbolized the era. The Reagan Revolution paved the way for a wave of mergers in the 1980s and ’90s. Unions didn’t do much better than trust-busters in the 1980s. Shortly into Reagan’s presidency, air traffic controllers went on strike. He had the FAA order them back to work and fired over 11k that refused.
Deregulation also impacted communications in 1986 when the FCC (Federal Communications Commission) removed their 1949 Fairness Doctrine requiring TV and radio broadcasts to be fair and “tell both sides of a story.” Such a rule was arguably a violation of the First Amendment, but its retraction fragmented news into what it is today, where most conservatives and liberals just listen to spins on their own tribe’s websites, networks or radio shows, with little center of gravity in the middle to rely on for “straight news.” Critiquing what he later called the “Four Corners of Deceit” — science, government, media, and academia — Rush Limbaugh launched his radio show in 1988, two years after the deregulation. The Sinclair Broadcast Group brought a right-leaning if subtler approach to local news at the hundreds of network affiliates it bought in the South and Midwest. Most news now is what people used to refer to as op-ed (for opinions and editorials) within each tribe’s echo chamber or filter bubble. If deregulation opened up television and radio, the advent of the Internet obliterated any hope of an agreed-upon reality. That can be true even when opposing sides consult the same source, as is the case with the Washington Post’s Fatal Force crime database, based on FBI statistics. People can Tweet® or vent on blogs with like-minded people, or have profanity-laced exchanges with faceless adversaries in comment boxes that go nowhere. Mostly we just talk past each other, as shown in this image that, contrary to first impressions, wasn’t taken by the Hubble Telescope but rather purportedly mapped 2016 election-oriented traffic in the Twitterverse:
In addition to the basic responsibilities of citizenship, like voting and paying taxes, Americans now have to filter news to find out what’s going on. Unfortunately, many of us aren’t up to the task, with the result that we not only disagree — perfectly normal and healthy, even necessary, in a democracy — but that we aren’t disagreeing based on agreed-upon facts. For instance, the 2015 Iran Deal would’ve been complicated enough without huge swaths of Americans thinking that Obama’s administration paid Iran $150 billion in U.S. taxpayer money (ransom, in effect) instead of the truth that they unfroze ~$100 billion in Iranian assets locked up in overseas banks. During Merrick Garland’s failed Supreme Court nomination, many Americans believed in a mythical Biden Rule. Likewise, Sarah Palin’s death panel hoax warped an already complicated Obamacare debate, with ~30% of Americans believing it. In Spring 2020, the same percentage thought COVID-19 was created deliberately in a lab, with suspects ranging from the Chinese to Microsoft founder Bill Gates (PEW), while their arsonous kindred spirits in Europe scapegoated 5G phone towers. Mikki Willis’ Plandemic (2020) video misinformed millions more Americans precisely when they needed solid information, including dubious claims that masks and flu vaccines can cause coronavirus. Facing public pressure in July 2020, the Sinclair Group refrained from running a documentary on local stations blaming NIAID Director Anthony Fauci for starting the pandemic. Journalist/historian Garrett Graff wrote that the media is “so polluted that it should be an Environmental Protection Agency Superfund site.”
The new media business model often doesn’t go far beyond maximizing “eyeballs” or “hits,” with little money to be made teaching reality or encouraging intelligent debate because many people find that boring. Right-wing media profited from amplifying the death panel hoax in ways that would’ve been illegal prior to 1986. Since it’s human nature to suffer from confirmation bias and easier to confirm preconceptions, most people just choose their “truths” from a virtual buffet table of options and some politicians learned from Palin to exploit the “post-truth” era. At a 2019 Trump rally in Mississippi, a reporter asked a supporter if it mattered whether what he was hearing was true. The candid young man thought for a moment then said: “He tells you what you want to hear…And I don’t know if it’s true or not—but it sounds good, so [expletive] it.” (Atlantic)
In 2018, Trump’s chief strategist Steve Bannon told journalist Michael Lewis: “Anger and fear is what gets people to the polls…the Democrats don’t matter. The real opposition is the media. And the way to deal with them is to flood the zone with s**t.” Trump launched his political career by promoting the Birther Movement that questioned Obama’s American citizenship, and he accused Obama of founding the ISIS terrorist group. During Obama’s presidency, the number of Americans who thought that he was Muslim grew from ~20% to ~30% (he is a Hawaiian-born non-denominational Christian with a Kenyan father and white mother from Kansas). Billy Graham’s son, Franklin, was another prominent Birther. When he took office, Trump flipped the script by accusing his critics of fake news, unwittingly echoing George Orwell in 1984 when telling his supporters that “what you’re seeing and what you’re reading is not what’s happening” (some people now use the term fake interchangeably with bias). As president, Trump falsely accused Hillary Clinton of rigging the 2016 election, suggested that conservative Supreme Court judge Antonin Scalia was murdered, and forwarded other dubious Tweets®, including one claiming that Navy SEAL Team Six conspired to fake Osama bin Laden’s assassination and another of a Muslim mob throwing a little boy off a building in London. When confronted about the fake Muslim video, Trump’s Press Secretary Sarah Huckabee Sanders said: “Whether it’s a real video, the threat is real and that is what the president is talking about.” Likewise, college football coach Mike Leach Tweeted® a fake montage of an Obama speech — re-ordered, edited, and spliced into a different speech because, as he told ESPN’s E:60, he preferred the fake video to the real Obama and thought it would stimulate a better discussion on the proper role of government.
The post-truth era has truly arrived. Similarly, if more along the lines of old-fashioned misrepresentation as opposed to outright falsehoods, Bernie Sanders patched together chunks of video to form a 2020 commercial suggesting that Obama supported him, though the two mostly didn’t get along. If the real Obama was too centrist for our liking, we can at least take solace that fictionalized Obamas will live on in perpetuity, like Colonel Sanders on Kentucky Fried Chicken commercials.
Limbaugh admired Trump’s knack for re-posting hoax conspiracies then distancing himself the next day by saying that he was merely sharing others’ views. In keeping with that strategy, after pushback on the SEAL Team Six accusation, Trump distinguished between originating fake news and spreading it, arguing that the latter helped people decide for themselves. But his Secretary of State Rex Tillerson warned about declining standards in a Spring 2018 graduation speech at the Virginia Military Institute: “If our leaders seek to conceal the truth or we as people become accepting of alternative realities that are no longer grounded in facts, then we as American citizens are on a pathway to relinquishing our freedom.” Disinformation and unsubstantiated conspiracy theories are a problem across the political spectrum. At the height of the Truther Movement, many Americans believed that George W. Bush or the U.S. was behind the 9/11 attacks, or at least knew of them in advance, despite no quality evidence to support that (Polls). While Bannon and Trump’s campaign digital media director Brad Parscale (right) pioneered and refined sketchy micro-targeting on social media, Democrats and other Republicans will no doubt follow suit if such tactics continue to be successful. Facebook even provides coaching for politicians interested in micro-targeting and CEO Mark Zuckerberg initially pledged to allow disinformation in future elections to support free speech and to avoid having tech companies moderate content. During Roy Moore’s 2017 senatorial campaign in Alabama, one liberal posted disinformation indicating Moore’s support for banning alcohol (he lost the special election), triggering a debate within the Democratic Party as to whether they’d need to “fight fire with fire” to compete in the new environment.
Always remember to judge theories based on the quality of evidence, not the quantity. The potential quantity of evidence for wrong theories is infinite, but the potential quality is not (See Rear Defogger #25 on the tab above). Quality media is part of the Anglo-American buffet table, too, as suggested by Spanish actor Vanessa Otero in the chart below. A more thorough site, referenced by sources as wide-ranging as Huffington Post and Breitbart News, is Media Bias/Fact Check. The point here isn’t the accuracy of Otero’s 2016 diagram — with which presumably nearly anyone with a pulse would quibble (underscoring my point) — but rather that many Americans only feed off the edges of the buffet table then share posts on social media that friends and relatives either already agree with or ignore. (It’s also dated as Trump’s presidency has pushed some formerly centrist/near-left outlets like CNN further to the left.) There’s little constructive debate on the edges of the spectrum and more often cherry-picking or even bogus disinformation and misinformation, aka fake news.
If you’ll pardon some soapbox advice, consult fact-checking sites like the Pulitzer Prize-winning Politifact or Snopes before posting politically-volatile material on social media. Politifact includes a page for Facebook hoaxes and another called Punditfact that assesses journalists instead of politicians. If you care about the United States, it’s your job to self-regulate by cross-checking. A handful of us at least should stay grounded in reality just to man the fort while others indulge in fantasy. Are these fact-checking sites perfect? Undoubtedly not, but what is? Don’t let perfection be the enemy of good. Politifact and Snopes call people out across the political spectrum. Those brave souls daring to venture outside their bubbles should try out the daily newsletter Flipside, that curates the best sources on the right and left. The Week also curates partisan spins on key news items.
The term fake news was commonly used in the early 20th century by people across the media spectrum (see optional article below), which is why historians see this as a return to form. Irresponsible media spinning was out of control in free-speech Britain as 17th-century newspapers misinformed English readers of fake or exaggerated Catholic atrocities, stoking the Irish Confederate Wars. Students who’ve taken HIST 1301 might remember the epic mudslinging of the 1800 presidential election, when John Adams had his voters worrying that Thomas Jefferson was going to confiscate their Bibles, while Jefferson’s camp paid for favorable newspaper editorials and called Adams a “howling hermaphrodite” — a hideous hermaphroditical character which has neither the force nor firmness of a man nor the gentleness and sensibility of a woman.” Joseph Pulitzer and William Randolph Hearst frequently accused each other of publishing fake news (sometimes accurately) as they competed for New York’s sensationalized Yellow Journalism market. Really the idea, if not the exact term, has roots as old as journalism itself 10k years ago. The better part of the 20th century was, in retrospect, a unique time in journalism when readers valued objectivity. Most newspapers self-identified as Democratic or Republican with op-ed pages that leaned left or right, but the rest of the paper was neutral and factual or at least aimed to be. Exceptions were sensationalist tabloids near the grocery check-out. That ideal of objectivity has been more the exception than the rule, historically, though. From the late 18th to the early 20th centuries, no one made much pretense toward objectivity.
In modern America, media fragmentation is happening among an increasingly diverse population with access to more democratic and potentially deceptive technology. Unfortunately, this might get more complicated in the future as video footage gets less reliable in the deepfake era. Still, there’s an unseen benefit to all this if you have time and the fortitude to stomach it. If you can transcend your filter bubble and expose yourself to a wide spectrum of media, at least higher-grade versions of it, you can be better informed today than someone who simply watched the “straight” nightly news fifty years ago. We’ll revisit the media landscape more below in our discussion of amped-up partisanship, as the two reinforce each other.
Drawing congressional district boundaries also took on greater importance in American politics in the late 20th century. As with media bias, this too has a longer history, is impacted by technology, and transcends the Reagan-era conservative resurgence. Gerrymandering, named after Revolutionary-era Massachusetts politician Elbridge Gerry’s salamander-shaped district (right), is as old as American politics and stems from the problem of how to divide up a state’s congressional districts. Senators don’t present this problem because each state gets two drawn from all the state’s voters. However, with the House of Representatives, there is no perfect or fair way to map districts given the irregular shape of most states and the shifting population within them. Even in a square state like Wyoming, dividing into four even squares wouldn’t be even in terms of population distribution, although Wyoming’s population is so low that it’s not a problem anyway because they have only one representative in the House; the whole state is one district. Gerrymandering manipulates district borders to maximize one party’s capacity to win votes by herding opponents’ voters into as few districts as possible, or by spreading and diluting those votes across districts. This diagram, if a bit hyperbolic in using “steal” in its title, shows two simplified versions with right angles:
We could discuss gerrymandering at any point in the course but, by the late 20th century, computer software enhanced the efficiency of redistricting. Also, like the media deregulation mentioned in the previous section, enhanced gerrymandering has amplified partisanship. Both phenomena divide and sort us. Per the Voting Rights Act of 1965, courts have generally struck down gerrymanders aimed at racial discrimination but sanctioned those aimed at partisan discrimination (unless it’s too extreme). Several states voted to transfer redistricting duties from legislatures to bipartisan commissions in the 2018 mid-terms (add more green to the map below), but those were swing states like Michigan, Ohio, and Missouri. California, though, is a blue (Democratic) state that unlike red (Republican) Texas has also transferred redistricting to a bipartisan commission.
Allowing the party in power to draw up their own redistricting lines, as many states currently do, exacerbates the problem, creating a situation where “politicians pick voters” nearly as much as voters pick politicians. GOP strategist Thomas Hofeller, who redrew North Carolina’s lines, talked openly about that strategy, saying that “redistricting is like an election in reverse.” He advised other Republican strategists to cover their tracks by avoiding email but, when he died in 2018, his liberal daughter found evidence on his hard drive that, for him, the point of gerrymandering in North Carolina and potentially requiring a citizenship question on the census was to suppress minority voting (NYT) — evidence that could come into play the next time the higher court revisits the topic (lower courts already shot down partisan gerrymandering in North Carolina in 2019). In 2020, the 53 Republicans serving in the Senate received 24 million fewer votes than their 47 Democratic counterparts. Today, about two-thirds of the most egregiously gerrymandered states are controlled by Republicans (e.g. Texas and Pennsylvania), but they aren’t alone. In Democrat-controlled Ilinois, their 4th Congressional District as of 2017 barely met the requirement that districts must be contiguous (below).
Texas, one of the 38 yellow states above, is a notorious example. Democrats once drew up the lines to favor themselves and, since the Reagan Revolution of the 1980s, Republicans have done likewise. Texas Attorney General Ken Paxton (R) insists that, in compliance with Gill v. Whitford (2018), gerrymandering in the Lone Star state is merely intended to discriminate against Democrats rather than minorities. But there’s a lot of overlap since minorities have tended to vote Democrat since 1964. Austin, for instance, is the “bluest” (most liberal) city in Texas and one of the most liberal in the country — a “blueberry in a bowl of tomato soup” as comedian Jon Stewart called it (writer Lawrence Wright says conservative Texans see Austin as a “spore of the California fungus that is destroying America”). Under former Congressman Tom Delay (R), the GOP created the 25th District, aka the “Fajita Strip,” to condense Democratic voters into one district — writing that district off but limiting Democrats’ overall impact.
When courts shot down the Fajita Strip, the GOP created a map whereby five of Austin’s six Congressional representatives were Republican. The two basic ways to Gerrymander are “packing and cracking” and the GOP has used both on Austin. To pack is to herd like-minded constituents into one district such as the Fajita Strip, to minimize their impact. A second way is to divide a city like Austin by cracking it into the tips of multiple wedges that fan out into large enough conservative districts that the end result is a liberal city like Austin represented by mostly conservative congressman (see Districts #10 & #17 below).
When packed rather than cracked, Gerrymanders result in clear red and blue districts whose voters don’t elect centrist candidates. In the House of Representatives, that results in a diverse collection of staunch partisans who are elected for the very purpose of doing battle with the opposing party and not compromising. Many conservative voters/constituents see compromise as a sign of weakness and force their candidates to pledge that they won’t, just as many Barack Obama supporters — and perhaps, subconsciously, even his opponents — saw the former president’s willingness to compromise as a weakness. Gerrymandering also increases the motivation to either cheat or rig the system in those few swing counties that aren’t red or blue. If voters themselves aren’t more divided than they were in the past, they’re at least better sorted. They choose their own sets of experts and facts from the media buffet table, vote for candidates from districts deliberately made as partisan as possible and, according to some studies (but not all) have increasingly come to self-segregate by moving to “red and blue” areas to be around like-minded people. Consequently, congressional districts with “swing voters” have shrunk steadily over the last thirty-five years. In Gerrymandered non-swing districts, one’s biggest threat is often someone more stridently conservative or liberal within his or her own party, not an opponent from the opposing party. This creates parties less open to compromise and debate across the aisle. The current chair of the Democratic National Committee, Tom Perez, for instance, has ruled that the right-to-choose (an abortion) isn’t open to negotiation among Democrats; pro-life Democrats, in other words, aren’t welcome to debate the issue at platform meetings or compromise with Republicans if elected. In Texas, only District #23 above, stretching from the western suburbs of San Antonio all the way to El Paso, was up for grabs before 2018 (more are now). This mostly Hispanic district is currently led by ex-CIA, African-American Republican and outspoken Trump critic Will Hurd. Hurd was student body president at Texas A&M during the bonfire tragedy of 1999.
Let’s return to the 1980s to explore more issues that impact us now. The post-Reagan era kicked off with what seemed then like a depressing campaign — one that served mainly to underscore the superficiality of media coverage and tendency in democracies for campaigners to manipulate voters by appealing to their worst instincts. That wasn’t because either of the candidates was bad. The 1988 race pitted Reagan’s VP, George H.W. Bush against Democrat Michael Dukakis of Massachusetts. Since New England is generally more liberal than the rest of the country, the Democrats balanced the ticket with Lloyd Bentsen of Texas, hoping to recapture the Austin-Boston magic of the 1960 Kennedy-Johnson ticket. Dukakis led midway through the summer, but Bush’s media consultant Roger Ailes (Chapter 16, future head of FOX News) and campaign manager, the forenamed Lee Atwater, came up with an ad attacking Dukakis’ weakest point besides his unfortunate photo-op in a tank.
As governor of Massachusetts, Dukakis oversaw a prison furlough program that allowed prisoners out on temporary weekend probations. One of the convicts, Willie Horton, broke into a home and raped a woman. Horton was black and a group called Bush for Americans flooded the airwaves with his mug shot (left), asking viewers if they wanted someone soft on crime. Atwater said the ad would “strip the bark off the little bastard” [Dukakis] and “make Horton Dukakis’ running mate” (for VP). Dukakis made the soft-on-crime spin worse by answering no to a tough debate question from CNN’s Bernard Shaw over whether he’d favor the death penalty for someone who raped and murdered his wife. While his answer was clear and he backed it up by arguing that studies showed the death penalty was not a deterrent, viewers were put off that the question hadn’t stirred deeper emotions.
Bush campaigners also made up stories that Dukakis burned American flags to protest the Vietnam War and that his wife was mentally ill, though Bush distanced himself from the false smears. When Atwater was diagnosed with brain cancer a couple of years later, he converted to Catholicism and issued an apology to Dukakis for the “naked cruelty” of the 1988 campaign. At the time, though, it was enough to pull his client ahead in the race and Bush won the election, with no real help from his old boss Reagan, whom he broke with. The Horton ads, along with the Rodney King arrest and riots in 1991-92 and the O.J. Simpson murder trial of 1994-95 that we read about in the Civil Rights chapter, served as unpleasant reminders of America’s ongoing racial conflicts and undertones. In political campaigns, though, even thinly veiled racism tapered off for the most part between the 1990s and 2016, at least among the politicians themselves.
Crime & Punishment
Since such shenanigans are typical of political campaigns, none of this would normally be worth mentioning. However, like the 1964 campaign during the escalation of the Vietnam War, the 1988 campaign had a lasting impact that transcended merely just deciding the next president. Fellow Democrats learned from Dukakis’ Willie Horton fiasco and vowed to be tougher on crime. In a case of bipartisan agreement, they “crossed the aisle” and by and large supported Republican policies of tougher sentencing (Democrats had always supported increased funding for staffing more police). Consequently, American prisons filled with small-time offenders through mandatory minimum sentencing, whereby judges didn’t have the discretion to lower sentences, with the proportion of African American prisoners growing substantially. Unfortunately, the 1994 Crime Bill also eliminated funding for college education within prisons, the thinking being that taxpayers shouldn’t fund free tuition when they were struggling to pay their own. That was short-sighted because studies showed that taxpayers had saved ~$5 for every $1 spent on prisoner tuition due to AA and BA degrees, drastically reducing recidivism rates. Education is a key component in reforming criminals.
However, another rare bipartisan consensus emerged by the late 2010s to reduce prison populations, which are higher in the U.S. than anywhere in the world and are expensive (ca. 2019, prisoners cost ~ $65k/yr). There are many factors contributing to crowded prisons. While incarceration for drug possession is commonly blamed, only 15-20% of today’s inmates are in for drug-related offenses. Federal judges now offer Drug Court treatment options for non-violent offenders. But many people that would’ve been in mental hospitals forty years ago are homeless or in prison. Moreover, petty offenders are taking up too much space, forcing judges to reduce sentences and grant earlier paroles for violent criminals. Supporters of tougher sentencing point to charts like the one on the right and see direct causation: crime has gone down precisely because more criminals are behind bars. In that scenario, the cost of criminality has been shifted (and spread more evenly) from individual victims to taxpayers at large and the streets are safer. Another theory for reduced crime rates is the lead-crime hypothesis, which posits that lead poisoning can increase violent behavior and that crime rates in America rose and fell with increased use of lead in gas, paint, and pipes and fell accordingly with lead’s abatement.
There are other factors impacting serious crime, including better surveillance (especially in the post-9/11 era), GPS-enhanced phone records, more thorough cross-checking of databases with small-time criminals having already entered their fingerprints, and broken windows strategies to reduce serious crime by curbing petty crime and cleaning up garbage. Studies and experience have shown that improving physical surroundings (garbage, graffiti, broken windows, etc.) and strictly enforcing minor crimes lowers the rate of serious crimes like murder, rape, and armed robbery. On the other hand, it’s a short step from curbing minor crimes to racial profiling. In short: the flip-side of “broken windows” strategies is “stop-and-frisk.” Undoubtedly, numerous factors act simultaneously on something as complex as crime rates. Despite the big crime drop-off between 1990 and 2010, the U.S. was still the most violent developed nation, especially in the South.
Either way, increased prison populations had libertarians on the right and civil rights progressives on the left calling for penal reform, and the idea of using prisons just for serious criminals became popular among Democrats and Republicans looking for wedge issue among working classes. In a rare recent case of bipartisan legislation, Congress passed the First Step Act in 2019, relaxing, among other things, minimum mandatory sentencing for non-violent offenders.
George H.W. Bush (hereafter referred to as Bush 41, since he was the 41st president) was popular midway through his presidency because of his handling of the Gulf War — so popular that most front-runners in the Democratic Party chose not to run in 1992. However, having inherited a large deficit from his predecessor Reagan, Bush broke a famous “read my lips” campaign promise and raised taxes to balance the budget. Cynical Democrats goaded him into it as an act of fiscal responsibility, then used it against him. Bush also alienated the Republicans by beefing up the Clean Air Act (1963- ), expanding federal research on pollution and stemming acid rain and ozone depletion. Mostly ignored by conservatives and liberals alike was Bush’s help in fighting HIV-AIDS. Like his son George W. Bush, who later authorized relief to African countries when he was president (2001-09), Bush the elder aided American cities most afflicted by the disease through the 1990 Ryan White CARE Act. Bush 41 also signed congressional legislation outlawing an important type of discrimination with the 1990 Americans with Disabilities Act. According to the younger Bush’s speechwriter, Michael Gerson, his Emergency Plan for AIDS Relief (PEPFAR) resulted from an alliance of liberal global-health advocates and evangelicals who held sway with key congressional Republicans. Bush 43 explained and defended the program by referencing Luke 12:48: “To whom much is given, much is required.”
The country dipped into a mild recession in 1991 and Bush seemed to some detached in his response, possibly because he was smart and realized that mild recessions happen all the time and aren’t the presidents’ fault. Americans, overall, overestimate what presidents can do to revive or harm the economy. It’s not run or handled by a person and, if it were, that person would be Chair of the Federal Reserve. Bush’s problem was more of a public relations crisis than anything else. He tried to connect to the regular people by going to a grocery store but, when he got to the counter and had never seen a scanner, it just added to his image of being a detached blue blood. Bush stumbled into the 1992 campaign vulnerable, facing off against a little-known, young Democratic governor from Arkansas, Bill Clinton, and a colorful computer mogul independent from Texas, Ross Perot (right), who supported protectionism (tariffs) and opposed free trade.
Bill Clinton was prominent in the Democratic Leadership Council, a group of “third way” policy wonks who argued that the Democrats should stay competitive by moving to the right (toward the center) and acquiescing in the Reagan Revolution. Other than Jimmy Carter, the Democrats had been getting pummeled since 1968. At its best, the DLC represented a refreshing and realistic change beyond the usual partisan tire ruts, helping to rid the Democrats of their unconditional support for the welfare state and anti-business reputation. At its worst, it just meant that Democrats would sell out to corporations and high finance, giving Wall Street virtual control over both parties. Liberals organized an anti-DLC conference in Washington under the motto Because One Republican Party Is Enough. While they advocated slightly higher taxes for the rich than Republicans, the new Democrats appealed as much to the leisured classes as the inner-city poor. Charles Schumer (D) of New York defended tax loopholes for hedge fund managers claiming that, like all Senators, he needed to look out for the constituents in his state — in this case, not defined by the 99% of New Yorkers who aren’t billionaires.
A key difference for the new Clinton-led Democrats was their support of free trade. Unions had weakened a lot by the late 20th century and Democrats needed other sources of money, so they turned increasingly to corporate contributions. For leftists, the answer would’ve been to strengthen unions by resisting free trade; for centrist Democrats, the answer was to get realistic and start winning elections, then do their best to barricade against the more extreme aspects of the Reagan Revolution. Clinton Democrats also tacked toward the center culturally. Clinton pledged to stop illegal immigration and, in an interview with civil rights leader Jesse Jackson, Clinton cleverly distanced himself from more radical elements in his own party by condemning rap singer Sister Souljah’s purported call for Blacks to kill Whites during the 1992 L.A. Riots. This technique, not invented in 1992, is now known as a “Sister Souljah Moment” and has been employed by George W. Bush, John McCain, and Barack Obama.
For leftists leery of big cuts to welfare and privatization of student loans, Clinton seemed like “Republican Lite.” For right-wingers, their goal was to spin the Arkansas governor like a communist anyway to prevent the Democrats from making inroads among independent, centrist voters still known as “Reagan Democrats.” That, too, is a time-honored political tactic. Later, Barack Obama would be stuck in the same political purgatory, spun alternatively as a commie or a corporate shill depending on who was doing the spinning. Clinton defined himself as an “Eisenhower Republican,” dedicated to balanced budgets, strong markets, and free trade. Indeed, Clinton’s long-term historical role was ratifying Reagan’s conservative revolution as a Democrat in the same way that Eisenhower ratified FDR’s New Deal as a Republican. Clinton saw globalization (free trade) as good for the American economy and a key to diffusing world conflict, yet the Democrats’ support of globalization opened a window of opportunity for Ross Perot (I), whose protectionist-fueled populism foreshadowed Donald Trump’s nomination among Republicans 24 years later. The Democrats were quietly withdrawing their uncompromised support for unionized labor and hoping that workers wouldn’t notice.
Politics Get Personal
The 1992 campaign also kicked off an era when the public and media have been consumed with the personal biographies of their candidates more than ever before. Previous politicians were from the WWII era and came of age when journalists, for the most part, didn’t pry into politicians’ sex lives or religions. In JFK’s case, his Catholicism was controversial, but not his rampant adultery. By 1992, though, Baby Boomers like Clinton were running for office and people wanted to know more about their backgrounds. Clinton was an inveterate philanderer and women came forward claiming to have had affairs with him. Some were no doubt gold diggers, but there was too much smoke for no fire. And what was Clinton doing in the 1960s? Was he in Vietnam? Was he a protester? Clinton was at Oxford on a Rhodes Scholarship and had smoked pot, but “not inhaled.” He went into the National Guard toward the end of the Vietnam War and got a deferment to study another year overseas. Republicans jumped on Clinton for being a draft dodger but it came back to haunt them eight years later when George W. Bush (Bush 43) ran for president and people discovered he’d also gone into the Guard. Donald Trump got college deferments and a medical disqualification (1-Y/4-F) for short-term heel spurs. Unlike today, the Guard didn’t fight overseas in the 1960s. The GOP had a photo of Clinton from the early ’70s with a beard — enough to spin him as anti-establishment.
For his part, Clinton embraced more universally popular visions of the 1960s, including that of John Kennedy and the Civil Rights movement. The argument over which sixties cut both ways, though, because many voters had grown up in that era and not all of them necessarily held it against Clinton that he’d had a beard, smoked pot, or hadn’t fought in Vietnam. In the three-way race, Clinton won the Electoral College and a plurality of popular votes (43%), with independent Ross Perot garnering nearly 20% despite dropping out two weeks prior to the election. While many commentators assume that Perot’s presence hurt Bush, there’s no solid evidence to support him swinging the election toward Clinton. Perot drew voters from both parties but he attacked Bush more relentlessly than Clinton during the campaign.
A crusade against Clinton commenced with his election, more passionate and well-funded than those normally launched by political opponents. The same would occur with his successors, but never really had on this scale with his predecessors since the 1790s. Such venomous and one-sided dialogue became a hallmark of American politics from the 1990s forward, thriving in a deregulated era of cable TV and the Internet. By and large, conservatives have used radio more effectively while liberals dominate comedy. Clinton’s biggest nemeses were Limbaugh and (during his second term) the Drudge Report, that blended fiction and reality. Clinton’s team was young and inexperienced and the Republicans were able to “throw him off message” once he got in the White House by distracting him from the issues that made him popular among voters, such as his centrist business policies. They forced the issue of gays in the military since they knew it was a no-win issue among the public. Clinton endorsed equal rights for gays in the military during his campaign, but then hedged in office and came up with the muddled don’t ask, don’t tell policy, pleasing neither side.
After stumbling out of the gate on that, Clinton made another strategic error, which was to grant too much responsibility to his wife Hillary in crafting healthcare insurance legislation. Hillary was plenty capable, but some opponents were leery of her as an educated career woman, symbolized by her pantsuits. The couple had even talked of a “two-for-one deal” on the campaign trail and Hillary was the first First Lady to have an office in the West Wing alongside other senior staff. Other than Eleanor Roosevelt and Edith Wilson in 1920 (after Woodrow’s stroke), First Ladies had stayed in the East Wing and avoided a meaningful role in policy. While their role is ambiguous, First Ladies (or future First Gentlemen) are not elected officials. It’s been understood that the president’s wife should limit herself to innocuous things like promoting literacy or physical fitness or discouraging drug use among kids, with maybe some ribbon-cutting here and there, occasionally necessitating a hard hat. At the same time, American healthcare insurance was a system sorely in need of repair (more in Chapter 21).
1994 Mid-Terms & 1996 Election
Clinton stumbled out of the gate for two reasons in 1993. First, like the other state governors who became president in the post-Watergate era (especially Jimmy Carter), he brought with him to Washington a fairly inexperienced team. Second, his administration’s attempt to repair the nation’s troubled health insurance system, about which you’ll read in the following chapter, mostly failed. The Speaker of the House of Representatives, Newt Gingrich of Georgia, seized on Clinton’s first-term problems and spearheaded the Republican Revolution of 1994, whereby the GOP took over both houses of Congress in the mid-term elections. The Republicans hadn’t controlled the House since the late 1940s. Its leaders were predominantly southern, including the Georgian Gingrich and Texans Tom Delay and Dick Armey. Presaging future threats against Barack Obama, North Carolina Senator Jesse Helms warned Clinton to not visit his state “without a bodyguard.” Just as Nixon presided during Lyndon Johnson’s Great Society era even though he was a Republican, the Democrat Clinton was presiding over a Republican congress during the ongoing Reagan Revolution. Partisan gridlock began to kick in as the growing debt issue inherited from previous administrations reared its head.
The U.S. had been “in the red” since 1969 and, as Senator and ex-Democrat Richard Shelby (AL-R) pointed out, President Reagan had run the country further into debt. Working across the aisle with Republicans like John Kasich of Ohio and taking advantage of the post-Cold War peace dividend, Bill Clinton was the first president since 1969 to balance annual budgets (receipts=outlays), but the overall, long-term debt didn’t go away. Deficits are annual shortfalls whereas the debt is the running total. Republicans wanted a balanced budget amendment but weren’t specific as to where they’d make cuts. Democrats laid out more specific ideas for cuts but wouldn’t chain themselves to an amendment, citing circumstances like wars or the Louisiana Purchase where governments need to run deficits. The U.S. can’t really get its long-term budget under control without moderate tax hikes or cuts to entitlements, but a wide swath of Americans in both parties like Medicare Social Security. Meanwhile, conservative policy wonks like Grover Norquist force politicians to sign pledges against raising taxes. In a democracy, blame for reckless budgets ultimately falls on the divided citizenry that, collectively, wants more for less in an atmosphere that discourages compromise.
Gingrich promised a Contract with America that would pass a balanced budget amendment, reform (reduce) welfare, make the day-to-day workings of Congress more efficient, roll back Social Security, and cut funding for environmental initiatives like the Superfund and Safe Drinking Water Act. He shortened the workweek of Congress to three days so that they could spend the other two “dialing for dollars” (raising money). Gingrich favored cutting benefits for the working poor and taxes for the rich, blaming increasing wealth disparity on “radical seculars.” In cutting back on the size of congressional committees and limiting the power of senior committee chairs, Gingrich’s reforms seemingly made Congress leaner and more transparent. However, in a classic case of how reform can have unforeseen consequences, by the early 21st century, it became harder for party leaders (or anyone, for that matter) to assert leadership in Congress as increased transparency made everyone less likely to make deals that might alienate their constituents. They didn’t respond to party “whips” and majority leaders so much as lobbyists and the voters back home watching on C-SPAN. That’s good in a way, and more democratic, but it disrupts dealmaking in the proverbial “smoke-filled back room.”
These reforms in the House of Representatives built on earlier attempts by the so-called “Watergate Baby” Democrats elected in the wake of the namesake scandal. They, too, objected to the glacial pace of the House with its seniority-dominated committees and unwritten codes against young Congressmen speaking up and initiating bills or amendments. After all, the supposed point of American democracy is to enable voters to effect change, especially in the lower house of Congress. But historian John Lawrence traces some of today’s lack of compromise to their infusion of political rights into the debate, which conservative Republicans later countered with their own rights:
Veterans of the long struggles for civil rights, for women, for children, for the environment, for people with disabilities, these new legislators articulated their agenda not merely as policy objectives but as constitutional and ethical “rights” with a profoundly moral dimension: a right to an abortion, a right to clean water and air, a right to consumer safety. While the practice occurred among liberals in the class of 1974, it increasingly appeared among conservatives as well: a right to gun ownership, a right to life for unborn fetuses, a right to lower taxes, a right to less government, a right to freedom from government regulation. Elevating policy goals to the status of rights would prove to be a crucial step in the evolution of ideological partisanship in the United States. The application of such a moral dimension to the framing of public issues served to diminish the attractiveness of compromise in pursuit of a common objective.
Rights are as American as apple pie and one thing nearly all Americans appreciate; they just appreciate different ones. In their purest forms, rights don’t lend themselves to compromise when applied to too many specific issues, creating gridlock. Bipartisanship — the idea of the competing political parties working together “across the aisle” (of the Senate or House floors) on compromise — has generally been seen as a positive thing in American history, even though conflict and disagreement are natural and healthy parts of the system, too. In 2013, the Arizona Republican Party censured longtime senator, 2008 presidential candidate and Vietnam veteran John McCain for having cast too many bipartisan votes in his career; he was officially declared guilty, in other words, of having cooperated across the aisle and voting his mind.
The 1980s, in retrospect, were the last decade (so far) in American history when politicians of opposing parties socialized together. Newt Gingrich’s popularity in the mid-1990s signaled a new type of confrontational politics that demonized opponents and discouraged bipartisanship and compromise of the sort that made Great Society and Reagan Revolution legislation possible in the 1960s and ’80s. The key was to hammer home simple messages on selected media that misrepresent the opposition and assume they aren’t acting in good faith. Gone were the days of Democratic-Republican whiskey, poker, and golf, as new rules sent Congressmen home on the weekends, ostensibly on behalf of staying in touch with citizens. Gingrich re-arranged schedules so that congressmen would fly home to their constituents over the weekend rather than stay in Washington socializing with colleagues. He hoped that, instead of compromising, Republicans could win and hang on to all the branches, asserting one-party rule.
As a history Ph.D. and former professor, Gingrich understood how politicians can help hammer home peoples’ worldviews through repetition. In 1990, he and his GOPAC action committee issued a list of negative terminology: corrupt, betray, bizarre, cheat, devour, disgrace, greed, steal, sick, traitors, shallow, radical, pathetic, anti-flag, advising Republicans to never speak of Democrats without associating them with those terms and never interpret a national tragedy or crime without tying it to Democrats. Democrats, ideally, would become people that reasonable Americans couldn’t respect enough to compromise with. Gingrich spread conspiracy theories and smears against House Speaker Jim Wright (D-TX) until he was tossed out of office. GOPAC resisted all bipartisan legislation on the hopes that the public would blame Democrats if Republicans sabotaged Congress by making it more dysfunctional. Henceforth, threat of a government shutdown would loom over future negotiations between Congress and the president. Gingrich correctly linked his fame to “noise” and conflict rather than legislation, since that’s what made the news. Echoing Ayn Rand, he declared to one reporter that “people like me are what stand between us and Auschwitz [Nazi concentration camp],” perpetuating a longstanding equation among the fringe right between Democrats and Nazis. Later, conservative commentators Glenn Beck and Anne Coulter warned that Obamacare would lead to another Holocaust.
“Newtspeak” (a play on Orwell’s Newspeak) mandated calling the opposition the “Democrat Party” or “Dems” because Gingrich feared the adjective Democratic had positive connotations. Even today he avoids the term, blaming congressional dysfunction and partisanship on the party “of socialism and anti-Semitism.” The comic strip Doonesbury called Gingrich’s memo the Magna Carta of attack politics. Gingrich’s strategy dovetailed with the proliferation of cable TV as Roger Ailes aligned FOX News with the GOP (Ailes also mentored MSNBC’s Rachel Maddow and once offered to pay her a full-time salary to not work). Such partisanship, regardless of which side of the aisle it originates on, is usually accompanied by disingenuous complaints that the opposition is who doesn’t want to cooperate and criticism is deflected with whataboutism, as in what about when your side did something similar or worse? (Latin tu quoque for “you also”). George Washington warned in his 1797 Farewell Address about how partisanship makes the country more vulnerable to foreign interference:
It serves always to distract the public councils and enfeeble the public administration. It agitates the community with ill-founded jealousies and false alarms, kindles the animosity of one part against another, foments occasionally riot and insurrection. It opens the door to foreign influence and corruption, which finds a facilitated access to the government itself through the channels of party passions. Thus the policy and the will of one country are subjected to the policy and will of another.
Gingrich’s embrace of partisanship, combined with aforementioned media fragmentation and the decline of swing districts proved remarkably effective at getting citizens to subconsciously place a higher premium on cultural and political tribalism than patriotism or policy. The merit of any given policy is often reduced to whether or not it has a (D) or (R) attached to it. One big advantage of negativity and fear-mongering is that it raises more money than promises to craft legislation through compromise. While partisanship runs deeper in American history than most people realize — the 1790s Newspaper War and elections of 1800 and 1824 come to mind, for instance — politics no doubt grew increasingly toxic starting in the 1990s, or at least toxic politics grew more mainstream and sensationalist media grew more profitable into what commentators called the “outrage-industrial complex.”
Gingrich, then retired, led the infamous meeting at the Caucus Room Steakhouse on Obama’s inauguration night in January 2009, at which the GOP resolved to block Obama’s attempts at bipartisanship (Frank Luntz organized the dinner). Obama had campaigned on unity and naively thought that he could work across the aisle to craft legislation. Instead, the GOP rebuffed him and blamed him for being divisive, leading Obama to give up such efforts midway through his second term and circumvent Congress with executive orders. The Tea Party’s perception that Congressmen John Boehner and Eric Cantor were negotiating a budget deal with President Obama ruined their careers, as well.
In the Gingrich era of negative partisanship, we don’t just use partisanship as a tool to get things accomplished (policy), but actually elevate hatred toward the other tribe as the highest ideal, or at least the only common denominator holding parties together. Partisanship is no longer a means to an end; hyper-partisanship is the end goal because it’s the most profitable in media and in campaign fundraising.
It’s easy to see how unwary listeners reared in this toxic environment could believe in conspiracies like Trutherism, Birthirism, Pizzagate, Sandy Hook Denial, or Jade Helm 15 just to name a few of dozens. Such rumors are as old as politics itself, but in the late 20th and early 21st centuries, America’s Overton Window of acceptable mainstream discourse widened as the market grew of people disposed to think along those lines, with the Internet and social media fueling the fire and no agreed-upon go-to anchor of neutral information. In other words, these are conspiracies that would’ve been relegated to tabloids like National Enquirer as recently as a generation ago. Future scholars will have to further unravel how it was that we became so unhinged. Freedom of the press is a right that citizens should honor and be grateful for — the kind Americans have fought and died for — but we need to up our game when it comes to filtering news. As Rex Tillerson suggested, our republic depends on it.
Back to the 1990s. What Gingrich’s Contract didn’t promise was to crack down on corporate lobbying, as many people riding the “reform” wave into Washington were there to cash in themselves. The new House Majority Whip in 1994 was Tom Delay of suburban Houston, who went into office promising to reform Washington and left as one of the most corrupt politicians in the modern era. Gingrich overstretched a bit with his Contract, not taking into account that only 38% of Americans had voted in the 1994 midterm elections. Clinton cherry-picked the popular portions of the Contract (welfare reform and the balanced budget — not as an amendment, but at least as a reality for a few years) and held firm against the rest.
Clinton backed Gingrich down and rode the momentum to victory in the 1996 election. He had good economic tailwinds at his back, including improving information technology, the forenamed post-Cold War “peace dividend” of reduced military spending, heavy worker immigration, and Baby Boomers passing through peak years of productivity. And Clinton played to the centrist popularity that helped get him elected in 1992 by beefing up police forces and reforming the worst abuses of the welfare system. Welfare recipients now faced benefit limits, had to look harder for a job, and couldn’t have more kids while on welfare. Clinton defeated Bob Dole, a WWII vet who ran a clean election. At the GOP’s summer convention, Dole played on his seniority and credibility, promising a “bridge to the past.” The Democrats held their convention two weeks later and promised a “bridge to the 21st century.” When it comes to conventions, it sometimes helps to go second.
The 4th Exam for In-Class Lecture Students and 5th Exam for Distance Learning Doesn’t Currently Cover Information From The Sections Below on the Lewinsky Scandal and 2000 Election. The Rest of The Chapter is Optional, but read the conclusion to recap the Reagan Revolution.
Republicans stayed riveted on Bill Clinton’s sex life and purported criminal dealings throughout his two-term presidency. The Judiciary’s Independent Counsel, a special branch formed in the 1970s after Nixon corrupted the regular judiciary during Watergate, investigated Clinton. Congress took over the investigation from the Justice Department because Clinton’s attorney general, Janet Reno, had originally appointed Democrat Robert Fiske as lead investigator. They focused on Whitewater, a 1978 real estate investment the Clintons had made in the Arkansas Ozarks with a shady developer, then later the associated suicide of their friend Vince Foster and, finally, Clinton’s unsettled sexual harassment case with Paula Jones. Anxious to put the issues behind him, Clinton himself authorized the creation of the Independent Counsel investigation. Fiske’s replacement, Republican Ken Starr, suspected all along that the original Whitewater charges against Clinton were bogus. Later, when defending Trump in his 2020 Senate impeachment trial, Starr lamented that impeachments had become partisan. However, outside sources like Richard Mellon Scaife, hoping to get Clinton impeached before his term expired, bankrolled the investigation anyway through the Arkansas Project. Having run up against a dead end on Whitewater and Foster, the counsel’s hope was that if they investigated Clinton’s extramarital affairs (e.g. Paula Jones), he might have already revealed, or reveal in the future, something they didn’t know yet about Whitewater during “pillow talk” with a woman after sex. If a waste of taxpayer money, it at least spared Americans the boredom and effort of confronting al-Qaeda, climate change, and Wall Street instability.
For his part, Clinton couldn’t keep his drawers zipped up, even though he knew his opponents were pining to catch him in an affair. He struck up a relationship with a twenty-five-year-old intern named Monica Lewinsky. She told a co-worker about it, who told Starr’s counsel. They subpoenaed Clinton and videotaped his testimony so that he could remain in the White House while the grand jury viewed him, then slipped the tape to FOX News “by accident.” Under oath in a separate deposition involving Paula Jones, Clinton had claimed he “did not have sexual relations” with Lewinsky, but a semen stain on one of her dresses obtained under warrant suggested otherwise. Thus, Clinton was guilty of perjury, depending on how one defined sex exactly. Did a semen stain prove he’d had sex? Clinton admitted to an “improper physical relationship.” He claimed he hadn’t lied when he testified that he wasn’t in a relationship with Lewinsky because he wasn’t when he was asked, famously saying “it depends on what your meaning of is is.” Clinton was notorious for this kind of ambiguity, as evidenced by the non-inhaling aspect of his marijuana smoking. The perjury and obstruction of justice charges hinged on whether the actual allegations of oral sex constituted sex in the way that intercourse did. The House of Representatives determined that it did and impeached Clinton, sending the trial to the Senate.
Here’s where things started going awry for the Republicans. They’d been trying to get rid of Clinton for two terms, which was virtually unprecedented in American history. Normally the opposing party just counters policy through the accepted checks-and-balances system and tries to win the next election. The Constitution doesn’t authorize impeaching presidents because of disagreement over policies or unpopularity. The GOP would get their wish, though, if perjury could be defined as a high crime or misdemeanor, the Constitutional bar for impeachment. But the public was put off by the lurid details of the Starr Report, not understanding why Clinton’s investigators had taken such a detailed interest in his sex life.
Then journalists discovered that the Republican’s ringleader, Newt Gingrich (R-GA), was having an affair with his intern as well and that he’d asked his wife for a divorce as she was recovering from cancer. Ironically, Gingrich became the first victim of the Lewinsky scandal — yet another politician impaled by his own sword. After he stepped down, replacement Bob Livingston (R-ID) revealed that he, too, was sleeping with an intern, and resigned in a remorseful press conference. Years later, it came out that Livingston’s replacement Dennis Hastert (R-IL) paid $1.7 million in blackmail to a wrestler he’d sexually abused as a high school coach. He went to prison on child molestation charges in 2016.
By now, the public was also starting to realize that the Starr Commission had been a waste of time because they knew all along there was nothing to Whitewater, and the Clinton trial was looking like a witch-hunt. Did the U.S. really want to set the precedent whereby private citizens like Scaife could finance impeachment campaigns, launched before anyone had any knowledge of criminal wrongdoing? After all, Clinton’s perjury didn’t cause the original investigation; it resulted from it. Why was there an investigation to start with? No good reason, as it turned out, and at the most, nothing that had anything substantive to do with his womanizing. The Senate sensed which way public sentiment was leaning and voted in Clinton’s favor, 55-45, as his approval ratings shot up to the highest of his presidency (73%). All the Democrats voted to acquit and ten Republicans broke ranks and joined them. But Clinton’s goose was cooked, politically. He’d let down many Democratic voters, his VP Al Gore (whom he lied to about Lewinsky) and his wife, Hillary. He likely spent the remainder of his presidency sleeping on the couch.
The Lewinsky Scandal had substantive political fallout. First, it prevented Clinton from working with the GOP on reforming Social Security. Some commentators think that Clinton and the Republicans were working on a deal to partially privatize Social Security prior to Clinton’s impeachment. Second, Republicans remembered that liberals gave Bill Clinton a pass on sexual transgressions, which reemerged in 2016-17 as multiple women accused Donald Trump of harassment and he was even caught on tape bragging about adultery and groping (Hollywood Access Tape). Then news broke that he’d paid hush money to porn star Stormy Daniels. When liberals cried foul, conservatives naturally retorted with whataboutism. Then, in the 2017 #MeToo movement, many women came forward with complaints against prominent actors, politicians, journalists, etc., leading commentators to reconsider the Clinton scandal. In 2017 vocabulary, two of the four recent presidents, Clinton and Trump, were sexual predators (depending on how one defines the term, so too were several past presidents, including Thomas Jefferson and John Kennedy). In retrospect, it might have been best for Democrats to force Clinton to step down. That would’ve allowed VP Al Gore to continue with their agenda and perhaps put Gore in a better position to win the 2000 election since incumbents often have an advantage.
When Al Gore, Jr. ran for president in 2000, in fact, he distanced himself from Clinton because of the Lewinsky Scandal. His opponents were George W. Bush (R) and Ralph Nader, the old Nader’s Raiders lobbyist from the 1960s who brought you seat belts (Chapter 16). Nader wanted to clean up Wall Street corruption, reduce CO2 emissions and, ironically, kick lobbyists out of Washington. Gore had to pander to farmers and coal miners to win the Democratic nomination, so the future subject of An Inconvenient Truth (2006) was ambiguous about reducing carbon emissions. Nader ended up stealing some votes from Gore in the critical state of Florida though only 10% as many as the Democrats in that state who voted for Bush. But, what made this election famous was the controversy on election night over who won and when. The media are in a tough spot calling states (naming the winner) after only 2-3% of the votes are in because each network wants to be first but none wants to be wrong.
Florida is a tough-to-predict “swing state,” with a sizable 22 all-or-nothing electoral votes at stake. No president has won an election without winning Florida since Calvin Coolidge in 1924. Like Ohio, it has a northern and southern contingent culturally speaking, except that in Florida’s case, their South is up north in the panhandle, and their North is down south with Yankee retirees. Complicating matters are Cuban refugees in south Florida that are swing voters themselves, except that in this case they were voting more Republican to punish Democrats for deporting six-year-old refugee Elián Gonzáles back to Cuba. Elián escaped Cuba with his mother and others but their boat capsized, killing everyone except Elián, whom fishermen found floating in an inner tube. Cuban Miami wanted him to stay in Florida with extended relatives but Bill Clinton’s administration and Attorney General Janet Reno returned him to Cuba to reunite with his closest living relative, his father. Gonzáles became a controversial symbol that politicians and media from all over the country argued about even if no one actually cared about the kid. He also likely cost Gore the presidency since Gore inherited blame as Clinton’s VP.
The networks called Florida in Bush’s favor too early, and Gore even called Bush to congratulate him on his win since the entire close election got down to Florida. Gore didn’t hold a press conference to concede the race, though, and he called Bush back to rescind his earlier concession when he started to catch up in Florida. With all the networks using blue to denote states Gore won on their maps, and red for Bush, the terms red and blue states entered the political vocabulary to represent Republican and Democrat, respectively, though earlier elections assigned the two colors randomly. As you can see from the county map above, the GOP dominated sparsely-populated rural areas, while the Democrats did well on the coasts and in cities, and in areas with predominantly Hispanic and black voters. Gore still won counties in south Florida, but not as decisively as he would have without the uproar over Elián Gonzáles (counties don’t go all the way blue or red; all that matters to the electoral college is the state popular vote).
By midnight, Florida had narrowed and was too close to call. Bush wisely spun it as if he’d already won and was being challenged by Gore. Gore never should’ve made the first concession call, though, and the networks shouldn’t have called it in Bush’s favor. Having relatives at FOX, his brother Jeb as governor of Florida, and Florida’s Secretary of State Katherine Harris as his campaign co-chair helped Bush spin the election as a challenged victory rather than what it really was: a virtual tie. In a glaring if unplanned conflict of interest, the person in charge of the recount, Harris, worked directly for Bush. Gore wanted to recount either the entire state or at least certain disputed Democratically-leaning counties and Bush wanted to recount nothing, revealing a lot about what his camp suspected was the case. When they did recount Palm Beach County, they found many Democrats had accidentally voted for Republican Pat Buchanan because of their ballots’ confusing “butterfly” design.
Statewide, many voters of both parties had partially poked a hole through the circle next to their candidate, but accidentally left the “hanging (or dimpled) chad” — the little circle that dangles on the ballot without falling off when counted by antiquated IBM punched card machines. In the meantime, minorities were claiming that white cops intimidated them away from the polls (similar charges arose in Ohio and in South Dakota toward Indians in future campaigns).
The darkest moment — for those of that value the American republic and see it as a global leader in democracy — came in Miami-Dade county, where Bush’s camp hired protesters to crash the county office and successfully intimidate officials into stopping their recount. In their time of need, the GOP called on Watergate alumnus and future Trump operative Roger Stone, who understood the time-honored strategy of rationalizing any coup by blaming the other side of a coup. He later gloated that “one man’s dirty trick is another man’s civic participation.” Bush’s legal adviser, James Baker, flew rioters in from out-of-state to create enough havoc that citizens would want to stop the recount to save democracy and restore law-and-order. Miami’s mayor Alex Penales, still miffed at his fellow Democrats over the Elián Gonzáles affair, was nowhere to be seen to order police to clear the county office of the “Brooks Brothers Riot,” so-named because Republicans are thought to dress better than Democrats. The key to any real law-and-order campaign worthy of mention is to incite criminal activity, as seen by the AK-47-wielding Boogaloo Bois‘ role in Minneapolis in 2020. With no police protection, the election officials freaked out and said that they wouldn’t have time, despite a large team, to count 10k votes in four days and called a halt to the recount.
The Supreme Court bought into the Bush campaign’s spin. SCOTUS ruled 5-4 in Bush v. Gore to stop the recounts ordered by Florida’s Supreme Court because inconsistent recount standards would violate 14th Amendment rights to equal protection and their result “might not reflect well on Bush’s victory,” causing the new president “irreparable harm.” Why he was considered the victor in the first place went unexplained. The ruling was strictly along partisan lines, with five conservative judges and four liberal. It was tortured logic and hypocritical given the GOP’s usual emphasis on states’ rights. When they stopped the recount, Bush was ahead by 537 votes.
But if the Republicans’ intent was to subvert democracy by preventing a recount, Bush might’ve actually won anyway had justice been served according to later independent recounts. When the Miami Herald and USA Today recounted the disputed counties, they found that Bush won Florida by a small margin. When the National Opinion Research Center, a consortium of media outlets, counted all the counties — what should have been done in the first place — they found that naming the winner would’ve depended on how the ballots themselves were counted (the hanging chads, etc.). Gore would’ve won the most restrictive but consistent scenario by 127 votes, and Bush would’ve won the most inclusive but consistent scenario by 110 votes. These studies don’t take voter intimidation into account. If there was even a shred of truth to those charges, that would’ve thrown the election off by more than these razor-thin margins. It was a tough loss for Gore, who defeated Bush nationally in the popular vote. He is one of five candidates in history to win the popular vote and lose the presidency.
The heated controversy surrounding Bill Clinton and the contested 2000 election dovetailed well with an increasingly confrontational media landscape and increased partisanship.
Conclusion: the Reagan Revolution
The Reagan Revolution shaped Republican Party politics from 1980 to 2016 and pushed the Democrats to the right economically. In some respects, we live in the shadow of the Reagan Revolution today. Just as conservatives won office and impacted policy after Roosevelt’s New Deal, from 1933-1980, liberals haven’t been silent since 1980 and two Democrats have won the presidency. But Republicans Eisenhower and Nixon were okay with the New Deal and even expanded the federal government with departments like Health & Human Services (1953-), Environmental Protection Agency (1970-), and Drug Enforcement Administration (1973-). Democrats Clinton and Obama, meanwhile, were no doubt liberal in some respects but pandered to Wall Street.
Language is one of the best indicators of overall trends. No aspiring Democratic politician called himself or herself a liberal today between 1980 and 2012, while Republicans knocked each other out claiming who was most conservative. In the 2000 primary, fellow Democrats accused New Jersey Senator Bill Bradley of being a liberal and he dropped out of the race shortly thereafter. That’s a far cry from progressive Democratic candidates accusing Joe Biden of being liberal in 2019-20 because he was too conservative. Republicans had the overall momentum and upper hand from 1980 to at least 2008, with Democrats on the defensive.
As for the actual size and role of government, it hasn’t budged far in either direction since Reagan took office in 1980. In 2002, it consolidated a number of existing agencies under the Department of Homeland Security and added some smaller banking and consumer protection agencies after the financial meltdown of 2008. Beyond that, the biggest growth in government at the national level has been the post-9/11 increase of the National Security Agency (1952-) in domestic eavesdropping and Obamacare mandating insurers to expand coverage and citizens to be insured (the GOP dropped the mandate in 2017). There have been no dramatic increases or reductions in entitlement programs like Social Security beyond prescription drug coverage for Medicare (Bush 43), and tax rates stabilized with the top bracket oscillating between 36-39.6%.
Up until the 2008 economic crises, it was understood that regulation is bad and deregulation is good. There were even people who blamed the 2008 financial meltdown on too much regulation even though the derivatives markets that imploded had virtually no oversight. Federal income taxes show no signs of increasing to anything near pre-1980 levels. Corporations that contributed around 30% of the country’s revenues in the 1950s now pay 6%, with some of the bigger firms on welfare (in 2010, General Electric earned $14 billion and paid Uncle Sam $-3.2 billion). The U.S. now ranks last among developed nations in upward mobility.
Because its jobs pay better, high finance lures more of the country’s top graduates than law, medicine, science, or industry. Rather than just fueling the economy by lending to other businesses and encouraging constructive investments, finance itself is the biggest business in America, and most investing is purely speculative, high-frequency, and short-term. These changes aren’t just the result of presidential administrations, but rather overall structural changes in the economy, including trends toward globalization, outsourcing, etc. But the corporate-friendly politics of Reagan and his successors in both parties contributed to an era more financially conservative than postwar America — back in the direction of the late 19th century, but not as extreme. Has this historic shift run its course? With that very question at stake, pundits on the contested cable, radio, and blogosphere outlets fight to shape our minds and future with information and misinformation.
Supplement: The New Yellow Journalism
Backstory, “Stuck: A History of Gridlock” (Virginia Foundation for the Humanities)
H.W. Brands, “What Reagan Learned From FDR,” History News Network, 5.15
Robert Barnes, “Efforts To Limit Partisan Gerrymandering Falter At the Supreme Court,” Washington Post, 6.18.2018
Wade Goodwyn, “Texas Governor Deploys State Guard to Stave Off Obama Takeover,” NPR, 5.2.15
McKay Coppins, “The Man Who Broke Politics,” Atlantic, 11.18
Peter Wehner, “The Party of Reagan is No More,” TIME, 3.10.16
Henry Olsen, “How the Right Gets Reagan Wrong,” Politico, 6.26.2017
Matthew Jordan, “A Century Ago, Progressives Were the Ones Shouting Fake News,” Conversation, 2.1.18
John Lawrence, “How the Watergate Babies Broke American Politics,” Politico, 6.5.18
Maggie Astor & K.K. Rebecca Lai, “What’s Stronger Than A Blue Wave? Gerrymandered Districts,” New York Times, 11.29.18
Dan MacGuill, “Texas Governor May Have Emboldened Russian Misinformation Efforts,” Snopes, 5.3.18
Ford Fessenden & John Broder, “Examining the Vote [2000 Election],” New York Times, 11.12.2001
Amanda Robb, “Pizzagate: Anatomy of a Fake News Scandal,” Rolling Stone, 11.16.17
Media Bias / Fact Check