You are looking at 301-320 of 375 articles
The concept of the Third World emerged after 1945 as a way to refer to the developing regions of the world, most often encompassing Africa, Asia, the Middle East, and Latin America. As a descriptive shorthand, the Third World entered common usage to contrast these regions from the capitalist “First World” and the communist “Second World,” even though some nations in these regions overtly aligned with one of the superpowers while others did not ascribe to such classifications. The term thus defies easy categorization and was used by historical actors to reflect different political and economic understandings of their geopolitical status. In the latter half of the 20th century, the Third World also gained purchase among some political leaders to describe non-Western, anti-imperialist, and anti-racist nations that had gained their independence from colonial rule and worked together to resist Cold War alignment. However, Third World leaders struggled to sustain their transnational solidarity, sometimes dividing along the lines of the broader superpower rivalry, regional or sectarian conflicts, and differing aspirations for world order.
US relations with the Third World were often fraught and complex. The United States did not maintain a single policy toward the Third World as such, reflective of its enormous diversity in language, culture, and politics. Instead, the United States devised foreign policies toward Third World nations according to different perceived imperatives and interests. A major theme in the history of US relations with the Third World was the US government’s overarching effort after World War II to undermine the spread of international communism, against which it deployed a wide range of military, political, social, and economic tools. In kind, Third World political leaders often strove to evade direct control by the United States and entered into diplomatic relations cognizant of the power asymmetries at work in the international system. In the 21st century, scholars and policymakers continue to use the Third World term, though it is often used interchangeably with the Global South and the underdeveloped, developing, and non-Western world.
The relationship between the United States and the island of Ireland combines nostalgic sentimentality and intervention in the sectarian conflict known as the “Troubles.” Irish migration to the United States remains a celebrated and vital part of the American saga, while Irish American interest—and involvement—in the “Troubles” during the second half of the 20th century was a problematic issue in transatlantic relations and for those seeking to establish a peaceful political consensus on the Irish question. Paradoxically, much of the historiography of American–Irish relations addresses the social, economic, and cultural consequences of the Irish in America, yet the major political issue—namely the United States’ approach to the “Troubles”—has only recently become subject of thorough historiographical inquiry. As much as the Irish have contributed to developments in American history, the American contribution to the Anglo-Irish process, and ultimate peace process, in order to end conflict in Northern Ireland is an example of the peacemaking potential of US foreign policy.
The United States was heavily involved in creating the United Nations in 1945 and drafting its charter. The United States continued to exert substantial clout in the organization after its founding, though there have been periods during which U.S. officials have met with significant opposition inside the United Nations, in Congress, and in American electoral politics, all of which produced struggles to gain support for America’s international policy goals. U.S. influence in the international organization has thus waxed and waned. The early postwar years witnessed the zenith of American prestige on the global stage. Starting in the mid- to late 1950s, as decolonization and the establishment of newly independent nations quickened, the United States began to lose influence in the United Nations owing to the spreading perception that its alliances with the European colonial powers placed it on the wrong side of history. As U.N. membership skyrocketed, the organization became more responsive to the needs and interests of the decolonizing states. During the 1970s and early 1980s, the American public responded to declining U.S. influence in the United Nations with calls to defund the organization and to pursue a unilateral approach to international challenges. The role of the United States in the United Nations was shaped by the politics of the Cold War competition with the Soviet Union. Throughout the nearly five decades of the Cold War, the United Nations served as a forum for the political and ideological rivalry between the United States and the Soviet Union, which frequently inhibited the organization from fulfilling what most considered to be its primary mission: the maintenance of global security and stability. After the collapse of the Soviet Union and the peaceful end of the Cold War, the United States enjoyed a brief period of unrivaled global hegemony. During this period, U.S. officials pursued a closer relationship with the United Nations and sought to use the organization to build support for its international policy agenda and military interventionism.
Andrew J. Gawthorpe
From 1965 to 1973, the United States attempted to prevent the absorption of the non-Communist state of South Vietnam by Communist North Vietnam as part of its Cold War strategy of containment. In doing so, the United States had to battle both the North Vietnamese military and guerrillas indigenous to South Vietnam. The Johnson administration entered the war without a well-thought-out strategy for victory, and the United States quickly became bogged down in a bloody stalemate. A major Communist assault in 1968 known as the Tet Offensive convinced US leaders of the need to seek a negotiated solution. This task fell to the Nixon administration, which carried on peace talks while simultaneously seeking ways to escalate the conflict and force North Vietnam to make concessions. Eventually it was Washington that made major concessions, allowing North Vietnam to keep its forces in the South and leaving South Vietnam in an untenable position. US troops left in 1973 and Hanoi successfully invaded the South in 1975. The two Vietnams were formally unified in 1976.
The war devastated much of Vietnam and came at a huge cost to the United States in terms of lives, resources, and political division at home. It gave birth to the largest mass movement against a war in US history, motivated by opposition both to conscription and to the damage that protesters perceived the war was doing to the United States. It also raised persistent questions about the wisdom of both military intervention and nation-building as tools of US foreign policy. The war has remained a touchstone for national debate and partisan division even as the United States and Vietnam moved to normalize diplomatic relations with the end of the Cold War.
James Graham Wilson
The Cold War may have ended on the evening of November 9, 1989, when East German border guards opened up checkpoints and allowed their fellow citizens to stream into West Berlin; it certainly was over by January 28, 1992, when U.S. president George H. W. Bush delivered his annual State of the Union Address one month after President Mikhail Gorbachev had announced his resignation and the end of the Soviet Union. After the Berlin Wall came down, Bush and Gorbachev spoke of the Cold War in the past tense in person and on the telephone. The reunification of Germany and U.S. military campaign in the Persian Gulf confirmed that reality. In January 1991, polls indicated that, for the first time, a majority of Americans believed that the Cold War was over. However, the poll results obscured the substantial foreign and domestic crises, challenges, and opportunities created by the end of the Cold War that occupied President Bush and his national-security team between November 1989 and Bush’s defeat in the 1992 presidential inauguration and the inauguration of William Jefferson Clinton as America’s first post–Cold War president in January 1993.
President Abraham Lincoln signed the law that established the Department of Agriculture in 1862 and in 1889, President Grover Cleveland signed the law that raised the Department to Cabinet status. Thus, by 1900 the US Department of Agriculture had been established for nearly four decades, had been a Cabinet-level department for one, and was recognized as a rising star among agricultural science institutions. Over the first half of the next century, the USDA would grow beyond its scientific research roots to assume a role in supporting rural and farm life more broadly, with a presence that reached across the nation. The Department acquired regulatory responsibilities in plant and animal health and food safety and quality, added research in farm management and agricultural economics, provided extension services to reach farms and rural communities in all regions, and created conservation and forestry programs to protect natural resources and prevent soil erosion and flooding across the geographical diversity of rural America. The Department gained additional responsibility for delivering credit, price supports, supply management, and rural rehabilitation programs during the severe economic depression that disrupted the agricultural economy and rural life from 1920 to 1940, while building efficient systems for encouraging production and facilitating distribution of food during the crises of World War I and World War II that bounded those decades. In the process, the Department became a pioneer in developing the regulatory state as well as in piloting programs and bureaucratic systems that empowered cooperative leadership at the federal, state, and local levels and democratic participation in implementing programs in local communities.
Lindsay M. Chervinsky
From 1775 to 1815, empire served as the most pressing foreign relationship problem for the United States. Would the new nation successfully break free from the British Empire? What would an American empire look like? How would it be treated by other empires? And could Americans hold their own against European superpowers? These questions dominated the United States’ first few decades of existence and shaped its interactions with American Indian, Haitian, Spanish, British, and French peoples. The US government—first the Continental Congress, then the Confederation Congress, and finally the federal administration under the new Constitution—grappled with five key issues. First, they sought international recognition of their independence and negotiated trade deals during the Revolutionary War to support the war effort. Second, they obtained access to the Mississippi River and Port of New Orleans from Spain and France to facilitate trade and western settlement. Third, they grappled with ongoing conflict with Indian nations over white settlement on Indian lands and demands from white communities for border security. Fourth, they defined and protected American neutrality, negotiated a trade policy that required European recognition of American independence, and denied recognition to Haiti. Lastly, they fought a quasi-war with France and real war with Great Britain in 1812.
Paul V. Murphy
Americans grappled with the implications of industrialization, technological progress, urbanization, and mass immigration with startling vigor and creativity in the 1920s even as wide numbers kept their eyes as much on the past as on the future. American industrial engineers and managers were global leaders in mass production, and millions of citizens consumed factory-made products, including electric refrigerators and vacuum cleaners, technological marvels like radios and phonographs, and that most revolutionary of mass-produced durables, the automobile. They flocked to commercial amusements (movies, sporting events, amusement parks) and absorbed mass culture in their homes, through the radio and commercial recordings. In the major cities, skyscrapers drew Americans upward while thousands of new miles of roads scattered them across the country. Even while embracing the dynamism of modernity, Americans repudiated many of the progressive impulses of the preceding era. The transition from war to peace in 1919 and 1920 was tumultuous, marked by class conflict, a massive strike wave, economic crisis, and political repression. Exhausted by reform, war, and social experimentation, millions of Americans recoiled from central planning and federal power and sought determinedly to bypass traditional politics in the 1920s. This did not mean a retreat from active and engaged citizenship; Americans fought bitterly over racial equality, immigration, religion, morals, Prohibition, economic justice, and politics. In a greatly divided nation, citizens experimented with new forms of nationalism, cultural identity, and social order that could be alternatively exclusive and pluralistic. Whether repressive or tolerant, such efforts held the promise of unity amid diversity; even those in the throes of reaction sought new ways of integration. The result was a nation at odds with itself, embracing modernity, sometimes heedlessly, while seeking desperately to retain a grip on the past.
C. J. Alvarez
The region that today constitutes the United States–Mexico borderland has evolved through various systems of occupation over thousands of years. Beginning in time immemorial, the land was used and inhabited by ancient peoples whose cultures we can only understand through the archeological record and the beliefs of their living descendants. Spain, then Mexico and the United States after it, attempted to control the borderlands but failed when confronted with indigenous power, at least until the late 19th century when American capital and police established firm dominance. Since then, borderland residents have often fiercely contested this supremacy at the local level, but the borderland has also, due to the primacy of business, expressed deep harmonies and cooperation between the U.S. and Mexican federal governments. It is a majority minority zone in the United States, populated largely by Mexican Americans. The border is both a porous membrane across which tremendous wealth passes and a territory of interdiction in which noncitizens and smugglers are subject to unusually concentrated police attention. All of this exists within a particularly harsh ecosystem characterized by extreme heat and scarce water.
After World War II, the United States backed multinational private oil companies known as the “Seven Sisters”—five American companies (including Standard Oil of New Jersey and Texaco), one British (British Petroleum), and one Anglo-Dutch (Shell)—in their efforts to control Middle East oil and feed rising demand for oil products in the West. In 1960 oil-producing states in Latin America and the Middle East formed the Organization of the Petroleum Exporting Countries (OPEC) to protest what they regarded as the inequitable dominance of the private oil companies. Between 1969 and 1973 changing geopolitical and economic conditions shifted the balance of power from the Seven Sisters to OPEC. Following the first “oil shock” of 1973–1974, OPEC assumed control over the production and price of oil, ending the rule of the companies and humbling the United States, which suddenly found itself dependent upon OPEC for its energy security. Yet this dependence was complicated by a close relationship between the United States and major oil producers such as Saudi Arabia, which continued to adopt pro-US strategic positions even as they squeezed out the companies. Following the Iranian Revolution (1978–1979), the Iran–Iraq War (1980–1988), and the First Iraq War (1990–1991), the antagonism that colored US relations with OPEC evolved into a more comfortable, if wary, recognition of the new normal, where OPEC supplied the United States with crude oil while acknowledging the United States’ role in maintaining the security of the international energy system.
Michael R. Anderson
American strategy in the Asia Pacific over the past two centuries has been marked by strong and often contradictory impulses. On the one hand, the western Pacific has served as a fertile ground for Christian missionaries, an alluring destination for American commercial enterprises, and eventually a critical launchpad for U.S. global power projection. Yet on the other hand, American policymakers at times have subordinated Asian strategy to European-based interests, or have found themselves embroiled in area conflicts that have hampered efforts to extend U.S. regional hegemony. Furthermore, leading countries in the Asia-Pacific region at times have challenged U.S. economic and military objectives, and the assertion of “Asian values” in recent years has undermined efforts to expand Western political and cultural norms. The United States’s professed “pivot to Asia” has opened a new chapter in a centuries-long relationship, one that will determine the geopolitical fault lines of the 21st century.
Risa L. Goluboff and Adam Sorensen
The crime of vagrancy has deep historical roots in American law and legal culture. Originating in 16th-century England, vagrancy laws came to the New World with the colonists and soon proliferated throughout the British colonies and, later, the United States. Vagrancy laws took myriad forms, generally making it a crime to be poor, idle, dissolute, immoral, drunk, lewd, or suspicious. Vagrancy laws often included prohibitions on loitering—wandering around without any apparent lawful purpose—though some jurisdictions criminalized loitering separately. Taken together, vaguely worded vagrancy, loitering, and suspicious persons laws targeted objectionable “out of place” people rather than any particular conduct. They served as a ubiquitous tool for maintaining hierarchy and order in American society. Their application changed alongside perceived threats to the social fabric, at different times and places targeting the unemployed, labor activists, radical orators, cultural and sexual nonconformists, racial and religious minorities, civil rights protesters, and the poor. By the mid-20th century, vagrancy laws served as the basis for hundreds of thousands of arrests every year. But over the course of just two decades, the crime of vagrancy, virtually unquestioned for four hundred years, unraveled. Profound social upheaval in the 1960s produced a concerted effort against the vagrancy regime, and in 1972, the US Supreme Court invalidated the laws. Local authorities have spent the years since looking for alternatives to the many functions vagrancy laws once served.
Jeffrey F. Taffet
In the first half of the 20th century, and more actively in the post–World War II period, the United States government used economic aid programs to advance its foreign policy interests. US policymakers generally believed that support for economic development in poorer countries would help create global stability, which would limit military threats and strengthen the global capitalist system. Aid was offered on a country-by-country basis to guide political development; its implementation reflected views about how humanity had advanced in richer countries and how it could and should similarly advance in poorer regions. Humanitarianism did play a role in driving US aid spending, but it was consistently secondary to political considerations. Overall, while funding varied over time, amounts spent were always substantial. Between 1946 and 2015, the United States offered almost $757 billion in economic assistance to countries around the world—$1.6 trillion in inflation-adjusted 2015 dollars. Assessing the impact of this spending is difficult; there has long been disagreement among scholars and politicians about how much economic growth, if any, resulted from aid spending and similar disputes about its utility in advancing US interests. Nevertheless, for most political leaders, even without solid evidence of successes, aid often seemed to be the best option for constructively engaging poorer countries and trying to create the kind of world in which the United States could be secure and prosperous.
The transformation of post-industrial American life in the late 20th and early 21st centuries includes several economically robust metropolitan centers that stand as new models of urban and economic life, featuring well-educated populations that engage in professional practices in education, medical care, design and legal services, and artistic and cultural production. By the early 21st century, these cities dominated the nation’s consciousness economically and culturally, standing in for the most dynamic and progressive sectors of the economy, driven by collections of technical and creative spark. The origins of these academic and knowledge centers are rooted in the political economy, including investments shaped by federal policy and philanthropic ambition. Education and health care communities were and remain frequently economically robust but also rife with racial, economic, and social inequality, and riddled with resulting political tensions over development. These information communities fundamentally incubated and directed the proceeds of the new economy, but also constrained who accessed this new mode of wealth in the knowledge economy.
Christopher P. Loss
Until World War II, American universities were widely regarded as good but not great centers of research and learning. This changed completely in the press of wartime, when the federal government pumped billions into military research, anchored by the development of the atomic bomb and radar, and into the education of returning veterans under the GI Bill of 1944. The abandonment of decentralized federal–academic relations marked the single most important development in the history of the modern American university. While it is true that the government had helped to coordinate and fund the university system prior to the war—most notably the country’s network of public land-grant colleges and universities—government involvement after the war became much more hands-on, eventually leading to direct financial support to and legislative interventions on behalf of core institutional activities, not only the public land grants but the nation’s mix of private institutions as well. However, the reliance on public subsidies and legislative and judicial interventions of one kind or another ended up being a double-edged sword: state action made possible the expansion in research and in student access that became the hallmarks of the post-1945 American university; but it also created a rising tide of expectations for continued support that has proven challenging in fiscally stringent times and in the face of ongoing political fights over the government’s proper role in supporting the sector.
Megan Kate Nelson
During the American Civil War, Union and Confederate commanders made the capture and destruction of enemy cities a central feature of their military campaigns. They did so for two reasons. First, most mid-19th-century cities had factories, foundries, and warehouses within their borders, churning out and storing war materiel; military officials believed that if they interrupted or incapacitated the enemy’s ability to arm or clothe themselves, the war would end. Second, it was believed that the widespread destruction of property—especially in major or capital cities—would also damage civilians’ morale, undermining their political convictions and decreasing their support for the war effort.
Both Union and Confederate armies bombarded and burned cities with these goals in mind. Sometimes they fought battles on city streets but more often, Union troops initiated long-term sieges in order to capture Confederate cities and demoralize their inhabitants. Soldiers on both sides were motivated by vengeance when they set fire to city businesses and homes; these acts were controversial, as was defensive burning—the deliberate destruction of one’s own urban center in order to keep its war materiel out of the hands of the enemy.
Urban destruction, particularly long-term sieges, took a psychological toll on (mostly southern) city residents. Many were wounded, lost property, or were forced to become refugees. Because of this, the destruction of cities during the American Civil War provoked widespread discussions about the nature of “civilized warfare” and the role that civilians played in military strategy. Both soldiers and civilians tried to make sense of the destruction of cities in writing, and also in illustrations and photographs; images in particular shaped both northern and southern memories of the war and its costs.
While colonial New Englanders gathered around town commons, settlers in the Southern colonials sprawled out on farms and plantations. The distinctions had more to do with the varying objectives of these colonial settlements and the geography of deep-flowing rivers in the South than with any philosophical predilections. The Southern colonies did indeed sprout towns, but these were places of planters’ residences, planters’ enslaved Africans, and the plantation economy, an axis that would persist through the antebellum period. Still, the aspirations of urban Southerners differed little from their Northern counterparts in the decades before the Civil War. The institution of slavery and an economy emphasizing commercial agriculture hewed the countryside close to the urban South, not only in economics, but also in politics. The devastation of the Civil War rendered the ties between city and country in the South even tighter. The South participated in the industrial revolution primarily to the extent of processing crops. Factories were often located in small towns and did not typically contribute to urbanization. City boosters aggressively sought and subsidized industrial development, but a poorly educated labor force and the scarcity of capital restricted economic development. Southern cities were more successful in legalizing the South’s culture of white supremacy through legal segregation and the memorialization of the Confederacy. But the dislocations triggered by World War II and the billions of federal dollars poured into Southern urban infrastructure and industries generated hope among civic leaders for a postwar boom. The civil rights movement after 1950, with many of its most dramatic moments focused on the South’s cities, loosened the connection between Southern city and region as cities chose development rather than the stagnation that was certain to occur without a moderation of race relations. The predicted economic bonanza occurred. Young people left the rural areas and small towns of the South for the larger cities to find work in the postindustrial economy and, for the first time in over a century, the urban South received migrants in appreciable numbers from other parts of the country and the world. The lingering impact of spatial distinctions and historical differences (particularly those related to the Civil War) linger in Southern cities, but exceptionalism is a fading characteristic.
Between 1880 and 1929, industrialization and urbanization expanded in the United States faster than ever before. Industrialization, meaning manufacturing in factory settings using machines plus a labor force with unique, divided tasks to increase production, stimulated urbanization, meaning the growth of cities in both population and physical size. During this period, urbanization spread out into the countryside and up into the sky, thanks to new methods of building taller buildings. Having people concentrated into small areas accelerated economic activity, thereby producing more industrial growth. Industrialization and urbanization thus reinforced one another, augmenting the speed with which such growth would have otherwise occurred.
Industrialization and urbanization affected Americans everywhere, but especially in the Northeast and Midwest. Technological developments in construction, transportation, and illumination, all connected to industrialization, changed cities forever, most immediately those north of Washington, DC and east of Kansas City. Cities themselves fostered new kinds of industrial activity on large and small scales. Cities were also the places where businessmen raised the capital needed to industrialize the rest of the United States. Later changes in production and transportation made urbanization less acute by making it possible for people to buy cars and live further away from downtown areas in new suburban areas after World War II ended.
James J. Connolly
The convergence of mass politics and the growth of cities in 19th-century America produced sharp debates over the character of politics in urban settings. The development of what came to be called machine politics, primarily in the industrial cities of the East and Midwest, generated sharp criticism of its reliance on the distribution of patronage and favor trading, its emphatic partisanship, and the plebian character of the “bosses” who practiced it. Initially, upper- and middle-class businessmen spearheaded opposition to this kind of politics, but during the late nineteenth and early 20th centuries, labor activists, women reformers, and even some ethnic spokespersons confronted “boss rule” as well. These challenges did not succeed in bringing an end to machine politics where it was well established, but the reforms they generated during the Progressive Era reshaped local government in most cities. In the West and Southwest, where cities were younger and partisan organizations less entrenched, business leaders implemented Progressive municipal reforms to consolidate their power. Whether dominated by reform regime or a party machine, urban politics and governance became more centralized by 1940 and less responsive to the concerns and demands of workers and immigrants.
Urban politics provides a means to understand the major political and economic trends and transformations of the last seventy years in American cities. The growth of the federal government; the emergence of new powerful identity- and neighborhood-based social movements; and large-scale economic restructuring have characterized American cities since 1945. The postwar era witnessed the expansion of scope and scale of the federal government, which had a direct impact on urban space and governance, particularly as urban renewal fundamentally reshaped the urban landscape and power configurations. Urban renewal and liberal governance, nevertheless, spawned new and often violent tensions and powerful opposition movements among old and new residents. These movements engendered a generation of city politicians who assumed power in the 1970s. Yet all of these figures were forced to grapple with the larger forces of capital flight, privatization, the war on drugs, mass incarceration, immigration, and gentrification. This confluence of factors meant that as many American cities and their political representatives became demographically more diverse by the 1980s and 1990s, they also became increasingly separated by neighborhood boundaries and divided by the forces of class and economic inequality.