You are looking at 181-200 of 380 articles
Nicolas G. Rosenthal
An important relationship has existed between Native Americans and cities from pre-Columbian times to the early 21st century. Long before Europeans arrived in the Americas, indigenous peoples developed societies characterized by dense populations, large-scale agriculture, monumental architecture, and complex social hierarchies. Following European and American conquest and colonization, Native Americans played a crucial role in the development of towns and cities throughout North America, often on the site of former indigenous settlements.
Beginning in the early 20th century, Native Americans began migrating from reservations to U.S. cities in large numbers and formed new intertribal communities. By 1970, the majority of the Native American population lived in cities and the numbers of urban American Indians have been growing ever since. Indian Country in the early 21st century continues to be influenced by the complex and evolving ties between Native Americans and cities.
In the years following the US Civil War, the federal government implemented a campaign to assimilate Native peoples into an expanding American nation and a modernizing American society. As policymakers and social reformers understood it, assimilation required a transformation in Native gender roles, and as a result, Native American women were the targets of several assimilationist initiatives. Native women navigated federal interventions strategically, embracing what was useful, accommodating what was necessary, and discarding what was not. As mothers, grandmothers, and healers, women provided stability for families and communities enduring disruption and coerced change. In the 20th century, Native women embraced new economic and political roles even as they adapted long-standing customs. Many began working for wages; although often confined to menial labor such as domestic service in other women’s homes, growing numbers of Native women also pursued white-collar occupations in the Bureau of Indian Affairs and later in tribal governments. As tribal governance evolved over the course of the century, some women obtained positions on tribal councils and tribal courts. Native women have also made intellectual contributions—as tribal members and ultimately as American citizens—to modern understandings of democracy, citizenship, sovereignty, and feminism. Since the late 20th century, Native women have been at the forefront of movements to revitalize Indigenous languages and cultures.
Wendy L. Wall
The New Deal generally refers to a set of domestic policies implemented by the administration of Franklin Delano Roosevelt in response to the crisis of the Great Depression. Propelled by that economic cataclysm, Roosevelt and his New Dealers pushed through legislation that regulated the banking and securities industries, provided relief for the unemployed, aided farmers, electrified rural areas, promoted conservation, built national infrastructure, regulated wages and hours, and bolstered the power of unions. The Tennessee Valley Authority prevented floods and brought electricity and economic progress to seven states in one of the most impoverished parts of the nation. The Works Progress Administration offered jobs to millions of unemployed Americans and launched an unprecedented federal venture into the arena of culture. By providing social insurance to the elderly and unemployed, the Social Security Act laid the foundation for the U.S. welfare state.
The benefits of the New Deal were not equitably distributed. Many New Deal programs—farm subsidies, work relief projects, social insurance, and labor protection programs—discriminated against racial minorities and women, while profiting white men disproportionately. Nevertheless, women achieved symbolic breakthroughs, and African Americans benefited more from Roosevelt’s policies than they had from any past administration since Abraham Lincoln’s. The New Deal did not end the Depression—only World War II did that—but it did spur economic recovery. It also helped to make American capitalism less volatile by extending federal regulation into new areas of the economy.
Although the New Deal most often refers to policies and programs put in place between 1933 and 1938, some scholars have used the term more expansively to encompass later domestic legislation or U.S. actions abroad that seemed animated by the same values and impulses—above all, a desire to make individuals more secure and a belief in institutional solutions to long-standing problems. In order to pass his legislative agenda, Roosevelt drew many Catholic and Jewish immigrants, industrial workers, and African Americans into the Democratic Party. Together with white Southerners, these groups formed what became known as the “New Deal coalition.” This unlikely political alliance endured long after Roosevelt’s death, supporting the Democratic Party and a “liberal” agenda for nearly half a century. When the coalition finally cracked in 1980, historians looked back on this extended epoch as reflecting a “New Deal order.”
In late 19th- and early 20th-century America, a new image of womanhood emerged that began to shape public views and understandings of women’s role in society.
Identified by contemporaries as a Gibson Girl, a suffragist, a Progressive reformer, a bohemian feminist, a college girl, a bicyclist, a flapper, a working-class militant, or a Hollywood vamp, all of these images came to epitomize the New Woman, an umbrella term for modern understandings of femininity. Referring both to real, flesh-and-blood women, and also to an abstract idea or a visual archetype, the New Woman represented a generation of women who came of age between 1890 and 1920 and challenged gender norms and structures by asserting a new public presence through work, education, entertainment, and politics, while also denoting a distinctly modern appearance that contrasted with Victorian ideals. The New Woman became associated with the rise of feminism and the campaign for women’s suffrage, as well as with the rise of consumerism, mass culture, and freer expressions of sexuality that defined the first decades of the 20th century. Emphasizing youth, mobility, freedom, and modernity, the image of the New Woman varied by age, class, race, ethnicity, and geographical region, offering a spectrum of behaviors and appearances with which different women could identify. At times controversial, the New Woman image provided women with opportunities to negotiate new social roles and to promote ideas of equality and freedom that would later become mainstream.
Luke A. Nichter
Assessments of President Richard Nixon’s foreign policy continue to evolve as scholars tap new possibilities for research. Due to the long wait before national security records are declassified by the National Archives and made available to researchers and the public, only in recent decades has the excavation of the Nixon administration’s engagement with the world started to become well documented. As more records are released by the National Archives (including potentially 700 hours of Nixon’s secret White House tapes that remain closed), scholarly understanding of the Nixon presidency is likely to continue changing. Thus far, historians have pointed to four major legacies of Nixon’s foreign policy: tendencies to use American muscle abroad on a more realistic scale, to reorient the focus of American foreign policy to the Pacific, to reduce the chance that the Cold War could turn hot, and, inadvertently, to contribute to the later rise of Ronald Reagan and the Republican right wing—many of whom had been part of Nixon’s “silent majority.” While earlier works focused primarily on subjects like Vietnam, China, and the Soviet Union, the historiography today is much more diverse – now there is at least one work covering most major aspects of Nixon’s foreign policy.
Nicole Etcheson and Cortney Cantrell
During the Civil War, the entire North constituted the homefront, an area largely removed from the din and horror of combat. With a few exceptions of raids and battles such as Gettysburg, civilians in the North experienced the war indirectly. The people on the homefront mobilized for war, sent their menfolk off to fight, supplied the soldiers and the army, coped without their breadwinners, and suffered the loss or maiming of men they loved. All the while, however, the homefront was crucially important to the course of the war. The mobilization of northern resources—not just men, but the manufacture of the arms and supplies needed to fight a war—enabled the North to conduct what some have called a total war, one on which the Union expended money and manpower at unprecedented levels. Confederate strategists hoped to break the will of the northern homefront to secure southern independence. Despite the hardships endured in the North, this strategy failed.
On the homefront, women struggled to provide for their families as well as to serve soldiers and the army by sending care packages and doing war work. Family letters reveal the impact of the war on children who lost their fathers either temporarily or permanently. Communities rallied to aid soldiers’ families but were riven by dissension over issues such as conscription and emancipation. Immigrants and African Americans sought a new place in U.S. society by exploiting the opportunities the war offered to prove their worth. Service in the Union army certainly advanced the status of some groups, but was not the only means to that end. Nuns who nursed the wounded improved the reputation of the Catholic Church and northern African Americans used the increasingly emancipationist war goals to improve their legal status in the North. The Civil War altered race relations most radically, but change came to everyone on the northern homefront.
The development of military arms harnessing nuclear energy for mass destruction has inspired continual efforts to control them. Since 1945, the United States, the Soviet Union, the United Kingdom, France, the People’s Republic of China (PRC), Israel, India, Pakistan, North Korea, and South Africa acquired control over these powerful weapons, though Pretoria dismantled its small cache in 1989 and Russia inherited the Soviet arsenal in 1996. Throughout this period, Washington sought to limit its nuclear forces in tandem with those of Moscow, prevent new states from fielding them, discourage their military use, and even permit their eventual abolition.
Scholars disagree about what explains the United States’ distinct approach to nuclear arms control. The history of U.S. nuclear policy treats intellectual theories and cultural attitudes alongside technical advances and strategic implications. The central debate is one of structure versus agency: whether the weapons’ sheer power, or historical actors’ attitudes toward that power, drove nuclear arms control. Among those who emphasize political responsibility, there are two further disagreements: (1) the relative influence of domestic protest, culture, and politics; and (2) whether U.S. nuclear arms control aimed first at securing the peace by regulating global nuclear forces or at bolstering American influence in the world.
The intensity of nuclear arms control efforts tended to rise or fall with the likelihood of nuclear war. Harry Truman’s faith in the country’s monopoly on nuclear weapons caused him to sabotage early initiatives, while Dwight Eisenhower’s belief in nuclear deterrence led in a similar direction. Fears of a U.S.-Soviet thermonuclear exchange mounted in the late 1950s, stoked by atmospheric nuclear testing and widespread radioactive fallout, which stirred protest movements and diplomatic initiatives. The spread of nuclear weapons to new states motivated U.S. presidents (John Kennedy in the vanguard) to mount a concerted campaign against “proliferation,” climaxing with the 1968 Treaty on the Non-Proliferation of Nuclear Weapons (NPT). Richard Nixon was exceptional. His reasons for signing the Strategic Arms Limitation Treaty (SALT I) and Anti-Ballistic Missile Treaty (ABM) with Moscow in 1972 were strategic: to buttress the country’s geopolitical position as U.S. armed forces withdrew from Southeast Asia. The rise of protest movements and Soviet economic difficulties after Ronald Reagan entered the Oval Office brought about two more landmark U.S.-Soviet accords—the 1987 Intermediate Ballistic Missile Treaty (INF) and the 1991 Strategic Arms Reduction Treaty (START)—the first occasions on which the superpowers eliminated nuclear weapons through treaty. The country’s attention swung to proliferation after the Soviet collapse in December 1991, as failed states, regional disputes, and non-state actors grew more prominent. Although controversies over Iraq, North Korea, and Iran’s nuclear programs have since erupted, Washington and Moscow continued to reduce their arsenals and refine their nuclear doctrines even as President Barack Obama proclaimed his support for a nuclear-free world.
Christopher J. Castañeda
The modern oil industry began in 1859 with Edwin Drake’s discovery of oil at Titusville, Pennsylvania. Since then, this dynamic industry has experienced dramatic episodes of growth, aggressive competition for market share, various forms of corporate organization and cartel-like agreements, and governmental efforts at regulation and control, as well as monopoly, mergers, and consolidation. The history of the oil industry reflects its capital-intensive nature. Immense sums of money are spent on oil discovery, production, and refining projects. Marketing, transportation, and distribution systems likewise require enormous amounts of financing and logistical planning. Although oil is often produced in conjunction with, or in wells pressurized by, natural gas, the oil industry is distinct from the related natural gas industry. Since its origins in the mid-19th century, the oil industry has developed an industrial structure that emphasizes scale and scope to maximize profits. Profits can be huge, which attracts entrepreneurial efforts on individual, corporate, and national scales. By the late 20th through early 21st century, the oil industry had begun confronting questions about long-term viability, combined with an increasingly influential environmental movement that seeks to reduce fossil fuel consumption and prevent its toxic waste and by-products from polluting human, animal habitats, and natural habitats.
The relationship between organized labor and the civil rights movement proceeded along two tracks. At work, the two groups were adversaries, as civil rights groups criticized employment discrimination by the unions. But in politics, they allied. Unions and civil rights organizations partnered to support liberal legislation and to oppose conservative southern Democrats, who were as militant in opposing unions as they were fervent in supporting white supremacy.
At work, unions dithered in their efforts to root out employment discrimination. Their initial enthusiasm for Title VII of the 1964 Civil Rights Act, which outlawed employment discrimination, waned the more the new law violated foundational union practices by infringing on the principle of seniority, emphasizing the rights of the individual over the group, and inserting the courts into the workplace. The two souls of postwar liberalism— labor solidarity represented by unions and racial justice represented by the civil rights movement—were in conflict at work.
Although the unions and civil rights activists were adversaries over employment discrimination, they united in trying to register southern blacks to vote. Black enfranchisement would end the South’s exceptionalism and the veto it exercised over liberal legislation in Congress. But the two souls of liberalism that were at odds over the meaning of fairness at work would also diverge at the ballot box. As white workers began to defect from the Democratic Party, the political coalition of black and white workers that union leaders had hoped to build was undermined from below. The divergence between the two souls of liberalism in the 1960s—economic justice represented by unions and racial justice represented by civil rights—helps explain the resurgence of conservatism that followed.
Jessica M. Chapman
The origins of the Vietnam War can be traced to France’s colonization of Indochina in the late 1880s. The Viet Minh, led by Ho Chi Minh, emerged as the dominant anti-colonial movement by the end of World War II, though Viet Minh leaders encountered difficulties as they tried to consolidate their power on the eve of the First Indochina War against France. While that war was, initially, a war of decolonization, it became a central battleground of the Cold War by 1950. The lines of future conflict were drawn that year when the Peoples Republic of China and the Soviet Union recognized and provided aid to the Democratic Republic of Vietnam in Hanoi, followed almost immediately by Washington’s recognition of the State of Vietnam in Saigon. From that point on, American involvement in Vietnam was most often explained in terms of the Domino Theory, articulated by President Dwight D. Eisenhower on the eve of the Geneva Conference of 1954. The Franco-Viet Minh ceasefire reached at Geneva divided Vietnam in two at the 17th parallel, with countrywide reunification elections slated for the summer of 1956. However, the United States and its client, Ngo Dinh Diem, refused to participate in talks preparatory to those elections, preferring instead to build South Vietnam as a non-communist bastion. While the Vietnamese communist party, known as the Vietnam Worker’s Party in Hanoi, initially hoped to reunify the country by peaceful means, it reached the conclusion by 1959 that violent revolution would be necessary to bring down the “American imperialists and their lackeys.” In 1960, the party formed the National Liberation Front for Vietnam and, following Diem’s assassination in 1963, passed a resolution to wage all-out war in the south in an effort to claim victory before the United States committed combat troops. After President John F. Kennedy took office in 1961, he responded to deteriorating conditions in South Vietnam by militarizing the American commitment, though he stopped short of introducing dedicated ground troops. After Diem and Kennedy were assassinated in quick succession in November 1963, Lyndon Baines Johnson took office determined to avoid defeat in Vietnam, but hoping to prevent the issue from interfering with his domestic political agenda. As the situation in South Vietnam became more dire, LBJ found himself unable to maintain the middle-of-the-road approach that Kennedy had pursued. Forced to choose between escalation and withdrawal, he chose the former in March 1965 by launching a sustained campaign of aerial bombardment, coupled with the introduction of the first officially designated U.S. combat forces to Vietnam.
Michael E. Donoghue
The United States’ construction and operation of the Panama Canal began as an idea and developed into a reality after prolonged diplomatic machinations to acquire the rights to build the waterway. Once the canal was excavated, a century-long struggle ensued to hold it in the face of Panamanian nationalism. Washington used considerable negotiation and finally gunboat diplomacy to achieve its acquisition of the Canal. The construction of the channel proved a titanic effort with large regional, global, and cultural ramifications. The importance of the Canal as a geostrategic and economic asset was magnified during the two world wars. But rising Panamanian frustration over the U.S. creation of a state-within-a-state via the Canal Zone, one with a discriminatory racial structure, fomented a local movement to wrest control of the Canal from the Americans. The explosion of the 1964 anti-American uprising drove this process forward toward the 1977 Carter-Torrijos treaties that established a blueprint for eventual U.S. retreat and transfer of the channel to Panama at the century’s end. But before that historic handover, the Noriega crisis and the 1989 U.S. invasion nearly upended the projected transition of U.S. retreat from the management and control of the Canal.
Early historians emphasized high politics, economics, and military considerations in the U.S. acquisition of the Canal. They concentrated on high-status actors, economic indices, and major political contingencies in establishing the U.S. colonial order on the isthmus. Panamanian scholars brought a legalistic and nationalist critique, stressing that Washington did not create Panama and that local voices in the historical debate have largely been ignored in the grand narrative of the Canal as a great act of progressive civilization. More recent U.S. scholarship has focused on American imperialism in Panama, on the role of race, culture, labor, and gender as major factors that shaped the U.S. presence, the structure of the Canal Zone, as well as Panamanian resistance to its occupation. The role of historical memory, of globalization, representation, and how the Canal fits into notions of U.S. empire have also figured more prominently in recent scholarly examination of this relationship. Contemporary research on the Panama Canal has been supported by numerous archives in the United States and Panama, as well as a variety of newspapers, magazines, novels, and films.
The creation and evolution of urban parks is in some ways a familiar story, especially given the attention that Frederick Law Olmsted’s work has commanded since the early 1970s. Following the success of Central Park, cities across the United States began building parks to meet the recreational needs of residents, and during the second half of the 19th century, Olmsted and his partners designed major parks or park systems in thirty cities. Yet, even that story is incomplete. To be sure, Olmsted believed that every city should have a large rural park as an alternative to the density of building and crowding of the modern metropolis, a place to provide for an “unbending of the faculties,” a process of recuperation from the stresses and strains of urban life. But, even in the mid-1860s he sought to create alternative spaces for other types of recreation. Olmsted and his partner Calvert Vaux successfully persuaded the Prospect Park commission, in Brooklyn, New York, to acquire land for a parade ground south of the park as a place for military musters and athletics; moreover, in 1868 they prepared a plan for a park system in Buffalo, New York, that consisted of three parks, linked by parkways, that served different functions and provided for different forms of recreation. As the decades progressed, Olmsted became a champion of parks designed for active recreation; gymnasiums for women as well as men, especially in working-class areas of cities; and playgrounds for small children. He did so in part to relieve pressure on the large landscape parks to accommodate uses he believed would be inappropriate, but also because he recognized the legitimate demands for new forms of recreation. In later years, other park designers and administrators would similarly add facilities for active recreation, though sometimes in ways that compromised what Olmsted considered the primary purpose of a public park. Urban parks are, in important ways, a microcosm of the nation’s cities. Battles over location, financing, political patronage, and use have been a constant. Through it all, parks have evolved to meet the changing recreational needs of residents. And, as dominant a figure as Olmsted has been, this is a story that antedates his professional career and that includes the many voices that have shaped public parks in U.S. cities in the 20th century.
Peace activism in the United States between 1945 and the 2010s focused mostly on opposition to U.S. foreign policy, efforts to strengthen and foster international cooperation, and support for nuclear nonproliferation and arms control. The onset of the Cold War between the United States and the Soviet Union marginalized a reviving postwar American peace movement emerging from concerns about atomic and nuclear power and worldwide nationalist politics that everywhere seemed to foster conflict, not peace. Still, peace activism continued to evolve in dynamic ways and to influence domestic politics and international relations.
Most significantly, peace activists pioneered the use of Gandhian nonviolence in the United States and provided critical assistance to the African American civil rights movement, led the postwar antinuclear campaign, played a major role in the movement against the war in Vietnam, helped to move the liberal establishment (briefly) toward a more dovish foreign policy in the early 1970s, and helped to shape the political culture of American radicalism. Despite these achievements, the peace movement never regained the political legitimacy and prestige it held in the years before World War II, and it struggled with internal divisions about ideology, priorities, and tactics.
Peace activist histories in the 20th century tended to emphasize organizational or biographical approaches that sometimes carried hagiographic overtones. More recently, historians have applied the methods of cultural history, examining the role of religion, gender, and race in structuring peace activism. The transnational and global turn in the historical discipline has also begun to make inroads in peace scholarship. These are promising new directions because they situate peace activism within larger historical and cultural developments and relate peace history to broader historiographical debates and trends.
Spencer D. Bakich
The Persian Gulf War of 1990–1991 was something of a paradox. From the American perspective, the war had the hallmarks of a resounding victory. Responding to a flagrant case of interstate aggression by Iraq against Kuwait, the George H. W. Bush administration assembled a substantial international coalition to deter further Iraqi attacks against its neighbors in the Gulf and to compel Saddam Hussein into quitting Kuwait, to avoid war. When the latter proved infeasible, the United States led that coalition in forcibly ousting Iraq’s military from Kuwait, substantially degrading Iraqi combat power in the process. The war’s outcome resulted from an auspiciously altered geopolitical landscape at the end of the Cold War, the overwhelming superiority of American power vis-à-vis Iraq, and a US decision-making process that tightly knitted military and diplomatic objectives into a coherent—and coherently executed—wartime strategy. However, America’s historically lopsided victory in the Persian Gulf War proved fleeting. Iraq’s surviving military forces retained the capacity to crush domestic challenges to the Ba’athist regime and to threaten its Gulf neighbors. President Bush’s vision of a post-war new world order notwithstanding, Gulf security depended heavily on continuing military missions years after the Persian Gulf War ended. Despite wartime tactical and strategic successes, grand strategic success eluded the United States in the years after the war.
In the years after the Civil War, Polish immigrants became an important part of the American working class. They actively participated in the labor movement and played key roles in various industrial strikes ranging from the 1877 Railroad Strike through the rise of the CIO and the post-1945 era of prosperity. Over time, the Polish American working class became acculturated and left its largely immigrant past behind while maintaining itself as an ethnic community. It also witnessed a good deal of upward mobility, especially over several generations. This ethnic community, however, continued to be refreshed with immigrants throughout the 20th century.
As with the larger American working class, Polish American workers were hard hit by changes in the industrial structure of the United States. Deindustrialization turned the centers of much of the Polish American community into the Rust Belt. This, despite a radical history, caused many to react by turning toward conservative causes in the late 20th and early 21st centuries.
The reproductive experiences of women and girls in the 20th-century United States followed historical patterns shaped by the politics of race and class. Laws and policies governing reproduction generally regarded white women as legitimate reproducers and potentially fit mothers and defined women of color as unfit for reproduction and motherhood; regulations provided for rewards and punishments accordingly. In addition, public policy and public rhetoric defined “population control” as the solution to a variety of social and political problems in the United States, including poverty, immigration, the “quality” of the population, environmental degradation, and “overpopulation.” Throughout the century, nonetheless, women, communities of color, and impoverished persons challenged official efforts, at times reducing or even eliminating barriers to reproductive freedom and community survival.
Between 1900 and 1930, decades marked by increasing urbanization, industrialization, and immigration, eugenic fears of “race suicide” (concerns that white women were not having enough babies) fueled a reproductive control regime that pressured middle-class white women to reproduce robustly. At the same time, the state enacted anti-immigrant laws, undermined the integrity of Native families, and protected various forms of racial segregation and white supremacy, all of which attacked the reproductive dignity of millions of women. Also in these decades, many African American women escaped the brutal and sexually predatory Jim Crow culture of the South, and middle-class white women gained greater sexual freedom and access to reproductive health care, including contraceptive services.
During the Great Depression, the government devised the Aid to Dependent Children program to provide destitute “worthy” white mothers with government aid while often denying such supports to women of color forced to subordinate their motherhood to agricultural and domestic labor. Following World War II, as the Civil Rights movement gathered form, focus, and adherents, and as African American and other women of color claimed their rights to motherhood and social provision, white policymakers railed against “welfare queens” and defined motherhood as a class privilege, suitable only for those who could afford to give their children “advantages.” The state, invoking the “population bomb,” fought to reduce the birth rates of poor women and women of color through sterilization and mandatory contraception, among other strategies. Between 1960 and 1980, white feminists employed the consumerist language of “choice” as part of the campaign for legalized abortion, even as Native, black, Latina, immigrant, and poor women struggled to secure the right to give birth to and raise their children with dignity and safety. The last decades of the 20th century saw severe cuts in social programs designed to aid low-income mothers and their children, cuts to funding for public education and housing, court decisions that dramatically reduced poor women’s access to reproductive health care including abortion, and the emergence of a powerful, often violent, anti-abortion movement. In response, in 1994 a group of women of color activists articulated the theory of reproductive justice, splicing together “social justice” and “reproductive rights.” The resulting Reproductive Justice movement, which would become increasingly influential in the 21st century, defined reproductive health, rights, and justice as human rights due to all persons and articulated what each individual requires to achieve these rights: the right not to have children, the right to have children, and the right to the social, economic, and environmental conditions necessary to raise children in healthy, peaceful, and sustainable households and communities.
Rosina A. Lozano
Language rights are an integral part of civil rights. They provide the tools that permit individuals to engage with and participate in society. The broad use of the Spanish language in the United States by both citizens and immigrants—it is the second-most-spoken language in the country by far—has a long history. Spanish was the first European governing language in parts of the future United States that included the Southwest, portions of the Louisiana Purchase, and Florida. The use of the language did not disappear when these regions became part of the United States, but rather persisted in some locales as a politically important language. In the 20th century, Spanish-speaking immigrants entered not just the Southwest and Florida, but also Chicago, New York, the South, Michigan, and other locales across the country in large numbers. Throughout the 20th century and into the 21st century, Spanish speakers and their advocates have reasserted their cultural preference by fighting for monolingual speakers’ right to use Spanish in legal settings, in public, as voters, as elected officials, at work, and in education. The politics of the Spanish language have only grown in importance as the largest influx of Spanish-speaking immigrants ever has entered the United States. This demographic shift makes the longer history of Spanish a crucial backstory for future language-policy decisions.
Andrew J. Falk
Americans in and out of government have relied on media and popular culture to construct the national identity, frame debates on military interventions, communicate core values abroad, and motivate citizens around the world to act in prescribed ways. During the late 19th century, as the United States emerged as a world power and expanded overseas, Americans adopted an ethos of worldliness in their everyday lives, even as some expressed worry about the nation’s position on war and peace. During the interwar period of the 1920s and 1930s, though America failed to join the League of Nations and retreated from foreign engagements, the nation also increased cultural interactions with the rest of the world through the export of motion pictures, music, consumer products, food, fashion, and sports. The policies and character of the Second World War were in part shaped by propaganda that evolved from earlier information campaigns. As the United States confronted communism during the Cold War, the government sanitized its cultural weapons to win the hearts and minds of Americans, allies, enemies, and nonaligned nations. But some cultural producers dissented from America’s “containment policy,” refashioned popular media for global audiences, and sparked a change in Washington’s cultural-diplomacy programs. An examination of popular culture also shows how people in the “Third World” deftly used the media to encourage superpower action. In the 21st century, activists and revolutionaries can be considered the inheritors of this tradition because they use social media to promote their political agendas. In short, understanding the roles popular culture played as America engaged the world greatly expands our understanding of modern American foreign relations.
The People’s (or Populist) Party represented the last major third-party effort to prevent the emergence of large-scale corporate capitalism in the United States. Founded in 1891, the party sought to unite the producers of wealth—farmers and workers—into a political coalition dedicated to breaking the hold of private bankers over the nation’s monetary system, controlling monopolies through government ownership, and opening up unused land to actual settlers. Industrial workers and their unions were initially wary of the new party, but things changed after the traumatic labor unrest of 1894: Coxey’s March, the nationwide coal strike, and the Pullman boycott. At that time, the American Federation of Labor (AFL) debated some form of alliance with the Populists. Although the Federation rejected such an alliance in both 1894 and 1895 by the slimmest of margins, it did elect a labor Populist—John McBride of the United Mine Workers of America (UMWA)—to the presidency in 1894. This Populist insurgency represents the closest that the main body of the nation’s labor movement ever came to forming a labor party resembling those that arose in industrialized Europe, and its failure helps explain why American workers were unable to mobilize politically to challenge the emerging economic order dominated by large corporate enterprises.
While the agrarian leaders of the People’s Party at first sought the backing of industrial workers, especially those associated with the AFL, they shunned labor’s support after the trauma of 1894. Party officials like Herman Taubeneck, James Weaver, and Tom Watson feared that labor’s support would taint the party with radicalism and violence, warned that trade unionists sought to control the party, and took steps designed to alienate industrial workers. They even justified their retreat from the broad-based Omaha Platform (1892) on the grounds that it would drive the trade unionists they called “socialists” from the party.
Courts and legislatures in colonial America and the early American republic developed and refined a power to compel civilians to assist peace and law enforcement officers in arresting wrongdoers, keeping the peace, and other matters of law enforcement. This power to command civilian cooperation was known as the posse comitatus or “power of the county.” Rooted in early modern English countryside law enforcement, the posse comitatus became an important police institution in 18th- and 19th-century America. The posse comitatus was typically composed of able-bodied white male civilians who were temporarily deputized to aid a sheriff or constable. But if this “power of the county” was insufficient, law enforcement officers were often authorized to call on the military to serve as the posse comitatus.
The posse comitatus proved particularly important in buttressing slavery in the American South. Slaveholders pushed for and especially benefited from laws that required citizens to assist in the recapture of local runaway slaves and fugitive slaves who crossed into states without slavery. Though slave patrols were rooted in the posse comitatus, the posse comitatus originated as a compulsory and noncompensated institution. Slaveholders in the American South later added financial incentives for those who acted in the place of a posse to recapture slaves on the run from their owners.
The widespread use of the posse comitatus in southern slave law became part of the national discussion about slavery during the early American republic as national lawmakers contemplated how to deal with the problem of fugitive slaves who fled to free states. This dialogue culminated with the Fugitive Slave Law of 1850, in which the US Congress authorized officials to “summon and call to their aid the bystanders, or posse comitatus” and declared that “all good citizens are hereby commanded to aid and assist in the prompt and efficient execution of this law, whenever their services may be required.” During Reconstruction, the Radical Republican Congress used the posse comitatus to enforce laws that targeted conquered Confederates. After the end of Reconstruction in 1877, Southern states pushed Congress to create what would come to be known as the “Posse Comitatus Act,” which prohibited the use of federal military forces for law enforcement. The history of the posse comitatus in early America is thus best understood as a story about and an example of the centralization of government authority and its ramifications.