You are looking at 181-200 of 341 articles
Substantial numbers of Asian Americans and Asian immigrants moved into suburbs across the United States after World War II, bringing distinctive everyday lifeways, identities, worldviews, family types, and community norms that remade much of American suburbia. Although Asian Americans were excluded from suburbs on racial grounds since the late 19th century, American Cold War objectives in Asia and the Pacific and domestic American civil rights struggles afforded Asian Americans increased access to suburban housing in the 1950s, especially Chinese and Japanese Americans. Following passage of the Immigration Act of 1965 and the Fair Housing Act of 1968, new groups of Asian Americans, particularly Filipino, Vietnamese, Thai, Korean, and South Asian Indian, joined Chinese and Japanese Americans in settling in earnest into all kinds of suburban neighborhoods. At the turn of the 21st century, a majority of Asians resided in the suburbs, which also became the preferred gateway communities for new immigrants who often bypassed urban cores and moved straight to the suburbs when they arrived.
Entrance into highly racialized postwar suburbs defined by white middle-class norms and segregated white privilege did not, however, mean that Asian Americans gained entry or assimilated into whiteness. While many certainly aspired to and reinforced long-standing white suburban ideals, others revamped, contested, and outright fractured dominant notions of the suburban good life. By the 1980s Asian Americans of various ethnic and national backgrounds had transformed the sights, sounds, and smells of suburban landscapes throughout the country. They made claims on suburban space and asserted a “right to the suburb” through a range of social and cultural practices, often in physical places, especially shopping plazas, grocery stores, restaurants, religious centers, and schools. Yet as Asian Americans tried to become full-fledged participants in suburban culture and life, their presence, ethnic expressions, and ways of life sparked tensions with other mostly white suburbanites that led to heated debates over immigration, race, multiculturalism, and assimilation in American society.
The history of post-World War II Asian American suburban cultures highlights suburbia as a principal setting for Asian American experiences and the making of Asian American identity during the second half of the 20th century. More broadly, the Asian American experience reveals how control over the suburban ideal and the making of suburban space in the United States was and remains a contested, layered process. It also underscores the racial and ethnic diversification of metropolitan America and how pressing social, political, economic, and cultural issues in US society played out increasingly on the suburban stage. Moreover, Asian Americans built communities and social networks precisely the moment in which the authentic “American” community was supposedly in decline, providing a powerful counterpunch to those who lament nonwhite populations, particularly immigrants, for fracturing an otherwise unified American culture or sense of togetherness.
American cities expanded during the late 19th century, as industrial growth was fueled by the arrival of millions of immigrants and migrants. Poverty rates escalated, overwhelming existing networks of private charities. Progressive reformers created relief organizations and raised public awareness of urban poverty. The devastating effects of the Great Depression inspired greater focus on poverty from state and federal agencies. The Social Security Act, the greatest legacy of the New Deal, would provide a safety net for millions of Americans. During the postwar era of general prosperity, federal housing policies often reinforced and deepened racial and socioeconomic inequality and segregation. The 1960s War on Poverty created vital aid programs that expanded access to food, housing, and health care. These programs also prompted a rising tide of conservative backlash against perceived excesses. Fueled by such critical sentiments, the Reagan administration implemented dramatic cuts to assistance programs. Later, the Clinton administration further reformed welfare by tying aid to labor requirements. Throughout the 20th century, the urban homeless struggled to survive in hostile environments. Skid row areas housed the homeless for decades, providing shelter, food, and social interaction within districts that were rarely visited by the middle and upper classes. The loss of such spaces to urban renewal and gentrification in many cities left many of the homeless unsheltered and dislocated.
Robert G. Parkinson
According to David Ramsay, one of the first historians of the American Revolution, “in establishing American independence, the pen and press had merit equal to that of the sword.” Because of the unstable and fragile notions of unity among the thirteen American colonies, print acted as a binding agent that mitigated the chances that the colonies would not support one another when war with Britain broke out in 1775.
Two major types of print dealt with the political process of the American Revolution: pamphlets and newspapers. Pamphlets were one of the most important conveyors of ideas during the imperial crisis. Often written by elites under pseudonyms and published by booksellers, they have long been held by historians as the lifeblood of the American Revolution. There were also three dozen newspaper printers in the American mainland colonies at the start of the Revolution, each producing a four-page issue every week. These weekly papers, or one-sheet broadsides that appeared in American cities even more frequently, were the most important communication avenue to keep colonists informed of events hundreds of miles away. Because of the structure of the newspaper business in the 18th century, the stories that appeared in each paper were “exchanged” from other papers in different cities, creating a uniform effect akin to a modern news wire. The exchange system allowed for the same story to appear across North America, and it provided the Revolutionaries with a method to shore up that fragile sense of unity. It is difficult to imagine American independence—as a popular idea let alone a possible policy decision—without understanding how print worked in colonial America in the mid-18th century.
Steven A. Riess
Professional sports teams are athletic organizations comprising talented, expert players hired by club owners whose revenues originally derived from admission fees charged to spectators seeing games in enclosed ballparks or indoor arenas. Teams are usually members of a league that schedules a championship season, although independent teams also can arrange their own contests. The first professional baseball teams emerged in the east and Midwest in 1860s, most notably the all-salaried undefeated Cincinnati Red Stockings of 1869. The first league was the haphazardly organized National Association of Professional Base Ball Players (1871), supplanted five years later by the more profit-oriented National League (NL) that set up strict rules for franchise locations, financing, and management–employee relations (including a reserve clause in 1879, which bound players to their original employer), and barred African Americans after 1884. Once the NL prospered, rival major leagues also sprang up, notably the American Association in 1882 and the American League in 1901.
Major League Baseball (MLB) became a model for the professionalization of football, basketball, and hockey, which all had short-lived professional leagues around the turn of the century. The National Football League and the National Hockey League of the 1920s were underfinanced regional operations, and their teams often went out of business, while the National Basketball Association was not even organized until 1949.
Professional team sports gained considerable popularity after World War II. The leagues dealt with such problems as franchise relocations and nationwide expansion, conflicts with interlopers, limiting player salaries, and racial integration. The NFL became the most successful operation by securing rich national television contracts, supplanting baseball as the national pastime in the 1970s. All these leagues became lucrative investments. With the rise of “free agency,” professional team athletes became extremely well paid, currently averaging more than $2 million a year.
Maureen A. Flanagan
The decades from the 1890s into the 1920s produced reform movements in the United States that resulted in significant changes to the country’s social, political, cultural, and economic institutions. The impulse for reform emanated from a pervasive sense that the country’s democratic promise was failing. Political corruption seemed endemic at all levels of government. An unregulated capitalist industrial economy exploited workers and threatened to create a serious class divide, especially as the legal system protected the rights of business over labor. Mass urbanization was shifting the country from a rural, agricultural society to an urban, industrial one characterized by poverty, disease, crime, and cultural clash. Rapid technological advancements brought new, and often frightening, changes into daily life that left many people feeling that they had little control over their lives. Movements for socialism, woman suffrage, and rights for African Americans, immigrants, and workers belied the rhetoric of the United States as a just and equal democratic society for all its members.
Responding to the challenges presented by these problems, and fearful that without substantial change the country might experience class upheaval, groups of Americans proposed undertaking significant reforms. Underlying all proposed reforms was a desire to bring more justice and equality into a society that seemed increasingly to lack these ideals. Yet there was no agreement among these groups about the exact threat that confronted the nation, the means to resolve problems, or how to implement reforms. Despite this lack of agreement, all so-called Progressive reformers were modernizers. They sought to make the country’s democratic promise a reality by confronting its flaws and seeking solutions. All Progressivisms were seeking a via media, a middle way between relying on older ideas of 19th-century liberal capitalism and the more radical proposals to reform society through either social democracy or socialism. Despite differences among Progressives, the types of Progressivisms put forth, and the successes and failures of Progressivism, this reform era raised into national discourse debates over the nature and meaning of democracy, how and for whom a democratic society should work, and what it meant to be a forward-looking society. It also led to the implementation of an activist state.
Laura A. Belmonte
From the revolutionary era to the post-9/11 years, public and private actors have attempted to shape U.S. foreign relations by persuading mass audiences to embrace particular policies, people, and ways of life. Although the U.S. government conducted wartime propaganda activities prior to the 20th century, it had no official propaganda agency until the Committee on Public Information (CPI) was formed in 1917. For the next two years, CPI aimed to generate popular support for the United States and its allies in World War I. In 1938, as part of its Good Neighbor Policy, the Franklin Roosevelt administration launched official informational and cultural exchanges with Latin America. Following American entry into World War II, the U.S. government created a new propaganda agency, the Office of War Information (OWI). Like CPI, OWI was disbanded once hostilities ended. But in the fall of 1945, to combat the threats of anti-Americanism and communism, President Harry S. Truman broke with precedent and ordered the continuation of U.S. propaganda activities in peacetime. After several reorganizations within the Department of State, all U.S. cultural and information activities came under the purview of the newly created U.S. Information Agency (USIA) in 1953. Following the dissolution of USIA in 1999, the State Department reassumed authority over America’s international information and cultural programs through its Office of International Information Programs.
Commercialized sexuality became a prominent feature of American urban settings in the early 19th century when young men migrated far from the watchful eyes of family as soldiers and laborers. Concentrated in large populations, and unable to afford the comforts of marriage, these men constituted a reliable pool of customers for women who sold sexual access to their bodies. These women turned to prostitution on a casual or steady basis as a survival strategy in a sex segregated labor market that paid women perilously low wages, or in response to family disruptions such as paternal or spousal abandonment. Prostitution could be profitable and it provided some women with a path towards economic independence, although it brought risks of venereal disease, addiction, violence, harassment by law enforcement, and unintended pregnancy. By mid-century most American cities tolerated red-light districts where brothels thrived as part of the urban sporting culture. Fears that white women were being coerced into prostitution led to the “white slavery” scare of the 1910s, spurring a concerted attack on brothels by progressive reformers. These reformers used the emergency of World War I to close public brothels, pushing America’s sex markets into clandestine spaces and empowering pimps’ control over women’s sexual labor. World War II raised concerns about soldiers’ venereal health that prompted the US military to experiment with different schemes for regulating prostitution that had been developed earlier during the Spanish–American War, as well as in the Philippines and Puerto Rico. After the war, the introduction of antibiotics and the celebration of marriage and family nudged prostitution into the margins of society, where women who sold sex were seen as psychologically deviant, yet men who purchased sex were thought to be sexually liberated. The dawning of second-wave feminism gave birth to the sex workers’ rights movement and a new critique of the criminalization of prostitution. Nevertheless, attitudes about prostitution continue to divide activists, and sex workers still bear the brunt of criminalization.
It is virtually impossible to understand the history of the American experience without Protestantism. The theological and religious descendants of the Protestant Reformation arrived in the United States in the early 17th century, shaped American culture in the 18th century, grew dramatically in the 19th century, and continued to be the guardians of American religious life in the 20th century. Protestantism, of course, is not monolithic. In fact, the very idea at the heart of Protestantism—the translation of the Bible into vernacular languages so it can be read and interpreted by all men and women—has resulted in thousands of different denominations, all claiming to be true to the teachings of scripture.
Protestantism, with its emphasis on the belief that human beings can access God as individuals, flourished in a nation that celebrated democracy and freedom. During the period of British colonization, especially following the so-called Glorious Revolution of 1688, Protestantism went hand in hand with British concepts of political liberty. As the British people celebrated their rights-oriented philosophy of government and compared their freedoms with the tyranny of France and other absolute monarchies in Europe, they also extolled the religious freedom that they had to read and interpret the Bible for themselves. Following the American Revolution, this historic connection between political liberty and Protestant liberty proved to be compatible with the kind of democratic individualism that emerged in the decades preceding the Civil War and, in many respects, continues to define American political culture.
Protestantism, of course, is first and foremost a religious movement. The proliferation of Protestant denominations provides the best support for G. K. Chesterton’s quip that “America is a nation with the soul of a church.” Spiritual individualism, a commitment to the authority of an inspired Bible, and the idea that faith in the Christian gospel is all that is needed to be saved from eternal punishment, has transformed the lives of millions and millions of ordinary Americans over the course of the last four hundred years.
Public authorities are agencies created by governments to engage directly in the economy for public purposes. They differ from standard agencies in that they operate outside the administrative framework of democratically accountable government. Since they generate their own operating income by charging users for goods and services and borrow for capital expenses based on projections of future revenues, they can avoid the input from voters and the regulations that control public agencies funded by tax revenues.
Institutions built on the public authority model exist at all levels of government and in every state. A few of these enterprises, such as the Tennessee Valley Authority and the Port Authority of New York and New Jersey, are well known. Thousands more toil in relative obscurity, operating toll roads and bridges, airports, transit systems, cargo ports, entertainment venues, sewer and water systems, and even parking garages. Despite their ubiquity, these agencies are not well understood. Many release little information about their internal operations. It is not even possible to say conclusively how many exist, since experts disagree about how to define them, and states do not systematically track them.
One thing we do know about public authorities is that, over the course of the 20th century, these institutions have become a major component of American governance. Immediately following the Second World War, they played a minor role in public finance. But by the early 21st century, borrowing by authorities constituted well over half of all public borrowing at the sub-federal level. This change means that increasingly the leaders of these entities, rather than elected officials, make key decisions about where and how to build public infrastructure and steer economic development in the United States
D. Bradford Hunt
Public housing emerged during the New Deal as a progressive effort to end the scourge of dilapidated housing in American cities. Reformers argued that the private market had failed to provide decent, safe, and affordable housing, and they convinced Congress to provide deep subsidies to local housing authorities to build and manage modern, low-cost housing projects for the working poor. Well-intentioned but ultimately misguided policy decisions encouraged large-scale developments, concentrated poverty and youth, and starved public housing of needed resources. Further, the antipathy of private interests to public competition and the visceral resistance of white Americans to racial integration saddled public housing with many enemies and few friends. While residents often formed tight communities and fought for improvements, stigmatization and neglect undermined the success of many projects; a sizable fraction became disgraceful and tangible symbols of systemic racism toward the nation’s African American poor. Federal policy had few answers and retreated in the 1960s, eventually making a neoliberal turn to embrace public-private partnerships for delivering affordable housing. Housing vouchers and tax credits effectively displaced the federal public housing program. In the 1990s, the Clinton administration encouraged the demolition and rebuilding of troubled projects using vernacular “New Urbanist” designs to house “mixed-income” populations. Policy problems, political weakness, and an ideology of homeownership in the United States meant that a robust, public-centered program of housing for use rather than profit could not be sustained.
Adam M. Sowards
For more than a century after the republic’s founding in the 1780s, American law reflected the ideal that the commons—the public domain—should be turned into private property. As Americans became concerned about resource scarcity, waste, and monopolies at the end of the 19th century, reform-minded bureaucrats and scientists convinced Congress to maintain in perpetuity some of the nation’s land as public. This shift offered a measure of protection and an alternative to private property regimes. The federal agencies that primarily manage these lands today—U.S. Forest Service (USFS), National Park Service (NPS), U.S. Fish and Wildlife Service (USFWS), and Bureau of Land Management (BLM)—have worked since their origins in the early decades of the 20th century to fulfill their diverse, competing, evolving missions. Meanwhile, the public and Congress have continually demanded new and different goals as the land itself has functioned and responded in interdependent ways. In the mid-20th century, the agencies intensified their management, hoping they could satisfy the rising—and often conflicting—demands American citizens placed on the public lands. This intensification often worsened public lands’ ecology and increased political conflict, resulting in a series of new laws in the 1960s and 1970s. Those laws strengthened the role of science and the public in influencing agency practices while providing more opportunities for litigation. Predictably, since the late 1970s, these developments have polarized public lands’ politics. The economies, but also the identities, of many Americans remain entwined with the public lands, making political standoffs—over endangered species, oil production, privatizing land, and more—common and increasingly intractable. Because the public lands are national in scope but used by local people for all manner of economic and recreational activities, they have been and remain microcosms of the federal democratic system and all its conflicted nature.
Nicholas J. Cull
Public opinion has been part of US foreign relations in two key ways. As one would expect in a democracy, the American public has shaped the foreign policy of its government. No less significantly, the United States has sought to influence foreign public opinion as a tool of its diplomacy, now known as public diplomacy. The US public has also been a target of foreign attempts at influence with varying degrees of success. While analysis across the span of US history reveals a continuity of issues and approaches, issues of public opinion gained unprecedented salience in the second decade of the 21st century. This salience was not matched by scholarship.
Joseph E. Hower
Government employees are an essential part of the early-21st-century labor movement in the United States. Teachers, firefighters, and police officers are among the most heavily unionized occupations in America, but public-sector union members also include street cleaners and nurses, janitors and librarians, zookeepers and engineers. Despite cultural stereotypes that continue to associate unions with steel or auto workers, public employees are five times more likely to be members of unions than workers in private industry. Today, nearly half of all union members work for federal, state, or local governments.
It was not always so. Despite a long, rich history of workplace and ballot box activism, government workers were marginal to the broader labor movement until the second half of the 20th century. Excluded from the legal breakthroughs that reshaped American industry in the 1930s, government workers lacked the basic organizing and bargaining rights extended to their private-sector counterparts. A complicated, and sometimes convoluted, combination of discourse and doctrine held that government employees were, as union leader Jerry Wurf later put it, a “servant to a master” rather than “a worker with a boss.” Inspired by the material success of workers in mass industry and moved by the moral clarity of the Black Freedom struggle, government workers demanded an end to their second-class status through one of the most consequential, and least recognized, social movements of late 20th century. Yet their success at improving the pay, benefits, and conditions of government work also increased the cost of government services, imposing new obligations at a time of dramatic change in the global economy. In the resulting crunch, unionized public workers came under political pressure, particularly from fiscal conservatives who charged that their bargaining rights and political power were incompatible with a new age of austerity and limits.
Jessica Ellen Sewell
From 1800 to 2000, cities grew enormously, and saw an expansion of public spaces to serve the varied needs of a diverse population living in ever more cramped and urban circumstances. While a wide range of commercial semipublic spaces became common in the late 19th century, parks and streets were the best examples of truly public spaces with full freedom of access. Changes in the design and management of streets, sidewalks, squares, parks, and plazas during this period reflect changing ideas about the purpose of public space and how it should be used.
Streets shifted from being used for a wide range of activities, including vending, playing games, and storing goods, to becoming increasingly specialized spaces of movement, designed and managed by the early twentieth century for automobile traffic. Sidewalks, which in the early nineteenth century were paid for and liberally used by adjacent businesses, were similarly specialized as spaces of pedestrian movement. However, the tradition of using streets and sidewalks as a space of public celebration and public speech remained strong throughout the period. During parades and protests, streets and sidewalks were temporarily remade as spaces of the performance of the public, and the daily activities of circulation and commerce were set aside.
In 1800, the main open public spaces in cities were public squares or commons, often used for militia training and public celebration. In the second half of the 19th century, these were augmented by large picturesque parks. Designed as an antidote to urbanity, these parks served the public as a place for leisure, redefining public space as a polite leisure amenity, rather than a place for people to congregate as a public. The addition of playgrounds, recreational spaces, and public plazas in the 20th century served both the physical and mental health of the public. In the late 20th century, responding to neoliberal ideas and urban fiscal crises, the ownership and management of public parks and plazas was increasingly privatized, further challenging public accessibility.
Puerto Rican migrants have resided in the United States since before the Spanish-Cuban-American War of 1898, when the United States took possession of the island of Puerto Rico as part of the Treaty of Paris. After the war, groups of Puerto Ricans began migrating to the United States as contract laborers, first to sugarcane plantations in Hawaii, and then to other destinations on the mainland. After the Jones Act of 1917 extended U.S. citizenship to islanders, Puerto Ricans migrated to the United States in larger numbers, establishing their largest base in New York City. Over the course of the 1920s and 1930s, a vibrant and heterogeneous colonia developed there, and Puerto Ricans participated actively both in local politics and in the increasingly contentious politics of their homeland, whose status was indeterminate until it became a commonwealth in 1952. The Puerto Rican community in New York changed dramatically after World War II, accommodating up to fifty thousand new migrants per year during the peak of the “great migration” from the island. Newcomers faced intense discrimination and marginalization in this era, defined by both a Cold War ethos and liberal social scientists’ interest in the “Puerto Rican problem.”
Puerto Rican migrant communities in the 1950s and 1960s—now rapidly expanding into the Midwest, especially Chicago, and into New Jersey, Connecticut, and Philadelphia—struggled with inadequate housing and discrimination in the job market. In local schools, Puerto Rican children often faced a lack of accommodation of their need for English language instruction. Most catastrophic for Puerto Rican communities, on the East Coast particularly, was the deindustrialization of the labor market over the course of the 1960s. By the late 1960s, in response to these conditions and spurred by the civil rights, Black Power, and other social movements, young Puerto Ricans began organizing and protesting in large numbers. Their activism combined a radical approach to community organizing with Puerto Rican nationalism and international anti-imperialism. The youth were not the only activists in this era. Parents in New York had initiated, together with their African American neighbors, a “community control” movement that spanned the late 1960s and early 1970s; and many other adult activists pushed the politics of the urban social service sector—the primary institutions in many impoverished Puerto Rican communities—further to the left.
By the mid-1970s, urban fiscal crises and the rising conservative backlash in national politics dealt another blow to many Puerto Rican communities in the United States. The Puerto Rican population as a whole was now widely considered part of a national “underclass,” and much of the political energy of Puerto Rican leaders focused on addressing the paucity of both basic material stability and social equality in their communities. Since the 1980s, however, Puerto Ricans have achieved some economic gains, and a growing college-educated middle class has managed to gain more control over the cultural representations of their communities. More recently, the political salience of Puerto Ricans as a group has begun to shift. For the better part of the 20th century, Puerto Ricans in the United States were considered numerically insignificant or politically impotent (or both); but in the last two presidential elections (2008 and 2012), their growing populations in the South, especially in Florida, have drawn attention to their demographic significance and their political sensibilities.
The Puritans were a group of people loosely defined through their shared adherence to the reformed theological tradition, largely following the work of John Calvin. Beginning in the 16th century, the Puritan movement took root in specific regional locales throughout Germany, Scotland, the Low Countries, and England. Following Queen Elizabeth’s settlement of 1559, which mandated conformity with the Church of England, the church’s authority splintered further as Protestants clashed with the episcopal polity, or church hierarchy. Religious conflict intensified from the 1580s through the end of James I’s reign, through repeated appeals to antiquity and patristics (writings from early Christian fathers) as pleas for further reform. Religious tension and persecution under the repressive regime of Archbishop Laud caused Puritans to leave England in search of new lands and communities.
When the Pilgrims and Puritans migrated to North America in 1620 and 1630, respectively, they did so with the intention of contesting the power of the crown to mandate religious uniformity. They believed in a Calvinist-based religion that espoused a separation of church and state, but that also privileged the spiritual authority of the individual to such a degree as to leave no clear signposts about how the disparate individuals practicing these faiths should form communities. Puritan congregations in New England allowed laymen as well as women new forms of spiritual self-discovery as they orally translated the evidence of grace recorded upon their souls into communal knowledge and a corporate identity that fashioned itself as a spiritual beacon to the world. Missionary encounters soon redefined Puritan faith, theology, and pious practices. Puritan identity in 17th century North America reconstituted itself through a particular confluence of interaction with foreign landscapes, native tribes, Africans, and new models of community and social interaction.
“Twenty and odd” Africans arrived in Virginia aboard a Dutch vessel in 1619 shortly after permanent colonization of the English Americas began. There has been significant academic debate about whether the enslavement of peoples of African descent in England’s early 17th-century colonies was an inevitable or “unthinking decision” and about the nature and degree of anti-black racism during the 17th century. The legal and social status of African peoples was more flexible at first in the English colonies than it later became. Some Africans managed to escape permanent enslavement and a few Africans, such as Anthony Johnson, even owned servants of their own. There was no legal basis for enslavement in the British Americas for the first several decades of settlement and slave and servant codes emerged only gradually. Labor systems operated by custom rather than through any legal mechanisms of coercion. Most workers in the Americas experienced degrees of coercion. In the earliest years of plantation production, peoples from Africa, Europe, and the Americas often toiled alongside each other in the fields. Large numbers of Native Americans were captured and forced to work on plantations in the English Americas and many whites worked in agricultural fields as indentured and convict laborers. There were a wide variety of different kinds of coerced labor beyond enslavement in the 17th century and ideas about racial difference had yet to become as determinative as they would later be. As the staple crop plantation system matured and became entrenched on the North American mainland in the late 17th and early 18th centuries and planters required a large and regular supply of slaves, African laborers became synonymous with large-scale plantation production. The permeable boundaries between slavery and freedom disappeared, dehumanizing racism became more entrenched and U.S.-based planters developed slave codes premised on racial distinctions and legal mechanisms of coercion that were modeled on Caribbean precedents.
Courtney Q. Shah
A concerted movement to promote sex education in America emerged in the early 20th century as part of a larger public health movement that also responded to the previous century’s concerns about venereal disease, prostitution, “seduction,” and “white slavery.” Sex education, therefore, offered a way to protect people (especially privileged women) from sexual activity of all kinds—consensual and coerced. A widespread introduction into public schools did not occur until after World War I. Sex education programs in schools tended to focus on training for heterosexual marriage at a time when high school attendance spiked in urban and suburban areas. Teachers often segregated male and female students.
Beyond teaching boys about male anatomy and girls about female anatomy, reformers and educators often conveyed different messages and used different materials, depending on the race of their students. Erratic desegregation efforts during the Civil Rights movement renewed a crisis in sex education programs. Parents and administrators considered sexuality education even more dangerous in the context of a racially integrated classroom. The backlash against sex education in the schools kept pace with the backlash against integration, with each often used to bolster the other. Opponents of integration and sex education, for example, often used racial language to scare parents about what kids were learning, and with whom.
In the 1980s and 1990s, the political power of the evangelical movement in the United States attracted support for “abstinence-only” curricula that relied on scare tactics and traditional assumptions about gender and sexuality. The ever-expanding acceptance (both legal and social) of lesbian, gay, bisexual, or transgender identity directly challenged the conservative turn of abstinence-until-marriage sex education programs. The politics of gender, race, class, and sexual orientation have consistently shaped and limited sex education.
Susanah Shaw Romney
On the mid-Atlantic coast between 1624 and 1664, the Dutch developed a successful and expansive colony, one that depended on particular interactions among women and men from American, European, and African backgrounds. Unlike some other colonial efforts, such as Jamestown, New Netherland had white women colonists from its inception. In contrast to Plymouth and other English settler colonies, a population of African men and women did the crucial work of establishing the colony’s initial infrastructure in its first years. What is more, a thriving cross-cultural trade between Netherlanders and Munsee, Mahican, and Mohawk residents of the region nurtured the development of the infant colony. Looking at the colony’s establishment and growth reveals that complex interactions among ethnically distinct families gave New Netherland its particular form and character. As European and African populations took root, many households engaged in the frontier trading economy, creating a web of connections reaching into multiple indigenous villages. Women and men cooperated to sustain this trade over long distances by relying on marriage and the economic unit of the household to organize production and exchange. In addition, the colonial government used these households to stake claims to the ground and to define Dutch jurisdiction, just as they recognized that residence by Indian or English households determined where Dutch power ended. Thus ethnic and gender relations shaped not only the colony’s internal hierarchies, but also its economy and its very boundaries.
This is an advance summary of a forthcoming article in the Oxford Research Encyclopedia of American History. Please check back later for the full article.
Despite its cultivated reputation as the nation’s “white spot” in the early 20th century, Southern California was in fact home to diverse and numerous communities of color, some composed of relatively new immigrants and some long predating the era of Anglo settlement and conquest. In the years following World War II, the region engaged in suburban home construction on a mass scale and became a global symbol of what Dolores Hayden called the economically democratic but racially exclusive “sitcom suburb,” from the tax-lowering mechanism of its “Lakewood plan” to the car-friendly “Googie” architecture of the San Fernando Valley. Existing suburban communities of color, such as the colonias of agricultural laborers, were engulfed by new settlements, while upwardly mobile African Americans, Latinas/Latinos, and Asian Americans sought access to the expanding suburban dream of homeownership, with varying degrees of success. The political responses to suburban diversity in metropolitan Los Angeles ranged from Anglo resistance and flight to multiracial political coalitions and the incorporation of people of color at multiple levels of local government. The ascent by a number of suburbanites of color to positions of local and regional political power from the 1960s through the 1980s sometimes exposed intra-ethnic discord and sometimes the fragility of cross-race coalition as multiple actors sought to protect property values and to pursue economic security within the competitive constraints of shrinking municipal resources, aging infrastructure, and a receding suburban fringe. As a result, political conflicts over crime, immigration, education, and inequality emerged in many Los Angeles County suburbs by the 1970s and later in the more distant corporate suburbs of Orange, Ventura, Riverside, and San Bernardino Counties. The suburbanization of poverty, the role of suburbs as immigrant gateways, and the emergence of “majority-minority” suburbs—all national trends by the late 1990s and the first decade of the 20th century—were evident far earlier in the Los Angeles metropolitan region, where diverse suburbanites negotiated social and economic crises and innovated political responses.