Mark W. Deets
Since the founding of the United States of America, coinciding with the height of the Atlantic slave trade, U.S. officials have based their relations with West Africa primarily on economic interests. Initially, these interests were established on the backs of slaves, as the Southern plantation economy quickly vaulted the United States to prominence in the Atlantic world. After the U.S. abolition of the slave trade in 1808, however, American relations with West Africa focused on the establishment of the American colony of Liberia as a place of “return” for formerly enslaved persons. Following the turn to “legitimate commerce” in the Atlantic and the U.S. Civil War, the United States largely withdrew from large-scale interaction with West Africa. Liberia remained the notable exception, where prominent Pan-African leaders like Edward Blyden, W. E. B. DuBois, and Marcus Garvey helped foster cultural and intellectual ties between West Africa and the Diaspora in the early 1900s. These ties to Liberia were deepened in the 1920s when Firestone Rubber Corporation of Akron, Ohio established a long-term lease to harvest rubber. World War II marked a significant increase in American presence and influence in West Africa. Still focused on Liberia, the war years saw the construction of infrastructure that would prove essential to Allied war efforts and to American security interests during the Cold War. After 1945, the United States competed with the Soviet Union in West Africa for influence and access to important economic and national security resources as African nations ejected colonial regimes across most of the continent. West African independence quickly demonstrated a turn from nationalism to ethnic nationalism, as civil wars engulfed several countries in the postcolonial, and particularly the post-Cold War, era. After a decade of withdrawal, American interest in West Africa revived with the need for alternative sources of petroleum and concerns about transnational terrorism following the attacks of September 11, 2001.
Catherine A. Brekus
Historically, women in colonial North America and the United States have been deeply influenced by their religious traditions. Even though world religions like Judaism, Christianity, Buddhism, Hinduism, and Islam are based on scriptural traditions that portray women as subordinate to men, women have made up the majority of most religious groups in America. While some Americans have used religious arguments to limit women’s legal, political, and economic rights, others have drawn on scripture to defend women’s dignity and equality. Women’s religious beliefs have shaped every aspect of their lives, including their choices about how to structure their time, their attitudes toward sexuality and the body, and their understanding of suffering. Unlike early American Catholic women, who saw their highest religious calling as the sisterhood, most white colonial women identified their primary religious vocation as ministering to their families. In the 19th century, however, white Protestant women become increasingly involved in reform movements like temperance, abolitionism, and women’s suffrage, and African-American, Native American, Asian-American, and Latina women used religious arguments to challenge assumptions about white racial supremacy. In the 20th century, growing numbers of women from many different religious traditions have served as religious leaders, and in some cases they have also demanded ordination. Despite these dramatic changes in religious life, however, many religiously conservative women opposed the Equal Rights Amendment during the 1970s and early 1980s, and in the first decades of the 21st century they have continued to identify feminism and religion as antithetical.