text stringlengths 0 473k |
|---|
[SOURCE: https://en.wikipedia.org/wiki/United_States#cite_note-175] | [TOKENS: 17273] |
Contents United States The United States of America (USA), also known as the United States (U.S.) or America, is a country primarily located in North America. It is a federal republic of 50 states and a federal capital district, Washington, D.C. The 48 contiguous states border Canada to the north and Mexico to the south, with the semi-exclave of Alaska in the northwest and the archipelago of Hawaii in the Pacific Ocean. The United States also asserts sovereignty over five major island territories and various uninhabited islands in Oceania and the Caribbean.[j] It is a megadiverse country, with the world's third-largest land area[c] and third-largest population, exceeding 341 million.[k] Paleo-Indians first migrated from North Asia to North America at least 15,000 years ago, and formed various civilizations. Spanish colonization established Spanish Florida in 1513, the first European colony in what is now the continental United States. British colonization followed with the 1607 settlement of Virginia, the first of the Thirteen Colonies. Enslavement of Africans was practiced in all colonies by 1770 and supplied most of the labor for the Southern Colonies' plantation economy. Clashes with the British Crown began as a civil protest over the illegality of taxation without representation in Parliament and the denial of other English rights. They evolved into the American Revolution, which led to the Declaration of Independence and a society based on universal rights. Victory in the 1775–1783 Revolutionary War brought international recognition of U.S. sovereignty and fueled westward expansion, further dispossessing native inhabitants. As more states were admitted, a North–South division over slavery led the Confederate States of America to declare secession and fight the Union in the 1861–1865 American Civil War. With the United States' victory and reunification, slavery was abolished nationally. By the late 19th century, the U.S. economy outpaced the French, German and British economies combined. As of 1900, the country had established itself as a great power, a status solidified after its involvement in World War I. Following Japan's attack on Pearl Harbor in 1941, the U.S. entered World War II. Its aftermath left the U.S. and the Soviet Union as rival superpowers, competing for ideological dominance and international influence during the Cold War. The Soviet Union's collapse in 1991 ended the Cold War, leaving the U.S. as the world's sole superpower. The U.S. federal government is a representative democracy with a president and a constitution that grants separation of powers under three branches: legislative, executive, and judicial. The United States Congress is a bicameral national legislature composed of the House of Representatives (a lower house based on population) and the Senate (an upper house based on equal representation for each state). Federalism grants substantial autonomy to the 50 states. In addition, 574 Native American tribes have sovereignty rights, and there are 326 Native American reservations. Since the 1850s, the Democratic and Republican parties have dominated American politics. American ideals and values are based on a democratic tradition inspired by the American Enlightenment movement. A developed country, the U.S. ranks high in economic competitiveness, innovation, and higher education. Accounting for over a quarter of nominal global GDP, its economy has been the world's largest since about 1890. It is the wealthiest country, with the highest disposable household income per capita among OECD members, though its wealth inequality is highly pronounced. Shaped by centuries of immigration, the culture of the U.S. is diverse and globally influential. Making up more than a third of global military spending, the country has one of the strongest armed forces and is a designated nuclear state. A member of numerous international organizations, the U.S. plays a major role in global political, cultural, economic, and military affairs. Etymology Documented use of the phrase "United States of America" dates back to January 2, 1776. On that day, Stephen Moylan, a Continental Army aide to General George Washington, wrote a letter to Joseph Reed, Washington's aide-de-camp, seeking to go "with full and ample powers from the United States of America to Spain" to seek assistance in the Revolutionary War effort. The first known public usage is an anonymous essay published in the Williamsburg newspaper The Virginia Gazette on April 6, 1776. Sometime on or after June 11, 1776, Thomas Jefferson wrote "United States of America" in a rough draft of the Declaration of Independence, which was adopted by the Second Continental Congress on July 4, 1776. The term "United States" and its initialism "U.S.", used as nouns or as adjectives in English, are common short names for the country. The initialism "USA", a noun, is also common. "United States" and "U.S." are the established terms throughout the U.S. federal government, with prescribed rules.[l] "The States" is an established colloquial shortening of the name, used particularly from abroad; "stateside" is the corresponding adjective or adverb. "America" is the feminine form of the first word of Americus Vesputius, the Latinized name of Italian explorer Amerigo Vespucci (1454–1512);[m] it was first used as a place name by the German cartographers Martin Waldseemüller and Matthias Ringmann in 1507.[n] Vespucci first proposed that the West Indies discovered by Christopher Columbus in 1492 were part of a previously unknown landmass and not among the Indies at the eastern limit of Asia. In English, the term "America" usually does not refer to topics unrelated to the United States, despite the usage of "the Americas" to describe the totality of the continents of North and South America. History The first inhabitants of North America migrated from Siberia approximately 15,000 years ago, either across the Bering land bridge or along the now-submerged Ice Age coastline. Small isolated groups of hunter-gatherers are said to have migrated alongside herds of large herbivores far into Alaska, with ice-free corridors developing along the Pacific coast and valleys of North America in c. 16,500 – c. 13,500 BCE (c. 18,500 – c. 15,500 BP). The Clovis culture, which appeared around 11,000 BCE, is believed to be the first widespread culture in the Americas. Over time, Indigenous North American cultures grew increasingly sophisticated, and some, such as the Mississippian culture, developed agriculture, architecture, and complex societies. In the post-archaic period, the Mississippian cultures were located in the midwestern, eastern, and southern regions, and the Algonquian in the Great Lakes region and along the Eastern Seaboard, while the Hohokam culture and Ancestral Puebloans inhabited the Southwest. Native population estimates of what is now the United States before the arrival of European colonizers range from around 500,000 to nearly 10 million. Christopher Columbus began exploring the Caribbean for Spain in 1492, leading to Spanish-speaking settlements and missions from what are now Puerto Rico and Florida to New Mexico and California. The first Spanish colony in the present-day continental United States was Spanish Florida, chartered in 1513. After several settlements failed there due to starvation and disease, Spain's first permanent town, Saint Augustine, was founded in 1565. France established its own settlements in French Florida in 1562, but they were either abandoned (Charlesfort, 1578) or destroyed by Spanish raids (Fort Caroline, 1565). Permanent French settlements were founded much later along the Great Lakes (Fort Detroit, 1701), the Mississippi River (Saint Louis, 1764) and especially the Gulf of Mexico (New Orleans, 1718). Early European colonies also included the thriving Dutch colony of New Nederland (settled 1626, present-day New York) and the small Swedish colony of New Sweden (settled 1638 in what became Delaware). British colonization of the East Coast began with the Virginia Colony (1607) and the Plymouth Colony (Massachusetts, 1620). The Mayflower Compact in Massachusetts and the Fundamental Orders of Connecticut established precedents for local representative self-governance and constitutionalism that would develop throughout the American colonies. While European settlers in what is now the United States experienced conflicts with Native Americans, they also engaged in trade, exchanging European tools for food and animal pelts.[o] Relations ranged from close cooperation to warfare and massacres. The colonial authorities often pursued policies that forced Native Americans to adopt European lifestyles, including conversion to Christianity. Along the eastern seaboard, settlers trafficked Africans through the Atlantic slave trade, largely to provide manual labor on plantations. The original Thirteen Colonies[p] that would later found the United States were administered as possessions of the British Empire by Crown-appointed governors, though local governments held elections open to most white male property owners. The colonial population grew rapidly from Maine to Georgia, eclipsing Native American populations; by the 1770s, the natural increase of the population was such that only a small minority of Americans had been born overseas. The colonies' distance from Britain facilitated the entrenchment of self-governance, and the First Great Awakening, a series of Christian revivals, fueled colonial interest in guaranteed religious liberty. Following its victory in the French and Indian War, Britain began to assert greater control over local affairs in the Thirteen Colonies, resulting in growing political resistance. One of the primary grievances of the colonists was the denial of their rights as Englishmen, particularly the right to representation in the British government that taxed them. To demonstrate their dissatisfaction and resolve, the First Continental Congress met in 1774 and passed the Continental Association, a colonial boycott of British goods enforced by local "committees of safety" that proved effective. The British attempt to then disarm the colonists resulted in the 1775 Battles of Lexington and Concord, igniting the American Revolutionary War. At the Second Continental Congress, the colonies appointed George Washington commander-in-chief of the Continental Army, and created a committee that named Thomas Jefferson to draft the Declaration of Independence. Two days after the Second Continental Congress passed the Lee Resolution to create an independent, sovereign nation, the Declaration was adopted on July 4, 1776. The political values of the American Revolution evolved from an armed rebellion demanding reform within an empire to a revolution that created a new social and governing system founded on the defense of liberty and the protection of inalienable natural rights; sovereignty of the people; republicanism over monarchy, aristocracy, and other hereditary political power; civic virtue; and an intolerance of political corruption. The Founding Fathers of the United States, who included Washington, Jefferson, John Adams, Benjamin Franklin, Alexander Hamilton, John Jay, James Madison, Thomas Paine, and many others, were inspired by Classical, Renaissance, and Enlightenment philosophies and ideas. Though in practical effect since its drafting in 1777, the Articles of Confederation was ratified in 1781 and formally established a decentralized government that operated until 1789. After the British surrender at the siege of Yorktown in 1781, American sovereignty was internationally recognized by the Treaty of Paris (1783), through which the U.S. gained territory stretching west to the Mississippi River, north to present-day Canada, and south to Spanish Florida. The Northwest Ordinance (1787) established the precedent by which the country's territory would expand with the admission of new states, rather than the expansion of existing states. The U.S. Constitution was drafted at the 1787 Constitutional Convention to overcome the limitations of the Articles. It went into effect in 1789, creating a federal republic governed by three separate branches that together formed a system of checks and balances. George Washington was elected the country's first president under the Constitution, and the Bill of Rights was adopted in 1791 to allay skeptics' concerns about the power of the more centralized government. His resignation as commander-in-chief after the Revolutionary War and his later refusal to run for a third term as the country's first president established a precedent for the supremacy of civil authority in the United States and the peaceful transfer of power. In the late 18th century, American settlers began to expand westward in larger numbers, many with a sense of manifest destiny. The Louisiana Purchase of 1803 from France nearly doubled the territory of the United States. Lingering issues with Britain remained, leading to the War of 1812, which was fought to a draw. Spain ceded Florida and its Gulf Coast territory in 1819. The Missouri Compromise of 1820, which admitted Missouri as a slave state and Maine as a free state, attempted to balance the desire of northern states to prevent the expansion of slavery into new territories with that of southern states to extend it there. Primarily, the compromise prohibited slavery in all other lands of the Louisiana Purchase north of the 36°30′ parallel. As Americans expanded further into territory inhabited by Native Americans, the federal government implemented policies of Indian removal or assimilation. The most significant such legislation was the Indian Removal Act of 1830, a key policy of President Andrew Jackson. It resulted in the Trail of Tears (1830–1850), in which an estimated 60,000 Native Americans living east of the Mississippi River were forcibly removed and displaced to lands far to the west, causing 13,200 to 16,700 deaths along the forced march. Settler expansion as well as this influx of Indigenous peoples from the East resulted in the American Indian Wars west of the Mississippi. During the colonial period, slavery became legal in all the Thirteen colonies, but by 1770 it provided the main labor force in the large-scale, agriculture-dependent economies of the Southern Colonies from Maryland to Georgia. The practice began to be significantly questioned during the American Revolution, and spurred by an active abolitionist movement that had reemerged in the 1830s, states in the North enacted laws to prohibit slavery within their boundaries. At the same time, support for slavery had strengthened in Southern states, with widespread use of inventions such as the cotton gin (1793) having made slavery immensely profitable for Southern elites. The United States annexed the Republic of Texas in 1845, and the 1846 Oregon Treaty led to U.S. control of the present-day American Northwest. Dispute with Mexico over Texas led to the Mexican–American War (1846–1848). After the victory of the U.S., Mexico recognized U.S. sovereignty over Texas, New Mexico, and California in the 1848 Mexican Cession; the cession's lands also included the future states of Nevada, Colorado and Utah. The California gold rush of 1848–1849 spurred a huge migration of white settlers to the Pacific coast, leading to even more confrontations with Native populations. One of the most violent, the California genocide of thousands of Native inhabitants, lasted into the mid-1870s. Additional western territories and states were created. Throughout the 1850s, the sectional conflict regarding slavery was further inflamed by national legislation in the U.S. Congress and decisions of the Supreme Court. In Congress, the Fugitive Slave Act of 1850 mandated the forcible return to their owners in the South of slaves taking refuge in non-slave states, while the Kansas–Nebraska Act of 1854 effectively gutted the anti-slavery requirements of the Missouri Compromise. In its Dred Scott decision of 1857, the Supreme Court ruled against a slave brought into non-slave territory, simultaneously declaring the entire Missouri Compromise to be unconstitutional. These and other events exacerbated tensions between North and South that would culminate in the American Civil War (1861–1865). Beginning with South Carolina, 11 slave-state governments voted to secede from the United States in 1861, joining to create the Confederate States of America. All other state governments remained loyal to the Union.[q] War broke out in April 1861 after the Confederacy bombarded Fort Sumter. Following the Emancipation Proclamation on January 1, 1863, many freed slaves joined the Union army. The war began to turn in the Union's favor following the 1863 Siege of Vicksburg and Battle of Gettysburg, and the Confederates surrendered in 1865 after the Union's victory in the Battle of Appomattox Court House. Efforts toward reconstruction in the secessionist South had begun as early as 1862, but it was only after President Lincoln's assassination that the three Reconstruction Amendments to the Constitution were ratified to protect civil rights. The amendments codified nationally the abolition of slavery and involuntary servitude except as punishment for crimes, promised equal protection under the law for all persons, and prohibited discrimination on the basis of race or previous enslavement. As a result, African Americans took an active political role in ex-Confederate states in the decade following the Civil War. The former Confederate states were readmitted to the Union, beginning with Tennessee in 1866 and ending with Georgia in 1870. National infrastructure, including transcontinental telegraph and railroads, spurred growth in the American frontier. This was accelerated by the Homestead Acts, through which nearly 10 percent of the total land area of the United States was given away free to some 1.6 million homesteaders. From 1865 through 1917, an unprecedented stream of immigrants arrived in the United States, including 24.4 million from Europe. Most came through the Port of New York, as New York City and other large cities on the East Coast became home to large Jewish, Irish, and Italian populations. Many Northern Europeans as well as significant numbers of Germans and other Central Europeans moved to the Midwest. At the same time, about one million French Canadians migrated from Quebec to New England. During the Great Migration, millions of African Americans left the rural South for urban areas in the North. Alaska was purchased from Russia in 1867. The Compromise of 1877 is generally considered the end of the Reconstruction era, as it resolved the electoral crisis following the 1876 presidential election and led President Rutherford B. Hayes to reduce the role of federal troops in the South. Immediately, the Redeemers began evicting the Carpetbaggers and quickly regained local control of Southern politics in the name of white supremacy. African Americans endured a period of heightened, overt racism following Reconstruction, a time often considered the nadir of American race relations. A series of Supreme Court decisions, including Plessy v. Ferguson, emptied the Fourteenth and Fifteenth Amendments of their force, allowing Jim Crow laws in the South to remain unchecked, sundown towns in the Midwest, and segregation in communities across the country, which would be reinforced in part by the policy of redlining later adopted by the federal Home Owners' Loan Corporation. An explosion of technological advancement, accompanied by the exploitation of cheap immigrant labor, led to rapid economic expansion during the Gilded Age of the late 19th century. It continued into the early 20th, when the United States already outpaced the economies of Britain, France, and Germany combined. This fostered the amassing of power by a few prominent industrialists, largely by their formation of trusts and monopolies to prevent competition. Tycoons led the nation's expansion in the railroad, petroleum, and steel industries. The United States emerged as a pioneer of the automotive industry. These changes resulted in significant increases in economic inequality, slum conditions, and social unrest, creating the environment for labor unions and socialist movements to begin to flourish. This period eventually ended with the advent of the Progressive Era, which was characterized by significant economic and social reforms. Pro-American elements in Hawaii overthrew the Hawaiian monarchy; the islands were annexed in 1898. That same year, Puerto Rico, the Philippines, and Guam were ceded to the U.S. by Spain after the latter's defeat in the Spanish–American War. (The Philippines was granted full independence from the U.S. on July 4, 1946, following World War II. Puerto Rico and Guam have remained U.S. territories.) American Samoa was acquired by the United States in 1900 after the Second Samoan Civil War. The U.S. Virgin Islands were purchased from Denmark in 1917. The United States entered World War I alongside the Allies in 1917 helping to turn the tide against the Central Powers. In 1920, a constitutional amendment granted nationwide women's suffrage. During the 1920s and 1930s, radio for mass communication and early television transformed communications nationwide. The Wall Street Crash of 1929 triggered the Great Depression, to which President Franklin D. Roosevelt responded with the New Deal plan of "reform, recovery and relief", a series of unprecedented and sweeping recovery programs and employment relief projects combined with financial reforms and regulations. Initially neutral during World War II, the U.S. began supplying war materiel to the Allies of World War II in March 1941 and entered the war in December after Japan's attack on Pearl Harbor. Agreeing to a "Europe first" policy, the U.S. concentrated its wartime efforts on Japan's allies Italy and Germany until their final defeat in May 1945. The U.S. developed the first nuclear weapons and used them against the Japanese cities of Hiroshima and Nagasaki in August 1945, ending the war. The United States was one of the "Four Policemen" who met to plan the post-war world, alongside the United Kingdom, the Soviet Union, and China. The U.S. emerged relatively unscathed from the war, with even greater economic power and international political influence. The end of World War II in 1945 left the U.S. and the Soviet Union as superpowers, each with its own political, military, and economic sphere of influence. Geopolitical tensions between the two superpowers soon led to the Cold War. The U.S. implemented a policy of containment intended to limit the Soviet Union's sphere of influence; engaged in regime change against governments perceived to be aligned with the Soviets; and prevailed in the Space Race, which culminated with the first crewed Moon landing in 1969. Domestically, the U.S. experienced economic growth, urbanization, and population growth following World War II. The civil rights movement emerged, with Martin Luther King Jr. becoming a prominent leader in the early 1960s. The Great Society plan of President Lyndon B. Johnson's administration resulted in groundbreaking and broad-reaching laws, policies and a constitutional amendment to counteract some of the worst effects of lingering institutional racism. The counterculture movement in the U.S. brought significant social changes, including the liberalization of attitudes toward recreational drug use and sexuality. It also encouraged open defiance of the military draft (leading to the end of conscription in 1973) and wide opposition to U.S. intervention in Vietnam, with the U.S. totally withdrawing in 1975. A societal shift in the roles of women was significantly responsible for the large increase in female paid labor participation starting in the 1970s, and by 1985 the majority of American women aged 16 and older were employed. The Fall of Communism and the dissolution of the Soviet Union from 1989 to 1991 marked the end of the Cold War and left the United States as the world's sole superpower. This cemented the United States' global influence, reinforcing the concept of the "American Century" as the U.S. dominated international political, cultural, economic, and military affairs. The 1990s saw the longest recorded economic expansion in American history, a dramatic decline in U.S. crime rates, and advances in technology. Throughout this decade, technological innovations such as the World Wide Web, the evolution of the Pentium microprocessor in accordance with Moore's law, rechargeable lithium-ion batteries, the first gene therapy trial, and cloning either emerged in the U.S. or were improved upon there. The Human Genome Project was formally launched in 1990, while Nasdaq became the first stock market in the United States to trade online in 1998. In the Gulf War of 1991, an American-led international coalition of states expelled an Iraqi invasion force that had occupied neighboring Kuwait. The September 11 attacks on the United States in 2001 by the pan-Islamist militant organization al-Qaeda led to the war on terror and subsequent military interventions in Afghanistan and in Iraq. The U.S. housing bubble culminated in 2007 with the Great Recession, the largest economic contraction since the Great Depression. In the 2010s and early 2020s, the United States has experienced increased political polarization and democratic backsliding. The country's polarization was violently reflected in the January 2021 Capitol attack, when a mob of insurrectionists entered the U.S. Capitol and sought to prevent the peaceful transfer of power in an attempted self-coup d'état. Geography The United States is the world's third-largest country by total area behind Russia and Canada.[c] The 48 contiguous states and the District of Columbia have a combined area of 3,119,885 square miles (8,080,470 km2). In 2021, the United States had 8% of the Earth's permanent meadows and pastures and 10% of its cropland. Starting in the east, the coastal plain of the Atlantic seaboard gives way to inland forests and rolling hills in the Piedmont plateau region. The Appalachian Mountains and the Adirondack Massif separate the East Coast from the Great Lakes and the grasslands of the Midwest. The Mississippi River System, the world's fourth-longest river system, runs predominantly north–south through the center of the country. The flat and fertile prairie of the Great Plains stretches to the west, interrupted by a highland region in the southeast. The Rocky Mountains, west of the Great Plains, extend north to south across the country, peaking at over 14,000 feet (4,300 m) in Colorado. The supervolcano underlying Yellowstone National Park in the Rocky Mountains, the Yellowstone Caldera, is the continent's largest volcanic feature. Farther west are the rocky Great Basin and the Chihuahuan, Sonoran, and Mojave deserts. In the northwest corner of Arizona, carved by the Colorado River, is the Grand Canyon, a steep-sided canyon and popular tourist destination known for its overwhelming visual size and intricate, colorful landscape. The Cascade and Sierra Nevada mountain ranges run close to the Pacific coast. The lowest and highest points in the contiguous United States are in the State of California, about 84 miles (135 km) apart. At an elevation of 20,310 feet (6,190.5 m), Alaska's Denali (also called Mount McKinley) is the highest peak in the country and on the continent. Active volcanoes in the U.S. are common throughout Alaska's Alexander and Aleutian Islands. Located entirely outside North America, the archipelago of Hawaii consists of volcanic islands, physiographically and ethnologically part of the Polynesian subregion of Oceania. In addition to its total land area, the United States has one of the world's largest marine exclusive economic zones spanning approximately 4.5 million square miles (11.7 million km2) of ocean. With its large size and geographic variety, the United States includes most climate types. East of the 100th meridian, the climate ranges from humid continental in the north to humid subtropical in the south. The western Great Plains are semi-arid. Many mountainous areas of the American West have an alpine climate. The climate is arid in the Southwest, Mediterranean in coastal California, and oceanic in coastal Oregon, Washington, and southern Alaska. Most of Alaska is subarctic or polar. Hawaii, the southern tip of Florida and U.S. territories in the Caribbean and Pacific are tropical. The United States receives more high-impact extreme weather incidents than any other country. States bordering the Gulf of Mexico are prone to hurricanes, and most of the world's tornadoes occur in the country, mainly in Tornado Alley. Due to climate change in the country, extreme weather has become more frequent in the U.S. in the 21st century, with three times the number of reported heat waves compared to the 1960s. Since the 1990s, droughts in the American Southwest have become more persistent and more severe. The regions considered as the most attractive to the population are the most vulnerable. The U.S. is one of 17 megadiverse countries containing large numbers of endemic species: about 17,000 species of vascular plants occur in the contiguous United States and Alaska, and over 1,800 species of flowering plants are found in Hawaii, few of which occur on the mainland. The United States is home to 428 mammal species, 784 birds, 311 reptiles, 295 amphibians, and around 91,000 insect species. There are 63 national parks, and hundreds of other federally managed monuments, forests, and wilderness areas, administered by the National Park Service and other agencies. About 28% of the country's land is publicly owned and federally managed, primarily in the Western States. Most of this land is protected, though some is leased for commercial use, and less than one percent is used for military purposes. Environmental issues in the United States include debates on non-renewable resources and nuclear energy, air and water pollution, biodiversity, logging and deforestation, and climate change. The U.S. Environmental Protection Agency (EPA) is the federal agency charged with addressing most environmental-related issues. The idea of wilderness has shaped the management of public lands since 1964, with the Wilderness Act. The Endangered Species Act of 1973 provides a way to protect threatened and endangered species and their habitats. The United States Fish and Wildlife Service implements and enforces the Act. In 2024, the U.S. ranked 35th among 180 countries in the Environmental Performance Index. Government and politics The United States is a federal republic of 50 states and a federal capital district, Washington, D.C. The U.S. asserts sovereignty over five unincorporated territories and several uninhabited island possessions. It is the world's oldest surviving federation, and its presidential system of federal government has been adopted, in whole or in part, by many newly independent states worldwide following their decolonization. The Constitution of the United States serves as the country's supreme legal document. Most scholars describe the United States as a liberal democracy.[r] Composed of three branches, all headquartered in Washington, D.C., the federal government is the national government of the United States. The U.S. Constitution establishes a separation of powers intended to provide a system of checks and balances to prevent any of the three branches from becoming supreme. The three-branch system is known as the presidential system, in contrast to the parliamentary system where the executive is part of the legislative body. Many countries around the world adopted this aspect of the 1789 Constitution of the United States, especially in the postcolonial Americas. In the U.S. federal system, sovereign powers are shared between three levels of government specified in the Constitution: the federal government, the states, and Indian tribes. The U.S. also asserts sovereignty over five permanently inhabited territories: American Samoa, Guam, the Northern Mariana Islands, Puerto Rico, and the U.S. Virgin Islands. Residents of the 50 states are governed by their elected state government, under state constitutions compatible with the national constitution, and by elected local governments that are administrative divisions of a state. States are subdivided into counties or county equivalents, and (except for Hawaii) further divided into municipalities, each administered by elected representatives. The District of Columbia is a federal district containing the U.S. capital, Washington, D.C. The federal district is an administrative division of the federal government. Indian country is made up of 574 federally recognized tribes and 326 Indian reservations. They hold a government-to-government relationship with the U.S. federal government in Washington and are legally defined as domestic dependent nations with inherent tribal sovereignty rights. In addition to the five major territories, the U.S. also asserts sovereignty over the United States Minor Outlying Islands in the Pacific Ocean and the Caribbean. The seven undisputed islands without permanent populations are Baker Island, Howland Island, Jarvis Island, Johnston Atoll, Kingman Reef, Midway Atoll, and Palmyra Atoll. U.S. sovereignty over the unpopulated Bajo Nuevo Bank, Navassa Island, Serranilla Bank, and Wake Island is disputed. The Constitution is silent on political parties. However, they developed independently in the 18th century with the Federalist and Anti-Federalist parties. Since then, the United States has operated as a de facto two-party system, though the parties have changed over time. Since the mid-19th century, the two main national parties have been the Democratic Party and the Republican Party. The former is perceived as relatively liberal in its political platform while the latter is perceived as relatively conservative in its platform. The United States has an established structure of foreign relations, with the world's second-largest diplomatic corps as of 2024[update]. It is a permanent member of the United Nations Security Council and home to the United Nations headquarters. The United States is a member of the G7, G20, and OECD intergovernmental organizations. Almost all countries have embassies and many have consulates (official representatives) in the country. Likewise, nearly all countries host formal diplomatic missions with the United States, except Iran, North Korea, and Bhutan. Though Taiwan does not have formal diplomatic relations with the U.S., it maintains close unofficial relations. The United States regularly supplies Taiwan with military equipment to deter potential Chinese aggression. Its geopolitical attention also turned to the Indo-Pacific when the United States joined the Quadrilateral Security Dialogue with Australia, India, and Japan. The United States has a "Special Relationship" with the United Kingdom and strong ties with Canada, Australia, New Zealand, the Philippines, Japan, South Korea, Israel, and several European Union countries such as France, Italy, Germany, Spain, and Poland. The U.S. works closely with its NATO allies on military and national security issues, and with countries in the Americas through the Organization of American States and the United States–Mexico–Canada Free Trade Agreement. The U.S. exercises full international defense authority and responsibility for Micronesia, the Marshall Islands, and Palau through the Compact of Free Association. It has increasingly conducted strategic cooperation with India, while its ties with China have steadily deteriorated. Beginning in 2014, the U.S. had become a key ally of Ukraine. After Donald Trump was elected U.S. president in 2024, he sought to negotiate an end to the Russo-Ukrainian War. He paused all military aid to Ukraine in March 2025, although the aid resumed later. Trump also ended U.S. intelligence sharing with the country, but this too was eventually restored. The president is the commander-in-chief of the United States Armed Forces and appoints its leaders, the secretary of defense and the Joint Chiefs of Staff. The Department of Defense, headquartered at the Pentagon near Washington, D.C., administers five of the six service branches, which are made up of the U.S. Army, Marine Corps, Navy, Air Force, and Space Force. The Coast Guard is administered by the Department of Homeland Security in peacetime and can be transferred to the Department of the Navy in wartime. Total strength of the entire military is about 1.3 million active duty with an additional 400,000 in reserve. The United States spent $997 billion on its military in 2024, which is by far the largest amount of any country, making up 37% of global military spending and accounting for 3.4% of the country's GDP. The U.S. possesses 42% of the world's nuclear weapons—the second-largest stockpile after that of Russia. The U.S. military is widely regarded as the most powerful and advanced in the world. The United States has the third-largest combined armed forces in the world, behind the Chinese People's Liberation Army and Indian Armed Forces. The U.S. military operates about 800 bases and facilities abroad, and maintains deployments greater than 100 active duty personnel in 25 foreign countries. The United States has engaged in over 400 military interventions since its founding in 1776, with over half of these occurring between 1950 and 2019 and 25% occurring in the post-Cold War era. State defense forces (SDFs) are military units that operate under the sole authority of a state government. SDFs are authorized by state and federal law but are under the command of the state's governor. By contrast, the 54 U.S. National Guard organizations[t] fall under the dual control of state or territorial governments and the federal government; their units can also become federalized entities, but SDFs cannot be federalized. The National Guard personnel of a state or territory can be federalized by the president under the National Defense Act Amendments of 1933; this legislation created the Guard and provides for the integration of Army National Guard and Air National Guard units and personnel into the U.S. Army and (since 1947) the U.S. Air Force. The total number of National Guard members is about 430,000, while the estimated combined strength of SDFs is less than 10,000. There are about 18,000 U.S. police agencies from local to national level in the United States. Law in the United States is mainly enforced by local police departments and sheriff departments in their municipal or county jurisdictions. The state police departments have authority in their respective state, and federal agencies such as the Federal Bureau of Investigation (FBI) and the U.S. Marshals Service have national jurisdiction and specialized duties, such as protecting civil rights, national security, enforcing U.S. federal courts' rulings and federal laws, and interstate criminal activity. State courts conduct almost all civil and criminal trials, while federal courts adjudicate the much smaller number of civil and criminal cases that relate to federal law. There is no unified "criminal justice system" in the United States. The American prison system is largely heterogenous, with thousands of relatively independent systems operating across federal, state, local, and tribal levels. In 2025, "these systems hold nearly 2 million people in 1,566 state prisons, 98 federal prisons, 3,116 local jails, 1,277 juvenile correctional facilities, 133 immigration detention facilities, and 80 Indian country jails, as well as in military prisons, civil commitment centers, state psychiatric hospitals, and prisons in the U.S. territories." Despite disparate systems of confinement, four main institutions dominate: federal prisons, state prisons, local jails, and juvenile correctional facilities. Federal prisons are run by the Federal Bureau of Prisons and hold pretrial detainees as well as people who have been convicted of federal crimes. State prisons, run by the department of corrections of each state, hold people sentenced and serving prison time (usually longer than one year) for felony offenses. Local jails are county or municipal facilities that incarcerate defendants prior to trial; they also hold those serving short sentences (typically under a year). Juvenile correctional facilities are operated by local or state governments and serve as longer-term placements for any minor adjudicated as delinquent and ordered by a judge to be confined. In January 2023, the United States had the sixth-highest per capita incarceration rate in the world—531 people per 100,000 inhabitants—and the largest prison and jail population in the world, with more than 1.9 million people incarcerated. An analysis of the World Health Organization Mortality Database from 2010 showed U.S. homicide rates "were 7 times higher than in other high-income countries, driven by a gun homicide rate that was 25 times higher". Economy The U.S. has a highly developed mixed economy that has been the world's largest nominally since about 1890. Its 2024 gross domestic product (GDP)[e] of more than $29 trillion constituted over 25% of nominal global economic output, or 15% at purchasing power parity (PPP). From 1983 to 2008, U.S. real compounded annual GDP growth was 3.3%, compared to a 2.3% weighted average for the rest of the G7. The country ranks first in the world by nominal GDP, second when adjusted for purchasing power parities (PPP), and ninth by PPP-adjusted GDP per capita. In February 2024, the total U.S. federal government debt was $34.4 trillion. Of the world's 500 largest companies by revenue, 138 were headquartered in the U.S. in 2025, the highest number of any country. The U.S. dollar is the currency most used in international transactions and the world's foremost reserve currency, backed by the country's dominant economy, its military, the petrodollar system, its large U.S. treasuries market, and its linked eurodollar. Several countries use it as their official currency, and in others it is the de facto currency. The U.S. has free trade agreements with several countries, including the USMCA. Although the United States has reached a post-industrial level of economic development and is often described as having a service economy, it remains a major industrial power; in 2024, the U.S. manufacturing sector was the world's second-largest by value output after China's. New York City is the world's principal financial center, and its metropolitan area is the world's largest metropolitan economy. The New York Stock Exchange and Nasdaq, both located in New York City, are the world's two largest stock exchanges by market capitalization and trade volume. The United States is at the forefront of technological advancement and innovation in many economic fields, especially in artificial intelligence; electronics and computers; pharmaceuticals; and medical, aerospace and military equipment. The country's economy is fueled by abundant natural resources, a well-developed infrastructure, and high productivity. The largest trading partners of the United States are the European Union, Mexico, Canada, China, Japan, South Korea, the United Kingdom, Vietnam, India, and Taiwan. The United States is the world's largest importer and second-largest exporter.[u] It is by far the world's largest exporter of services. Americans have the highest average household and employee income among OECD member states, and the fourth-highest median household income in 2023, up from sixth-highest in 2013. With personal consumption expenditures of over $18.5 trillion in 2023, the U.S. has a heavily consumer-driven economy and is the world's largest consumer market. The U.S. ranked first in the number of dollar billionaires and millionaires in 2023, with 735 billionaires and nearly 22 million millionaires. Wealth in the United States is highly concentrated; in 2011, the richest 10% of the adult population owned 72% of the country's household wealth, while the bottom 50% owned just 2%. U.S. wealth inequality increased substantially since the late 1980s, and income inequality in the U.S. reached a record high in 2019. In 2024, the country had some of the highest wealth and income inequality levels among OECD countries. Since the 1970s, there has been a decoupling of U.S. wage gains from worker productivity. In 2016, the top fifth of earners took home more than half of all income, giving the U.S. one of the widest income distributions among OECD countries. There were about 771,480 homeless persons in the U.S. in 2024. In 2022, 6.4 million children experienced food insecurity. Feeding America estimates that around one in five, or approximately 13 million, children experience hunger in the U.S. and do not know where or when they will get their next meal. Also in 2022, about 37.9 million people, or 11.5% of the U.S. population, were living in poverty. The United States has a smaller welfare state and redistributes less income through government action than most other high-income countries. It is the only advanced economy that does not guarantee its workers paid vacation nationally and one of a few countries in the world without federal paid family leave as a legal right. The United States has a higher percentage of low-income workers than almost any other developed country, largely because of a weak collective bargaining system and lack of government support for at-risk workers. The United States has been a leader in technological innovation since the late 19th century and scientific research since the mid-20th century. Methods for producing interchangeable parts and the establishment of a machine tool industry enabled the large-scale manufacturing of U.S. consumer products in the late 19th century. By the early 20th century, factory electrification, the introduction of the assembly line, and other labor-saving techniques created the system of mass production. In the 21st century, the United States continues to be one of the world's foremost scientific powers, though China has emerged as a major competitor in many fields. The U.S. has the highest research and development expenditures of any country and ranks ninth as a percentage of GDP. In 2022, the United States was (after China) the country with the second-highest number of published scientific papers. In 2021, the U.S. ranked second (also after China) by the number of patent applications, and third by trademark and industrial design applications (after China and Germany), according to World Intellectual Property Indicators. In 2025 the United States ranked third (after Switzerland and Sweden) in the Global Innovation Index. The United States is considered to be a world leader in the development of artificial intelligence technology. In 2023, the United States was ranked the second most technologically advanced country in the world (after South Korea) by Global Finance magazine. The United States has maintained a space program since the late 1950s, beginning with the establishment of the National Aeronautics and Space Administration (NASA) in 1958. NASA's Apollo program (1961–1972) achieved the first crewed Moon landing with the 1969 Apollo 11 mission; it remains one of the agency's most significant milestones. Other major endeavors by NASA include the Space Shuttle program (1981–2011), the Voyager program (1972–present), the Hubble and James Webb space telescopes (launched in 1990 and 2021, respectively), and the multi-mission Mars Exploration Program (Spirit and Opportunity, Curiosity, and Perseverance). NASA is one of five agencies collaborating on the International Space Station (ISS); U.S. contributions to the ISS include several modules, including Destiny (2001), Harmony (2007), and Tranquility (2010), as well as ongoing logistical and operational support. The United States private sector dominates the global commercial spaceflight industry. Prominent American spaceflight contractors include Blue Origin, Boeing, Lockheed Martin, Northrop Grumman, and SpaceX. NASA programs such as the Commercial Crew Program, Commercial Resupply Services, Commercial Lunar Payload Services, and NextSTEP have facilitated growing private-sector involvement in American spaceflight. In 2023, the United States received approximately 84% of its energy from fossil fuel, and its largest source of energy was petroleum (38%), followed by natural gas (36%), renewable sources (9%), coal (9%), and nuclear power (9%). In 2022, the United States constituted about 4% of the world's population, but consumed around 16% of the world's energy. The U.S. ranks as the second-highest emitter of greenhouse gases behind China. The U.S. is the world's largest producer of nuclear power, generating around 30% of the world's nuclear electricity. It also has the highest number of nuclear power reactors of any country. From 2024, the U.S. plans to triple its nuclear power capacity by 2050. The United States' 4 million miles (6.4 million kilometers) of road network, owned almost entirely by state and local governments, is the longest in the world. The extensive Interstate Highway System that connects all major U.S. cities is funded mostly by the federal government but maintained by state departments of transportation. The system is further extended by state highways and some private toll roads. The U.S. is among the top ten countries with the highest vehicle ownership per capita (850 vehicles per 1,000 people) in 2022. A 2022 study found that 76% of U.S. commuters drive alone and 14% ride a bicycle, including bike owners and users of bike-sharing networks. About 11% use some form of public transportation. Public transportation in the United States is well developed in the largest urban areas, notably New York City, Washington, D.C., Boston, Philadelphia, Chicago, and San Francisco; otherwise, coverage is generally less extensive than in most other developed countries. The U.S. also has many relatively car-dependent localities. Long-distance intercity travel is provided primarily by airlines, but travel by rail is more common along the Northeast Corridor, the only high-speed rail in the U.S. that meets international standards. Amtrak, the country's government-sponsored national passenger rail company, has a relatively sparse network compared to that of Western European countries. Service is concentrated in the Northeast, California, the Midwest, the Pacific Northwest, and Virginia/Southeast. The United States has an extensive air transportation network. U.S. civilian airlines are all privately owned. The three largest airlines in the world, by total number of passengers carried, are U.S.-based; American Airlines became the global leader after its 2013 merger with US Airways. Of the 50 busiest airports in the world, 16 are in the United States, as well as five of the top 10. The world's busiest airport by passenger volume is Hartsfield–Jackson Atlanta International in Atlanta, Georgia. In 2022, most of the 19,969 U.S. airports were owned and operated by local government authorities, and there are also some private airports. Some 5,193 are designated as "public use", including for general aviation. The Transportation Security Administration (TSA) has provided security at most major airports since 2001. The country's rail transport network, the longest in the world at 182,412.3 mi (293,564.2 km), handles mostly freight (in contrast to more passenger-centered rail in Europe). Because they are often privately owned operations, U.S. railroads lag behind those of the rest of the world in terms of electrification. The country's inland waterways are the world's fifth-longest, totaling 25,482 mi (41,009 km). They are used extensively for freight, recreation, and a small amount of passenger traffic. Of the world's 50 busiest container ports, four are located in the United States, with the busiest in the country being the Port of Los Angeles. Demographics The U.S. Census Bureau reported 331,449,281 residents on April 1, 2020,[v] making the United States the third-most-populous country in the world, after India and China. The Census Bureau's official 2025 population estimate was 341,784,857, an increase of 3.1% since the 2020 census. According to the Bureau's U.S. Population Clock, on July 1, 2024, the U.S. population had a net gain of one person every 16 seconds, or about 5400 people per day. In 2023, 51% of Americans age 15 and over were married, 6% were widowed, 10% were divorced, and 34% had never been married. In 2023, the total fertility rate for the U.S. stood at 1.6 children per woman, and, at 23%, it had the world's highest rate of children living in single-parent households in 2019. Most Americans live in the suburbs of major metropolitan areas. The United States has a diverse population; 37 ancestry groups have more than one million members. White Americans with ancestry from Europe, the Middle East, or North Africa form the largest racial and ethnic group at 57.8% of the United States population. Hispanic and Latino Americans form the second-largest group and are 18.7% of the United States population. African Americans constitute the country's third-largest ancestry group and are 12.1% of the total U.S. population. Asian Americans are the country's fourth-largest group, composing 5.9% of the United States population. The country's 3.7 million Native Americans account for about 1%, and some 574 native tribes are recognized by the federal government. In 2024, the median age of the United States population was 39.1 years. While many languages and dialects are spoken in the United States, English is by far the most commonly spoken and written. De facto, English is the official language of the United States, and in 2025, Executive Order 14224 declared English official. However, the U.S. has never had a de jure official language, as Congress has never passed a law to designate English as official for all three federal branches. Some laws, such as U.S. naturalization requirements, nonetheless standardize English. Twenty-eight states and the United States Virgin Islands have laws that designate English as the sole official language; 19 states and the District of Columbia have no official language. Three states and four U.S. territories have recognized local or indigenous languages in addition to English: Hawaii (Hawaiian), Alaska (twenty Native languages),[w] South Dakota (Sioux), American Samoa (Samoan), Puerto Rico (Spanish), Guam (Chamorro), and the Northern Mariana Islands (Carolinian and Chamorro). In total, 169 Native American languages are spoken in the United States. In Puerto Rico, Spanish is more widely spoken than English. According to the American Community Survey (2020), some 245.4 million people in the U.S. age five and older spoke only English at home. About 41.2 million spoke Spanish at home, making it the second most commonly used language. Other languages spoken at home by one million people or more include Chinese (3.40 million), Tagalog (1.71 million), Vietnamese (1.52 million), Arabic (1.39 million), French (1.18 million), Korean (1.07 million), and Russian (1.04 million). German, spoken by 1 million people at home in 2010, fell to 857,000 total speakers in 2020. America's immigrant population is by far the world's largest in absolute terms. In 2022, there were 87.7 million immigrants and U.S.-born children of immigrants in the United States, accounting for nearly 27% of the overall U.S. population. In 2017, out of the U.S. foreign-born population, some 45% (20.7 million) were naturalized citizens, 27% (12.3 million) were lawful permanent residents, 6% (2.2 million) were temporary lawful residents, and 23% (10.5 million) were unauthorized immigrants. In 2019, the top countries of origin for immigrants were Mexico (24% of immigrants), India (6%), China (5%), the Philippines (4.5%), and El Salvador (3%). In fiscal year 2022, over one million immigrants (most of whom entered through family reunification) were granted legal residence. The undocumented immigrant population in the U.S. reached a record high of 14 million in 2023. The First Amendment guarantees the free exercise of religion in the country and forbids Congress from passing laws respecting its establishment. Religious practice is widespread, among the most diverse in the world, and profoundly vibrant. The country has the world's largest Christian population, which includes the fourth-largest population of Catholics. Other notable faiths include Judaism, Buddhism, Hinduism, Islam, New Age, and Native American religions. Religious practice varies significantly by region. "Ceremonial deism" is common in American culture. The overwhelming majority of Americans believe in a higher power or spiritual force, engage in spiritual practices such as prayer, and consider themselves religious or spiritual. In the Southern United States' "Bible Belt", evangelical Protestantism plays a significant role culturally; New England and the Western United States tend to be more secular. Mormonism, a Restorationist movement founded in the U.S. in 1847, is the predominant religion in Utah and a major religion in Idaho. About 82% of Americans live in metropolitan areas, particularly in suburbs; about half of those reside in cities with populations over 50,000. In 2022, 333 incorporated municipalities had populations over 100,000, nine cities had more than one million residents, and four cities—New York City, Los Angeles, Chicago, and Houston—had populations exceeding two million. Many U.S. metropolitan populations are growing rapidly, particularly in the South and West. According to the Centers for Disease Control and Prevention (CDC), average U.S. life expectancy at birth reached 79.0 years in 2024, its highest recorded level. This was an increase of 0.6 years over 2023. The CDC attributed the improvement to a significant fall in the number of fatal drug overdoses in the country, noting that "heart disease continues to be the leading cause of death in the United States, followed by cancer and unintentional injuries." In 2024, life expectancy at birth for American men rose to 76.5 years (+0.7 years compared to 2023), while life expectancy for women was 81.4 years (+0.3 years). Starting in 1998, life expectancy in the U.S. fell behind that of other wealthy industrialized countries, and Americans' "health disadvantage" gap has been increasing ever since. The Commonwealth Fund reported in 2020 that the U.S. had the highest suicide rate among high-income countries. Approximately one-third of the U.S. adult population is obese and another third is overweight. The U.S. healthcare system far outspends that of any other country, measured both in per capita spending and as a percentage of GDP, but attains worse healthcare outcomes when compared to peer countries for reasons that are debated. The United States is the only developed country without a system of universal healthcare, and a significant proportion of the population that does not carry health insurance. Government-funded healthcare coverage for the poor (Medicaid) and for those age 65 and older (Medicare) is available to Americans who meet the programs' income or age qualifications. In 2010, then-President Obama passed the Patient Protection and Affordable Care Act.[x] Abortion in the United States is not federally protected, and is illegal or restricted in 17 states. American primary and secondary education, known in the U.S. as K–12 ("kindergarten through 12th grade"), is decentralized. School systems are operated by state, territorial, and sometimes municipal governments and regulated by the U.S. Department of Education. In general, children are required to attend school or an approved homeschool from the age of five or six (kindergarten or first grade) until they are 18 years old. This often brings students through the 12th grade, the final year of a U.S. high school, but some states and territories allow them to leave school earlier, at age 16 or 17. The U.S. spends more on education per student than any other country, an average of $18,614 per year per public elementary and secondary school student in 2020–2021. Among Americans age 25 and older, 92.2% graduated from high school, 62.7% attended some college, 37.7% earned a bachelor's degree, and 14.2% earned a graduate degree. The U.S. literacy rate is near-universal. The U.S. has produced the most Nobel Prize winners of any country, with 411 (having won 413 awards). U.S. tertiary or higher education has earned a global reputation. Many of the world's top universities, as listed by various ranking organizations, are in the United States, including 19 of the top 25. American higher education is dominated by state university systems, although the country's many private universities and colleges enroll about 20% of all American students. Local community colleges generally offer open admissions, lower tuition, and coursework leading to a two-year associate degree or a non-degree certificate. As for public expenditures on higher education, the U.S. spends more per student than the OECD average, and Americans spend more than all nations in combined public and private spending. Colleges and universities directly funded by the federal government do not charge tuition and are limited to military personnel and government employees, including: the U.S. service academies, the Naval Postgraduate School, and military staff colleges. Despite some student loan forgiveness programs in place, student loan debt increased by 102% between 2010 and 2020, and exceeded $1.7 trillion in 2022. Culture and society The United States is home to a wide variety of ethnic groups, traditions, and customs. The country has been described as having the values of individualism and personal autonomy, as well as a strong work ethic and competitiveness. Voluntary altruism towards others also plays a major role; according to a 2016 study by the Charities Aid Foundation, Americans donated 1.44% of total GDP to charity—the highest rate in the world by a large margin. Americans have traditionally been characterized by a unifying political belief in an "American Creed" emphasizing consent of the governed, liberty, equality under the law, democracy, social equality, property rights, and a preference for limited government. The U.S. has acquired significant hard and soft power through its diplomatic influence, economic power, military alliances, and cultural exports such as American movies, music, video games, sports, and food. The influence that the United States exerts on other countries through soft power is referred to as Americanization. Nearly all present Americans or their ancestors came from Europe, Africa, or Asia (the "Old World") within the past five centuries. Mainstream American culture is a Western culture largely derived from the traditions of European immigrants with influences from many other sources, such as traditions brought by slaves from Africa. More recent immigration from Asia and especially Latin America has added to a cultural mix that has been described as a homogenizing melting pot, and a heterogeneous salad bowl, with immigrants contributing to, and often assimilating into, mainstream American culture. Under the First Amendment to the Constitution, the United States is considered to have the strongest protections of free speech of any country. Flag desecration, hate speech, blasphemy, and lese majesty are all forms of protected expression. A 2016 Pew Research Center poll found that Americans were the most supportive of free expression of any polity measured. Additionally, they are the "most supportive of freedom of the press and the right to use the Internet without government censorship". The U.S. is a socially progressive country with permissive attitudes surrounding human sexuality. LGBTQ rights in the United States are among the most advanced by global standards. The American Dream, or the perception that Americans enjoy high levels of social mobility, plays a key role in attracting immigrants. Whether this perception is accurate has been a topic of debate. While mainstream culture holds that the United States is a classless society, scholars identify significant differences between the country's social classes, affecting socialization, language, and values. Americans tend to greatly value socioeconomic achievement, but being ordinary or average is promoted by some as a noble condition as well. The National Foundation on the Arts and the Humanities is an agency of the United States federal government that was established in 1965 with the purpose to "develop and promote a broadly conceived national policy of support for the humanities and the arts in the United States, and for institutions which preserve the cultural heritage of the United States." It is composed of four sub-agencies: Colonial American authors were influenced by John Locke and other Enlightenment philosophers. The American Revolutionary Period (1765–1783) is notable for the political writings of Benjamin Franklin, Alexander Hamilton, Thomas Paine, and Thomas Jefferson. Shortly before and after the Revolutionary War, the newspaper rose to prominence, filling a demand for anti-British national literature. An early novel is William Hill Brown's The Power of Sympathy, published in 1791. Writer and critic John Neal in the early- to mid-19th century helped advance America toward a unique literature and culture by criticizing predecessors such as Washington Irving for imitating their British counterparts, and by influencing writers such as Edgar Allan Poe, who took American poetry and short fiction in new directions. Ralph Waldo Emerson and Margaret Fuller pioneered the influential Transcendentalism movement; Henry David Thoreau, author of Walden, was influenced by this movement. The conflict surrounding abolitionism inspired writers, like Harriet Beecher Stowe, and authors of slave narratives, such as Frederick Douglass. Nathaniel Hawthorne's The Scarlet Letter (1850) explored the dark side of American history, as did Herman Melville's Moby-Dick (1851). Major American poets of the 19th century American Renaissance include Walt Whitman, Melville, and Emily Dickinson. Mark Twain was the first major American writer to be born in the West. Henry James achieved international recognition with novels like The Portrait of a Lady (1881). As literacy rates rose, periodicals published more stories centered around industrial workers, women, and the rural poor. Naturalism, regionalism, and realism were the major literary movements of the period. While modernism generally took on an international character, modernist authors working within the United States more often rooted their work in specific regions, peoples, and cultures. Following the Great Migration to northern cities, African-American and black West Indian authors of the Harlem Renaissance developed an independent tradition of literature that rebuked a history of inequality and celebrated black culture. An important cultural export during the Jazz Age, these writings were a key influence on Négritude, a philosophy emerging in the 1930s among francophone writers of the African diaspora. In the 1950s, an ideal of homogeneity led many authors to attempt to write the Great American Novel, while the Beat Generation rejected this conformity, using styles that elevated the impact of the spoken word over mechanics to describe drug use, sexuality, and the failings of society. Contemporary literature is more pluralistic than in previous eras, with the closest thing to a unifying feature being a trend toward self-conscious experiments with language. Twelve American laureates have won the Nobel Prize in Literature. Media in the United States is broadly uncensored, with the First Amendment providing significant protections, as reiterated in New York Times Co. v. United States. The four major broadcasters in the U.S. are the National Broadcasting Company (NBC), Columbia Broadcasting System (CBS), American Broadcasting Company (ABC), and Fox Broadcasting Company (Fox). The four major broadcast television networks are all commercial entities. The U.S. cable television system offers hundreds of channels catering to a variety of niches. In 2021, about 83% of Americans over age 12 listened to broadcast radio, while about 40% listened to podcasts. In the prior year, there were 15,460 licensed full-power radio stations in the U.S. according to the Federal Communications Commission (FCC). Much of the public radio broadcasting is supplied by National Public Radio (NPR), incorporated in February 1970 under the Public Broadcasting Act of 1967. U.S. newspapers with a global reach and reputation include The Wall Street Journal, The New York Times, The Washington Post, and USA Today. About 800 publications are produced in Spanish. With few exceptions, newspapers are privately owned, either by large chains such as Gannett or McClatchy, which own dozens or even hundreds of newspapers; by small chains that own a handful of papers; or, in an increasingly rare situation, by individuals or families. Major cities often have alternative newspapers to complement the mainstream daily papers, such as The Village Voice in New York City and LA Weekly in Los Angeles. The five most-visited websites in the world are Google, YouTube, Facebook, Instagram, and ChatGPT—all of them American-owned. Other popular platforms used include X (formerly Twitter) and Amazon. In 2025, the U.S. was the world's second-largest video game market by revenue (after China). In 2015, the U.S. video game industry consisted of 2,457 companies that employed around 220,000 jobs and generated $30.4 billion in revenue. There are 444 game publishers, developers, and hardware companies in California alone. According to the Game Developers Conference (GDC), the U.S. is the top location for video game development, with 58% of the world's game developers based there in 2025. The United States is well known for its theater. Mainstream theater in the United States derives from the old European theatrical tradition and has been heavily influenced by the British theater. By the middle of the 19th century, America had created new distinct dramatic forms in the Tom Shows, the showboat theater and the minstrel show. The central hub of the American theater scene is the Theater District in Manhattan, with its divisions of Broadway, off-Broadway, and off-off-Broadway. Many movie and television celebrities have gotten their big break working in New York productions. Outside New York City, many cities have professional regional or resident theater companies that produce their own seasons. The biggest-budget theatrical productions are musicals. U.S. theater has an active community theater culture. The Tony Awards recognizes excellence in live Broadway theater and are presented at an annual ceremony in Manhattan. The awards are given for Broadway productions and performances. One is also given for regional theater. Several discretionary non-competitive awards are given as well, including a Special Tony Award, the Tony Honors for Excellence in Theatre, and the Isabelle Stevenson Award. Folk art in colonial America grew out of artisanal craftsmanship in communities that allowed commonly trained people to individually express themselves. It was distinct from Europe's tradition of high art, which was less accessible and generally less relevant to early American settlers. Cultural movements in art and craftsmanship in colonial America generally lagged behind those of Western Europe. For example, the prevailing medieval style of woodworking and primitive sculpture became integral to early American folk art, despite the emergence of Renaissance styles in England in the late 16th and early 17th centuries. The new English styles would have been early enough to make a considerable impact on American folk art, but American styles and forms had already been firmly adopted. Not only did styles change slowly in early America, but there was a tendency for rural artisans there to continue their traditional forms longer than their urban counterparts did—and far longer than those in Western Europe. The Hudson River School was a mid-19th-century movement in the visual arts tradition of European naturalism. The 1913 Armory Show in New York City, an exhibition of European modernist art, shocked the public and transformed the U.S. art scene. American Realism and American Regionalism sought to reflect and give America new ways of looking at itself. Georgia O'Keeffe, Marsden Hartley, and others experimented with new and individualistic styles, which would become known as American modernism. Major artistic movements such as the abstract expressionism of Jackson Pollock and Willem de Kooning and the pop art of Andy Warhol and Roy Lichtenstein developed largely in the United States. Major photographers include Alfred Stieglitz, Edward Steichen, Dorothea Lange, Edward Weston, James Van Der Zee, Ansel Adams, and Gordon Parks. The tide of modernism and then postmodernism has brought global fame to American architects, including Frank Lloyd Wright, Philip Johnson, and Frank Gehry. The Metropolitan Museum of Art in Manhattan is the largest art museum in the United States and the fourth-largest in the world. American folk music encompasses numerous music genres, variously known as traditional music, traditional folk music, contemporary folk music, or roots music. Many traditional songs have been sung within the same family or folk group for generations, and sometimes trace back to such origins as the British Isles, mainland Europe, or Africa. The rhythmic and lyrical styles of African-American music in particular have influenced American music. Banjos were brought to America through the slave trade. Minstrel shows incorporating the instrument into their acts led to its increased popularity and widespread production in the 19th century. The electric guitar, first invented in the 1930s, and mass-produced by the 1940s, had an enormous influence on popular music, in particular due to the development of rock and roll. The synthesizer, turntablism, and electronic music were also largely developed in the U.S. Elements from folk idioms such as the blues and old-time music were adopted and transformed into popular genres with global audiences. Jazz grew from blues and ragtime in the early 20th century, developing from the innovations and recordings of composers such as W.C. Handy and Jelly Roll Morton. Louis Armstrong and Duke Ellington increased its popularity early in the 20th century. Country music developed in the 1920s, bluegrass and rhythm and blues in the 1940s, and rock and roll in the 1950s. In the 1960s, Bob Dylan emerged from the folk revival to become one of the country's most celebrated songwriters. The musical forms of punk and hip hop both originated in the United States in the 1970s. The United States has the world's largest music market, with a total retail value of $15.9 billion in 2022. Most of the world's major record companies are based in the U.S.; they are represented by the Recording Industry Association of America (RIAA). Mid-20th-century American pop stars, such as Frank Sinatra and Elvis Presley, became global celebrities and best-selling music artists, as have artists of the late 20th century, such as Michael Jackson, Madonna, Whitney Houston, and Mariah Carey, and of the early 21st century, such as Eminem, Britney Spears, Lady Gaga, Katy Perry, Taylor Swift and Beyoncé. The United States has the world's largest apparel market by revenue. Apart from professional business attire, American fashion is eclectic and predominantly informal. Americans' diverse cultural roots are reflected in their clothing; however, sneakers, jeans, T-shirts, and baseball caps are emblematic of American styles. New York, with its Fashion Week, is considered to be one of the "Big Four" global fashion capitals, along with Paris, Milan, and London. A study demonstrated that general proximity to Manhattan's Garment District has been synonymous with American fashion since its inception in the early 20th century. A number of well-known designer labels, among them Tommy Hilfiger, Ralph Lauren, Tom Ford and Calvin Klein, are headquartered in Manhattan. Labels cater to niche markets, such as preteens. New York Fashion Week is one of the most influential fashion shows in the world, and is held twice each year in Manhattan; the annual Met Gala, also in Manhattan, has been called the fashion world's "biggest night". The U.S. film industry has a worldwide influence and following. Hollywood, a district in central Los Angeles, the nation's second-most populous city, is also metonymous for the American filmmaking industry. The major film studios of the United States are the primary source of the most commercially successful movies selling the most tickets in the world. Largely centered in the New York City region from its beginnings in the late 19th century through the first decades of the 20th century, the U.S. film industry has since been primarily based in and around Hollywood. Nonetheless, American film companies have been subject to the forces of globalization in the 21st century, and an increasing number of films are made elsewhere. The Academy Awards, popularly known as "the Oscars", have been held annually by the Academy of Motion Picture Arts and Sciences since 1929, and the Golden Globe Awards have been held annually since January 1944. The industry peaked in what is commonly referred to as the "Golden Age of Hollywood", from the early sound period until the early 1960s, with screen actors such as John Wayne and Marilyn Monroe becoming iconic figures. In the 1970s, "New Hollywood", or the "Hollywood Renaissance", was defined by grittier films influenced by French and Italian realist pictures of the post-war period. The 21st century has been marked by the rise of American streaming platforms, which came to rival traditional cinema. Early settlers were introduced by Native Americans to foods such as turkey, sweet potatoes, corn, squash, and maple syrup. Of the most enduring and pervasive examples are variations of the native dish called succotash. Early settlers and later immigrants combined these with foods they were familiar with, such as wheat flour, beef, and milk, to create a distinctive American cuisine. New World crops, especially pumpkin, corn, potatoes, and turkey as the main course are part of a shared national menu on Thanksgiving, when many Americans prepare or purchase traditional dishes to celebrate the occasion. Characteristic American dishes such as apple pie, fried chicken, doughnuts, french fries, macaroni and cheese, ice cream, hamburgers, hot dogs, and American pizza derive from the recipes of various immigrant groups. Mexican dishes such as burritos and tacos preexisted the United States in areas later annexed from Mexico, and adaptations of Chinese cuisine as well as pasta dishes freely adapted from Italian sources are all widely consumed. American chefs have had a significant impact on society both domestically and internationally. In 1946, the Culinary Institute of America was founded by Katharine Angell and Frances Roth. This would become the United States' most prestigious culinary school, where many of the most talented American chefs would study prior to successful careers. The United States restaurant industry was projected at $899 billion in sales for 2020, and employed more than 15 million people, representing 10% of the nation's workforce directly. It is the country's second-largest private employer and the third-largest employer overall. The United States is home to over 220 Michelin star-rated restaurants, 70 of which are in New York City. Wine has been produced in what is now the United States since the 1500s, with the first widespread production beginning in what is now New Mexico in 1628. In the modern U.S., wine production is undertaken in all fifty states, with California producing 84 percent of all U.S. wine. With more than 1,100,000 acres (4,500 km2) under vine, the United States is the fourth-largest wine-producing country in the world, after Italy, Spain, and France. The classic American diner, a casual restaurant type originally intended for the working class, emerged during the 19th century from converted railroad dining cars made stationary. The diner soon evolved into purpose-built structures whose number expanded greatly in the 20th century. The American fast-food industry developed alongside the nation's car culture. American restaurants developed the drive-in format in the 1920s, which they began to replace with the drive-through format by the 1940s. American fast-food restaurant chains, such as McDonald's, Burger King, Chick-fil-A, Kentucky Fried Chicken, Dunkin' Donuts and many others, have numerous outlets around the world. The most popular spectator sports in the U.S. are American football, basketball, baseball, soccer, and ice hockey. Their premier leagues are, respectively, the National Football League, the National Basketball Association, Major League Baseball, Major League Soccer, and the National Hockey League, All these leagues enjoy wide-ranging domestic media coverage and, except for the MLS, all are considered the preeminent leagues in their respective sports in the world. While most major U.S. sports such as baseball and American football have evolved out of European practices, basketball, volleyball, skateboarding, and snowboarding are American inventions, many of which have become popular worldwide. Lacrosse and surfing arose from Native American and Native Hawaiian activities that predate European contact. The market for professional sports in the United States was approximately $69 billion in July 2013, roughly 50% larger than that of Europe, the Middle East, and Africa combined. American football is by several measures the most popular spectator sport in the United States. Although American football does not have a substantial following in other nations, the NFL does have the highest average attendance (67,254) of any professional sports league in the world. In the year 2024, the NFL generated over $23 billion, making them the most valued professional sports league in the United States and the world. Baseball has been regarded as the U.S. "national sport" since the late 19th century. The most-watched individual sports in the U.S. are golf and auto racing, particularly NASCAR and IndyCar. On the collegiate level, earnings for the member institutions exceed $1 billion annually, and college football and basketball attract large audiences, as the NCAA March Madness tournament and the College Football Playoff are some of the most watched national sporting events. In the U.S., the intercollegiate sports level serves as the main feeder system for professional and Olympic sports, with significant exceptions such as Minor League Baseball. This differs greatly from practices in nearly all other countries, where publicly and privately funded sports organizations serve this function. Eight Olympic Games have taken place in the United States. The 1904 Summer Olympics in St. Louis, Missouri, were the first-ever Olympic Games held outside of Europe. The Olympic Games will be held in the U.S. for a ninth time when Los Angeles hosts the 2028 Summer Olympics. U.S. athletes have won a total of 2,968 medals (1,179 gold) at the Olympic Games, the most of any country. In other international competition, the United States is the home of a number of prestigious events, including the America's Cup, World Baseball Classic, the U.S. Open, and the Masters Tournament. The U.S. men's national soccer team has qualified for eleven World Cups, while the women's national team has won the FIFA Women's World Cup and Olympic soccer tournament four and five times, respectively. The 1999 FIFA Women's World Cup was hosted by the United States. Its final match was attended by 90,185, setting the world record for largest women's sporting event crowd at the time. The United States hosted the 1994 FIFA World Cup and will co-host, along with Canada and Mexico, the 2026 FIFA World Cup. See also Notes References This article incorporates text from a free content work. Licensed under CC BY-SA IGO 3.0 (license statement/permission). Text taken from World Food and Agriculture – Statistical Yearbook 2023, FAO, FAO. External links 40°N 100°W / 40°N 100°W / 40; -100 (United States of America) |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Jews#cite_note-17] | [TOKENS: 15852] |
Contents Jews Jews (Hebrew: יְהוּדִים, ISO 259-2: Yehudim, Israeli pronunciation: [jehuˈdim]), or the Jewish people, are an ethnoreligious group and nation, originating from the Israelites of ancient Israel and Judah. They traditionally adhere to Judaism. Jewish ethnicity, religion, and community are highly interrelated, as Judaism is an ethnic religion, though many ethnic Jews do not practice it. Religious Jews regard converts to Judaism as members of the Jewish nation, pursuant to the long-standing conversion process. The Israelites emerged from the pre-existing Canaanite peoples to establish Israel and Judah in the Southern Levant during the Iron Age. Originally, Jews referred to the inhabitants of the kingdom of Judah and were distinguished from the gentiles and the Samaritans. According to the Hebrew Bible, these inhabitants predominately originate from the tribe of Judah, who were descendants of Judah, the fourth son of Jacob. The tribe of Benjamin were another significant demographic in Judah and were considered Jews too. By the late 6th century BCE, Judaism had evolved from the Israelite religion, dubbed Yahwism (for Yahweh) by modern scholars, having a theology that religious Jews believe to be the expression of the Mosaic covenant between God and the Jewish people. After the Babylonian exile, Jews referred to followers of Judaism, descendants of the Israelites, citizens of Judea, or allies of the Judean state. Jewish migration within the Mediterranean region during the Hellenistic period, followed by population transfers, caused by events like the Jewish–Roman wars, gave rise to the Jewish diaspora, consisting of diverse Jewish communities that maintained their sense of Jewish history, identity, and culture. In the following millennia, Jewish diaspora communities coalesced into three major ethnic subdivisions according to where their ancestors settled: the Ashkenazim (Central and Eastern Europe), the Sephardim (Iberian Peninsula), and the Mizrahim (Middle East and North Africa). While these three major divisions account for most of the world's Jews, there are other smaller Jewish groups outside of the three. Prior to World War II, the global Jewish population reached a peak of 16.7 million, representing around 0.7% of the world's population at that time. During World War II, approximately six million Jews throughout Europe were systematically murdered by Nazi Germany in a genocide known as the Holocaust. Since then, the population has slowly risen again, and as of 2021[update], was estimated to be at 15.2 million by the demographer Sergio Della Pergola or less than 0.2% of the total world population in 2012.[b] Today, over 85% of Jews live in Israel or the United States. Israel, whose population is 73.9% Jewish, is the only country where Jews comprise more than 2.5% of the population. Jews have significantly influenced and contributed to the development and growth of human progress in many fields, both historically and in modern times, including in science and technology, philosophy, ethics, literature, governance, business, art, music, comedy, theatre, cinema, architecture, food, medicine, and religion. Jews founded Christianity and had an indirect but profound influence on Islam. In these ways and others, Jews have played a significant role in the development of Western culture. Name and etymology The term "Jew" is derived from the Hebrew word יְהוּדִי Yehudi, with the plural יְהוּדִים Yehudim. Endonyms in other Jewish languages include the Ladino ג׳ודיו Djudio (plural ג׳ודיוס, Djudios) and the Yiddish ייִד Yid (plural ייִדן Yidn). Though Genesis 29:35 and 49:8 connect "Judah" with the verb yada, meaning "praise", scholars generally agree that "Judah" most likely derives from the name of a Levantine geographic region dominated by gorges and ravines. The gradual ethnonymic shift from "Israelites" to "Jews", regardless of their descent from Judah, although not contained in the Torah, is made explicit in the Book of Esther (4th century BCE) of the Tanakh. Some modern scholars disagree with the conflation, based on the works of Josephus, Philo and Apostle Paul. The English word "Jew" is a derivation of Middle English Gyw, Iewe. The latter was loaned from the Old French giu, which itself evolved from the earlier juieu, which in turn derived from judieu/iudieu which through elision had dropped the letter "d" from the Medieval Latin Iudaeus, which, like the New Testament Greek term Ioudaios, meant both "Jew" and "Judean" / "of Judea". The Greek term was a loan from Aramaic *yahūdāy, corresponding to Hebrew יְהוּדִי Yehudi. Some scholars prefer translating Ioudaios as "Judean" in the Bible since it is more precise, denotes the community's origins and prevents readers from engaging in antisemitic eisegesis. Others disagree, believing that it erases the Jewish identity of Biblical characters such as Jesus. Daniel R. Schwartz distinguishes "Judean" and "Jew". Here, "Judean" refers to the inhabitants of Judea, which encompassed southern Palestine. Meanwhile, "Jew" refers to the descendants of Israelites that adhere to Judaism. Converts are included in the definition. But Shaye J.D. Cohen argues that "Judean" is inclusive of believers of the Judean God and allies of the Judean state. Another scholar, Jodi Magness, wrote the term Ioudaioi refers to a "people of Judahite/Judean ancestry who worshipped the God of Israel as their national deity and (at least nominally) lived according to his laws." The etymological equivalent is in use in other languages, e.g., يَهُودِيّ yahūdī (sg.), al-yahūd (pl.), in Arabic, "Jude" in German, "judeu" in Portuguese, "Juif" (m.)/"Juive" (f.) in French, "jøde" in Danish and Norwegian, "judío/a" in Spanish, "jood" in Dutch, "żyd" in Polish etc., but derivations of the word "Hebrew" are also in use to describe a Jew, e.g., in Italian (Ebreo), in Persian ("Ebri/Ebrani" (Persian: عبری/عبرانی)) and Russian (Еврей, Yevrey). The German word "Jude" is pronounced [ˈjuːdə], the corresponding adjective "jüdisch" [ˈjyːdɪʃ] (Jewish) is the origin of the word "Yiddish". According to The American Heritage Dictionary of the English Language, fourth edition (2000), It is widely recognized that the attributive use of the noun Jew, in phrases such as Jew lawyer or Jew ethics, is both vulgar and highly offensive. In such contexts Jewish is the only acceptable possibility. Some people, however, have become so wary of this construction that they have extended the stigma to any use of Jew as a noun, a practice that carries risks of its own. In a sentence such as There are now several Jews on the council, which is unobjectionable, the substitution of a circumlocution like Jewish people or persons of Jewish background may in itself cause offense for seeming to imply that Jew has a negative connotation when used as a noun. Identity Judaism shares some of the characteristics of a nation, an ethnicity, a religion, and a culture, making the definition of who is a Jew vary slightly depending on whether a religious or national approach to identity is used.[better source needed] Generally, in modern secular usage, Jews include three groups: people who were born to a Jewish family regardless of whether or not they follow the religion, those who have some Jewish ancestral background or lineage (sometimes including those who do not have strictly matrilineal descent), and people without any Jewish ancestral background or lineage who have formally converted to Judaism and therefore are followers of the religion. In the context of biblical and classical literature, Jews could refer to inhabitants of the Kingdom of Judah, or the broader Judean region, allies of the Judean state, or anyone that followed Judaism. Historical definitions of Jewish identity have traditionally been based on halakhic definitions of matrilineal descent, and halakhic conversions. These definitions of who is a Jew date back to the codification of the Oral Torah into the Babylonian Talmud, around 200 CE. Interpretations by Jewish sages of sections of the Tanakh – such as Deuteronomy 7:1–5, which forbade intermarriage between their Israelite ancestors and seven non-Israelite nations: "for that [i.e. giving your daughters to their sons or taking their daughters for your sons,] would turn away your children from following me, to serve other gods"[failed verification] – are used as a warning against intermarriage between Jews and gentiles. Leviticus 24:10 says that the son in a marriage between a Hebrew woman and an Egyptian man is "of the community of Israel." This is complemented by Ezra 10:2–3, where Israelites returning from Babylon vow to put aside their gentile wives and their children. A popular theory is that the rape of Jewish women in captivity brought about the law of Jewish identity being inherited through the maternal line, although scholars challenge this theory citing the Talmudic establishment of the law from the pre-exile period. Another argument is that the rabbis changed the law of patrilineal descent to matrilineal descent due to the widespread rape of Jewish women by Roman soldiers. Since the anti-religious Haskalah movement of the late 18th and 19th centuries, halakhic interpretations of Jewish identity have been challenged. According to historian Shaye J. D. Cohen, the status of the offspring of mixed marriages was determined patrilineally in the Bible. He brings two likely explanations for the change in Mishnaic times: first, the Mishnah may have been applying the same logic to mixed marriages as it had applied to other mixtures (Kil'ayim). Thus, a mixed marriage is forbidden as is the union of a horse and a donkey, and in both unions the offspring are judged matrilineally. Second, the Tannaim may have been influenced by Roman law, which dictated that when a parent could not contract a legal marriage, offspring would follow the mother. Rabbi Rivon Krygier follows a similar reasoning, arguing that Jewish descent had formerly passed through the patrilineal descent and the law of matrilineal descent had its roots in the Roman legal system. Origins The prehistory and ethnogenesis of the Jews are closely intertwined with archaeology, biology, historical textual records, mythology, and religious literature. The ethnic origin of the Jews lie in the Israelites, a confederation of Iron Age Semitic-speaking tribes that inhabited a part of Canaan during the tribal and monarchic periods. Modern Jews are named after and also descended from the southern Israelite Kingdom of Judah. Gary A. Rendsburg links the early Canaanite nomadic pastoralists confederation to the Shasu known to the Egyptians around the 15th century BCE. According to the Hebrew Bible narrative, Jewish history begins with the Biblical patriarchs such as Abraham, his son Isaac, Isaac's son Jacob, and the Biblical matriarchs Sarah, Rebecca, Leah, and Rachel, who lived in Canaan. The twelve sons of Jacob subsequently gave birth to the Twelve Tribes. Jacob and his family migrated to Ancient Egypt after being invited to live with Jacob's son Joseph by the Pharaoh himself. Jacob's descendants were later enslaved until the Exodus, led by Moses. Afterwards, the Israelites conquered Canaan under Moses' successor Joshua, and went through the period of the Biblical judges after the death of Joshua. Through the mediation of Samuel, the Israelites were subject to a king, Saul, who was succeeded by David and then Solomon, after whom the United Monarchy ended and was split into a separate Kingdom of Israel and a Kingdom of Judah. The Kingdom of Judah is described as comprising the tribes of Judah, Benjamin and partially, Levi. They later assimilated remnants of other tribes who migrated there from the northern Kingdom of Israel. In the extra-biblical record, the Israelites become visible as a people between 1200 and 1000 BCE. There is well accepted archeological evidence referring to "Israel" in the Merneptah Stele, which dates to about 1200 BCE, and in the Mesha stele from 840 BCE. It is debated whether a period like that of the Biblical judges occurred and if there ever was a United Monarchy. There is further disagreement about the earliest existence of the Kingdoms of Israel and Judah and their extent and power. Historians agree that a Kingdom of Israel existed by c. 900 BCE,: 169–95 there is a consensus that a Kingdom of Judah existed by c. 700 BCE at least, and recent excavations in Khirbet Qeiyafa have provided strong evidence for dating the Kingdom of Judah to the 10th century BCE. In 587 BCE, Nebuchadnezzar II, King of the Neo-Babylonian Empire, besieged Jerusalem, destroyed the First Temple and deported parts of the Judahite population. Scholars disagree regarding the extent to which the Bible should be accepted as a historical source for early Israelite history. Rendsburg states that there are two approximately equal groups of scholars who debate the historicity of the biblical narrative, the minimalists who largely reject it, and the maximalists who largely accept it, with the minimalists being the more vocal of the two. Some of the leading minimalists reframe the biblical account as constituting the Israelites' inspiring national myth narrative, suggesting that according to the modern archaeological and historical account, the Israelites and their culture did not overtake the region by force, but instead branched out of the Canaanite peoples and culture through the development of a distinct monolatristic—and later monotheistic—religion of Yahwism centered on Yahweh, one of the gods of the Canaanite pantheon. The growth of Yahweh-centric belief, along with a number of cultic practices, gradually gave rise to a distinct Israelite ethnic group, setting them apart from other Canaanites. According to Dever, modern archaeologists have largely discarded the search for evidence of the biblical narrative surrounding the patriarchs and the exodus. According to the maximalist position, the modern archaeological record independently points to a narrative which largely agrees with the biblical account. This narrative provides a testimony of the Israelites as a nomadic people known to the Egyptians as belonging to the Shasu. Over time these nomads left the desert and settled on the central mountain range of the land of Canaan, in simple semi-nomadic settlements in which pig bones are notably absent. This population gradually shifted from a tribal lifestyle to a monarchy. While the archaeological record of the ninth century BCE provides evidence for two monarchies, one in the south under a dynasty founded by a figure named David with its capital in Jerusalem, and one in the north under a dynasty founded by a figure named Omri with its capital in Samaria. It also points to an early monarchic period in which these regions shared material culture and religion, suggesting a common origin. Archaeological finds also provide evidence for the later cooperation of these two kingdoms in their coalition against Aram, and for their destructions by the Assyrians and later by the Babylonians. Genetic studies on Jews show that most Jews worldwide bear a common genetic heritage which originates in the Middle East, and that they share certain genetic traits with other Gentile peoples of the Fertile Crescent. The genetic composition of different Jewish groups shows that Jews share a common gene pool dating back four millennia, as a marker of their common ancestral origin. Despite their long-term separation, Jewish communities maintained their unique commonalities, propensities, and sensibilities in culture, tradition, and language. History The earliest recorded evidence of a people by the name of Israel appears in the Merneptah Stele, which dates to around 1200 BCE. The majority of scholars agree that this text refers to the Israelites, a group that inhabited the central highlands of Canaan, where archaeological evidence shows that hundreds of small settlements were constructed between the 12th and 10th centuries BCE. The Israelites differentiated themselves from neighboring peoples through various distinct characteristics including religious practices, prohibition on intermarriage, and an emphasis on genealogy and family history. In the 10th century BCE, two neighboring Israelite kingdoms—the northern Kingdom of Israel and the southern Kingdom of Judah—emerged. Since their inception, they shared ethnic, cultural, linguistic and religious characteristics despite a complicated relationship. Israel, with its capital mostly in Samaria, was larger and wealthier, and soon developed into a regional power. In contrast, Judah, with its capital in Jerusalem, was less prosperous and covered a smaller, mostly mountainous territory. However, while in Israel the royal succession was often decided by a military coup d'état, resulting in several dynasty changes, political stability in Judah was much greater, as it was ruled by the House of David for the whole four centuries of its existence. Scholars also describe Biblical Jews as a 'proto-nation', in the modern nationalist sense, comparable to classical Greeks, the Gauls and the British Celts. Around 720 BCE, Kingdom of Israel was destroyed when it was conquered by the Neo-Assyrian Empire, which came to dominate the ancient Near East. Under the Assyrian resettlement policy, a significant portion of the northern Israelite population was exiled to Mesopotamia and replaced by immigrants from the same region. During the same period, and throughout the 7th century BCE, the Kingdom of Judah, now under Assyrian vassalage, experienced a period of prosperity and witnessed a significant population growth. This prosperity continued until the Neo-Assyrian king Sennacherib devastated the region of Judah in response to a rebellion in the area, ultimately halting at Jerusalem. Later in the same century, the Assyrians were defeated by the rising Neo-Babylonian Empire, and Judah became its vassal. In 587 BCE, following a revolt in Judah, the Babylonian king Nebuchadnezzar II besieged and destroyed Jerusalem and the First Temple, putting an end to the kingdom. The majority of Jerusalem's residents, including the kingdom's elite, were exiled to Babylon. According to the Book of Ezra, the Persian Cyrus the Great ended the Babylonian exile in 538 BCE, the year after he captured Babylon. The exile ended with the return under Zerubbabel the Prince (so called because he was a descendant of the royal line of David) and Joshua the Priest (a descendant of the line of the former High Priests of the Temple) and their construction of the Second Temple circa 521–516 BCE. As part of the Persian Empire, the former Kingdom of Judah became the province of Judah (Yehud Medinata), with a smaller territory and a reduced population. Judea was under control of the Achaemenids until the fall of their empire in c. 333 BCE to Alexander the Great. After several centuries under foreign imperial rule, the Maccabean Revolt against the Seleucid Empire resulted in an independent Hasmonean kingdom, under which the Jews once again enjoyed political independence for a period spanning from 110 to 63 BCE. Under Hasmonean rule the boundaries of their kingdom were expanded to include not only the land of the historical kingdom of Judah, but also the Galilee and Transjordan. In the beginning of this process the Idumeans, who had infiltrated southern Judea after the destruction of the First Temple, were converted en masse. In 63 BCE, Judea was conquered by the Romans. From 37 BCE to 6 CE, the Romans allowed the Jews to maintain some degree of independence by installing the Herodian dynasty as vassal kings. However, Judea eventually came directly under Roman control and was incorporated into the Roman Empire as the province of Judaea. The Jewish–Roman wars, a series of failed uprisings against Roman rule during the first and second centuries CE, had profound and devastating consequences for the Jewish population of Judaea. The First Jewish–Roman War (66–73/74 CE) culminated in the destruction of Jerusalem and the Second Temple, after which the significantly diminished Jewish population was stripped of political autonomy. A few generations later, the Bar Kokhba revolt (132–136 CE) erupted in response to Roman plans to rebuild Jerusalem as a Roman colony, and, possibly, to restrictions on circumcision. Its violent suppression by the Romans led to the near-total depopulation of Judea, and the demographic and cultural center of Jewish life shifted to Galilee. Jews were subsequently banned from residing in Jerusalem and the surrounding area, and the province of Judaea was renamed Syria Palaestina. These developments effectively ended Jewish efforts to restore political sovereignty in the region for nearly two millennia. Similar upheavals impacted the Jewish communities in the empire's eastern provinces during the Diaspora Revolt (115–117 CE), leading to the near-total destruction of Jewish diaspora communities in Libya, Cyprus and Egypt, including the highly influential community in Alexandria. The destruction of the Second Temple in 70 CE brought profound changes to Judaism. With the Temple's central place in Jewish worship gone, religious practices shifted towards prayer, Torah study (including Oral Torah), and communal gatherings in synagogues. Judaism also lost much of its sectarian nature.: 69 Two of the three main sects that flourished during the late Second Temple period, namely the Sadducees and Essenes, eventually disappeared, while Pharisaic beliefs became the foundational, liturgical, and ritualistic basis of Rabbinic Judaism, which emerged as the prevailing form of Judaism since late antiquity. The Jewish diaspora existed well before the destruction of the Second Temple in 70 CE and had been ongoing for centuries, with the dispersal driven by both forced expulsions and voluntary migrations. In Mesopotamia, a testimony to the beginnings of the Jewish community can be found in Joachin's ration tablets, listing provisions allotted to the exiled Judean king and his family by Nebuchadnezzar II, and further evidence are the Al-Yahudu tablets, dated to the 6th–5th centuries BCE and related to the exiles from Judea arriving after the destruction of the First Temple, though there is ample evidence for the presence of Jews in Babylonia even from 626 BCE. In Egypt, the documents from Elephantine reveal the trials of a community founded by a Persian Jewish garrison at two fortresses on the frontier during the 5th–4th centuries BCE, and according to Josephus the Jewish community in Alexandria existed since the founding of the city in the 4th century BCE by Alexander the Great. By 200 BCE, there were well established Jewish communities both in Egypt and Mesopotamia ("Babylonia" in Jewish sources) and in the two centuries that followed, Jewish populations were also present in Asia Minor, Greece, Macedonia, Cyrene, and, beginning in the middle of the first century BCE, in the city of Rome. Later, in the first centuries CE, as a result of the Jewish-Roman Wars, a large number of Jews were taken as captives, sold into slavery, or compelled to flee from the regions affected by the wars, contributing to the formation and expansion of Jewish communities across the Roman Empire as well as in Arabia and Mesopotamia. After the Bar Kokhba revolt, the Jewish population in Judaea—now significantly reduced— made efforts to recover from the revolt's devastating effects, but never fully regained its former strength. Between the second and fourth centuries CE, the region of Galilee emerged as the primary center of Jewish life in Syria Palaestina, experiencing both demographic growth and cultural development. It was during this period that two central rabbinic texts, the Mishnah and the Jerusalem Talmud, were composed. The Romans recognized the patriarchs—rabbinic sages such as Judah ha-Nasi—as representatives of the Jewish people, granting them a certain degree of autonomy. However, as the Roman Empire gave way to the Christianized Byzantine Empire under Constantine, Jews began to face persecution by both the Church and imperial authorities, Jews came to be persecuted by the church and the authorities, and many immigrated to communities in the diaspora. By the fourth century CE, Jews are believed to have lost their demographic majority in Syria Palaestina. The long-established Jewish community of Mesopotamia, which had been living under Parthian and later Sasanian rule, beyond the confines of the Roman Empire, became an important center of Jewish study as Judea's Jewish population declined. Estimates often place the Babylonian Jewish community of the 3rd to 7th centuries at around one million, making it the largest Jewish diaspora community of that period. Under the political leadership of the exilarch, who was regarded as a royal heir of the House of David, this community had an autonomous status and served as a place of refuge for the Jews of Syria Palaestina. A number of significant Talmudic academies, such as the Nehardea, Pumbedita, and Sura academies, were established in Mesopotamia, and many important Amoraim were active there. The Babylonian Talmud, a centerpiece of Jewish religious law, was compiled in Babylonia in the 3rd to 6th centuries. Jewish diaspora communities are generally described to have coalesced into three major ethnic subdivisions according to where their ancestors settled: the Ashkenazim (initially in the Rhineland and France), the Sephardim (initially in the Iberian Peninsula), and the Mizrahim (Middle East and North Africa). Romaniote Jews, Tunisian Jews, Yemenite Jews, Egyptian Jews, Ethiopian Jews, Bukharan Jews, Mountain Jews, and other groups also predated the arrival of the Sephardic diaspora. During the same period, Jewish communities in the Middle East thrived under Islamic rule, especially in cities like Baghdad, Cairo, and Damascus. In Babylonia, from the 7th to 11th centuries the Pumbedita and Sura academies led the Arab and to an extent the entire Jewish world. The deans and students of said academies defined the Geonic period in Jewish history. Following this period were the Rishonim who lived from the 11th to 15th centuries. Like their European counterparts, Jews in the Middle East and North Africa also faced periods of persecution and discriminatory policies, with the Almohad Caliphate in North Africa and Iberia issuing forced conversion decrees, causing Jews such as Maimonides to seek safety in other regions. Despite experiencing repeated waves of persecution, Ashkenazi Jews in Western Europe worked in a variety of fields, making an impact on their communities' economy and societies. In Francia, for example, figures like Isaac Judaeus and Armentarius occupied prominent social and economic positions. Francia also witnessed the development of a sophisticated tradition of biblical commentary, as exemplified by Rashi and the tosafists. In 1144, the first documented blood libel occurred in Norwich, England, marking an escalation in the pattern of discrimination and violence that Jews had already been subjected to throughout medieval Europe. During the 12th and 13th centuries, Jews faced frequent antisemitic legislation - including laws prescribing distinctive dress - alongside segregation, repeated blood libels, pogroms, and massacres such as the Rhineland Massacres (1066). The Jews of the Holy Roman Empire were designated Servi camerae regis (“servants of the imperial chamber”) by Frederick II, a status that afforded limited protection while simultaneously entangling them in the political struggles between the emperor and the German principalities and cities. Persecution intensified during the Black Death in the mid-14th century, when Jews were accused of poisoning wells and many communities were destroyed. These pressures, combined with major expulsions such as that from England in 1290, gradually pushed Ashkenazi Jewish populations eastward into Poland, Lithuania, and Russia. One of the largest Jewish communities of the Middle Ages was in the Iberian Peninsula, which for a time contained the largest Jewish population in Europe. Iberian Jewry endured discrimination under the Visigoths but saw its fortunes improve under Umayyad rule and later the Taifa kingdoms. During this period, the Jews of Muslim Spain entered a "Golden Age" marked by achievements in Hebrew poetry and literature, religious scholarship, grammar, medicine and science, with leading figures including Hasdai ibn Shaprut, Judah Halevi, Moses ibn Ezra and Solomon ibn Gabirol. Jews also rose to high office, most notably Samuel ibn Naghrillah, a scholar and poet who served as grand vizier and military commander of Granada. The Golden Age ended with the rise of the radical Almoravid and Almohad dynasties, whose persecutions drove many Jews from Iberia (including Maimonides), together with the advancing Reconquista. In 1391, widespread pogroms swept across Spain, leaving thousands dead and forcing mass conversions. The Spanish Inquisition was later established to pursue, torture and execute conversos who continued to practice Judaism in secret, while public disputations were staged to discredit Judaism. In 1492, after the Reconquista, Isabella I of Castile and Ferdinand II of Aragon decreed the expulsion of all Jews who refused conversion, sending an estimated 200,000 into exile in Portugal, Italy, North Africa, and the Ottoman Empire. In 1497, Portugal's Jews, about 30,000, were formally ordered expelled but instead were forcibly converted to retain their economic role. In 1498, some 3,500 Jews were expelled from Navarre. Many converts outwardly adopted Christianity while secretly preserving Jewish practices, becoming crypto-Jews (also known as marranos or anusim), who remained targets of the various Inquisitions for centuries. Following the expulsions from Spain and Portugal in the 1490s, Jewish exiles dispersed across the Mediterranean, Europe, and North Africa. Many settled in the Ottoman Empire—which, replacing the Iberian Peninsula, became home to the world's largest Jewish population—where new communities developed in Anatolia, the Balkans, and the Land of Israel. Cities such as Istanbul and Thessaloniki grew into major Jewish centers, while in 16th-century Safed a flourishing spiritual life took shape. There, Solomon Alkabetz, Moses Cordovero, and Isaac Luria developed influential new schools of Kabbalah, giving powerful impetus to Jewish mysticism, and Joseph Karo composed the Shulchan Aruch, which became a cornerstone of Jewish law. In the 17th century, Portuguese conversos who returned to Judaism and engaged in trade and banking helped establish Amsterdam as a prosperous Jewish center, while also forming communities in cities such as Antwerp and London. This period also witnessed waves of messianic fervor, most notably the rise of the Sabbatean movement in the 1660s, led by Sabbatai Zvi of İzmir, which reverberated throughout the Jewish world. In Eastern Europe, Poland–Lithuania became the principal center of Ashkenazi Jewry, eventually becoming home to the largest Jewish population in the world. Jewish life flourished there from in the early modern era, supported by relative stability, economic opportunity, and strong communal institutions. The mid-17th century brought devastation with the Cossack uprisings in Ukraine, which reversed migration flows and sent refugees westward, yet Poland–Lithuania remained the demographic and cultural heartland of Ashkenazic Jewry. Following the partitions of Poland, most of its Jews came under Russian rule and were confined to the "Pale of Settlement." The 18th century also witnessed new religious and intellectual currents. Hasidism, founded by Baal Shem Tov, emphasized mysticism and piety, while its opponents, the Misnagdim ("opponents") led by the Vilna Gaon, defended rabbinic scholarship and tradition. In Western Europe, during the 1760s and 1770s, the Haskalah (Jewish Enlightenment) emerged in German-speaking lands, where figures such as Moses Mendelssohn promoted secular learning, vernacular literacy, and integration into European society. Elsewhere, Jews began to be re-admitted to Western Europe, including England, where Menasseh ben Israel petitioned Oliver Cromwell for their return. In the Americas, Jews of Sephardic descent first arrived as conversos in Spanish and Portuguese colonies, where many faced trial by Inquisition tribunals for "judaizing." A more durable presence began in Dutch Brazil, where Jews openly practiced their religion and established the first synagogues in the New World, before the Portuguese reconquest forced their dispersal to Amsterdam, the Caribbean, and North America. Sephardic communities took root in Curaçao, Suriname, Jamaica, and Barbados, later joined by Ashkenazi migrants. In North America, Jews were present from the mid-17th century, with New Amsterdam hosting the first organized congregation in 1654. By the time of the American Revolution, small communities in New York, Newport, Philadelphia, Savannah, and Charleston played an active role in the struggle for independence. In the late 19th century, Jews in Western Europe gradually achieved legal emancipation, though social acceptance remained limited by persistent antisemitism and rising nationalism. In Eastern Europe, particularly within the Russian Empire's Pale of Settlement, Jews faced mounting legal restrictions and recurring pogroms. From this environment emerged Zionism, a national revival movement originating in Central and Eastern Europe that sought to re-establish a Jewish polity in the Land of Israel as a means of returning the Jewish people to their ancestral homeland and ending centuries of exile and persecution. This led to waves of Jewish migration to Ottoman-controlled Palestine. Theodor Herzl, who is considered the father of political Zionism, offered his vision of a future Jewish state in his 1896 book Der Judenstaat (The Jewish State); a year later, he presided over the First Zionist Congress. The antisemitism that inflicted Jewish communities in Europe also triggered a mass exodus of 2.8 million Jews to the United States between 1881 and 1924. Despite this, some Jews of Europe and the United States were able to make great achievements in various fields of science and culture. Among the most influential from this period are Albert Einstein in physics, Sigmund Freud in psychology, Franz Kafka in literature, and Irving Berlin in music. Many Nobel Prize winners at this time were Jewish, as is still the case. When Adolf Hitler and the Nazi Party came to power in Germany in 1933, the situation for Jews deteriorated rapidly as a direct result of Nazi policies. Many Jews fled from Europe to Mandatory Palestine, the United States, and the Soviet Union as a result of racial anti-Semitic laws, economic difficulties, and the fear of an impending war. World War II started in 1939, and by 1941, Hitler occupied almost all of Europe. Following the German invasion of the Soviet Union in 1941, the Final Solution—an extensive, organized effort with an unprecedented scope intended to annihilate the Jewish people—began, and resulted in the persecution and murder of Jews in Europe and North Africa. In Poland, three million were murdered in gas chambers in all concentration camps combined, with one million at the Auschwitz camp complex alone. The Holocaust is the name given to this genocide, in which six million Jews in total were systematically murdered. Before and during the Holocaust, enormous numbers of Jews immigrated to Mandatory Palestine. In 1944, the Jewish insurgency in Mandatory Palestine began with the aim of gaining full independence from the United Kingdom. On 14 May 1948, upon the termination of the mandate, David Ben-Gurion declared the creation of the State of Israel, a Jewish and democratic state. Immediately afterwards, all neighboring Arab states invaded, and were resisted by the newly formed Israel Defense Forces. In 1949, the war ended and Israel started building its state and absorbing waves of Aliyah, granting citizenship to Jews all over the world via the Law of Return passed in 1950. However, both the Israeli–Palestinian conflict and wider Arab–Israeli conflict continue to this day. Culture The Jewish people and the religion of Judaism are strongly interrelated. Converts to Judaism have a status within the Jewish people equal to those born into it. However, converts who go on to practice no Judaism are likely to be viewed with skepticism. Mainstream Judaism does not proselytize, and conversion is considered a difficult task. A significant portion of conversions are undertaken by children of mixed marriages, or would-be or current spouses of Jews. The Hebrew Bible, a religious interpretation of the traditions and early history of the Jews, established the first of the Abrahamic religions, which are now practiced by 54 percent of the world. Judaism guides its adherents in both practice and belief, and has been called not only a religion, but also a "way of life," which has made drawing a clear distinction between Judaism, Jewish culture, and Jewish identity rather difficult. Throughout history, in eras and places as diverse as the ancient Hellenic world, in Europe before and after The Age of Enlightenment (see Haskalah), in Islamic Spain and Portugal, in North Africa and the Middle East, India, China, or the contemporary United States and Israel, cultural phenomena have developed that are in some sense characteristically Jewish without being at all specifically religious. Some factors in this come from within Judaism, others from the interaction of Jews or specific communities of Jews with their surroundings, and still others from the inner social and cultural dynamics of the community, as opposed to from the religion itself. This phenomenon has led to considerably different Jewish cultures unique to their own communities. Hebrew is the liturgical language of Judaism (termed lashon ha-kodesh, "the holy tongue"), the language in which most of the Hebrew scriptures (Tanakh) were composed, and the daily speech of the Jewish people for centuries. By the 5th century BCE, Aramaic, a closely related tongue, joined Hebrew as the spoken language in Judea. By the 3rd century BCE, some Jews of the diaspora were speaking Greek. Others, such as in the Jewish communities of Asoristan, known to Jews as Babylonia, were speaking Hebrew and Aramaic, the languages of the Babylonian Talmud. Dialects of these same languages were also used by the Jews of Syria Palaestina at that time.[citation needed] For centuries, Jews worldwide have spoken the local or dominant languages of the regions they migrated to, often developing distinctive dialectal forms or branches that became independent languages. Yiddish is the Judaeo-German language developed by Ashkenazi Jews who migrated to Central Europe. Ladino is the Judaeo-Spanish language developed by Sephardic Jews who migrated to the Iberian Peninsula. Due to many factors, including the impact of the Holocaust on European Jewry, the Jewish exodus from Arab and Muslim countries, and widespread emigration from other Jewish communities around the world, ancient and distinct Jewish languages of several communities, including Judaeo-Georgian, Judaeo-Arabic, Judaeo-Berber, Krymchak, Judaeo-Malayalam and many others, have largely fallen out of use. For over sixteen centuries Hebrew was used almost exclusively as a liturgical language, and as the language in which most books had been written on Judaism, with a few speaking only Hebrew on the Sabbath. Hebrew was revived as a spoken language by Eliezer ben Yehuda, who arrived in Palestine in 1881. It had not been used as a mother tongue since Tannaic times. Modern Hebrew is designated as the "State language" of Israel. Despite efforts to revive Hebrew as the national language of the Jewish people, knowledge of the language is not commonly possessed by Jews worldwide and English has emerged as the lingua franca of the Jewish diaspora. Although many Jews once had sufficient knowledge of Hebrew to study the classic literature, and Jewish languages like Yiddish and Ladino were commonly used as recently as the early 20th century, most Jews lack such knowledge today and English has by and large superseded most Jewish vernaculars. The three most commonly spoken languages among Jews today are Hebrew, English, and Russian. Some Romance languages, particularly French and Spanish, are also widely used. Yiddish has been spoken by more Jews in history than any other language, but it is far less used today following the Holocaust and the adoption of Modern Hebrew by the Zionist movement and the State of Israel. In some places, the mother language of the Jewish community differs from that of the general population or the dominant group. For example, in Quebec, the Ashkenazic majority has adopted English, while the Sephardic minority uses French as its primary language. Similarly, South African Jews adopted English rather than Afrikaans. Due to both Czarist and Soviet policies, Russian has superseded Yiddish as the language of Russian Jews, but these policies have also affected neighboring communities. Today, Russian is the first language for many Jewish communities in a number of Post-Soviet states, such as Ukraine and Uzbekistan,[better source needed] as well as for Ashkenazic Jews in Azerbaijan, Georgia, and Tajikistan. Although communities in North Africa today are small and dwindling, Jews there had shifted from a multilingual group to a monolingual one (or nearly so), speaking French in Algeria, Morocco, and the city of Tunis, while most North Africans continue to use Arabic or Berber as their mother tongue.[citation needed] There is no single governing body for the Jewish community, nor a single authority with responsibility for religious doctrine. Instead, a variety of secular and religious institutions at the local, national, and international levels lead various parts of the Jewish community on a variety of issues. Today, many countries have a Chief Rabbi who serves as a representative of that country's Jewry. Although many Hasidic Jews follow a certain hereditary Hasidic dynasty, there is no one commonly accepted leader of all Hasidic Jews. Many Jews believe that the Messiah will act a unifying leader for Jews and the entire world. A number of modern scholars of nationalism support the existence of Jewish national identity in antiquity. One of them is David Goodblatt, who generally believes in the existence of nationalism before the modern period. In his view, the Bible, the parabiblical literature and the Jewish national history provide the base for a Jewish collective identity. Although many of the ancient Jews were illiterate (as were their neighbors), their national narrative was reinforced through public readings. The Hebrew language also constructed and preserved national identity. Although it was not widely spoken after the 5th century BCE, Goodblatt states: the mere presence of the language in spoken or written form could invoke the concept of a Jewish national identity. Even if one knew no Hebrew or was illiterate, one could recognize that a group of signs was in Hebrew script. ... It was the language of the Israelite ancestors, the national literature, and the national religion. As such it was inseparable from the national identity. Indeed its mere presence in visual or aural medium could invoke that identity. Anthony D. Smith, an historical sociologist considered one of the founders of the field of nationalism studies, wrote that the Jews of the late Second Temple period provide "a closer approximation to the ideal type of the nation [...] than perhaps anywhere else in the ancient world." He adds that this observation "must make us wary of pronouncing too readily against the possibility of the nation, and even a form of religious nationalism, before the onset of modernity." Agreeing with Smith, Goodblatt suggests omitting the qualifier "religious" from Smith's definition of ancient Jewish nationalism, noting that, according to Smith, a religious component in national memories and culture is common even in the modern era. This view is echoed by political scientist Tom Garvin, who writes that "something strangely like modern nationalism is documented for many peoples in medieval times and in classical times as well," citing the ancient Jews as one of several "obvious examples", alongside the classical Greeks and the Gaulish and British Celts. Fergus Millar suggests that the sources of Jewish national identity and their early nationalist movements in the first and second centuries CE included several key elements: the Bible as both a national history and legal source, the Hebrew language as a national language, a system of law, and social institutions such as schools, synagogues, and Sabbath worship. Adrian Hastings argued that Jews are the "true proto-nation", that through the model of ancient Israel found in the Hebrew Bible, provided the world with the original concept of nationhood which later influenced Christian nations. However, following Jerusalem's destruction in the first century CE, Jews ceased to be a political entity and did not resemble a traditional nation-state for almost two millennia. Despite this, they maintained their national identity through collective memory, religion and sacred texts, even without land or political power, and remained a nation rather than just an ethnic group, eventually leading to the rise of Zionism and the establishment of Israel. Steven Weitzman suggests that Jewish nationalist sentiment in antiquity was encouraged because under foreign rule (Persians, Greeks, Romans) Jews were able to claim that they were an ancient nation. This claim was based on the preservation and reverence of their scriptures, the Hebrew language, the Temple and priesthood, and other traditions of their ancestors. Doron Mendels further observes that the Hasmonean kingdom, one of the few examples of indigenous statehood at its time, significantly reinforced Jewish national consciousness. The memory of this period of independence contributed to the persistent efforts to revive Jewish sovereignty in Judea, leading to the major revolts against Roman rule in the 1st and 2nd centuries CE. Demographics Within the world's Jewish population there are distinct ethnic divisions, most of which are primarily the result of geographic branching from an originating Israelite population, and subsequent independent evolutions. An array of Jewish communities was established by Jewish settlers in various places around the Old World, often at great distances from one another, resulting in effective and often long-term isolation. During the millennia of the Jewish diaspora the communities would develop under the influence of their local environments: political, cultural, natural, and populational. Today, manifestations of these differences among the Jews can be observed in Jewish cultural expressions of each community, including Jewish linguistic diversity, culinary preferences, liturgical practices, religious interpretations, as well as degrees and sources of genetic admixture. Jews are often identified as belonging to one of two major groups: the Ashkenazim and the Sephardim. Ashkenazim are so named in reference to their geographical origins (their ancestors' culture coalesced in the Rhineland, an area historically referred to by Jews as Ashkenaz). Similarly, Sephardim (Sefarad meaning "Spain" in Hebrew) are named in reference their origins in Iberia. The diverse groups of Jews of the Middle East and North Africa are often collectively referred to as Sephardim together with Sephardim proper for liturgical reasons having to do with their prayer rites. A common term for many of these non-Spanish Jews who are sometimes still broadly grouped as Sephardim is Mizrahim (lit. 'easterners' in Hebrew). Nevertheless, Mizrahis and Sepharadim are usually ethnically distinct. Smaller groups include, but are not restricted to, Indian Jews such as the Bene Israel, Bnei Menashe, Cochin Jews, and Bene Ephraim; the Romaniotes of Greece; the Italian Jews ("Italkim" or "Bené Roma"); the Teimanim from Yemen; various African Jews, including most numerously the Beta Israel of Ethiopia; and Chinese Jews, most notably the Kaifeng Jews, as well as various other distinct but now almost extinct communities. The divisions between all these groups are approximate and their boundaries are not always clear. The Mizrahim for example, are a heterogeneous collection of North African, Central Asian, Caucasian, and Middle Eastern Jewish communities that are no closer related to each other than they are to any of the earlier mentioned Jewish groups. In modern usage, however, the Mizrahim are sometimes termed Sephardi due to similar styles of liturgy, despite independent development from Sephardim proper. Thus, among Mizrahim there are Egyptian Jews, Iraqi Jews, Lebanese Jews, Kurdish Jews, Moroccan Jews, Libyan Jews, Syrian Jews, Bukharian Jews, Mountain Jews, Georgian Jews, Iranian Jews, Afghan Jews, and various others. The Teimanim from Yemen are sometimes included, although their style of liturgy is unique and they differ in respect to the admixture found among them to that found in Mizrahim. In addition, there is a differentiation made between Sephardi migrants who established themselves in the Middle East and North Africa after the expulsion of the Jews from Spain and Portugal in the 1490s and the pre-existing Jewish communities in those regions. Ashkenazi Jews represent the bulk of modern Jewry, with at least 70 percent of Jews worldwide (and up to 90 percent prior to World War II and the Holocaust). As a result of their emigration from Europe, Ashkenazim also represent the overwhelming majority of Jews in the New World continents, in countries such as the United States, Canada, Argentina, Australia, and Brazil. In France, the immigration of Jews from Algeria (Sephardim) has led them to outnumber the Ashkenazim. Only in Israel is the Jewish population representative of all groups, a melting pot independent of each group's proportion within the overall world Jewish population. Y DNA studies tend to imply a small number of founders in an old population whose members parted and followed different migration paths. In most Jewish populations, these male line ancestors appear to have been mainly Middle Eastern. For example, Ashkenazi Jews share more common paternal lineages with other Jewish and Middle Eastern groups than with non-Jewish populations in areas where Jews lived in Eastern Europe, Germany, and the French Rhine Valley. This is consistent with Jewish traditions in placing most Jewish paternal origins in the region of the Middle East. Conversely, the maternal lineages of Jewish populations, studied by looking at mitochondrial DNA, are generally more heterogeneous. Scholars such as Harry Ostrer and Raphael Falk believe this indicates that many Jewish males found new mates from European and other communities in the places where they migrated in the diaspora after fleeing ancient Israel. In contrast, Behar has found evidence that about 40 percent of Ashkenazi Jews originate maternally from just four female founders, who were of Middle Eastern origin. The populations of Sephardi and Mizrahi Jewish communities "showed no evidence for a narrow founder effect." Subsequent studies carried out by Feder et al. confirmed the large portion of non-local maternal origin among Ashkenazi Jews. Reflecting on their findings related to the maternal origin of Ashkenazi Jews, the authors conclude "Clearly, the differences between Jews and non-Jews are far larger than those observed among the Jewish communities. Hence, differences between the Jewish communities can be overlooked when non-Jews are included in the comparisons." However, a 2025 genetic study on the Ashkenazi Jewish founder population supports the presence of a substantial Near Eastern component in the maternal lineages. Analyses of mitochondrial DNA (mtDNA) indicate that the core founder lineages, estimated at around 54, likely originated from the Near East, with these founder signatures appearing in multiple copies across the population. While later admixture introduced additional mtDNA lineages, these absorbed lineages are distinguishable from the original founders. The findings are consistent with genome-wide Identity-by-Descent and Lineage Extinction analyses, reinforcing the Near Eastern origin of the Ashkenazi maternal founders. A study showed that 7% of Ashkenazi Jews have the haplogroup G2c, which is mainly found in Pashtuns and on lower scales all major Jewish groups, Palestinians, Syrians, and Lebanese. Studies of autosomal DNA, which look at the entire DNA mixture, have become increasingly important as the technology develops. They show that Jewish populations have tended to form relatively closely related groups in independent communities, with most in a community sharing significant ancestry in common. For Jewish populations of the diaspora, the genetic composition of Ashkenazi, Sephardic, and Mizrahi Jewish populations show a predominant amount of shared Middle Eastern ancestry. According to Behar, the most parsimonious explanation for this shared Middle Eastern ancestry is that it is "consistent with the historical formulation of the Jewish people as descending from ancient Hebrew and Israelite residents of the Levant" and "the dispersion of the people of ancient Israel throughout the Old World". North African, Italian and others of Iberian origin show variable frequencies of admixture with non-Jewish historical host populations among the maternal lines. In the case of Ashkenazi and Sephardi Jews (in particular Moroccan Jews), who are closely related, the source of non-Jewish admixture is mainly Southern European, while Mizrahi Jews show evidence of admixture with other Middle Eastern populations. Behar et al. have remarked on a close relationship between Ashkenazi Jews and modern Italians. A 2001 study found that Jews were more closely related to groups of the Fertile Crescent (Kurds, Turks, and Armenians) than to their Arab neighbors, whose genetic signature was found in geographic patterns reflective of Islamic conquests. The studies also show that Sephardic Bnei Anusim (descendants of the "anusim" who were forced to convert to Catholicism), which comprise up to 19.8 percent of the population of today's Iberia (Spain and Portugal) and at least 10 percent of the population of Ibero-America (Hispanic America and Brazil), have Sephardic Jewish ancestry within the last few centuries. The Bene Israel and Cochin Jews of India, Beta Israel of Ethiopia, and a portion of the Lemba people of Southern Africa, despite more closely resembling the local populations of their native countries, have also been thought to have some more remote ancient Jewish ancestry. Views on the Lemba have changed and genetic Y-DNA analyses in the 2000s have established a partially Middle-Eastern origin for a portion of the male Lemba population but have been unable to narrow this down further. Although historically, Jews have been found all over the world, in the decades since World War II and the establishment of Israel, they have increasingly concentrated in a small number of countries. In 2021, Israel and the United States together accounted for over 85 percent of the global Jewish population, with approximately 45.3% and 39.6% of the world's Jews, respectively. More than half (51.2%) of world Jewry resides in just ten metropolitan areas. As of 2021, these ten areas were Tel Aviv, New York, Jerusalem, Haifa, Los Angeles, Miami, Philadelphia, Paris, Washington, and Chicago. The Tel Aviv metro area has the highest percent of Jews among the total population (94.8%), followed by Jerusalem (72.3%), Haifa (73.1%), and Beersheba (60.4%), the balance mostly being Israeli Arabs. Outside Israel, the highest percent of Jews in a metropolitan area was in New York (10.8%), followed by Miami (8.7%), Philadelphia (6.8%), San Francisco (5.1%), Washington (4.7%), Los Angeles (4.7%), Toronto (4.5%), and Baltimore (4.1%). As of 2010, there were nearly 14 million Jews around the world, roughly 0.2% of the world's population at the time. According to the 2007 estimates of The Jewish People Policy Planning Institute, the world's Jewish population is 13.2 million. This statistic incorporates both practicing Jews affiliated with synagogues and the Jewish community, and approximately 4.5 million unaffiliated and secular Jews.[citation needed] According to Sergio Della Pergola, a demographer of the Jewish population, in 2021 there were about 6.8 million Jews in Israel, 6 million in the United States, and 2.3 million in the rest of the world. Israel, the Jewish nation-state, is the only country in which Jews make up a majority of the citizens. Israel was established as an independent democratic and Jewish state on 14 May 1948. Of the 120 members in its parliament, the Knesset, as of 2016[update], 14 members of the Knesset are Arab citizens of Israel (not including the Druze), most representing Arab political parties. One of Israel's Supreme Court judges is also an Arab citizen of Israel. Between 1948 and 1958, the Jewish population rose from 800,000 to two million. Currently, Jews account for 75.4 percent of the Israeli population, or 6 million people. The early years of the State of Israel were marked by the mass immigration of Holocaust survivors in the aftermath of the Holocaust and Jews fleeing Arab lands. Israel also has a large population of Ethiopian Jews, many of whom were airlifted to Israel in the late 1980s and early 1990s. Between 1974 and 1979 nearly 227,258 immigrants arrived in Israel, about half being from the Soviet Union. This period also saw an increase in immigration to Israel from Western Europe, Latin America, and North America. A trickle of immigrants from other communities has also arrived, including Indian Jews and others, as well as some descendants of Ashkenazi Holocaust survivors who had settled in countries such as the United States, Argentina, Australia, Chile, and South Africa. Some Jews have emigrated from Israel elsewhere, because of economic problems or disillusionment with political conditions and the continuing Arab–Israeli conflict. Jewish Israeli emigrants are known as yordim. The waves of immigration to the United States and elsewhere at the turn of the 19th century, the founding of Zionism and later events, including pogroms in Imperial Russia (mostly within the Pale of Settlement in present-day Ukraine, Moldova, Belarus and eastern Poland), the massacre of European Jewry during the Holocaust, and the founding of the state of Israel, with the subsequent Jewish exodus from Arab lands, all resulted in substantial shifts in the population centers of world Jewry by the end of the 20th century. More than half of the Jews live in the Diaspora (see Population table). Currently, the largest Jewish community outside Israel, and either the largest or second-largest Jewish community in the world, is located in the United States, with 6 million to 7.5 million Jews by various estimates. Elsewhere in the Americas, there are also large Jewish populations in Canada (315,000), Argentina (180,000–300,000), and Brazil (196,000–600,000), and smaller populations in Mexico, Uruguay, Venezuela, Chile, Colombia and several other countries (see History of the Jews in Latin America). According to a 2010 Pew Research Center study, about 470,000 people of Jewish heritage live in Latin America and the Caribbean. Demographers disagree on whether the United States has a larger Jewish population than Israel, with many maintaining that Israel surpassed the United States in Jewish population during the 2000s, while others maintain that the United States still has the largest Jewish population in the world. Currently, a major national Jewish population survey is planned to ascertain whether or not Israel has overtaken the United States in Jewish population. Western Europe's largest Jewish community, and the third-largest Jewish community in the world, can be found in France, home to between 483,000 and 500,000 Jews, the majority of whom are immigrants or refugees from North African countries such as Algeria, Morocco, and Tunisia (or their descendants). The United Kingdom has a Jewish community of 292,000. In Eastern Europe, the exact figures are difficult to establish. The number of Jews in Russia varies widely according to whether a source uses census data (which requires a person to choose a single nationality among choices that include "Russian" and "Jewish") or eligibility for immigration to Israel (which requires that a person have one or more Jewish grandparents). According to the latter criteria, the heads of the Russian Jewish community assert that up to 1.5 million Russians are eligible for aliyah. In Germany, the 102,000 Jews registered with the Jewish community are a slowly declining population, despite the immigration of tens of thousands of Jews from the former Soviet Union since the fall of the Berlin Wall. Thousands of Israelis also live in Germany, either permanently or temporarily, for economic reasons. Prior to 1948, approximately 800,000 Jews were living in lands which now make up the Arab world (excluding Israel). Of these, just under two-thirds lived in the French-controlled Maghreb region, 15 to 20 percent in the Kingdom of Iraq, approximately 10 percent in the Kingdom of Egypt and approximately 7 percent in the Kingdom of Yemen. A further 200,000 lived in Pahlavi Iran and the Republic of Turkey. Today, around 26,000 Jews live in Muslim-majority countries, mainly in Turkey (14,200) and Iran (9,100), while Morocco (2,000), Tunisia (1,000), and the United Arab Emirates (500) host the largest communities in the Arab world. A small-scale exodus had begun in many countries in the early decades of the 20th century, although the only substantial aliyah came from Yemen and Syria. The exodus from Arab and Muslim countries took place primarily from 1948. The first large-scale exoduses took place in the late 1940s and early 1950s, primarily in Iraq, Yemen and Libya, with up to 90 percent of these communities leaving within a few years. The peak of the exodus from Egypt occurred in 1956. The exodus in the Maghreb countries peaked in the 1960s. Lebanon was the only Arab country to see a temporary increase in its Jewish population during this period, due to an influx of refugees from other Arab countries, although by the mid-1970s the Jewish community of Lebanon had also dwindled. In the aftermath of the exodus wave from Arab states, an additional migration of Iranian Jews peaked in the 1980s when around 80 percent of Iranian Jews left the country.[citation needed] Outside Europe, the Americas, the Middle East, and the rest of Asia, there are significant Jewish populations in Australia (112,500) and South Africa (70,000). There is also a 6,800-strong community in New Zealand. Since at least the time of the Ancient Greeks, a proportion of Jews have assimilated into the wider non-Jewish society around them, by either choice or force, ceasing to practice Judaism and losing their Jewish identity. Assimilation took place in all areas, and during all time periods, with some Jewish communities, for example the Kaifeng Jews of China, disappearing entirely. The advent of the Jewish Enlightenment of the 18th century (see Haskalah) and the subsequent emancipation of the Jewish populations of Europe and America in the 19th century, accelerated the situation, encouraging Jews to increasingly participate in, and become part of, secular society. The result has been a growing trend of assimilation, as Jews marry non-Jewish spouses and stop participating in the Jewish community. Rates of interreligious marriage vary widely: In the United States, it is just under 50 percent; in the United Kingdom, around 53 percent; in France, around 30 percent; and in Australia and Mexico, as low as 10 percent. In the United States, only about a third of children from intermarriages affiliate with Jewish religious practice. The result is that most countries in the Diaspora have steady or slightly declining religiously Jewish populations as Jews continue to assimilate into the countries in which they live.[citation needed] The Jewish people and Judaism have experienced various persecutions throughout their history. During Late Antiquity and the Early Middle Ages, the Roman Empire (in its later phases known as the Byzantine Empire) repeatedly repressed the Jewish population, first by ejecting them from their homelands during the pagan Roman era and later by officially establishing them as second-class citizens during the Christian Roman era. According to James Carroll, "Jews accounted for 10% of the total population of the Roman Empire. By that ratio, if other factors had not intervened, there would be 200 million Jews in the world today, instead of something like 13 million." Later in medieval Western Europe, further persecutions of Jews by Christians occurred, notably during the Crusades—when Jews all over Germany were massacred—and in a series of expulsions from the Kingdom of England, Germany, and France. Then there occurred the largest expulsion of all, when Spain and Portugal, after the Reconquista (the Catholic Reconquest of the Iberian Peninsula), expelled both unbaptized Sephardic Jews and the ruling Muslim Moors. In the Papal States, which existed until 1870, Jews were required to live only in specified neighborhoods called ghettos. Islam and Judaism have a complex relationship. Traditionally Jews and Christians living in Muslim lands, known as dhimmis, were allowed to practice their religions and administer their internal affairs, but they were subject to certain conditions. They had to pay the jizya (a per capita tax imposed on free adult non-Muslim males) to the Islamic state. Dhimmis had an inferior status under Islamic rule. They had several social and legal disabilities such as prohibitions against bearing arms or giving testimony in courts in cases involving Muslims. Many of the disabilities were highly symbolic. The one described by Bernard Lewis as "most degrading" was the requirement of distinctive clothing, not found in the Quran or hadith but invented in early medieval Baghdad; its enforcement was highly erratic. On the other hand, Jews rarely faced martyrdom or exile, or forced compulsion to change their religion, and they were mostly free in their choice of residence and profession. Notable exceptions include the massacre of Jews and forcible conversion of some Jews by the rulers of the Almohad dynasty in Al-Andalus in the 12th century, as well as in Islamic Persia, and the forced confinement of Moroccan Jews to walled quarters known as mellahs beginning from the 15th century and especially in the early 19th century. In modern times, it has become commonplace for standard antisemitic themes to be conflated with anti-Zionist publications and pronouncements of Islamic movements such as Hezbollah and Hamas, in the pronouncements of various agencies of the Islamic Republic of Iran, and even in the newspapers and other publications of Turkish Refah Partisi."[better source needed] Throughout history, many rulers, empires and nations have oppressed their Jewish populations or sought to eliminate them entirely. Methods employed ranged from expulsion to outright genocide; within nations, often the threat of these extreme methods was sufficient to silence dissent. The history of antisemitism includes the First Crusade which resulted in the massacre of Jews; the Spanish Inquisition (led by Tomás de Torquemada) and the Portuguese Inquisition, with their persecution and autos-da-fé against the New Christians and Marrano Jews; the Bohdan Chmielnicki Cossack massacres in Ukraine; the Pogroms backed by the Russian Tsars; as well as expulsions from Spain, Portugal, England, France, Germany, and other countries in which the Jews had settled. According to a 2008 study published in the American Journal of Human Genetics, 19.8 percent of the modern Iberian population has Sephardic Jewish ancestry, indicating that the number of conversos may have been much higher than originally thought. The persecution reached a peak in Nazi Germany's Final Solution, which led to the Holocaust and the slaughter of approximately 6 million Jews. Of the world's 16 million Jews in 1939, almost 40% were murdered in the Holocaust. The Holocaust—the state-led systematic persecution and genocide of European Jews (and certain communities of North African Jews in European controlled North Africa) and other minority groups of Europe during World War II by Germany and its collaborators—remains the most notable modern-day persecution of Jews. The persecution and genocide were accomplished in stages. Legislation to remove the Jews from civil society was enacted years before the outbreak of World War II. Concentration camps were established in which inmates were used as slave labour until they died of exhaustion or disease. Where the Third Reich conquered new territory in Eastern Europe, specialized units called Einsatzgruppen murdered Jews and political opponents in mass shootings. Jews and Roma were crammed into ghettos before being transported hundreds of kilometres by freight train to extermination camps where, if they survived the journey, the majority of them were murdered in gas chambers. Virtually every arm of Germany's bureaucracy was involved in the logistics of the mass murder, turning the country into what one Holocaust scholar has called "a genocidal nation." Throughout Jewish history, Jews have repeatedly been directly or indirectly expelled from both their original homeland, the Land of Israel, and many of the areas in which they have settled. This experience as refugees has shaped Jewish identity and religious practice in many ways, and is thus a major element of Jewish history. In summary, the pogroms in Eastern Europe, the rise of modern antisemitism, the Holocaust, as well as the rise of Arab nationalism, all served to fuel the movements and migrations of huge segments of Jewry from land to land and continent to continent until they arrived back in large numbers at their original historical homeland in Israel. In the Bible, the patriarch Abraham is described as a migrant to the land of Canaan from Ur of the Chaldees. His descendants, the Children of Israel, undertook the Exodus (meaning "departure" or "exit" in Greek) from ancient Egypt, as described in the Book of Exodus. The first movement documented in the historical record occurred with the resettlement policy of the Neo-Assyrian Empire, which mandated the deportation of conquered peoples, and it is estimated some 4,500,000 among its captive populations suffered this dislocation over three centuries of Assyrian rule. With regard to Israel, Tiglath-Pileser III claims he deported 80% of the population of Lower Galilee, some 13,520 people. Some 27,000 Israelites, 20 to 25% of the population of the Kingdom of Israel, were described as being deported by Sargon II, and were replaced by other deported populations and sent into permanent exile by Assyria, initially to the Upper Mesopotamian provinces of the Assyrian Empire. Between 10,000 and 80,000 people from the Kingdom of Judah were similarly exiled by Babylonia, but these people were then returned to Judea by Cyrus the Great of the Persian Achaemenid Empire. Many Jews were exiled again by the Roman Empire. The 2,000 year dispersion of the Jewish diaspora beginning under the Roman Empire, as Jews were spread throughout the Roman world and, driven from land to land, settled wherever they could live freely enough to practice their religion. Over the course of the diaspora the center of Jewish life moved from Babylonia to the Iberian Peninsula to Poland to the United States and, as a result of Zionism, back to Israel. There were also many expulsions of Jews during the Middle Ages and Enlightenment in Europe, including: 1290, 16,000 Jews were expelled from England, (see the Statute of Jewry); in 1396, 100,000 from France; in 1421, thousands were expelled from Austria. Many of these Jews settled in East-Central Europe, especially Poland. Following the Spanish Inquisition in 1492, the Spanish population of around 200,000 Sephardic Jews were expelled by the Spanish crown and Catholic church, followed by expulsions in 1493 in Sicily (37,000 Jews) and Portugal in 1496. The expelled Jews fled mainly to the Ottoman Empire, the Netherlands, and North Africa, others migrating to Southern Europe and the Middle East. During the 19th century, France's policies of equal citizenship regardless of religion led to the immigration of Jews (especially from Eastern and Central Europe). This contributed to the arrival of millions of Jews in the New World. Over two million Eastern European Jews arrived in the United States from 1880 to 1925. In the latest phase of migrations, the Islamic Revolution of Iran caused many Iranian Jews to flee Iran. Most found refuge in the US (particularly Los Angeles, California, and Long Island, New York) and Israel. Smaller communities of Persian Jews exist in Canada and Western Europe. Similarly, when the Soviet Union collapsed, many of the Jews in the affected territory (who had been refuseniks) were suddenly allowed to leave. This produced a wave of migration to Israel in the early 1990s. Israel is the only country with a Jewish population that is consistently growing through natural population growth, although the Jewish populations of other countries, in Europe and North America, have recently increased through immigration. In the Diaspora, in almost every country the Jewish population in general is either declining or steady, but Orthodox and Haredi Jewish communities, whose members often shun birth control for religious reasons, have experienced rapid population growth. Orthodox and Conservative Judaism discourage proselytism to non-Jews, but many Jewish groups have tried to reach out to the assimilated Jewish communities of the Diaspora in order for them to reconnect to their Jewish roots. Additionally, while in principle Reform Judaism favours seeking new members for the faith, this position has not translated into active proselytism, instead taking the form of an effort to reach out to non-Jewish spouses of intermarried couples. There is also a trend of Orthodox movements reaching out to secular Jews in order to give them a stronger Jewish identity so there is less chance of intermarriage. As a result of the efforts by these and other Jewish groups over the past 25 years, there has been a trend (known as the Baal teshuva movement) for secular Jews to become more religiously observant, though the demographic implications of the trend are unknown. Additionally, there is also a growing rate of conversion to Jews by Choice of gentiles who make the decision to head in the direction of becoming Jews. Contributions Jewish individuals have played a significant role in the development and growth of Western culture, advancing many fields of thought, science and technology, both historically and in modern times, including through discrete trends in Jewish philosophy, Jewish ethics and Jewish literature, as well as specific trends in Jewish culture, including in Jewish art, Jewish music, Jewish humor, Jewish theatre, Jewish cuisine and Jewish medicine. Jews have established various Jewish political movements, religious movements, and, through the authorship of the Hebrew Bible and parts of the New Testament, provided the foundation for Christianity and Islam. More than 20 percent of the awarded Nobel Prize have gone to individuals of Jewish descent. Philanthropic giving is a widespread core function among Jewish organizations. Notes References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Ponaturi] | [TOKENS: 243] |
Contents Ponaturi In Māori mythology, the Ponaturi are a group of hostile creatures (goblins) who live in a land beneath the sea by day, returning to shore each evening to sleep. They dread daylight, which is fatal to them. They appear in a number of stories, including: Kanae The kanae (or grey mullet) is represented as a companion of the Ponaturi in another version of Tāwhaki (Grey 1956:51). When the Ponaturi come up out of the water to their house Manawa-Tāne, Kanae comes with them. Tāwhaki and Karihi kill all the Ponaturi, in revenge for the death of Hemā, but the mullet escapes by leaping again and again until it gets back to the sea (Craig 1989:99, Grey 1855:40, Tregear 1891:122). In the story of Ruapupuke (or Rua-te-pupuke), the kanae is associated with similar creatures, the horde of Tangaroa, which are not overtly named as Ponaturi: Notes References |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Twenty-fifth_government_of_Israel] | [TOKENS: 233] |
Contents Twenty-fifth government of Israel The twenty-fifth government of Israel was formed by Yitzhak Rabin of the Labor Party on 13 July 1992, after the party's victory in the June elections. The coalition also contained the new Meretz party (an alliance of Ratz, Mapam, and Shinui) and Shas, and held 62 of the 120 seats in the Knesset. The government was also supported, but not joined, by Hadash and the Arab Democratic Party, which held an additional five seats between them. Shas left the government on 14 September 1993, but the coalition was joined by the new Yiud faction (a three-member breakaway from Tzomet) on 9 January 1995. Rabin was assassinated on 4 November 1995, with Shimon Peres taking over as Interim Prime Minister until forming the twenty-sixth government on 22 November. Cabinet members 1 When Rubinstein was appointed to the post in 1994, it was renamed the Minister of Education, Culture and Sport. 2 Died in office. References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/The_Initiative_(company)] | [TOKENS: 1952] |
Contents Perfect Dark (cancelled video game) Perfect Dark is a cancelled first-person shooter game that was under development by The Initiative and Crystal Dynamics for Windows and Xbox Series X/S. As a reboot of the Perfect Dark series, the game aimed to revitalize the franchise with modern gameplay mechanics and a new narrative. Set in a near-future world grappling with environmental collapse, players would have stepped into the role of agent Joanna Dark, navigating a world where megacorporations conceal secrets amidst efforts to restore ecological balance. Announced at The Game Awards in 2020, Perfect Dark was the first project by The Initiative, a first-party team established by Xbox Game Studios in 2018. Production was slowed by various problems, including significant personnel turnover and creative differences. A gameplay trailer was revealed at the Xbox Games Showcase in June 2024. In July 2025, it was announced that the game would cease production and that The Initiative would be closed as part of numerous rounds of layoffs by Microsoft. Gameplay Perfect Dark was intended to be a first-person shooter drawing inspiration from immersive sims and stealth games. Players would have had the option to engage in direct combat with enemies or employ stealth tactics, utilizing advanced gadgets for distraction and infiltration. The game features a parkour system, enabling the protagonist, Joanna, to navigate the environment by clambering down pipes and running along walls. Players can eliminate or incapacitate enemies using firearms and melee combat, with Joanna smoothly sliding and running through combat scenarios. The game allows for environmental interaction, such as shooting a fire extinguisher to create a temporary cover cloud. Stealth was supposed to be a significant component of the gameplay. Players would've been presented with options for how to deal with oncoming encounters and objectives, such as choosing to sneak past adversaries or using high-tech gadgets to deceive them. Joanna's HUD would have provided capabilities such as seeing enemies through walls and scanning them for details. Additionally, her voice analysis tool was intended to be capable of mimicking an enemy's voice to open voice-activated locks or distract foes with false communications. Perfect Dark would've followed a more linear, mission-based structure as opposed to open-world gameplay. The game's parkour mechanics were meant to enable Joanna to perform actions such as wall running, scaling buildings, and sliding along scaffolding, which would have enabled her to access otherwise unreachable locations. Premise Perfect Dark is set in a near-future world experiencing environmental collapse, with megacorporations attempting to address the crisis while concealing their own secrets. Players assume the role of agent Joanna Dark, a skilled operative seeking to uncover the truth behind these events. According to an article on Xbox Wire, the game's setting involves significant ecological disasters, collectively referred to as "The Cascade," causing widespread devastation on Earth. Core Mantis, a megacorporation, creates The GEN Network and deploys it in Cairo, Egypt, successfully restoring the region's ecosystem and making it habitable once again. Following this success, other major companies also get involved, but hidden agendas are at play. This reimagined futuristic version of Cairo serves as the main setting for the game. The protagonist, Joanna Dark, is a special operative working for the megacorporation dataDyne. The character is portrayed through performance capture by actress Alix Wilton Regan, with her facial likeness based on model Elissa Bibaud. Joanna is proficient in firearms, advanced gadgets, stealth, and close-quarters combat. Her mission involves investigating and capturing Daniel Carrington, the world's most wanted criminal, who is hiding in the restored city of Cairo with a radioactive device. Despite her expertise, Joanna is not infallible, and part of the game's narrative focuses on her growth and learning from her mistakes, reflecting the aspirational nature of the game's title. Development Perfect Dark was intended to be the debut project by The Initiative, an internally formed first-party team within Xbox Game Studios founded in 2018 and headed by Darrell Gallagher. At its formation, it aimed to create and produce big-budget titles. In addition to Gallagher, the studio recruited several industry veterans, including Christian Cantamessa (writer of Red Dead Redemption) and Brian Westergaard (senior producer of 2018's God of War), as well as talents from BioWare, Naughty Dog, Respawn Entertainment, Santa Monica Studio, Blizzard Entertainment, Insomniac Games and Rockstar Games. Gallagher, who had previously worked on the 2013 Tomb Raider reboot, chose to work on a Perfect Dark reboot after Microsoft presented him with a few opportunities. Plans for the game had already been discussed before Gallagher joined The Initiative, with Xbox head Phil Spencer stating that the game is seen as an opportunity for protagonist Joanna Dark to diversify the Xbox family. According to a source, the game will run on Unreal Engine 4 and feature "various weapons, gadgets, and a camera surveillance system". Design director Drew Murray revealed that the game is being envisioned as a spy shooter, and that the studio wants the player character's physicality to play a bigger role than in traditional first-person shooters. In February 2021, Murray left The Initiative to rejoin Insomniac Games. Shortly afterwards, God of War producer Rhonda Cox joined the company as senior producer for the game. In September 2021, The Initiative announced they would be partnering with Crystal Dynamics on the game's development. Daniel Neuburger, who had previously directed several Tomb Raider games at Crystal Dynamics, was the game director, until he left The Initiative in February 2022. In the previous 12 months, a substantial number of developers had also left the company, citing a lack of creative autonomy and slow development progress as the reason for their departures. In May 2022, after Embracer Group announced that they had entered an agreement with Square Enix to acquire Crystal Dynamics, The Initiative confirmed that they would still continue to co-develop the game with the studio. In November 2022, Xbox Studios boss Matt Booty revealed that, after the departure of several senior staff members at The Initiative, the Perfect Dark team has been slowly rebuilding while dealing with the challenges of the COVID-19 pandemic and starting up a new studio. By June 2023, Perfect Dark was still in pre-production and roughly two to three years away from release. It was also revealed that Certain Affinity was involved in its development before The Initiative announced that they would co-develop the game with Crystal Dynamics. Due to multiple factors, including poor management and creative differences between both studios, progress on the game stalled. In 2022, after The Initiative partnered with Crystal Dynamics, development restarted under Unreal Engine 5 and was a bit more productive. There are plans for the game to follow an episodic format, but it could still change before release. It was also reported that The Coalition has been supporting both studios with their Unreal Engine expertise. Grant Kirkhope, who composed most of the original Perfect Dark score, expressed his desire to work on the game, but a new composer was chosen instead. In October 2024, it was reported that former Insomniac Games developer Brian Horton had left the studio, where he was acting as creative director on Marvel's Wolverine, to join The Initiative as Perfect Dark's new creative director over the summer. In July 2025, it was announced that The Initiative would be closed, and that development of the project would be stopped as part of several rounds of layoffs by Microsoft. Joanna Dark actress Alix Wilton Regan called for fans to "speak up if [they] wanna see Perfect Dark survive", while actor Elias Toufexis, who voiced Adam Jensen in the Deus Ex series, said the cancellation resulted in a loss of "thousands of dollars" that he was expecting for his role in the game. EA Japan's general manager Shaun Noguchi expressed concerns about the numerous layoffs and the cancellation of Perfect Dark and Rare's Everwild, which had been in development for roughly seven and eight years respectively, stating: "That's a decade of work, potentially a quarter of someone's entire career completely lost. Even if the final product isn't what people originally expected, I think it still deserves to ship. Something is better than nothing for both the team and for the players. But also, don't announce games when they're still half baked." Xbox co-creator Seamus Blackley echoed similar sentiments, "Think of the number of great games that had troubled development histories. All of them? Now consider how often executives cancel troubled games. Smooth development comes only when you take no risks. Greatness comes only when great risks are braved." In September 2025, it was reported that Microsoft briefly held talks with Take-Two Interactive regarding the latter taking over development funding and publishing duties in an effort to revive the game, but said talks fell through due to disputes over ownership of the game and property, which in part led to additional layoffs at Crystal Dynamics. Wilton Regan later revealed that she had recorded entire chapters of the game's universe before its cancellation, and that Crystal Dynamics suffered considerable job losses. Marketing Although work on a Perfect Dark revival was rumored in early 2018, Perfect Dark was officially announced at The Game Awards 2020 with a cinematic trailer, after development on the game had been hinted by some sources earlier that year. A gameplay trailer was revealed at the Xbox Games Showcase on June 9, 2024. After the game was cancelled, former level designer Adam McDonald confirmed that most of the game mechanics showcased in the demo were not scripted, though the gameplay systems were still in a rudimentary stage. References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Cassini%E2%80%93Huygens] | [TOKENS: 9057] |
Contents Cassini–Huygens Cassini–Huygens (/kəˈsiːni ˈhɔɪɡənz/ kə-SEE-nee HOY-gənz), commonly called Cassini, was a space-research mission by NASA, the European Space Agency (ESA), and the Italian Space Agency (ASI) to send a space probe to study the planet Saturn and its system, including its rings and natural satellites. The Flagship-class robotic spacecraft comprised both NASA's Cassini space probe and ESA's Huygens lander, which landed on Saturn's largest moon, Titan. Cassini was the fourth space probe to visit Saturn and the first to enter its orbit, where it stayed from 2004 to 2017. The two craft took their names from the astronomers Giovanni Cassini and Christiaan Huygens. Launched aboard a Titan IVB/Centaur on October 15, 1997, Cassini was active in space for nearly 20 years, spending almost 7 years in transit and 13 years orbiting Saturn, studying the planet and its system after entering orbit on July 1, 2004. The voyage to Saturn included flybys of Venus (April 1998 and July 1999), Earth (August 1999), the asteroid 2685 Masursky, and Jupiter (December 2000). The mission ended on September 15, 2017, when Cassini's trajectory took it into Saturn's upper atmosphere and it burned up in order to prevent any risk of contaminating Saturn's moons, which might have offered habitable environments to stowaway terrestrial microbes on the spacecraft. The mission was successful beyond expectations – NASA's Planetary Science Division Director, Jim Green, described Cassini–Huygens as a "mission of firsts" that revolutionized human understanding of the Saturn system, including its moons and rings, and our understanding of where life might be found in the Solar System. Overview Scientists and individuals from 27 countries made up the joint team responsible for designing, building, flying and collecting data from the Cassini orbiter and the Huygens probe. Cassini's planners originally scheduled a mission of four years, from June 2004 to May 2008. The mission was extended for another two years until September 2010, branded the Cassini Equinox Mission. The mission was extended a second and final time with the Cassini Solstice Mission, lasting another seven years until September 15, 2017, on which date Cassini was de-orbited to burn up in Saturn's upper atmosphere. The Huygens module traveled with Cassini until its separation from the probe on December 25, 2004; Huygens landed by parachute. on Titan on January 14, 2005. The separation was facilitated by the SED (Spin/Eject device), which provided a relative separation speed of 0.35 metres per second (1.1 ft/s) and a spin rate of 7.5 rpm. It returned data to Earth for around 90 minutes, using the orbiter as a relay. This was the first landing ever accomplished in the outer Solar System and the first landing on a moon other than Earth's Moon. At the end of its mission, the Cassini spacecraft executed its "Grand Finale": a number of risky passes through the gaps between Saturn and its inner rings. This phase aimed to maximize Cassini's scientific outcome before the spacecraft was intentionally destroyed to prevent potential contamination of Saturn's moons if Cassini were to unintentionally crash into them when maneuvering the probe was no longer possible due to power loss or other communication issues at the end of its operational lifespan. Cassini's atmospheric entry on Saturn ended the mission, but analysis of the returned data will continue for many years. NASA's Jet Propulsion Laboratory, where the orbiter was assembled, managed the mission. The European Space Research and Technology Centre developed Huygens. The centre's prime contractor, Aérospatiale of France (which became part of Thales Alenia Space in 2005), assembled the probe with equipment and instruments supplied by many European countries (including Huygens's batteries and two scientific instruments from the United States). The Italian Space Agency (ASI) provided the Cassini orbiter's high-gain radio antenna, with the incorporation of a low-gain antenna (to ensure telecommunications with the Earth for the entire duration of the mission), a compact and lightweight radar, which also used the high-gain antenna and served as a synthetic-aperture radar, a radar altimeter, a radiometer, the radio science subsystem (RSS), and the visible-channel portion VIMS-V of VIMS spectrometer. NASA provided the VIMS infrared counterpart, as well as the Main Electronic Assembly, which included electronic sub-assemblies provided by CNES of France. On April 16, 2008, NASA announced a two-year extension of the funding for ground operations of this mission, at which point it was renamed the Cassini Equinox Mission. It was extended again in February 2010 as the Cassini Solstice Mission. Naming The mission consisted of two main elements: the ASI/NASA Cassini orbiter, named for the Italian astronomer Giovanni Domenico Cassini, the discoverer of Saturn's ring divisions and four of its satellites; and the ESA-developed Huygens probe, named for the Dutch astronomer, mathematician and physicist Christiaan Huygens, discoverer of Titan. The mission was commonly called Saturn Orbiter Titan Probe (SOTP) during gestation, both as a Mariner Mark II mission and generically. Cassini–Huygens was a Flagship-class mission to the outer planets. The other planetary flagships include Galileo, Voyager, and Viking. Objectives Cassini had several objectives, including: Cassini–Huygens was launched on October 15, 1997, from Cape Canaveral Air Force Station's Space Launch Complex 40 using a U.S. Air Force Titan IVB/Centaur rocket. The complete launcher was made up of a two-stage Titan IV booster rocket, two strap-on solid rocket engines, the Centaur upper stage, and a payload enclosure, or fairing. The total cost of this scientific exploration mission was about US$3.26 billion, including $1.4 billion for pre-launch development, $704 million for mission operations, $54 million for tracking and $422 million for the launch vehicle. The United States contributed $2.6 billion (80%), the ESA $500 million (15%), and the ASI $160 million (5%). However, these figures are from the press kit which was prepared in October 2000. They do not include inflation over the course of a very long mission, nor do they include the cost of the extended missions. The primary mission for Cassini was completed on July 30, 2008. The mission was extended to June 2010 (Cassini Equinox Mission). This studied the Saturn system in detail during the planet's equinox, which happened in August 2009. On February 3, 2010, NASA announced another extension for Cassini, lasting 61⁄2 years until 2017, ending at the time of summer solstice in Saturn's northern hemisphere (Cassini Solstice Mission). The extension enabled another 155 revolutions around the planet, 54 flybys of Titan and 11 flybys of Enceladus. In 2017, an encounter with Titan changed its orbit in such a way that, at closest approach to Saturn, it was only 3,000 km (1,900 mi) above the planet's cloudtops, below the inner edge of the D ring. This sequence of "proximal orbits" ended when its final encounter with Titan sent the probe into Saturn's atmosphere to be destroyed. Selected destinations (ordered largest to smallest but not to scale) History Cassini–Huygens's origins date to 1982, when the European Science Foundation and the American National Academy of Sciences formed a working group to investigate future cooperative missions. Two European scientists suggested a paired Saturn Orbiter and Titan Probe as a possible joint mission. In 1983, NASA's Solar System Exploration Committee recommended the same Orbiter and Probe pair as a core NASA project. NASA and the European Space Agency (ESA) performed a joint study of the potential mission from 1984 to 1985. ESA continued with its own study in 1986, while the American astronaut Sally Ride, in her influential 1987 report NASA Leadership and America's Future in Space, also examined and approved of the Cassini mission. While Ride's report described the Saturn orbiter and probe as a NASA solo mission, in 1988 the Associate Administrator for Space Science and Applications of NASA, Len Fisk, returned to the idea of a joint NASA and ESA mission. He wrote to his counterpart at ESA, Roger Bonnet, strongly suggesting that ESA choose the Cassini mission from the three candidates at hand and promising that NASA would commit to the mission as soon as ESA did. At the time, NASA was becoming more sensitive to the strain that had developed between the American and European space programs as a result of European perceptions that NASA had not treated it like an equal during previous collaborations. NASA officials and advisers involved in promoting and planning Cassini–Huygens attempted to correct this trend by stressing their desire to evenly share any scientific and technology benefits resulting from the mission. In part, this newfound spirit of cooperation with Europe was driven by a sense of competition with the Soviet Union, which had begun to cooperate more closely with Europe as ESA drew further away from NASA. Late in 1988, ESA chose Cassini–Huygens as its next major mission and the following year the program received major funding in the US. The collaboration not only improved relations between the two space programs but also helped Cassini–Huygens survive congressional budget cuts in the United States. Cassini–Huygens came under fire politically in both 1992 and 1994, but NASA successfully persuaded the United States Congress that it would be unwise to halt the project after ESA had already poured funds into development because frustration on broken space exploration promises might spill over into other areas of foreign relations. The project proceeded politically smoothly after 1994, although citizens' groups concerned about the potential environmental impact a launch failure might have (because of its plutonium power source) attempted to derail it through protests and lawsuits until and past its 1997 launch. Spacecraft design The spacecraft was planned to be the second three-axis stabilized, RTG-powered Mariner Mark II, a class of spacecraft developed for missions beyond the orbit of Mars, after the Comet Rendezvous Asteroid Flyby (CRAF) mission, but budget cuts and project rescopings forced NASA to terminate CRAF development to save Cassini. As a result, Cassini became more specialized. The Mariner Mark II series was cancelled. The combined orbiter and probe was at the time the third-largest uncrewed interplanetary spacecraft ever successfully launched, behind the Phobos 1 and 2 Mars probes, as well as being among the most complex; NASA's Europa Clipper became the new third-largest probe upon its launch in 2024. The orbiter had a mass of 2,150 kg (4,740 lb), the probe 350 kg (770 lb) including 30 kg (66 lb) of probe support equipment left on the orbiter. With the launch vehicle adapter and 3,132 kg (6,905 lb) of propellants at launch, the spacecraft had a mass of 5,600 kg (12,300 lb). The Cassini spacecraft was 6.8 meters (22 ft) high and 4 meters (13 ft) wide. Its bus was a dodecagonal prism atop a conical frustum connecting it to a cylinder containing the propellant tanks, to which the RTGs and Huygens were attached. Spacecraft complexity was increased by its trajectory (flight path) to Saturn, and by the ambitious science at its destination. Cassini had 1,630 interconnected electronic components, 22,000 wire connections, and 14 kilometers (8.7 mi) of cabling. The core control computer CPU was a redundant system using the MIL-STD-1750A instruction set architecture. The main propulsion system consisted of one prime and one backup R-4D bipropellant rocket engine. The thrust of each engine was 490 N (110 lbf) and the total spacecraft delta-v was 2,352 m/s (5,260 mph). Smaller monopropellant rockets provided attitude control. Cassini was powered by 32.7 kg (72 lb) of nuclear fuel, mainly plutonium dioxide (containing 28.3 kg (62 lb) of pure plutonium). The heat from the material's radioactive decay was turned into electricity. Huygens was supported by Cassini during cruise, but used chemical batteries when independent. The probe contained a DVD with more than 616,400 signatures from citizens in 81 countries, collected in a public campaign. Until September 2017 the Cassini probe continued orbiting Saturn at a distance of between 8.2 and 10.2 astronomical units (1.23×109 and 1.53×109 km; 760,000,000 and 950,000,000 mi) from the Earth. It took 68 to 84 minutes for radio signals to travel from Earth to the spacecraft, and vice versa. Thus ground controllers could not give "real-time" instructions for daily operations or for unexpected events. Even if response were immediate, more than two hours would have passed between the occurrence of a problem and the reception of the engineers' response by the satellite. Instruments Instruments: Cassini's instrumentation consisted of: a synthetic aperture radar mapper, a charge-coupled device imaging system, a visible/infrared mapping spectrometer, a composite infrared spectrometer, a cosmic dust analyzer, a radio and plasma wave experiment, a plasma spectrometer, an ultraviolet imaging spectrograph, a magnetospheric imaging instrument, a magnetometer and an ion/neutral mass spectrometer. Telemetry from the communications antenna and other special transmitters (an S-band transmitter and a dual-frequency Ka-band system) was also used to make observations of the atmospheres of Titan and Saturn and to measure the gravity fields of the planet and its satellites. The HSP channel is designed to observe starlight that passes through Saturn's rings (known as stellar occultations) in order to understand the structure and optical depth of the rings. Stellar occultation data from both the HSP and FUV channels confirmed the existence of water vapor plumes at the south pole of Enceladus, as well as characterized the composition of the plumes. Plutonium power source Because of Saturn's distance from the Sun, solar arrays were not feasible as power sources for this space probe. To generate enough power, such arrays would have been too large and too heavy. Instead, the Cassini orbiter was powered by three GPHS-RTG radioisotope thermoelectric generators, which use heat from the decay of about 33 kg (73 lb) of plutonium-238 (in the form of plutonium dioxide) to generate direct current electricity via thermoelectrics. The RTGs on the Cassini mission have the same design as those used on the New Horizons, Galileo, and Ulysses space probes, and they were designed to have very long operational lifetimes. At the end of the nominal 11-year Cassini mission, they were still able to produce 600 to 700 watts of electrical power. (Leftover hardware from the Cassini RTG Program was modified and used to power the New Horizons mission to Pluto and the Kuiper belt, which was designed and launched later.) Electric power distribution was accomplished by 192 solid-state power switches, which also functioned as circuit breakers in the event of an overload condition. The switches used MOSFETs that featured better efficiency and a longer lifetime as compared to conventional switches, while at the same time eliminating transients. However, these solid-state circuit breakers were prone to erroneous tripping (presumably from cosmic rays), requiring them to reset and causing losses in experimental data. To gain momentum while already in flight, the trajectory of the Cassini mission included several gravitational slingshot maneuvers: two fly-by passes of Venus, one more of the Earth, and then one of the planet Jupiter. The terrestrial flyby was the final instance when the probe posed any conceivable danger to human beings. The maneuver was successful, with Cassini passing by 1,171 km (728 mi) above the Earth on August 18, 1999. Had there been any malfunction causing the probe to collide with the Earth, NASA's complete environmental impact study estimated that, in the worst case (with an acute angle of entry in which Cassini would gradually burn up), a significant fraction of the 33 kg of nuclear fuel inside the RTGs would have been dispersed into the Earth's atmosphere so that up to five billion people (i.e. almost the entire terrestrial population) could have been exposed, causing up to an estimated 5,000 additional cancer deaths over the subsequent decades (0.0005 per cent, i.e. a fraction 0.000005, of a billion cancer deaths expected anyway from other causes; the product is incorrectly calculated elsewhere as 500,000 deaths). However, the chance of this happening were estimated to be less than one in one million, i.e. a chance of one person dying (assuming 5,000 deaths) as less than 1 in 200. NASA's risk analysis to use plutonium was publicly criticized by Michio Kaku on the grounds that casualties, property damage, and lawsuits resulting from a possible accident, as well as the potential use of other energy sources, such as solar and fuel cells, were underestimated. Telemetry The Cassini spacecraft was capable of transmitting in several different telemetry formats. The telemetry subsystem is perhaps the most important subsystem, because without it there could be no data return. The telemetry was developed from the ground up, due to the spacecraft using a more modern set of computers than previous missions. Therefore, Cassini was the first spacecraft to adopt mini-packets to reduce the complexity of the Telemetry Dictionary, and the software development process led to the creation of a Telemetry Manager for the mission. There were around 1088 channels (in 67 mini-packets) assembled in the Cassini Telemetry Dictionary. Out of these 67 lower complexity mini-packets, 6 mini-packets contained the subsystem covariance and Kalman gain elements (161 measurements), not used during normal mission operations. This left 947 measurements in 61 mini-packets. A total of seven telemetry maps corresponding to 7 AACS telemetry modes were constructed. These modes are: (1) Record; (2) Nominal Cruise; (3) Medium Slow Cruise; (4) Slow Cruise; (5) Orbital Ops; (6) Av; (7) ATE (Attitude Estimator) Calibration. These 7 maps cover all spacecraft telemetry modes. Huygens probe The Huygens probe, supplied by the European Space Agency (ESA) and named after the 17th century Dutch astronomer who first discovered Titan, Christiaan Huygens, scrutinized the clouds, atmosphere, and surface of Saturn's moon Titan in its descent on January 15, 2005. It was designed to enter and brake in Titan's atmosphere and parachute a fully instrumented robotic laboratory down to the surface. The probe system consisted of the probe itself which descended to Titan, and the probe support equipment (PSE) which remained attached to the orbiting spacecraft. The PSE includes electronics that track the probe, recover the data gathered during its descent, and process and deliver the data to the orbiter that transmits it to Earth. The core control computer CPU was a redundant MIL-STD-1750A control system. The data were transmitted by a radio link between Huygens and Cassini provided by Probe Data Relay Subsystem (PDRS). As the probe's mission could not be telecommanded from Earth because of the great distance, it was automatically managed by the Command Data Management Subsystem (CDMS). The PDRS and CDMS were provided by the Italian Space Agency (ASI). After Cassini's launch, it was discovered that data sent from the Huygens probe to Cassini orbiter (and then re-transmitted to Earth) would be largely unreadable. The cause was that the bandwidth of signal processing electronics was too narrow and the anticipated Doppler shift between the lander and the mother craft would put the signals out of the system's range. Thus, Cassini's receiver would be unable to receive the data from Huygens during its descent to Titan. A work-around was found to recover the mission. The trajectory of Cassini was altered to reduce the line of sight velocity and therefore the doppler shift. Cassini's subsequent trajectory was identical to the previously planned one, although the change replaced two orbits prior to the Huygens mission with three, shorter orbits. Selected events and discoveries The Cassini space probe performed two gravitational-assist flybys of Venus on April 26, 1998, and June 24, 1999. These flybys provided the space probe with enough momentum to travel all the way out to the asteroid belt, while the Sun's gravity pulled the space probe back into the inner Solar System. On August 18, 1999, at 03:28 UTC, the craft made a gravitational-assist flyby of the Earth. One hour and 20 minutes before closest approach, Cassini made its closest approach to the Earth's Moon at 377,000 kilometers, and it took a series of calibration photos. On January 23, 2000, Cassini performed a flyby of the asteroid 2685 Masursky at around 10:00 UTC. It took photos in the period five to seven hours before the flyby at a distance of 1.6×10^6 km (0.99×10^6 mi) and a diameter of 15 to 20 km (9.3 to 12.4 mi) was estimated for the asteroid. Cassini made its closest approach to Jupiter on December 30, 2000, at 9.7 million kilometers, and made many scientific measurements. About 26,000 images of Jupiter, its faint rings, and its moons were taken during the six-month flyby. It produced the most detailed global color portrait of the planet yet (see image at right), in which the smallest visible features are approximately 60 km (37 mi) across. A major finding of the flyby, announced on March 6, 2003, was of Jupiter's atmospheric circulation. Dark "belts" alternate with light "zones" in the atmosphere, and scientists had long considered the zones, with their pale clouds, to be areas of upwelling air, partly because many clouds on Earth form where air is rising. But analysis of Cassini imagery showed that individual storm cells of upwelling bright-white clouds, too small to see from Earth, pop up almost without exception in the dark belts. According to Anthony Del Genio of NASA's Goddard Institute for Space Studies, "the belts must be the areas of net-rising atmospheric motion on Jupiter, [so] the net motion in the zones has to be sinking". Other atmospheric observations included a swirling dark oval of high atmospheric haze, about the size of the Great Red Spot, near Jupiter's north pole. Infrared imagery revealed aspects of circulation near the poles, with bands of globe-encircling winds, with adjacent bands moving in opposite directions. The same announcement also discussed the nature of Jupiter's rings. Light scattering by particles in the rings showed the particles were irregularly shaped (rather than spherical) and likely originate as ejecta from micrometeorite impacts on Jupiter's moons, probably Metis and Adrastea. On October 10, 2003, the mission's science team announced the results of tests of Albert Einstein's general theory of relativity, performed by using radio waves transmitted from the Cassini space probe. The radio scientists measured a frequency shift in the radio waves to and from the spacecraft, as they passed close to the Sun. According to the general theory of relativity, a massive object like the Sun causes space-time to curve, causing a beam of radiowaves travelling out of its gravitational well to decrease in frequency and radiowaves travelling into the gravitational well to increase in frequency, referred to as gravitational redshift / blueshift. Although some measurable deviations from the values calculated using the general theory of relativity are predicted by some unusual cosmological models, no such deviations were found by this experiment. Previous tests using radiowaves transmitted by the Viking and Voyager space probes were in agreement with the calculated values from general relativity to within an accuracy of one part in one thousand. The more refined measurements from the Cassini space probe experiment improved this accuracy to about one part in 51,000.[a] The data firmly support Einstein's general theory of relativity. In total, the Cassini mission discovered seven new moons orbiting Saturn. Using images taken by Cassini, researchers discovered Methone, Pallene and Polydeuces in 2004, although later analysis revealed that Voyager 2 had photographed Pallene in its 1981 flyby of the ringed planet. On May 1, 2005, a new moon was discovered by Cassini in the Keeler gap. It was given the designation S/2005 S 1 before being named Daphnis. A fifth new moon was discovered by Cassini on May 30, 2007, and was provisionally labeled S/2007 S 4. It is now known as Anthe. A press release on February 3, 2009, showed a sixth new moon found by Cassini. The moon is approximately 500 m (0.3 mi) in diameter within the G-ring of the ring system of Saturn, and is now named Aegaeon (formerly S/2008 S 1). A press release on November 2, 2009, mentions the seventh new moon found by Cassini on July 26, 2009. It is presently labeled S/2009 S 1 and is approximately 300 m (980 ft) in diameter in the B-ring system. On April 14, 2014, NASA scientists reported the possible beginning of a new moon in Saturn's A Ring. On June 11, 2004, Cassini flew by the moon Phoebe. This was the first opportunity for close-up studies of this moon (Voyager 2 performed a distant flyby in 1981 but returned no detailed images). It also was Cassini's only possible flyby for Phoebe due to the mechanics of the available orbits around Saturn. The first close-up images were received on June 12, 2004, and mission scientists immediately realized that the surface of Phoebe looks different from asteroids visited by spacecraft. Parts of the heavily cratered surface look very bright in those pictures, and it is currently believed that a large amount of water ice exists under its immediate surface. In an announcement on June 28, 2004, Cassini program scientists described the measurement of the rotational period of Saturn. Because there are no fixed features on the surface that can be used to obtain this period, the repetition of radio emissions was used. This new data agreed with the latest values measured from Earth, and constituted a puzzle to the scientists. It turns out that the radio rotational period had changed since it was first measured in 1980 by Voyager 1, and it was now 6 minutes longer. This, however, does not indicate a change in the overall spin of the planet. It is thought to be due to variations in the upper atmosphere and ionosphere at the latitudes which are magnetically connected to the radio source region. In 2019 NASA announced Saturn's rotational period as 10 hours, 33 minutes, 38 seconds, calculated using Saturnian ring seismology. Vibrations from Saturn's interior cause oscillations in its gravitational field. This energy is absorbed by ring particles in specific locations, where it accumulates until it is released in a wave. Scientists used data from more than 20 of these waves to construct a family of models of Saturn's interior, providing basis for calculating its rotational period. On July 1, 2004, the spacecraft flew through the gap between the F and G rings and achieved orbit, after a seven-year voyage. It was the first spacecraft to orbit Saturn. The Saturn Orbital Insertion (SOI) maneuver performed by Cassini was complex, requiring the craft to orient its High-Gain Antenna away from Earth and along its flight path, to shield its instruments from particles in Saturn's rings. Once the craft crossed the ring plane, it had to rotate again to point its engine along its flight path, and then the engine fired to decelerate the craft by 622 m/s to allow Saturn to capture it. Cassini was captured by Saturn's gravity at around 8:54 pm Pacific Daylight Time on June 30, 2004. During the maneuver Cassini passed within 20,000 km (12,000 mi) of Saturn's cloud tops. When Cassini was in Saturnian orbit, departure from the Saturn system was evaluated in 2008 during end of mission planning.[clarification needed] Cassini had its first flyby of Saturn's largest moon, Titan, on July 2, 2004, a day after orbit insertion, when it approached to within 339,000 km (211,000 mi) of Titan. Images taken through special filters (able to see through the moon's global haze) showed south polar clouds thought to be composed of methane and surface features with widely differing brightness. On October 27, 2004, the spacecraft executed the first of the 45 planned close flybys of Titan when it passed a mere 1,200 km (750 mi) above the moon. Almost four gigabits of data were collected and transmitted to Earth, including the first radar images of the moon's haze-enshrouded surface. It revealed the surface of Titan (at least the area covered by radar) to be relatively level, with topography reaching no more than about 50 m (160 ft) in altitude. The flyby provided a remarkable increase in imaging resolution over previous coverage. Images with up to 100 times better resolution were taken and are typical of resolutions planned for subsequent Titan flybys. Cassini collected pictures of Titan and the lakes of methane were similar to the lakes of water on Earth. Cassini released the Huygens probe on December 25, 2004, by means of a spring and spiral rails intended to rotate the probe for greater stability. It entered the atmosphere of Titan on January 14, 2005, and after a two-and-a-half-hour descent landed on solid ground. Although Cassini successfully relayed 350 of the pictures that it received from Huygens of its descent and landing site, a malfunction in one of the communications channels resulted in the loss of a further 350 pictures. During the first two close flybys of the moon Enceladus in 2005, Cassini discovered a deflection in the local magnetic field that is characteristic for the existence of a thin but significant atmosphere. Other measurements obtained at that time point to ionized water vapor as its main constituent. Cassini also observed water ice geysers erupting from the south pole of Enceladus, which gives more credibility to the idea that Enceladus is supplying the particles of Saturn's E ring. Mission scientists began to suspect that there may be pockets of liquid water near the surface of the moon that fuel the eruptions. On March 12, 2008, Cassini made a close fly-by of Enceladus, passing within 50 km of the moon's surface. The spacecraft passed through the plumes extending from its southern geysers, detecting water, carbon dioxide and various hydrocarbons with its mass spectrometer, while also mapping surface features that are at much higher temperature than their surroundings with the infrared spectrometer. Cassini was unable to collect data with its cosmic dust analyzer due to an unknown software malfunction. On November 21, 2009, Cassini made its eighth flyby of Enceladus, this time with a different geometry, approaching within 1,600 km (990 mi) of the surface. The Composite Infrared Spectrograph (CIRS) instrument produced a map of thermal emissions from the Baghdad Sulcus 'tiger stripe'. The data returned helped create a detailed and high resolution mosaic image of the southern part of the moon's Saturn-facing hemisphere. On April 3, 2014, nearly ten years after Cassini entered Saturn's orbit, NASA reported evidence of a large salty internal ocean of liquid water in Enceladus. The presence of an internal salty ocean in contact with the moon's rocky core, places Enceladus "among the most likely places in the Solar System to host alien microbial life". On June 30, 2014, NASA celebrated ten years of Cassini exploring Saturn and its moons, highlighting the discovery of water activity on Enceladus among other findings. In September 2015, NASA announced that gravitational and imaging data from Cassini were used to analyze the librations of Enceladus's orbit and determined that the moon's surface is not rigidly joined to its core, concluding that the underground ocean must therefore be global in extent. On October 28, 2015, Cassini performed a close flyby of Enceladus, coming within 49 km (30 mi) of the surface, and passing through the icy plume above the south pole. On December 14, 2023, astronomers reported the first time discovery, in the plumes of Enceladus, of hydrogen cyanide, a possible chemical essential for life as we know it, as well as other organic molecules, some of which are yet to be better identified and understood. According to the researchers, "these [newly discovered] compounds could potentially support extant microbial communities or drive complex organic synthesis leading to the origin of life". In May 2005, Cassini began a series of radio occultation experiments, to measure the size-distribution of particles in Saturn's rings, and measure the atmosphere of Saturn itself. For over four months, the craft completed orbits designed for this purpose. During these experiments, it flew behind the ring plane of Saturn, as seen from Earth, and transmitted radio waves through the particles. The radio signals received on Earth were analyzed, for frequency, phase, and power shift of the signal to determine the structure of the rings. In images captured September 5, 2005, Cassini detected spokes in Saturn's rings, previously seen only by the visual observer Stephen James O'Meara in 1977 and then confirmed by the Voyager space probes in the early 1980s. Radar images obtained on July 21, 2006, appear to show lakes of liquid hydrocarbon (such as methane and ethane) in Titan's northern latitudes. This is the first discovery of currently existing lakes anywhere besides on Earth. The lakes range in size from one to one-hundred kilometers across. On March 13, 2007, the Jet Propulsion Laboratory announced that it had found strong evidence of seas of methane and ethane in the northern hemisphere of Titan. At least one of these is larger than any of the Great Lakes in North America. In November 2006, scientists discovered a storm at the south pole of Saturn with a distinct eyewall. This is characteristic of a hurricane on Earth and had never been seen on another planet before. Unlike a terrestrial hurricane, the storm appears to be stationary at the pole. The storm is 8,000 km (5,000 mi) across, and 70 km (43 mi) high, with winds blowing at 560 km/h (350 mph). On September 10, 2007, Cassini completed its flyby of the strange, two-toned, walnut-shaped moon, Iapetus. Images were taken from 1,600 km (1,000 mi) above the surface. As it was sending the images back to Earth, it was hit by a cosmic ray that forced it to temporarily enter safe mode. All of the data from the flyby was recovered. On April 15, 2008, Cassini received funding for a 27-month extended mission. It consisted of 60 more orbits of Saturn, with 21 more close Titan flybys, seven of Enceladus, six of Mimas, eight of Tethys, and one targeted flyby each of Dione, Rhea, and Helene. The extended mission began on July 1, 2008, and was renamed the Cassini Equinox Mission as the mission coincided with Saturn's equinox. A proposal was submitted to NASA for a second mission extension (September 2010 – May 2017), provisionally named the extended-extended mission or XXM. This ($60M pa) was approved in February 2010 and renamed the Cassini Solstice Mission. It included Cassini orbiting Saturn 155 more times, conducting 54 additional flybys of Titan and 11 more of Enceladus. On October 25, 2012, Cassini witnessed the aftermath of the massive Great White Spot storm that recurs roughly every 30 years on Saturn. Data from the composite infrared spectrometer (CIRS) instrument indicated a powerful discharge from the storm that caused a temperature spike in the stratosphere of Saturn 83 K (83 °C; 149 °F) above normal. Simultaneously, a huge increase in ethylene gas was detected by NASA researchers at Goddard Research Center in Greenbelt, Maryland. Ethylene is a colorless gas that is highly uncommon on Saturn and is produced both naturally and through man-made sources on Earth. The storm that produced this discharge was first observed by the spacecraft on December 5, 2010, in Saturn's northern hemisphere. The storm is the first of its kind to be observed by a spacecraft in orbit around Saturn as well as the first to be observed at thermal infrared wavelengths, allowing scientists to observe the temperature of Saturn's atmosphere and track phenomena that are invisible to the naked eye. The spike of ethylene gas that was produced by the storm reached levels that were 100 times more than those thought possible for Saturn. Scientists have also determined that the storm witnessed was the largest, hottest stratospheric vortex ever detected in the Solar System, initially being larger than Jupiter's Great Red Spot. On December 21, 2012, Cassini observed a transit of Venus across the Sun. The VIMS instrument analyzed sunlight passing through the Venusian atmosphere. VIMS previously observed the transit of exoplanet HD 189733 b. On July 19, 2013, the probe was pointed towards Earth to capture an image of the Earth and the Moon, as part of a natural light, multi-image portrait of the entire Saturn system. The event was unique as it was the first time NASA informed the public that a long-distance photo was being taken in advance. The imaging team said they wanted people to smile and wave to the skies, with Cassini scientist Carolyn Porco describing the moment as a chance to "celebrate life on the Pale Blue Dot". On February 10, 2015, the Cassini spacecraft visited Rhea more closely, coming within 47,000 km (29,000 mi). The spacecraft observed the moon with its cameras producing some of the highest resolution color images yet of Rhea. Cassini performed its latest flyby of Saturn's moon Hyperion on May 31, 2015, at a distance of about 34,000 km (21,000 mi). Cassini performed its last flyby of Saturn's moon Dione on August 17, 2015, at a distance of about 475 km (295 mi). A previous flyby was performed on June 16. Between 2012 and 2016, the persistent hexagonal cloud pattern at Saturn's north pole changed from a mostly blue color to more of a golden color. One theory for this is a seasonal change: extended exposure to sunlight may be creating haze as the pole swivels toward the Sun. It was previously noted that there was less blue color overall on Saturn between 2004 and 2008. Grand Finale and destruction Cassini's end, named the Grand Finale, involved a series of close Saturn passes, approaching within the rings, then an entry into Saturn's atmosphere on September 15, 2017, to destroy the spacecraft. This method was chosen to ensure protection and prevent biological contamination to any of the moons of Saturn thought to offer potential habitability. In 2008, a number of options were evaluated to achieve this goal, each with varying funding, scientific, and technical challenges. A short period Saturn impact for an end of mission was rated "excellent" with the reasons "D-ring option satisfies unachieved AO goals;[definition needed] cheap and easily achievable" while collision with an icy moon was rated "good" for being "cheap and achievable anywhere/time". There were problems in 2013–14 about NASA receiving U.S. government funding for the Grand Finale. The two phases of the Grand Finale ended up being the equivalent of having two separate Discovery-class missions in that the Grand Finale was completely different from the main Cassini regular mission. The U.S. government in late 2014 approved the Grand Finale at the cost of $200 million. This was far cheaper than building two new probes in separate Discovery-class missions. On November 29, 2016, the spacecraft performed a Titan flyby that took it to the gateway of F-ring orbits: This was the start of the Grand Finale phase culminating in its impact with the planet. A final Titan flyby on April 22, 2017, changed the orbit again to fly through the gap between Saturn and its inner ring days later on April 26. Cassini passed about 3,100 km (1,900 mi) above Saturn's cloud layer and 320 km (200 mi) from the visible edge of the inner ring; it successfully took images of Saturn's atmosphere and began returning data the next day. After a further 22 orbits through the gap, the mission was ended with a dive into Saturn's atmosphere on September 15; signal was lost at 11:55:46 UTC on September 15, 2017, 30 seconds later than predicted. It is estimated that the spacecraft burned up about 45 seconds after the last transmission. In September 2018, NASA won an Emmy Award for Outstanding Original Interactive Program for its presentation of the Cassini mission's Grand Finale at Saturn. In December 2018, Netflix aired "NASA's Cassini Mission" on their series 7 Days Out documenting the final days of work on the Cassini mission before the spacecraft crashed into Saturn to complete its Grand Finale. In January 2019, new research using data collected during Cassini's Grand Finale phase was published: The spacecraft operation was organized around a series of missions. Each was structured according to a certain amount of funding, goals, etc. At least 260 scientists from 17 countries have worked on the Cassini–Huygens mission; in addition thousands of people overall worked to design, manufacture, and launch the mission. Glossary See also Notes References External links Official websites Media and telecommunications |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Containment] | [TOKENS: 3923] |
Contents Containment Containment was a geopolitical strategic foreign policy pursued by the United States during the Cold War to prevent the spread of communism after the end of World War II. The name was loosely related to the term cordon sanitaire, which was containment of the Soviet Union in the interwar period. Containment represented a middle-ground position between détente (relaxation of relations) and rollback (actively replacing a regime). The basis of the doctrine was articulated in a 1946 cable by U.S. diplomat George F. Kennan during the post-World War II term of U.S. President Harry S. Truman. As a description of U.S. foreign policy, the word originated in a report Kennan submitted to US Defense Secretary James Forrestal in 1947, which was later used in a Foreign Affairs article. In a broader context, the term is employed to denote a strategy designed to limit or hinder an opponent's capacity for international power projection. China used this term to characterize the United States' efforts to impede its global ascent. Earlier uses of term Both Americans and Europeans were aware of significant historical antecedents. In the 1850s, anti-slavery forces in the United States developed a free soil strategy of containment to stop the expansion of slavery until it later collapsed. Historian James Oakes explains the strategy: The Federal government would surround the south with free states, free territories, and free waters, building what they called a 'cordon of freedom' around slavery, hemming it in until the system's own internal weaknesses forced the slave states one by one to abandon slavery. Between 1873 and 1877, Germany repeatedly intervened in the internal affairs of France's neighbors. In Belgium, Spain, and Italy, Chancellor Otto von Bismarck exerted strong and sustained political pressure to support the election or appointment of liberal, anticlerical governments. That was part of an integrated strategy to promote republicanism in France by strategically and ideologically isolating the clerical-monarchist regime of President Patrice de MacMahon. It was hoped that by surrounding France with a number of liberal states, French Republicans could defeat MacMahon and his reactionary supporters. The modern concept of containment provides a useful model for understanding the dynamics of this policy. After the 1917 October Revolution in Russia, there were calls by Western leaders to isolate the Bolshevik government, which seemed intent on promoting worldwide revolution. In March 1919, French Premier Georges Clemenceau called for a cordon sanitaire, a ring of non-communist states, to isolate Soviet Russia. Translating that phrase, US President Woodrow Wilson called for a "quarantine." The World War I allies launched an incursion into Russia, as after the Bolshevik Revolution, Vladimir Lenin withdrew the country from the First World War, allowing Germany to reallocate troops to face the Allied forces on the Western Front. Concurrently, President Wilson became increasingly aware of the human rights violations perpetuated by the new Russian Soviet Federative Socialist Republic, and opposed the new regime's militant atheism and advocacy of a command economy. He also was concerned that Marxism–Leninism would spread to the remainder of the Western world, and intended his landmark Fourteen Points partially to provide liberal democracy as an alternative worldwide ideology to Communism. Despite reservations, the United States, as a result of the fear of Japanese expansion into Russian-held territory and their support for the Allied-aligned Czech Legion, sent a small number of troops to Northern Russia and Siberia. The United States also provided indirect aid such as food and supplies to the White Army. The incursion was unpopular at home and lacked a cohesive strategy, leading the allies to ultimately withdraw from Russia. The U.S. initially refused to recognize the Soviet Union, but President Franklin D. Roosevelt reversed the policy in 1933 in the hope to expand American export markets. The Munich Agreement of 1938 was a failed attempt to contain Nazi expansion in Europe. The U.S. tried to contain Japanese expansion in Asia from 1937 to 1941, and Japan reacted with its attack on Pearl Harbor. After Germany invaded the Soviet Union in 1941 during World War II, the U.S. and the Soviet Union found themselves allied against Germany and used rollback to defeat the Axis powers: Germany, Italy, and Japan. Origin (1944–1947) Key State Department personnel grew increasingly frustrated with and suspicious of the Soviets as the war drew to a close. Averell Harriman, U.S. Ambassador in Moscow, once a "confirmed optimist" regarding U.S.–Soviet relations, was disillusioned by what he saw as the Soviet betrayal of the 1944 Warsaw Uprising as well as by violations of the February 1945 Yalta Agreement concerning Poland. Harriman would later have a significant influence in forming Truman's views on the Soviet Union. In February 1946, the U.S. State Department asked George F. Kennan, then at the U.S. Embassy in Moscow, why the Russians opposed the creation of the World Bank and the International Monetary Fund. He responded with a wide-ranging analysis of Russian policy now called the Long Telegram: Soviet power, unlike that of Hitlerite Germany, is neither schematic nor adventuristic. It does not work with fixed plans. It does not take unnecessary risks. Impervious to the logic of reason, and it is highly sensitive to the logic of force. For this reason, it can easily withdraw—and usually does when strong resistance is encountered at any point. Kennan's cable was hailed in the State Department as "the appreciation of the situation that had long been needed." Kennan himself attributed the enthusiastic reception to timing: "Six months earlier the message would probably have been received in the State Department with raised eyebrows and lips pursed in disapproval. Six months later, it would probably have sounded redundant." Clark Clifford and George Elsey produced a report elaborating on the Long Telegram and proposing concrete policy recommendations based on its analysis. This report, which recommended "restraining and confining" Soviet influence, was presented to Truman on September 24, 1946. In January 1947, Kennan drafted an essay entitled "The Sources of Soviet Conduct." Navy Secretary James Forrestal gave permission for the report to be published in the journal Foreign Affairs under the pseudonym "X." Biographer Douglas Brinkley has dubbed Forrestal "godfather of containment" on account of his work in distributing Kennan's writing. The use of the word "containment" originates from this so-called "X Article": "In these circumstances, it is clear that the main element of any United States policy toward the Soviet Union must be that of long-term, patient but firm and vigilant containment of Russian expansive tendencies." Kennan later turned against the containment policy and noted several deficiencies in his X Article. He later said that by containment he meant not the containment of Soviet Power "by military means of a military threat, but the political containment of a political threat." Second, Kennan admitted a failure in the article to specify the geographical scope of "containment", and that containment was not something he believed the United States could necessarily achieve everywhere successfully. Harry S. Truman After Republicans gained control of Congress in the 1946 elections, President Truman, a Democrat, made a dramatic speech that is often considered to mark the beginning of the Cold War. In March 1947, he requested that Congress appropriate $400 million in aid to the Greek and Turkish governments, which were fighting communist subversion. Truman pledged to, "support free peoples who are resisting attempted subjugation by armed minorities or by outside pressures." This pledge became known as the Truman Doctrine. Portraying the issue as a mighty clash between "totalitarian regimes" and "free peoples", the speech marks the adoption of containment as official US policy. Congress appropriated the money. Truman's motives on that occasion have been the subject of considerable scholarship and several schools of interpretation. In the orthodox explanation of Herbert Feis, a series of aggressive Soviet actions in 1945–1947 in Poland, Iran, Turkey, and elsewhere awakened the American public to the new danger to freedom to which Truman responded. In the revisionist view of William Appleman Williams, Truman's speech was an expression of longstanding American expansionism. In the realpolitik view of Lynn E. Davis, Truman was a naive idealist who unnecessarily provoked the Soviets by couching disputes in terms like democracy and freedom that were alien to the communist vision. According to a psychological analysis by Deborah Larson, Truman felt a need to prove his decisiveness and feared that aides would make unfavorable comparisons between him and his predecessor, Roosevelt. "I am here to make decisions, and whether they prove right or wrong I am going to take them", he once said. The drama surrounding the announcement of the Truman Doctrine catered to the president's self-image of a strong and decisive leader, but his real decision-making process was more complex and gradual. The timing of the speech was not a response to any particular Soviet action but to the fact that the Republican Party had just gained control of Congress. Truman was little involved in drafting the speech and did not himself adopt the hard-line attitude that it suggested until several months later. The British, with their own position weakened by economic distress, urgently called on the U.S. to take over the traditional British role in Greece. Undersecretary of State Dean Acheson took the lead in Washington, warning congressional leaders in late February 1947 that if the United States did not take over from the British, the result most probably would be a "Soviet breakthrough" that "might open three continents to Soviet penetration." Truman was explicit about the challenge of communism taking control of Greece. He won wide support from both parties as well as experts in foreign policy inside and outside the government. It was strongly opposed by the left, notably by former Vice President Henry A. Wallace, who ran against Truman in the 1948 presidential campaign. Truman, under the guidance of Acheson, followed up his speech with a series of measures to contain Soviet influence in Europe, including the Marshall Plan, or European Recovery Program, and NATO, a 1949 military alliance between the U.S. and Western European nations. Because containment required detailed information about communist moves, the government relied increasingly on the Central Intelligence Agency (CIA). Established by the National Security Act of 1947, the CIA conducted espionage in foreign lands, some of it visible, more of it secret. Truman approved a classified statement of containment policy called NSC 20/4 in November 1948, the first comprehensive statement of security policy ever created by the United States. The Soviet Union's first nuclear test in 1949 prompted the National Security Council to formulate a revised security doctrine. Completed in April 1950, it became known as NSC 68. It concluded that a massive military buildup was necessary to deal with the Soviet threat. According to the report, drafted by Paul Nitze and others: In the words of the Federalist (No. 28) "The means to be employed must be proportioned to the extent of the mischief." The mischief may be a global war or it may be a Soviet campaign for limited objectives. In either case, we should take no avoidable initiative which would cause it to become a war of annihilation, and if we have the forces to defeat a Soviet drive for limited objectives it may well be to our interest not to let it become a global war. Alternative strategies There were three alternative policies to containment under discussion in the late 1940s. The first was a return to isolationism, minimizing American involvement with the rest of the world, a policy that was supported by conservative Republicans, especially from the Midwest, including former President Herbert Hoover and Senator Robert A. Taft. However, many other Republicans, led by Senator Arthur H. Vandenberg, said that policy had helped cause World War II and so was too dangerous to revive. The second policy was a continuation of the détente policies that aimed at friendly relationships with the Soviet Union, especially trade. Roosevelt had been the champion of détente, but he was dead, and most of his inner circle had left the government by 1946. The chief proponent of détente was Henry Wallace, a former vice president and the Secretary of Commerce under Truman. Wallace's position was supported by far-left elements of the CIO, but they were purged in 1947 and 1948. Wallace ran against Truman on the Progressive Party ticket in 1948, but his campaign was increasingly dominated by Communists, which helped to discredit détente. The third policy was rollback, an aggressive effort to undercut or destroy the Soviet Union itself. Military rollback against the Soviet Union was proposed by James Burnham and other conservative strategists in the late 1940s. After 1954, Burnham and like-minded strategists became editors and regular contributors to William F. Buckley Jr.'s National Review magazine. Truman himself adopted a rollback strategy in the Korean War after the success of the Inchon landings in September 1950, only to reverse himself after the Chinese counterattack two months later and revert to containment. General Douglas MacArthur called on Congress to continue the rollback policy, but Truman fired him for insubordination. Under President Dwight D. Eisenhower, a rollback strategy was considered against communism in Eastern Europe from 1953 to 1956. Eisenhower agreed to a propaganda campaign to roll back the influence of communism psychologically, but he refused to intervene in the 1956 Hungarian Revolution, mainly for fear that it would cause World War III. Since late 1949, when the Soviets had successfully tested an atomic bomb, they had been known to possess nuclear weapons. Korea The U.S. followed containment when it first entered the Korean War to defend South Korea from a communist invasion by North Korea. Initially, this directed the action of the U.S. to only push back North Korea across the 38th Parallel and restore South Korea's sovereignty, thereby allowing North Korea's survival as a state. However, the success of the Inchon landing inspired the U.S. and the United Nations to adopt a rollback strategy instead and to overthrow communist North Korea, thus allowing nationwide elections under UN auspices. General Douglas MacArthur then advanced across the 38th Parallel into North Korea. The Chinese, fearful of a possible U.S. presence on their border or even an invasion by them, then sent in a large army and defeated the UN forces, pushing them back below the 38th parallel. Truman publicly hinted that he might use his "ace in the hole" of the atomic bomb, but Mao was unmoved. The episode was used to support the wisdom of the containment doctrine as opposed to rollback. The Communists were later pushed back to roughly around the original border, with minimal changes. Truman criticized MacArthur's focus on absolute victory and adopted a "limited war" policy. His focus shifted to negotiating a settlement, which was finally reached in 1953. For his part, MacArthur denounced Truman's "no-win policy." Dulles Many Republicans, including John Foster Dulles, were concerned that Truman had been too timid. In 1952, Dulles called for rollback and the eventual liberation of Eastern Europe. Dulles was named secretary of state by incoming President Eisenhower, but Eisenhower's decision not to intervene during the 1956 Hungarian Revolution, which was put down by the Soviet Army, made containment a bipartisan doctrine. Eisenhower relied on clandestine CIA actions to undermine hostile governments and used economic and military foreign aid to strengthen governments supporting the American position in the Cold War. Cuba In the Cuban Missile Crisis of 1962, the top officials in Washington debated using rollback to get rid of Soviet nuclear missiles, which were threatening the United States. There was fear of a nuclear war until a deal was reached in which the Soviets would publicly remove their nuclear weapons, the United States would secretly remove its missiles from Turkey and to avoid invading Cuba. The policy of containing Cuba was put into effect by President John F. Kennedy and continued until 2015. Vietnam Senator Barry Goldwater, the Republican candidate for president in 1964, challenged containment and asked, "Why not victory?" President Lyndon Johnson, the Democratic nominee, answered that rollback risked nuclear war. Johnson explained containment doctrine by quoting the Bible: "Hitherto shalt thou come, but not further." Goldwater lost to Johnson in the 1964 election by a wide margin. Johnson adhered closely to containment during the Vietnam War. Rejecting proposals by General William Westmoreland for U.S. ground forces to advance into Laos and cut communist supply lines, Johnson gathered a group of elder statesmen called The Wise Men. The group included Kennan, Acheson and other former Truman advisors. Rallies in support of the troops were discouraged for fear that a patriotic response would lead to demands for victory and rollback. Military responsibility was divided among three generals so that no powerful theater commander could emerge to challenge Johnson as MacArthur had challenged Truman. Nixon, who replaced Johnson in 1969, referred to his foreign policy as détente, a relaxation of tension. Although it continued to aim at restraining the Soviet Union, it was based on political realism, thinking in terms of national interest, as opposed to crusades against communism or for democracy. Emphasis was placed on talks with the Soviet Union concerning nuclear weapons called the Strategic Arms Limitation Talks. Nixon reduced U.S. military presence in Vietnam to the minimum required to contain communist advances, in a policy called Vietnamization. As the war continued, it grew less popular. A Democratic Congress forced Nixon, a Republican, to abandon the policy in 1973 by enacting the Case–Church Amendment, which ended U.S. military involvement in Vietnam and led to successful communist invasions of South Vietnam, Laos, and Cambodia. Afghanistan President Jimmy Carter came to office in 1977 and was committed to a foreign policy that emphasized human rights. However, in response to the Soviet invasion of Afghanistan, containment was again made a priority. The wording of the Carter Doctrine (1980) intentionally echoed that of the Truman Doctrine. Reagan Doctrine Following the communist victory in Vietnam, Democrats began to view further communist advances as inevitable, but Republicans returned to the rollback doctrine. Ronald Reagan, a long-time advocate of rollback, was elected U.S. president in 1980. He took a more aggressive approach to dealings with the Soviets and believed that détente was misguided and peaceful coexistence was tantamount to surrender. When the Soviet Union invaded Afghanistan in 1979, American policymakers worried that the Soviets were making a run for control of the Persian Gulf. Throughout the 1980s, under a policy that came to be known as the Reagan Doctrine, the United States provided technical and economic assistance to the Afghan guerrillas (mujahideen) fighting against the Soviet army. After the Cold War The conclusion of the Cold War in 1992 marked the official end of the containment policy, but the U.S. kept its bases in the areas around Russia, such as those in Iceland, Germany, and Turkey. Much of the policy later helped influence U.S. foreign policy towards China in the 21st century. See also References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Middle_East#cite_note-69] | [TOKENS: 6152] |
Contents Middle East The Middle East[b] is a geopolitical region encompassing the Arabian Peninsula, Egypt, Iran, Iraq, the Levant, and Turkey. The term came into widespread usage by Western European nations in the early 20th century as a replacement of the term Near East (both were in contrast to the Far East). The term "Middle East" has led to some confusion over its changing definitions. Since the late 20th century, it has been criticized as being too Eurocentric. The region includes the vast majority of the territories included in the closely associated definition of West Asia, but without the South Caucasus. It also includes all of Egypt (not just the Sinai region) and all of Turkey (including East Thrace). Most Middle Eastern countries (13 out of 18) are part of the Arab world. The three most populous countries in the region are Egypt, Iran, and Turkey, while Saudi Arabia is the largest Middle Eastern country by area. The history of the Middle East dates back to ancient times, and it was long considered the "cradle of civilization". The geopolitical importance of the region has been recognized and competed for during millennia. The Abrahamic religions (Judaism, Christianity, and Islam) have their origins in the Middle East. Arabs constitute the main ethnic group in the region, followed by Turks, Persians, Kurds, Jews, and Assyrians. The Middle East generally has a hot, arid climate, especially in the Arabian and Egyptian regions. Several major rivers provide irrigation to support agriculture in limited areas here, such as the Nile Delta in Egypt, the Tigris and Euphrates watersheds of Mesopotamia, and the basin of the Jordan River that spans most of the Levant. These regions are collectively known as the Fertile Crescent, and comprise the core of what historians had long referred to as the cradle of civilization; multiple regions of the world have since been classified as also having developed independent, original civilizations. Conversely, the Levantine coast and most of Turkey have relatively temperate climates typical of the Mediterranean, with dry summers and cool, wet winters. Most of the countries that border the Persian Gulf have vast reserves of petroleum. Monarchs of the Arabian Peninsula in particular have benefitted economically from petroleum exports. Because of the arid climate and dependence on the fossil fuel industry, the Middle East is both a major contributor to climate change and a region that is expected to be severely adversely affected by it. Other concepts of the region exist, including the broader Middle East and North Africa (MENA), which includes states of the Maghreb and the Sudan. The term the "Greater Middle East" also includes Afghanistan, Mauritania, Pakistan, as well as parts of East Africa, and sometimes Central Asia and the South Caucasus. Terminology The term "Middle East" may have originated in the 1850s in the British India Office. However, it became more widely known when United States naval strategist Alfred Thayer Mahan used the term in 1902 to "designate the area between Arabia and India". During this time the British and Russian empires were vying for influence in Central Asia, a rivalry that would become known as the Great Game. Mahan realized not only the strategic importance of the region, but also of its center, the Persian Gulf. He labeled the area surrounding the Persian Gulf as the Middle East. He said that, beyond Egypt's Suez Canal, the Gulf was the most important passage for Britain to control in order to keep the Russians from advancing towards British India. Mahan first used the term in his article "The Persian Gulf and International Relations", published in September 1902 in the National Review, a British journal. The Middle East, if I may adopt a term which I have not seen, will some day need its Malta, as well as its Gibraltar; it does not follow that either will be in the Persian Gulf. Naval force has the quality of mobility which carries with it the privilege of temporary absences; but it needs to find on every scene of operation established bases of refit, of supply, and in case of disaster, of security. The British Navy should have the facility to concentrate in force if occasion arise, about Aden, India, and the Persian Gulf. Mahan's article was reprinted in The Times and followed in October by a 20-article series entitled "The Middle Eastern Question", written by Sir Ignatius Valentine Chirol. During this series, Sir Ignatius expanded the definition of Middle East to include "those regions of Asia which extend to the borders of India or command the approaches to India." After the series ended in 1903, The Times removed quotation marks from subsequent uses of the term. Until World War II, it was customary to refer to areas centered on Turkey and the eastern shore of the Mediterranean as the "Near East", while the "Far East" centered on China, India and Japan. The Middle East was then defined as the area from Mesopotamia to Burma; namely, the area between the Near East and the Far East. This area broadly corresponds to South Asia. In the late 1930s, the British established the Middle East Command, which was based in Cairo, for its military forces in the region. After that time, the term "Middle East" gained broader usage in Europe and the United States. Following World War II, for example, the Middle East Institute was founded in Washington, D.C. in 1946. The corresponding adjective is Middle Eastern and the derived noun is Middle Easterner. While non-Eurocentric terms such as "Southwest Asia" or "Swasia" have been sparsely used, the classification of the African country, Egypt, among those counted in the Middle East challenges the usefulness of using such terms. The description Middle has also led to some confusion over changing definitions. Before the First World War, "Near East" was used in English to refer to the Balkans and the Ottoman Empire, while "Middle East" referred to the Caucasus, Persia, and Arabian lands, and sometimes Afghanistan, India and others. In contrast, "Far East" referred to the countries of East Asia (e.g. China, Japan, and Korea). With the collapse of the Ottoman Empire in 1918, "Near East" largely fell out of common use in English, while "Middle East" came to be applied to the emerging independent countries of the Islamic world. However, the usage "Near East" was retained by a variety of academic disciplines, including archaeology and ancient history. In their usage, the term describes an area identical to the term Middle East, which is not used by these disciplines (see ancient Near East).[citation needed] The first official use of the term "Middle East" by the United States government was in the 1957 Eisenhower Doctrine, which pertained to the Suez Crisis. Secretary of State John Foster Dulles defined the Middle East as "the area lying between and including Libya on the west and Pakistan on the east, Syria and Iraq on the North and the Arabian peninsula to the south, plus the Sudan and Ethiopia." In 1958, the State Department explained that the terms "Near East" and "Middle East" were interchangeable, and defined the region as including only Egypt, Syria, Israel, Lebanon, Jordan, Iraq, Saudi Arabia, Kuwait, Bahrain, and Qatar. Since the late 20th century, scholars and journalists from the region, such as journalist Louay Khraish and historian Hassan Hanafi have criticized the use of "Middle East" as a Eurocentric and colonialist term. The Associated Press Stylebook of 2004 says that Near East formerly referred to the farther west countries while Middle East referred to the eastern ones, but that now they are synonymous. It instructs: Use Middle East unless Near East is used by a source in a story. Mideast is also acceptable, but Middle East is preferred. European languages have adopted terms similar to Near East and Middle East. Since these are based on a relative description, the meanings depend on the country and are generally different from the English terms. In German the term Naher Osten (Near East) is still in common use (nowadays the term Mittlerer Osten is more and more common in press texts translated from English sources, albeit having a distinct meaning). In the four Slavic languages, Russian Ближний Восток or Blizhniy Vostok, Bulgarian Близкия Изток, Polish Bliski Wschód or Croatian Bliski istok (terms meaning Near East are the only appropriate ones for the region). However, some European languages do have "Middle East" equivalents, such as French Moyen-Orient, Swedish Mellanöstern, Spanish Oriente Medio or Medio Oriente, Greek is Μέση Ανατολή (Mesi Anatoli), and Italian Medio Oriente.[c] Perhaps because of the political influence of the United States and Europe, and the prominence of Western press, the Arabic equivalent of Middle East (Arabic: الشرق الأوسط ash-Sharq al-Awsaṭ) has become standard usage in the mainstream Arabic press. It comprises the same meaning as the term "Middle East" in North American and Western European usage. The designation, Mashriq, also from the Arabic root for East, also denotes a variously defined region around the Levant, the eastern part of the Arabic-speaking world (as opposed to the Maghreb, the western part). Even though the term originated in the West, countries of the Middle East that use languages other than Arabic also use that term in translation. For instance, the Persian equivalent for Middle East is خاورمیانه (Khāvar-e miyāneh), the Hebrew is המזרח התיכון (hamizrach hatikhon), and the Turkish is Orta Doğu. Countries and territory Traditionally included within the Middle East are Arabia, Asia Minor, East Thrace, Egypt, Iran, the Levant, Mesopotamia, and the Socotra Archipelago. The region includes 17 UN-recognized countries and one British Overseas Territory. Various concepts are often paralleled to the Middle East, most notably the Near East, Fertile Crescent, and Levant. These are geographical concepts, which refer to large sections of the modern-day Middle East, with the Near East being the closest to the Middle East in its geographical meaning. Due to it primarily being Arabic speaking, the Maghreb region of North Africa is sometimes included. "Greater Middle East" is a political term coined by the second Bush administration in the first decade of the 21st century to denote various countries, pertaining to the Muslim world, specifically Afghanistan, Iran, Pakistan, and Turkey. Various Central Asian countries are sometimes also included. History The Middle East lies at the juncture of Africa and Eurasia and of the Indian Ocean and the Mediterranean Sea (see also: Indo-Mediterranean). It is the birthplace and spiritual center of religions such as Christianity, Islam, Judaism, Manichaeism, Yezidi, Druze, Yarsan, and Mandeanism, and in Iran, Mithraism, Zoroastrianism, Manicheanism, and the Baháʼí Faith. Throughout its history the Middle East has been a major center of world affairs; a strategically, economically, politically, culturally, and religiously sensitive area. The region is one of the regions where agriculture was independently discovered, and from the Middle East it was spread, during the Neolithic, to different regions of the world such as Europe, the Indus Valley and Eastern Africa. Prior to the formation of civilizations, advanced cultures formed all over the Middle East during the Stone Age. The search for agricultural lands by agriculturalists, and pastoral lands by herdsmen meant different migrations took place within the region and shaped its ethnic and demographic makeup. The Middle East is widely and most famously known as the cradle of civilization. The world's earliest civilizations, Mesopotamia (Sumer, Akkad, Assyria and Babylonia), ancient Egypt and Kish in the Levant, all originated in the Fertile Crescent and Nile Valley regions of the ancient Near East. These were followed by the Hittite, Greek, Hurrian and Urartian civilisations of Asia Minor; Elam, Persia and Median civilizations in Iran, as well as the civilizations of the Levant (such as Ebla, Mari, Nagar, Ugarit, Canaan, Aramea, Mitanni, Phoenicia and Israel) and the Arabian Peninsula (Magan, Sheba, Ubar). The Near East was first largely unified under the Neo Assyrian Empire, then the Achaemenid Empire followed later by the Macedonian Empire and after this to some degree by the Iranian empires (namely the Parthian and Sassanid Empires), the Roman Empire and Byzantine Empire. The region served as the intellectual and economic center of the Roman Empire and played an exceptionally important role due to its periphery on the Sassanid Empire. Thus, the Romans stationed up to five or six of their legions in the region for the sole purpose of defending it from Sassanid and Bedouin raids and invasions. From the 4th century CE onwards, the Middle East became the center of the two main powers at the time, the Byzantine Empire and the Sassanid Empire. However, it would be the later Islamic Caliphates of the Middle Ages, or Islamic Golden Age which began with the Islamic conquest of the region in the 7th century AD, that would first unify the entire Middle East as a distinct region and create the dominant Islamic Arab ethnic identity that largely (but not exclusively) persists today. The 4 caliphates that dominated the Middle East for more than 600 years were the Rashidun Caliphate, the Umayyad caliphate, the Abbasid caliphate and the Fatimid caliphate. Additionally, the Mongols would come to dominate the region, the Kingdom of Armenia would incorporate parts of the region to their domain, the Seljuks would rule the region and spread Turko-Persian culture, and the Franks would found the Crusader states that would stand for roughly two centuries. Josiah Russell estimates the population of what he calls "Islamic territory" as roughly 12.5 million in 1000 – Anatolia 8 million, Syria 2 million, and Egypt 1.5 million. From the 16th century onward, the Middle East came to be dominated, once again, by two main powers: the Ottoman Empire and the Safavid dynasty. The modern Middle East began after World War I, when the Ottoman Empire, which was allied with the Central Powers, was defeated by the Allies and partitioned into a number of separate nations, initially under British and French Mandates. Other defining events in this transformation included the establishment of Israel in 1948 and the eventual departure of European powers, notably Britain and France by the end of the 1960s. They were supplanted in some part by the rising influence of the United States from the 1970s onwards. In the 20th century, the region's significant stocks of crude oil gave it new strategic and economic importance. Mass production of oil began around 1945, with Saudi Arabia, Iran, Kuwait, Iraq, and the United Arab Emirates having large quantities of oil. Estimated oil reserves, especially in Saudi Arabia and Iran, are some of the highest in the world, and the international oil cartel OPEC is dominated by Middle Eastern countries. During the Cold War, the Middle East was a theater of ideological struggle between the two superpowers and their allies: NATO and the United States on one side, and the Soviet Union and Warsaw Pact on the other, as they competed to influence regional allies. Besides the political reasons there was also the "ideological conflict" between the two systems. Moreover, as Louise Fawcett argues, among many important areas of contention, or perhaps more accurately of anxiety, were, first, the desires of the superpowers to gain strategic advantage in the region, second, the fact that the region contained some two-thirds of the world's oil reserves in a context where oil was becoming increasingly vital to the economy of the Western world [...] Within this contextual framework, the United States sought to divert the Arab world from Soviet influence. Throughout the 20th and 21st centuries, the region has experienced both periods of relative peace and tolerance and periods of conflict particularly between Sunnis and Shiites. Geography In 2018, the MENA region emitted 3.2 billion tonnes of carbon dioxide and produced 8.7% of global greenhouse gas emissions (GHG) despite making up only 6% of the global population. These emissions are mostly from the energy sector, an integral component of many Middle Eastern and North African economies due to the extensive oil and natural gas reserves that are found within the region. The Middle East region is one of the most vulnerable to climate change. The impacts include increase in drought conditions, aridity, heatwaves and sea level rise. Sharp global temperature and sea level changes, shifting precipitation patterns and increased frequency of extreme weather events are some of the main impacts of climate change as identified by the Intergovernmental Panel on Climate Change (IPCC). The MENA region is especially vulnerable to such impacts due to its arid and semi-arid environment, facing climatic challenges such as low rainfall, high temperatures and dry soil. The climatic conditions that foster such challenges for MENA are projected by the IPCC to worsen throughout the 21st century. If greenhouse gas emissions are not significantly reduced, part of the MENA region risks becoming uninhabitable before the year 2100. Climate change is expected to put significant strain on already scarce water and agricultural resources within the MENA region, threatening the national security and political stability of all included countries. Over 60 percent of the region's population lives in high and very high water-stressed areas compared to the global average of 35 percent. This has prompted some MENA countries to engage with the issue of climate change on an international level through environmental accords such as the Paris Agreement. Law and policy are also being established on a national level amongst MENA countries, with a focus on the development of renewable energies. Economy Middle Eastern economies range from being very poor (such as Gaza and Yemen) to extremely wealthy nations (such as Qatar and UAE). According to the International Monetary Fund, the three largest Middle Eastern economies in nominal GDP in 2023 were Saudi Arabia ($1.06 trillion), Turkey ($1.03 trillion), and Israel ($0.54 trillion). For nominal GDP per person, the highest ranking countries are Qatar ($83,891), Israel ($55,535), the United Arab Emirates ($49,451) and Cyprus ($33,807). Turkey ($3.6 trillion), Saudi Arabia ($2.3 trillion), and Iran ($1.7 trillion) had the largest economies in terms of GDP PPP. For GDP PPP per person, the highest-ranking countries are Qatar ($124,834), the United Arab Emirates ($88,221), Saudi Arabia ($64,836), Bahrain ($60,596) and Israel ($54,997). The lowest-ranking country in the Middle East, in terms of GDP nominal per capita, is Yemen ($573). The economic structure of Middle Eastern nations are different because while some are heavily dependent on export of only oil and oil-related products (Saudi Arabia, the UAE and Kuwait), others have a highly diverse economic base (such as Cyprus, Israel, Turkey and Egypt). Industries of the Middle Eastern region include oil and oil-related products, agriculture, cotton, cattle, dairy, textiles, leather products, surgical instruments, defence equipment (guns, ammunition, tanks, submarines, fighter jets, UAVs, and missiles). Banking is an important sector, especially for UAE and Bahrain. With the exception of Cyprus, Turkey, Egypt, Lebanon and Israel, tourism has been a relatively undeveloped area of the economy, in part because of the socially conservative nature of the region as well as political turmoil in certain regions. Since the end of the COVID pandemic however, countries such as the UAE, Bahrain, and Jordan have begun attracting greater numbers of tourists because of improving tourist facilities and the relaxing of tourism-related restrictive policies. Unemployment is high in the Middle East and North Africa region, particularly among people aged 15–29, a demographic representing 30% of the region's population. The total regional unemployment rate in 2025 is 10.8%, and among youth is as high as 28%. Demographics Arabs constitute the largest ethnic group in the Middle East, followed by various Iranian peoples and then by Turkic peoples (Turkish, Azeris, Syrian Turkmen, and Iraqi Turkmen). Native ethnic groups of the region include, in addition to Arabs, Arameans, Assyrians, Baloch, Berbers, Copts, Druze, Greek Cypriots, Jews, Kurds, Lurs, Mandaeans, Persians, Samaritans, Shabaks, Tats, and Zazas. European ethnic groups that form a diaspora in the region include Albanians, Bosniaks, Circassians (including Kabardians), Crimean Tatars, Greeks, Franco-Levantines, Italo-Levantines, and Iraqi Turkmens. Among other migrant populations are Chinese, Filipinos, Indians, Indonesians, Pakistanis, Pashtuns, Romani, and Afro-Arabs. "Migration has always provided an important vent for labor market pressures in the Middle East. For the period between the 1970s and 1990s, the Arab states of the Persian Gulf in particular provided a rich source of employment for workers from Egypt, Yemen and the countries of the Levant, while Europe had attracted young workers from North African countries due both to proximity and the legacy of colonial ties between France and the majority of North African states." According to the International Organization for Migration, there are 13 million first-generation migrants from Arab nations in the world, of which 5.8 reside in other Arab countries. Expatriates from Arab countries contribute to the circulation of financial and human capital in the region and thus significantly promote regional development. In 2009 Arab countries received a total of US$35.1 billion in remittance in-flows and remittances sent to Jordan, Egypt and Lebanon from other Arab countries are 40 to 190 per cent higher than trade revenues between these and other Arab countries. In Somalia, the Somali Civil War has greatly increased the size of the Somali diaspora, as many of the best educated Somalis left for Middle Eastern countries as well as Europe and North America. Non-Arab Middle Eastern countries such as Turkey, Israel and Iran are also subject to important migration dynamics. A fair proportion of those migrating from Arab nations are from ethnic and religious minorities facing persecution and are not necessarily ethnic Arabs, Iranians or Turks.[citation needed] Large numbers of Kurds, Jews, Assyrians, Greeks and Armenians as well as many Mandeans have left nations such as Iraq, Iran, Syria and Turkey for these reasons during the last century. In Iran, many religious minorities such as Christians, Baháʼís, Jews and Zoroastrians have left since the Islamic Revolution of 1979. The Middle East is very diverse when it comes to religions, many of which originated there. Islam is the largest religion in the Middle East, but other faiths that originated there, such as Judaism and Christianity, are also well represented. Christian communities have played a vital role in the Middle East, and they represent 78% of Cyprus population, and 40.5% of Lebanon, where the Lebanese president, half of the cabinet, and half of the parliament follow one of the various Lebanese Christian rites. There are also important minority religions like the Baháʼí Faith, Yarsanism, Yazidism, Zoroastrianism, Mandaeism, Druze, and Shabakism, and in ancient times the region was home to Mesopotamian religions, Canaanite religions, Manichaeism, Mithraism and various monotheist gnostic sects. The six top languages, in terms of numbers of speakers, are Arabic, Persian, Turkish, Kurdish, Modern Hebrew and Greek. About 20 minority languages are also spoken in the Middle East. Arabic, with all its dialects, is the most widely spoken language in the Middle East, with Literary Arabic being official in all North African and in most West Asian countries. Arabic dialects are also spoken in some adjacent areas in neighbouring Middle Eastern non-Arab countries. It is a member of the Semitic branch of the Afro-Asiatic languages. Several Modern South Arabian languages such as Mehri and Soqotri are also spoken in Yemen and Oman. Another Semitic language is Aramaic and its dialects are spoken mainly by Assyrians and Mandaeans, with Western Aramaic still spoken in two villages near Damascus, Syria. There is also an Oasis Berber-speaking community in Egypt where the language is also known as Siwa. It is a non-Semitic Afro-Asiatic sister language. Persian is the second most spoken language. While it is primarily spoken in Iran and some border areas in neighbouring countries, the country is one of the region's largest and most populous. It belongs to the Indo-Iranian branch of the family of Indo-European languages. Other Western Iranic languages spoken in the region include Achomi, Daylami, Kurdish dialects, Semmani, Lurish, amongst many others. The close third-most widely spoken language, Turkish, is largely confined to Turkey, which is also one of the region's largest and most populous countries, but it is present in areas in neighboring countries. It is a member of the Turkic languages, which have their origins in East Asia. Another Turkic language, Azerbaijani, is spoken by Azerbaijanis in Iran. The fourth-most widely spoken language, Kurdish, is spoken in the countries of Iran, Iraq, Syria and Turkey, Sorani Kurdish is the second official language in Iraq (instated after the 2005 constitution) after Arabic. Hebrew is the official language of Israel, with Arabic given a special status after the 2018 Basic law lowered its status from an official language prior to 2018. Hebrew is spoken and used by over 80% of Israel's population, the other 20% using Arabic. Modern Hebrew only began being spoken in the 20th century after being revived in the late 19th century by Elizer Ben-Yehuda (Elizer Perlman) and European Jewish settlers, with the first native Hebrew speaker being born in 1882. Greek is one of the two official languages of Cyprus, and the country's main language. Small communities of Greek speakers exist all around the Middle East; until the 20th century it was also widely spoken in Asia Minor (being the second most spoken language there, after Turkish) and Egypt. During the antiquity, Ancient Greek was the lingua franca for many areas of the western Middle East and until the Muslim expansion it was widely spoken there as well. Until the late 11th century, it was also the main spoken language in Asia Minor; after that it was gradually replaced by the Turkish language as the Anatolian Turks expanded and the local Greeks were assimilated, especially in the interior. English is one of the official languages of Akrotiri and Dhekelia. It is also commonly taught and used as a foreign second language, in countries such as Egypt, Jordan, Iran, Iraq, Qatar, Bahrain, United Arab Emirates and Kuwait. It is also a main language in some Emirates of the United Arab Emirates. It is also spoken as native language by Jewish immigrants from Anglophone countries (UK, US, Australia) in Israel and understood widely as second language there. French is taught and used in many government facilities and media in Lebanon, and is taught in some primary and secondary schools of Egypt and Syria. Maltese, a Semitic language mainly spoken in Europe, is used by the Franco-Maltese diaspora in Egypt. Due to widespread immigration of French Jews to Israel, it is the native language of approximately 200,000 Jews in Israel. Armenian speakers are to be found in the region. Georgian is spoken by the Georgian diaspora. Russian is spoken by a large portion of the Israeli population, because of emigration in the late 1990s. Russian today is a popular unofficial language in use in Israel; news, radio and sign boards can be found in Russian around the country after Hebrew and Arabic. Circassian is also spoken by the diaspora in the region and by almost all Circassians in Israel who speak Hebrew and English as well. The largest Romanian-speaking community in the Middle East is found in Israel, where as of 1995[update] Romanian is spoken by 5% of the population.[d] Bengali, Hindi and Urdu are widely spoken by migrant communities in many Middle Eastern countries, such as Saudi Arabia (where 20–25% of the population is South Asian), the United Arab Emirates (where 50–55% of the population is South Asian), and Qatar, which have large numbers of Pakistani, Bangladeshi and Indian immigrants. Culture The Middle East has recently become more prominent in hosting global sport events due to its wealth and desire to diversify its economy. The South Asian diaspora is a major backer of cricket in the region. See also Notes References Further reading External links 29°N 41°E / 29°N 41°E / 29; 41 |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Georges_Cuvier] | [TOKENS: 9106] |
Contents Georges Cuvier Jean Léopold Nicolas Frédéric, baron Cuvier (23 August 1769 – 13 May 1832), known as Georges Cuvier (/ˈkjuːvieɪ/; French: [ʒɔʁʒ(ə) kyvje]), was a French naturalist and zoologist, sometimes referred to as the "founding father of paleontology". Cuvier was a major figure in natural sciences research in the early 19th century and was instrumental in establishing the fields of comparative anatomy and paleontology through his work in comparing living animals with fossils. Cuvier's work is considered the foundation of vertebrate paleontology, and he expanded Linnaean taxonomy by grouping classes into phyla and incorporating both fossils and living species into the classification. Cuvier is also known for establishing extinction as a scientific fact—at the time, extinction was considered by many of Cuvier's contemporaries to be merely controversial speculation. In his Essay on the Theory of the Earth (1813), Cuvier proposed that now-extinct species had been wiped out by periodic catastrophic flooding events. In this way, Cuvier became the most influential proponent of catastrophism in geology in the early 19th century. His study of the strata of the Paris basin with Alexandre Brongniart established the basic principles of biostratigraphy. Among his other accomplishments, Cuvier established that elephant-like bones found in North America belonged to an extinct animal he later would name a "mastodon", and that a large skeleton dug up in present-day Argentina was of a giant, prehistoric ground sloth, which he named Megatherium. He also established two ungulate genera from the Paris Basin named Palaeotherium and Anoplotherium based on fragmentary remains alone, although more complete remains were later uncovered. He named the pterosaur Pterodactylus and described (but did not discover or name) the aquatic reptile Mosasaurus. Cuvier is also remembered for strongly opposing theories of evolution, which at the time (before Darwin's theory) were mainly proposed by Jean-Baptiste de Lamarck and Geoffroy Saint-Hilaire. Cuvier believed there was no evidence for evolution, but rather evidence for cyclical creations and destructions of life forms by global extinction events such as deluges. In 1830, Cuvier and Geoffroy engaged in a famous debate, which is said to exemplify the two major deviations in biological thinking at the time – whether animal structure was due to function or (evolutionary) morphology. Cuvier supported function and rejected Lamarck's thinking. Cuvier also conducted racial studies which provided part of the foundation for scientific racism, and published work on the supposed differences between racial groups' physical properties and mental abilities. Cuvier subjected Sarah Baartman to examinations alongside other French naturalists during a period in which she was held captive in a state of neglect. Cuvier examined Baartman shortly before her death, and conducted a dissection following her death that disparagingly compared her physical features to those of monkeys. His most famous work among the general public was the Preliminary Discourse of the Recherches sur les ossemens fossiles de 1812, which was published as a separate in 1821 and in book form in 1825, with the title Discours sur les révolutions de la surface du Globe. The evolution of his ideas on Comparative Anatomy, Paleontology, and pre-Darwinian Natural History can be seen by comparing these works in the form of a Variorum. Another important work was Le Règne Animal (1817; English: The Animal Kingdom). In 1819, he was created a peer for life in honour of his scientific contributions. Thereafter, he was known as Baron Cuvier. He died in Paris during an epidemic of cholera. Some of Cuvier's most influential followers were Louis Agassiz on the continent and in the United States, and Richard Owen in Britain. His name is one of the 72 names inscribed on the Eiffel Tower. Biography Jean Léopold Nicolas Frédéric Cuvier was born in Montbéliard, where his Protestant ancestors had lived since the time of the Reformation. His mother was Anne Clémence Chatel; his father, Jean-Georges Cuvier, was a lieutenant in the Swiss Guards and a bourgeois of the town of Montbéliard. At the time, the town, which would be annexed to France on 10 October 1793, belonged to the Sovereign County of Montbéliard (in personal union with the Duchy of Württemberg). His mother, who was much younger than his father, tutored him diligently throughout his early years, so he easily surpassed the other children at school. During his gymnasium years, he had little trouble acquiring Latin and Greek, and was always at the head of his class in mathematics, history, and geography. According to Lee, "The history of mankind was, from the earliest period of his life, a subject of the most indefatigable application; and long lists of sovereigns, princes, and the driest chronological facts, once arranged in his memory, were never forgotten." At the age of 10, soon after entering the gymnasium, he encountered a copy of Conrad Gessner's Historiae Animalium, the work that first sparked his interest in natural history. He then began frequent visits to the home of a relative, where he could borrow volumes of the Comte de Buffon's massive Histoire Naturelle. All of these he read and reread, retaining so much of the information, that by the age of 12, "he was as familiar with quadrupeds and birds as a first-rate naturalist." He remained at the gymnasium for four years. Cuvier spent an additional four years at the Caroline Academy in Stuttgart, where he excelled in all of his coursework. Although he knew no German on his arrival, after only nine months of study, he managed to win the school prize for that language. Cuvier's German education exposed him to the work of the geologist Abraham Gottlob Werner (1750–1817), whose Neptunism and emphasis on the importance of rigorous, direct observation of three-dimensional, structural relationships of rock formations to geological understanding provided models for Cuvier's scientific theories and methods. Upon graduation, he had no money on which to live as he awaited an appointment to an academic office. So in July 1788, he took a job at Fiquainville chateau in Normandy as tutor to the only son of the Comte d'Héricy, a Protestant noble. There, during the early 1790s, he began his comparisons of fossils with extant forms. Cuvier regularly attended meetings held at the nearby town of Valmont for the discussion of agricultural topics. There, he became acquainted with Henri Alexandre Tessier (1741–1837), who had assumed a false identity. Previously, he had been a physician and well-known agronomist, who had fled the Terror in Paris. After hearing Tessier speak on agricultural matters, Cuvier recognized him as the author of certain articles on agriculture in the Encyclopédie Méthodique and addressed him as M. Tessier. Tessier replied in dismay, "I am known, then, and consequently lost."—"Lost!" replied M. Cuvier, "no; you are henceforth the object of our most anxious care." They soon became intimate and Tessier introduced Cuvier to his colleagues in Paris"I have just found a pearl in the dunghill of Normandy", he wrote his friend Antoine-Augustin Parmentier. As a result, Cuvier entered into correspondence with several leading naturalists of the day and was invited to Paris. Arriving in the spring of 1795, at the age of 26, he soon became the assistant of Jean-Claude Mertrud (1728–1802), who had been appointed to the chair of Animal Anatomy at the Jardin des Plantes. When Mertrud died in 1802, Cuvier replaced him in office and the Chair changed its name to Chair of Comparative Anatomy. The Institut de France was founded in the same year, and he was elected a member of its Academy of Sciences. On 4 April 1796 he began to lecture at the École Centrale du Pantheon and, at the opening of the National Institute in April, he read his first paleontological paper, which subsequently was published in 1800 under the title Mémoires sur les espèces d'éléphants vivants et fossiles. In this paper, he analyzed skeletal remains of Indian and African elephants, as well as mammoth fossils, and a fossil skeleton known at that time as the "Ohio animal". In his second paper in 1796, he described and analyzed a large skeleton found in Paraguay, which he would name Megatherium. He concluded this skeleton represented yet another extinct animal and, by comparing its skull with living species of tree-dwelling sloths, that it was a kind of ground-dwelling giant sloth. Together, these two 1796 papers were a seminal or landmark event, becoming a turning point in the history of paleontology, and in the development of comparative anatomy, as well. They also greatly enhanced Cuvier's personal reputation and they essentially ended what had been a long-running debate about the reality of extinction. In 1799, he succeeded Daubenton as professor of natural history in the Collège de France. In 1802, he became titular professor at the Jardin des Plantes; and in the same year, he was appointed commissary of the institute to accompany the inspectors general of public instruction. In this latter capacity, he visited the south of France, but in the early part of 1803, he was chosen permanent secretary of the department of physical sciences of the Academy, and he consequently abandoned the earlier appointment and returned to Paris. In 1806, he became a foreign member of the Royal Society, and in 1812, a foreign member of the Royal Swedish Academy of Sciences. In 1812, he became a correspondent for the Royal Institute of the Netherlands, and became a member in 1827. Cuvier was elected a Foreign Honorary Member of the American Academy of Arts and Sciences in 1822. Cuvier then devoted himself more especially to three lines of inquiry: (i) the structure and classification of the Mollusca; (ii) the comparative anatomy and systematic arrangement of the fishes; (iii) fossil mammals and reptiles and, secondarily, the osteology of living forms belonging to the same groups. In 1812, Cuvier made what the cryptozoologist Bernard Heuvelmans called his "Rash dictum": he remarked that it was unlikely that any large animal remained undiscovered. Ten years after his death, the word "dinosaur" would be coined by Richard Owen in 1842. During his lifetime, Cuvier served as an imperial councillor under Napoleon, president of the Council of Public Instruction and chancellor of the university under the restored Bourbons, Grand Officer of the Legion of Honour, a Peer of France, Minister of the Interior, and president of the Council of State under Louis Philippe. He was eminent in all these capacities, and yet the dignity given by such high administrative positions was as nothing compared to his leadership in natural science. Cuvier was by birth, education, and conviction a devout Lutheran, and remained Protestant throughout his life while regularly attending church services. Despite this, he regarded his personal faith as a private matter; he evidently identified himself with his confessional minority group when he supervised governmental educational programs for Protestants. He also was very active in founding the Parisian Biblical Society in 1818, where he later served as a vice president. From 1822 until his death in 1832, Cuvier was Grand Master of the Protestant Faculties of Theology of the French University. Scientific ideas and their impact Cuvier was critical of theories of evolution, in particular those proposed by his contemporaries Lamarck and Geoffroy Saint-Hilaire, which involved the gradual transmutation of one form into another. He repeatedly emphasized that his extensive experience with fossil material indicated one fossil form does not, as a rule, gradually change into a succeeding, distinct fossil form. A deep-rooted source of his opposition to the gradual transformation of species was his goal of creating an accurate taxonomy based on principles of comparative anatomy. Such a project would become impossible if species were mutable, with no clear boundaries between them. According to the University of California Museum of Paleontology, "Cuvier did not believe in organic evolution, for any change in an organism's anatomy would have rendered it unable to survive. He studied the mummified cats and ibises that Geoffroy had brought back from Napoleon's invasion of Egypt, and showed they were no different from their living counterparts; Cuvier used this to support his claim that life forms did not evolve over time." He also observed that Napoleon's expedition to Egypt had retrieved animals mummified thousands of years previously that seemed no different from their modern counterparts. "Certainly", Cuvier wrote, "one cannot detect any greater difference between these creatures and those we see, than between the human mummies and the skeletons of present-day men." Lamarck dismissed this conclusion, arguing that evolution happened much too slowly to be observed over just a few thousand years. Cuvier, however, in turn criticized how Lamarck and other naturalists conveniently introduced hundreds of thousands of years "with a stroke of a pen" to uphold their theory. Instead, he argued that one may judge what a long time would produce only by multiplying what a lesser time produces. Since a lesser time produced no organic changes, neither, he argued, would a much longer time. Moreover, his commitment to the principle of the correlation of parts caused him to doubt that any mechanism could ever gradually modify any part of an animal in isolation from all the other parts (in the way Lamarck proposed), without rendering the animal unable to survive. In his Éloge de M. de Lamarck (Praise for M. de Lamarck), Cuvier wrote that Lamarck's theory of evolution rested on two arbitrary suppositions; the one, that it is the seminal vapour which organizes the embryo; the other, that efforts and desires may engender organs. A system established on such foundations may amuse the imagination of a poet; a metaphysician may derive from it an entirely new series of systems; but it cannot for a moment bear the examination of anyone who has dissected a hand, a viscus, or even a feather. Instead, he said, the typical form makes an abrupt appearance in the fossil record, and persists unchanged to the time of its extinction. Cuvier attempted to explain this paleontological phenomenon he envisioned (which would be readdressed more than a century later by "punctuated equilibrium") and to harmonize it with the Bible. He attributed the different time periods he was aware of as intervals between major catastrophes, the last of which is found in Genesis. Cuvier's claim that new fossil forms appear abruptly in the geological record and then continue without alteration in overlying strata was used by later critics of evolution to support creationism, to whom the abruptness seemed consistent with special divine creation (although Cuvier's finding that different types made their paleontological debuts in different geological strata clearly did not). The lack of change was consistent with the supposed sacred immutability of "species", but, again, the idea of extinction, of which Cuvier was the great proponent, obviously was not. Many writers have unjustly accused Cuvier of obstinately maintaining that fossil human beings could never be found. In his Essay on the Theory of the Earth, he did say, "no human bones have yet been found among fossil remains", but he made it clear exactly what he meant: "When I assert that human bones have not been hitherto found among extraneous fossils, I must be understood to speak of fossils, or petrifactions, properly so called". Petrified bones, which have had time to mineralize and turn to stone, are typically far older than bones found to that date. Cuvier's point was that all human bones found that he knew of, were of relatively recent age because they had not been petrified and had been found only in superficial strata. He was not dogmatic in this claim, however; when new evidence came to light, he included in a later edition an appendix describing a skeleton that he freely admitted was an "instance of a fossil human petrifaction". The harshness of his criticism and the strength of his reputation, however, continued to discourage naturalists from speculating about the gradual transmutation of species, until Charles Darwin published On the Origin of Species more than two decades after Cuvier's death. Early in his tenure at the National Museum in Paris, Cuvier published studies of fossil bones in which he argued that they belonged to large, extinct quadrupeds. His first two such publications were those identifying mammoth and mastodon fossils as belonging to extinct species rather than modern elephants and the study in which he identified the Megatherium as a giant, extinct species of sloth. His primary evidence for his identifications of mammoths and mastodons as separate, extinct species was the structure of their jaws and teeth. His primary evidence that the Megatherium fossil had belonged to a massive sloth came from his comparison of its skull with those of extant sloth species. Cuvier wrote of his paleontological method that "the form of the tooth leads to the form of the condyle, that of the scapula to that of the nails, just as an equation of a curve implies all of its properties; and, just as in taking each property separately as the basis of a special equation we are able to return to the original equation and other associated properties, similarly, the nails, the scapula, the condyle, the femur, each separately reveal the tooth or each other; and by beginning from each of them the thoughtful professor of the laws of organic economy can reconstruct the entire animal." However, Cuvier's actual method was heavily dependent on the comparison of fossil specimens with the anatomy of extant species in the necessary context of his vast knowledge of animal anatomy and access to unparalleled natural history collections in Paris. This reality, however, did not prevent the rise of a popular legend that Cuvier could reconstruct the entire bodily structures of extinct animals given only a few fragments of bone. At the time Cuvier presented his 1796 paper on living and fossil elephants, it was still widely believed that no species of animal had ever become extinct. Authorities such as Buffon had claimed that fossils found in Europe of animals such as the woolly rhinoceros and the mammoth were remains of animals still living in the tropics (i.e. rhinoceros and elephants), which had shifted out of Europe and Asia as the earth became cooler. Thereafter, Cuvier performed a pioneering research study on some elephant fossils excavated around Paris. The bones he studied, however, were remarkably different from the bones of elephants currently thriving in India and Africa. This discovery led Cuvier to denounce the idea that fossils came from those that are currently living. The idea that these bones belonged to elephants living – but hiding – somewhere on Earth seemed ridiculous to Cuvier, because it would be nearly impossible to miss them due to their enormous size. The Megatherium provided another compelling data point for this argument. Ultimately, his repeated identification of fossils as belonging to species unknown to man, combined with mineralogical evidence from his stratigraphical studies in Paris, drove Cuvier to the proposition that the abrupt changes the Earth underwent over a long period of time caused some species to go extinct. Cuvier's theory on extinction has met opposition from other notable natural scientists like Darwin and Charles Lyell. Unlike Cuvier, they didn't believe that extinction was a sudden process; they believed that like the Earth, animals collectively undergo gradual change as a species. This differed widely from Cuvier's theory, which seemed to propose that animal extinction was catastrophic. However, Cuvier's theory of extinction is still justified in the case of mass extinctions that occurred in the last 600 million years, when approximately half of all living species went completely extinct within a short geological span of two million years, due in part by volcanic eruptions, asteroids, and rapid fluctuations in sea level. At this time, new species rose and others fell, precipitating the arrival of human beings. Cuvier's early work demonstrated conclusively that extinction was indeed a credible natural global process. Cuvier's thinking on extinctions was influenced by his extensive readings in Greek and Latin literature; he gathered every ancient report known in his day relating to discoveries of petrified bones of remarkable size in the Mediterranean region. Influence on Cuvier's theory of extinction was his collection of specimens from the New World, many of them obtained from Native Americans. He also maintained an archive of Native American observations, legends, and interpretations of immense fossilized skeletal remains, sent to him by informants and friends in the Americas. He was impressed that most of the Native American accounts identified the enormous bones, teeth, and tusks as animals of the deep past that had been destroyed by catastrophe. Cuvier came to believe that most, if not all, the animal fossils he examined were remains of species that had become extinct. Near the end of his 1796 paper on living and fossil elephants, he said: Contrary to many natural scientists' beliefs at the time, Cuvier believed that animal extinction was not a product of anthropogenic causes. Instead, he proposed that humans were around long enough to indirectly maintain the fossilized records of ancient Earth. He also attempted to verify the water catastrophe by analyzing records of various cultural backgrounds. Though he found many accounts of the water catastrophe unclear, he did believe that such an event occurred at the brink of human history nonetheless. This led Cuvier to become an active proponent of the geological school of thought called catastrophism, which maintained that many of the geological features of the earth and the history of life could be explained by catastrophic events that had caused the extinction of many species of animals. Over the course of his career, Cuvier came to believe there had not been a single catastrophe, but several, resulting in a succession of different faunas. He wrote about these ideas many times, in particular, he discussed them in great detail in the preliminary discourse (an introduction) to a collection of his papers, Recherches sur les ossements fossiles de quadrupèdes (Researches on quadruped fossil bones), on quadruped fossils published in 1812. Cuvier's own explanation for such a catastrophic event is derived from two different sources, including those from Jean-André Deluc and Déodat de Dolomieu. The former proposed that the continents existing ten millennia ago collapsed, allowing the ocean floors to rise higher than the continental plates and become the continents that now exist today. The latter proposed that a massive tsunami hit the globe, leading to mass extinction. Whatever the case was, he believed that the deluge happened quite recently in human history. In fact, he believed that Earth's existence was limited and not as extended as many natural scientists, like Lamarck, believed it to be. Much of the evidence he used to support his catastrophist theories has been taken from his fossil records. He strongly suggested that the fossils he found were evidence of the world's first reptiles, followed chronologically by mammals and humans. Cuvier didn't wish to delve much into the causation of all the extinction and introduction of new animal species but rather focused on the sequential aspects of animal history on Earth. In a way, his chronological dating of Earth's history somewhat reflected Lamarck's transformationist theories. Cuvier also worked alongside Alexandre Brongniart in analyzing the Parisian rock cycle. Using stratigraphical methods, they were both able to extrapolate key information regarding Earth history from studying these rocks. These rocks contained remnants of molluscs, bones of mammals, and shells. From these findings, Cuvier and Brongniart concluded that many environmental changes occurred in quick catastrophes, though Earth itself was often placid for extended periods of time in between sudden disturbances. The 'Preliminary Discourse' became very well known and, unauthorized translations were made into English, German, and Italian (and in the case of those in English, not entirely accurately). In 1826, Cuvier published a revised version under the name, Discours sur les révolutions de la surface du globe (Discourse on the upheavals of the surface of the globe). After Cuvier's death, the catastrophic school of geological thought lost ground to uniformitarianism, as championed by Charles Lyell and others, which claimed that the geological features of the earth were best explained by currently observable forces, such as erosion and volcanism, acting gradually over an extended period of time. The increasing interest in the topic of mass extinction starting in the late twentieth century, however, has led to a resurgence of interest among historians of science and other scholars in this aspect of Cuvier's work. Cuvier collaborated for several years with Alexandre Brongniart, an instructor at the Paris mining school, to produce a monograph on the geology of the region around Paris. They published a preliminary version in 1808 and the final version was published in 1811. In this monograph, they identified characteristic fossils of different rock layers that they used to analyze the geological column, the ordered layers of sedimentary rock, of the Paris basin. They concluded that the layers had been laid down over an extended period during which there clearly had been faunal succession and that the area had been submerged under sea water at times and at other times under fresh water. Along with William Smith's work during the same period on a geological map of England, which also used characteristic fossils and the principle of faunal succession to correlate layers of sedimentary rock, the monograph helped establish the scientific discipline of stratigraphy. It was a major development in the history of paleontology and the history of geology. In 1800 and working only from a drawing, Cuvier was the first to correctly identify in print, a fossil found in Bavaria as a small flying reptile, which he named the Ptero-Dactyle in 1809, (later Latinized as Pterodactylus antiquus)—the first known member of the diverse order of pterosaurs. In 1808 Cuvier identified a fossil found in Maastricht as a giant marine lizard, the first known mosasaur. Cuvier speculated correctly that there had been a time when reptiles rather than mammals had been the dominant fauna. This speculation was confirmed over the two decades following his death by a series of spectacular finds, mostly by English geologists and fossil collectors such as Mary Anning, William Conybeare, William Buckland, and Gideon Mantell, who found and described the first ichthyosaurs, plesiosaurs, and dinosaurs. In a 1798 paper on the fossil remains of an animal found in some plaster quarries near Paris, Cuvier states what is known as the principle of the correlation of parts. He writes: This idea is referred to as Cuvier's principle of correlation of parts, which states that all organs in an animal's body are deeply interdependent. Species' existence relies on the way in which these organs interact. For example, a species whose digestive tract is best suited to digesting flesh but whose body is best suited to foraging for plants cannot survive. Thus in all species, the functional significance of each body part must be correlated to the others, or else the species cannot sustain itself. Cuvier believed that the power of his principle came in part from its ability to aid in the reconstruction of fossils. In most cases, fossils of quadrupeds were not found as complete, assembled skeletons, but rather as scattered pieces that needed to be put together by anatomists. To make matters worse, deposits often contained the fossilized remains of several species of animals mixed together. Anatomists reassembling these skeletons ran the risk of combining remains of different species, producing imaginary composite species. However, by examining the functional purpose of each bone and applying the principle of correlation of parts, Cuvier believed that this problem could be avoided. This principle's ability to aid in the reconstruction of fossils was also helpful to Cuvier's work in providing evidence in favour of extinction. The strongest evidence Cuvier could provide in favour of extinction would be to prove that the fossilized remains of an animal belonged to a species that no longer existed. By applying Cuvier's principle of correlation of parts, it would be easier to verify that a fossilized skeleton had been authentically reconstructed, thus validating any observations drawn from comparing it to skeletons of existing species. In addition to helping anatomists reconstruct fossilized remains, Cuvier believed that his principle also held enormous predictive power. For example, when he discovered a fossil that resembled a marsupial in the gypsum quarries of Montmartre, he correctly predicted that the fossil would contain bones commonly found in marsupials in its pelvis as well. Cuvier hoped that his principles of anatomy would provide the law-based framework that would elevate natural history to the truly scientific level occupied by physics and chemistry thanks to the laws established by Isaac Newton (1643 – 1727) and Antoine Lavoisier (1743 – 1794), respectively. He expressed confidence in the introduction to Le Règne Animal that someday anatomy would be expressed as laws as simple, mathematical, and predictive as Newton's laws of physics, and he viewed his principle as an important step in that direction. To him, the predictive capabilities of his principles demonstrated in his prediction of the existence of marsupial pelvic bones in the gypsum quarries of Montmartre demonstrated that these goals were not only in reach, but imminent. The principle of correlation of parts was also Cuvier's way of understanding function in a non-evolutionary context, without invoking a divine creator. In the same 1798 paper on the fossil remains of an animal found in plaster quarries near Paris, Cuvier emphasizes the predictive power of his principle, writing, Today comparative anatomy has reached such a point of perfection that, after inspecting a single bone, one can often determine the class, and sometimes even the genus of the animal to which it belonged, above all if that bone belonged to the head or the limbs ... This is because the number, direction, and shape of the bones that compose each part of an animal's body are always in a necessary relation to all the other parts, in such a way that—up to a point—one can infer the whole from any one of them and vice versa. Though Cuvier believed that his principle's major contribution was that it was a rational, mathematical way to reconstruct fossils and make predictions, in reality, it was difficult for Cuvier to use his principle. The functional significance of many body parts was still unknown at the time, and so relating those body parts to other body parts using Cuvier's principle was impossible. Though Cuvier was able to make accurate predictions about fossil finds, in practice, the accuracy of his predictions came not from application of his principle, but rather from his vast knowledge of comparative anatomy. However, despite Cuvier's exaggerations of the power of his principle, the basic concept is central to comparative anatomy and paleontology. Scientific work At the Paris Museum, Cuvier furthered his studies on the anatomical classification of animals. He believed that classification should be based on how organs collectively function, a concept he called functional integration. Cuvier reinforced the idea of subordinating less vital body parts to more critical organ systems as part of anatomical classification. He included these ideas in his 1817 book, The Animal Kingdom. In his anatomical studies, Cuvier believed function played a bigger role than form in the field of taxonomy. His scientific beliefs rested in the idea of the principles of the correlation of parts and of the conditions of existence. The former principle accounts for the connection between organ function and its practical use for an organism to survive. The latter principle emphasizes the animal's physiological function in relation to its surrounding environment. These findings were published in his scientific readings, including Leçons d'anatomie comparée (Lessons on Comparative Anatomy) between 1800 and 1805,[a] and The Animal Kingdom in 1817. Ultimately, Cuvier developed four embranchements, or branches, through which he classified animals based on his taxonomical and anatomical studies. He later performed groundbreaking work in classifying animals in vertebrate and invertebrate groups by subdividing each category. For instance, he proposed that the invertebrates could be segmented into three individual categories, including Mollusca, Radiata, and Articulata. He also articulated that species cannot move across these categories, a theory called transmutation. He reasoned that organisms cannot acquire or change their physical traits over time and still retain optimal survival. As a result, he often conflicted with Geoffroy Saint-Hilaire and Jean-Baptiste Lamarck's theories of transmutation. In 1798, Cuvier published his first independent work, the Tableau élémentaire de l'histoire naturelle des animaux, which was an abridgement of his course of lectures at the École du Pantheon and may be regarded as the foundation and first statement of his natural classification of the animal kingdom. Cuvier categorized snails, cockles, and cuttlefish into one category he called molluscs (Mollusca), an embranchment. Though he noted how all three of these animals were outwardly different in terms of shell shape and diet, he saw a noticeable pattern pertaining to their overall physical appearance. Cuvier began his intensive studies of molluscs during his time in Normandy – the first time he had ever seen the sea – and his papers on the so-called Mollusca began appearing as early as 1792. However, most of his memoirs on this branch were published in the Annales du museum between 1802 and 1815; they were subsequently collected as Mémoires pour servir à l'histoire et à l'anatomie des mollusques, published in one volume at Paris in 1817. Cuvier's researches on fish, begun in 1801, finally culminated in the publication of the Histoire naturelle des poissons, which contained descriptions of 5,000 species of fishes, and was a joint production with Achille Valenciennes. Cuvier's work on this project extended over the years 1828–1831. In palaeontology, Cuvier published a long list of memoirs, partly relating to the bones of extinct animals, and partly detailing the results of observations on the skeletons of living animals, specially examined with a view toward throwing light upon the structure and affinities of the fossil forms. Among living forms he published papers relating to the osteology of the Rhinoceros indicus, the tapir, Hyrax capensis, the hippopotamus, the sloths, the manatee, etc. He produced an even larger body of work on fossils, dealing with the extinct mammals of the Eocene beds of Montmartre and other localities near Paris, such as the Buttes Chaumont, the fossil species of hippopotamus, Palaeotherium, Anoplotherium, a marsupial (which he called Didelphys gypsorum), the Megalonyx, the Megatherium, the cave-hyena, the pterodactyl, the extinct species of rhinoceros, the cave bear, the mastodon, the extinct species of elephant, fossil species of manatee and seals, fossil forms of crocodilians, chelonians, fish, birds, etc. If his identification of fossil animals was dependent upon comparison with the osteology of extant animals whose anatomy was poorly known, Cuvier would often publish a thorough documentation of the relevant extant species' anatomy before publishing his analyses of the fossil specimens. The department of palaeontology dealing with the Mammalia may be said to have been essentially created and established by Cuvier. The results of Cuvier's principal palaeontological and geological investigations ultimately were given to the world in the form of two separate works: Recherches sur les ossemens fossiles de quadrupèdes (Paris, 1812; later editions in 1821 and 1825); and Discours sur les revolutions de la surface du globe (Paris, 1825). In this latter work he expounded a scientific theory of Catastrophism. Cuvier's most admired work was his Le Règne Animal. It appeared in four octavo volumes in 1817; a second edition in five volumes was brought out in 1829–1830. In this classic work, Cuvier presented the results of his life's research into the structure of living and fossil animals. With the exception of the section on insects, in which he was assisted by his friend Latreille, the whole of the work was his own. It was translated into English many times, often with substantial notes and supplementary material updating the book in accordance with the expansion of knowledge. Cuvier was a Protestant and a believer in monogenism, who held that all men descended from the biblical Adam, although his position usually was confused as polygenist. Some writers who have studied his racial work have dubbed his position as "quasi-polygenist", and most of his racial studies have influenced scientific racism. Cuvier believed there were three distinct races: the Caucasian (white), Mongolian (yellow), and the Ethiopian (black). Cuvier claimed that Adam and Eve were Caucasian, the original race of mankind. The other two races originated from survivors escaping in different directions after a major catastrophe hit the earth 5,000 years ago, with those survivors then living in complete isolation from each other. Cuvier categorized these divisions he identified into races according to his perception of the beauty or ugliness of their skulls and the quality of their civilizations. Cuvier's racial studies held the supposed features of polygenism, namely fixity of species; limits on environmental influence; unchanging underlying type; anatomical and cranial measurement differences in races; and physical and mental differences between distinct races. Alongside other French naturalists, Cuvier subjected Sarah Baartman, a South African Khokhoi woman exhibited in European freak shows as the "Hottentot Venus", to examinations. At the time that Cuvier interacted with Baartman, Baartman's "existence was really quite miserable and extraordinarily poor. Sara was literally [sic] treated like an animal." In 1815, while Baartman was very ill, Cuvier commissioned a nude painting of her. She died shortly afterward, aged 26. Following Baartman's death, Cuvier sought out and received permission to dissect her body, focusing on her genitalia, buttocks and skull shape. In his examination, Cuvier concluded that many of Baartman's features more closely resembled the anatomy of a monkey than a human. Her remains were displayed in the Musée de l'Homme in Paris until 1970, then were put into storage. Her remains were returned to South Africa in 2002. Taxa described by him Official and public work Apart from his own original investigations in zoology and paleontology Cuvier carried out a vast amount of work as perpetual secretary of the National Institute, and as an official connected with public education generally; and much of this work appeared ultimately in a published form. Thus, in 1808 he was placed by Napoleon upon the council of the Imperial University, and in this capacity he presided (in the years 1809, 1811, and 1813) over commissions charged to examine the state of the higher educational establishments in the districts beyond the Alps and the Rhine that had been annexed to France, and to report upon the means by which these could be affiliated with the central university. He published three separate reports on this subject. In his capacity, again, of perpetual secretary of the Institute, he not only prepared a number of éloges historiques on deceased members of the Academy of Sciences, but was also the author of a number of reports on the history of the physical and natural sciences, the most important of these being the Rapport historique sur le progrès des sciences physiques depuis 1789, published in 1810. Prior to the fall of Napoleon (1814) he had been admitted to the council of state, and his position remained unaffected by the restoration of the Bourbons. He was elected chancellor of the university, in which capacity he acted as interim president of the council of public instruction, while he also, as a Lutheran, superintended the faculty of Protestant theology. In 1819 he was appointed president of the committee of the interior, an office he retained until his death. In 1826 he was made grand officer of the Legion of Honour; he subsequently was appointed president of the council of state. He served as a member of the Académie des Inscriptions et Belles-Lettres from 1830 to his death. A member of the Doctrinaires, he was nominated to the ministry of the interior in the beginning of 1832. Commemorations Cuvier is commemorated in the naming of several animals; they include Cuvier's beaked whale (which he first thought to be extinct), Cuvier's gazelle, Cuvier's toucan, Cuvier's bichir, Cuvier's dwarf caiman, and Galeocerdo cuvier (tiger shark). Cuvier is commemorated in the scientific name of the following reptiles: Anolis cuvieri (a lizard from Puerto Rico), Bachia cuvieri (a synonym of Bachia alleni), and Oplurus cuvieri. The fish Hepsetus cuvieri, sometimes known as the African pike or Kafue pike characin, which is a predatory freshwater fish found in southern Africa was named after him. There also are some extinct animals named after Cuvier, such as the South American giant sloth Catonyx cuvieri. Cuvier Island in New Zealand was named after Cuvier by D'Urville. The professor of English Wayne Glausser argues at length that the Aubrey-Maturin series of 21 novels (1970–2004) by Patrick O'Brian make the character Stephen Maturin "an advocate of the neo-classical paradigm articulated .. by Georges Cuvier." Cuvier is referenced in Edgar Allan Poe's short story The Murders in the Rue Morgue as having written a description of the orangutan. Arthur Conan Doyle also refers to Cuvier in The Five Orange Pips, in which Sherlock Holmes compares Cuvier's methods to his own. There is a statue of Cuvier standing in front of the Hôtel de Ville in Montbéliard. Works Cuvier also collaborated on the Dictionnaire des sciences naturelles (61 volumes, 1816–1845) and on the Biographie universelle (45 volumes, 1843–18??) Taxa named in his honour See also References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Comparative_anatomy] | [TOKENS: 929] |
Contents Comparative anatomy Comparative anatomy is a study of similarities and differences in the anatomy of different species. It is closely related to evolutionary biology and phylogeny (the evolution of species). The science began in the classical era, continuing in the early modern period with work by Pierre Belon who noted the similarities of the skeletons of birds and humans. Comparative anatomy has provided evidence of common descent, and has assisted in the classification of animals. History The first specifically anatomical investigation separate from a surgical or medical procedure is associated by Alcmaeon of Croton. Leonardo da Vinci made notes for a planned anatomical treatise in which he intended to compare the hands of various animals including bears. Pierre Belon, a French naturalist born in 1517, conducted research and held discussions on dolphin embryos as well as the comparisons between the skeletons of birds to the skeletons of humans. His research led to modern comparative anatomy. Around the same time, Andreas Vesalius was also making some strides of his own. A young anatomist of Flemish descent made famous by a penchant for amazing charts, he was systematically investigating and correcting the anatomical knowledge of the Greek physician Galen. He noticed that many of Galen's observations were not even based on actual humans. Instead, they were based on other animals such as non-human apes, monkeys, and oxen. In fact, he entreated his students to do the following, in substitution for human skeletons, as cited by Edward Tyson : "If you can't happen to see any of these, dissect an Ape, carefully view each Bone, &c. ..." Then he advises what sort of Apes to make the choice of, as most resembling a Man : And conclude "One ought to know the Structure of all the Bones either in a Humane Body or in an Apes ; 'tis best in both ; and then to go to the Anatomy of the Muscles." Up until that point, Galen and his teachings had been the authority on human anatomy. The irony is that Galen himself had emphasized the fact that one should make one's own observations instead of using those of another, but this advice was lost during the numerous translations of his work. As Vesalius began to uncover these mistakes, other physicians of the time began to trust their own observations more than those of Galen. An interesting observation made by some of these physicians was the presence of homologous structures in a wide variety of animals, even including humans. These observations were later used by Darwin as he formed his theory of Natural Selection. Edward Tyson is regarded as the founder of modern comparative anatomy. He is credited with determining that whales and dolphins are, in fact, mammals. Also, he concluded that chimpanzees are more similar to humans than to monkeys because of their arms. Marco Aurelio Severino also compared various animals, including birds, in his Zootomia democritaea, one of the first works of comparative anatomy. In the 18th and 19th century, great anatomists like George Cuvier, Richard Owen and Thomas Henry Huxley revolutionized our understanding of the basic build and systematics of vertebrates, laying the foundation for Charles Darwin's work on evolution. An example of a 20th-century comparative anatomist is Victor Negus, who worked on the structure and evolution of the larynx. Until the advent of genetic techniques like DNA sequencing, comparative anatomy together with embryology were the primary tools for understanding phylogeny, as exemplified by the work of Alfred Romer.[citation needed] Concepts Two major concepts of comparative anatomy are: Uses Comparative anatomy has long served as evidence for evolution, now joined in that role by comparative genomics; it indicates that organisms share a common ancestor. It also assists scientists in classifying organisms based on similar characteristics of their anatomical structures. A common example of comparative anatomy is the similar bone structures in forelimbs of cats, whales, bats, and humans. All of these appendages consist of the same basic parts; yet, they serve completely different functions. The skeletal parts which form a structure used for swimming, such as a fin, would not be ideal for forming a wing, which is better suited for flight. One explanation for the forelimbs' similar composition is descent with modification. Through random mutations and natural selection, each organism's anatomical structures gradually adapted to suit their respective habitats. The rules for development of special characteristics which differ significantly from general homology were listed by Karl Ernst von Baer as the laws now named after him. See also References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Mars#cite_note-Petropoulos_Longuski_Bonfiglio_2000_pp._776–783-202] | [TOKENS: 11899] |
Contents Mars Mars is the fourth planet from the Sun. It is also known as the "Red Planet", for its orange-red appearance. Mars is a desert-like rocky planet with a tenuous atmosphere that is primarily carbon dioxide (CO2). At the average surface level the atmospheric pressure is a few thousandths of Earth's, atmospheric temperature ranges from −153 to 20 °C (−243 to 68 °F), and cosmic radiation is high. Mars retains some water, in the ground as well as thinly in the atmosphere, forming cirrus clouds, fog, frost, larger polar regions of permafrost and ice caps (with seasonal CO2 snow), but no bodies of liquid surface water. Its surface gravity is roughly a third of Earth's or double that of the Moon. Its diameter, 6,779 km (4,212 mi), is about half the Earth's, or twice the Moon's, and its surface area is the size of all the dry land of Earth. Fine dust is prevalent across the surface and the atmosphere, being picked up and spread at the low Martian gravity even by the weak wind of the tenuous atmosphere. The terrain of Mars roughly follows a north-south divide, the Martian dichotomy, with the northern hemisphere mainly consisting of relatively flat, low lying plains, and the southern hemisphere of cratered highlands. Geologically, the planet is fairly active with marsquakes trembling underneath the ground, but also hosts many enormous volcanoes that are extinct (the tallest is Olympus Mons, 21.9 km or 13.6 mi tall), as well as one of the largest canyons in the Solar System (Valles Marineris, 4,000 km or 2,500 mi long). Mars has two natural satellites that are small and irregular in shape: Phobos and Deimos. With a significant axial tilt of 25 degrees, Mars experiences seasons, like Earth (which has an axial tilt of 23.5 degrees). A Martian solar year is equal to 1.88 Earth years (687 Earth days), a Martian solar day (sol) is equal to 24.6 hours. Mars formed along with the other planets approximately 4.5 billion years ago. During the martian Noachian period (4.5 to 3.5 billion years ago), its surface was marked by meteor impacts, valley formation, erosion, the possible presence of water oceans and the loss of its magnetosphere. The Hesperian period (beginning 3.5 billion years ago and ending 3.3–2.9 billion years ago) was dominated by widespread volcanic activity and flooding that carved immense outflow channels. The Amazonian period, which continues to the present, is the currently dominating and remaining influence on geological processes. Because of Mars's geological history, the possibility of past or present life on Mars remains an area of active scientific investigation, with some possible traces needing further examination. Being visible with the naked eye in Earth's sky as a red wandering star, Mars has been observed throughout history, acquiring diverse associations in different cultures. In 1963 the first flight to Mars took place with Mars 1, but communication was lost en route. The first successful flyby exploration of Mars was conducted in 1965 with Mariner 4. In 1971 Mariner 9 entered orbit around Mars, being the first spacecraft to orbit any body other than the Moon, Sun or Earth; following in the same year were the first uncontrolled impact (Mars 2) and first successful landing (Mars 3) on Mars. Probes have been active on Mars continuously since 1997. At times, more than ten probes have simultaneously operated in orbit or on the surface, more than at any other planet beyond Earth. Mars is an often proposed target for future crewed exploration missions, though no such mission is currently planned. Natural history Scientists have theorized that during the Solar System's formation, Mars was created as the result of a random process of run-away accretion of material from the protoplanetary disk that orbited the Sun. Mars has many distinctive chemical features caused by its position in the Solar System. Elements with comparatively low boiling points, such as chlorine, phosphorus, and sulfur, are much more common on Mars than on Earth; these elements were probably pushed outward by the young Sun's energetic solar wind. After the formation of the planets, the inner Solar System may have been subjected to the so-called Late Heavy Bombardment. About 60% of the surface of Mars shows a record of impacts from that era, whereas much of the remaining surface is probably underlain by immense impact basins caused by those events. However, more recent modeling has disputed the existence of the Late Heavy Bombardment. There is evidence of an enormous impact basin in the Northern Hemisphere of Mars, spanning 10,600 by 8,500 kilometres (6,600 by 5,300 mi), or roughly four times the size of the Moon's South Pole–Aitken basin, which would be the largest impact basin yet discovered if confirmed. It has been hypothesized that the basin was formed when Mars was struck by a Pluto-sized body about four billion years ago. The event, thought to be the cause of the Martian hemispheric dichotomy, created the smooth Borealis basin that covers 40% of the planet. A 2023 study shows evidence, based on the orbital inclination of Deimos (a small moon of Mars), that Mars may once have had a ring system 3.5 billion years to 4 billion years ago. This ring system may have been formed from a moon, 20 times more massive than Phobos, orbiting Mars billions of years ago; and Phobos would be a remnant of that ring. Epochs: The geological history of Mars can be split into many periods, but the following are the three primary periods: Geological activity is still taking place on Mars. The Athabasca Valles is home to sheet-like lava flows created about 200 million years ago. Water flows in the grabens called the Cerberus Fossae occurred less than 20 million years ago, indicating equally recent volcanic intrusions. The Mars Reconnaissance Orbiter has captured images of avalanches. Physical characteristics Mars is approximately half the diameter of Earth or twice that of the Moon, with a surface area only slightly less than the total area of Earth's dry land. Mars is less dense than Earth, having about 15% of Earth's volume and 11% of Earth's mass, resulting in about 38% of Earth's surface gravity. Mars is the only presently known example of a desert planet, a rocky planet with a surface akin to that of Earth's deserts. The red-orange appearance of the Martian surface is caused by iron(III) oxide (nanophase Fe2O3) and the iron(III) oxide-hydroxide mineral goethite. It can look like butterscotch; other common surface colors include golden, brown, tan, and greenish, depending on the minerals present. Like Earth, Mars is differentiated into a dense metallic core overlaid by less dense rocky layers. The outermost layer is the crust, which is on average about 42–56 kilometres (26–35 mi) thick, with a minimum thickness of 6 kilometres (3.7 mi) in Isidis Planitia, and a maximum thickness of 117 kilometres (73 mi) in the southern Tharsis plateau. For comparison, Earth's crust averages 27.3 ± 4.8 km in thickness. The most abundant elements in the Martian crust are silicon, oxygen, iron, magnesium, aluminum, calcium, and potassium. Mars is confirmed to be seismically active; in 2019, it was reported that InSight had detected and recorded over 450 marsquakes and related events. Beneath the crust is a silicate mantle responsible for many of the tectonic and volcanic features on the planet's surface. The upper Martian mantle is a low-velocity zone, where the velocity of seismic waves is lower than surrounding depth intervals. The mantle appears to be rigid down to the depth of about 250 km, giving Mars a very thick lithosphere compared to Earth. Below this the mantle gradually becomes more ductile, and the seismic wave velocity starts to grow again. The Martian mantle does not appear to have a thermally insulating layer analogous to Earth's lower mantle; instead, below 1050 km in depth, it becomes mineralogically similar to Earth's transition zone. At the bottom of the mantle lies a basal liquid silicate layer approximately 150–180 km thick. The Martian mantle appears to be highly heterogenous, with dense fragments up to 4 km across, likely injected deep into the planet by colossal impacts ~4.5 billion years ago; high-frequency waves from eight marsquakes slowed as they passed these localized regions, and modeling indicates the heterogeneities are compositionally distinct debris preserved because Mars lacks plate tectonics and has a sluggishly convecting interior that prevents complete homogenization. Mars's iron and nickel core is at least partially molten, and may have a solid inner core. It is around half of Mars's radius, approximately 1650–1675 km, and is enriched in light elements such as sulfur, oxygen, carbon, and hydrogen. The temperature of the core is estimated to be 2000–2400 K, compared to 5400–6230 K for Earth's solid inner core. In 2025, based on data from the InSight lander, a group of researchers reported the detection of a solid inner core 613 kilometres (381 mi) ± 67 kilometres (42 mi) in radius. Mars is a terrestrial planet with a surface that consists of minerals containing silicon and oxygen, metals, and other elements that typically make up rock. The Martian surface is primarily composed of tholeiitic basalt, although parts are more silica-rich than typical basalt and may be similar to andesitic rocks on Earth, or silica glass. Regions of low albedo suggest concentrations of plagioclase feldspar, with northern low albedo regions displaying higher than normal concentrations of sheet silicates and high-silicon glass. Parts of the southern highlands include detectable amounts of high-calcium pyroxenes. Localized concentrations of hematite and olivine have been found. Much of the surface is deeply covered by finely grained iron(III) oxide dust. The Phoenix lander returned data showing Martian soil to be slightly alkaline and containing elements such as magnesium, sodium, potassium and chlorine. These nutrients are found in soils on Earth, and are necessary for plant growth. Experiments performed by the lander showed that the Martian soil has a basic pH of 7.7, and contains 0.6% perchlorate by weight, concentrations that are toxic to humans. Streaks are common across Mars and new ones appear frequently on steep slopes of craters, troughs, and valleys. The streaks are dark at first and get lighter with age. The streaks can start in a tiny area, then spread out for hundreds of metres. They have been seen to follow the edges of boulders and other obstacles in their path. The commonly accepted hypotheses include that they are dark underlying layers of soil revealed after avalanches of bright dust or dust devils. Several other explanations have been put forward, including those that involve water or even the growth of organisms. Environmental radiation levels on the surface are on average 0.64 millisieverts of radiation per day, and significantly less than the radiation of 1.84 millisieverts per day or 22 millirads per day during the flight to and from Mars. For comparison the radiation levels in low Earth orbit, where Earth's space stations orbit, are around 0.5 millisieverts of radiation per day. Hellas Planitia has the lowest surface radiation at about 0.342 millisieverts per day, featuring lava tubes southwest of Hadriacus Mons with potentially levels as low as 0.064 millisieverts per day, comparable to radiation levels during flights on Earth. Although Mars has no evidence of a structured global magnetic field, observations show that parts of the planet's crust have been magnetized, suggesting that alternating polarity reversals of its dipole field have occurred in the past. This paleomagnetism of magnetically susceptible minerals is similar to the alternating bands found on Earth's ocean floors. One hypothesis, published in 1999 and re-examined in October 2005 (with the help of the Mars Global Surveyor), is that these bands suggest plate tectonic activity on Mars four billion years ago, before the planetary dynamo ceased to function and the planet's magnetic field faded. Geography and features Although better remembered for mapping the Moon, Johann Heinrich von Mädler and Wilhelm Beer were the first areographers. They began by establishing that most of Mars's surface features were permanent and by more precisely determining the planet's rotation period. In 1840, Mädler combined ten years of observations and drew the first map of Mars. Features on Mars are named from a variety of sources. Albedo features are named for classical mythology. Craters larger than roughly 50 km are named for deceased scientists and writers and others who have contributed to the study of Mars. Smaller craters are named for towns and villages of the world with populations of less than 100,000. Large valleys are named for the word "Mars" or "star" in various languages; smaller valleys are named for rivers. Large albedo features retain many of the older names but are often updated to reflect new knowledge of the nature of the features. For example, Nix Olympica (the snows of Olympus) has become Olympus Mons (Mount Olympus). The surface of Mars as seen from Earth is divided into two kinds of areas, with differing albedo. The paler plains covered with dust and sand rich in reddish iron oxides were once thought of as Martian "continents" and given names like Arabia Terra (land of Arabia) or Amazonis Planitia (Amazonian plain). The dark features were thought to be seas, hence their names Mare Erythraeum, Mare Sirenum and Aurorae Sinus. The largest dark feature seen from Earth is Syrtis Major Planum. The permanent northern polar ice cap is named Planum Boreum. The southern cap is called Planum Australe. Mars's equator is defined by its rotation, but the location of its Prime Meridian was specified, as was Earth's (at Greenwich), by choice of an arbitrary point; Mädler and Beer selected a line for their first maps of Mars in 1830. After the spacecraft Mariner 9 provided extensive imagery of Mars in 1972, a small crater (later called Airy-0), located in the Sinus Meridiani ("Middle Bay" or "Meridian Bay"), was chosen by Merton E. Davies, Harold Masursky, and Gérard de Vaucouleurs for the definition of 0.0° longitude to coincide with the original selection. Because Mars has no oceans, and hence no "sea level", a zero-elevation surface had to be selected as a reference level; this is called the areoid of Mars, analogous to the terrestrial geoid. Zero altitude was defined by the height at which there is 610.5 Pa (6.105 mbar) of atmospheric pressure. This pressure corresponds to the triple point of water, and it is about 0.6% of the sea level surface pressure on Earth (0.006 atm). For mapping purposes, the United States Geological Survey divides the surface of Mars into thirty cartographic quadrangles, each named for a classical albedo feature it contains. In April 2023, The New York Times reported an updated global map of Mars based on images from the Hope spacecraft. A related, but much more detailed, global Mars map was released by NASA on 16 April 2023. The vast upland region Tharsis contains several massive volcanoes, which include the shield volcano Olympus Mons. The edifice is over 600 km (370 mi) wide. Because the mountain is so large, with complex structure at its edges, giving a definite height to it is difficult. Its local relief, from the foot of the cliffs which form its northwest margin to its peak, is over 21 km (13 mi), a little over twice the height of Mauna Kea as measured from its base on the ocean floor. The total elevation change from the plains of Amazonis Planitia, over 1,000 km (620 mi) to the northwest, to the summit approaches 26 km (16 mi), roughly three times the height of Mount Everest, which in comparison stands at just over 8.8 kilometres (5.5 mi). Consequently, Olympus Mons is either the tallest or second-tallest mountain in the Solar System; the only known mountain which might be taller is the Rheasilvia peak on the asteroid Vesta, at 20–25 km (12–16 mi). The dichotomy of Martian topography is striking: northern plains flattened by lava flows contrast with the southern highlands, pitted and cratered by ancient impacts. It is possible that, four billion years ago, the Northern Hemisphere of Mars was struck by an object one-tenth to two-thirds the size of Earth's Moon. If this is the case, the Northern Hemisphere of Mars would be the site of an impact crater 10,600 by 8,500 kilometres (6,600 by 5,300 mi) in size, or roughly the area of Europe, Asia, and Australia combined, surpassing Utopia Planitia and the Moon's South Pole–Aitken basin as the largest impact crater in the Solar System. Mars is scarred by 43,000 impact craters with a diameter of 5 kilometres (3.1 mi) or greater. The largest exposed crater is Hellas, which is 2,300 kilometres (1,400 mi) wide and 7,000 metres (23,000 ft) deep, and is a light albedo feature clearly visible from Earth. There are other notable impact features, such as Argyre, which is around 1,800 kilometres (1,100 mi) in diameter, and Isidis, which is around 1,500 kilometres (930 mi) in diameter. Due to the smaller mass and size of Mars, the probability of an object colliding with the planet is about half that of Earth. Mars is located closer to the asteroid belt, so it has an increased chance of being struck by materials from that source. Mars is more likely to be struck by short-period comets, i.e., those that lie within the orbit of Jupiter. Martian craters can[discuss] have a morphology that suggests the ground became wet after the meteor impact. The large canyon, Valles Marineris (Latin for 'Mariner Valleys, also known as Agathodaemon in the old canal maps), has a length of 4,000 kilometres (2,500 mi) and a depth of up to 7 kilometres (4.3 mi). The length of Valles Marineris is equivalent to the length of Europe and extends across one-fifth the circumference of Mars. By comparison, the Grand Canyon on Earth is only 446 kilometres (277 mi) long and nearly 2 kilometres (1.2 mi) deep. Valles Marineris was formed due to the swelling of the Tharsis area, which caused the crust in the area of Valles Marineris to collapse. In 2012, it was proposed that Valles Marineris is not just a graben, but a plate boundary where 150 kilometres (93 mi) of transverse motion has occurred, making Mars a planet with possibly a two-tectonic plate arrangement. Images from the Thermal Emission Imaging System (THEMIS) aboard NASA's Mars Odyssey orbiter have revealed seven possible cave entrances on the flanks of the volcano Arsia Mons. The caves, named after loved ones of their discoverers, are collectively known as the "seven sisters". Cave entrances measure from 100 to 252 metres (328 to 827 ft) wide and they are estimated to be at least 73 to 96 metres (240 to 315 ft) deep. Because light does not reach the floor of most of the caves, they may extend much deeper than these lower estimates and widen below the surface. "Dena" is the only exception; its floor is visible and was measured to be 130 metres (430 ft) deep. The interiors of these caverns may be protected from micrometeoroids, UV radiation, solar flares and high energy particles that bombard the planet's surface. Martian geysers (or CO2 jets) are putative sites of small gas and dust eruptions that occur in the south polar region of Mars during the spring thaw. "Dark dune spots" and "spiders" – or araneiforms – are the two most visible types of features ascribed to these eruptions. Similarly sized dust will settle from the thinner Martian atmosphere sooner than it would on Earth. For example, the dust suspended by the 2001 global dust storms on Mars only remained in the Martian atmosphere for 0.6 years, while the dust from Mount Pinatubo took about two years to settle. However, under current Martian conditions, the mass movements involved are generally much smaller than on Earth. Even the 2001 global dust storms on Mars moved only the equivalent of a very thin dust layer – about 3 μm thick if deposited with uniform thickness between 58° north and south of the equator. Dust deposition at the two rover sites has proceeded at a rate of about the thickness of a grain every 100 sols. Atmosphere Mars lost its magnetosphere 4 billion years ago, possibly because of numerous asteroid strikes, so the solar wind interacts directly with the Martian ionosphere, lowering the atmospheric density by stripping away atoms from the outer layer. Both Mars Global Surveyor and Mars Express have detected ionized atmospheric particles trailing off into space behind Mars, and this atmospheric loss is being studied by the MAVEN orbiter. Compared to Earth, the atmosphere of Mars is quite rarefied. Atmospheric pressure on the surface today ranges from a low of 30 Pa (0.0044 psi) on Olympus Mons to over 1,155 Pa (0.1675 psi) in Hellas Planitia, with a mean pressure at the surface level of 600 Pa (0.087 psi). The highest atmospheric density on Mars is equal to that found 35 kilometres (22 mi) above Earth's surface. The resulting mean surface pressure is only 0.6% of Earth's 101.3 kPa (14.69 psi). The scale height of the atmosphere is about 10.8 kilometres (6.7 mi), which is higher than Earth's 6 kilometres (3.7 mi), because the surface gravity of Mars is only about 38% of Earth's. The atmosphere of Mars consists of about 96% carbon dioxide, 1.93% argon and 1.89% nitrogen along with traces of oxygen and water. The atmosphere is quite dusty, containing particulates about 1.5 μm in diameter which give the Martian sky a tawny color when seen from the surface. It may take on a pink hue due to iron oxide particles suspended in it. Despite repeated detections of methane on Mars, there is no scientific consensus as to its origin. One suggestion is that methane exists on Mars and that its concentration fluctuates seasonally. The existence of methane could be produced by non-biological process such as serpentinization involving water, carbon dioxide, and the mineral olivine, which is known to be common on Mars, or by Martian life. Compared to Earth, its higher concentration of atmospheric CO2 and lower surface pressure may be why sound is attenuated more on Mars, where natural sources are rare apart from the wind. Using acoustic recordings collected by the Perseverance rover, researchers concluded that the speed of sound there is approximately 240 m/s for frequencies below 240 Hz, and 250 m/s for those above. Auroras have been detected on Mars. Because Mars lacks a global magnetic field, the types and distribution of auroras there differ from those on Earth; rather than being mostly restricted to polar regions as is the case on Earth, a Martian aurora can encompass the planet. In September 2017, NASA reported radiation levels on the surface of the planet Mars were temporarily doubled, and were associated with an aurora 25 times brighter than any observed earlier, due to a massive, and unexpected, solar storm in the middle of the month. Mars has seasons, alternating between its northern and southern hemispheres, similar to on Earth. Additionally the orbit of Mars has, compared to Earth's, a large eccentricity and approaches perihelion when it is summer in its southern hemisphere and winter in its northern, and aphelion when it is winter in its southern hemisphere and summer in its northern. As a result, the seasons in its southern hemisphere are more extreme and the seasons in its northern are milder than would otherwise be the case. The summer temperatures in the south can be warmer than the equivalent summer temperatures in the north by up to 30 °C (54 °F). Martian surface temperatures vary from lows of about −110 °C (−166 °F) to highs of up to 35 °C (95 °F) in equatorial summer. The wide range in temperatures is due to the thin atmosphere which cannot store much solar heat, the low atmospheric pressure (about 1% that of the atmosphere of Earth), and the low thermal inertia of Martian soil. The planet is 1.52 times as far from the Sun as Earth, resulting in just 43% of the amount of sunlight. Mars has the largest dust storms in the Solar System, reaching speeds of over 160 km/h (100 mph). These can vary from a storm over a small area, to gigantic storms that cover the entire planet. They tend to occur when Mars is closest to the Sun, and have been shown to increase global temperature. Seasons also produce dry ice covering polar ice caps. Hydrology While Mars contains water in larger amounts, most of it is dust covered water ice at the Martian polar ice caps. The volume of water ice in the south polar ice cap, if melted, would be enough to cover most of the surface of the planet with a depth of 11 metres (36 ft). Water in its liquid form cannot persist on the surface due to Mars's low atmospheric pressure, which is less than 1% that of Earth. Only at the lowest of elevations are the pressure and temperature high enough for liquid water to exist for short periods. Although little water is present in the atmosphere, there is enough to produce clouds of water ice and different cases of snow and frost, often mixed with snow of carbon dioxide dry ice. Landforms visible on Mars strongly suggest that liquid water has existed on the planet's surface. Huge linear swathes of scoured ground, known as outflow channels, cut across the surface in about 25 places. These are thought to be a record of erosion caused by the catastrophic release of water from subsurface aquifers, though some of these structures have been hypothesized to result from the action of glaciers or lava. One of the larger examples, Ma'adim Vallis, is 700 kilometres (430 mi) long, much greater than the Grand Canyon, with a width of 20 kilometres (12 mi) and a depth of 2 kilometres (1.2 mi) in places. It is thought to have been carved by flowing water early in Mars's history. The youngest of these channels is thought to have formed only a few million years ago. Elsewhere, particularly on the oldest areas of the Martian surface, finer-scale, dendritic networks of valleys are spread across significant proportions of the landscape. Features of these valleys and their distribution strongly imply that they were carved by runoff resulting from precipitation in early Mars history. Subsurface water flow and groundwater sapping may play important subsidiary roles in some networks, but precipitation was probably the root cause of the incision in almost all cases. Along craters and canyon walls, there are thousands of features that appear similar to terrestrial gullies. The gullies tend to be in the highlands of the Southern Hemisphere and face the Equator; all are poleward of 30° latitude. A number of authors have suggested that their formation process involves liquid water, probably from melting ice, although others have argued for formation mechanisms involving carbon dioxide frost or the movement of dry dust. No partially degraded gullies have formed by weathering and no superimposed impact craters have been observed, indicating that these are young features, possibly still active. Other geological features, such as deltas and alluvial fans preserved in craters, are further evidence for warmer, wetter conditions at an interval or intervals in earlier Mars history. Such conditions necessarily require the widespread presence of crater lakes across a large proportion of the surface, for which there is independent mineralogical, sedimentological and geomorphological evidence. Further evidence that liquid water once existed on the surface of Mars comes from the detection of specific minerals such as hematite and goethite, both of which sometimes form in the presence of water. The chemical signature of water vapor on Mars was first unequivocally demonstrated in 1963 by spectroscopy using an Earth-based telescope. In 2004, Opportunity detected the mineral jarosite. This forms only in the presence of acidic water, showing that water once existed on Mars. The Spirit rover found concentrated deposits of silica in 2007 that indicated wet conditions in the past, and in December 2011, the mineral gypsum, which also forms in the presence of water, was found on the surface by NASA's Mars rover Opportunity. It is estimated that the amount of water in the upper mantle of Mars, represented by hydroxyl ions contained within Martian minerals, is equal to or greater than that of Earth at 50–300 parts per million of water, which is enough to cover the entire planet to a depth of 200–1,000 metres (660–3,280 ft). On 18 March 2013, NASA reported evidence from instruments on the Curiosity rover of mineral hydration, likely hydrated calcium sulfate, in several rock samples including the broken fragments of "Tintina" rock and "Sutton Inlier" rock as well as in veins and nodules in other rocks like "Knorr" rock and "Wernicke" rock. Analysis using the rover's DAN instrument provided evidence of subsurface water, amounting to as much as 4% water content, down to a depth of 60 centimetres (24 in), during the rover's traverse from the Bradbury Landing site to the Yellowknife Bay area in the Glenelg terrain. In September 2015, NASA announced that they had found strong evidence of hydrated brine flows in recurring slope lineae, based on spectrometer readings of the darkened areas of slopes. These streaks flow downhill in Martian summer, when the temperature is above −23 °C, and freeze at lower temperatures. These observations supported earlier hypotheses, based on timing of formation and their rate of growth, that these dark streaks resulted from water flowing just below the surface. However, later work suggested that the lineae may be dry, granular flows instead, with at most a limited role for water in initiating the process. A definitive conclusion about the presence, extent, and role of liquid water on the Martian surface remains elusive. Researchers suspect much of the low northern plains of the planet were covered with an ocean hundreds of meters deep, though this theory remains controversial. In March 2015, scientists stated that such an ocean might have been the size of Earth's Arctic Ocean. This finding was derived from the ratio of protium to deuterium in the modern Martian atmosphere compared to that ratio on Earth. The amount of Martian deuterium (D/H = 9.3 ± 1.7 10−4) is five to seven times the amount on Earth (D/H = 1.56 10−4), suggesting that ancient Mars had significantly higher levels of water. Results from the Curiosity rover had previously found a high ratio of deuterium in Gale Crater, though not significantly high enough to suggest the former presence of an ocean. Other scientists caution that these results have not been confirmed, and point out that Martian climate models have not yet shown that the planet was warm enough in the past to support bodies of liquid water. Near the northern polar cap is the 81.4 kilometres (50.6 mi) wide Korolev Crater, which the Mars Express orbiter found to be filled with approximately 2,200 cubic kilometres (530 cu mi) of water ice. In November 2016, NASA reported finding a large amount of underground ice in the Utopia Planitia region. The volume of water detected has been estimated to be equivalent to the volume of water in Lake Superior (which is 12,100 cubic kilometers). During observations from 2018 through 2021, the ExoMars Trace Gas Orbiter spotted indications of water, probably subsurface ice, in the Valles Marineris canyon system. Orbital motion Mars's average distance from the Sun is roughly 230 million km (143 million mi), and its orbital period is 687 (Earth) days. The solar day (or sol) on Mars is only slightly longer than an Earth day: 24 hours, 39 minutes, and 35.244 seconds. A Martian year is equal to 1.8809 Earth years, or 1 year, 320 days, and 18.2 hours. The gravitational potential difference and thus the delta-v needed to transfer between Mars and Earth is the second lowest for Earth. The axial tilt of Mars is 25.19° relative to its orbital plane, which is similar to the axial tilt of Earth. As a result, Mars has seasons like Earth, though on Mars they are nearly twice as long because its orbital period is that much longer. In the present day, the orientation of the north pole of Mars is close to the star Deneb. Mars has a relatively pronounced orbital eccentricity of about 0.09; of the seven other planets in the Solar System, only Mercury has a larger orbital eccentricity. It is known that in the past, Mars has had a much more circular orbit. At one point, 1.35 million Earth years ago, Mars had an eccentricity of roughly 0.002, much less than that of Earth today. Mars's cycle of eccentricity is 96,000 Earth years compared to Earth's cycle of 100,000 years. Mars has its closest approach to Earth (opposition) in a synodic period of 779.94 days. It should not be confused with Mars conjunction, where the Earth and Mars are at opposite sides of the Solar System and form a straight line crossing the Sun. The average time between the successive oppositions of Mars, its synodic period, is 780 days; but the number of days between successive oppositions can range from 764 to 812. The distance at close approach varies between about 54 and 103 million km (34 and 64 million mi) due to the planets' elliptical orbits, which causes comparable variation in angular size. At their furthest Mars and Earth can be as far as 401 million km (249 million mi) apart. Mars comes into opposition from Earth every 2.1 years. The planets come into opposition near Mars's perihelion in 2003, 2018 and 2035, with the 2020 and 2033 events being particularly close to perihelic opposition. The mean apparent magnitude of Mars is +0.71 with a standard deviation of 1.05. Because the orbit of Mars is eccentric, the magnitude at opposition from the Sun can range from about −3.0 to −1.4. The minimum brightness is magnitude +1.86 when the planet is near aphelion and in conjunction with the Sun. At its brightest, Mars (along with Jupiter) is second only to Venus in apparent brightness. Mars usually appears distinctly yellow, orange, or red. When farthest away from Earth, it is more than seven times farther away than when it is closest. Mars is usually close enough for particularly good viewing once or twice at 15-year or 17-year intervals. Optical ground-based telescopes are typically limited to resolving features about 300 kilometres (190 mi) across when Earth and Mars are closest because of Earth's atmosphere. As Mars approaches opposition, it begins a period of retrograde motion, which means it will appear to move backwards in a looping curve with respect to the background stars. This retrograde motion lasts for about 72 days, and Mars reaches its peak apparent brightness in the middle of this interval. Moons Mars has two relatively small (compared to Earth's) natural moons, Phobos (about 22 km (14 mi) in diameter) and Deimos (about 12 km (7.5 mi) in diameter), which orbit at 9,376 km (5,826 mi) and 23,460 km (14,580 mi) around the planet. The origin of both moons is unclear, although a popular theory states that they were asteroids captured into Martian orbit. Both satellites were discovered in 1877 by Asaph Hall and were named after the characters Phobos (the deity of panic and fear) and Deimos (the deity of terror and dread), twins from Greek mythology who accompanied their father Ares, god of war, into battle. Mars was the Roman equivalent to Ares. In modern Greek, the planet retains its ancient name Ares (Aris: Άρης). From the surface of Mars, the motions of Phobos and Deimos appear different from that of the Earth's satellite, the Moon. Phobos rises in the west, sets in the east, and rises again in just 11 hours. Deimos, being only just outside synchronous orbit – where the orbital period would match the planet's period of rotation – rises as expected in the east, but slowly. Because the orbit of Phobos is below a synchronous altitude, tidal forces from Mars are gradually lowering its orbit. In about 50 million years, it could either crash into Mars's surface or break up into a ring structure around the planet. The origin of the two satellites is not well understood. Their low albedo and carbonaceous chondrite composition have been regarded as similar to asteroids, supporting a capture theory. The unstable orbit of Phobos would seem to point toward a relatively recent capture. But both have circular orbits near the equator, which is unusual for captured objects, and the required capture dynamics are complex. Accretion early in the history of Mars is plausible, but would not account for a composition resembling asteroids rather than Mars itself, if that is confirmed. Mars may have yet-undiscovered moons, smaller than 50 to 100 metres (160 to 330 ft) in diameter, and a dust ring is predicted to exist between Phobos and Deimos. A third possibility for their origin as satellites of Mars is the involvement of a third body or a type of impact disruption. More-recent lines of evidence for Phobos having a highly porous interior, and suggesting a composition containing mainly phyllosilicates and other minerals known from Mars, point toward an origin of Phobos from material ejected by an impact on Mars that reaccreted in Martian orbit, similar to the prevailing theory for the origin of Earth's satellite. Although the visible and near-infrared (VNIR) spectra of the moons of Mars resemble those of outer-belt asteroids, the thermal infrared spectra of Phobos are reported to be inconsistent with chondrites of any class. It is also possible that Phobos and Deimos were fragments of an older moon, formed by debris from a large impact on Mars, and then destroyed by a more recent impact upon the satellite. More recently, a study conducted by a team of researchers from multiple countries suggests that a lost moon, at least fifteen times the size of Phobos, may have existed in the past. By analyzing rocks which point to tidal processes on the planet, it is possible that these tides may have been regulated by a past moon. Human observations and exploration The history of observations of Mars is marked by oppositions of Mars when the planet is closest to Earth and hence is most easily visible, which occur every couple of years. Even more notable are the perihelic oppositions of Mars, which are distinguished because Mars is close to perihelion, making it even closer to Earth. The ancient Sumerians named Mars Nergal, the god of war and plague. During Sumerian times, Nergal was a minor deity of little significance, but, during later times, his main cult center was the city of Nineveh. In Mesopotamian texts, Mars is referred to as the "star of judgement of the fate of the dead". The existence of Mars as a wandering object in the night sky was also recorded by the ancient Egyptian astronomers and, by 1534 BCE, they were familiar with the retrograde motion of the planet. By the period of the Neo-Babylonian Empire, the Babylonian astronomers were making regular records of the positions of the planets and systematic observations of their behavior. For Mars, they knew that the planet made 37 synodic periods, or 42 circuits of the zodiac, every 79 years. They invented arithmetic methods for making minor corrections to the predicted positions of the planets. In Ancient Greece, the planet was known as Πυρόεις. Commonly, the Greek name for the planet now referred to as Mars, was Ares. It was the Romans who named the planet Mars, for their god of war, often represented by the sword and shield of the planet's namesake. In the fourth century BCE, Aristotle noted that Mars disappeared behind the Moon during an occultation, indicating that the planet was farther away. Ptolemy, a Greek living in Alexandria, attempted to address the problem of the orbital motion of Mars. Ptolemy's model and his collective work on astronomy was presented in the multi-volume collection later called the Almagest (from the Arabic for "greatest"), which became the authoritative treatise on Western astronomy for the next fourteen centuries. Literature from ancient China confirms that Mars was known by Chinese astronomers by no later than the fourth century BCE. In the East Asian cultures, Mars is traditionally referred to as the "fire star" (火星) based on the Wuxing system. In 1609 Johannes Kepler published a 10 year study of Martian orbit, using the diurnal parallax of Mars, measured by Tycho Brahe, to make a preliminary calculation of the relative distance to the planet. From Brahe's observations of Mars, Kepler deduced that the planet orbited the Sun not in a circle, but in an ellipse. Moreover, Kepler showed that Mars sped up as it approached the Sun and slowed down as it moved farther away, in a manner that later physicists would explain as a consequence of the conservation of angular momentum.: 433–437 In 1610 the first use of a telescope for astronomical observation, including Mars, was performed by Italian astronomer Galileo Galilei. With the telescope the diurnal parallax of Mars was again measured in an effort to determine the Sun-Earth distance. This was first performed by Giovanni Domenico Cassini in 1672. The early parallax measurements were hampered by the quality of the instruments. The only occultation of Mars by Venus observed was that of 13 October 1590, seen by Michael Maestlin at Heidelberg. By the 19th century, the resolution of telescopes reached a level sufficient for surface features to be identified. On 5 September 1877, a perihelic opposition to Mars occurred. The Italian astronomer Giovanni Schiaparelli used a 22-centimetre (8.7 in) telescope in Milan to help produce the first detailed map of Mars. These maps notably contained features he called canali, which, with the possible exception of the natural canyon Valles Marineris, were later shown to be an optical illusion. These canali were supposedly long, straight lines on the surface of Mars, to which he gave names of famous rivers on Earth. His term, which means "channels" or "grooves", was popularly mistranslated in English as "canals". Influenced by the observations, the orientalist Percival Lowell founded an observatory which had 30- and 45-centimetre (12- and 18-in) telescopes. The observatory was used for the exploration of Mars during the last good opportunity in 1894, and the following less favorable oppositions. He published several books on Mars and life on the planet, which had a great influence on the public. The canali were independently observed by other astronomers, like Henri Joseph Perrotin and Louis Thollon in Nice, using one of the largest telescopes of that time. The seasonal changes (consisting of the diminishing of the polar caps and the dark areas formed during Martian summers) in combination with the canals led to speculation about life on Mars, and it was a long-held belief that Mars contained vast seas and vegetation. As bigger telescopes were used, fewer long, straight canali were observed. During observations in 1909 by Antoniadi with an 84-centimetre (33 in) telescope, irregular patterns were observed, but no canali were seen. The first spacecraft from Earth to visit Mars was Mars 1 of the Soviet Union, which flew by in 1963, but contact was lost en route. NASA's Mariner 4 followed and became the first spacecraft to successfully transmit from Mars; launched on 28 November 1964, it made its closest approach to the planet on 15 July 1965. Mariner 4 detected the weak Martian radiation belt, measured at about 0.1% that of Earth, and captured the first images of another planet from deep space. Once spacecraft visited the planet during the 1960s and 1970s, many previous concepts of Mars were radically broken. After the results of the Viking life-detection experiments, the hypothesis of a dead planet was generally accepted. The data from Mariner 9 and Viking allowed better maps of Mars to be made. Until 1997 and after Viking 1 shut down in 1982, Mars was only visited by three unsuccessful probes, two flying past without contact (Phobos 1, 1988; Mars Observer, 1993), and one (Phobos 2 1989) malfunctioning in orbit before reaching its destination Phobos. In 1997 Mars Pathfinder became the first successful rover mission beyond the Moon and started together with Mars Global Surveyor (operated until late 2006) an uninterrupted active robotic presence at Mars that has lasted until today. It produced complete, extremely detailed maps of the Martian topography, magnetic field and surface minerals. Starting with these missions a range of new improved crewless spacecraft, including orbiters, landers, and rovers, have been sent to Mars, with successful missions by the NASA (United States), Jaxa (Japan), ESA, United Kingdom, ISRO (India), Roscosmos (Russia), the United Arab Emirates, and CNSA (China) to study the planet's surface, climate, and geology, uncovering the different elements of the history and dynamic of the hydrosphere of Mars and possible traces of ancient life. As of 2023[update], Mars is host to ten functioning spacecraft. Eight are in orbit: 2001 Mars Odyssey, Mars Express, Mars Reconnaissance Orbiter, MAVEN, ExoMars Trace Gas Orbiter, the Hope orbiter, and the Tianwen-1 orbiter. Another two are on the surface: the Mars Science Laboratory Curiosity rover and the Perseverance rover. Collected maps are available online at websites including Google Mars. NASA provides two online tools: Mars Trek, which provides visualizations of the planet using data from 50 years of exploration, and Experience Curiosity, which simulates traveling on Mars in 3-D with Curiosity. Planned missions to Mars include: As of February 2024[update], debris from these types of missions has reached over seven tons. Most of it consists of crashed and inactive spacecraft as well as discarded components. In April 2024, NASA selected several companies to begin studies on providing commercial services to further enable robotic science on Mars. Key areas include establishing telecommunications, payload delivery and surface imaging. Habitability and habitation During the late 19th century, it was widely accepted in the astronomical community that Mars had life-supporting qualities, including the presence of oxygen and water. However, in 1894 W. W. Campbell at Lick Observatory observed the planet and found that "if water vapor or oxygen occur in the atmosphere of Mars it is in quantities too small to be detected by spectroscopes then available". That observation contradicted many of the measurements of the time and was not widely accepted. Campbell and V. M. Slipher repeated the study in 1909 using better instruments, but with the same results. It was not until the findings were confirmed by W. S. Adams in 1925 that the myth of the Earth-like habitability of Mars was finally broken. However, even in the 1960s, articles were published on Martian biology, putting aside explanations other than life for the seasonal changes on Mars. The current understanding of planetary habitability – the ability of a world to develop environmental conditions favorable to the emergence of life – favors planets that have liquid water on their surface. Most often this requires the orbit of a planet to lie within the habitable zone, which for the Sun is estimated to extend from within the orbit of Earth to about that of Mars. During perihelion, Mars dips inside this region, but Mars's thin (low-pressure) atmosphere prevents liquid water from existing over large regions for extended periods. The past flow of liquid water demonstrates the planet's potential for habitability. Recent evidence has suggested that any water on the Martian surface may have been too salty and acidic to support regular terrestrial life. The environmental conditions on Mars are a challenge to sustaining organic life: the planet has little heat transfer across its surface, it has poor insulation against bombardment by the solar wind due to the absence of a magnetosphere and has insufficient atmospheric pressure to retain water in a liquid form (water instead sublimes to a gaseous state). Mars is nearly, or perhaps totally, geologically dead; the end of volcanic activity has apparently stopped the recycling of chemicals and minerals between the surface and interior of the planet. Evidence suggests that the planet was once significantly more habitable than it is today, but whether living organisms ever existed there remains unknown. The Viking probes of the mid-1970s carried experiments designed to detect microorganisms in Martian soil at their respective landing sites and had positive results, including a temporary increase in CO2 production on exposure to water and nutrients. This sign of life was later disputed by scientists, resulting in a continuing debate, with NASA scientist Gilbert Levin asserting that Viking may have found life. A 2014 analysis of Martian meteorite EETA79001 found chlorate, perchlorate, and nitrate ions in sufficiently high concentrations to suggest that they are widespread on Mars. UV and X-ray radiation would turn chlorate and perchlorate ions into other, highly reactive oxychlorines, indicating that any organic molecules would have to be buried under the surface to survive. Small quantities of methane and formaldehyde detected by Mars orbiters are both claimed to be possible evidence for life, as these chemical compounds would quickly break down in the Martian atmosphere. Alternatively, these compounds may instead be replenished by volcanic or other geological means, such as serpentinite. Impact glass, formed by the impact of meteors, which on Earth can preserve signs of life, has also been found on the surface of the impact craters on Mars. Likewise, the glass in impact craters on Mars could have preserved signs of life, if life existed at the site. The Cheyava Falls rock discovered on Mars in June 2024 has been designated by NASA as a "potential biosignature" and was core sampled by the Perseverance rover for possible return to Earth and further examination. Although highly intriguing, no definitive final determination on a biological or abiotic origin of this rock can be made with the data currently available. Several plans for a human mission to Mars have been proposed, but none have come to fruition. The NASA Authorization Act of 2017 directed NASA to study the feasibility of a crewed Mars mission in the early 2030s; the resulting report concluded that this would be unfeasible. In addition, in 2021, China was planning to send a crewed Mars mission in 2033. Privately held companies such as SpaceX have also proposed plans to send humans to Mars, with the eventual goal to settle on the planet. As of 2024, SpaceX has proceeded with the development of the Starship launch vehicle with the goal of Mars colonization. In plans shared with the company in April 2024, Elon Musk envisions the beginning of a Mars colony within the next twenty years. This would be enabled by the planned mass manufacturing of Starship and initially sustained by resupply from Earth, and in situ resource utilization on Mars, until the Mars colony reaches full self sustainability. Any future human mission to Mars will likely take place within the optimal Mars launch window, which occurs every 26 months. The moon Phobos has been proposed as an anchor point for a space elevator. Besides national space agencies and space companies, groups such as the Mars Society and The Planetary Society advocate for human missions to Mars. In culture Mars is named after the Roman god of war (Greek Ares), but was also associated with the demi-god Heracles (Roman Hercules) by ancient Greek astronomers, as detailed by Aristotle. This association between Mars and war dates back at least to Babylonian astronomy, in which the planet was named for the god Nergal, deity of war and destruction. It persisted into modern times, as exemplified by Gustav Holst's orchestral suite The Planets, whose famous first movement labels Mars "The Bringer of War". The planet's symbol, a circle with a spear pointing out to the upper right, is also used as a symbol for the male gender. The symbol dates from at least the 11th century, though a possible predecessor has been found in the Greek Oxyrhynchus Papyri. The idea that Mars was populated by intelligent Martians became widespread in the late 19th century. Schiaparelli's "canali" observations combined with Percival Lowell's books on the subject put forward the standard notion of a planet that was a drying, cooling, dying world with ancient civilizations constructing irrigation works. Many other observations and proclamations by notable personalities added to what has been termed "Mars Fever". In the present day, high-resolution mapping of the surface of Mars has revealed no artifacts of habitation, but pseudoscientific speculation about intelligent life on Mars still continues. Reminiscent of the canali observations, these speculations are based on small scale features perceived in the spacecraft images, such as "pyramids" and the "Face on Mars". In his book Cosmos, planetary astronomer Carl Sagan wrote: "Mars has become a kind of mythic arena onto which we have projected our Earthly hopes and fears." The depiction of Mars in fiction has been stimulated by its dramatic red color and by nineteenth-century scientific speculations that its surface conditions might support not just life but intelligent life. This gave way to many science fiction stories involving these concepts, such as H. G. Wells's The War of the Worlds, in which Martians seek to escape their dying planet by invading Earth; Ray Bradbury's The Martian Chronicles, in which human explorers accidentally destroy a Martian civilization; as well as Edgar Rice Burroughs's series Barsoom, C. S. Lewis's novel Out of the Silent Planet (1938), and a number of Robert A. Heinlein stories before the mid-sixties. Since then, depictions of Martians have also extended to animation. A comic figure of an intelligent Martian, Marvin the Martian, appeared in Haredevil Hare (1948) as a character in the Looney Tunes animated cartoons of Warner Brothers, and has continued as part of popular culture to the present. After the Mariner and Viking spacecraft had returned pictures of Mars as a lifeless and canal-less world, these ideas about Mars were abandoned; for many science-fiction authors, the new discoveries initially seemed like a constraint, but eventually the post-Viking knowledge of Mars became itself a source of inspiration for works like Kim Stanley Robinson's Mars trilogy. See also Notes References Further reading External links Solar System → Local Interstellar Cloud → Local Bubble → Gould Belt → Orion Arm → Milky Way → Milky Way subgroup → Local Group → Local Sheet → Local Volume → Virgo Supercluster → Laniakea Supercluster → Pisces–Cetus Supercluster Complex → Local Hole → Observable universe → UniverseEach arrow (→) may be read as "within" or "part of". |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/United_States#cite_note-176] | [TOKENS: 17273] |
Contents United States The United States of America (USA), also known as the United States (U.S.) or America, is a country primarily located in North America. It is a federal republic of 50 states and a federal capital district, Washington, D.C. The 48 contiguous states border Canada to the north and Mexico to the south, with the semi-exclave of Alaska in the northwest and the archipelago of Hawaii in the Pacific Ocean. The United States also asserts sovereignty over five major island territories and various uninhabited islands in Oceania and the Caribbean.[j] It is a megadiverse country, with the world's third-largest land area[c] and third-largest population, exceeding 341 million.[k] Paleo-Indians first migrated from North Asia to North America at least 15,000 years ago, and formed various civilizations. Spanish colonization established Spanish Florida in 1513, the first European colony in what is now the continental United States. British colonization followed with the 1607 settlement of Virginia, the first of the Thirteen Colonies. Enslavement of Africans was practiced in all colonies by 1770 and supplied most of the labor for the Southern Colonies' plantation economy. Clashes with the British Crown began as a civil protest over the illegality of taxation without representation in Parliament and the denial of other English rights. They evolved into the American Revolution, which led to the Declaration of Independence and a society based on universal rights. Victory in the 1775–1783 Revolutionary War brought international recognition of U.S. sovereignty and fueled westward expansion, further dispossessing native inhabitants. As more states were admitted, a North–South division over slavery led the Confederate States of America to declare secession and fight the Union in the 1861–1865 American Civil War. With the United States' victory and reunification, slavery was abolished nationally. By the late 19th century, the U.S. economy outpaced the French, German and British economies combined. As of 1900, the country had established itself as a great power, a status solidified after its involvement in World War I. Following Japan's attack on Pearl Harbor in 1941, the U.S. entered World War II. Its aftermath left the U.S. and the Soviet Union as rival superpowers, competing for ideological dominance and international influence during the Cold War. The Soviet Union's collapse in 1991 ended the Cold War, leaving the U.S. as the world's sole superpower. The U.S. federal government is a representative democracy with a president and a constitution that grants separation of powers under three branches: legislative, executive, and judicial. The United States Congress is a bicameral national legislature composed of the House of Representatives (a lower house based on population) and the Senate (an upper house based on equal representation for each state). Federalism grants substantial autonomy to the 50 states. In addition, 574 Native American tribes have sovereignty rights, and there are 326 Native American reservations. Since the 1850s, the Democratic and Republican parties have dominated American politics. American ideals and values are based on a democratic tradition inspired by the American Enlightenment movement. A developed country, the U.S. ranks high in economic competitiveness, innovation, and higher education. Accounting for over a quarter of nominal global GDP, its economy has been the world's largest since about 1890. It is the wealthiest country, with the highest disposable household income per capita among OECD members, though its wealth inequality is highly pronounced. Shaped by centuries of immigration, the culture of the U.S. is diverse and globally influential. Making up more than a third of global military spending, the country has one of the strongest armed forces and is a designated nuclear state. A member of numerous international organizations, the U.S. plays a major role in global political, cultural, economic, and military affairs. Etymology Documented use of the phrase "United States of America" dates back to January 2, 1776. On that day, Stephen Moylan, a Continental Army aide to General George Washington, wrote a letter to Joseph Reed, Washington's aide-de-camp, seeking to go "with full and ample powers from the United States of America to Spain" to seek assistance in the Revolutionary War effort. The first known public usage is an anonymous essay published in the Williamsburg newspaper The Virginia Gazette on April 6, 1776. Sometime on or after June 11, 1776, Thomas Jefferson wrote "United States of America" in a rough draft of the Declaration of Independence, which was adopted by the Second Continental Congress on July 4, 1776. The term "United States" and its initialism "U.S.", used as nouns or as adjectives in English, are common short names for the country. The initialism "USA", a noun, is also common. "United States" and "U.S." are the established terms throughout the U.S. federal government, with prescribed rules.[l] "The States" is an established colloquial shortening of the name, used particularly from abroad; "stateside" is the corresponding adjective or adverb. "America" is the feminine form of the first word of Americus Vesputius, the Latinized name of Italian explorer Amerigo Vespucci (1454–1512);[m] it was first used as a place name by the German cartographers Martin Waldseemüller and Matthias Ringmann in 1507.[n] Vespucci first proposed that the West Indies discovered by Christopher Columbus in 1492 were part of a previously unknown landmass and not among the Indies at the eastern limit of Asia. In English, the term "America" usually does not refer to topics unrelated to the United States, despite the usage of "the Americas" to describe the totality of the continents of North and South America. History The first inhabitants of North America migrated from Siberia approximately 15,000 years ago, either across the Bering land bridge or along the now-submerged Ice Age coastline. Small isolated groups of hunter-gatherers are said to have migrated alongside herds of large herbivores far into Alaska, with ice-free corridors developing along the Pacific coast and valleys of North America in c. 16,500 – c. 13,500 BCE (c. 18,500 – c. 15,500 BP). The Clovis culture, which appeared around 11,000 BCE, is believed to be the first widespread culture in the Americas. Over time, Indigenous North American cultures grew increasingly sophisticated, and some, such as the Mississippian culture, developed agriculture, architecture, and complex societies. In the post-archaic period, the Mississippian cultures were located in the midwestern, eastern, and southern regions, and the Algonquian in the Great Lakes region and along the Eastern Seaboard, while the Hohokam culture and Ancestral Puebloans inhabited the Southwest. Native population estimates of what is now the United States before the arrival of European colonizers range from around 500,000 to nearly 10 million. Christopher Columbus began exploring the Caribbean for Spain in 1492, leading to Spanish-speaking settlements and missions from what are now Puerto Rico and Florida to New Mexico and California. The first Spanish colony in the present-day continental United States was Spanish Florida, chartered in 1513. After several settlements failed there due to starvation and disease, Spain's first permanent town, Saint Augustine, was founded in 1565. France established its own settlements in French Florida in 1562, but they were either abandoned (Charlesfort, 1578) or destroyed by Spanish raids (Fort Caroline, 1565). Permanent French settlements were founded much later along the Great Lakes (Fort Detroit, 1701), the Mississippi River (Saint Louis, 1764) and especially the Gulf of Mexico (New Orleans, 1718). Early European colonies also included the thriving Dutch colony of New Nederland (settled 1626, present-day New York) and the small Swedish colony of New Sweden (settled 1638 in what became Delaware). British colonization of the East Coast began with the Virginia Colony (1607) and the Plymouth Colony (Massachusetts, 1620). The Mayflower Compact in Massachusetts and the Fundamental Orders of Connecticut established precedents for local representative self-governance and constitutionalism that would develop throughout the American colonies. While European settlers in what is now the United States experienced conflicts with Native Americans, they also engaged in trade, exchanging European tools for food and animal pelts.[o] Relations ranged from close cooperation to warfare and massacres. The colonial authorities often pursued policies that forced Native Americans to adopt European lifestyles, including conversion to Christianity. Along the eastern seaboard, settlers trafficked Africans through the Atlantic slave trade, largely to provide manual labor on plantations. The original Thirteen Colonies[p] that would later found the United States were administered as possessions of the British Empire by Crown-appointed governors, though local governments held elections open to most white male property owners. The colonial population grew rapidly from Maine to Georgia, eclipsing Native American populations; by the 1770s, the natural increase of the population was such that only a small minority of Americans had been born overseas. The colonies' distance from Britain facilitated the entrenchment of self-governance, and the First Great Awakening, a series of Christian revivals, fueled colonial interest in guaranteed religious liberty. Following its victory in the French and Indian War, Britain began to assert greater control over local affairs in the Thirteen Colonies, resulting in growing political resistance. One of the primary grievances of the colonists was the denial of their rights as Englishmen, particularly the right to representation in the British government that taxed them. To demonstrate their dissatisfaction and resolve, the First Continental Congress met in 1774 and passed the Continental Association, a colonial boycott of British goods enforced by local "committees of safety" that proved effective. The British attempt to then disarm the colonists resulted in the 1775 Battles of Lexington and Concord, igniting the American Revolutionary War. At the Second Continental Congress, the colonies appointed George Washington commander-in-chief of the Continental Army, and created a committee that named Thomas Jefferson to draft the Declaration of Independence. Two days after the Second Continental Congress passed the Lee Resolution to create an independent, sovereign nation, the Declaration was adopted on July 4, 1776. The political values of the American Revolution evolved from an armed rebellion demanding reform within an empire to a revolution that created a new social and governing system founded on the defense of liberty and the protection of inalienable natural rights; sovereignty of the people; republicanism over monarchy, aristocracy, and other hereditary political power; civic virtue; and an intolerance of political corruption. The Founding Fathers of the United States, who included Washington, Jefferson, John Adams, Benjamin Franklin, Alexander Hamilton, John Jay, James Madison, Thomas Paine, and many others, were inspired by Classical, Renaissance, and Enlightenment philosophies and ideas. Though in practical effect since its drafting in 1777, the Articles of Confederation was ratified in 1781 and formally established a decentralized government that operated until 1789. After the British surrender at the siege of Yorktown in 1781, American sovereignty was internationally recognized by the Treaty of Paris (1783), through which the U.S. gained territory stretching west to the Mississippi River, north to present-day Canada, and south to Spanish Florida. The Northwest Ordinance (1787) established the precedent by which the country's territory would expand with the admission of new states, rather than the expansion of existing states. The U.S. Constitution was drafted at the 1787 Constitutional Convention to overcome the limitations of the Articles. It went into effect in 1789, creating a federal republic governed by three separate branches that together formed a system of checks and balances. George Washington was elected the country's first president under the Constitution, and the Bill of Rights was adopted in 1791 to allay skeptics' concerns about the power of the more centralized government. His resignation as commander-in-chief after the Revolutionary War and his later refusal to run for a third term as the country's first president established a precedent for the supremacy of civil authority in the United States and the peaceful transfer of power. In the late 18th century, American settlers began to expand westward in larger numbers, many with a sense of manifest destiny. The Louisiana Purchase of 1803 from France nearly doubled the territory of the United States. Lingering issues with Britain remained, leading to the War of 1812, which was fought to a draw. Spain ceded Florida and its Gulf Coast territory in 1819. The Missouri Compromise of 1820, which admitted Missouri as a slave state and Maine as a free state, attempted to balance the desire of northern states to prevent the expansion of slavery into new territories with that of southern states to extend it there. Primarily, the compromise prohibited slavery in all other lands of the Louisiana Purchase north of the 36°30′ parallel. As Americans expanded further into territory inhabited by Native Americans, the federal government implemented policies of Indian removal or assimilation. The most significant such legislation was the Indian Removal Act of 1830, a key policy of President Andrew Jackson. It resulted in the Trail of Tears (1830–1850), in which an estimated 60,000 Native Americans living east of the Mississippi River were forcibly removed and displaced to lands far to the west, causing 13,200 to 16,700 deaths along the forced march. Settler expansion as well as this influx of Indigenous peoples from the East resulted in the American Indian Wars west of the Mississippi. During the colonial period, slavery became legal in all the Thirteen colonies, but by 1770 it provided the main labor force in the large-scale, agriculture-dependent economies of the Southern Colonies from Maryland to Georgia. The practice began to be significantly questioned during the American Revolution, and spurred by an active abolitionist movement that had reemerged in the 1830s, states in the North enacted laws to prohibit slavery within their boundaries. At the same time, support for slavery had strengthened in Southern states, with widespread use of inventions such as the cotton gin (1793) having made slavery immensely profitable for Southern elites. The United States annexed the Republic of Texas in 1845, and the 1846 Oregon Treaty led to U.S. control of the present-day American Northwest. Dispute with Mexico over Texas led to the Mexican–American War (1846–1848). After the victory of the U.S., Mexico recognized U.S. sovereignty over Texas, New Mexico, and California in the 1848 Mexican Cession; the cession's lands also included the future states of Nevada, Colorado and Utah. The California gold rush of 1848–1849 spurred a huge migration of white settlers to the Pacific coast, leading to even more confrontations with Native populations. One of the most violent, the California genocide of thousands of Native inhabitants, lasted into the mid-1870s. Additional western territories and states were created. Throughout the 1850s, the sectional conflict regarding slavery was further inflamed by national legislation in the U.S. Congress and decisions of the Supreme Court. In Congress, the Fugitive Slave Act of 1850 mandated the forcible return to their owners in the South of slaves taking refuge in non-slave states, while the Kansas–Nebraska Act of 1854 effectively gutted the anti-slavery requirements of the Missouri Compromise. In its Dred Scott decision of 1857, the Supreme Court ruled against a slave brought into non-slave territory, simultaneously declaring the entire Missouri Compromise to be unconstitutional. These and other events exacerbated tensions between North and South that would culminate in the American Civil War (1861–1865). Beginning with South Carolina, 11 slave-state governments voted to secede from the United States in 1861, joining to create the Confederate States of America. All other state governments remained loyal to the Union.[q] War broke out in April 1861 after the Confederacy bombarded Fort Sumter. Following the Emancipation Proclamation on January 1, 1863, many freed slaves joined the Union army. The war began to turn in the Union's favor following the 1863 Siege of Vicksburg and Battle of Gettysburg, and the Confederates surrendered in 1865 after the Union's victory in the Battle of Appomattox Court House. Efforts toward reconstruction in the secessionist South had begun as early as 1862, but it was only after President Lincoln's assassination that the three Reconstruction Amendments to the Constitution were ratified to protect civil rights. The amendments codified nationally the abolition of slavery and involuntary servitude except as punishment for crimes, promised equal protection under the law for all persons, and prohibited discrimination on the basis of race or previous enslavement. As a result, African Americans took an active political role in ex-Confederate states in the decade following the Civil War. The former Confederate states were readmitted to the Union, beginning with Tennessee in 1866 and ending with Georgia in 1870. National infrastructure, including transcontinental telegraph and railroads, spurred growth in the American frontier. This was accelerated by the Homestead Acts, through which nearly 10 percent of the total land area of the United States was given away free to some 1.6 million homesteaders. From 1865 through 1917, an unprecedented stream of immigrants arrived in the United States, including 24.4 million from Europe. Most came through the Port of New York, as New York City and other large cities on the East Coast became home to large Jewish, Irish, and Italian populations. Many Northern Europeans as well as significant numbers of Germans and other Central Europeans moved to the Midwest. At the same time, about one million French Canadians migrated from Quebec to New England. During the Great Migration, millions of African Americans left the rural South for urban areas in the North. Alaska was purchased from Russia in 1867. The Compromise of 1877 is generally considered the end of the Reconstruction era, as it resolved the electoral crisis following the 1876 presidential election and led President Rutherford B. Hayes to reduce the role of federal troops in the South. Immediately, the Redeemers began evicting the Carpetbaggers and quickly regained local control of Southern politics in the name of white supremacy. African Americans endured a period of heightened, overt racism following Reconstruction, a time often considered the nadir of American race relations. A series of Supreme Court decisions, including Plessy v. Ferguson, emptied the Fourteenth and Fifteenth Amendments of their force, allowing Jim Crow laws in the South to remain unchecked, sundown towns in the Midwest, and segregation in communities across the country, which would be reinforced in part by the policy of redlining later adopted by the federal Home Owners' Loan Corporation. An explosion of technological advancement, accompanied by the exploitation of cheap immigrant labor, led to rapid economic expansion during the Gilded Age of the late 19th century. It continued into the early 20th, when the United States already outpaced the economies of Britain, France, and Germany combined. This fostered the amassing of power by a few prominent industrialists, largely by their formation of trusts and monopolies to prevent competition. Tycoons led the nation's expansion in the railroad, petroleum, and steel industries. The United States emerged as a pioneer of the automotive industry. These changes resulted in significant increases in economic inequality, slum conditions, and social unrest, creating the environment for labor unions and socialist movements to begin to flourish. This period eventually ended with the advent of the Progressive Era, which was characterized by significant economic and social reforms. Pro-American elements in Hawaii overthrew the Hawaiian monarchy; the islands were annexed in 1898. That same year, Puerto Rico, the Philippines, and Guam were ceded to the U.S. by Spain after the latter's defeat in the Spanish–American War. (The Philippines was granted full independence from the U.S. on July 4, 1946, following World War II. Puerto Rico and Guam have remained U.S. territories.) American Samoa was acquired by the United States in 1900 after the Second Samoan Civil War. The U.S. Virgin Islands were purchased from Denmark in 1917. The United States entered World War I alongside the Allies in 1917 helping to turn the tide against the Central Powers. In 1920, a constitutional amendment granted nationwide women's suffrage. During the 1920s and 1930s, radio for mass communication and early television transformed communications nationwide. The Wall Street Crash of 1929 triggered the Great Depression, to which President Franklin D. Roosevelt responded with the New Deal plan of "reform, recovery and relief", a series of unprecedented and sweeping recovery programs and employment relief projects combined with financial reforms and regulations. Initially neutral during World War II, the U.S. began supplying war materiel to the Allies of World War II in March 1941 and entered the war in December after Japan's attack on Pearl Harbor. Agreeing to a "Europe first" policy, the U.S. concentrated its wartime efforts on Japan's allies Italy and Germany until their final defeat in May 1945. The U.S. developed the first nuclear weapons and used them against the Japanese cities of Hiroshima and Nagasaki in August 1945, ending the war. The United States was one of the "Four Policemen" who met to plan the post-war world, alongside the United Kingdom, the Soviet Union, and China. The U.S. emerged relatively unscathed from the war, with even greater economic power and international political influence. The end of World War II in 1945 left the U.S. and the Soviet Union as superpowers, each with its own political, military, and economic sphere of influence. Geopolitical tensions between the two superpowers soon led to the Cold War. The U.S. implemented a policy of containment intended to limit the Soviet Union's sphere of influence; engaged in regime change against governments perceived to be aligned with the Soviets; and prevailed in the Space Race, which culminated with the first crewed Moon landing in 1969. Domestically, the U.S. experienced economic growth, urbanization, and population growth following World War II. The civil rights movement emerged, with Martin Luther King Jr. becoming a prominent leader in the early 1960s. The Great Society plan of President Lyndon B. Johnson's administration resulted in groundbreaking and broad-reaching laws, policies and a constitutional amendment to counteract some of the worst effects of lingering institutional racism. The counterculture movement in the U.S. brought significant social changes, including the liberalization of attitudes toward recreational drug use and sexuality. It also encouraged open defiance of the military draft (leading to the end of conscription in 1973) and wide opposition to U.S. intervention in Vietnam, with the U.S. totally withdrawing in 1975. A societal shift in the roles of women was significantly responsible for the large increase in female paid labor participation starting in the 1970s, and by 1985 the majority of American women aged 16 and older were employed. The Fall of Communism and the dissolution of the Soviet Union from 1989 to 1991 marked the end of the Cold War and left the United States as the world's sole superpower. This cemented the United States' global influence, reinforcing the concept of the "American Century" as the U.S. dominated international political, cultural, economic, and military affairs. The 1990s saw the longest recorded economic expansion in American history, a dramatic decline in U.S. crime rates, and advances in technology. Throughout this decade, technological innovations such as the World Wide Web, the evolution of the Pentium microprocessor in accordance with Moore's law, rechargeable lithium-ion batteries, the first gene therapy trial, and cloning either emerged in the U.S. or were improved upon there. The Human Genome Project was formally launched in 1990, while Nasdaq became the first stock market in the United States to trade online in 1998. In the Gulf War of 1991, an American-led international coalition of states expelled an Iraqi invasion force that had occupied neighboring Kuwait. The September 11 attacks on the United States in 2001 by the pan-Islamist militant organization al-Qaeda led to the war on terror and subsequent military interventions in Afghanistan and in Iraq. The U.S. housing bubble culminated in 2007 with the Great Recession, the largest economic contraction since the Great Depression. In the 2010s and early 2020s, the United States has experienced increased political polarization and democratic backsliding. The country's polarization was violently reflected in the January 2021 Capitol attack, when a mob of insurrectionists entered the U.S. Capitol and sought to prevent the peaceful transfer of power in an attempted self-coup d'état. Geography The United States is the world's third-largest country by total area behind Russia and Canada.[c] The 48 contiguous states and the District of Columbia have a combined area of 3,119,885 square miles (8,080,470 km2). In 2021, the United States had 8% of the Earth's permanent meadows and pastures and 10% of its cropland. Starting in the east, the coastal plain of the Atlantic seaboard gives way to inland forests and rolling hills in the Piedmont plateau region. The Appalachian Mountains and the Adirondack Massif separate the East Coast from the Great Lakes and the grasslands of the Midwest. The Mississippi River System, the world's fourth-longest river system, runs predominantly north–south through the center of the country. The flat and fertile prairie of the Great Plains stretches to the west, interrupted by a highland region in the southeast. The Rocky Mountains, west of the Great Plains, extend north to south across the country, peaking at over 14,000 feet (4,300 m) in Colorado. The supervolcano underlying Yellowstone National Park in the Rocky Mountains, the Yellowstone Caldera, is the continent's largest volcanic feature. Farther west are the rocky Great Basin and the Chihuahuan, Sonoran, and Mojave deserts. In the northwest corner of Arizona, carved by the Colorado River, is the Grand Canyon, a steep-sided canyon and popular tourist destination known for its overwhelming visual size and intricate, colorful landscape. The Cascade and Sierra Nevada mountain ranges run close to the Pacific coast. The lowest and highest points in the contiguous United States are in the State of California, about 84 miles (135 km) apart. At an elevation of 20,310 feet (6,190.5 m), Alaska's Denali (also called Mount McKinley) is the highest peak in the country and on the continent. Active volcanoes in the U.S. are common throughout Alaska's Alexander and Aleutian Islands. Located entirely outside North America, the archipelago of Hawaii consists of volcanic islands, physiographically and ethnologically part of the Polynesian subregion of Oceania. In addition to its total land area, the United States has one of the world's largest marine exclusive economic zones spanning approximately 4.5 million square miles (11.7 million km2) of ocean. With its large size and geographic variety, the United States includes most climate types. East of the 100th meridian, the climate ranges from humid continental in the north to humid subtropical in the south. The western Great Plains are semi-arid. Many mountainous areas of the American West have an alpine climate. The climate is arid in the Southwest, Mediterranean in coastal California, and oceanic in coastal Oregon, Washington, and southern Alaska. Most of Alaska is subarctic or polar. Hawaii, the southern tip of Florida and U.S. territories in the Caribbean and Pacific are tropical. The United States receives more high-impact extreme weather incidents than any other country. States bordering the Gulf of Mexico are prone to hurricanes, and most of the world's tornadoes occur in the country, mainly in Tornado Alley. Due to climate change in the country, extreme weather has become more frequent in the U.S. in the 21st century, with three times the number of reported heat waves compared to the 1960s. Since the 1990s, droughts in the American Southwest have become more persistent and more severe. The regions considered as the most attractive to the population are the most vulnerable. The U.S. is one of 17 megadiverse countries containing large numbers of endemic species: about 17,000 species of vascular plants occur in the contiguous United States and Alaska, and over 1,800 species of flowering plants are found in Hawaii, few of which occur on the mainland. The United States is home to 428 mammal species, 784 birds, 311 reptiles, 295 amphibians, and around 91,000 insect species. There are 63 national parks, and hundreds of other federally managed monuments, forests, and wilderness areas, administered by the National Park Service and other agencies. About 28% of the country's land is publicly owned and federally managed, primarily in the Western States. Most of this land is protected, though some is leased for commercial use, and less than one percent is used for military purposes. Environmental issues in the United States include debates on non-renewable resources and nuclear energy, air and water pollution, biodiversity, logging and deforestation, and climate change. The U.S. Environmental Protection Agency (EPA) is the federal agency charged with addressing most environmental-related issues. The idea of wilderness has shaped the management of public lands since 1964, with the Wilderness Act. The Endangered Species Act of 1973 provides a way to protect threatened and endangered species and their habitats. The United States Fish and Wildlife Service implements and enforces the Act. In 2024, the U.S. ranked 35th among 180 countries in the Environmental Performance Index. Government and politics The United States is a federal republic of 50 states and a federal capital district, Washington, D.C. The U.S. asserts sovereignty over five unincorporated territories and several uninhabited island possessions. It is the world's oldest surviving federation, and its presidential system of federal government has been adopted, in whole or in part, by many newly independent states worldwide following their decolonization. The Constitution of the United States serves as the country's supreme legal document. Most scholars describe the United States as a liberal democracy.[r] Composed of three branches, all headquartered in Washington, D.C., the federal government is the national government of the United States. The U.S. Constitution establishes a separation of powers intended to provide a system of checks and balances to prevent any of the three branches from becoming supreme. The three-branch system is known as the presidential system, in contrast to the parliamentary system where the executive is part of the legislative body. Many countries around the world adopted this aspect of the 1789 Constitution of the United States, especially in the postcolonial Americas. In the U.S. federal system, sovereign powers are shared between three levels of government specified in the Constitution: the federal government, the states, and Indian tribes. The U.S. also asserts sovereignty over five permanently inhabited territories: American Samoa, Guam, the Northern Mariana Islands, Puerto Rico, and the U.S. Virgin Islands. Residents of the 50 states are governed by their elected state government, under state constitutions compatible with the national constitution, and by elected local governments that are administrative divisions of a state. States are subdivided into counties or county equivalents, and (except for Hawaii) further divided into municipalities, each administered by elected representatives. The District of Columbia is a federal district containing the U.S. capital, Washington, D.C. The federal district is an administrative division of the federal government. Indian country is made up of 574 federally recognized tribes and 326 Indian reservations. They hold a government-to-government relationship with the U.S. federal government in Washington and are legally defined as domestic dependent nations with inherent tribal sovereignty rights. In addition to the five major territories, the U.S. also asserts sovereignty over the United States Minor Outlying Islands in the Pacific Ocean and the Caribbean. The seven undisputed islands without permanent populations are Baker Island, Howland Island, Jarvis Island, Johnston Atoll, Kingman Reef, Midway Atoll, and Palmyra Atoll. U.S. sovereignty over the unpopulated Bajo Nuevo Bank, Navassa Island, Serranilla Bank, and Wake Island is disputed. The Constitution is silent on political parties. However, they developed independently in the 18th century with the Federalist and Anti-Federalist parties. Since then, the United States has operated as a de facto two-party system, though the parties have changed over time. Since the mid-19th century, the two main national parties have been the Democratic Party and the Republican Party. The former is perceived as relatively liberal in its political platform while the latter is perceived as relatively conservative in its platform. The United States has an established structure of foreign relations, with the world's second-largest diplomatic corps as of 2024[update]. It is a permanent member of the United Nations Security Council and home to the United Nations headquarters. The United States is a member of the G7, G20, and OECD intergovernmental organizations. Almost all countries have embassies and many have consulates (official representatives) in the country. Likewise, nearly all countries host formal diplomatic missions with the United States, except Iran, North Korea, and Bhutan. Though Taiwan does not have formal diplomatic relations with the U.S., it maintains close unofficial relations. The United States regularly supplies Taiwan with military equipment to deter potential Chinese aggression. Its geopolitical attention also turned to the Indo-Pacific when the United States joined the Quadrilateral Security Dialogue with Australia, India, and Japan. The United States has a "Special Relationship" with the United Kingdom and strong ties with Canada, Australia, New Zealand, the Philippines, Japan, South Korea, Israel, and several European Union countries such as France, Italy, Germany, Spain, and Poland. The U.S. works closely with its NATO allies on military and national security issues, and with countries in the Americas through the Organization of American States and the United States–Mexico–Canada Free Trade Agreement. The U.S. exercises full international defense authority and responsibility for Micronesia, the Marshall Islands, and Palau through the Compact of Free Association. It has increasingly conducted strategic cooperation with India, while its ties with China have steadily deteriorated. Beginning in 2014, the U.S. had become a key ally of Ukraine. After Donald Trump was elected U.S. president in 2024, he sought to negotiate an end to the Russo-Ukrainian War. He paused all military aid to Ukraine in March 2025, although the aid resumed later. Trump also ended U.S. intelligence sharing with the country, but this too was eventually restored. The president is the commander-in-chief of the United States Armed Forces and appoints its leaders, the secretary of defense and the Joint Chiefs of Staff. The Department of Defense, headquartered at the Pentagon near Washington, D.C., administers five of the six service branches, which are made up of the U.S. Army, Marine Corps, Navy, Air Force, and Space Force. The Coast Guard is administered by the Department of Homeland Security in peacetime and can be transferred to the Department of the Navy in wartime. Total strength of the entire military is about 1.3 million active duty with an additional 400,000 in reserve. The United States spent $997 billion on its military in 2024, which is by far the largest amount of any country, making up 37% of global military spending and accounting for 3.4% of the country's GDP. The U.S. possesses 42% of the world's nuclear weapons—the second-largest stockpile after that of Russia. The U.S. military is widely regarded as the most powerful and advanced in the world. The United States has the third-largest combined armed forces in the world, behind the Chinese People's Liberation Army and Indian Armed Forces. The U.S. military operates about 800 bases and facilities abroad, and maintains deployments greater than 100 active duty personnel in 25 foreign countries. The United States has engaged in over 400 military interventions since its founding in 1776, with over half of these occurring between 1950 and 2019 and 25% occurring in the post-Cold War era. State defense forces (SDFs) are military units that operate under the sole authority of a state government. SDFs are authorized by state and federal law but are under the command of the state's governor. By contrast, the 54 U.S. National Guard organizations[t] fall under the dual control of state or territorial governments and the federal government; their units can also become federalized entities, but SDFs cannot be federalized. The National Guard personnel of a state or territory can be federalized by the president under the National Defense Act Amendments of 1933; this legislation created the Guard and provides for the integration of Army National Guard and Air National Guard units and personnel into the U.S. Army and (since 1947) the U.S. Air Force. The total number of National Guard members is about 430,000, while the estimated combined strength of SDFs is less than 10,000. There are about 18,000 U.S. police agencies from local to national level in the United States. Law in the United States is mainly enforced by local police departments and sheriff departments in their municipal or county jurisdictions. The state police departments have authority in their respective state, and federal agencies such as the Federal Bureau of Investigation (FBI) and the U.S. Marshals Service have national jurisdiction and specialized duties, such as protecting civil rights, national security, enforcing U.S. federal courts' rulings and federal laws, and interstate criminal activity. State courts conduct almost all civil and criminal trials, while federal courts adjudicate the much smaller number of civil and criminal cases that relate to federal law. There is no unified "criminal justice system" in the United States. The American prison system is largely heterogenous, with thousands of relatively independent systems operating across federal, state, local, and tribal levels. In 2025, "these systems hold nearly 2 million people in 1,566 state prisons, 98 federal prisons, 3,116 local jails, 1,277 juvenile correctional facilities, 133 immigration detention facilities, and 80 Indian country jails, as well as in military prisons, civil commitment centers, state psychiatric hospitals, and prisons in the U.S. territories." Despite disparate systems of confinement, four main institutions dominate: federal prisons, state prisons, local jails, and juvenile correctional facilities. Federal prisons are run by the Federal Bureau of Prisons and hold pretrial detainees as well as people who have been convicted of federal crimes. State prisons, run by the department of corrections of each state, hold people sentenced and serving prison time (usually longer than one year) for felony offenses. Local jails are county or municipal facilities that incarcerate defendants prior to trial; they also hold those serving short sentences (typically under a year). Juvenile correctional facilities are operated by local or state governments and serve as longer-term placements for any minor adjudicated as delinquent and ordered by a judge to be confined. In January 2023, the United States had the sixth-highest per capita incarceration rate in the world—531 people per 100,000 inhabitants—and the largest prison and jail population in the world, with more than 1.9 million people incarcerated. An analysis of the World Health Organization Mortality Database from 2010 showed U.S. homicide rates "were 7 times higher than in other high-income countries, driven by a gun homicide rate that was 25 times higher". Economy The U.S. has a highly developed mixed economy that has been the world's largest nominally since about 1890. Its 2024 gross domestic product (GDP)[e] of more than $29 trillion constituted over 25% of nominal global economic output, or 15% at purchasing power parity (PPP). From 1983 to 2008, U.S. real compounded annual GDP growth was 3.3%, compared to a 2.3% weighted average for the rest of the G7. The country ranks first in the world by nominal GDP, second when adjusted for purchasing power parities (PPP), and ninth by PPP-adjusted GDP per capita. In February 2024, the total U.S. federal government debt was $34.4 trillion. Of the world's 500 largest companies by revenue, 138 were headquartered in the U.S. in 2025, the highest number of any country. The U.S. dollar is the currency most used in international transactions and the world's foremost reserve currency, backed by the country's dominant economy, its military, the petrodollar system, its large U.S. treasuries market, and its linked eurodollar. Several countries use it as their official currency, and in others it is the de facto currency. The U.S. has free trade agreements with several countries, including the USMCA. Although the United States has reached a post-industrial level of economic development and is often described as having a service economy, it remains a major industrial power; in 2024, the U.S. manufacturing sector was the world's second-largest by value output after China's. New York City is the world's principal financial center, and its metropolitan area is the world's largest metropolitan economy. The New York Stock Exchange and Nasdaq, both located in New York City, are the world's two largest stock exchanges by market capitalization and trade volume. The United States is at the forefront of technological advancement and innovation in many economic fields, especially in artificial intelligence; electronics and computers; pharmaceuticals; and medical, aerospace and military equipment. The country's economy is fueled by abundant natural resources, a well-developed infrastructure, and high productivity. The largest trading partners of the United States are the European Union, Mexico, Canada, China, Japan, South Korea, the United Kingdom, Vietnam, India, and Taiwan. The United States is the world's largest importer and second-largest exporter.[u] It is by far the world's largest exporter of services. Americans have the highest average household and employee income among OECD member states, and the fourth-highest median household income in 2023, up from sixth-highest in 2013. With personal consumption expenditures of over $18.5 trillion in 2023, the U.S. has a heavily consumer-driven economy and is the world's largest consumer market. The U.S. ranked first in the number of dollar billionaires and millionaires in 2023, with 735 billionaires and nearly 22 million millionaires. Wealth in the United States is highly concentrated; in 2011, the richest 10% of the adult population owned 72% of the country's household wealth, while the bottom 50% owned just 2%. U.S. wealth inequality increased substantially since the late 1980s, and income inequality in the U.S. reached a record high in 2019. In 2024, the country had some of the highest wealth and income inequality levels among OECD countries. Since the 1970s, there has been a decoupling of U.S. wage gains from worker productivity. In 2016, the top fifth of earners took home more than half of all income, giving the U.S. one of the widest income distributions among OECD countries. There were about 771,480 homeless persons in the U.S. in 2024. In 2022, 6.4 million children experienced food insecurity. Feeding America estimates that around one in five, or approximately 13 million, children experience hunger in the U.S. and do not know where or when they will get their next meal. Also in 2022, about 37.9 million people, or 11.5% of the U.S. population, were living in poverty. The United States has a smaller welfare state and redistributes less income through government action than most other high-income countries. It is the only advanced economy that does not guarantee its workers paid vacation nationally and one of a few countries in the world without federal paid family leave as a legal right. The United States has a higher percentage of low-income workers than almost any other developed country, largely because of a weak collective bargaining system and lack of government support for at-risk workers. The United States has been a leader in technological innovation since the late 19th century and scientific research since the mid-20th century. Methods for producing interchangeable parts and the establishment of a machine tool industry enabled the large-scale manufacturing of U.S. consumer products in the late 19th century. By the early 20th century, factory electrification, the introduction of the assembly line, and other labor-saving techniques created the system of mass production. In the 21st century, the United States continues to be one of the world's foremost scientific powers, though China has emerged as a major competitor in many fields. The U.S. has the highest research and development expenditures of any country and ranks ninth as a percentage of GDP. In 2022, the United States was (after China) the country with the second-highest number of published scientific papers. In 2021, the U.S. ranked second (also after China) by the number of patent applications, and third by trademark and industrial design applications (after China and Germany), according to World Intellectual Property Indicators. In 2025 the United States ranked third (after Switzerland and Sweden) in the Global Innovation Index. The United States is considered to be a world leader in the development of artificial intelligence technology. In 2023, the United States was ranked the second most technologically advanced country in the world (after South Korea) by Global Finance magazine. The United States has maintained a space program since the late 1950s, beginning with the establishment of the National Aeronautics and Space Administration (NASA) in 1958. NASA's Apollo program (1961–1972) achieved the first crewed Moon landing with the 1969 Apollo 11 mission; it remains one of the agency's most significant milestones. Other major endeavors by NASA include the Space Shuttle program (1981–2011), the Voyager program (1972–present), the Hubble and James Webb space telescopes (launched in 1990 and 2021, respectively), and the multi-mission Mars Exploration Program (Spirit and Opportunity, Curiosity, and Perseverance). NASA is one of five agencies collaborating on the International Space Station (ISS); U.S. contributions to the ISS include several modules, including Destiny (2001), Harmony (2007), and Tranquility (2010), as well as ongoing logistical and operational support. The United States private sector dominates the global commercial spaceflight industry. Prominent American spaceflight contractors include Blue Origin, Boeing, Lockheed Martin, Northrop Grumman, and SpaceX. NASA programs such as the Commercial Crew Program, Commercial Resupply Services, Commercial Lunar Payload Services, and NextSTEP have facilitated growing private-sector involvement in American spaceflight. In 2023, the United States received approximately 84% of its energy from fossil fuel, and its largest source of energy was petroleum (38%), followed by natural gas (36%), renewable sources (9%), coal (9%), and nuclear power (9%). In 2022, the United States constituted about 4% of the world's population, but consumed around 16% of the world's energy. The U.S. ranks as the second-highest emitter of greenhouse gases behind China. The U.S. is the world's largest producer of nuclear power, generating around 30% of the world's nuclear electricity. It also has the highest number of nuclear power reactors of any country. From 2024, the U.S. plans to triple its nuclear power capacity by 2050. The United States' 4 million miles (6.4 million kilometers) of road network, owned almost entirely by state and local governments, is the longest in the world. The extensive Interstate Highway System that connects all major U.S. cities is funded mostly by the federal government but maintained by state departments of transportation. The system is further extended by state highways and some private toll roads. The U.S. is among the top ten countries with the highest vehicle ownership per capita (850 vehicles per 1,000 people) in 2022. A 2022 study found that 76% of U.S. commuters drive alone and 14% ride a bicycle, including bike owners and users of bike-sharing networks. About 11% use some form of public transportation. Public transportation in the United States is well developed in the largest urban areas, notably New York City, Washington, D.C., Boston, Philadelphia, Chicago, and San Francisco; otherwise, coverage is generally less extensive than in most other developed countries. The U.S. also has many relatively car-dependent localities. Long-distance intercity travel is provided primarily by airlines, but travel by rail is more common along the Northeast Corridor, the only high-speed rail in the U.S. that meets international standards. Amtrak, the country's government-sponsored national passenger rail company, has a relatively sparse network compared to that of Western European countries. Service is concentrated in the Northeast, California, the Midwest, the Pacific Northwest, and Virginia/Southeast. The United States has an extensive air transportation network. U.S. civilian airlines are all privately owned. The three largest airlines in the world, by total number of passengers carried, are U.S.-based; American Airlines became the global leader after its 2013 merger with US Airways. Of the 50 busiest airports in the world, 16 are in the United States, as well as five of the top 10. The world's busiest airport by passenger volume is Hartsfield–Jackson Atlanta International in Atlanta, Georgia. In 2022, most of the 19,969 U.S. airports were owned and operated by local government authorities, and there are also some private airports. Some 5,193 are designated as "public use", including for general aviation. The Transportation Security Administration (TSA) has provided security at most major airports since 2001. The country's rail transport network, the longest in the world at 182,412.3 mi (293,564.2 km), handles mostly freight (in contrast to more passenger-centered rail in Europe). Because they are often privately owned operations, U.S. railroads lag behind those of the rest of the world in terms of electrification. The country's inland waterways are the world's fifth-longest, totaling 25,482 mi (41,009 km). They are used extensively for freight, recreation, and a small amount of passenger traffic. Of the world's 50 busiest container ports, four are located in the United States, with the busiest in the country being the Port of Los Angeles. Demographics The U.S. Census Bureau reported 331,449,281 residents on April 1, 2020,[v] making the United States the third-most-populous country in the world, after India and China. The Census Bureau's official 2025 population estimate was 341,784,857, an increase of 3.1% since the 2020 census. According to the Bureau's U.S. Population Clock, on July 1, 2024, the U.S. population had a net gain of one person every 16 seconds, or about 5400 people per day. In 2023, 51% of Americans age 15 and over were married, 6% were widowed, 10% were divorced, and 34% had never been married. In 2023, the total fertility rate for the U.S. stood at 1.6 children per woman, and, at 23%, it had the world's highest rate of children living in single-parent households in 2019. Most Americans live in the suburbs of major metropolitan areas. The United States has a diverse population; 37 ancestry groups have more than one million members. White Americans with ancestry from Europe, the Middle East, or North Africa form the largest racial and ethnic group at 57.8% of the United States population. Hispanic and Latino Americans form the second-largest group and are 18.7% of the United States population. African Americans constitute the country's third-largest ancestry group and are 12.1% of the total U.S. population. Asian Americans are the country's fourth-largest group, composing 5.9% of the United States population. The country's 3.7 million Native Americans account for about 1%, and some 574 native tribes are recognized by the federal government. In 2024, the median age of the United States population was 39.1 years. While many languages and dialects are spoken in the United States, English is by far the most commonly spoken and written. De facto, English is the official language of the United States, and in 2025, Executive Order 14224 declared English official. However, the U.S. has never had a de jure official language, as Congress has never passed a law to designate English as official for all three federal branches. Some laws, such as U.S. naturalization requirements, nonetheless standardize English. Twenty-eight states and the United States Virgin Islands have laws that designate English as the sole official language; 19 states and the District of Columbia have no official language. Three states and four U.S. territories have recognized local or indigenous languages in addition to English: Hawaii (Hawaiian), Alaska (twenty Native languages),[w] South Dakota (Sioux), American Samoa (Samoan), Puerto Rico (Spanish), Guam (Chamorro), and the Northern Mariana Islands (Carolinian and Chamorro). In total, 169 Native American languages are spoken in the United States. In Puerto Rico, Spanish is more widely spoken than English. According to the American Community Survey (2020), some 245.4 million people in the U.S. age five and older spoke only English at home. About 41.2 million spoke Spanish at home, making it the second most commonly used language. Other languages spoken at home by one million people or more include Chinese (3.40 million), Tagalog (1.71 million), Vietnamese (1.52 million), Arabic (1.39 million), French (1.18 million), Korean (1.07 million), and Russian (1.04 million). German, spoken by 1 million people at home in 2010, fell to 857,000 total speakers in 2020. America's immigrant population is by far the world's largest in absolute terms. In 2022, there were 87.7 million immigrants and U.S.-born children of immigrants in the United States, accounting for nearly 27% of the overall U.S. population. In 2017, out of the U.S. foreign-born population, some 45% (20.7 million) were naturalized citizens, 27% (12.3 million) were lawful permanent residents, 6% (2.2 million) were temporary lawful residents, and 23% (10.5 million) were unauthorized immigrants. In 2019, the top countries of origin for immigrants were Mexico (24% of immigrants), India (6%), China (5%), the Philippines (4.5%), and El Salvador (3%). In fiscal year 2022, over one million immigrants (most of whom entered through family reunification) were granted legal residence. The undocumented immigrant population in the U.S. reached a record high of 14 million in 2023. The First Amendment guarantees the free exercise of religion in the country and forbids Congress from passing laws respecting its establishment. Religious practice is widespread, among the most diverse in the world, and profoundly vibrant. The country has the world's largest Christian population, which includes the fourth-largest population of Catholics. Other notable faiths include Judaism, Buddhism, Hinduism, Islam, New Age, and Native American religions. Religious practice varies significantly by region. "Ceremonial deism" is common in American culture. The overwhelming majority of Americans believe in a higher power or spiritual force, engage in spiritual practices such as prayer, and consider themselves religious or spiritual. In the Southern United States' "Bible Belt", evangelical Protestantism plays a significant role culturally; New England and the Western United States tend to be more secular. Mormonism, a Restorationist movement founded in the U.S. in 1847, is the predominant religion in Utah and a major religion in Idaho. About 82% of Americans live in metropolitan areas, particularly in suburbs; about half of those reside in cities with populations over 50,000. In 2022, 333 incorporated municipalities had populations over 100,000, nine cities had more than one million residents, and four cities—New York City, Los Angeles, Chicago, and Houston—had populations exceeding two million. Many U.S. metropolitan populations are growing rapidly, particularly in the South and West. According to the Centers for Disease Control and Prevention (CDC), average U.S. life expectancy at birth reached 79.0 years in 2024, its highest recorded level. This was an increase of 0.6 years over 2023. The CDC attributed the improvement to a significant fall in the number of fatal drug overdoses in the country, noting that "heart disease continues to be the leading cause of death in the United States, followed by cancer and unintentional injuries." In 2024, life expectancy at birth for American men rose to 76.5 years (+0.7 years compared to 2023), while life expectancy for women was 81.4 years (+0.3 years). Starting in 1998, life expectancy in the U.S. fell behind that of other wealthy industrialized countries, and Americans' "health disadvantage" gap has been increasing ever since. The Commonwealth Fund reported in 2020 that the U.S. had the highest suicide rate among high-income countries. Approximately one-third of the U.S. adult population is obese and another third is overweight. The U.S. healthcare system far outspends that of any other country, measured both in per capita spending and as a percentage of GDP, but attains worse healthcare outcomes when compared to peer countries for reasons that are debated. The United States is the only developed country without a system of universal healthcare, and a significant proportion of the population that does not carry health insurance. Government-funded healthcare coverage for the poor (Medicaid) and for those age 65 and older (Medicare) is available to Americans who meet the programs' income or age qualifications. In 2010, then-President Obama passed the Patient Protection and Affordable Care Act.[x] Abortion in the United States is not federally protected, and is illegal or restricted in 17 states. American primary and secondary education, known in the U.S. as K–12 ("kindergarten through 12th grade"), is decentralized. School systems are operated by state, territorial, and sometimes municipal governments and regulated by the U.S. Department of Education. In general, children are required to attend school or an approved homeschool from the age of five or six (kindergarten or first grade) until they are 18 years old. This often brings students through the 12th grade, the final year of a U.S. high school, but some states and territories allow them to leave school earlier, at age 16 or 17. The U.S. spends more on education per student than any other country, an average of $18,614 per year per public elementary and secondary school student in 2020–2021. Among Americans age 25 and older, 92.2% graduated from high school, 62.7% attended some college, 37.7% earned a bachelor's degree, and 14.2% earned a graduate degree. The U.S. literacy rate is near-universal. The U.S. has produced the most Nobel Prize winners of any country, with 411 (having won 413 awards). U.S. tertiary or higher education has earned a global reputation. Many of the world's top universities, as listed by various ranking organizations, are in the United States, including 19 of the top 25. American higher education is dominated by state university systems, although the country's many private universities and colleges enroll about 20% of all American students. Local community colleges generally offer open admissions, lower tuition, and coursework leading to a two-year associate degree or a non-degree certificate. As for public expenditures on higher education, the U.S. spends more per student than the OECD average, and Americans spend more than all nations in combined public and private spending. Colleges and universities directly funded by the federal government do not charge tuition and are limited to military personnel and government employees, including: the U.S. service academies, the Naval Postgraduate School, and military staff colleges. Despite some student loan forgiveness programs in place, student loan debt increased by 102% between 2010 and 2020, and exceeded $1.7 trillion in 2022. Culture and society The United States is home to a wide variety of ethnic groups, traditions, and customs. The country has been described as having the values of individualism and personal autonomy, as well as a strong work ethic and competitiveness. Voluntary altruism towards others also plays a major role; according to a 2016 study by the Charities Aid Foundation, Americans donated 1.44% of total GDP to charity—the highest rate in the world by a large margin. Americans have traditionally been characterized by a unifying political belief in an "American Creed" emphasizing consent of the governed, liberty, equality under the law, democracy, social equality, property rights, and a preference for limited government. The U.S. has acquired significant hard and soft power through its diplomatic influence, economic power, military alliances, and cultural exports such as American movies, music, video games, sports, and food. The influence that the United States exerts on other countries through soft power is referred to as Americanization. Nearly all present Americans or their ancestors came from Europe, Africa, or Asia (the "Old World") within the past five centuries. Mainstream American culture is a Western culture largely derived from the traditions of European immigrants with influences from many other sources, such as traditions brought by slaves from Africa. More recent immigration from Asia and especially Latin America has added to a cultural mix that has been described as a homogenizing melting pot, and a heterogeneous salad bowl, with immigrants contributing to, and often assimilating into, mainstream American culture. Under the First Amendment to the Constitution, the United States is considered to have the strongest protections of free speech of any country. Flag desecration, hate speech, blasphemy, and lese majesty are all forms of protected expression. A 2016 Pew Research Center poll found that Americans were the most supportive of free expression of any polity measured. Additionally, they are the "most supportive of freedom of the press and the right to use the Internet without government censorship". The U.S. is a socially progressive country with permissive attitudes surrounding human sexuality. LGBTQ rights in the United States are among the most advanced by global standards. The American Dream, or the perception that Americans enjoy high levels of social mobility, plays a key role in attracting immigrants. Whether this perception is accurate has been a topic of debate. While mainstream culture holds that the United States is a classless society, scholars identify significant differences between the country's social classes, affecting socialization, language, and values. Americans tend to greatly value socioeconomic achievement, but being ordinary or average is promoted by some as a noble condition as well. The National Foundation on the Arts and the Humanities is an agency of the United States federal government that was established in 1965 with the purpose to "develop and promote a broadly conceived national policy of support for the humanities and the arts in the United States, and for institutions which preserve the cultural heritage of the United States." It is composed of four sub-agencies: Colonial American authors were influenced by John Locke and other Enlightenment philosophers. The American Revolutionary Period (1765–1783) is notable for the political writings of Benjamin Franklin, Alexander Hamilton, Thomas Paine, and Thomas Jefferson. Shortly before and after the Revolutionary War, the newspaper rose to prominence, filling a demand for anti-British national literature. An early novel is William Hill Brown's The Power of Sympathy, published in 1791. Writer and critic John Neal in the early- to mid-19th century helped advance America toward a unique literature and culture by criticizing predecessors such as Washington Irving for imitating their British counterparts, and by influencing writers such as Edgar Allan Poe, who took American poetry and short fiction in new directions. Ralph Waldo Emerson and Margaret Fuller pioneered the influential Transcendentalism movement; Henry David Thoreau, author of Walden, was influenced by this movement. The conflict surrounding abolitionism inspired writers, like Harriet Beecher Stowe, and authors of slave narratives, such as Frederick Douglass. Nathaniel Hawthorne's The Scarlet Letter (1850) explored the dark side of American history, as did Herman Melville's Moby-Dick (1851). Major American poets of the 19th century American Renaissance include Walt Whitman, Melville, and Emily Dickinson. Mark Twain was the first major American writer to be born in the West. Henry James achieved international recognition with novels like The Portrait of a Lady (1881). As literacy rates rose, periodicals published more stories centered around industrial workers, women, and the rural poor. Naturalism, regionalism, and realism were the major literary movements of the period. While modernism generally took on an international character, modernist authors working within the United States more often rooted their work in specific regions, peoples, and cultures. Following the Great Migration to northern cities, African-American and black West Indian authors of the Harlem Renaissance developed an independent tradition of literature that rebuked a history of inequality and celebrated black culture. An important cultural export during the Jazz Age, these writings were a key influence on Négritude, a philosophy emerging in the 1930s among francophone writers of the African diaspora. In the 1950s, an ideal of homogeneity led many authors to attempt to write the Great American Novel, while the Beat Generation rejected this conformity, using styles that elevated the impact of the spoken word over mechanics to describe drug use, sexuality, and the failings of society. Contemporary literature is more pluralistic than in previous eras, with the closest thing to a unifying feature being a trend toward self-conscious experiments with language. Twelve American laureates have won the Nobel Prize in Literature. Media in the United States is broadly uncensored, with the First Amendment providing significant protections, as reiterated in New York Times Co. v. United States. The four major broadcasters in the U.S. are the National Broadcasting Company (NBC), Columbia Broadcasting System (CBS), American Broadcasting Company (ABC), and Fox Broadcasting Company (Fox). The four major broadcast television networks are all commercial entities. The U.S. cable television system offers hundreds of channels catering to a variety of niches. In 2021, about 83% of Americans over age 12 listened to broadcast radio, while about 40% listened to podcasts. In the prior year, there were 15,460 licensed full-power radio stations in the U.S. according to the Federal Communications Commission (FCC). Much of the public radio broadcasting is supplied by National Public Radio (NPR), incorporated in February 1970 under the Public Broadcasting Act of 1967. U.S. newspapers with a global reach and reputation include The Wall Street Journal, The New York Times, The Washington Post, and USA Today. About 800 publications are produced in Spanish. With few exceptions, newspapers are privately owned, either by large chains such as Gannett or McClatchy, which own dozens or even hundreds of newspapers; by small chains that own a handful of papers; or, in an increasingly rare situation, by individuals or families. Major cities often have alternative newspapers to complement the mainstream daily papers, such as The Village Voice in New York City and LA Weekly in Los Angeles. The five most-visited websites in the world are Google, YouTube, Facebook, Instagram, and ChatGPT—all of them American-owned. Other popular platforms used include X (formerly Twitter) and Amazon. In 2025, the U.S. was the world's second-largest video game market by revenue (after China). In 2015, the U.S. video game industry consisted of 2,457 companies that employed around 220,000 jobs and generated $30.4 billion in revenue. There are 444 game publishers, developers, and hardware companies in California alone. According to the Game Developers Conference (GDC), the U.S. is the top location for video game development, with 58% of the world's game developers based there in 2025. The United States is well known for its theater. Mainstream theater in the United States derives from the old European theatrical tradition and has been heavily influenced by the British theater. By the middle of the 19th century, America had created new distinct dramatic forms in the Tom Shows, the showboat theater and the minstrel show. The central hub of the American theater scene is the Theater District in Manhattan, with its divisions of Broadway, off-Broadway, and off-off-Broadway. Many movie and television celebrities have gotten their big break working in New York productions. Outside New York City, many cities have professional regional or resident theater companies that produce their own seasons. The biggest-budget theatrical productions are musicals. U.S. theater has an active community theater culture. The Tony Awards recognizes excellence in live Broadway theater and are presented at an annual ceremony in Manhattan. The awards are given for Broadway productions and performances. One is also given for regional theater. Several discretionary non-competitive awards are given as well, including a Special Tony Award, the Tony Honors for Excellence in Theatre, and the Isabelle Stevenson Award. Folk art in colonial America grew out of artisanal craftsmanship in communities that allowed commonly trained people to individually express themselves. It was distinct from Europe's tradition of high art, which was less accessible and generally less relevant to early American settlers. Cultural movements in art and craftsmanship in colonial America generally lagged behind those of Western Europe. For example, the prevailing medieval style of woodworking and primitive sculpture became integral to early American folk art, despite the emergence of Renaissance styles in England in the late 16th and early 17th centuries. The new English styles would have been early enough to make a considerable impact on American folk art, but American styles and forms had already been firmly adopted. Not only did styles change slowly in early America, but there was a tendency for rural artisans there to continue their traditional forms longer than their urban counterparts did—and far longer than those in Western Europe. The Hudson River School was a mid-19th-century movement in the visual arts tradition of European naturalism. The 1913 Armory Show in New York City, an exhibition of European modernist art, shocked the public and transformed the U.S. art scene. American Realism and American Regionalism sought to reflect and give America new ways of looking at itself. Georgia O'Keeffe, Marsden Hartley, and others experimented with new and individualistic styles, which would become known as American modernism. Major artistic movements such as the abstract expressionism of Jackson Pollock and Willem de Kooning and the pop art of Andy Warhol and Roy Lichtenstein developed largely in the United States. Major photographers include Alfred Stieglitz, Edward Steichen, Dorothea Lange, Edward Weston, James Van Der Zee, Ansel Adams, and Gordon Parks. The tide of modernism and then postmodernism has brought global fame to American architects, including Frank Lloyd Wright, Philip Johnson, and Frank Gehry. The Metropolitan Museum of Art in Manhattan is the largest art museum in the United States and the fourth-largest in the world. American folk music encompasses numerous music genres, variously known as traditional music, traditional folk music, contemporary folk music, or roots music. Many traditional songs have been sung within the same family or folk group for generations, and sometimes trace back to such origins as the British Isles, mainland Europe, or Africa. The rhythmic and lyrical styles of African-American music in particular have influenced American music. Banjos were brought to America through the slave trade. Minstrel shows incorporating the instrument into their acts led to its increased popularity and widespread production in the 19th century. The electric guitar, first invented in the 1930s, and mass-produced by the 1940s, had an enormous influence on popular music, in particular due to the development of rock and roll. The synthesizer, turntablism, and electronic music were also largely developed in the U.S. Elements from folk idioms such as the blues and old-time music were adopted and transformed into popular genres with global audiences. Jazz grew from blues and ragtime in the early 20th century, developing from the innovations and recordings of composers such as W.C. Handy and Jelly Roll Morton. Louis Armstrong and Duke Ellington increased its popularity early in the 20th century. Country music developed in the 1920s, bluegrass and rhythm and blues in the 1940s, and rock and roll in the 1950s. In the 1960s, Bob Dylan emerged from the folk revival to become one of the country's most celebrated songwriters. The musical forms of punk and hip hop both originated in the United States in the 1970s. The United States has the world's largest music market, with a total retail value of $15.9 billion in 2022. Most of the world's major record companies are based in the U.S.; they are represented by the Recording Industry Association of America (RIAA). Mid-20th-century American pop stars, such as Frank Sinatra and Elvis Presley, became global celebrities and best-selling music artists, as have artists of the late 20th century, such as Michael Jackson, Madonna, Whitney Houston, and Mariah Carey, and of the early 21st century, such as Eminem, Britney Spears, Lady Gaga, Katy Perry, Taylor Swift and Beyoncé. The United States has the world's largest apparel market by revenue. Apart from professional business attire, American fashion is eclectic and predominantly informal. Americans' diverse cultural roots are reflected in their clothing; however, sneakers, jeans, T-shirts, and baseball caps are emblematic of American styles. New York, with its Fashion Week, is considered to be one of the "Big Four" global fashion capitals, along with Paris, Milan, and London. A study demonstrated that general proximity to Manhattan's Garment District has been synonymous with American fashion since its inception in the early 20th century. A number of well-known designer labels, among them Tommy Hilfiger, Ralph Lauren, Tom Ford and Calvin Klein, are headquartered in Manhattan. Labels cater to niche markets, such as preteens. New York Fashion Week is one of the most influential fashion shows in the world, and is held twice each year in Manhattan; the annual Met Gala, also in Manhattan, has been called the fashion world's "biggest night". The U.S. film industry has a worldwide influence and following. Hollywood, a district in central Los Angeles, the nation's second-most populous city, is also metonymous for the American filmmaking industry. The major film studios of the United States are the primary source of the most commercially successful movies selling the most tickets in the world. Largely centered in the New York City region from its beginnings in the late 19th century through the first decades of the 20th century, the U.S. film industry has since been primarily based in and around Hollywood. Nonetheless, American film companies have been subject to the forces of globalization in the 21st century, and an increasing number of films are made elsewhere. The Academy Awards, popularly known as "the Oscars", have been held annually by the Academy of Motion Picture Arts and Sciences since 1929, and the Golden Globe Awards have been held annually since January 1944. The industry peaked in what is commonly referred to as the "Golden Age of Hollywood", from the early sound period until the early 1960s, with screen actors such as John Wayne and Marilyn Monroe becoming iconic figures. In the 1970s, "New Hollywood", or the "Hollywood Renaissance", was defined by grittier films influenced by French and Italian realist pictures of the post-war period. The 21st century has been marked by the rise of American streaming platforms, which came to rival traditional cinema. Early settlers were introduced by Native Americans to foods such as turkey, sweet potatoes, corn, squash, and maple syrup. Of the most enduring and pervasive examples are variations of the native dish called succotash. Early settlers and later immigrants combined these with foods they were familiar with, such as wheat flour, beef, and milk, to create a distinctive American cuisine. New World crops, especially pumpkin, corn, potatoes, and turkey as the main course are part of a shared national menu on Thanksgiving, when many Americans prepare or purchase traditional dishes to celebrate the occasion. Characteristic American dishes such as apple pie, fried chicken, doughnuts, french fries, macaroni and cheese, ice cream, hamburgers, hot dogs, and American pizza derive from the recipes of various immigrant groups. Mexican dishes such as burritos and tacos preexisted the United States in areas later annexed from Mexico, and adaptations of Chinese cuisine as well as pasta dishes freely adapted from Italian sources are all widely consumed. American chefs have had a significant impact on society both domestically and internationally. In 1946, the Culinary Institute of America was founded by Katharine Angell and Frances Roth. This would become the United States' most prestigious culinary school, where many of the most talented American chefs would study prior to successful careers. The United States restaurant industry was projected at $899 billion in sales for 2020, and employed more than 15 million people, representing 10% of the nation's workforce directly. It is the country's second-largest private employer and the third-largest employer overall. The United States is home to over 220 Michelin star-rated restaurants, 70 of which are in New York City. Wine has been produced in what is now the United States since the 1500s, with the first widespread production beginning in what is now New Mexico in 1628. In the modern U.S., wine production is undertaken in all fifty states, with California producing 84 percent of all U.S. wine. With more than 1,100,000 acres (4,500 km2) under vine, the United States is the fourth-largest wine-producing country in the world, after Italy, Spain, and France. The classic American diner, a casual restaurant type originally intended for the working class, emerged during the 19th century from converted railroad dining cars made stationary. The diner soon evolved into purpose-built structures whose number expanded greatly in the 20th century. The American fast-food industry developed alongside the nation's car culture. American restaurants developed the drive-in format in the 1920s, which they began to replace with the drive-through format by the 1940s. American fast-food restaurant chains, such as McDonald's, Burger King, Chick-fil-A, Kentucky Fried Chicken, Dunkin' Donuts and many others, have numerous outlets around the world. The most popular spectator sports in the U.S. are American football, basketball, baseball, soccer, and ice hockey. Their premier leagues are, respectively, the National Football League, the National Basketball Association, Major League Baseball, Major League Soccer, and the National Hockey League, All these leagues enjoy wide-ranging domestic media coverage and, except for the MLS, all are considered the preeminent leagues in their respective sports in the world. While most major U.S. sports such as baseball and American football have evolved out of European practices, basketball, volleyball, skateboarding, and snowboarding are American inventions, many of which have become popular worldwide. Lacrosse and surfing arose from Native American and Native Hawaiian activities that predate European contact. The market for professional sports in the United States was approximately $69 billion in July 2013, roughly 50% larger than that of Europe, the Middle East, and Africa combined. American football is by several measures the most popular spectator sport in the United States. Although American football does not have a substantial following in other nations, the NFL does have the highest average attendance (67,254) of any professional sports league in the world. In the year 2024, the NFL generated over $23 billion, making them the most valued professional sports league in the United States and the world. Baseball has been regarded as the U.S. "national sport" since the late 19th century. The most-watched individual sports in the U.S. are golf and auto racing, particularly NASCAR and IndyCar. On the collegiate level, earnings for the member institutions exceed $1 billion annually, and college football and basketball attract large audiences, as the NCAA March Madness tournament and the College Football Playoff are some of the most watched national sporting events. In the U.S., the intercollegiate sports level serves as the main feeder system for professional and Olympic sports, with significant exceptions such as Minor League Baseball. This differs greatly from practices in nearly all other countries, where publicly and privately funded sports organizations serve this function. Eight Olympic Games have taken place in the United States. The 1904 Summer Olympics in St. Louis, Missouri, were the first-ever Olympic Games held outside of Europe. The Olympic Games will be held in the U.S. for a ninth time when Los Angeles hosts the 2028 Summer Olympics. U.S. athletes have won a total of 2,968 medals (1,179 gold) at the Olympic Games, the most of any country. In other international competition, the United States is the home of a number of prestigious events, including the America's Cup, World Baseball Classic, the U.S. Open, and the Masters Tournament. The U.S. men's national soccer team has qualified for eleven World Cups, while the women's national team has won the FIFA Women's World Cup and Olympic soccer tournament four and five times, respectively. The 1999 FIFA Women's World Cup was hosted by the United States. Its final match was attended by 90,185, setting the world record for largest women's sporting event crowd at the time. The United States hosted the 1994 FIFA World Cup and will co-host, along with Canada and Mexico, the 2026 FIFA World Cup. See also Notes References This article incorporates text from a free content work. Licensed under CC BY-SA IGO 3.0 (license statement/permission). Text taken from World Food and Agriculture – Statistical Yearbook 2023, FAO, FAO. External links 40°N 100°W / 40°N 100°W / 40; -100 (United States of America) |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Thirty-seventh_government_of_Israel#cite_note-13] | [TOKENS: 9915] |
Contents Thirty-seventh government of Israel The thirty-seventh government of Israel is the current cabinet of Israel, formed on 29 December 2022, following the Knesset election the previous month. The coalition government currently consists of five parties — Likud, Shas, Otzma Yehudit, Religious Zionist Party and New Hope — and is led by Benjamin Netanyahu, who took office as the prime minister of Israel for the sixth time. The government is widely regarded as the most right-wing government in the country's history, and includes far-right politicians. Several of the government's policy proposals have led to controversies, both within Israel and abroad, with the government's attempts at reforming the judiciary leading to a wave of demonstrations across the country. Following the outbreak of the Gaza war, opposition leader Yair Lapid initiated discussions with Netanyahu on the formation of an emergency government. On 11 October 2023, National Unity MKs Benny Gantz, Gadi Eisenkot, Gideon Sa'ar, Hili Tropper, and Yifat Shasha-Biton joined the Security Cabinet of Israel to form an emergency national unity government. Their accession to the Security Cabinet and to the government (as ministers without portfolio) was approved by the Knesset the following day. Gantz, Netanyahu, and Defense Minister Yoav Gallant became part of the newly formed Israeli war cabinet, with Eisenkot and Ron Dermer serving as observers. National Unity left the government in June 2024. New Hope rejoined the government in September. Otzma Yehudit announced on 19 January 2025 that it had withdrawn from the government, which took effect on 21 January, following the cabinet's acceptance of the three-phase Gaza war ceasefire proposal, though it rejoined two months later. United Torah Judaism left the government in July 2025 over dissatisfaction with the government's draft conscription law. Shas left the government several days later, though it remains part of the coalition. Background The right-wing bloc of parties, led by Benjamin Netanyahu, known in Israel as the national camp, won 64 of the 120 seats in the elections for the Knesset, while the coalition led by the incumbent prime minister Yair Lapid won 51 seats. The new majority has been variously described as the most right-wing government in Israeli history, as well as Israel's most religious government. Shortly after the elections, Lapid conceded to Netanyahu, and congratulated him, wishing him luck "for the sake of the Israeli people". On 15 November, the swearing-in ceremony for the newly elected members of the 25th Knesset was held during the opening session. The vote to appoint a new Speaker of the Knesset, which is usually conducted at the opening session, as well as the swearing in of cabinet members were postponed since ongoing coalition negotiations had not yet resulted in agreement on these positions. Government formation Yair Lapid Yesh Atid Benjamin Netanyahu Likud On 3 November 2022, Netanyahu told his aide Yariv Levin to begin informal coalition talks with allied parties, after 97% of the vote was counted. The leader of the Shas party Aryeh Deri met with Yitzhak Goldknopf, the leader of United Torah Judaism and its Agudat Yisrael faction, on 4 November. The two parties agreed to cooperate as members of the next government. The Degel HaTorah faction of United Torah Judaism stated on 5 November that it will maintain its ideological stance about not seeking any ministerial posts, as per the instruction of its spiritual leader Rabbi Gershon Edelstein, but will seek other senior posts like Knesset committee chairmen and deputy ministers. Netanyahu himself started holding talks on 6 November. He first met with Moshe Gafni, the leader of Degel HaTorah, and then with Goldknopf. Meanwhile, the Religious Zionist Party leader Bezalel Smotrich and the leader of its Otzma Yehudit faction Itamar Ben-Gvir pledged that they would not enter the coalition without the other faction. Gafni later met with Smotrich for coalition talks. Smotrich then met with Netanyahu. On 7 November, Netanyahu met with Ben-Gvir who demanded the Ministry of Public Security with expanded powers for himself and the Ministry of Education or Transport and Road Safety for Yitzhak Wasserlauf. A major demand among all of Netanyahu's allies was that the Knesset be allowed to ignore the rulings of the Supreme Court. Netanyahu met with the Noam faction leader and its sole MK Avi Maoz on 8 November after he threatened to boycott the coalition. He demanded complete control of the Western Wall by the Haredi rabbinate and removal of what he considered as anti-Zionist and anti-Jewish content in schoolbooks. President Isaac Herzog began consultations with heads of all the political parties on 9 November after the election results were certified. During the consultations, he expressed his reservations about Ben-Gvir becoming a member in the next government. Shas met with Likud for coalition talks on 10 November. By 11 November, Netanyahu had secured recommendations from 64 MKs, which constituted a majority. He was given the mandate to form the thirty-seventh government of Israel by President Herzog on 13 November. Otzma Yehudit and Noam officially split from Religious Zionism on 20 November as per a pre-election agreement. On 25 November, Otzma Yehudit and Likud signed a coalition agreement, under which Ben-Gvir will assume the newly created position of National Security Minister, whose powers would be more expansive than that of the Minister of Public Security, including overseeing the Israel Police and the Israel Border Police in the West Bank, as well as giving powers to authorities to shoot thieves stealing from military bases. Yitzhak Wasserlauf was given the Ministry for the Development of the Negev and the Galilee with expanded powers to regulate new West Bank settlements, while separating it from the "Periphery" portfolio, which will be given to Shas. The deal also includes giving the Ministry of Heritage to Amihai Eliyahu, separating it from the "Jerusalem Affairs" portfolio, the chairmanship of the Knesset's Public Security Committee to Zvika Fogel and that of the Special Committee for the Israeli Citizens' Fund to Limor Son Har-Melech, the post of Deputy Economic Minister to Almog Cohen, establishment of a national guard, and expansion of mobilization of reservists in the Border Police. Netanyahu and Maoz signed a coalition agreement on 27 November, under which the latter would become a deputy minister, would head an agency on Jewish identity in the Prime Minister's Office, and would also head Nativ, which processes the aliyah from the former Soviet Union. The agency for Jewish identity would have authority over educational content taught outside the regular curriculum in schools, in addition to the department of the Ministry of Education overseeing external teaching and partnerships, which would bring nonofficial organisations permitted to teach and lecture at schools under its purview. Likud signed a coalition agreement with the Religious Zionist Party on 1 December. Under the deal, Smotrich would serve as the Minister of Finance in rotation with Aryeh Deri, and the party will receive the post of a minister within the Ministry of Defense with control over the departments administering settlement and open lands under the Coordinator of Government Activities in the Territories, in addition to another post of a deputy minister. The deal also includes giving the post of Minister of Aliyah and Integration to Ofir Sofer, the newly created National Missions Ministry to Orit Strook, and the chairmanship of the Knesset's Constitution, Law and Justice Committee to Simcha Rothman. Likud and United Torah Judaism signed a coalition agreement on 6 December, to allow request for an extension to the deadline. Under it, the party would receive the Ministry of Construction and Housing, the chairmanship of the Knesset Finance Committee which will be given to Moshe Gafni, the Ministry of Jerusalem and Tradition (which would replace the Ministry of Jerusalem Affairs and Heritage), in addition to several posts of deputy ministers and chairmanships of Knesset committees. Likud also signed a deal with Shas by 8 December, securing interim coalition agreements with all of their allies. Under the deal, Deri will first serve as the Minister of Interior and Health, before rotating posts with Smotrich after two years. The party will also receive the Ministry of Religious Services and Welfare Ministries, as well as posts of deputy ministers in the Ministry of Education and Interior. The vote to replace then-incumbent Knesset speaker Mickey Levy was scheduled for 13 December, after Likud and its allies secured the necessary number of signatures for it. Yariv Levin of Likud was elected as an interim speaker by 64 votes, while his opponents Merav Ben-Ari of Yesh Atid and Ayman Odeh of Hadash received 45 and five votes respectively. Netanyahu asked Herzog for a 14-day extension after the agreement with Shas to finalise the roles his allied parties would play. Herzog on 9 December extended the deadline to 21 December. On that date, Netanyahu informed Herzog that he had succeeded in forming a coalition, with the new government expected to be sworn in by 2 January 2023. The government was sworn in on 29 December 2022. Timeline Israeli law stated that people convicted of crimes cannot serve in the government. An amendment to that law was made in late 2022, known colloquially as the Deri Law, to allow those who had been convicted without prison time to serve. This allowed Deri to be appointed to the cabinet. Shas leader Aryeh Deri was appointed to be Minister of Health, Minister of the Interior, and Vice Prime Minister in December 2022. He was fired in January 2023, following a Supreme Court decision that his appointment was unreasonable, since he had been convicted of fraud, and had promised not to seek government roles through a plea deal. In March 2023, Defence Minister Yoav Gallant called on the government to delay legislation related to the judicial reform. Prime Minister Netanyahu announced that he had been dismissed from his position, leading to the continuation of mass protests across the country (which had started in January in Tel Aviv). Gallant continued to serve as a minister as he had not received formal notice of dismissal, and two weeks later it was announced that Netanyahu had reversed his decision. Public Safety Minister Itamar Ben-Gvir (Otzma Yehudit leader) and Minister of Justice Yariv Levin (Likud) both threatened to resign if the judicial reform was delayed.[better source needed] After the outbreak of the Gaza war, five members of the National Unity party joined the government as ministers without portfolio, with leader Benny Gantz being made a member of the new Israeli war cabinet (along with Netanyahu and Gallant). As the war progressed, minister of national security Itamar Ben-Gvir threatened to leave the government if the war was ended. A month later in mid December, he again threatened to leave if the war did not maintain "full strength". Gideon Sa'ar stated on 16 March that his New Hope party would resign from the government and join the opposition if Prime Minister Benjamin Netanyahu did not appoint him to the Israeli war cabinet. Netanyahu did not do so, resulting in Sa'ar's New Hope party leaving the government nine days later, reducing the size of the coalition from 76 MKs to 72. Ben-Gvir and Bezalel Smotrich, of the National Religious Party–Religious Zionism party, have indicated that they will withdraw their parties from the government if the January 2025 Gaza war ceasefire is adopted, which would bring down the government. Ben-Gvir announced on 5 June that the members of his party would be allowed to vote as they wish, though his party resumed support on 9 June. On 18 May, Gantz set an 8 June deadline for withdrawal from the coalition, which was delayed by a day following the 2024 Nuseirat rescue operation. Gantz and his party left the government on 9 June, giving the government 64 seats in the Knesset. Sa'ar and his New Hope party rejoined the Netanyahu government on 30 September, increasing the number of seats held by the government to 68. The High Court of Justice ruled on 28 March 2024 that yeshiva funds would no longer be available for students who are "eligible for enlistment", effectively allowing ultra-Orthodox Jews to be drafted into the IDF. Attorney general Gali Baharav-Miara indicated on 31 March that the conscription process must begin on 1 April. The court ruled on 25 June that the IDF must begin to draft yeshiva students. Likud announced on 7 July that it would not put forward any legislation after Shas and United Torah Judaism said that they would boycott the plenary session over the lack of legislation dealing with the Haredi draft. The Ultra-Orthodox boycott continued for a second day, with UTJ briefly ending its boycott on 9 July to unsuccessfully vote in favor of a bill which would have weakened the Law of Return. Yuli Edelstein, who was replaced by Boaz Bismuth on the Foreign Affairs and Defense Committee in early August, published a draft version of the conscription law shortly before his ouster. Bismuth cancelled the work on the draft law in September 2025, which Edelstein called "a shame." Bismuth released the official version of the draft law in late November 2025. It weakened penalties for draft evaders, with Edelstein saying it was "the exact opposite" of the bill which he attempted to pass. Members of Otzma Yehudit resigned from the government on 19 January 2025 over the January 2025 Gaza war ceasefire, which took effect on 21 January. The members rejoined in March, following the "resumption" of the war in Gaza. Avi Maoz of the Noam party left the government in March 2025. On 4 June 2025, senior rabbis for United Torah Judaism Dov Lando and Moshe Hillel Hirsch instructed the party's MKs to pass a bill which would dissolve the Knesset. Yesh Atid, Yisrael Beytenu and The Democrats announced that they will "submit a bill" for dissolution on 11 June, with Yesh Atid tabling the bill on 4 June. There were also reports that Shas would vote in favor of Knesset dissolution amidst division within the governing coalition on Haredi conscription. This jeopardized the coalition's majority and would have triggered new elections if the bill passed. The following day, Agudat Yisrael, one of the United Torah Judaism factions, confirmed that it would submit a bill to dissolve the Knesset. Asher Medina, a Shas spokesman, indicated on 9 June that the party would vote in favor of a preliminary bill to dissolve the Knesset. The rabbis of Degel HaTorah instructed the parties' MKs on 12 June 2025 to oppose the dissolution of the Knesset, which was followed by Yuli Edelstein and the Shas and Degel HaTorah parties announcing that a deal had been reached, with "rabbinical leaders" telling their parties to delay the dissolution vote by a week. Shas and Degel HaTorah voted against the dissolution bill, which led to the bill failing its preliminary reading in a vote of 61 against and 53 in favor. MKs Ya'akov Tessler and Moshe Roth of Agudat Yisrael voted in favor of dissolution. Another dissolution bill will be unable to be brought forward for six months. If the bill had passed its preliminary reading, in addition to three more readings, an election would have been held in approximately three months; The Jerusalem Post posited it would have been held in October. Degel HaTorah announced on 14 July 2025 that it would leave the government because members of the party were dissatisfied after viewing the proposed draft bill by Yuli Edelstein regarding Haredi exemptions from the Israeli draft. Several hours later, Agudat Yisrael announced that it would also leave the government. Deputy Transportation Minister Uri Maklev, Moshe Gafni, the head of the Knesset Finance Committee, Ya'akov Asher, the head of the Knesset Interior and Environment Protection Committee and Jerusalem Affairs minister Meir Porush all submitted their resignations, with their resignations taking effect in 48 hours. Sports Minister Ya'akov Tessler and "Special Committee for Public Petitions Chair" Yitzhak Pindrus also submitted resignations. Yisrael Eichler submitted his resignation as the "head of the Knesset Labor and Welfare Committee" the same day. The resignations will leave Netanyahu's government with a 60-seat majority in the Knesset, as Avi Maoz, of the Noam party, left the government in March 2025. Despite Edelstein's ouster in August, a spokesman for UTJ head Yitzhak Goldknopf remarked that it would not change the faction's withdrawal from the government. The religious council for Shas, called the Moetzet Chachmei HaTorah, instructed the party on 16 July to leave the government, but stay in the coalition. The following day, various cabinet ministers submitted their resignations, including "Interior Minister Moshe Arbel, Social Affairs Minister Ya'akov Margi and Religious Services Minister Michael Malchieli." Malchieli reportedly has postponed his resignation so he could attend a 20 July meeting of the panel investigating whether attorney general Gali Baharav-Miara should be dismissed. Deputy Minister of Agriculture Moshe Abutbul, Minister of Health Uriel Buso and Haim Biton, a minister in the Education Ministry, also submitted their resignation letters, while Arbel retracted his resignation letter. The last cabinet member from the party to submit it was Labor Minister Yoav Ben-Tzur. The ministers who resigned will return to the Knesset, replacing MKs Moshe Roth, Yitzhak Pindrus and Eliyahu Baruchi. Members of government Listed below are the current ministers in the government: Principles and priorities According to the agreements signed between Likud and each of its coalition partners, and the incoming government's published guideline principles, its stated priorities are to combat the cost of living, further centralize Orthodox control over the state religious services, pass judicial reforms which include legislation to reduce judicial controls on executive and legislative power, expand settlements in the West Bank, and consider an annexation of the West Bank. Before the vote of confidence in his new government in the Knesset, Netanyahu presented three top priorities for the new government: internal security and governance, halting the nuclear program of Iran, and the development of infrastructure, with a focus on further connecting the center of the country with its periphery. Policies The government's flagship program, centered around reforms in the judicial branch, drew widespread criticism. Critics said it would have negative effects on the separation of powers, the office of the Attorney General, the economy, public health, women and minorities, workers' rights, scientific research, the overall strength of Israel's democracy and its foreign relations. After weeks of public protests on Israel's streets, joined by a growing number of military reservists, Minister of Defense Yoav Gallant spoke against the reform on 25 March, calling for a halt of the legislative process "for the sake of Israel's security". The next day, Netanyahu announced that he would be removed from his post, sparking another wave of protest across Israel and ultimately leading to Netanyahu agreeing to pause the legislation. On 10 April, Netanyahu announced that Gallant would keep his post. On 27 March 2023, after the public protests and general strikes, Netanyahu announced a pause in the reform process to allow for dialogue with opposition parties. However, negotiations aimed at reaching a compromise collapsed in June, and the government resumed its plans to unilaterally pass parts of the legislation. On 24 July 2023, the Knesset passed a bill that curbs the power of the Supreme Court to declare government decisions unreasonable; on 1 January 2024, the Supreme Court struck the bill down. The Knesset passed a "watered-down" version of the judicial reform package in late March 2025 which "changes the composition" of the judicial selection committee. In December 2022 Minister of National Security Itamar Ben-Gvir sought to amend the law that regulates the operations of the Israel Police, such that the ministry will have more direct control of its forces and policies, including its investigative priorities. Attorney General Gali Baharav-Miara objected to the draft proposal, raising concerns that the law would enable the politicization of police work, and the draft was amended to partially address those concerns. Nevertheless, in March 2023 Deputy Attorney General Gil Limon stated that the Attorney General's fears had been realized, referring to several instances of ministerial involvement in the day-to-day work of the otherwise independent police force – statements that were repeated by the Attorney General herself two days later. Separately, Police Commissioner Kobi Shabtai instructed Deputy Commissioners to avoid direct communication with the minister, later stating that "the Israel Police will remain apolitical, and act only according to law". Following appeals by the Association for Civil Rights in Israel and the Movement for Quality Government in Israel, the High Court of Justice instructed Ben-Gvir "to refrain from giving operational directions to the police... [especially] as regards to protests and demonstrations against the government." As talks of halting the judicial reform gained wind during March 2023, Minister of National Security Itamar Ben-Gvir threatened to resign if the legislation implementing the changes was suspended. To appease Ben-Gvir, Prime Minister Netanyahu announced that the government would promote the creation of a new National Guard, to be headed by Ben-Gvir. On 29 March, thousands of Israelis demonstrated in Tel Aviv, Haifa and Jerusalem against this decision. On 1 April, the New York Times quoted Gadeer Nicola, head of the Arab department at the Association for Civil Rights in Israel, as saying "If this thing passes, it will be an imminent danger to the rights of Arab citizens in this country. This will create two separate systems of applying the law. The regular police which will operate against Jewish citizens — and a militarized militia to deal only with Arab citizens." The same day, while speaking on Israel's Channel 13 about those whom he'd like to see enlist in the National Guard, Ben-Gvir specifically mentioned La Familia, the far-right fan club of the Beitar Jerusalem soccer team. On 2 April, Israel's cabinet approved the establishment of a law enforcement body that would operate independently of the police, under Ben-Gvir's authority. According to the decision, the Minister was to establish a committee chaired by the Director General of the Ministry of National Security, with representatives of the ministries of defense, justice and finance, as well as the police and the IDF, to outline the operations of the new organization. The committee's recommendations will be submitted to the government for consideration. Addressing a conference on 4 April, Police Commissioner Kobi Shabtai said that he is not opposed to the establishment of a security body which would answer to the police, but "a separate body? Absolutely not." The police chief said he had warned Ben-Gvir that the establishment of a security body separate from the police is "unnecessary, with extremely high costs that may harm citizens' personal security." During a press conference on 10 April, Prime Minister Netanyahu said, in what has been seen by some news outlets as a concession to the protesters, that "This will not be anyone's militia, it will be a security body, orderly, professional, that will be subordinate to one of the [existing] security bodies." The committee established by the government recommended the government to order the establishment of the National Guard immediately while allocating budgets. The National Guard, under whose command will be a superintendent of the police, will not be subordinate to Ben-Gvir. It will be subordinate to the police commissioner and will be part of Israel Border Police. The Ministry of Defense and Finance opposed the conclusions. The Israeli National Security Council called for further discussion on this. The coalition's efforts to expand the purview of Rabbinical courts; force some organizations, such as hospitals, to enforce certain religious practices; amend the Law Prohibiting Discrimination to allow gender segregation and discrimination on the grounds of religious belief; expand funding for religious causes; and put into law the exemption of yeshiva and kolel students from conscription have drawn criticism. According to the Haaretz op-ed of 7 March 2023, "the current coalition is interested... in modifying the public space so it suits the religious lifestyle. The legal coup is meant to castrate anyone who can prevent it, most of all the HCJ." Several banks and institutional investors, including the Israel Discount Bank and AIG have committed to avoid investing in, or providing credit to any organization that will discriminate against others on ground of religion, race, gender or sexual orientation. A series of technology companies and investment firms including Wiz, Intel Israel, Salesforce and Microsoft Israel Research and Development, have criticized the proposed changes to the Law Prohibiting Discrimination, with Wiz stating that it will require its suppliers to commit to preventing discrimination. Over sixty prominent law firms pledged that they will neither represent, nor do business with discriminating individuals and organizations. Insight Partners, a major private equity fund operating in Israel, released a statement warning against intolerance and any attempt to harm personal liberties. Orit Lahav, chief executive of the women's rights organization Mavoi Satum ("Dead End"), said that "the Rabbinical courts are the most discriminatory institution in the State of Israel... Limiting the HCJ[d] while expanding the jurisdiction of the Rabbinical courts would... cause significant harm to women." Anat Thon Ashkenazy, Director of the Center for Democratic Values and Institutions at the Israel Democracy Institute, said that "almost every part of the reform could harm women... the meaning of an override clause is that even if the court says that the law on gender segregation is illegitimate, is harmful, the Knesset could say 'Okay, we say otherwise'". She added that "there is a very broad institutional framework here, after which there will come legislation that harms women's right and we will have no way of protecting or stopping it." During July 2023, 20 professional medical associations signed a letter of position warning against the ramifications to public health that would result from the exclusion of women from the public sphere. They cited, among others, a rise in prevalence of risk factors for cardiovascular disease, pregnancy-related ailments, psychological distress, and the risk of suicide. On 30 July the Knesset passed an amendment to penal law adding sexual offenses to those offenses whose penalty can be doubled if done on grounds of "nationalistic terrorism, racism or hostility towards a certain community". According to MK Limor Son Har-Melech, the bill is meant to penalize any individual who "[intends to] harm a woman sexually based on her Jewishness". The law was criticized by MK Gilad Kariv as "populist, nationalistic, and dangerous towards the Arab citizens of Israel", and by MK Ahmad Tibi as a "race law", and was objected to by legal advisors at the Ministry of Justice and the Knesset Committee on National Security. Activist Orit Kamir wrote that "the amendment... is neither feminist, equal, nor progressive, but the opposite: it subordinates women's sexuality to the nationalistic, racist patriarchy. It hijacks the Law for Prevention of Sexual Harassment to serve a world view that tags women as sexual objects that personify the nation's honor." Yael Sherer, director of the Lobby to Combat Sexual Violence, criticized the law as being informed by dated ideas about sexual assault, and proposed that MKs "dedicate a session... to give victims of sexual assault an opportunity to come out of the darkness... instead of [submitting] declarative bills that change nothing and are not meant but for grabbing headlines". In Israel, during 2022, 24 women "were murdered because they were women," which was an increase of 50% compared to 2021. A law permitting courts to order men subject to a restraining order following domestic violence offenses to wear electronic tags was drafted during the previous Knesset and had passed its first reading unanimously. On 22 March 2023, the Knesset voted to reject the bill. It had been urged to do so by National Security Minister Itamar Ben-Gvir, who said that the bill was unfair to men. Earlier in the week, Ben-Gvir had blocked the measure from advancing in the ministerial legislative committee. The MKs voting against the bill included Prime Minister Netanyahu. The Association of Families of Murder Victims said that by rejecting the law, National Security Minister Itamar Ben-Gvir "brings joy to violent men and abandons the women threatened with murder… unsupervised restraining orders endanger women's lives even more. They give women the illusion of being protected, and then they are murdered." MK Pnina Tamano-Shata, chairwoman of the Knesset Committee on the Status of Women and Gender Equality, said that "the coalition proved today that it despises women's lives." The NGO Amutat Bat Melech [he], which assists Orthodox and ultra-Orthodox women who suffer from domestic violence, said that: "Rejecting the electronic bracelet bill is disconnected from the terrible reality of seven femicides since the beginning of the year. This is an effective tool of the first degree that could have saved lives and reduced the threat to women suffering from domestic violence. This is a matter of life and death, whose whole purpose is to provide a solution to defend women." The agreement signed by the coalition parties includes the setting up of a committee to draft changes to the Law of Return. Israeli religious parties have long demanded that the "grandchild clause" of the Law of Return be cancelled. This clause grants citizenship to anyone with at least one Jewish grandparent, as long as they do not practice another religion. If the grandchild clause were to be removed from the Law of Return then around 3 million people who are currently eligible for aliyah would no longer be eligible. The heads of the Jewish Agency, the Jewish Federations of North America, the World Zionist Organization and Keren Hayesod sent a joint letter to Prime Minister Netanyahu, expressing their "deep concern" about any changes to the Law of Return, adding that "Any change in the delicate and sensitive status quo on issues such as the Law of Return or conversion could threaten to unravel the ties between us and keep us away from each other." The Executive Council of Australian Jewry and the Zionist Federation of Australia issued a joint statement saying "We… view with deep concern… proposals in relation to religious pluralism and the law of return that risk damaging Israel's… relationship with Diaspora Jewry." On 19 March 2023, Israeli Finance Minister Bezalel Smotrich spoke in Paris at a memorial service for a Likud activist. The lectern at which Smotrich spoke was covered with a flag depicting the 'Greater Land of Israel,' encompassing the whole of Mandatory Palestine, as well as Trans-Jordan. During his speech, Smotrich said that "there's no such thing as Palestinians because there's no such thing as a Palestinian people." He added that the Palestinian people are a fictitious nation invented only to fight the Zionist movement, asking "Is there a Palestinian history or culture? There isn't any." The event received widespread media coverage. On 21 March, a spokesman for the US State Department sharply criticized Smotrich's comments. "The comments, which were delivered at a podium adorned with an inaccurate and provocative map, are offensive, they are deeply concerning, and, candidly, they're dangerous. The Palestinians have a rich history and culture, and the United States greatly values our partnership with the Palestinian people," he said. The Jordanian Foreign Ministry also voiced disapproval: "The Israeli Minister of Finance's use, during his participation in an event held yesterday in Paris, of a map of Israel that includes the borders of the Hashemite Kingdom of Jordan and the occupied Palestinian territories represents a reckless inflammatory act, and a violation of international norms and the Jordanian-Israeli peace treaty." Additionally, a map encompassing Mandatory Palestine and Trans-Jordan with a Jordanian flag on it was placed on a central lectern in the Jordanian Parliament. Jordan's parliament voted to expel the Israeli ambassador. Israel's Ministry of Foreign Affairs released a clarification relating to the matter, stating that "Israel is committed to the 1994 peace agreement with Jordan. There has been no change in the position of the State of Israel, which recognizes the territorial integrity of the Hashemite Kingdom of Jordan". Ahead of a Europe Day event due to take place on 9 May 2023, far-right wing National Security Minister Itamar Ben-Gvir was assigned as a representative of the government and a speaker at the event by the government secretariat, which deals with placing ministers at receptions on the occasion of the national days of the foreign embassies. The European Union requested that Ben-Gvir not attend, but the government did not make changes to the plan. On 8 May, the European delegation to Israel cancelled the reception, stating that: "The EU Delegation to Israel is looking forward to celebrating Europe Day on May 9, as it does every year. Regrettably, this year we have decided to cancel the diplomatic reception, as we do not want to offer a platform to someone whose views contradict the values the European Union stands for. However, the Europe Day cultural event for the Israeli public will be maintained to celebrate with our friends and partners in Israel the strong and constructive bilateral relationship". Israel's Opposition Leader Yair Lapid stated: "Sending Itamar Ben-Gvir to a gathering of EU ambassadors is a serious professional mistake. The government is embarrassing a large group of friendly countries, jeopardizing future votes in international institutions, and damaging our foreign relations. Last year, after a decade of efforts, we succeeded in signing an economic-political agreement with the European Union that will contribute to the Israeli economy and our foreign relations. Why risk it, and for what? Ben-Gvir is not a legitimate person in the international community (and not really in Israel either), and sometimes you have to be both wise and just and simply send someone else". On 23 February 2023, Defense Minister Gallant signed an agreement assigning governmental powers in the West Bank to a body to be headed by Minister Bezalel Smotrich, who will effectively become the governor of the West Bank, controlling almost all areas of life in the area, including planning, building and infrastructure. Israeli governments have hitherto been careful to keep the occupation as a military government. The temporary holding of power by an occupying military force, pending a negotiated settlement, is a principle of international law – an expression of the prohibition against obtaining sovereignty through conquest that was introduced in the wake of World War II. An editorial in Haaretz noted that the assignment of governmental powers in the West Bank to a civilian governor, alongside the plan to expand the dual justice system so that Israeli law will apply fully to settlers in the West Bank, constitutes de jure annexation of the West Bank. On 26 February 2023, following the 2023 Huwara shooting in which two Israelis were killed by an unidentified attacker, hundreds of Israeli settlers attacked the Palestinian town of Huwara and three nearby villages, setting alight hundreds of Palestinian homes (some with people in them), businesses, a school, and numerous vehicles, killing one Palestinian man and injuring 100 others. Bezalel Smotrich subsequently called on Twitter for Huwara to be "wiped out" by the Israeli government. Zvika Fogel MK, of the ultra-nationalist Otzma Yehudit, which forms part of the governing coalition, said that he "looks very favorably upon" the results of the rampage. Members of the coalition proposed an amendment to the Disengagement Law, which would allow Israelis to resettle settlements vacated during the 2005 Israeli disengagement from Gaza and the northern West Bank. The evacuated settlements were considered illegal under international law, according to most countries. The proposal was approved for voting by the Foreign Affairs and Defense Committee on 9 March 2023, while the committee was still waiting for briefing materials from the NSS, IDF, MFA and Shin Bet, and was passed on 21 March. The US has requested clarification from Israeli ambassador Michael Herzog. A US State Department spokesman stated that "The U.S. strongly urges Israel to refrain from allowing the return of settlers to the area covered by the legislation, consistent with both former Prime Minister Sharon and the current Israeli Government's commitment to the United States," noting that the actions represent a clear violation of undertakings given by the Sharon government to the Bush administration in 2005 and Netanyahu's far-right coalition to the Biden administration the previous week. Minister of Communication Shlomo Karhi had initially intended to cut the funding of the Israeli Public Broadcasting Corporation (also known by its blanket branding Kan) by 400 million shekels – roughly half of its total budget – closing several departments, and privatizing content creation. In response, the Director-General of the European Broadcasting Union, Noel Curran, sent two urgent letters to Netanyahu, expressing his concerns and calling on the Israeli government to "safeguard the independence of our Member KAN and ensure it is allowed to operate in a sustainable way, with funding that is both stable, adequate, fair, and transparent." On 25 January 2023, nine journalist organizations representing some of Kan's competitors issued a statement of concern, acknowledging the "important contribution of public broadcasting in creating a worthy, unbiased and non-prejudicial journalistic platform", and noting that "the existence of the [broadcasting] corporation as a substantial public broadcast organization strengthens media as a whole, adding to the competition in the market rather than weakening it." They also expressed their concern that the "real reason" for the proposal was actually "an attempt to silence voices from which... [the Minister] doesn't always draw satisfaction". The same day, hundreds of journalists, actors and filmmakers protested in Tel Aviv. The proposal was eventually put on hold. On 22 February 2023 it was reported that Prime Minister Netanyahu was attempting to appoint his close associate Yossi Shelley as the deputy to the National Statistician — a highly sensitive position in charge of providing accurate data for decision makers. The appointment of Shelley, who did not possess the required qualifications for the role, was withdrawn following publication. In its daily editorial, Haaretz tied this attempt with the judicial reform: "once they take control of the judiciary, law enforcement and public media, they wish to control the state's data base, the dry numerical data it uses to plan its future". Netanyahu also proposed Avi Simhon for the role, and eventually froze all appointments at the Israel Central Bureau of Statistics. Also on 22 February 2023, it was revealed that Yoav Kish, the Minister of Education, was promoting a draft government decision change to the National Library of Israel board of directors which would grant him more power over the institution. In response, the Hebrew University — which owned the library until 2008 – announced that if the draft is accepted, it will withdraw its collections from the library. The university's collections, which according to the university constitute some 80% of the library's collection, include the Agnon archive, the original manuscript of Hatikvah, and the Rothschild Haggadah, the oldest known Haggadah. A group of 300 authors and poets signed an open letter against the move, further noting their objection against "political takeover" of public broadcasting, as well as "any legislation that will castrate the judiciary and damage the democratic foundations of the state of Israel". Several days later, it was reported that a series of donors decided to withhold their donations to the library, totaling some 80 million shekels. On 3 March a petition against the move by 1,500 academics, including Israel Prize laureates, was sent to Kish. The proposal has been seen by some as retribution against Shai Nitzan, the former State Attorney and the library's current rector. On 5 March it was reported that the Legal Advisor to the Ministry of Finance, Asi Messing, was withholding the proposal. According to Messing, the proposal – which was being promoted as part of the Economic Arrangements Law – "was not reviewed... by the qualified personnel in the Ministry of Finance, does not align with any of the common goals of the economic plan, was not agreed to by myself and was not approved by the Attorney General." As of February 2023, the government has been debating several proposals that will significantly weaken the Ministry of Environmental Protection, including reducing the environmental regulation of planning and development and electricity production. One of the main proposals, the transferal of a 3 billion shekel fund meant to finance waste management plants from the Ministry of Environmental Protection to the Ministry of the Interior, was eventually withdrawn. The Minister of Environmental Protection, Idit Silman, has been criticized for using for meeting with climate change denialists, for wasteful and personally-motivated travel on the ministry's expense, for politicizing the role, and for engaging in political activity on the ministry's time. The government has been noted for an unusually high number of dismissals and resignations of senior career civil servants, and for the frequent attempts to replace them with candidates with known political associations, who are often less competent. According to sources, Netanyahu and people in his vicinity are seeking out civil servants who were appointed by the previous government, intent on replacing them with people loyal to him. Governmental nominees for various positions have been criticized for lack of expertise. In addition to the nominee to the position of Deputy National Statistician (see above), the Director General of the Ministry of Finance, Shlomi Heisler; the Director General of the Ministry of Justice, Itamar Donenfeld; and the Director General of Ministry of Transport, Moshe Ben Zaken, have all been criticized for incompetence, lack of familiarity with their Ministries' subject matter, lack of interest in the job, or lack of experience in managing large organizations. It has been reported that in some ministries, senior officials were enacting slowdowns as a means for dealing with the new ministers and director generals. On 28 July the director general of the Ministry of Education resigned, citing as reason the societal "rift". Asaf Zalel, a retired Air Force Brigadier General, was appointed in January. When asked about attempts to appoint his personal friend and attorney to the board of directors of a state-owned company, Minister David Amsalem replied: "that is my job, due to my authority to appoint directors. I put forward people that I know and hold in esteem". Under Minister of Transport Miri Regev, the ministry has either dismissed or lost the heads of the National Public Transport Authority, Israel Airports Authority, National Road Safety Authority, Israel Railways, and several officials in Netivei Israel. The current chair of Netivei Israel is Likud member and Regev associate Yigal Amadi, and the legal counsel is Einav Abuhzira, daughter of a former Likud branch chair. Abuhzira was appointed instead of Elad Berdugo, nephew of Netanyahu surrogate Yaakov Bardugo, after he was disqualified for the role by the Israel Government Companies Authority. In July 2023 the Ministry of Communications, Shlomo Karhi, and the minister in charge of the Israel Government Companies Authority, Dudi Amsalem, deposed the chair of the Israel Postal Company, Michael Vaknin. The chair, who was hired to lead the company's financial recovery after years of operational loss and towards privatization, has gained the support of officials at the Authority and at the Ministry of Finance; nevertheless, the ministers claimed that his performance is inadequate, and nominated in his place Yiftah Ron-Tal, who has known ties to Netanyahu and Smotrich. They also nominated four new directors, two of which have known political associations, and a third who was a witness in Netanyahu's trial. The coalition is allowed to spend a portion of the state's budget on a discretionary basis, meant to coax member parties to reach an agreement on the budget. As of May 2023, the government was pushing an allocation of over 13 billion shekels over two years - almost seven times the amount allocated by the previous government. Most of the funds will be allocated for uses associated with the religious, orthodox and settler communities. The head of the Budget Department at the Ministry of Finance, Yoav Gardos, objected to the allocations, claiming they would exacerbate unemployment in the Orthodox community, which is projected to cost the economy a total of 6.7 trillion shekels in lost produce by 2065. At the onset of the Gaza war and the declaration of a state of national emergency, Minister of Finance Bezalel Smotrich instructed government agencies to continue with the planned distribution of discretionary funds. Corruption During March 2023, the government was promoting an amendment to the Law on Public Service (Gifts) that would allow Netanyahu to receive donations to fund his legal defense. The amendment follows a decision by the High Court of Justice (HCJ) that forced Netanyahu to refund US$270,000 given to him and his wife by his late cousin, Nathan Mileikowsky, for their legal defense. This is in contrast to past statements by Minister of Justice Yariv Levin, who spoke against the possible conflict of interests that can result from such transactions. The bill was opposed by the Attorney General Gali Baharav-Miara, who stressed that it could "create a real opportunity for governmental corruption", and was eventually withdrawn at the end of March. As of March 2023, the coalition was promoting a bill that would prevent judicial review of ministerial appointments. The bill is intended to prevent the HCJ from reviewing the appointment of the twice-convicted chairman of Shas, Aryeh Deri (convicted of bribery, fraud, and breach of trust), to a ministerial position, after his previous appointment was annulled on grounds of unreasonableness. The bill follows on the heels of another amendment, that relaxed the ban on the appointment of convicted criminals, so that Deri - who was handed a suspended sentence after his second conviction - could be appointed. The bill is opposed by the Attorney General, as well as by the Knesset Legal Adviser, Sagit Afik. Israeli law allows for declaring a Prime Minister (as well as several other high-ranking public officials) to be temporarily or permanently incapacitated, but does not specify the conditions which can lead to a declaration of incapacitation. In the case of the Prime Minister, the authority to do so is given to the Attorney General. In March 2023, the coalition advanced a bill that passes this authority from the Attorney General to the government with the approval of the Knesset committee, and clarified that incapacitation can only result from medical or mental conditions. On 3 January 2024, the Supreme Court ruled by a majority of 6 out of 11 that the validity of the law will be postponed to the next Knesset because the bill in its immediate application is a personal law and is intended to serve a distinct personal purpose. Later, the court rejected a petition regarding the definition of Netanyahu as an incapacitated prime minister due to his ongoing trial and conflict of interests. Notes References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Somali_Civil_War] | [TOKENS: 7859] |
Contents Somali Civil War Ongoing Hizbul Islam (until 2010; 2012–2013) Alleged non-state allies: Houthis (de facto control over most of Yemen's population) Somali pirates Battles The Somali Civil War (Somali: Dagaalkii Sokeeye ee Soomaaliya; Arabic: الحرب الأهلية الصومالية al-ḥarb al-’ahliyya aṣ-ṣūmāliyya) is an ongoing civil war that is taking place in Somalia. It grew out of resistance to the military junta which was led by Siad Barre during the 1980s. From 1988 to 1990, the Somali Armed Forces began engaging in combat against various armed rebel groups, including the Somali Salvation Democratic Front in the northeast, the Somali National Movement in the Somaliland War of Independence in the northwest, and the United Somali Congress in the south. The clan-based armed opposition groups overthrew the Barre government in 1991. Various armed factions began competing for influence in the power vacuum and turmoil that followed, particularly in the south. In 1990–92, customary law temporarily collapsed, and factional fighting proliferated. In the absence of a central government, Somalia became a "failed state". This precipitated the arrival of UNOSOM I UN military observers in July 1992, followed by the larger UNITAF and UNOSOM II missions. Following an armed conflict between Somali insurgents and UNOSOM II troops during 1993, the UN withdrew from Somalia in 1995. After the central government's collapse and the withdrawal of UN forces, there was some return to customary and religious law in most regions. In 1991 and 1998, two autonomous regional governments were also established in the northern part of the country: Somaliland and Puntland. In the south Islamic Sharia courts began proliferating in response to lawlessness. This led to a relative decrease in the intensity of the fighting, with the Stockholm International Peace Research Institute removing Somalia from its list of major armed conflicts for 1997 and 1998. In 2000, the Transitional National Government was established, followed by the Transitional Federal Government (TFG) in 2004. The trend toward reduced conflict halted in 2005, and sustained and destructive conflict took place in the south in 2005–07, but the battle was of a much lower scale and intensity than in the early 1990s. In 2006, Ethiopian troops invaded Somalia to depose the Islamic Courts Union (ICU) and install the TFG. The ICU effectively disintegrated, and soon after a large scale insurgency began against the occupation as other Islamist groups formed and established themselves as independent actors. Most notably Al-Shabaab rose to prominence in this period, and has since been fighting the Somali government and the AU-mandated AMISOM peacekeeping force for control of the country. Somalia topped the annual Fragile States Index for six years from 2008 up to and including 2013. In October 2011, following preparatory meetings, Kenyan troops entered southern Somalia ("Operation Linda Nchi") to fight al-Shabaab and establish a buffer zone inside Somalia. Kenyan troops were formally integrated into the multinational force in February 2012. The Federal Government of Somalia was established in August 2012, constituting the country's first permanent central government since the start of the civil war. In 2023, the Las Anod conflict broke out in the northern part of Somalia between SSC-Khatumo and the Somaliland Army. International stakeholders and analysts subsequently began to describe Somalia as a "fragile state" that is making some progress toward stability. Decline and Fall of Siad Barre's government (1978–91) After Somalia lost the Ogaden War in March 1978, the president's popularity with Somalis plummeted and widespread discontent among his generals led to an attempted coup d'état on 10 April 1978. Most of the coup's ringleaders were rounded up and executed but some escaped and formed the Somali Salvation Democratic Front, starting the rebellion that eventually toppled Siad Barre from power 13 years later. In May 1986, Barre suffered serious injuries in a car crash near Mogadishu, when the car transporting him smashed into the back of a bus during a heavy rainstorm. He was treated in a hospital in Saudi Arabia for head injuries, broken ribs and shock for a month. Lieutenant General Mohamed Ali Samatar, then Vice President, served as de facto head of state for the next several months. Although Barre managed to recover enough to present himself for reelection to a seven-year term on December 23, 1986, his poor health and advanced age led to speculation about who would succeed him. Possible contenders included his son-in-law General Ahmed Suleiman Abdille, then the Minister of the Interior, in addition to Samatar. In an effort to hold on to power, Barre's ruling Supreme Revolutionary Council (SRC) became increasingly totalitarian and arbitrary. This caused opposition to his government to grow. Barre tried to quell the unrest by abandoning appeals to nationalism, relying more and more on his own inner circle, and exploiting historical clan animosities. By the mid-1980s, more resistance movements supported by Ethiopia's communist Derg administration had sprung up across the country. Barre responded by ordering punitive measures against those he perceived as supporting the guerrillas, especially in the north. The clampdown included the bombing of cities, with the northwestern administrative center of Hargeisa, a Somali National Movement (SNM) stronghold, among the targeted areas in 1988. In December 1981, unrest was triggered in Northern Somalia by the arrest of 30 Isaaq professionals in Hargeisa who created a self-help group to improve local facilities. This was followed by the systematic efforts to remove all Isaaqs from positions of power including the military, judiciary and security services, as well as harsh policies enacted against the Isaaq, including a declaration of economic warfare on the Isaaq. The transfer of power to non-Isaaq pro-government individuals further pushed Isaaq communities to rebel against Barre's regime and was one of the main causes of the breakout of the Somaliland War of Independence. In 1988, Siad Barre and Ethiopian dictator Mengistu Haile Mariam agreed to a secret deal whereby each would cease hosting insurgencies of one another. This prompted the Somali National Movement (SNM) to launch an offensive on Northern Somalia from its bases on the Ethiopian border. Barre's regime responded with "systematic" human rights abuses and the genocide of thousands of Isaaq tribesmen resulting in up to 200,000 civilians slaughtered and 500,000 more people seeking refuge in neighbouring Ethiopia. The clampdown initiated by Barre's government extended its reach beyond the initial bombings in the north to encompass various regions across the country. This reproduction of aggressive strategies aimed at stifling dissent and retaining authority over the populace was a hallmark of the government's repressive actions in the South. One of the most notable instances occurred in 1991, when Barre's regime initiated a ruthless aerial assault that led to the deaths of numerous innocent individuals in the town of Beledwene, situated in southern Somalia. The cruelty and magnitude of this atrocity highlighted the degree to which the government was prepared to go to quash any sort of opposition or resistance, displaying a blatant disregard for human rights and the worth of human life. Another notable instance of Barre's repressive policies occurred in the city of Baidoa, which earned the nickname 'the city of death' due to the tragic events that unfolded there during the famine and civil war. Hundreds of thousands of individuals lost their lives as a consequence of governmental strategies specifically aimed at the Rahanweyn community residing in these areas. In response to these humanitarian abuses, Western aid donors cut funding to the Somali regime which, at the time, was heavily reliant on foreign aid. This resulted in a rapid "retreat of the state", accompanied by a severe drop in value for the Somali Shilling and mass military desertion by Somali army units. In 1990, as fighting intensified, Somalia's first President, Aden Abdullah Osman Daar, and about 100 other Somali politicians signed a manifesto advocating reconciliation. A number of the signatories were subsequently arrested. Barre's heavy-handed tactics further strengthened the appeal of the various rebel movements, although these groups' only common goal was the overthrow of his government. In the north, fighting continued between SNM rebels and heavily armed pro-government militia in places like Awdal. In January 1991, in one of the final episodes of the civil war in the north, SNM militia gave chase to retreating government forces (26th Division) to the town of Dilla, where a battle took place and the town was destroyed. SNM militia then continued into Borama, the capital and largest town of Awdal, but the SNM leadership withdrew units within 24 hours to allow discussions to take place without the threat of occupation. By February 4, SNM's control extended to the entire north of Somalia, and all prisoners and pro-government ex-soldiers were released and ordered to return to their regions of origin (mainly Ethiopia), except for Hawiye ex-soldiers and ex-civil servants, who were permitted to remain in Burco since their lives would have been at risk if they had traveled through hostile pro-Barre country on their return to Mogadishu. On February 9, Ismail Omar Guelleh, then chief of staff of Djibouti's secret service, attempted to annex Zeyla in Awdal, Somaliland to Djibouti in the 1991 Zeila incursion during the Somaliland War of Independence, however the Djiboutian-backed United Somali Front was shortly routed from the area by Somali National Movement (SNM) forces. United Somali Congress topples Barre By mid-1990, United Somali Congress (USC) rebels had captured most towns and villages surrounding Mogadishu, which prompted some to give Barre the ironic title 'Mayor of Mogadishu.' In December the USC entered Mogadishu. Four weeks of battle between Barre's remaining troops and the USC ensued, during which the USC brought more forces into the city. By January 1991, USC rebels defeated the Red Berets, Barre's special forces, toppling Barre's hold on the government. The remainder of the government's forces then finally collapsed. Some became irregular regional forces and clan militias. After the USC's victory over Barre's troops, the other rebel groups declined to cooperate with it, as each instead drew primary support from its own constituency. Among these other opposition movements were the Somali Patriotic Movement (SPM) and Somali Democratic Alliance (SDA), a Gadabuursi group which had been formed in the northwest to counter the Somali National Movement Isaaq militia. For its part, the SNM initially refused to accept the legitimacy of the provisional government that the USC had established, but in March 1991 the SNM's former leader Ahmed Mohamed Silanyo proposed a power-sharing framework between the SNM and USC under a new transitional government. Many of the opposition groups subsequently began competing for influence in the power vacuum that followed the ousting of Barre's government. In the south, armed factions led by USC commanders General Mohamed Farah Aidid and Ali Mahdi Mohamed, in particular, clashed as each sought to exert authority over the capital. In the northwest, at the Burao conference of April–May 1991, the SNM declared an independent Republic of Somaliland in the region that had constituted the British Somaliland before independence and unification with the former colony of Italian Somaliland in 1960 electing Abdirahman Ahmed Ali Tuur as president. Violence flared up in Mogadishu on 17 November 1991, when the Aidid-aligned faction of the USC attacked Mahdi-aligned forces in the city. They seized part of the city, but could not push Mahdi's forces out of northern Mogadishu. In 1992, after four months of heavy fighting for control of Mogadishu, a ceasefire was agreed between Ali Mahdi Mohamed and Mohamed Farah Aideed. Neither had seized control of the capital, and as a result, a 'greenline' was established between east and west that divided their areas of control. United Nations intervention (1992–1995) UN Security Council Resolution 733 and UN Security Council Resolution 746 led to the creation of the United Nations Operation in Somalia I (UNOSOM I), to provide humanitarian relief and help restore order in Somalia after the dissolution of its central government. The political state was described by the UN’s summary as being chaos with a deteriorating security system and widespread death and destruction. United Nations Security Council Resolution 794 was unanimously passed on December 3, 1992, which approved a coalition of United Nations peacekeepers led by the United States. Forming the Unified Task Force (UNITAF), the alliance was tasked with assuring security until humanitarian efforts aimed at stabilizing the situation were transferred to the UN. Landing in 1993, the UN peacekeeping coalition started the two-year United Nations Operation in Somalia II (UNOSOM II) primarily in the south. UNITAF's original mandate was to use "all necessary means" to guarantee the delivery of humanitarian aid in accordance to Chapter VII of the United Nations Charter. Whereas the aim of UNOSOM 1 was primarily humanitarian UNITAFs mission statement to restore "peace, stability, law and order" suggests their belief of Somalia’s incapacity to secure the safety of the population without assistance from international military. During negotiations from 1993 to 1995, Somali principals had some success in reconciliation and establishment of public authorities. Among these initiatives was the Mudug peace agreement of June 1993 between Aidid's forces and the SSDF, which established a ceasefire between the Haber Gedir and the Majeerteen clans, opened the trade routes, and formalized the withdrawal of militants from Galkayo; the UNOSOM-mediated Hirab reconciliation of January 1994 in Mogadishu between elders of the rival Abgal and Haber Gedir clans, which was backed by politicians from these constituencies and concluded with a pact to end hostilities, dismantle the green line partitioning the city, and remove road blocks; the UNOSOM-mediated Kismayo initiative of 1994 between the SNA, SPM, SSDF, and representatives of nineteen clans from the southern Lower Juba and Middle Juba regions; the 1994 Bardhere conference between the Marehan and Rahanweyn (Digil and Mirifle), which resolved conflicts over local resources; and the short-lived Digil-Mirifle Governing Council for the southern Bay and Bakool regions, which was established in March 1995. Some of the militias that were then competing for power saw UNOSOM's presence as a threat to their hegemony. Consequently, gun battles took place in Mogadishu between local gunmen and peacekeepers. Among these was the Battle of Mogadishu in October 1993, part of an unsuccessful operation by U.S. troops to apprehend Somali National Alliance faction leader Mohamed Farah Aidid. UN soldiers eventually withdrew altogether from the country on March 3, 1995, having incurred more significant casualties. The UN stated that their withdrawal without completing their mandate was due to a lack of progress towards peace and little cooperation with Somali parties over security issues which were continually undermined. They received significant backlash after this withdrawal prompting them to state they were not abandoning Somalia however provided little international military support until the formation of the more local military operation AMISOM in 2007. USC/SSA (1995–2000) After UNOSOM II's departure in March 1995, military clashes between local factions became shorter, generally less intense, and more localized. This was in part due to the large-scale UN military intervention that had helped to curb the intense fighting between the major factions, who then began to focus on consolidating gains that they had made. The local peace and reconciliation initiatives that had been undertaken in the south-central part of the country between 1993 and 1995 also generally had a positive impact. Aidid subsequently declared himself President of Somalia on June 15, 1995. However, his declaration received no recognition, as his rival Ali Mahdi Muhammad had already been elected interim President at a conference in Djibouti and recognized as such by the international community. Consequently, Aidid's faction continued its quest for hegemony in the south. In September 1995, militia forces loyal to him attacked and occupied the city of Baidoa. Aidid's forces remained in control of Baidoa from September 1995 to at least January 1996, while the local Rahanweyn Resistance Army militia continued to engage his forces in the town's environs. Fighting continued in the later half of 1995 in southern Kismayo and the Juba Valley, as well as southwestern and central Somalia. However, despite these pockets of conflict, the Gedo and Middle Shabelle regions, and northwestern parts of the country remained relatively peaceful. A number of the regional and district administrations that had been locally established in the preceding few years continued to operate in these areas. In 1994–95, factions contending for power in the newly-declared-independent Somaliland region included the United Somalia Front, the Somalia Democratic Front, the Somali National Movement, and the United Somali Party. In March 1996, Ali Mahdi was elected chairman of the United Somali Congress/Somali Salvation Alliance (USC/SSA), based in northern Mogadishu. In the southern part of city, Aidid's forces battled those of Osman Atto for control of the port of Merca as well as strategic areas in Mogadishu. Fighting in Merca eventually ended after elders intervened, but continued in Mogadishu. In August 1996, Aidid died from wounds incurred during combat in the Medina area. In 1998, a homegrown constitutional conference was held in the northeastern town of Garowe over a period of three months. It was attended by the area's political elite, traditional elders (Issims), members of the business community, intellectuals and other civil society representatives. The Puntland State of Somalia was subsequently established. In 1999, Eritrea was alleged to be supporting Somali National Alliance forces led by the late Aidid's son Hussein Farrah Aidid. Aidid Jr. denied the claims, saying that the Ethiopian Prime Minister Meles Zenawi had requested that he mediate between Ethiopia and Eritrea in their separate conflict. However the International Institute for Strategic Studies separately reported that Hussein Aideed himself had acknowledged support from both Eritrea and Uganda. Aideed's forces occupied Huddur and Baidoa. However, they were driven out by the Rahanweyn Resistance Army in June 1999. By the end of the year, the Rahanweyn Resistance Army had taken control of the Bay and Bakool provinces. The RRA's leader Hasan Muhammad Nur Shatigadud subsequently established the Southwestern State of Somalia regional administration. In 2000, Ali Mahdi participated in another conference in Djibouti. He lost a re-election bid there to Barre's former Interior Minister Abdiqasim Salad Hassan. TFG, Islamic Courts Union, and Ethiopia (2006–2009) In 2000, the Transitional National Government (TNG) was established. The Transitional Federal Government (TFG) was formed in Nairobi in 2004. Selection of members of parliament was underway by June, over two hundred members of parliament (MPs) took the oath of office in August, and Abdullahi Yusuf Ahmed was elected president by the parliament in October 2004. However, in March 2005 the TFG split after a brawl in parliament over deployment of peacekeepers and relocation to an interim capital. The parliamentary speaker led some members to Mogadishu while the president and others remained in Nairobi. In June 2005, under pressure from Kenya, the remainder of the TFG left Nairobi for Jowhar. In February 2006, the TFG parliament met in Baidoa for the first time since March 2005. (Interpeace, 104) A battle for Mogadishu followed in the first half of 2006 in which the ARPCT, a coalition of U.S.-backed militia leaders, confronted the ascendant Islamic Courts Union (ICU). However, the ICU won a decisive victory in June of that year. It then rapidly expanded and consolidated its power throughout southern Somalia. By August 2006, the TFG was confined to Baidoa under Ethiopian protection. (Interpeace, 104) In December 2006, Ethiopian troops entered Somalia to assist the TFG against the advancing Islamic Courts Union, initially winning the Battle of Baidoa. With their support, Somali government forces recaptured the capital from the ICU. The offensive helped the TFG solidify its rule. On January 8, 2007, as the Battle of Ras Kamboni raged, TFG President and founder Abdullahi Yusuf Ahmed entered Mogadishu for the first time since being elected to office. But as Meckhaus writes, the TFG was seen "by most of the Mogadishu population as a puppet of Ethiopia, and uncontrolled TFG security forces became the principal sources of insecurity for the local population, engaging in kidnapping, assaults, and worse." Within weeks, an armed insurgency subsequently arose in the capital against the TFG and its Ethiopian allies. The government then relocated to the capital from its interim location in Baidoa. The arms embargo on Somalia was amended in February 2007 to allow states to supply weapons to the TFG's security forces, provided that they received prior approval from the UN's Somalia Sanctions Committee. After long discussions, the African Union approved the initial deployment of the African Union Mission to Somalia (AMISOM) in March 2007. It established a "small triangle of protection" around Mogadishu's airport, seaport, and the Villa Somalia, and began to adopt a low-key negotiating profile with key actors. In November 2008, following repeated violations of the weapons blockade, the Security Council decided that an arms embargo could be imposed on entities involved in such breaches. After a two-year consultation process, the TFG was formed in 2004 by Somali politicians in Nairobi under the auspices of the Intergovernmental Authority on Development (IGAD). The process also led to the establishment of the Transitional Federal Institutions (TFIs), and concluded in October 2004 with the election of Abdullahi Yusuf Ahmed as president. The TFG thereafter became Somalia's internationally recognized government. Following their defeat, the Islamic Courts Union splintered into several different factions. Some of the more radical elements, including Al-Shabaab, regrouped to continue their insurgency against the TFG and oppose the Ethiopian military's presence in Somalia. Throughout 2007 and 2008, Al-Shabaab scored military victories, seizing control of key towns and ports in both central and southern Somalia. At the end of 2008, the group had captured Baidoa but not Mogadishu. On May 1, 2008, the U.S. made an airstrike on Dhusamareb, and followed on 3 May with another airstrike on the border town of Dobley. According to the International Crisis Group, Ethiopia's leaders were surprised by the insurgency's persistence and strength and frustrated at the TFG's chronic internal problems. By January 2009, Al-Shabaab and other militias had forced the Ethiopian troops to retreat, leaving behind an understaffed African Union peacekeeping force. Due to a lack of funding and human resources, an arms embargo that made it difficult to re-establish a national security force, and general indifference on the part of the international community,[citation needed] President Yusuf found himself obliged to deploy thousands of troops from Puntland to Mogadishu to sustain the battle against insurgent elements in the southern part of the country. Financial support for this effort was provided by the autonomous region's government. This left little revenue for Puntland's own security forces and civil service employees, leaving the territory vulnerable to piracy and terrorist attacks. On December 29, 2008, Abdullahi Yusuf Ahmed announced before a united parliament in Baidoa his resignation as President of Somalia. In his speech, which was broadcast on national radio, Yusuf expressed regret at failing to end the country's seventeen-year conflict as his government had mandated to do. He also blamed the international community for its failure to support the government, and said that the speaker of parliament would succeed him in office per the charter of the Transitional Federal Government. Coalition government (2009) Between May 31 and June 9, 2008, representatives of Somalia's federal government and the Alliance for the Re-liberation of Somalia (ARS) participated in peace talks in Djibouti brokered by Ahmedou Ould-Abdallah, the Special Representative of the Secretary General for Somalia. The conference ended with a signed agreement calling for the withdrawal of Ethiopian troops in exchange for the cessation of armed confrontation. Parliament was subsequently expanded to 550 seats to accommodate ARS members, which then elected Sheikh Sharif Sheikh Ahmed, the former ARS chairman, to office. President Sharif shortly afterwards appointed Omar Abdirashid Ali Sharmarke, the son of slain former President Abdirashid Ali Sharmarke, as the nation's new Prime Minister. With the help of AMISOM, the coalition government also began a counteroffensive in February 2009 to assume full control of the southern half of the country. To solidify its rule, the TFG formed an alliance with the Islamic Courts Union, other members of the Alliance for the Re-liberation of Somalia, and Ahlu Sunna Waljama'a, a moderate Sufi militia. Furthermore, Al-Shabaab and Hizbul Islam, the two main Islamist groups in opposition, began to fight amongst themselves in mid-2009. As a truce, in March 2009, Somalia's coalition government announced that it would re-implement shari'a as the nation's official judicial system. However, conflict continued in the southern and central parts of the country. Within months, the coalition government had gone from holding about 70% of south-central Somalia's conflict zones, territory which it had inherited from the previous Yusuf administration, to losing control of over 80% of the disputed territory to the Islamist insurgents. Since 2009 In November 2010, a new technocratic government was elected to office, which enacted numerous reforms. Among these, in its first 50 days in office, the new administration completed its first monthly payment of stipends to government soldiers. This government subsequently began to push back Al-Shabaab over the following years. On August 6, 2011, Al-Shabaab was forced to withdraw from most of Mogadishu. Al-Shabaab did still retain a foothold in the northern outskirts of the capital, but by January 2012, the combined efforts of Somali government and AMISOM forces had expelled them from the city completely. An ideological rift within Al-Shabaab's leadership also emerged after the 2011 drought and the assassination of top officials in the organization. With the majority of Mogadishu secure, the Somali Armed Forces and Kenya Defence Forces next launched Operation Linda Nchi, a joint advance against Al-Shabaab, in October 2011. This operation had reportedly been planned for nearly two years, during which time Kenyan officials sought U.S. support for the mission. After the successful conclusion of Operation Linda Nchi in May 2012, Kenyan troops were formally integrated into AMISOM in June. After Operation Linda Nchi, the port city of Kismayo was the last major stronghold that remained in Al-Shabaab's control. A Kenyan-led AMISOM force, backed by the Raskamboni movement, then launched an offensive against Kismayo on September 28, 2012. After a three-day battle, Somali government forces were able to gain control of the city. The month of September 2012 also saw the establishment of the Federal Government of Somalia. In January 2013, AMISOM's mandate was extended for another year following the adoption of UNSC Resolution 2093. The Security Council also unanimously voted to suspend Somalia's arms embargo on light weapons for one year and welcomed the Federal Government's development of a new national security strategy, urging the central authorities to accelerate the plan's implementation, further define the Somali national security forces' composition, and identify capability gaps to assist their international partners in better addressing them. While many urban areas had been seized, Al-Shabaab still controlled many rural areas, where a number of their operatives disappeared into local communities in order to more effectively exploit any mistakes by the central authorities. In October 2013, the U.S. Africa Command began establishing the Mogadishu Coordinating Cell, which became fully operational in late December. The unit was formed at the request of the Somali government and AMISOM, who had approached U.S. Department of Defense Secretary Chuck Hagel in September about the possibility. It consists of a team of fewer than five advisers, including planners and communicators between the Somali authorities and AMISOM. The cell is intended to provide consultative and planning support to the allied forces to enhance their capacity and promote peace and security throughout the country and region. In November 2013, a senior Ethiopian government official announced that Ethiopia's troops deployed in Somalia would soon join AMISOM, having already forwarded a request to do so. At the time, an estimated 8,000 Ethiopian soldiers were stationed in the country. The Somali Foreign Ministry welcomed the decision, asserting that the move would galvanize AMISOM's campaign against Al-Shabaab. Following the adoption of UN Security Council Resolution 2124, which authorized the deployment of 4,000 additional troops to augment AMISOM's 22,126 strong force, Ethiopian troops formally joined the mission in January 2014. They are mandated to work alongside the Somali National Army, with responsibility for the allied forces' operations in the southern Gedo, Bakool and Bay regions. The Ethiopian troops represent AMISOM's sixth contingent after the Djibouti, Burundi, Sierra Leone, Kenya and Uganda units. In January 2014, at an African Union Summit in Addis Ababa, President Hassan Sheikh Mohamud requested an extension of the UN Security Council's weapons purchasing mandate for Somalia after its March expiration, saying that the Somali defence forces required better military equipment and arms to more effectively combat militants. The following month, the UN Somalia and Eritrea Monitoring Group reported that systematic abuses by Somali government officials had allowed weapons to be diverted away from Somalia's security forces into the hands of faction leaders and Al-Shabaab militants. The panel had observed various problems with the management of weapons and ammunition stockpiles, including difficulties by monitors in accessing local weapons stockpiles and obtaining information about the arms. The monitors also suggested that one key adviser to the president was involved in planning arms deliveries to Al-Shabaab and that shipments of weapons from Djibouti and Uganda could not be accounted for. Somali Chief of Army Dahir Adan Elmi made a pro forma denial of the allegations. He also said that a UN monitoring team had twice visited the government's weapons and ammunition storage facilities in Mogadishu, where it was shown the arms stockpiles for inspection and had expressed satisfaction. Elmi said that the government had twice purchased weapons since the arms embargo on Somalia was partially lifted. He also asserted that Al-Shabaab already had an adequate supply of weapons and mainly utilized explosive devices and sophisticated bombs. In February 2014, a delegation led by Prime Minister of Somalia Abdiweli Sheikh Ahmed met in Addis Ababa with Ethiopian Prime Minister Hailemariam Desalegn to discuss strengthening relations between the countries. Ahmed commended Ethiopia's role in the ongoing peace and stabilization process in Somalia as well as its opposition to Al-Shabaab, and welcomed the Ethiopian military's decision to join AMISOM. Desalegn in turn pledged his administration's continued support for Somalia's peace and stabilization efforts, as well as its preparedness to assist in initiatives aiming to build up Somali security forces through experience-sharing and training. The meeting concluded with a tripartite Memorandum of Understanding agreeing to promote partnership and cooperation, including a cooperative agreement to develop the police force. On 5 March 2014, the UN Security Council unanimously voted to extend the partial easing of the arms embargo on Somalia to 25 October. The resolution permitted the Somali government to purchase light weapons, with the stipulation that all member states must take steps to prevent the direct or indirect supply, transfer or sale of arms and military equipment to individuals or entities outside of the Somali security forces. The Somali government was also required to routinely report on the structural status of the military, as well as provide information on the extant infrastructure and protocols designed to ensure the military equipment's safe delivery, storage and maintenance. In early March 2014, AMISOM, supported by Somali militias, launched another operation to remove Al-Shabaab from its remaining areas of control in southern Somalia. According to Prime Minister Abdiweli Sheikh Ahmed, the government subsequently launched stabilization efforts in the newly liberated areas, which included Rab Dhuure, Hudur, Wajid and Burdhubo. The Ministry of Defense provided ongoing reassurance and security to local residents, and supplied logistical and security support. Additionally, the Ministry of Interior was prepared to support and put into place programs to assist local administration and security. A Deputy Minister and several religious scholars were dispatched to all four towns to coordinate and supervise the federal government's stabilization initiatives. By March 26, the allied forces had liberated ten towns, including Qoryoley and El Buur. UN Special Representative for Somalia Nicholas Kay described the military advance as the most significant and geographically extensive offensive since AU troops began operations in 2007. In August 2014, the Somali government-led Operation Indian Ocean was launched, aiming to reduce insurgent-held areas along the coastline. On 1 September 2014, a U.S. drone strike carried out as part of the broader mission killed Al-Shabaab leader Muktar Abu Zubeyr. U.S. authorities hailed the raid as a major symbolic and operational loss for Al-Shabaab, and the Somali government offered a 45-day amnesty to all moderate members of the militant group. On 15 December 2018 there were demonstrations in the city of Baidoa by supporters of Mukhtar Rowbow, a presidential candidate who had been arrested two days earlier by government forces and transferred to Mogadishu. Rowbow was a senior member of al-Shabaab. AMISOM announced that its forces did not assist in Rowbow's arrest and his transfer to Mogadishu. From the beginning of 2020, humanitarian researchers and local medical personnel became increasingly concerned that the COVID-19 pandemic could be catastrophic for Somalis, because of the damage the civil war has wrought on Somalia's health care, and weak provision since the 1980s. On 25 November 2020, it was reported that a CIA officer had been killed in Somalia. The death came as the US administration under Donald Trump was making plans to withdraw more than 600 troops from Somalia. President Donald Trump ordered the Department of Defense to remove the majority of the 700 U.S. troops in Somalia (many from Special Operations Command Africa) in December 2020. He changed the mission of American troops to assist the Somali Armed Forces in its fight against al-Shabaab. When President Mohamed Abdullahi Mohamed's term expired in February 2021, dates had not been set for the election of a successor, and fighting subsequently broke out in Mogadishu. This fighting continued until May 2021, when the government and opposition agreed to hold elections within 60 days; after further negotiation, the presidential election was scheduled for October 10, and was actually held on 15 May 2022, with Hassan Sheikh Mohamud being declared the new president in a peaceful transition of power. In March 2022, the 14-year long AMISOM mission came to an end. It was replaced by a Somali-led operation, the African Union Transition Mission in Somalia (ATMIS), and later by the African Union Support and Stabilization Mission in Somalia (AUSSOM) In 2023, fighting broke out between the Khatumo State and Somaliland leading to the 2023 Las Anod conflict. On May 14, 2025, in a post on social medias, Donald Trump claimed that he would "support the Somali People, who should not allow the Houthis to embed.", claiming that this initiative was "to end terrorism, and bring prosperity to their Country”. Casualties According to Necrometrics, around 500,000 people are estimated to have been killed in Somalia since the start of the civil war in 1991. The Armed Conflict Location & Event Dataset estimates that 3,300 people were killed during the conflict in 2012, with the number of fatalities dropping slightly in 2013 to 3,150. The United Nations Assistance Mission in Somalia reported at least 596 civilian casualties by August 2020. However, officials have struggled to maintain an accurate count due to flooding and COVID-19 deaths in the area. See also Notes References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Python_(programming_language)#cite_ref-12] | [TOKENS: 4314] |
Contents Python (programming language) Python is a high-level, general-purpose programming language. Its design philosophy emphasizes code readability with the use of significant indentation. Python is dynamically type-checked and garbage-collected. It supports multiple programming paradigms, including structured (particularly procedural), object-oriented and functional programming. Guido van Rossum began working on Python in the late 1980s as a successor to the ABC programming language. Python 3.0, released in 2008, was a major revision and not completely backward-compatible with earlier versions. Beginning with Python 3.5, capabilities and keywords for typing were added to the language, allowing optional static typing. As of 2026[update], the Python Software Foundation supports Python 3.10, 3.11, 3.12, 3.13, and 3.14, following the project's annual release cycle and five-year support policy. Python 3.15 is currently in the alpha development phase, and the stable release is expected to come out in October 2026. Earlier versions in the 3.x series have reached end-of-life and no longer receive security updates. Python has gained widespread use in the machine learning community. It is widely taught as an introductory programming language. Since 2003, Python has consistently ranked in the top ten of the most popular programming languages in the TIOBE Programming Community Index, which ranks based on searches in 24 platforms. History Python was conceived in the late 1980s by Guido van Rossum at Centrum Wiskunde & Informatica (CWI) in the Netherlands. It was designed as a successor to the ABC programming language, which was inspired by SETL, capable of exception handling and interfacing with the Amoeba operating system. Python implementation began in December 1989. Van Rossum first released it in 1991 as Python 0.9.0. Van Rossum assumed sole responsibility for the project, as the lead developer, until 12 July 2018, when he announced his "permanent vacation" from responsibilities as Python's "benevolent dictator for life" (BDFL); this title was bestowed on him by the Python community to reflect his long-term commitment as the project's chief decision-maker. (He has since come out of retirement and is self-titled "BDFL-emeritus".) In January 2019, active Python core developers elected a five-member Steering Council to lead the project. The name Python derives from the British comedy series Monty Python's Flying Circus. (See § Naming.) Python 2.0 was released on 16 October 2000, featuring many new features such as list comprehensions, cycle-detecting garbage collection, reference counting, and Unicode support. Python 2.7's end-of-life was initially set for 2015, and then postponed to 2020 out of concern that a large body of existing code could not easily be forward-ported to Python 3. It no longer receives security patches or updates. While Python 2.7 and older versions are officially unsupported, a different unofficial Python implementation, PyPy, continues to support Python 2, i.e., "2.7.18+" (plus 3.11), with the plus signifying (at least some) "backported security updates". Python 3.0 was released on 3 December 2008, and was a major revision and not completely backward-compatible with earlier versions, with some new semantics and changed syntax. Python 2.7.18, released in 2020, was the last release of Python 2. Several releases in the Python 3.x series have added new syntax to the language, and made a few (considered very minor) backward-incompatible changes. As of January 2026[update], Python 3.14.3 is the latest stable release. All older 3.x versions had a security update down to Python 3.9.24 then again with 3.9.25, the final version in 3.9 series. Python 3.10 is, since November 2025, the oldest supported branch. Python 3.15 has an alpha released, and Android has an official downloadable executable available for Python 3.14. Releases receive two years of full support followed by three years of security support. Design philosophy and features Python is a multi-paradigm programming language. Object-oriented programming and structured programming are fully supported, and many of their features support functional programming and aspect-oriented programming – including metaprogramming and metaobjects. Many other paradigms are supported via extensions, including design by contract and logic programming. Python is often referred to as a 'glue language' because it is purposely designed to be able to integrate components written in other languages. Python uses dynamic typing and a combination of reference counting and a cycle-detecting garbage collector for memory management. It uses dynamic name resolution (late binding), which binds method and variable names during program execution. Python's design offers some support for functional programming in the "Lisp tradition". It has filter, map, and reduce functions; list comprehensions, dictionaries, sets, and generator expressions. The standard library has two modules (itertools and functools) that implement functional tools borrowed from Haskell and Standard ML. Python's core philosophy is summarized in the Zen of Python (PEP 20) written by Tim Peters, which includes aphorisms such as these: However, Python has received criticism for violating these principles and adding unnecessary language bloat. Responses to these criticisms note that the Zen of Python is a guideline rather than a rule. The addition of some new features had been controversial: Guido van Rossum resigned as Benevolent Dictator for Life after conflict about adding the assignment expression operator in Python 3.8. Nevertheless, rather than building all functionality into its core, Python was designed to be highly extensible via modules. This compact modularity has made it particularly popular as a means of adding programmable interfaces to existing applications. Van Rossum's vision of a small core language with a large standard library and easily extensible interpreter stemmed from his frustrations with ABC, which represented the opposite approach. Python claims to strive for a simpler, less-cluttered syntax and grammar, while giving developers a choice in their coding methodology. Python lacks do .. while loops, which Rossum considered harmful. In contrast to Perl's motto "there is more than one way to do it", Python advocates an approach where "there should be one – and preferably only one – obvious way to do it". In practice, however, Python provides many ways to achieve a given goal. There are at least three ways to format a string literal, with no certainty as to which one a programmer should use. Alex Martelli is a Fellow at the Python Software Foundation and Python book author; he wrote that "To describe something as 'clever' is not considered a compliment in the Python culture." Python's developers typically prioritize readability over performance. For example, they reject patches to non-critical parts of the CPython reference implementation that would offer increases in speed that do not justify the cost of clarity and readability.[failed verification] Execution speed can be improved by moving speed-critical functions to extension modules written in languages such as C, or by using a just-in-time compiler like PyPy. Also, it is possible to transpile to other languages. However, this approach either fails to achieve the expected speed-up, since Python is a very dynamic language, or only a restricted subset of Python is compiled (with potential minor semantic changes). Python is meant to be a fun language to use. This goal is reflected in the name – a tribute to the British comedy group Monty Python – and in playful approaches to some tutorials and reference materials. For instance, some code examples use the terms "spam" and "eggs" (in reference to a Monty Python sketch), rather than the typical terms "foo" and "bar". A common neologism in the Python community is pythonic, which has a broad range of meanings related to program style: Pythonic code may use Python idioms well; be natural or show fluency in the language; or conform with Python's minimalist philosophy and emphasis on readability. Syntax and semantics Python is meant to be an easily readable language. Its formatting is visually uncluttered and often uses English keywords where other languages use punctuation. Unlike many other languages, it does not use curly brackets to delimit blocks, and semicolons after statements are allowed but rarely used. It has fewer syntactic exceptions and special cases than C or Pascal. Python uses whitespace indentation, rather than curly brackets or keywords, to delimit blocks. An increase in indentation comes after certain statements; a decrease in indentation signifies the end of the current block. Thus, the program's visual structure accurately represents its semantic structure. This feature is sometimes termed the off-side rule. Some other languages use indentation this way; but in most, indentation has no semantic meaning. The recommended indent size is four spaces. Python's statements include the following: The assignment statement (=) binds a name as a reference to a separate, dynamically allocated object. Variables may subsequently be rebound at any time to any object. In Python, a variable name is a generic reference holder without a fixed data type; however, it always refers to some object with a type. This is called dynamic typing—in contrast to statically-typed languages, where each variable may contain only a value of a certain type. Python does not support tail call optimization or first-class continuations; according to Van Rossum, the language never will. However, better support for coroutine-like functionality is provided by extending Python's generators. Before 2.5, generators were lazy iterators; data was passed unidirectionally out of the generator. From Python 2.5 on, it is possible to pass data back into a generator function; and from version 3.3, data can be passed through multiple stack levels. Python's expressions include the following: In Python, a distinction between expressions and statements is rigidly enforced, in contrast to languages such as Common Lisp, Scheme, or Ruby. This distinction leads to duplicating some functionality, for example: A statement cannot be part of an expression; because of this restriction, expressions such as list and dict comprehensions (and lambda expressions) cannot contain statements. As a particular case, an assignment statement such as a = 1 cannot be part of the conditional expression of a conditional statement. Python uses duck typing, and it has typed objects but untyped variable names. Type constraints are not checked at definition time; rather, operations on an object may fail at usage time, indicating that the object is not of an appropriate type. Despite being dynamically typed, Python is strongly typed, forbidding operations that are poorly defined (e.g., adding a number and a string) rather than quietly attempting to interpret them. Python allows programmers to define their own types using classes, most often for object-oriented programming. New instances of classes are constructed by calling the class, for example, SpamClass() or EggsClass()); the classes are instances of the metaclass type (which is an instance of itself), thereby allowing metaprogramming and reflection. Before version 3.0, Python had two kinds of classes, both using the same syntax: old-style and new-style. Current Python versions support the semantics of only the new style. Python supports optional type annotations. These annotations are not enforced by the language, but may be used by external tools such as mypy to catch errors. Python includes a module typing including several type names for type annotations. Also, mypy supports a Python compiler called mypyc, which leverages type annotations for optimization. 1.33333 frozenset() Python includes conventional symbols for arithmetic operators (+, -, *, /), the floor-division operator //, and the modulo operator %. (With the modulo operator, a remainder can be negative, e.g., 4 % -3 == -2.) Also, Python offers the ** symbol for exponentiation, e.g. 5**3 == 125 and 9**0.5 == 3.0. Also, it offers the matrix‑multiplication operator @ . These operators work as in traditional mathematics; with the same precedence rules, the infix operators + and - can also be unary, to represent positive and negative numbers respectively. Division between integers produces floating-point results. The behavior of division has changed significantly over time: In Python terms, the / operator represents true division (or simply division), while the // operator represents floor division. Before version 3.0, the / operator represents classic division. Rounding towards negative infinity, though a different method than in most languages, adds consistency to Python. For instance, this rounding implies that the equation (a + b)//b == a//b + 1 is always true. Also, the rounding implies that the equation b*(a//b) + a%b == a is valid for both positive and negative values of a. As expected, the result of a%b lies in the half-open interval [0, b), where b is a positive integer; however, maintaining the validity of the equation requires that the result must lie in the interval (b, 0] when b is negative. Python provides a round function for rounding a float to the nearest integer. For tie-breaking, Python 3 uses the round to even method: round(1.5) and round(2.5) both produce 2. Python versions before 3 used the round-away-from-zero method: round(0.5) is 1.0, and round(-0.5) is −1.0. Python allows Boolean expressions that contain multiple equality relations to be consistent with general usage in mathematics. For example, the expression a < b < c tests whether a is less than b and b is less than c. C-derived languages interpret this expression differently: in C, the expression would first evaluate a < b, resulting in 0 or 1, and that result would then be compared with c. Python uses arbitrary-precision arithmetic for all integer operations. The Decimal type/class in the decimal module provides decimal floating-point numbers to a pre-defined arbitrary precision with several rounding modes. The Fraction class in the fractions module provides arbitrary precision for rational numbers. Due to Python's extensive mathematics library and the third-party library NumPy, the language is frequently used for scientific scripting in tasks such as numerical data processing and manipulation. Functions are created in Python by using the def keyword. A function is defined similarly to how it is called, by first providing the function name and then the required parameters. Here is an example of a function that prints its inputs: To assign a default value to a function parameter in case no actual value is provided at run time, variable-definition syntax can be used inside the function header. Code examples "Hello, World!" program: Program to calculate the factorial of a non-negative integer: Libraries Python's large standard library is commonly cited as one of its greatest strengths. For Internet-facing applications, many standard formats and protocols such as MIME and HTTP are supported. The language includes modules for creating graphical user interfaces, connecting to relational databases, generating pseudorandom numbers, arithmetic with arbitrary-precision decimals, manipulating regular expressions, and unit testing. Some parts of the standard library are covered by specifications—for example, the Web Server Gateway Interface (WSGI) implementation wsgiref follows PEP 333—but most parts are specified by their code, internal documentation, and test suites. However, because most of the standard library is cross-platform Python code, only a few modules must be altered or rewritten for variant implementations. As of 13 March 2025,[update] the Python Package Index (PyPI), the official repository for third-party Python software, contains over 614,339 packages. Development environments Most[which?] Python implementations (including CPython) include a read–eval–print loop (REPL); this permits the environment to function as a command line interpreter, with which users enter statements sequentially and receive results immediately. Also, CPython is bundled with an integrated development environment (IDE) called IDLE, which is oriented toward beginners.[citation needed] Other shells, including IDLE and IPython, add additional capabilities such as improved auto-completion, session-state retention, and syntax highlighting. Standard desktop IDEs include PyCharm, Spyder, and Visual Studio Code; there are web browser-based IDEs, such as the following environments: Implementations CPython is the reference implementation of Python. This implementation is written in C, meeting the C11 standard since version 3.11. Older versions use the C89 standard with several select C99 features, but third-party extensions are not limited to older C versions—e.g., they can be implemented using C11 or C++. CPython compiles Python programs into an intermediate bytecode, which is then executed by a virtual machine. CPython is distributed with a large standard library written in a combination of C and native Python. CPython is available for many platforms, including Windows and most modern Unix-like systems, including macOS (and Apple M1 Macs, since Python 3.9.1, using an experimental installer). Starting with Python 3.9, the Python installer intentionally fails to install on Windows 7 and 8; Windows XP was supported until Python 3.5, with unofficial support for VMS. Platform portability was one of Python's earliest priorities. During development of Python 1 and 2, even OS/2 and Solaris were supported; since that time, support has been dropped for many platforms. All current Python versions (since 3.7) support only operating systems that feature multithreading, by now supporting not nearly as many operating systems (dropping many outdated) than in the past. All alternative implementations have at least slightly different semantics. For example, an alternative may include unordered dictionaries, in contrast to other current Python versions. As another example in the larger Python ecosystem, PyPy does not support the full C Python API. Creating an executable with Python often is done by bundling an entire Python interpreter into the executable, which causes binary sizes to be massive for small programs, yet there exist implementations that are capable of truly compiling Python. Alternative implementations include the following: Stackless Python is a significant fork of CPython that implements microthreads. This implementation uses the call stack differently, thus allowing massively concurrent programs. PyPy also offers a stackless version. Just-in-time Python compilers have been developed, but are now unsupported: There are several compilers/transpilers to high-level object languages; the source language is unrestricted Python, a subset of Python, or a language similar to Python: There are also specialized compilers: Some older projects existed, as well as compilers not designed for use with Python 3.x and related syntax: A performance comparison among various Python implementations, using a non-numerical (combinatorial) workload, was presented at EuroSciPy '13. In addition, Python's performance relative to other programming languages is benchmarked by The Computer Language Benchmarks Game. There are several approaches to optimizing Python performance, despite the inherent slowness of an interpreted language. These approaches include the following strategies or tools: Language Development Python's development is conducted mostly through the Python Enhancement Proposal (PEP) process; this process is the primary mechanism for proposing major new features, collecting community input on issues, and documenting Python design decisions. Python coding style is covered in PEP 8. Outstanding PEPs are reviewed and commented on by the Python community and the steering council. Enhancement of the language corresponds with development of the CPython reference implementation. The mailing list python-dev is the primary forum for the language's development. Specific issues were originally discussed in the Roundup bug tracker hosted by the foundation. In 2022, all issues and discussions were migrated to GitHub. Development originally took place on a self-hosted source-code repository running Mercurial, until Python moved to GitHub in January 2017. CPython's public releases have three types, distinguished by which part of the version number is incremented: Many alpha, beta, and release-candidates are also released as previews and for testing before final releases. Although there is a rough schedule for releases, they are often delayed if the code is not ready yet. Python's development team monitors the state of the code by running a large unit test suite during development. The major academic conference on Python is PyCon. Also, there are special Python mentoring programs, such as PyLadies. Naming Python's name is inspired by the British comedy group Monty Python, whom Python creator Guido van Rossum enjoyed while developing the language. Monty Python references appear frequently in Python code and culture; for example, the metasyntactic variables often used in Python literature are spam and eggs, rather than the traditional foo and bar. Also, the official Python documentation contains various references to Monty Python routines. Python users are sometimes referred to as "Pythonistas". Languages influenced by Python See also Notes References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Zoophytes] | [TOKENS: 988] |
Contents Zoophyte A zoophyte (animal-plant) is an obsolete term for an organism thought to be intermediate between animals and plants, or an animal with plant-like attributes or appearance. In the 19th century they were reclassified as Radiata which included various taxa, a term superseded by Coelenterata referring more narrowly to the animal phyla Cnidaria (coral animals, true jellies, sea anemones, sea pens, and their allies), sponges, and Ctenophora (comb jellies). A group of strange creatures that exist somewhere on, or between, the boundaries of plants and animals kingdoms were the subject of considerable debate in the eighteenth century. Some naturalists believed that they were a blend of plant and animal; other naturalists considered them to be entirely either plant or animal (such as sea anemones). Ancient and medieval to early modern era In Eastern cultures such as Ancient China fungi were classified as plants in the Traditional Chinese Medicine texts, and cordyceps, and in particular Ophiocordyceps sinensis, were considered zoophytes. Zoophytes are common in medieval and renaissance era herbals, notable examples including the Tartar Lamb, a legendary plant which grew sheep as fruit. Zoophytes appeared in many influential early medical texts, such as Dioscorides's De Materia Medica and subsequent adaptations and commentaries on that work, notably Mattioli's Discorsi. Zoophytes are frequently seen as medieval attempts to explain the origins of exotic, unknown plants with strange properties (such as cotton, in the case of the Tartar Lamb as theorized by Henry Lee, Fellow of the Linnean Society in the book The Vegetable Lamb of Tartary). Reports of zoophytes continued into the seventeenth century and were commented on by many influential thinkers of the time period, including Francis Bacon. It was not until 1646 that claims of zoophytes began to be concretely refuted, and skepticism towards claims of zoophytes mounted throughout the seventeenth and eighteenth centuries. 18th to 19th century, natural history As natural history and natural philosophy developed in the 18th century, there was considerable debate and disagreements between naturalists about organisms on or near the boundary between the animal and plant kingdoms, and how to relate them in taxonomy. Interest in the topic began in the 1730s with the research by Abraham Trembley into polyps. When Carl Linnaeus published the 10th edition of Systema Naturae in 1758, marking the start of zoological nomenclature, he set out three divisions of the Kingdom of Nature: rocks, plants and animals, "though all three exist in the lithophytes", the corals. He defined zoophytes as "a composite small organism, with both animal and plant characteristics". He acknowledged contributions from the coralline expert Ellis by describing him as a "lynx-eyed discoverer of zoophytes". In 1761 he wrote to Ellis that "zoophytes have a mere vegetable life, and are increased every year under their bark, like trees" as shown by growth rings on the trunk of Gorgonia, they are "therefore vegetables, with flowers like small animals. As zoophytes are, many of them, covered with a stony coat, the Creator has been pleased that they should receive nourishment by their naked flowers. He has therefore furnished each with a pore, which we call a mouth." After wide research, in 1786 Ellis was still unconvinced "what or where the link is that divides the animal and vegetable kingdoms of Nature", and pressed Linnaeus to classify most as animals. He subsequently proposed that the animals of the corals construct their own structures, in a book completed by Daniel Solander. Georges Cuvier in his Le Règne Animal of 1817 titled one of his four divisions (Embranchements) of the animal kingdom "Les Zoophytes ou Animaux Rayonnés". An 1834 English translation uses the term Radiata, and titles the division "The Zoophytes, or Animalia Radiata", an expanded 1840 translation notes that "Neither of these names is literally applicable, for all the animals in the division are not radiated; and the very name Zoophyte, 'plant - animal,' is a contradiction. In England, the term Zoophyte is much more restricted than in France, but it is equally inapplicable, excepting, perhaps, to those species, about which there are still disputes as to whether they are animals or vegetables." Despite its scientific obsolescence, Charles Darwin continued to use the term throughout his studies. References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Wandjina] | [TOKENS: 1127] |
Contents Wandjina The Wandjina, also written Wanjina and Wondjina and also known as Gulingi, are cloud and rain spirits from the Wanjina Wunggurr cultural bloc of Aboriginal Australians, depicted prominently in rock art in northwestern Australia. Some of the artwork in the Kimberley region of Western Australia dates back to approximately 4,000 years ago. Another closely related spirit entity is the creator being Wunngurr, a being analogous to the Rainbow Serpent in other Aboriginal peoples' belief systems, but with a different interpretation. The stories of the Wandjina and the artwork depicting them remain important to the Mowanjum Community of Aboriginal people, and are one of the basic cultural elements of the Wanjina Wunggurr cultural bloc, which includes four Aboriginal peoples in the Kimberley. Dreamtime legends Some Dreamtime stories say the Wandjina created the landscape and its inhabitants, and continue to have influence over both. When the spirits found the place they would die, they painted their images on cave walls and entered a nearby waterhole. These paintings were then refreshed by Aboriginal people as a method of regenerating life force. The Wandjina can punish those who break the law with floods, lightning and cyclones. The Wandjina and Wunggurr spirits are essential elements of the life of the cultural bloc known as the Wanjina Wunggurr, consisting of the Worrorra (and neighbouring Ngardi), Wunambal and Ngarinyin peoples of the Kimberley. Rock paintings depicting Wanjina, as well as the Gwion Gwion ("Bradshaw") paintings, are evidence of the shared culture. The Wunambal people in the Mitchell Plateau area refer to Wandjina as Gulingi. Wunggurr is a variant on the Rainbow Serpent creator being belief, while the wandjina are local spirits, attached to places, and associated with particular clans. Although some local expressions use the two terms interchangeably, wungurr is a "more diffuse life force animating and underlying the particular manifestations of its power that find expression in all species of things, including the wandjina". One facet of wungurr is embodied in a rock python known as Wanjad. Traditional rock artwork The broad-stroke artwork of the Wandjina rock art dates to around 3800–4000 years ago. The emergence of this art style follows the end of a millennium-long drought that gave way to a wetter climate characterised by regular monsoons. The Wandjina paintings have common colours of black, red and yellow on a white background. The spirits are depicted alone or in groups, vertically or horizontally depending on the dimensions of the rock, and are sometimes depicted with figures and objects like the Rainbow Serpent or yams. Common composition is with large upper bodies and heads that may show eyes and nose, but typically no mouth. Two explanations have been given for this: they are so powerful they do not require speech and if they had mouths, the rain would never cease. Around the heads of Wandjina are lines or blocks of color, depicting lighting coming out of transparent helmets. Today, the paintings are still believed to possess these powers and therefore are to be approached and treated respectfully. Each site and painting has a name. Indigenous people of the Mowanjum community repaint the images to ensure the continuity of the Wandjina's presence. Annual repainting in December or January also ensures the arrival of the monsoon rains, according to Mowanjum belief. Repainting has occurred so often that at one site the paint is over 40 layers deep. The painting style has evolved during this process: more recent figures are stockier and some now possess eyelashes. Depictions on bark In the late 1960s and early 1970s several Mowanjum artists depicted traditional Wandjina on pieces of string bark. These bark paintings were sold mainly through the mission at Kalumbaru. Some of the important artists from this region include Alec Mingelmanganu, Charlie Numbelmoore and Jack Karedada. These artworks are now in major museum collections around the world. Other depictions Wandjina were the inspiration for a 1966 children's fantasy television series, Wandjina!, produced by ABC Television in 1966. In 2007, graffiti depictions of Wandjina appeared in Perth, Western Australia. Styles ranged from stencil-work to a spray painted Wandjina driving a pink car. Using Flickr and blogs, several people engaged in "Wandjina watching", documenting the Wandjina graffiti they found. These "wandering Wandjina" angered and upset some Indigenous people who said that only certain artists from their people are permitted to depict the Wandjina, without saying who these people are. A short film, Who Paintin' Dis Wandjina, discussed the Aboriginal reaction. Images of the Wandjina are displayed on the walls of the Ringwood Magistrates Court in Victoria; these are referenced as produced the National Gallery of Victoria.[citation needed] In 2016, during the Vivid Sydney festival, artwork of Wandjina by artist Donny Woolagoodja was projected onto the Sydney Opera House as part of its Lighting of the Sails celebration. In 2023, Wandjina was added to the popular mobile game Fate/Grand Order, initially as an antagonist and later as a playable character as a Foreigner-class Servant. See also References |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Ammar_al-Hakim] | [TOKENS: 635] |
Contents Ammar al-Hakim Sayyid Ammar al-Hakim (Arabic: عمار الحكيم; b. 1971) is an Iraqi cleric and politician who led the Islamic Supreme Council of Iraq (ISCI), from 2009 to 2017. He is currently the head of the National Wisdom Movement which is a political coalition in Iraq that was formed to contest the 2018 general election. Early life Al-Hakim was born in 1971 in Najaf, to cleric, Abdul Aziz Al-Hakim, who preceded him as leader of ISCI, and the grandson of grand Ayatollah Muhsin al-Hakim. Exile He went into exile in Iran in 1979 with his father after the government of Saddam Hussein had executed seven of his uncles and sixty two of his relatives. He attended private schools in Tehran and graduated from the Islamic Arabic University in Qom. He taught Arabic language, Islamic jurisprudence, logic, philosophy and the science of the Qur'an at the same university for several years. Career He supervises the establishment and management of many organizations, scientific and cultural institutions in the exile, including the House of Wisdom for Islamic Science. In 2003 he established the Al-Hakim Foundation under the supervision of his uncle Ayatollah Sayyid Muhammad Baqir Al-Hakim. The Al-Hakim Foundation later received consultative status with the Economic and Social Council of the United Nations. Headquartered in Najaf, it is now the largest institution of civil society in Iraq, with over 80 offices in all Governorates of Iraq, and working in the field of humanitarian assistance, development, human rights, and dialogue among religions. The Foundation also oversees several schools, colleges and cultural and scientific centers. The foundation holds numerous symposiums, conferences and social and cultural events, in addition to publishing a number of magazines and specialized publications. Detainment On February 23, 2007, he was detained by U.S. forces at a border when he was returning from Iran for 12 hours and he was released. The U.S. ambassador Zalmay Khalilzad apologized for the arrest and stressed that Washington did not mean any disrespect to al-Hakim or his family. Islamic Supreme Council of Iraq He succeeded to the leadership of the Islamic Supreme Council of Iraq (ISCI), a leading Shiite Arab party, on the death of his father, Abdul Aziz al-Hakim, in 2009. He led the ISCI-led coalition, Citizen Alliance, in the 2014 general elections, gaining 29 seats and becoming the third largest coalition. They joined the government of Haider al-Abadi, with senior ISCI member Adil Abdul-Mahdi becoming oil minister. In August 2017, in advance of the upcoming 2018 general elections, he announced his resignation from ISCI and the creation of a new coalition called the National Wisdom Movement (Al-Hikma). All except 5 of the existing MPs from the Citizens Alliance joined Al-Hikma. See also References |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Somali_diaspora] | [TOKENS: 2372] |
Contents Somali diaspora Somali diaspora (Somali: 𐒎𐒚𐒇𐒁𐒖𐒃𐒝𐒌𐒂𐒖 𐒈𐒝𐒑𐒛𐒐𐒘𐒕𐒜𐒆, romanized: Qurbajoogta Soomaaliyeed ) refers to ethnic Somalis who reside in areas of the world outside of Greater Somalia. The Somali Civil War greatly increased the size of the Somali diaspora, as many Somalis moved from Somalia primarily to Europe, North America, Oceania, North Africa, West Africa, Middle East, South Asia and Southern Africa. There are also Somali populations in Asia. The UN estimates that in 2015, approximately 2 million people from Somalia were living outside of the country's borders. Global distribution The distribution of Somalis abroad is uncertain, primarily due to confusion between the number of ethnic Somalis and the number of Somalia nationals. Whereas most recent Somali migrants in the diaspora emigrated as refugees and asylum seekers, many have since obtained either permanent residence or citizenship. A 2008 publication estimated the ethnic Somali international migrant population as 1,010,000, including around 300,000 in East and South Africa, 250,000 in the United States and Canada, 250,000 in Europe, 200,000 in the Middle East, and 10,000 in Australia, New Zealand and Asia. By the start of the Yemeni Civil War, the number of Somalis in Yemen was estimated to be 500,000. UN migration estimates of the international migrant stock 2015 suggest that 1,998,764 people from Somalia were living abroad. As of 31 March 2020[update], the number of refugees from Somalia registered with the UNHCR was 763,933. This is down from a peak of almost 990,000 in June 2013. The majority of these individuals were registered in Kenya (256,408), Yemen (253,755) and Ethiopia (198,670). According to USAID, many of the displaced persons in these adjacent territories are Bantus and other minorities.[citation needed] Africa About two thirds of all Somalis who live outside Somalia live in the neighbouring countries of Djibouti, Ethiopia, Kenya, and Yemen. Kenya hosted more than 313,000 in April 2017. About 255,000 were registered as refugees in Ethiopia. Besides their traditional areas of inhabitation in Greater Somalia (the former Italian Somaliland, British Somaliland, French Somaliland, the Ogaden, and Northeastern Kenya), a Somali community mainly consisting of businesspeople, academics and students also exists in Egypt. In addition, there is a historical Somali community in the general Sudan area. Primarily concentrated in the north and Khartoum, the expatriate community mainly consists of students, as well as some businesspeople. More recently, Somali entrepreneurs have also established themselves in South Africa, where they provide most of the retail trade in informal settlements around the Western Cape province. Europe While the distribution of Somalis per country in Europe is difficult to measure since the Somali expatriate community on the continent has grown so quickly in recent years, there are significant Somali communities in the United Kingdom: 108,000 (2018); Sweden: 66,369 (2016); Norway: 43,196 (2016); the Netherlands: 39,465 (2016); Germany: 38,675 (2016); Denmark: 21,050 (2016); and Finland: 24,365 (2022). Most Somalis in Denmark emigrated from Somalia following the start of the Somali Civil War (1986-), in the period between 1995 and 2000. According to Statistics Denmark, as of 2017, there are a total 21,204 persons of Somali origin living in Denmark. Of those individuals, 11,832 are Somalia-born immigrants and 9,372 are descendants of Somalia-born persons. 8,852 individuals are citizens of Somalia (4,730 men, 4,122 women). Between 2008 and 2013, employment among Somalis fell from 38% to 26%. By December 2018, nearly 1000 Somalis in Denmark lost their residence permits after the Danish Immigration Service started a review of the permits in 2017. The permits were revoked as parts of Somalia are safe enough for refugees to return. In 2018, analysis showed about 44% of Somalis in Denmark live in a parallel society and were the immigrant group which were lagging behind the most in terms of integration into Danish society. In 2018, nearly 50% were in long-term unemployment (4 years or more). Somalis are one of the largest ethnic minorities in Finland, and the largest group of people of non-European origin. In 2009, there were 5,570 Somali citizens, but an equal number may have received Finnish citizenship. In 2014 there were 16,721 Somali speakers in Finland.[citation needed] According to the Finnish newspaper Helsingin Sanomat, the number of Somali-speaking people in Finland in 2010 rose by nearly 10% in a year. In 2022, there were 24,365 people resident in Finland with a Somali background. Somalis in Germany are citizens and residents of Germany who are of Somali descent. According to the Federal Statistical Office of Germany, as of 2020, there are a total 47,495 Somalia-born immigrants living in Germany. From 1989 to 1998, the Netherlands was the second-most common European destination for Somali immigrants, only slightly behind the United Kingdom and more than double the total of the next-most common destination, Denmark. Between 2000 and 2005, there was a significant exodus of Somalis from the Netherlands to the United Kingdom, unofficially estimated to be as large as 20,000 people. The exodus was partly the result of anti-Muslim sentiments in the Netherlands in the wake of the rise of and assassination of right-wing politician Pim Fortuyn. According to the Somali European Forum, the exodus started earlier due to the Somalis not wanting to adapt to the assimilation process Dutch authorities imposed on them which they felt tried to change their culture. In 2012, unemployment was particularly high among Somali immigrants at 37%, while unemployment for all non-Western population was at 16%. According to official statistics on immigrants, Somalis had higher unemployment and dependence on social welfare while the crime rate of young male Somalis was high. Unemployment among immigrants rose higher than for the indigenous population in the aftermath of the 2008 financial crisis. According to a 2018 report using 2015 data, Islam takes a central role in the lives of nearly all Somalis and in many ways their religiosity increased from the already high levels in 2009. Somalis in Norway are citizens and residents of Norway who are of Somali descent. They are the biggest African migration group in Norway. 36.5% of Somalis in Norway live in the capital Oslo. Almost all Somali in Norway have come to Norway as refugees from the Somali Civil War. In 2016, Somalis were the largest non-European migrant group in Norway. Somalis in Sweden (Swedish: Somalier i Sverige; Somali: Soomaalida Iswiidhan) are citizens and residents of Sweden who are of Somali ancestry or are Somali citizens. A large proportion of Somalis in Sweden emigrated to Sweden due to the Somali Civil War. Most Somalis in Sweden arrived to the country after the year 2006. Although most Somalis in the United Kingdom are recent arrivals, the first Somalis to arrive were seamen and traders who settled in port cities in the late 19th century. By 2001, the UK census reported 43,532 Somali-born residents, making the Somali community in Britain the largest Somali expatriate population in Europe. The Office for National Statistics (ONS) estimate of 2018 indicates that 108,000 Somalis live in the UK. There has also been some secondary migration of Somalis from mainland European countries to the United Kingdom. According to the 2011 UK Census, 71.5% of Somalia-born residents in England and Wales hold a UK passport. Established Somali communities are found in London, Birmingham, Liverpool, Cardiff and Bristol, and newer ones have formed in Manchester, Sheffield and Leicester. The Somali population in London alone accounts for roughly 78% of Britain's Somali residents. North America The earliest ethnic Somali immigrants to the United States were sailors who arrived in the 1920s, mainly from British Somaliland. In the late 1970s, more Somali immigrants followed. Not until the 1990s when the civil war broke out in Somalia did the majority of Somalis come to the US. In 2008–2012, the largest concentration of Somalia-born people in the United States were in the Minneapolis-St. Paul-Bloomington area (17,320) of Minnesota. Other metropolitan areas with significant numbers of Somali Americans include Columbus, Ohio (10,280), Seattle-Tacoma-Bellevue in Washington (7,850), San Diego-Carlsbad-San Marcos in California (2,845), Washington, D.C.-Arlington-Alexandria in the Virginia-D.C. area (2,715), Atlanta-Sandy Springs-Marietta in Georgia (2,305), Phoenix-Mesa-Glendale in Arizona (1,965), Portland-Vancouver-Hillsboro in Oregon (1,480), Nashville-Davidson-Murfreesboro-Franklin in Tennessee (1,420), Boston-Cambridge-Quincy in Massachusetts (1,380), and other areas (28,650). Canada hosts one of the largest Somali populations in the Western world, with the 2011 National Household Survey reporting 44,995 people claiming Somali descent, though an unofficial estimate placed the figure as high as 150,000 residents. Somalis tend to be concentrated in the southern part of the province of Ontario, especially the Ottawa and Toronto areas. The Albertan cities of Calgary and Edmonton have also seen a significant increase in their respective Somali communities over the past five years. In addition, the neighbourhood of Rexdale in the Toronto borough of Etobicoke has one of the largest Somali populations in the country. Statistics Canada's 2006 Census ranks people of Somali descent as the 69th largest ethnic group in Canada. Middle East There is a sizable Somali community in the United Arab Emirates. Somali-owned businesses line the streets of Deira, the Dubai city centre, with only Iranians exporting more products from the city at large. Relations between the modern-day territories of Somalia and Yemen stretch back to antiquity. A number of Somali clans trace descent to the latter region. During the colonial period, disgruntled Yemenis from the Hadhrami wars sought and received asylum in various Somali towns. Yemen in turn unconditionally opened its borders to Somali nationals following the outbreak of the civil war in Somalia in the early 1990s. In 2015, after the Saudi-led military intervention in Yemen, many returning Somali expatriates as well as various foreign nationals began emigrating from Yemen to Somaliland. See also References Bibliography External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Input_device] | [TOKENS: 946] |
Contents Input device In computing, an input device is a piece of equipment used to enter data and control signals to an information processing system, such as a computer or information appliance. Examples of input devices include keyboards, computer mice, scanners, cameras, joysticks, and microphones. Input devices can be categorized based on: Keyboard A keyboard is a human interface device which is represented as a matrix of buttons. Each button, or key, can be used to either input an alphanumeric character to a computer, or to call upon a particular function of the computer. It acts as the main text entry interface for most users. Keyboards are available in many form factors, depending on the use case. Standard keyboards can be categorized by its size and number of keys, and the type of switch it employs. Other keyboards cater to specific use cases, such as a numeric keypad or a keyer. Desktop keyboards are typically large, often have full key travel distance, and features such as multimedia keys and a numeric keypad. Keyboards on laptops and tablets typically compromise on comfort to achieve a thin figure. There are various switch technologies used in modern keyboards, such as mechanical switches (which use springs), scissor switches (usually found on a laptop keyboard), or a membrane. Other keyboards do not have physical keys, such as a virtual keyboard, or a projection keyboard. Pointing device A pointing device allows a user to input spatial data to a computer. It is commonly used as a simple and intuitive way to select items on a computer screen on a graphical user interface (GUI), either by moving a mouse pointer, or, in the case of a touch screen, by physically touching the item on screen. Common pointing devices include mice, touchpads, and touch screens. Whereas mice operate by detecting their displacement on a surface, analog devices, such as 3D mice, joysticks, or pointing sticks, function by reporting their angle of deflection. Pointing devices can be classified on: Direct input is almost necessarily absolute, but indirect input may be either absolute or relative. For example, digitizing graphics tablets that do not have an embedded screen involve indirect input and sense absolute positions and are often run in an absolute input mode, but they may also be set up to simulate a relative input mode like that of a touchpad, where the stylus or puck can be lifted and repositioned. Embedded LCD tablets, which are also referred to as graphics tablet monitors, are the extension of digitizing graphics tablets. They enable users to see the real-time positions via the screen while being used. Sensors A sensor is an input device which produces data based on physical properties. Sensors are commonly found in mobile devices to detect their physical orientation and acceleration, but may also be found in desktop computers in the form of a thermometer used to monitor system temperature. Some sensors can be built with MEMS, which allows them to be microscopic in size. High-degree of freedom input devices Some devices allow many continuous degrees of freedom as input. These can be used as pointing devices, but are generally used in ways that don't involve pointing to a location in space, such as the control of a camera angle while in 3D applications. These kinds of devices are typically used in virtual reality systems (CAVEs), where input that registers six degrees of freedom is required.[citation needed] Composite devices Input devices, such as buttons and joysticks, can be combined on a single physical device that could be thought of as a composite device. Many gaming devices have controllers like this. Technically mice are composite devices, as they both track movement and provide buttons for clicking, but composite devices are generally considered to have more than two different forms of input. Video input devices Video input devices are used to digitize images or video from the outside world into the computer. The information can be stored in a multitude of formats depending on the user's requirement. Many video input devices use a camera sensor. Voice recorder Voice input devices are used to capture sound. In some cases, an audio output device can be used as an input device, in order to capture produced sound. Audio input devices allow a user to send audio info to a computer for processing, recording, or carrying out commands. Devices such as microphones allow users to speak to the computer in order to record a voice message or navigate software. Aside from recording, audio input devices are also used with speech recognition software. Punched paper Punched cards and punched tapes were used often in the 20th century. A punched hole represented a one; its absence represented a zero. A mechanical or optical reader was used to input a punched card or tape. Other types See also References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/w/index.php?title=Social_network&action=edit§ion=30] | [TOKENS: 1430] |
Editing Social network (section) Copy and paste: – — ° ′ ″ ≈ ≠ ≤ ≥ ± − × ÷ ← → · § Cite your sources: <ref></ref> {{}} {{{}}} | [] [[]] [[Category:]] #REDIRECT [[]] <s></s> <sup></sup> <sub></sub> <code></code> <pre></pre> <blockquote></blockquote> <ref></ref> <ref name="" /> {{Reflist}} <references /> <includeonly></includeonly> <noinclude></noinclude> {{DEFAULTSORT:}} <nowiki></nowiki> <!-- --> <span class="plainlinks"></span> Symbols: ~ | ¡ ¿ † ‡ ↔ ↑ ↓ • ¶ # ∞ ‹› «» ¤ ₳ ฿ ₵ ¢ ₡ ₢ $ ₫ ₯ € ₠ ₣ ƒ ₴ ₭ ₤ ℳ ₥ ₦ ₧ ₰ £ ៛ ₨ ₪ ৳ ₮ ₩ ¥ ♠ ♣ ♥ ♦ 𝄫 ♭ ♮ ♯ 𝄪 © ¼ ½ ¾ Latin: A a Á á À à  â Ä ä Ǎ ǎ Ă ă Ā ā à ã Å å Ą ą Æ æ Ǣ ǣ B b C c Ć ć Ċ ċ Ĉ ĉ Č č Ç ç D d Ď ď Đ đ Ḍ ḍ Ð ð E e É é È è Ė ė Ê ê Ë ë Ě ě Ĕ ĕ Ē ē Ẽ ẽ Ę ę Ẹ ẹ Ɛ ɛ Ǝ ǝ Ə ə F f G g Ġ ġ Ĝ ĝ Ğ ğ Ģ ģ H h Ĥ ĥ Ħ ħ Ḥ ḥ I i İ ı Í í Ì ì Î î Ï ï Ǐ ǐ Ĭ ĭ Ī ī Ĩ ĩ Į į Ị ị J j Ĵ ĵ K k Ķ ķ L l Ĺ ĺ Ŀ ŀ Ľ ľ Ļ ļ Ł ł Ḷ ḷ Ḹ ḹ M m Ṃ ṃ N n Ń ń Ň ň Ñ ñ Ņ ņ Ṇ ṇ Ŋ ŋ O o Ó ó Ò ò Ô ô Ö ö Ǒ ǒ Ŏ ŏ Ō ō Õ õ Ǫ ǫ Ọ ọ Ő ő Ø ø Œ œ Ɔ ɔ P p Q q R r Ŕ ŕ Ř ř Ŗ ŗ Ṛ ṛ Ṝ ṝ S s Ś ś Ŝ ŝ Š š Ş ş Ș ș Ṣ ṣ ß T t Ť ť Ţ ţ Ț ț Ṭ ṭ Þ þ U u Ú ú Ù ù Û û Ü ü Ǔ ǔ Ŭ ŭ Ū ū Ũ ũ Ů ů Ų ų Ụ ụ Ű ű Ǘ ǘ Ǜ ǜ Ǚ ǚ Ǖ ǖ V v W w Ŵ ŵ X x Y y Ý ý Ŷ ŷ Ÿ ÿ Ỹ ỹ Ȳ ȳ Z z Ź ź Ż ż Ž ž ß Ð ð Þ þ Ŋ ŋ Ə ə Greek: Ά ά Έ έ Ή ή Ί ί Ό ό Ύ ύ Ώ ώ Α α Β β Γ γ Δ δ Ε ε Ζ ζ Η η Θ θ Ι ι Κ κ Λ λ Μ μ Ν ν Ξ ξ Ο ο Π π Ρ ρ Σ σ ς Τ τ Υ υ Φ φ Χ χ Ψ ψ Ω ω {{Polytonic|}} Cyrillic: А а Б б В в Г г Ґ ґ Ѓ ѓ Д д Ђ ђ Е е Ё ё Є є Ж ж З з Ѕ ѕ И и І і Ї ї Й й Ј ј К к Ќ ќ Л л Љ љ М м Н н Њ њ О о П п Р р С с Т т Ћ ћ У у Ў ў Ф ф Х х Ц ц Ч ч Џ џ Ш ш Щ щ Ъ ъ Ы ы Ь ь Э э Ю ю Я я ́ IPA: t̪ d̪ ʈ ɖ ɟ ɡ ɢ ʡ ʔ ɸ β θ ð ʃ ʒ ɕ ʑ ʂ ʐ ç ʝ ɣ χ ʁ ħ ʕ ʜ ʢ ɦ ɱ ɳ ɲ ŋ ɴ ʋ ɹ ɻ ɰ ʙ ⱱ ʀ ɾ ɽ ɫ ɬ ɮ ɺ ɭ ʎ ʟ ɥ ʍ ɧ ʼ ɓ ɗ ʄ ɠ ʛ ʘ ǀ ǃ ǂ ǁ ɨ ʉ ɯ ɪ ʏ ʊ ø ɘ ɵ ɤ ə ɚ ɛ œ ɜ ɝ ɞ ʌ ɔ æ ɐ ɶ ɑ ɒ ʰ ʱ ʷ ʲ ˠ ˤ ⁿ ˡ ˈ ˌ ː ˑ ̪ {{IPA|}} This page is a member of 6 hidden categories (help): |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Amnon_Rubinstein] | [TOKENS: 2304] |
Contents Amnon Rubinstein Amnon Rubinstein (Hebrew: אמנון רובינשטיין; 5 September 1931 – 18 January 2024) was an Israeli legal scholar, politician and recipient of the Israel Prize. A member of the Knesset between 1977 and 2002, he served in several ministerial positions. He is referred to as the “founding father of Israeli Constitutional Law” In later life he was dean of the Interdisciplinary Center (IDC) in Herzliya and a patron of Liberal International. Biography Rubinstein was born in Tel Aviv to Aharon, a building contractor and founder of the "Rubinstein Group", and Rachel, who immigrated from Poland to Israel in the early 1920s. His family opposed Labor Zionism and in his youth, he supported the Irgun. Rubinstein was active in public life from a young age. In the early 1950s, he was a member of the Volunteers' Row, a public organization aimed at fighting corruption and helping new immigrants. He attended the Geula Commercial High School in Tel Aviv and later at a high school in Baltimore, United States. During his mandatory service in the Israel Defense Forces, he served as an officer in the Artillery Corps. He later retrained and served in the anti-aircraft unit, continuing his service there in the reserves. After his military service, Rubinstein studied in law, economics, and international relations at the Hebrew University of Jerusalem. He did his legal internship at the State Attorney's Office and also with attorney Haim Yosef Zadok, who later became the Minister of Justice. Rubinstein was licensed as a lawyer in 1961. He received his Ph.D. in law from the London School of Economics in 1960. His doctoral dissertation, titled "Void and Voidable Actions in Administrative Law", was published as a book by Oxford University Press. At the age of 25, during his studies, Rubinstein married Roni a fellow student and granddaughter of Jacob and Perla Shlush, founders of Tel Aviv. They met at Kibbutz Afikim during fortification work prior to the Suez Crisis. Roni later became a lawyer, and the couple had a son and a daughter. Rubinstein returned to public activism after the Yom Kippur War. In March 1974, he co-founded the Shinui Party. Shinui merged with other liberal groups to form Democratic Movement for Change (Dash) in November 1976. Dash won 15 seats in the 1977 elections, but did not achieve its aim of being a decisive political force as Menachem Begin formed a government without its support. When Dash joined Begin’s government in November of that year, Rubinstein refused to join the coalition, arguing that its platform was insufficiently represented in the agreement. Dash split into three parties on September 14, 1978. Rubinstein led a faction called the Movement for Change and Initiative, later renamed Shinui – Center Party. The new Shinui was elected to the 10th Knesset, the 11th Knesset, and the 12th Knesset. In 1992, Shinui merged with the Ratz and Mapam parties to form Meretz. Rubinstein significantly advanced the Israeli Constitution Project. During the 12th Knesset, he initiated the proposal for the Basic Law on Human Rights and suggested splitting it into four separate proposals, including the Basic Law: Human Dignity and Liberty and the Basic Law: Freedom of Occupation. This strategy led to the successful passage of both laws. Elected to the 9th Knesset in the 1977 elections, Rubinstein and served continuously until 2002. He held various leadership roles, including chairing the Constitution, Law and Justice Committee (in the 15th Knesset), the Economic Affairs Committee (in the 14th Knesset), and the State Control Committee (in the 15th Knesset). He was also a member of the Judicial Selection Committee. In 1999, Avraham Burg, the Speaker of the Knesset, received a false message from Zalman Shoshi claiming that Rubinstein had died. In response, the Knesset members held a moment of silence in Rubinstein's memory, despite the fact that he was still alive. During the session, Burg delivered a eulogy and recited the El Malei Rachamim. Rubinstein announced his resignation from the Knesset in July 2002, which took effect in October of that year. Rubinstein was a Liberal International patron and previously as a vice president of the organization. In 1984, Rubinstein was appointed Minister of Communications in the National Unity Government led by Shimon Peres. He held this position until 1987, serving in the 21st and 22nd Israeli governments. During his tenure, he introduced several reforms, including the establishment of commercial television, cable television, and regional radio stations. Following the 1992 elections and the formation of Yitzhak Rabin's government, Rubinstein was appointed Minister of Science and Technology and Minister of Energy and Infrastructure. During this period, he initiated the Electricity Sector Law, 1996, which led to the creation of the Public Utilities Authority – Electricity. He also worked to introduce competition in the fuel industry by breaking the monopoly of fuel companies on imports. In 1994, Rubinstein was appointed Minister of Education, Culture, and Sport after Shulamit Aloni's resignation. He held the position until 1996. As Minister of Education, Rubinstein gained popularity for his liberal views, particularly on high school matriculation exams and his criticism of the Psychometric Entrance Test, remarking that he might not have been able to study law had such a test existed during his time. Rubinstein introduced two major reforms during his tenure. The first focused on restructuring matriculation exams. Under his leadership, the Ministry of Education increased the matriculation pass rate by implementing measures such as eliminating external exams for some core subjects. Between 1994 and 1996, the matriculation eligibility rate rose by over 5%, reaching approximately 45% of each student cohort. The second reform involved the establishment of academic colleges. This initiative included both state-funded institutions, such as the Tel Aviv–Yafo Academic College, and privately funded ones, like the Interdisciplinary Center Herzliya. After retiring from politics, Rubinstein became a vocal critic of the Israeli left. He accused the left of becoming increasingly radical in efforts to accommodate Palestinian rights. Following the signing of the Geneva Initiative by then-Meretz leader Yossi Beilin, Rubinstein left Meretz, opposing its compromise on the Law of Return. In 2007, during a legal debate over proposed judicial reforms by Justice Minister Daniel Friedmann, Rubinstein supported several of Friedmann's initiatives. He co-initiated a petition endorsing these reforms, stating: "My aim was to dispel the impression that the professors in the universities had created, as if the entire academia opposed Friedmann's initiatives. Every matter should be examined on its own merit, not through personal attacks." In 1963, Rubinstein became the first dean of the Faculty of Law at Tel Aviv University, a position he held until 1970. His research focused on the constitutional law of Israel, exploring topics such as the state's nature, Zionism, the status of the Law of Return, and the civil rights of Palestinian citizens of Israel. In 2006, Rubinstein was awarded the Israel Prize for legal research. The judging panel described him as: "The father of Israeli constitutional law. Through his deep academic writing and diverse public activities, he promotes democracy, equality, and human rights. Few can match his contribution to the state of Israel as Prof. Amnon Rubinstein – as a public figure, as a member of the legislative and executive branches, as a researcher, and as a brilliant jurist." After retiring from the Knesset in 2002, Rubinstein served as the dean of the Radzyner School of Law at the Interdisciplinary Center Herzliya from 2002 to 2004 and briefly as the president of the Interdisciplinary Center. He also held positions as a visiting professor at international academic institutions, including Stanford University and Columbia University. Rubinstein had a media career as a columnist and commentator. In the 1960s, he wrote articles criticizing religious coercion and concessions made to religious political parties. He gained fame as a popular columnist and interviewee. Between 1969 and 1972, he hosted a debate program on Israel Educational Television titled Boomerang. From 1964 until 2004, he was a member of the editorial board of Haaretz and a regular contributor, focusing on issues of religion and state and addressing antisemitism, particularly that which he attributed to the political left in Europe. He also wrote regularly for the Ma'ariv weekend supplement and international publications, including The New York Times. In 2005, Rubinstein published his first novel, The Blanket, which explored interconnected stories of Israeli characters from pre-state days to the present, addressing central themes in Israeli society. He followed this with Highway No. 5 (2006), The Sea Above Us (2007), Separate Entrance (2009), and Forbidden Loves (2010). In 2022, he published the futuristic novel Methuselah. Rubinstein hosted a radio program on Kol HaMusika, A Musical Morning for Truck and Bus Drivers, from 2014 to 2015. Rubinstein wrote extensively on the investigation and prosecution of public figures in Israel, arguing that enforcement bodies wield excessive and unchecked power over elected officials. He highlighted cases where such power appeared to have been misused. In 2020, Rubinstein participated in The Future Archive, a documentation project highlighting the generation of intellectuals who significantly influenced early Israeli culture. Rubinstein lived in Tel Aviv and died on January 18, 2024. He was survived by a son and a daughter. Awards In 2006, Rubinstein won the Israel Prize, for law. The Israel Prize award committee provided the following endorsement for its decision: "[Amnon Rubinstein is] the founding father of Israeli constitutional law… In the legal and public arena in Israel, there are few who can equal Prof. Amnon Rubinstein’s contribution to the State of Israel…" In 2003, Rubinstein was named a Knight of Quality Government in the Lifetime Achievement category by the Movement for Quality Government. He received the Hashin Prize for Academic Excellence in Law in 2010, the Gorney Prize for Public Law (2016) for his contribution to public law, the ISEF Award for social contribution in 2013, the Herzog Prize for unique contributions to Israel (2018), and the Ramat Gan Literary Award (2015). In 2022, he was awarded the EMET Prize for Law. In 2022, the "Rubinstein Center for Constitutional Challenges" was established in his name at Reichman University. Selected works by Amnon Rubinstein Non-fiction books: Novels: Further reading See also References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Animal#cite_note-160] | [TOKENS: 6011] |
Contents Animal Animals are multicellular, eukaryotic organisms belonging to the biological kingdom Animalia (/ˌænɪˈmeɪliə/). With few exceptions, animals consume organic material, breathe oxygen, have myocytes and are able to move, can reproduce sexually, and grow from a hollow sphere of cells, the blastula, during embryonic development. Animals form a clade, meaning that they arose from a single common ancestor. Over 1.5 million living animal species have been described, of which around 1.05 million are insects, over 85,000 are molluscs, and around 65,000 are vertebrates. It has been estimated there are as many as 7.77 million animal species on Earth. Animal body lengths range from 8.5 μm (0.00033 in) to 33.6 m (110 ft). They have complex ecologies and interactions with each other and their environments, forming intricate food webs. The scientific study of animals is known as zoology, and the study of animal behaviour is known as ethology. The animal kingdom is divided into five major clades, namely Porifera, Ctenophora, Placozoa, Cnidaria and Bilateria. Most living animal species belong to the clade Bilateria, a highly proliferative clade whose members have a bilaterally symmetric and significantly cephalised body plan, and the vast majority of bilaterians belong to two large clades: the protostomes, which includes organisms such as arthropods, molluscs, flatworms, annelids and nematodes; and the deuterostomes, which include echinoderms, hemichordates and chordates, the latter of which contains the vertebrates. The much smaller basal phylum Xenacoelomorpha have an uncertain position within Bilateria. Animals first appeared in the fossil record in the late Cryogenian period and diversified in the subsequent Ediacaran period in what is known as the Avalon explosion. Nearly all modern animal phyla first appeared in the fossil record as marine species during the Cambrian explosion, which began around 539 million years ago (Mya), and most classes during the Ordovician radiation 485.4 Mya. Common to all living animals, 6,331 groups of genes have been identified that may have arisen from a single common ancestor that lived about 650 Mya during the Cryogenian period. Historically, Aristotle divided animals into those with blood and those without. Carl Linnaeus created the first hierarchical biological classification for animals in 1758 with his Systema Naturae, which Jean-Baptiste Lamarck expanded into 14 phyla by 1809. In 1874, Ernst Haeckel divided the animal kingdom into the multicellular Metazoa (now synonymous with Animalia) and the Protozoa, single-celled organisms no longer considered animals. In modern times, the biological classification of animals relies on advanced techniques, such as molecular phylogenetics, which are effective at demonstrating the evolutionary relationships between taxa. Humans make use of many other animal species for food (including meat, eggs, and dairy products), for materials (such as leather, fur, and wool), as pets and as working animals for transportation, and services. Dogs, the first domesticated animal, have been used in hunting, in security and in warfare, as have horses, pigeons and birds of prey; while other terrestrial and aquatic animals are hunted for sports, trophies or profits. Non-human animals are also an important cultural element of human evolution, having appeared in cave arts and totems since the earliest times, and are frequently featured in mythology, religion, arts, literature, heraldry, politics, and sports. Etymology The word animal comes from the Latin noun animal of the same meaning, which is itself derived from Latin animalis 'having breath or soul'. The biological definition includes all members of the kingdom Animalia. In colloquial usage, the term animal is often used to refer only to nonhuman animals. The term metazoa is derived from Ancient Greek μετα meta 'after' (in biology, the prefix meta- stands for 'later') and ζῷᾰ zōia 'animals', plural of ζῷον zōion 'animal'. A metazoan is any member of the group Metazoa. Characteristics Animals have several characteristics that they share with other living things. Animals are eukaryotic, multicellular, and aerobic, as are plants and fungi. Unlike plants and algae, which produce their own food, animals cannot produce their own food, a feature they share with fungi. Animals ingest organic material and digest it internally. Animals have structural characteristics that set them apart from all other living things: Typically, there is an internal digestive chamber with either one opening (in Ctenophora, Cnidaria, and flatworms) or two openings (in most bilaterians). Animal development is controlled by Hox genes, which signal the times and places to develop structures such as body segments and limbs. During development, the animal extracellular matrix forms a relatively flexible framework upon which cells can move about and be reorganised into specialised tissues and organs, making the formation of complex structures possible, and allowing cells to be differentiated. The extracellular matrix may be calcified, forming structures such as shells, bones, and spicules. In contrast, the cells of other multicellular organisms (primarily algae, plants, and fungi) are held in place by cell walls, and so develop by progressive growth. Nearly all animals make use of some form of sexual reproduction. They produce haploid gametes by meiosis; the smaller, motile gametes are spermatozoa and the larger, non-motile gametes are ova. These fuse to form zygotes, which develop via mitosis into a hollow sphere, called a blastula. In sponges, blastula larvae swim to a new location, attach to the seabed, and develop into a new sponge. In most other groups, the blastula undergoes more complicated rearrangement. It first invaginates to form a gastrula with a digestive chamber and two separate germ layers, an external ectoderm and an internal endoderm. In most cases, a third germ layer, the mesoderm, also develops between them. These germ layers then differentiate to form tissues and organs. Repeated instances of mating with a close relative during sexual reproduction generally leads to inbreeding depression within a population due to the increased prevalence of harmful recessive traits. Animals have evolved numerous mechanisms for avoiding close inbreeding. Some animals are capable of asexual reproduction, which often results in a genetic clone of the parent. This may take place through fragmentation; budding, such as in Hydra and other cnidarians; or parthenogenesis, where fertile eggs are produced without mating, such as in aphids. Ecology Animals are categorised into ecological groups depending on their trophic levels and how they consume organic material. Such groupings include carnivores (further divided into subcategories such as piscivores, insectivores, ovivores, etc.), herbivores (subcategorised into folivores, graminivores, frugivores, granivores, nectarivores, algivores, etc.), omnivores, fungivores, scavengers/detritivores, and parasites. Interactions between animals of each biome form complex food webs within that ecosystem. In carnivorous or omnivorous species, predation is a consumer–resource interaction where the predator feeds on another organism, its prey, who often evolves anti-predator adaptations to avoid being fed upon. Selective pressures imposed on one another lead to an evolutionary arms race between predator and prey, resulting in various antagonistic/competitive coevolutions. Almost all multicellular predators are animals. Some consumers use multiple methods; for example, in parasitoid wasps, the larvae feed on the hosts' living tissues, killing them in the process, but the adults primarily consume nectar from flowers. Other animals may have very specific feeding behaviours, such as hawksbill sea turtles which mainly eat sponges. Most animals rely on biomass and bioenergy produced by plants and phytoplanktons (collectively called producers) through photosynthesis. Herbivores, as primary consumers, eat the plant material directly to digest and absorb the nutrients, while carnivores and other animals on higher trophic levels indirectly acquire the nutrients by eating the herbivores or other animals that have eaten the herbivores. Animals oxidise carbohydrates, lipids, proteins and other biomolecules in cellular respiration, which allows the animal to grow and to sustain basal metabolism and fuel other biological processes such as locomotion. Some benthic animals living close to hydrothermal vents and cold seeps on the dark sea floor consume organic matter produced through chemosynthesis (via oxidising inorganic compounds such as hydrogen sulfide) by archaea and bacteria. Animals originated in the ocean; all extant animal phyla, except for Micrognathozoa and Onychophora, feature at least some marine species. However, several lineages of arthropods begun to colonise land around the same time as land plants, probably between 510 and 471 million years ago, during the Late Cambrian or Early Ordovician. Vertebrates such as the lobe-finned fish Tiktaalik started to move on to land in the late Devonian, about 375 million years ago. Other notable animal groups that colonized land environments are Mollusca, Platyhelmintha, Annelida, Tardigrada, Onychophora, Rotifera, Nematoda. Animals occupy virtually all of earth's habitats and microhabitats, with faunas adapted to salt water, hydrothermal vents, fresh water, hot springs, swamps, forests, pastures, deserts, air, and the interiors of other organisms. Animals are however not particularly heat tolerant; very few of them can survive at constant temperatures above 50 °C (122 °F) or in the most extreme cold deserts of continental Antarctica. The collective global geomorphic influence of animals on the processes shaping the Earth's surface remains largely understudied, with most studies limited to individual species and well-known exemplars. Diversity The blue whale (Balaenoptera musculus) is the largest animal that has ever lived, weighing up to 190 tonnes and measuring up to 33.6 metres (110 ft) long. The largest extant terrestrial animal is the African bush elephant (Loxodonta africana), weighing up to 12.25 tonnes and measuring up to 10.67 metres (35.0 ft) long. The largest terrestrial animals that ever lived were titanosaur sauropod dinosaurs such as Argentinosaurus, which may have weighed as much as 73 tonnes, and Supersaurus which may have reached 39 metres. Several animals are microscopic; some Myxozoa (obligate parasites within the Cnidaria) never grow larger than 20 μm, and one of the smallest species (Myxobolus shekel) is no more than 8.5 μm when fully grown. The following table lists estimated numbers of described extant species for the major animal phyla, along with their principal habitats (terrestrial, fresh water, and marine), and free-living or parasitic ways of life. Species estimates shown here are based on numbers described scientifically; much larger estimates have been calculated based on various means of prediction, and these can vary wildly. For instance, around 25,000–27,000 species of nematodes have been described, while published estimates of the total number of nematode species include 10,000–20,000; 500,000; 10 million; and 100 million. Using patterns within the taxonomic hierarchy, the total number of animal species—including those not yet described—was calculated to be about 7.77 million in 2011.[a] 3,000–6,500 4,000–25,000 Evolutionary origin Evidence of animals is found as long ago as the Cryogenian period. 24-Isopropylcholestane (24-ipc) has been found in rocks from roughly 650 million years ago; it is only produced by sponges and pelagophyte algae. Its likely origin is from sponges based on molecular clock estimates for the origin of 24-ipc production in both groups. Analyses of pelagophyte algae consistently recover a Phanerozoic origin, while analyses of sponges recover a Neoproterozoic origin, consistent with the appearance of 24-ipc in the fossil record. The first body fossils of animals appear in the Ediacaran, represented by forms such as Charnia and Spriggina. It had long been doubted whether these fossils truly represented animals, but the discovery of the animal lipid cholesterol in fossils of Dickinsonia establishes their nature. Animals are thought to have originated under low-oxygen conditions, suggesting that they were capable of living entirely by anaerobic respiration, but as they became specialised for aerobic metabolism they became fully dependent on oxygen in their environments. Many animal phyla first appear in the fossil record during the Cambrian explosion, starting about 539 million years ago, in beds such as the Burgess Shale. Extant phyla in these rocks include molluscs, brachiopods, onychophorans, tardigrades, arthropods, echinoderms and hemichordates, along with numerous now-extinct forms such as the predatory Anomalocaris. The apparent suddenness of the event may however be an artefact of the fossil record, rather than showing that all these animals appeared simultaneously. That view is supported by the discovery of Auroralumina attenboroughii, the earliest known Ediacaran crown-group cnidarian (557–562 mya, some 20 million years before the Cambrian explosion) from Charnwood Forest, England. It is thought to be one of the earliest predators, catching small prey with its nematocysts as modern cnidarians do. Some palaeontologists have suggested that animals appeared much earlier than the Cambrian explosion, possibly as early as 1 billion years ago. Early fossils that might represent animals appear for example in the 665-million-year-old rocks of the Trezona Formation of South Australia. These fossils are interpreted as most probably being early sponges. Trace fossils such as tracks and burrows found in the Tonian period (from 1 gya) may indicate the presence of triploblastic worm-like animals, roughly as large (about 5 mm wide) and complex as earthworms. However, similar tracks are produced by the giant single-celled protist Gromia sphaerica, so the Tonian trace fossils may not indicate early animal evolution. Around the same time, the layered mats of microorganisms called stromatolites decreased in diversity, perhaps due to grazing by newly evolved animals. Objects such as sediment-filled tubes that resemble trace fossils of the burrows of wormlike animals have been found in 1.2 gya rocks in North America, in 1.5 gya rocks in Australia and North America, and in 1.7 gya rocks in Australia. Their interpretation as having an animal origin is disputed, as they might be water-escape or other structures. Phylogeny Animals are monophyletic, meaning they are derived from a common ancestor. Animals are the sister group to the choanoflagellates, with which they form the Choanozoa. Ros-Rocher and colleagues (2021) trace the origins of animals to unicellular ancestors, providing the external phylogeny shown in the cladogram. Uncertainty of relationships is indicated with dashed lines. The animal clade had certainly originated by 650 mya, and may have come into being as much as 800 mya, based on molecular clock evidence for different phyla. Holomycota (inc. fungi) Ichthyosporea Pluriformea Filasterea The relationships at the base of the animal tree have been debated. Other than Ctenophora, the Bilateria and Cnidaria are the only groups with symmetry, and other evidence shows they are closely related. In addition to sponges, Placozoa has no symmetry and was often considered a "missing link" between protists and multicellular animals. The presence of hox genes in Placozoa shows that they were once more complex. The Porifera (sponges) have long been assumed to be sister to the rest of the animals, but there is evidence that the Ctenophora may be in that position. Molecular phylogenetics has supported both the sponge-sister and ctenophore-sister hypotheses. In 2017, Roberto Feuda and colleagues, using amino acid differences, presented both, with the following cladogram for the sponge-sister view that they supported (their ctenophore-sister tree simply interchanging the places of ctenophores and sponges): Porifera Ctenophora Placozoa Cnidaria Bilateria Conversely, a 2023 study by Darrin Schultz and colleagues uses ancient gene linkages to construct the following ctenophore-sister phylogeny: Ctenophora Porifera Placozoa Cnidaria Bilateria Sponges are physically very distinct from other animals, and were long thought to have diverged first, representing the oldest animal phylum and forming a sister clade to all other animals. Despite their morphological dissimilarity with all other animals, genetic evidence suggests sponges may be more closely related to other animals than the comb jellies are. Sponges lack the complex organisation found in most other animal phyla; their cells are differentiated, but in most cases not organised into distinct tissues, unlike all other animals. They typically feed by drawing in water through pores, filtering out small particles of food. The Ctenophora and Cnidaria are radially symmetric and have digestive chambers with a single opening, which serves as both mouth and anus. Animals in both phyla have distinct tissues, but these are not organised into discrete organs. They are diploblastic, having only two main germ layers, ectoderm and endoderm. The tiny placozoans have no permanent digestive chamber and no symmetry; they superficially resemble amoebae. Their phylogeny is poorly defined, and under active research. The remaining animals, the great majority—comprising some 29 phyla and over a million species—form the Bilateria clade, which have a bilaterally symmetric body plan. The Bilateria are triploblastic, with three well-developed germ layers, and their tissues form distinct organs. The digestive chamber has two openings, a mouth and an anus, and in the Nephrozoa there is an internal body cavity, a coelom or pseudocoelom. These animals have a head end (anterior) and a tail end (posterior), a back (dorsal) surface and a belly (ventral) surface, and a left and a right side. A modern consensus phylogenetic tree for the Bilateria is shown below. Xenacoelomorpha Ambulacraria Chordata Ecdysozoa Spiralia Having a front end means that this part of the body encounters stimuli, such as food, favouring cephalisation, the development of a head with sense organs and a mouth. Many bilaterians have a combination of circular muscles that constrict the body, making it longer, and an opposing set of longitudinal muscles, that shorten the body; these enable soft-bodied animals with a hydrostatic skeleton to move by peristalsis. They also have a gut that extends through the basically cylindrical body from mouth to anus. Many bilaterian phyla have primary larvae which swim with cilia and have an apical organ containing sensory cells. However, over evolutionary time, descendant spaces have evolved which have lost one or more of each of these characteristics. For example, adult echinoderms are radially symmetric (unlike their larvae), while some parasitic worms have extremely simplified body structures. Genetic studies have considerably changed zoologists' understanding of the relationships within the Bilateria. Most appear to belong to two major lineages, the protostomes and the deuterostomes. It is often suggested that the basalmost bilaterians are the Xenacoelomorpha, with all other bilaterians belonging to the subclade Nephrozoa. However, this suggestion has been contested, with other studies finding that xenacoelomorphs are more closely related to Ambulacraria than to other bilaterians. Protostomes and deuterostomes differ in several ways. Early in development, deuterostome embryos undergo radial cleavage during cell division, while many protostomes (the Spiralia) undergo spiral cleavage. Animals from both groups possess a complete digestive tract, but in protostomes the first opening of the embryonic gut develops into the mouth, and the anus forms secondarily. In deuterostomes, the anus forms first while the mouth develops secondarily. Most protostomes have schizocoelous development, where cells simply fill in the interior of the gastrula to form the mesoderm. In deuterostomes, the mesoderm forms by enterocoelic pouching, through invagination of the endoderm. The main deuterostome taxa are the Ambulacraria and the Chordata. Ambulacraria are exclusively marine and include acorn worms, starfish, sea urchins, and sea cucumbers. The chordates are dominated by the vertebrates (animals with backbones), which consist of fishes, amphibians, reptiles, birds, and mammals. The protostomes include the Ecdysozoa, named after their shared trait of ecdysis, growth by moulting, Among the largest ecdysozoan phyla are the arthropods and the nematodes. The rest of the protostomes are in the Spiralia, named for their pattern of developing by spiral cleavage in the early embryo. Major spiralian phyla include the annelids and molluscs. History of classification In the classical era, Aristotle divided animals,[d] based on his own observations, into those with blood (roughly, the vertebrates) and those without. The animals were then arranged on a scale from man (with blood, two legs, rational soul) down through the live-bearing tetrapods (with blood, four legs, sensitive soul) and other groups such as crustaceans (no blood, many legs, sensitive soul) down to spontaneously generating creatures like sponges (no blood, no legs, vegetable soul). Aristotle was uncertain whether sponges were animals, which in his system ought to have sensation, appetite, and locomotion, or plants, which did not: he knew that sponges could sense touch and would contract if about to be pulled off their rocks, but that they were rooted like plants and never moved about. In 1758, Carl Linnaeus created the first hierarchical classification in his Systema Naturae. In his original scheme, the animals were one of three kingdoms, divided into the classes of Vermes, Insecta, Pisces, Amphibia, Aves, and Mammalia. Since then, the last four have all been subsumed into a single phylum, the Chordata, while his Insecta (which included the crustaceans and arachnids) and Vermes have been renamed or broken up. The process was begun in 1793 by Jean-Baptiste de Lamarck, who called the Vermes une espèce de chaos ('a chaotic mess')[e] and split the group into three new phyla: worms, echinoderms, and polyps (which contained corals and jellyfish). By 1809, in his Philosophie Zoologique, Lamarck had created nine phyla apart from vertebrates (where he still had four phyla: mammals, birds, reptiles, and fish) and molluscs, namely cirripedes, annelids, crustaceans, arachnids, insects, worms, radiates, polyps, and infusorians. In his 1817 Le Règne Animal, Georges Cuvier used comparative anatomy to group the animals into four embranchements ('branches' with different body plans, roughly corresponding to phyla), namely vertebrates, molluscs, articulated animals (arthropods and annelids), and zoophytes (radiata) (echinoderms, cnidaria and other forms). This division into four was followed by the embryologist Karl Ernst von Baer in 1828, the zoologist Louis Agassiz in 1857, and the comparative anatomist Richard Owen in 1860. In 1874, Ernst Haeckel divided the animal kingdom into two subkingdoms: Metazoa (multicellular animals, with five phyla: coelenterates, echinoderms, articulates, molluscs, and vertebrates) and Protozoa (single-celled animals), including a sixth animal phylum, sponges. The protozoa were later moved to the former kingdom Protista, leaving only the Metazoa as a synonym of Animalia. In human culture The human population exploits a large number of other animal species for food, both of domesticated livestock species in animal husbandry and, mainly at sea, by hunting wild species. Marine fish of many species are caught commercially for food. A smaller number of species are farmed commercially. Humans and their livestock make up more than 90% of the biomass of all terrestrial vertebrates, and almost as much as all insects combined. Invertebrates including cephalopods, crustaceans, insects—principally bees and silkworms—and bivalve or gastropod molluscs are hunted or farmed for food, fibres. Chickens, cattle, sheep, pigs, and other animals are raised as livestock for meat across the world. Animal fibres such as wool and silk are used to make textiles, while animal sinews have been used as lashings and bindings, and leather is widely used to make shoes and other items. Animals have been hunted and farmed for their fur to make items such as coats and hats. Dyestuffs including carmine (cochineal), shellac, and kermes have been made from the bodies of insects. Working animals including cattle and horses have been used for work and transport from the first days of agriculture. Animals such as the fruit fly Drosophila melanogaster serve a major role in science as experimental models. Animals have been used to create vaccines since their discovery in the 18th century. Some medicines such as the cancer drug trabectedin are based on toxins or other molecules of animal origin. People have used hunting dogs to help chase down and retrieve animals, and birds of prey to catch birds and mammals, while tethered cormorants have been used to catch fish. Poison dart frogs have been used to poison the tips of blowpipe darts. A wide variety of animals are kept as pets, from invertebrates such as tarantulas, octopuses, and praying mantises, reptiles such as snakes and chameleons, and birds including canaries, parakeets, and parrots all finding a place. However, the most kept pet species are mammals, namely dogs, cats, and rabbits. There is a tension between the role of animals as companions to humans, and their existence as individuals with rights of their own. A wide variety of terrestrial and aquatic animals are hunted for sport. The signs of the Western and Chinese zodiacs are based on animals. In China and Japan, the butterfly has been seen as the personification of a person's soul, and in classical representation the butterfly is also the symbol of the soul. Animals have been the subjects of art from the earliest times, both historical, as in ancient Egypt, and prehistoric, as in the cave paintings at Lascaux. Major animal paintings include Albrecht Dürer's 1515 The Rhinoceros, and George Stubbs's c. 1762 horse portrait Whistlejacket. Insects, birds and mammals play roles in literature and film, such as in giant bug movies. Animals including insects and mammals feature in mythology and religion. The scarab beetle was sacred in ancient Egypt, and the cow is sacred in Hinduism. Among other mammals, deer, horses, lions, bats, bears, and wolves are the subjects of myths and worship. See also Notes References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/New_General_Catalogue] | [TOKENS: 1326] |
Contents New General Catalogue The New General Catalogue of Nebulae and Clusters of Stars (abbreviated NGC) is an astronomical catalogue of deep-sky objects compiled by John Louis Emil Dreyer in 1888. The NGC contains 7,840 objects, including galaxies, star clusters and emission nebulae. Dreyer published two supplements to the NGC in 1895 and 1908, known as the Index Catalogues (abbreviated IC), describing a further 5,386 astronomical objects. Thousands of these objects are best known by their NGC or IC numbers, which remain in widespread use. The NGC consolidated the cataloguing work of William and Caroline Herschel, and John Herschel's General Catalogue of Nebulae and Clusters of Stars. Objects south of the celestial equator are catalogued somewhat less thoroughly, but many were included based on observation by John Herschel or James Dunlop. The NGC contained multiple errors, but attempts to eliminate them were made by the Revised New General Catalogue (RNGC) by Jack W. Sulentic and William G. Tifft in 1973, NGC2000.0 by Roger W. Sinnott in 1988, and the NGC/IC Project in 1993. A Revised New General Catalogue and Index Catalogue (abbreviated as RNGC/IC) was compiled in 2009 by Wolfgang Steinicke and updated in 2019 with 13,957 objects. Original catalogue The original New General Catalogue was compiled during the 1880s by John Louis Emil Dreyer using observations from William Herschel and his son John, among others. Dreyer had already published a supplement to Herschel's General Catalogue of Nebulae and Clusters (GC), containing about 1,000 new objects. In 1886, he suggested building a second supplement to the General Catalogue, but the Royal Astronomical Society asked Dreyer to compile a new version instead. This led to the publication of the New General Catalogue in the Memoirs of the Royal Astronomical Society in 1888. Assembling the NGC was a challenge, as Dreyer had to deal with many contradictory and unclear reports made with a variety of telescopes with apertures ranging from 2 to 72 inches. While he did check some himself, the sheer number of objects meant Dreyer had to accept them as published by others for the purpose of his compilation. The catalogue contained several errors, mostly relating to position and descriptions, but Dreyer referenced the catalogue, which allowed later astronomers to review the original references and publish corrections to the original NGC. Index Catalogue The first major update to the NGC is the Index Catalogue of Nebulae and Clusters of Stars (abbreviated as IC), published in two parts by Dreyer in 1895 (IC I, containing 1,520 objects) and 1908 (IC II, containing 3,866 objects). It serves as a supplement to the NGC, and contains an additional 5,386 objects, collectively known as the IC objects. It summarizes the discoveries of galaxies, clusters and nebulae between 1888 and 1907, most of them made possible by photography. A list of corrections to the IC was published in 1912. Revised New General Catalogue The Revised New Catalogue of Nonstellar Astronomical Objects (abbreviated as RNGC) was compiled by Sulentic and Tifft in the early 1970s, and was published in 1973, as an update to the NGC. The work did not incorporate several previously published corrections to the NGC data (including corrections published by Dreyer himself), and introduced some new errors. For example, the well-known compact galaxy group Copeland Septet in the Leo constellation appears as non-existent in the RNGC. Nearly 800 objects are listed as "non-existent" in the RNGC. The designation is applied to objects which are duplicate catalogue entries, those which were not detected in subsequent observations, and a number of objects catalogued as star clusters which in subsequent studies were regarded as coincidental groupings. A 1993 monograph considered the 229 star clusters called non-existent in the RNGC. They had been "misidentified or have not been located since their discovery in the 18th and 19th centuries". It found that one of the 229—NGC 1498—was not actually in the sky. Five others were duplicates of other entries, 99 existed "in some form", and the other 124 required additional research to resolve. As another example, reflection nebula NGC 2163 in Orion was classified "non-existent" due to a transcription error by Dreyer. Dreyer corrected his own mistake in the Index Catalogues, but the RNGC preserved the original error, and additionally reversed the sign of the declination, resulting in NGC 2163 being classified as non-existent. The Revised New General Catalogue and Index Catalogue (abbreviated as RNGC/IC) is a compilation made by Wolfgang Steinicke in 2009. It is a comprehensive and authoritative treatment of the NGC and IC catalogues. The number of objects with status of "not found" in this catalogue is 301 objects (2.3%). The brightest star in this catalogue is NGC 771 with magnitude of 4.0. NGC 2000.0 NGC 2000.0 (also known as the Complete New General Catalog and Index Catalog of Nebulae and Star Clusters) is a 1988 compilation of the NGC and IC made by Roger W. Sinnott, using the J2000.0 coordinates. It incorporates several corrections and errata made by astronomers over the years. NGC/IC Project The NGC/IC Project was a collaboration among professional and amateur astronomers formed by Steve Gottlieb in 1990, although Steve Gottlieb already started to observe and record NGC objects as early as 1979. Other primary team members were Harold G. Corwin Jr., Malcolm Thomson, Robert E. Erdmann and Jeffrey Corder. The project was completed by 2017. This project identified all NGC and IC objects, corrected mistakes, collected images and basic astronomical data and checked all historical data related to the objects. See also References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/URL#cite_note-rfc2396-22] | [TOKENS: 957] |
Contents URL A uniform resource locator (URL), colloquially known as web address, is a reference to a resource on the World Wide Web. A URL specifies the location of a resource on a computer network and a mechanism for retrieving it. A URL is a specific type of Uniform Resource Identifier (URI), although many people use the two terms interchangeably.[a] A URL is most commonly used to reference a web page (HTTP/HTTPS) but is also used for file transfer (FTP), email (mailto), database access (JDBC), and many other applications. Most web browsers display the URL of a web page above the page in an address bar. As an example of a web page URL, https://www.example.com/index.html indicates protocol https, hostname www.example.com, and file name index.html. History The Uniform Resource Locator was defined in RFC 1738 in 1994 by Tim Berners-Lee, the inventor of the World Wide Web, and the URI working group of the Internet Engineering Task Force (IETF), as an outcome of collaboration started at the IETF Living Documents birds of a feather session in 1992. The format combines the pre-existing system of domain names (created in 1985) with file path syntax, where slashes are used to separate directory and filenames. Conventions already existed where server names could be prefixed to complete file paths, preceded by a double slash (//). Berners-Lee later expressed regret at the use of dots to separate the parts of the domain name within URIs, wishing he had used slashes throughout, and also said that, given the colon following the first component of a URI, the two slashes before the domain name were unnecessary. Early WorldWideWeb collaborators, including Berners-Lee, originally proposed the use of UDIs: Universal Document Identifiers. An early (1993) draft of the HTML Specification referred to "Universal" Resource Locators. This was dropped some time between June 1994 and October 1994. In his book Weaving the Web, Berners-Lee emphasizes his preference for the original inclusion of "universal" in the expansion rather than the word "uniform", to which it was later changed, and he gives a brief account of the contention that led to the change. Syntax Every HTTP URL conforms to the syntax of a generic URI. The URI generic syntax consists of five components organized hierarchically in order of decreasing significance from left to right:: §3 A component is undefined if it has an associated delimiter and the delimiter does not appear in the URI; the scheme and path components are always defined.: §5.2.1 A component is empty if it has no characters; the scheme component is always non-empty.: §3 The authority component consists of subcomponents: This is represented in a syntax diagram as: The URI comprises: A web browser will usually dereference a URL by performing an HTTP request to the specified host, by default on port number 80. URLs using the https scheme require that requests and responses be made over a secure connection to the website. Internationalized URL Internet users are distributed throughout the world using a wide variety of languages and alphabets, and expect to be able to create URLs in their own local alphabets. An Internationalized Resource Identifier (IRI) is a form of URL that includes Unicode characters. All modern browsers support IRIs. The parts of the URL requiring special treatment for different alphabets are the domain name and path. The domain name in the IRI is known as an Internationalized Domain Name (IDN). Web and Internet software automatically convert the domain name into punycode usable by the Domain Name System; for example, the Chinese URL http://例子.卷筒纸 becomes http://xn--fsqu00a.xn--3lr804guic/. The xn-- indicates that the character was not originally ASCII. The URL path name can also be specified by the user in the local writing system. If not already encoded, it is converted to UTF-8, and any characters not part of the basic URL character set are escaped as hexadecimal using percent-encoding; for example, the Japanese URL http://example.com/引き割り.html becomes http://example.com/%E5%BC%95%E3%81%8D%E5%89%B2%E3%82%8A.html. The target computer decodes the address and displays the page. Protocol-relative URLs Protocol-relative links (PRL), also known as protocol-relative URLs (PRURL), are URLs that have no protocol specified. For example, //example.com will use the protocol of the current page, typically HTTP or HTTPS. See also Notes Citations References External links |
======================================== |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.