text
stringlengths
10
951k
source
stringlengths
39
44
Demographics of the Cocos (Keeling) Islands This article is about the demographic features of the population of the Cocos (Keeling) Islands, including population density, ethnicity, education level, health of the populace, economic status, religious affiliations and other aspects of the population. The following demographic statistics are from the CIA World Factbook, unless otherwise indicated.
https://en.wikipedia.org/wiki?curid=5523
Conspiracy theory A conspiracy theory is an explanation for an event or situation that invokes a conspiracy by sinister and powerful groups, often political in motivation, when other explanations are more probable. The term has a pejorative connotation, implying that the appeal to a conspiracy is based on prejudice or insufficient evidence. Conspiracy theories resist falsification and are reinforced by circular reasoning: both evidence against the conspiracy and an absence of evidence for it are re-interpreted as evidence of its truth, whereby the conspiracy becomes a matter of faith rather than something that can be proved or disproved. Research suggests that conspiracist ideation—belief in conspiracy theories—can be psychologically harmful or pathological and that it is correlated with psychological projection, paranoia and Machiavellianism. Conspiracy theories once limited to fringe audiences have become commonplace in mass media, emerging as a cultural phenomenon of the late 20th and early 21st centuries. Psychologists attribute finding a conspiracy where there is none to a mental illness called "illusory pattern perception". The "Oxford English Dictionary" defines "conspiracy theory" as "the theory that an event or phenomenon occurs as a result of a conspiracy between interested parties; "spec." a belief that some covert but influential agency (typically political in motivation and oppressive in intent) is responsible for an unexplained event". It cites a 1909 article in "The American Historical Review" as the earliest usage example, although it also appears in journals as early as April 1870. The word "conspiracy" derives from the Latin "con-" ("with, together") and "spirare" ("to breathe"). Robert Blaskiewicz comments that examples of the term were used as early as the nineteenth century and states that its usage has always been derogatory. According to a study by Andrew McKenzie-McHarg, in contrast, in the nineteenth century the term "conspiracy theory" simply "suggests a plausible postulate of a conspiracy" and "did not, at this stage, carry any connotations, either negative or positive", though sometimes a postulate so-labeled was criticized. Lance deHaven-Smith suggested that the term entered everyday language in the United States after 1964, the year in which the Warren Commission shared its findings, with "The New York Times" running five stories that year using the term. A conspiracy theory is not simply a conspiracy. Barkun writes that conspiracies are "actual covert plots planned and/or carried out by two or more persons". A conspiracy theory, on the other hand, is "an intellectual construct", a "template imposed upon the world to give the appearance of order to events". Positing that "some small and hidden group" has manipulated events, a conspiracy theory can be local or international, focused on single events or covering multiple incidents and entire countries, regions and periods of history. Conspiracy theorists see themselves as having privileged access to special knowledge or a special mode of thought that separates them from the masses who believe the official account. A conspiracy theory may take any matter as its subject, but certain subjects attract greater interest than others. Favored subjects include famous deaths and assassinations, morally dubious government activities, suppressed technologies, and “false flag” terrorism. Among the longest-standing and most widely recognized conspiracy theories are notions concerning the assassination of John F. Kennedy, the 1969 Apollo moon landings, and the 9/11 terrorist attacks, as well as numerous theories pertaining to alleged plots for world domination by various groups both real and imaginary. Scholars argue that conspiracy theories once limited to fringe audiences have become commonplace in mass media, contributing to conspiracism emerging as a cultural phenomenon in the United States of the late 20th and early 21st centuries. The general predisposition to believe conspiracy theories cuts across partisan and ideological lines. Conspiratorial thinking is correlated with antigovernmental orientations and a low sense of political efficacy, with conspiracy believers perceiving a governmental threat to individual rights and displaying a deep skepticism that who one votes for really matters. According to anthropologists Todd Sanders and Harry G. West, a broad cross-section of Americans today gives credence to at least some conspiracy theories. For instance, a study conducted in 2016 found that 10% of Americans think the chemtrail conspiracy theory is "completely true" and 20-30% think it is "somewhat true". This puts "the equivalent of 120 million Americans in the 'chemtrails are real' camp." Belief in conspiracy theories has therefore become a topic of interest for sociologists, psychologists and experts in folklore. Conspiracy theories are widely present on the Web in the form of blogs and YouTube videos, as well as on social media. Whether the Web has increased the prevalence of conspiracy theories or not is an open research question. The presence and representation of conspiracy theories in search engine results has been monitored and studied, showing significant variation across different topics, and a general absence of reputable, high-quality links in the results. One conspiracy that propagated through former US President Barack Obama’s time in office was a theory that he was born in Kenya instead of Hawaii- where he has said he was born. Former governor of Arkansas and political opponent of Obama, Mike Huckabee made headlines in 2011 when he, among other members of Republican leadership, continued to question Obama's citizenship status. Jesse Walker (2013) has identified five kinds of conspiracy theories: Michael Barkun has identified three classifications of conspiracy theory: Murray Rothbard argues in favor of a model that contrasts "deep" conspiracy theories to "shallow" ones. According to Rothbard, a "shallow" theorist observes an event and asks "Cui bono"? ("Who benefits?"), jumping to the conclusion that a posited beneficiary is responsible for covertly influencing events. On the other hand, the "deep" conspiracy theorist begins with a hunch and then seeks out evidence. Rothbard describes this latter activity as a matter of confirming with certain facts one's initial paranoia. Belief in conspiracy theories is generally based not on evidence, but in the faith of the believer. Noam Chomsky contrasts conspiracy theory to institutional analysis which focuses mostly on the public, long-term behavior of publicly known institutions, as recorded in, for example, scholarly documents or mainstream media reports. Conspiracy theory conversely posits the existence of secretive coalitions of individuals and speculates on their alleged activities. Clare Birchall at King's College London describes conspiracy theory as a "form of popular knowledge or interpretation". The use of the word 'knowledge' here suggests ways in which conspiracy theory may be considered in relation to legitimate modes of knowing. The relationship between legitimate and illegitimate knowledge, Birchall claims, is closer than common dismissals of conspiracy theory contend. Theories involving multiple conspirators that are proven to be correct, such as the Watergate scandal, are usually referred to as "investigative journalism" or "historical analysis" rather than conspiracy theory. By contrast, the term "Watergate conspiracy theory" is used to refer to a variety of hypotheses in which those convicted in the conspiracy were in fact the victims of a deeper conspiracy. The historian Richard Hofstadter addressed the role of paranoia and conspiracism throughout U.S. history in his 1964 essay "The Paranoid Style in American Politics". Bernard Bailyn's classic "The Ideological Origins of the American Revolution" (1967) notes that a similar phenomenon could be found in North America during the time preceding the American Revolution. Conspiracism labels people's attitudes as well as the type of conspiracy theories that are more global and historical in proportion. The term "conspiracism" was further popularized by academic Frank P. Mintz in the 1980s. According to Mintz, conspiracism denotes "belief in the primacy of conspiracies in the unfolding of history": Justin Fox of "Time" magazine argues that Wall Street traders are among the most conspiracy-minded group of people, and ascribes this to the reality of some financial market conspiracies, and to the ability of conspiracy theories to provide necessary orientation in the market's day-to-day movements. Matthew Gray has noted that conspiracy theories are a prevalent feature of Arab culture and politics. Variants include conspiracies involving colonialism, Zionism, superpowers, oil, and the war on terrorism, which may be referred to as a war against Islam. For example, " The Protocols of the Elders of Zion", an infamous hoax document purporting to be a Jewish plan for world domination, is commonly read and promoted in the Muslim world. Roger Cohen has suggested that the popularity of conspiracy theories in the Arab world is "the ultimate refuge of the powerless". Al-Mumin Said has noted the danger of such theories, for they "keep us not only from the truth but also from confronting our faults and problems". Harry G. West and others have noted that while conspiracy theorists may often be dismissed as a fringe minority, certain evidence suggests that a wide range of the U.S. maintains a belief in conspiracy theories. West also compares those theories to hypernationalism and religious fundamentalism. Theologian Robert Jewett and philosopher John Shelton Lawrence attribute the enduring popularity of conspiracy theories in the U.S. to the Cold War, McCarthyism, and counterculture rejection of authority. They state that among both the left-wing and right-wing, there remains a willingness to use real events, such as Soviet plots, inconsistencies in the Warren Report, and the 9/11 attacks, to support the existence of unverified and ongoing large-scale conspiracies. The Watergate scandal has also been used to bestow legitimacy to other conspiracy theories, with Richard Nixon himself commenting that it served as a "Rorschach ink blot" which invited others to fill in the underlying pattern. Historian Kathryn S. Olmsted cites three reasons why Americans are prone to believing in government conspiracies theories: The widespread belief in conspiracy theories has become a topic of interest for sociologists, psychologists, and experts in folklore since at least the 1960s, when a number of conspiracy theories arose regarding the assassination of U.S. President John F. Kennedy. Sociologist Türkay Salim Nefes underlines the political nature of conspiracy theories. He suggests that one of the most important characteristics of these accounts is their attempt to unveil the "real but hidden" power relations in social groups. Research suggests, on a psychological level, conspiracist ideation—belief in conspiracy theories—can be harmful or pathological, and is highly correlated with psychological projection, as well as with paranoia, which is predicted by the degree of a person's Machiavellianism. The propensity to believe in conspiracy theories is strongly associated with the mental health disorder of schizotypy. Conspiracy theories once limited to fringe audiences have become commonplace in mass media, emerging as a cultural phenomenon of the late 20th and early 21st centuries. Some research has suggested that “analytical thinking” aids in reducing belief in conspiracy theories, in part because it emphasizes rational and critical cognition. Some psychological scientists assert that explanations related to conspiracy theories can be, and often are “internally consistent” with strong beliefs that had previously been held prior to the event that sparked the conspiracy. The political scientist Michael Barkun, discussing the usage of "conspiracy theory" in contemporary American culture, holds that this term is used for a belief that explains an event as the result of a secret plot by exceptionally powerful and cunning conspirators to achieve a malevolent end. According to Barkun, the appeal of conspiracism is threefold: This third point is supported by research of Roland Imhoff, professor in Social Psychology at the Johannes Gutenberg University Mainz. The research suggests that the smaller the minority believing in a specific theory, the more attractive it is to conspiracy theorists. Humanistic psychologists argue that even if a posited cabal behind an alleged conspiracy is almost always perceived as hostile, there often remains an element of reassurance for theorists. This is because it is a consolation to imagine that difficulties in human affairs are created by humans, and remain within human control. If a cabal can be implicated, there may be a hope of breaking its power or of joining it. Belief in the power of a cabal is an implicit assertion of human dignity—an unconscious affirmation that man is responsible for his own destiny. People formulate conspiracy theories to explain, for example, power relations in social groups and the perceived existence of evil forces. Proposed psychological origins of conspiracy theorising include projection; the personal need to explain "a significant event [with] a significant cause;" and the product of various kinds and stages of thought disorder, such as paranoid disposition, ranging in severity to diagnosable mental illnesses. Some people prefer socio-political explanations over the insecurity of encountering random, unpredictable, or otherwise inexplicable events. According to Berlet and Lyons, "Conspiracism is a particular narrative form of scapegoating that frames demonized enemies as part of a vast insidious plot against the common good, while it valorizes the scapegoater as a hero for sounding the alarm". Some psychologists believe that a search for meaning is common in conspiracism. Once cognized, confirmation bias and avoidance of cognitive dissonance may reinforce the belief. In a context where a conspiracy theory has become embedded within a social group, communal reinforcement may also play a part. Inquiry into possible motives behind the accepting of irrational conspiracy theories has linked these beliefs to distress resulting from an event that occurred, such as the events of 9/11. Additionally, research done by Manchester Metropolitan University suggests that “delusional ideation” is the most likely condition that would indicate an elevated belief in conspiracy theories. Studies also show that an increased attachment to these irrational beliefs lead to a decrease in desire for civic engagement Some historians have argued that there is an element of psychological projection in conspiracism. This projection, according to the argument, is manifested in the form of attribution of undesirable characteristics of the self to the conspirators. Historian Richard Hofstadter stated that: Hofstadter also noted that "sexual freedom" is a vice frequently attributed to the conspiracist's target group, noting that "very often the fantasies of true believers reveal strong sadomasochistic outlets, vividly expressed, for example, in the delight of anti-Masons with the cruelty of Masonic punishments." In addition to psychological factors such as conspiracist ideation, sociological factors also help account for who believes in which conspiracy theories. Such theories tend to get more traction among election losers in society, for example, and the emphasis of conspiracy theories by elites and leaders tends to increase belief among followers who have higher levels of conspiracy thinking. Christopher Hitchens described conspiracy theory as the "exhaust fumes of democracy": the unavoidable result of a large amount of information circulating among a large number of people. Conspiracy theories may be emotionally satisfying, by assigning blame to a group to which the theorist does not belong and so absolving the theorist of moral or political responsibility in society. Likewise, Roger Cohen writing for "The New York Times" has said that, "captive minds; ... resort to conspiracy theory because it is the ultimate refuge of the powerless. If you cannot change your own life, it must be that some greater force controls the world." Sociological historian Holger Herwig found in studying German explanations for the origins of World War I, "Those events that are most important are hardest to understand because they attract the greatest attention from myth makers and charlatans." French sociologist Bruno Latour suggests that the widespread popularity of conspiracy theories in mass culture may be due, in part, to the pervasive presence of Marxist-inspired critical theory and similar ideas in academia since the 1970s. Latour notes that about 90% of contemporary social criticism in academia displays one of two approaches, which he terms "the "fact position" and the "fairy position"". Latour concludes that each of these two approaches in academia has led to a polarized, inefficient atmosphere highlighted (in both approaches) by its causticness. "Do you see now why it feels so good to be a critical mind?" asks Latour: no matter which position you take, "You're always right!" Latour notes that such social criticism has been appropriated by those he describes as conspiracy theorists, including climate-change denialists and the 9/11 Truth movement: "Maybe I am taking conspiracy theories too seriously, but I am worried to detect, in those mad mixtures of knee-jerk disbelief, punctilious demands for proofs, and free use of powerful explanation from the social neverland, many of the weapons of social critique." Michael Kelly, a "Washington Post" journalist and critic of anti-war movements on both the left and right, coined the term "fusion paranoia" to refer to a political convergence of left-wing and right-wing activists around anti-war issues and civil liberties, which he said were motivated by a shared belief in conspiracism or shared anti-government views. Barkun has adopted this term to refer to how the synthesis of paranoid conspiracy theories, which were once limited to American fringe audiences, has given them mass appeal and enabled them to become commonplace in mass media, thereby inaugurating an unrivaled period of people actively preparing for apocalyptic or millenarian scenarios in the United States of the late 20th and early 21st centuries. Barkun notes the occurrence of lone-wolf conflicts with law enforcement acting as proxy for threatening the established political powers. The physicist David Robert Grimes estimated the time it would take for a conspiracy to be exposed based on the number of people involved. His calculations used data from the PRISM surveillance program, the Tuskegee syphilis experiment, and the FBI forensic scandal. Grimes estimated that: In his book "The Open Society and Its Enemies", the philosopher Karl Popper used the term "the conspiracy theory of society" to denote a conception of social phenomena that he found to be defective—namely, that social phenomena such as "war, unemployment, poverty, shortages ... [are] the result of direct design by some powerful individuals and groups." Popper argued that totalitarianism was founded on "conspiracy theories" which drew on imaginary plots which were driven by paranoid scenarios predicated on tribalism, chauvinism, or racism. Popper acknowledged that genuine conspiracies do exist, but noted how infrequently conspirators have been able to achieve their goal. The historian Bruce Cumings similarly rejects the notion that history is controlled by conspiracies, stating that where real conspiracies have appeared they have usually had little effect on history and have had unforeseen consequences for the conspirators. Cumings concludes that history is instead "moved by the broad forces and large structures of human collectivities". In a 2009 article, the legal scholars Cass Sunstein and Adrian Vermeule considered a number of possible government responses to conspiracy theories, including censorship and taxation. They concluded that the authorities ought to engage in counter-speech and dialogue, which they termed "cognitive infiltration". Alex Jones referenced numerous conspiracy theories for convincing his supporters to endorse Ron Paul over Mitt Romney and Donald Trump over Hillary Clinton.
https://en.wikipedia.org/wiki?curid=5530
Coral Sea Islands The Coral Sea Islands Territory is an external territory of Australia which comprises a group of small and mostly uninhabited tropical islands and reefs in the Coral Sea, northeast of Queensland, Australia. The only inhabited island is Willis Island. The territory covers , most of which is ocean, extending east and south from the outer edge of the Great Barrier Reef and includes Heralds Beacon Island, Osprey Reef, the Willis Group and fifteen other reef/island groups. Cato Island is the highest point in the Territory. The Coral Sea Islands were first charted in 1803. In the 1870s and 1880s the islands were mined for guano but the absence of a reliable supply of fresh water prevented long-term habitation. The Coral Sea Islands became an Australian external territory in 1969 by the "Coral Sea Islands Act" (prior to that, the area was considered a part of Queensland) and extended in 1997 to include Elizabeth Reef and Middleton Reef nearly 800 km further south. The two latter reefs are much closer to Lord Howe Island, New South Wales, (about ) than to the southernmost island of the rest of the territory, Cato Island. The islands, cays and reefs of the Great Barrier Reef are not part of the territory, belonging to Queensland instead. The outer edge of the Great Barrier Reef is the boundary between Queensland and the Coral Sea Islands Territory. The territory is a possession or external territory of Australia, administered from Canberra by the Department of Infrastructure, Regional Development and Cities. Previously it was administered by the Attorney-General's Department and the Department of Transport and Regional Services). Defence is the responsibility of Australia, and the territory is visited regularly by the Royal Australian Navy. Australia maintains automatic weather stations on many of the isles and reefs, and claims a exclusive fishing zone. There is no economic activity (except for a significant but as yet unquantified charter fishing and diving industry), and only a staff of three or four people to run the meteorological station on Willis Island (South Islet), established in 1921. In November 2011, the Australian government announced that a protected area was planned in the Coral Sea. The Supreme Court of Norfolk Island has jurisdiction over the islands, however, the laws of the Australian Capital Territory apply. The territory's FIPS 10-4 code is CR, whereas ISO 3166 includes it in Australia (AU). In 2004, the Gay and Lesbian Kingdom of the Coral Sea Islands was declared as a sovereign state as a symbolic political protest by a group of gay rights activists based in Australia in response to the Australian government's refusal to recognise same-sex marriages. The Kingdom was dissolved on 17 November 2017 following results of the Australian Marriage Law Postal Survey indicated a majority of Australians supported same-sex marriage. There are about 30 separate reefs and atolls, twelve being wholly submerged or drying only during low tide, and 18 others with a total of about 51 islets and cays (18 alone on the atoll Lihou Reef), some of which are vegetated. The atolls exhibit a wide range of size, from a few kilometres in diameter to perhaps the second largest atoll in the world by total area (including lagoon): Lihou Reef, with a lagoon size of and an area of , which compares to a combined land area of the 18 individual islets of only . The islands are all very low. The Willis Islets are important nesting areas for birds and turtles but contain negligible natural resources. They comprise less than of land. There is no port or harbour, only offshore anchorage. Most of the atolls fall into two groups, while Mellish Reef to the east, and Middleton Reef and Elizabeth Reef to the south are grouped separately: The atolls of the Northwestern Group, except Osprey Reef and Shark Reef in the north, and Marion Reef in the south, are located on the Coral Sea Plateau (Queensland Plateau), a contiguous area of depths less than 1000 m. The Nature Reserves were created to protect wildlife in the respective areas of the territory; together they form the Coral Sea Reserves Ramsar Site. Elizabeth and Middleton Reefs, together with reefs around Lord Howe Island (New South Wales) 150 km to the south, are regarded as the southernmost coral reefs in the world. Their location, where tropical and temperate ocean currents meet, contributes to an unusually diverse assemblage of marine species. These mostly submerged atolls which dry only during low tide were added to the territory only in 1989. They are located on the Lord Howe Rise. Already on 23 December 1987, they were protected as the Elizabeth and Middleton Reefs Marine National Nature Reserve, which has an area of 1880 km2. Automatic, unmanned weather stations are located on the following reefs or atolls: Lighthouses are located on following reefs or islands: Willis Island, the only inhabited island, has a number of structures.
https://en.wikipedia.org/wiki?curid=5541
Costa Rica Costa Rica (, ; ; literally "Rich Coast"), officially the Republic of Costa Rica (), is a country in Central America, bordered by Nicaragua to the north, the Caribbean Sea to the northeast, Panama to the southeast, the Pacific Ocean to the southwest, and Ecuador to the south of Cocos Island. It has a population of around 5 million in a land area of 51,060 square kilometers (19,714 square miles). An estimated 333,980 people live in the capital and largest city, San José, with around 2 million people in the surrounding metropolitan area. The sovereign state is a unitary presidential constitutional republic. It is known for its long-standing and stable democracy, and for its highly educated workforce, most of whom speak English. The country spends roughly 6.9% of its budget (2016) on education, compared to a global average of 4.4%. Its economy, once heavily dependent on agriculture, has diversified to include sectors such as finance, corporate services for foreign companies, pharmaceuticals, and ecotourism. Many foreign manufacturing and services companies operate in Costa Rica's Free Trade Zones (FTZ) where they benefit from investment and tax incentives. Costa Rica was sparsely inhabited by indigenous peoples before coming under Spanish rule in the 16th century. It remained a peripheral colony of the empire until independence as part of the First Mexican Empire, followed by membership in the United Provinces of Central America, from which it formally declared independence in 1847. Following the brief Costa Rican Civil War in 1948, it permanently abolished its army in 1949, becoming one of only a few sovereign nations without a standing army. The country has consistently performed favorably in the Human Development Index (HDI), placing 68th in the world , and fifth in Latin America. It has also been cited by the United Nations Development Programme (UNDP) as having attained much higher human development than other countries at the same income levels, with a better record on human development and inequality than the median of the region. Costa Rica also has progressive environmental policies. It is the only country to meet all five UNDP criteria established to measure environmental sustainability. It was ranked 42nd in the world, and third in the Americas, in the 2016 Environmental Performance Index, and was twice ranked the best performing country in the New Economics Foundation's (NEF) Happy Planet Index, which measures environmental sustainability, and was identified by the NEF as the greenest country in the world in 2009. Costa Rica plans to become a carbon-neutral country by 2021. By 2019, 99.62% of its electricity was generated from green sources particularly hydro, wind, geothermal and solar. Historians have classified the indigenous people of Costa Rica as belonging to the Intermediate Area, where the peripheries of the Mesoamerican and Andean native cultures overlapped. More recently, pre-Columbian Costa Rica has also been described as part of the Isthmo-Colombian Area. Stone tools, the oldest evidence of human occupation in Costa Rica, are associated with the arrival of various groups of hunter-gatherers about 10,000 to 7,000 years BCE in the Turrialba Valley. The presence of Clovis culture type spearheads and arrows from South America opens the possibility that, in this area, two different cultures coexisted. Agriculture became evident in the populations that lived in Costa Rica about 5,000 years ago. They mainly grew tubers and roots. For the first and second millennia BCE there were already settled farming communities. These were small and scattered, although the timing of the transition from hunting and gathering to agriculture as the main livelihood in the territory is still unknown. The earliest use of pottery appears around 2,000 to 3,000 BCE. Shards of pots, cylindrical vases, platters, gourds and other forms of vases decorated with grooves, prints, and some modelled after animals have been found. The impact of indigenous peoples on modern Costa Rican culture has been relatively small compared to other nations, since the country lacked a strong native civilization to begin with. Most of the native population was absorbed into the Spanish-speaking colonial society through inter-marriage, except for some small remnants, the most significant of which are the Bribri and Boruca tribes who still inhabit the mountains of the Cordillera de Talamanca, in the southeastern part of Costa Rica, near the frontier with Panama. The name , meaning "rich coast" in the Spanish language, was in some accounts first applied by Christopher Columbus, who sailed to the eastern shores of Costa Rica during his final voyage in 1502, and reported vast quantities of gold jewelry worn by natives. The name may also have come from conquistador Gil González Dávila, who landed on the west coast in 1522, encountered natives, and obtained some of their gold, sometimes by violent theft and sometimes as gifts from local leaders. During most of the colonial period, Costa Rica was the southernmost province of the Captaincy General of Guatemala, nominally part of the Viceroyalty of New Spain. In practice, the captaincy general was a largely autonomous entity within the Spanish Empire. Costa Rica's distance from the capital of the captaincy in Guatemala, its legal prohibition under Spanish law from trade with its southern neighbor Panama, then part of the Viceroyalty of New Granada (i.e. Colombia), and lack of resources such as gold and silver, made Costa Rica into a poor, isolated, and sparsely-inhabited region within the Spanish Empire. Costa Rica was described as "the poorest and most miserable Spanish colony in all America" by a Spanish governor in 1719. Another important factor behind Costa Rica's poverty was the lack of a significant indigenous population available for (forced labor), which meant most of the Costa Rican settlers had to work on their own land, preventing the establishment of large (plantations). For all these reasons, Costa Rica was, by and large, unappreciated and overlooked by the Spanish Crown and left to develop on its own. The circumstances during this period are believed to have led to many of the idiosyncrasies for which Costa Rica has become known, while concomitantly setting the stage for Costa Rica's development as a more egalitarian society than the rest of its neighbors. Costa Rica became a "rural democracy" with no oppressed mestizo or indigenous class. It was not long before Spanish settlers turned to the hills, where they found rich volcanic soil and a milder climate than that of the lowlands. Like the rest of Central America, Costa Rica never fought for independence from Spain. On 15 September 1821, after the final Spanish defeat in the Mexican War of Independence (1810–21), the authorities in Guatemala declared the independence of all of Central America. That date is still celebrated as Independence Day in Costa Rica even though, technically, under the Spanish Constitution of 1812 that had been readopted in 1820, Nicaragua and Costa Rica had become an autonomous province with its capital in León. Upon independence, Costa Rican authorities faced the issue of officially deciding the future of the country. Two bands formed, the Imperialists, defended by Cartago and Heredia cities which were in favor of joining the Mexican Empire, and the Republicans, represented by the cities of San José and Alajuela who defended full independence. Because of the lack of agreement on these two possible outcomes, the first civil war of Costa Rica occurred. The Battle of Ochomogo took place on the Hill of Ochomogo, located in the Central Valley in 1823. The conflict was won by the Republicans and, as a consequence, the city of Cartago lost its status as the capital, which moved to San José. In 1838, long after the Federal Republic of Central America ceased to function in practice, Costa Rica formally withdrew and proclaimed itself sovereign. The considerable distance and poor communication routes between Guatemala City and the Central Plateau, where most of the Costa Rican population lived then and still lives now, meant the local population had little allegiance to the federal government in Guatemala. From colonial times to now, Costa Rica's reluctance to become economically tied with the rest of Central America has been a major obstacle to efforts for greater regional integration. Until 1849, when it became part of Panama, Chiriquí was part of Costa Rica. Costa Rican pride was assuaged for the loss of this eastern (or southern) territory with the acquisition of Guanacaste, in the north. Coffee was first planted in Costa Rica in 1808, and by the 1820s, it surpassed tobacco, sugar, and cacao as a primary export. Coffee production remained Costa Rica's principal source of wealth well into the 20th century, creating a wealthy class of growers, the so-called Coffee Barons. The revenue helped to modernize the country. Most of the coffee exported was grown around the main centers of population in the Central Plateau and then transported by oxcart to the Pacific port of Puntarenas after the main road was built in 1846. By the mid-1850s the main market for coffee was Britain. It soon became a high priority to develop an effective transportation route from the Central Plateau to the Atlantic Ocean. For this purpose, in the 1870s, the Costa Rican government contracted with U.S. businessman Minor C. Keith to build a railroad from San José to the Caribbean port of Limón. Despite enormous difficulties with construction, disease, and financing, the railroad was completed in 1890. Most Afro-Costa Ricans descend from Jamaican immigrants who worked in the construction of that railway and now make up about 3% of Costa Rica's population. U.S. convicts, Italians and Chinese immigrants also participated in the construction project. In exchange for completing the railroad, the Costa Rican government granted Keith large tracts of land and a lease on the train route, which he used to produce bananas and export them to the United States. As a result, bananas came to rival coffee as the principal Costa Rican export, while foreign-owned corporations (including the United Fruit Company later) began to hold a major role in the national economy and eventually became a symbol of the exploitative export economy. The major labor dispute between the peasants and the United Fruit Company (The Great Banana Strike) was a major event in the country's history and was an important step that would eventually lead to the formation of effective trade unions in Costa Rica, as the company was required to sign a collective agreement with its workers in 1938. Historically, Costa Rica has generally enjoyed greater peace and more consistent political stability than many of its fellow Latin American nations. Since the late 19th century, however, Costa Rica has experienced two significant periods of violence. In 1917–19, General Federico Tinoco Granados ruled as a military dictator until he was overthrown and forced into exile. The unpopularity of Tinoco's regime led, after he was overthrown, to a considerable decline in the size, wealth, and political influence of the Costa Rican military. In 1948, José Figueres Ferrer led an armed uprising in the wake of a disputed presidential election between Rafael Ángel Calderón Guardia (who had been president between 1940 and 1944) and Otilio Ulate Blanco. With more than 2,000 dead, the resulting 44-day Costa Rican Civil War was the bloodiest event in Costa Rica during the 20th century. The victorious rebels formed a government junta that abolished the military altogether, and oversaw the drafting of a new constitution by a democratically elected assembly. Having enacted these reforms, the junta transferred power to Ulate on 8 November 1949. After the "coup d'état", Figueres became a national hero, winning the country's first democratic election under the new constitution in 1953. Since then, Costa Rica has held 14 presidential elections, the latest in 2018. With uninterrupted democracy dating back to at least 1948, the country is the region's most stable. Costa Rica borders the Caribbean Sea to the east, and the Pacific Ocean to the west. Costa Rica also borders Nicaragua to the north and Panama to the south. The highest point in the country is Cerro Chirripó, at . The highest volcano in the country is the Irazú Volcano () and the largest lake is Lake Arenal. There are 14 known volcanoes in Costa Rica, and six of them have been active in the last 75 years. Costa Rica experiences a tropical climate year round. There are two seasons. The "summer" or dry season is December to April, and "winter" or rainy season is May to November. There is a rich variety of plants and Costa Rican wildlife. One national park, the Corcovado National Park, is internationally renowned among ecologists for its biodiversity (including big cats and tapirs) and is where visitors can expect to see an abundance of wildlife. Corcovado is the one park in Costa Rica where all four Costa Rican monkey species can be found. These include the white-headed capuchin, the mantled howler, the endangered Geoffroy's spider monkey, and the Central American squirrel monkey, found only on the Pacific coast of Costa Rica and a small part of Panama, and considered endangered until 2008, when its status was upgraded to vulnerable. Deforestation, illegal pet-trading, and hunting are the main reasons for its threatened status. The country has been considered economically stable with moderate inflation, estimated at 2.6% in 2017, and moderately high growth in GDP, which increased from US$41.3 billion in 2011 to US$52.6 billion in 2015. The estimated GDP for 2017 is US$61.5 billion and the estimated GDP per capita (purchasing power parity) is US$12,382. The growing debt and budget deficit are the country's primary concerns. A 2017 study by the Organisation for Economic Co-operation and Development warned that reducing the foreign debt must be a very high priority for the government. Other fiscal reforms were also recommended to moderate the budget deficit. Many foreign companies (manufacturing and services) operate in Costa Rica's Free Trade Zones (FTZ) where they benefit from investment and tax incentives. Well over half of that type of investment has come from the U.S. According to the government, the zones supported over 82,000 direct jobs and 43,000 indirect jobs in 2015. Companies with facilities in the America Free Zone in Heredia, for example, include Intel, Dell, HP, Bayer, Bosch, DHL, IBM and Okay Industries. Of the GDP, 5.5% is generated by agriculture, 18.6% by industry and 75.9% by services. (2016) Agriculture employs 12.9% of the labor force, industry 18.57%, services 69.02% (2016) For the region, its unemployment level is moderately high (8.2% in 2016, according to the IMF). Although 20.5% of the population lives below the poverty line (2017), Costa Rica has one of the highest standards of living in Central America. High quality health care is provided by the government at low cost to the users. Housing is also very affordable. Costa Rica is recognized in Latin America for the quality of its educational system. Because of its educational system, Costa Rica has one of the highest literacy rates in Latin America, 97%. General Basic Education is mandatory and provided without cost to the user. A US government report confirms that the country has "historically placed a high priority on education and the creation of a skilled work force" but notes that the high school drop-out rate is increasing. As well, Costa Rica would benefit from more courses in languages such as English, Portuguese, Mandarin and French and also in Science, Technology, Engineering and Math (STEM). Costa Rica has free trade agreements with many countries, including the US. There are no significant trade barriers that would affect imports and the country has been lowering its tariffs in accordance with other Central American countries. The country's Free Trade Zones provide incentives for manufacturing and service industries to operate in Costa Rica. In 2015, the zones supported over 82 thousand direct jobs and 43 thousand indirect jobs in 2015 and average wages in the FTZ were 1.8 times greater than the average for private enterprise work in the rest of the country. In 2016, Amazon.com for example, had some 3,500 employees in Costa Rica and planned to increase that by 1,500 in 2017, making it an important employer. The central location provides access to American markets and direct ocean access to Europe and Asia. The most important exports in 2015 (in order of dollar value) were medical instruments, bananas, tropical fruits, integrated circuits and orthopedic appliances. Total imports in that year were US$15 billion. The most significant products imported in 2015 (in order of dollar value) were refined petroleum, automobiles, packaged medications, broadcasting equipment and computers. The total exports were US$12.6 billion for a trade deficit of US$2.39 billion in 2015. Pharmaceuticals, financial outsourcing, software development, and ecotourism have become the prime industries in Costa Rica's economy. High levels of education among its residents make the country an attractive investing location. Since 1999, tourism earns more foreign exchange than the combined exports of the country's three main cash crops: bananas and pineapples especially, but also other crops, including coffee. Coffee production played a key role in Costa Rica's history and in 2006, was the third cash crop export. As a small country, Costa Rica now provides under 1% of the world's coffee production. In 2015, the value of coffee exports was US$305.9 million, a small part of the total agricultural exports of US$2.7 billion. Coffee production increased by 13.7% percent in 2015–16, declined by 17.5% in 2016–17, but was expected to increase by about 15% in the subsequent year. Costa Rica has developed a system of payments for environmental services. Similarly, Costa Rica has a tax on water pollution to penalize businesses and homeowners that dump sewage, agricultural chemicals, and other pollutants into waterways. In May 2007, the Costa Rican government announced its intentions to become 100% carbon neutral by 2021. By 2015, 93 percent of the country's electricity came from renewable sources. In 2019, the country produced 99.62% of its electricity from renewable sources and ran completely on renewable sources for 300 continuous days. In 1996, the Forest Law was enacted to provide direct financial incentives to landowners for the provision of environmental services. This helped reorient the forestry sector away from commercial timber production and the resulting deforestation, and helped create awareness of the services it provides for the economy and society (i.e., carbon fixation, hydrological services such as producing fresh drinking water, biodiversity protection, and provision of scenic beauty). A 2016 report by the U.S. government report identifies other challenges facing Costa Rica as it works to expand its economy by working with companies from the US (and probably from other countries). The major concerns identified were as follows: Costa Rica is the most-visited nation in the Central American region, with 2.9 million foreign visitors in 2016, up 10% from 2015. In 2015, the tourism sector was responsible for 5.8% of the country's GDP, or $3.4 billion. In 2016, the highest number of tourists came from the United States, with 1,000,000 visitors, followed by Europe with 434,884 arrivals. According to Costa Rica Vacations, once tourists arrive in the country, 22% go to Tamarindo, 18% go to Arenal, 17% pass through Liberia (where the Daniel Oduber Quirós International Airport is located), 16% go to San José, the country's capital (passing through Juan Santamaría International Airport), while 18% choose Manuel Antonio and 7% Monteverde. By 2004, tourism was generating more revenue and foreign exchange than bananas and coffee combined. In 2016, the World Travel & Tourism Council's estimates indicated a direct contribution to the GDP of 5.1% and 110,000 direct jobs in Costa Rica; the total number of jobs indirectly supported by tourism was 271,000. A pioneer of ecotourism, Costa Rica draws many tourists to its extensive series of national parks and other protected areas. In the 2011 Travel and Tourism Competitiveness Index, Costa Rica ranked 44th in the world and second among Latin American countries after Mexico in 2011. By the time of the 2017 report, the country had reached 38th place, slightly behind Panama. The Ethical Traveler group's ten countries on their 2017 list of The World's Ten Best Ethical Destinations includes Costa Rica. The country scored highest in environmental protection among the winners. Costa Rica began reversing deforestation in the 1990s, and they are moving towards using only renewable energy. Costa Rica is composed of seven provinces, which in turn are divided into 81 cantons (, plural ), each of which is directed by a mayor. Mayors are chosen democratically every four years by each canton. There are no provincial legislatures. The cantons are further divided into 473 districts (). Costa Rica is an active member of the United Nations and the Organization of American States. The Inter-American Court of Human Rights and the United Nations University of Peace are based in Costa Rica. It is also a member of many other international organizations related to human rights and democracy, such as the Community of Democracies. A main foreign policy objective of Costa Rica is to foster human rights and sustainable development as a way to secure stability and growth. Costa Rica is a member of the International Criminal Court, without a Bilateral Immunity Agreement of protection for the United States military (as covered under Article 98). Costa Rica is an observer of the Organisation internationale de la Francophonie. On 10 September 1961, some months after Fidel Castro declared Cuba a socialist state, Costa Rican President Mario Echandi ended diplomatic relations with Cuba through "Executive Decree Number 2". This freeze lasted 47 years until President Óscar Arias Sánchez re-established normal relations on 18 March 2009, saying, "If we have been able to turn the page with regimes as profoundly different to our reality as occurred with the USSR or, more recently, with the Republic of China, how would we not do it with a country that is geographically and culturally much nearer to Costa Rica?" Arias announced that both countries would exchange ambassadors. Costa Rica has a long-term disagreement with Nicaragua over the San Juan River, which defines the border between the two countries, and Costa Rica's rights of navigation on the river. In 2010, there was also a dispute around Isla Calero, and the impact of Nicaraguan dredging of the river in that area. On 14 July 2009, the International Court of Justice in the Hague upheld Costa Rica's navigation rights for commercial purposes to subsistence fishing on their side of the river. An 1858 treaty extended navigation rights to Costa Rica, but Nicaragua denied passenger travel and fishing were part of the deal; the court ruled Costa Ricans on the river were not required to have Nicaraguan tourist cards or visas as Nicaragua argued, but, in a nod to the Nicaraguans, ruled that Costa Rican boats and passengers must stop at the first and last Nicaraguan port along their route. They must also have an identity document or passport. Nicaragua can also impose timetables on Costa Rican traffic. Nicaragua may require Costa Rican boats to display the flag of Nicaragua, but may not charge them for departure clearance from its ports. These were all specific items of contention brought to the court in the 2005 filing. On 1 June 2007, Costa Rica broke diplomatic ties with Taiwan, switching recognition to the People's Republic of China. Costa Rica was the first of the Central American nations to do so. President Óscar Arias Sánchez admitted the action was a response to economic exigency. In response, the PRC built a new, $100 million, state-of-the-art football stadium in Parque la Sabana, in the province of San José. Approximately 600 Chinese engineers and laborers took part in this project, and it was inaugurated in March 2011, with a match between the national teams of Costa Rica and China. Costa Rica finished a term on the United Nations Security Council, having been elected for a nonrenewable, two-year term in the 2007 election. Its term expired on 31 December 2009; this was Costa Rica's third time on the Security Council. Elayne Whyte Gómez is the Permanent Representative of Costa Rica to the UN Office at Geneva (2017) and President of the United Nations Conference to Negotiate a Legally Binding Instrument to Prohibit Nuclear Weapons. On December 1, 1948, Costa Rica abolished its military force. In 1949, the abolition of the military was introduced in Article 12 of the Costa Rican Constitution. The budget previously dedicated to the military is now dedicated to providing health care services and education. According to DW, "Costa Rica is known for its stable democracy, progressive social policies, such as free, compulsory public education, high social well-being, and emphasis on environmental protection." In 2017, Costa Rica signed the UN treaty on the Prohibition of Nuclear Weapons. The 2011 census counted a population of 4.3 million people distributed among the following groups: 83.6% whites or mestizos, 6.7% mulattoes, 2.4% Native American, 1.1% black or Afro-Caribbean; the census showed 1.1% as Other, 2.9% (141,304 people) as None, and 2.2% (107,196 people) as unspecified. By 2016, the UN estimation for the population was around /1e6 round 1 million. In 2011, there were over 104,000 Native American or indigenous inhabitants, representing 2.4% of the population. Most of them live in secluded reservations, distributed among eight ethnic groups: Quitirrisí (in the Central Valley), Matambú or Chorotega (Guanacaste), Maleku (northern Alajuela), Bribri (southern Atlantic), Cabécar (Cordillera de Talamanca), Guaymí (southern Costa Rica, along the Panamá border), Boruca (southern Costa Rica) and (southern Costa Rica). The population includes European Costa Ricans (of European ancestry), primarily of Spanish descent, with significant numbers of Italian, German, English, Dutch, French, Irish, Portuguese, and Polish families, as well a sizable Jewish community. The majority of the Afro-Costa Ricans are Creole English-speaking descendants of 19th century black Jamaican immigrant workers. The 2011 census classified 83.6% of the population as white or Mestizo; the latter are persons of combined European and Amerindian descent. The Mulatto segment (mix of white and black) represented 6.7% and indigenous people made up 2.4% of the population. Native and European mixed blood populations are far less than in other Latin American countries. Exceptions are Guanacaste, where almost half the population is visibly mestizo, a legacy of the more pervasive unions between Spanish colonists and Chorotega Amerindians through several generations, and Limón, where the vast majority of the Afro-Costa Rican community lives. Costa Rica hosts many refugees, mainly from Colombia and Nicaragua. As a result of that and illegal immigration, an estimated 10–15% (400,000–600,000) of the Costa Rican population is made up of Nicaraguans. Some Nicaraguans migrate for seasonal work opportunities and then return to their country. Costa Rica took in many refugees from a range of other Latin American countries fleeing civil wars and dictatorships during the 1970s and 1980s, notably from Chile and Argentina, as well as people from El Salvador who fled from guerrillas and government death squads. According to the World Bank, in 2010 about 489,200 immigrants lived in the country, many from Nicaragua, Panama, El Salvador, Honduras, Guatemala, and Belize, while 125,306 Costa Ricans live abroad in the United States, Panama, Nicaragua, Spain, Mexico, Canada, Germany, Venezuela, Dominican Republic, and Ecuador. The number of migrants declined in later years but in 2015, there were some 420,000 immigrants in Costa Rica and the number of asylum seekers (mostly from Honduras, El Salvador, Guatemala and Nicaragua) rose to more than 110,000, a fivefold increase from 2012. In 2016, the country was called a "magnet" for migrants from South and Central America and other countries who were hoping to reach the U.S. Christianity is Costa Rica's predominant religion, with Roman Catholicism being the official state religion according to the 1949 Constitution, which at the same time guarantees freedom of religion. It is the only state in the Americas which established Roman Catholicism as its state religion; other such countries are microstates in Europe: Liechtenstein, Monaco, the Vatican City and Malta. The Latinobarómetro survey of 2017 found that 57% of the population identify themselves as Roman Catholics, 25% are Evangelical Protestants, 15% report that they do not have a religion, and 2% declare that they belong to another religion. This survey indicated a decline in the share of Catholics and rise in the share of Protestants and irreligious. A University of Costa Rica survey of 2018 show similar rates; 52% Catholics, 22% Protestants, 17% irreligious and 3% other. The rate of secularism is high by Latin American standards. Due to small, but continuous, immigration from Asia and the Middle East, other religions have grown, the most popular being Buddhism, with about 100,000 practitioners (over 2% of the population). Most Buddhists are members of the Han Chinese community of about 40,000 with some new local converts. There is also a small Muslim community of about 500 families, or 0.001% of the population. The Sinagoga Shaarei Zion synagogue is near La Sabana Metropolitan Park in San José. Several homes in the neighborhood east of the park display the Star of David and other Jewish symbols. The Church of Jesus Christ of Latter-day Saints claims more than 35,000 members, and has a temple in San José that served as a regional worship center for Costa Rica. However, they represent less than 1% of the population. The primary language spoken in Costa Rica is Spanish, which features characteristics distinct to the country, a form of Central American Spanish. Costa Rica is a linguistically diverse country and home to at least five living local indigenous languages spoken by the descendants of pre-Columbian peoples: Maléku, Cabécar, Bribri, Guaymí, and Buglere. Of native languages still spoken, primarily in indigenous reservations, the most numerically important are the Bribri, Maléku, Cabécar and Ngäbere languages; some of these have several thousand speakers in Costa Rica while others have a few hundred. Some languages, such as Teribe and Boruca, have fewer than a thousand speakers. The Buglere language and the closely related Guaymí are spoken by some in southeast Puntarenas. A Creole-English language, Jamaican "patois" (also known as Mekatelyu), is an English-based Creole language spoken by the Afro-Carib immigrants who have settled primarily in Limón Province along the Caribbean coast. About 10.7% of Costa Rica's adult population (18 or older) also speaks English, 0.7% French, and 0.3% speaks Portuguese or German as a second language. Costa Rica was the point where the Mesoamerican and South American native cultures met. The northwest of the country, the Nicoya peninsula, was the southernmost point of Nahuatl cultural influence when the Spanish conquerors (conquistadores) came in the 16th century. The central and southern portions of the country had Chibcha influences. The Atlantic coast, meanwhile, was populated with African workers during the 17th and 18th centuries. As a result of the immigration of Spaniards, their 16th-century Spanish culture and its evolution marked everyday life and culture until today, with Spanish language and the Catholic religion as primary influences. The Department of Culture, Youth, and Sports is in charge of the promotion and coordination of cultural life. The work of the department is divided into Direction of Culture, Visual Arts, Scenic Arts, Music, Patrimony and the System of Libraries. Permanent programs, such as the National Symphony Orchestra of Costa Rica and the Youth Symphony Orchestra, are conjunctions of two areas of work: Culture and Youth. Dance-oriented genres, such as "soca", "salsa", "bachata", "merengue", "cumbia" and Costa Rican swing are enjoyed increasingly by older rather than younger people. The guitar is popular, especially as an accompaniment to folk dances; however, the marimba was made the national instrument. In November 2017, "National Geographic" magazine named Costa Rica as the happiest country in the world. The article included this summary: "Costa Ricans enjoy the pleasure of living daily life to the fullest in a place that mitigates stress and maximizes joy". It is not surprising then that one of the most recognizable phrases among "Ticos" is ""Pura Vida"", pure life in a literal translation. It reflects the inhabitant's philosophy of life, denoting a simple life, free of stress, a positive, relaxed feeling. The expression is used in various contexts in conversation. Often, people walking down the streets, or buying food at shops say hello by saying "Pura Vida". It can be phrased as a question or as an acknowledgement of one's presence. A recommended response to "How are you?" would be ""Pura Vida"." In that usage, it might be translated as "awesome", indicating that all is very well. When used as a question, the connotation would be "everything is going well?" or "how are you?". Costa Rica rates 12th on the 2017 Happy Planet Index in the World Happiness Report by the UN but the country is said to be the happiest in Latin America. Reasons include the high level of social services, the caring nature of its inhabitants, long life expectancy and relatively low corruption. Costa Rican cuisine is a blend of Native American, Spanish, African and many other cuisine origins. Dishes such as the very traditional tamale and many others made of corn are the most representative of its indigenous inhabitants, and similar to other neighboring Mesoamerican countries. Spaniards brought many new ingredients to the country from other lands, especially spices and domestic animals. And later in the 19th century, the African flavor lent its presence with influence from other Caribbean mixed flavors. This is how Costa Rican cuisine today is very varied, with every new ethnic group who had recently become part of the country's population influencing the country's cuisine. Costa Rica entered the Summer Olympics for the first time in 1936 with the fencer Bernardo de la Guardia and the Winter Olympics for the first time in 1980 with the skier Arturo Kinch. All four of Costa Rica's Olympic medals were won by the sisters Silvia and Claudia Poll in swimming, with Claudia winning the only gold medal in 1996. Football is the most popular sport in Costa Rica. The national team has played in five FIFA World Cup tournaments and reached the quarter-finals for the first time in 2014. Its best performance in the regional CONCACAF Gold Cup was runner-up in 2002. Paulo Wanchope, a forward who played for three clubs in England's Premier League in the late 1990s and early 2000s, is credited with enhancing foreign recognition of Costa Rican football. Basketball is also a popular sport in Costa Rica even though the country's national team has not yet qualified for a major international tournament such as the FIBA AmeriCup or the FIBA World Cup. The literacy rate in Costa Rica is approximately 97 percent and English is widely spoken primarily due to Costa Rica's tourism industry. When the army was abolished in 1949, it was said that the "army would be replaced with an army of teachers". Universal public education is guaranteed in the constitution; primary education is obligatory, and both preschool and high school are free. Students who finish 11th grade receive a Costa Rican Bachillerato Diploma accredited by the Costa Rican Ministry of Education. There are both state and private universities. The University of Costa Rica has been awarded the title "Meritorious Institution of Costa Rican Education and Culture". A 2016 report by the U.S. government report identifies the current challenges facing the education system, including the high dropout rate among high school students. The country needs even more workers who are fluent in English and languages such as Portuguese, Mandarin and French. It would also benefit from more graduates in Science, Technology, Engineering and Math (STEM) programs, according to the report. According to the UNDP, in 2010 the life expectancy at birth for Costa Ricans was 79.3 years. The Nicoya Peninsula is considered one of the Blue Zones in the world, where people commonly live active lives past the age of 100 years. The New Economics Foundation (NEF) ranked Costa Rica first in its 2009 Happy Planet Index, and once again in 2012. The index measures the health and happiness they produce per unit of environmental input. According to NEF, Costa Rica's lead is due to its very high life expectancy which is second highest in the Americas, and higher than the United States. The country also experienced well-being higher than many richer nations and a per capita ecological footprint one-third the size of the United States. In 2002, there were 0.58 new general practitioner (medical) consultations and 0.33 new specialist consultations per capita, and a hospital admission rate of 8.1%. Preventive health care is also successful. In 2002, 96% of Costa Rican women used some form of contraception, and antenatal care services were provided to 87% of all pregnant women. All children under one have access to well-baby clinics, and the immunization coverage rate in 2002 was above 91% for all antigens. Costa Rica has a very low malaria incidence of 48 per 100,000 in 2000 and no reported cases of measles in 2002. The perinatal mortality rate dropped from 12.0 per 1000 in 1972 to 5.4 per 1000 in 2001. Costa Rica has been cited as Central America's great health success story. Its healthcare system is ranked higher than that of the United States, despite having a fraction of its GDP. Prior to 1940, government hospitals and charities provided most health care. But since the 1941 creation of the Social Insurance Administration ("Caja Costarricense de Seguro Social" – CCSS), Costa Rica has provided universal health care to its wage-earning residents, with coverage extended to dependants over time. In 1973, the CCSS took over administration of all 29 of the country's public hospitals and all health care, also launching a Rural Health Program ("Programa de Salud Rural") for primary care to rural areas, later extended to primary care services nationwide. In 1993, laws were passed to enable elected health boards that represented health consumers, social insurance representatives, employers, and social organizations. By the year 2000, social health insurance coverage was available to 82% of the Costa Rican population. Each health committee manages an area equivalent to one of the 83 administrative cantons of Costa Rica. There is limited use of private, for-profit services (around 14.4% of the national total health expenditure). About 7% of GDP is allocated to the health sector, and over 70% is government funded. Primary health care facilities in Costa Rica include health clinics, with a general practitioner, nurse, clerk, pharmacist and a primary health technician. In 2008, there were five specialty national hospitals, three general national hospitals, seven regional hospitals, 13 peripheral hospitals, and 10 major clinics serving as referral centers for primary care clinics, which also deliver biopsychosocial services, family and community medical services and promotion and prevention programs. Patients can choose private health care to avoid waiting lists. Costa Rica is among the Latin America countries that have become popular destinations for medical tourism. In 2006, Costa Rica received 150,000 foreigners that came for medical treatment. Costa Rica is particularly attractive to Americans due to geographic proximity, high quality of medical services, and lower medical costs. Since 2012, smoking in Costa Rica is subject to some of the most restrictive regulations in the world.
https://en.wikipedia.org/wiki?curid=5551
History of Costa Rica The first indigenous peoples of Costa Rica were hunters and gatherers, and when the Spanish conquerors arrived, Costa Rica was divided in two distinct cultural areas due to its geographical location in the Intermediate Area, between Mesoamerican and the Andean cultures, with influences of both cultures. Christopher Columbus first dropped anchor in Costa Rica in 1502 at Isla Uvita. Soon after, his forces overcame the indigenous people. He incorporated the territory into the Captaincy General of Guatemala as a province of New Spain in 1524. For the next 300 years, Costa Rica was a colony of Spain. As a result, Costa Rica's culture has been greatly influenced by the culture of Spain. During this period, Costa Rica remained sparsely developed and impoverished. Following the Mexican War of Independence (1810–1821), Costa Rica became part of the independent Mexican Empire in 1821. Subsequently, Costa Rica was part of the Federal Republic of Central America in 1813, before gaining full independence in 1821. Its economy struggled due to lack of connections with European suppliers. In 1856, Costa Rica resisted United States settlers from mounting a take-over of the government. After 1869, Costa Rica established a democratic government. After the Costa Rican Civil War in 1948, the government drafted a new constitution, guaranteeing universal suffrage and the dismantling of the military. Today, Costa Rica is a democracy that relies on technology and eco-tourism for its economy. Although poverty has declined since the turn of the 21st century, economic problems still exist. Costa Rica is facing problems of underemployment, foreign and internal debt, and a trade deficiency. The oldest evidence of human occupation in Costa Rica is associated with the arrival of groups of hunter-gatherers about 10,000 to 7,000 years BC, with ancient archaeological evidence (stone tool making) located in the Turrialba Valley, at sites called Guardiria and Florence, with matching quarry and workshop areas with presence of type clovis spearheads and South American inspired arrows. All this suggests the possibility that in this area two different cultures coexisted. The people of this era were nomadic. They were organized in family-based bands of about 20 to 30 members. Their usual prey animals were called megafauna, such as giant armadillos and sloths, mastodons, etc. These became extinct about 8,000 years before the modern era. The first settlers had to adapt to hunting smaller animals and develop appropriate strategies to adjust to the new conditions. In Pre-Columbian times, the native peoples in what is now Costa Rica were divided in two cultural areas due to its geographical location in the Intermediate Area, between the Mesoamerican and the Andean cultural regions. The northwest of the country, the Nicoya Peninsula, was the southernmost point of Mesoamerican cultural influence when the Spanish conquerors came in the sixteenth century. The Nicoya culture was the largest "cacicazgo" on the Pacific coast of Costa Rica. The central and southern portions of the country belonged to the Isthmo-Colombian cultural area with strong Muisca influences, as these were part of territories occupied predominantly by speakers of the Chibchan languages. The Diquis culture flourished from 700 CE to 1530 CE and were well known for their crafts in metal and stonework. The indigenous people have influenced modern Costa Rican culture to a relatively small degree. In the years soon after European encounter, many of the people died due to infectious diseases, such as measles and smallpox, which were endemic among the Europeans but to which they had no immunity. The colonial period began when Christopher Columbus reached the eastern coast of Costa Rica on his fourth voyage on September 18, 1502. Numerous subsequent Spanish expeditions followed, eventually leading to the first Spanish colony in Costa Rica, , founded in 1524. During most of the colonial period, Costa Rica was the southernmost province of the Captaincy General of Guatemala, which was nominally part of the Viceroyalty of New Spain (i.e., Mexico). In practice it operated as a largely autonomous entity within the Spanish Empire. Costa Rica's distance from the capital in Guatemala, its legal prohibition under Spanish law against trading with its southern neighbors in Panama, then part of the Viceroyalty of New Granada (i.e., Colombia), and the lack of resources such as gold and silver, resulted in Costa Rica attracting few inhabitants. It was a poor, isolated, and sparsely inhabited region within the Spanish Empire. a Spanish governor in 1719 described Costa Rica as "the poorest and most miserable Spanish colony in all America." Many historians say that the area suffered a lack of indigenous population available for forced labor, which meant that most of the Costa Rican settlers had to work their own land. This prevented the establishment of large "haciendas". For all these reasons Costa Rica was by and large unappreciated and overlooked by the Spanish Crown and left to develop on its own. The small landowners' relative poverty, the lack of a large indigenous labor force, the population's ethnic and linguistic homogeneity, and Costa Rica's isolation from the Spanish colonial centers in Mexico and the Andes, all contributed to the development of an autonomous and individualistic agrarian society. Even the Governor had to farm his own crops and tend to his own garden due to his poverty. The failure to build a colonial society based on indigenous and slave labor led to a peasant economy in the 1700s. During the time of conquest, as many as twenty distinct indigenous societies, numbering in the hundreds of thousands and speaking many different languages, inhabited the area. The Spanish conquest of Costa Rica lasted more than half a century after it started 1510. The genocidal enslavement of the indigenous societies of Nicoya on the Pacific North coast was the conquest's first stage. Its second phase began with fruitless attempts to consolidate a Spanish settlement on the country's Caribbean side. In the process, Spaniards reduced the indigenous population to the point of extinction through disease, war, reprisals, relocation and brutal exploitation. The Native American population stood at about 120,000 in 1569 and had fallen to 10,000 by 1611. In the early 19th century, Napoleon's occupation of Spain led to the outbreak of revolts all across Spanish America. In New Spain, all of the fighting by those seeking independence was done in the center of that area from 1810 to 1821, what today is central Mexico. Once the Viceroy was defeated in the capital city—today Mexico City—in 1821, the news of independence was sent to all the territories of New Spain, including the Intendencies of the former Captaincy General of Guatemala. Costa Rica joined the other Central American Intendancies in a joint declaration of independence from Spain, the 1821 Act of Independence. On October 13, 1821 the documents arrived at Cartago, and an emergency meeting was called upon by Governor . There were many ideas on what to do upon gaining independence, such as joining Mexico, joining Guatemala or Nueva Granada (today Colombia). A group was declared (Junta de Legados), which created the temporary . Meanwhile, "the clouds clear up" ("Mientras se aclaraban los nublados del día"), was a famous phrase of the events of the day. Independence from Spain was acknowledged and ratified on October 29, 1821 by the colonial authorities. It was then ratified in the cities of San José on November 1, 1821, at Cartago on November 3rd, 1821, at Heredia on November 11, 1821, and Alajuela on November 25, 1821. After the declaration of independence, the New Spain parliament intended to establish a commonwealth whereby the King of Spain, Ferdinand VII, would also be Emperor of New Spain, but in which both countries were to be governed by separate laws and with their own legislative offices. Should the king refuse the position, the law provided for a member of the House of Bourbon to accede to the New Spain throne. Ferdinand VII did not recognize the colony's independence and said that Spain would not allow any other European prince to take the throne of New Spain. By request of Parliament, the president of the regency, Agustín de Iturbide, was proclaimed emperor of New Spain, which was renamed Mexico. The Mexican Empire was the official name given to this monarchical regime from 1821 to 1823. The territory of the Mexican Empire included the continental intendancies and provinces of New Spain proper (including those of the former Captaincy General of Guatemala) (See: History of Central America). On 5 April 1823 the Battle of Ochomogo was fought between forces from Cartago who wanted to join the Mexican Empire and those who preferred to remain independent. The Republicans won and the capital was moved from Cartago to San José. In 1823, a revolution in Mexico ousted Emperor Agustín de Iturbide. A new Mexican congress voted to allow the Central American Intendancies to decide their own fate. That year, the United Provinces of Central America was formed of the five Central American Intendancies under General Manuel José Arce. The Intendancies took the new name of States. The United Provinces federation, not strongly united to begin with, rapidly disintegrated under the pressures of intra-provincial rivalries. Following full independence in 1838, Costa Rica had no regular trade routes established to export their coffee to European markets. Lack of infrastructure caused problems in transportation: the coffee-growing areas were mainly in the Central Valley and had access only to the port of Puntarenas on the Pacific coast. Before the Panama Canal opened, ships from Europe had to sail around Cape Horn in order to get to the Pacific Coast. In 1843, the country established a trade route to Europe with the help of William Le Lacheur, a Guernsey merchant and shipowner. In 1856, William Walker, an American filibuster, began incursions into Central America. After landing in Nicaragua, he proclaimed himself as president of Nicaragua and re-instated slavery, which had been abolished. He intended to expand into Costa Rica and after he entered that territory, the country declared war against his forces. Led by Commander in Chief of the Army of Costa Rica, President Juan Rafael Mora Porras, the filibusters were defeated and forced out of the country. Costa Rican forces followed the filibusters into Rivas, Nicaragua, where in a final battle, William Walker and his forces were finally pushed back. In this final battle, Juan Santamaría, a drummer boy from Alajuela, lost his life torching the filibusters' stronghold. He is today remembered as a national hero. An era of peaceful democracy in Costa Rica began in 1869 with elections. Costa Rica has avoided much of the violence that has plagued Central America. Since the late nineteenth century, only two brief periods of violence have marred its republican development. In 1917–19, Federico Tinoco Granados ruled as a dictator. In 1948, José Figueres Ferrer led an armed uprising in the wake of a disputed presidential election. "With more than 2,000 dead, the 44-day Costa Rican Civil War resulting from this uprising was the bloodiest event in twentieth-century Costa Rican history." The victorious junta drafted a constitution guaranteeing free elections with universal suffrage and the abolition of the military. Figueres became a national hero, winning the first election under the new constitution in 1953. Since then Costa Rica has been one of the few democracies to operate without a standing army. The nation has held 16 successive presidential elections, all peaceful, the latest being in 2018. Costa Rica's economy went under a transformation in 1978. The country went from being "an economic development success story" to entering a severe socio-economic crisis. Costa Rica relied on the exportation of bananas and coffee. In 1978, coffee prices dropped, and its revenues declined. In 1979, the price of oil, a main imported item, increased sharply and rapidly, plunging the country into crisis. In order to help improve the economy, President Rodrigo Carazo continued to borrow money internationally. This led the country into further debt. Once a largely agricultural country, Costa Rica has transformed to relying on technology industry and services, and eco-tourism. Costa Rica's major source of export income is technology-based. Microsoft, Motorola, Intel and other technology-related firms have established operations in Costa Rica. Local companies create and export software as well as other computer-related products. Tourism is growing at an accelerated pace, and many believe that income from this tourism may soon become the major contributor to the nation's GDP. Traditional agriculture, particularly coffee and bananas, continues to be an important part of Costa Rica's exports. General:
https://en.wikipedia.org/wiki?curid=5552
Geography of Costa Rica Costa Rica is located on the Central American Isthmus, surrounding the point 10° north of the equator and 84° west of the prime meridian. It has 212 km of Caribbean Sea coastline and 1,016 on the North Pacific Ocean. The area is 51,100 km² of which 40 km² is water. It is slightly smaller than the U.S. state of West Virginia. Costa Rica is located on the Caribbean Plate. It borders the Cocos Plate in the Pacific Ocean which is being subducted beneath it. This forms the volcanoes in Costa Rica, also known as the Central America Volcanic Arc. The Caribbean Plate began its eastward migration during the Late Cretaceous. During the Late Paleocene, a local sea-level low-stand assisted by the continental uplift of the western margin of South America, resulted in a land bridge over which several groups of mammals apparently took part in an interchange. Many earthquakes in Costa Rica have occurred. Costa Rica shares a 313 km border with Nicaragua to the north, and a 348 km with Panama to the south. Costa Rica claims an exclusive economic zone of with and a territorial sea of . Land use: Arable land: 4.8%. Permanent crops: 6.66%. Other: 88.54%. Administrative divisions of Costa Rica include 7 provinces, 82 cantons,and 478 districts. There are also 24 indigenous territories. There are many islands of Costa Rica, the most remote being Cocos Island and the largest being Isla Calero. The nation's coastal plain separated by the Cordillera Central and the Cordillera de Talamanca, which form the spine of the country and separate the Pacific and Caribbean drainage divides. The Cordillera de Guanacaste is in the north near the border with Nicaragua and forms part of the Continental Divide of the Americas. Much of the Cordillera de Talamanca is included in the La Amistad International Park, which is shared between Costa Rica and Panama. It contains the country's highest peaks: the Cerro Chirripó and the Cerro Kamuk. Much of the region is covered by the Talamancan montane forests. It also includes the Cerros de Escazú which borders the Costa Rican Central Valley to the south. Irrigated land covers 1,031 km². Rivers of Costa Rica all drain into the Caribbean or the Pacific. Cocos Island is the southwestern extreme of the country. Otherwise to the north it's Peñas Blancas, to the south and east the Panama border, and to the west the Santa Elena Peninsula. The lowest point is sea level, and the highest is Cerro Chirripo: at 3810 m. The climate is tropical and subtropical. Dry season (December to April); rainy season (May to November); cooler in highlands. Because Costa Rica is located between 8 and 12 degrees north of the Equator, the climate is tropical year round. However, the country has many microclimates depending on elevation, rainfall, topography, and by the geography of each particular region. Costa Rica's seasons are defined by how much rain falls during a particular period. The year can be split into two periods, the dry season known to the residents as summer (), and the rainy season, known locally as winter (). The "summer" or dry season goes from December to April, and "winter" or rainy season goes from May to November, which almost coincides with the Atlantic hurricane season, and during this time, it rains constantly in some regions. The location receiving the most rain is the Caribbean slopes of the Cordillera Central mountains, with an annual rainfall of over . Humidity is also higher on the Caribbean side than on the Pacific side. The mean annual temperature on the coastal lowlands is around , in the main populated areas of the Cordillera Central, and below on the summits of the highest mountains. Costa Rica is a biodiversity hotspot. While the country has only about 0.03% of the world's landmass, it contains 5% of the world's biodiversity. It is home to about 12,119 species of plants, of which 950 are endemic. There are 117 native trees and more than 1,400 types of orchids; a third of them can be found in the Monteverde Cloud Forest Reserve. Almost a half of the country's land is covered by forests, though only 3.5% is covered by primary forests. Deforestation in Costa Rica has been reduced from some of the worst rates in the world from 1973 to 1989, to almost zero by 2005. The diversity of wildlife in Costa Rica is very high; there are 441 species of amphibians and reptiles, 838 species of birds, 232 species of mammals and 181 species of fresh water fish. Costa Rica has high levels of endemism; 81 species of amphibians and reptiles, 17 species of birds and 7 species of mammals are endemic to the country. However, many species are endangered. According to the World Conservation Monitoring Centre, 209 species of birds, mammals, reptiles, amphibians and plants are endangered. Some of the country's most endangered species are the Harpy eagle, the Giant anteater, the Golden toad and the Jaguar. IUCN reports the Golden toad as extinct. Over 25% of Costa Rica's national territory is protected by SINAC (the National System of Conservation Areas), which oversees all of the country's protected areas. There 29 national parks of Costa Rica many conservation areas of Costa Rica. Together protected areas comprise over one-fourth of Costa Rican territory. 9.3% of the country is protected under IUCN categories I-V. Around 25% of the country's land area is in protected national parks and protected areas, the largest percentage of protected areas in the world (developing world average 13%, developed world average 8%). Tortuguero National Park is home to monkeys, sloths, birds; and a variety of reptiles. The Monteverde Cloud Forest Reserve is home to about 2,000 plant species, including numerous orchids. Over 400 types of birds and more than 100 species of mammals can be found there. Over 840 species of birds have been identified in Costa Rica. As is the case in much of Central America, the avian species in Costa Rica are a mix of North and South American species. The country's abundant fruit trees, many of which bear fruit year round, are hugely important to the birds, some of whom survive on diets that consist only of one or two types of fruit. Some of the country's most notable avian species include the resplendent quetzal, scarlet macaw, three-wattled bellbird, bare-necked umbrellabird, and the keel-billed toucan. The Instituto Nacional de Biodiversidad is allowed to collect royalties on any biological discoveries of medical importance. Costa Rica is a center of biological diversity for reptiles and amphibians, including the world's fastest running lizard, the spiny-tailed iguana ("Ctenosaura similis"). Hydropower from Lake Arenal, the largest lake in Costa Rica. Total renewable water resources is 112.4 km³. Freshwater withdrawal is 5.77 km³/year (15%/9%/77%), or per capita: 1,582 m³/year. Agriculture is the largest water user demanding around 53% of total supplies while the sector contributes 6.5% to the Costa Rica GDP. Both total and per capita water usage is very high in comparison to other Central American countries but when measured against available freshwater sources, Costa Rica uses only 5% of its available supply. Increasing urbanization will put pressure on water resources management in Costa Rica.
https://en.wikipedia.org/wiki?curid=5553
Demographics of Costa Rica This article is about the demographic features of the population of Costa Rica, including population density, ethnicity, education level, health of the populace, economic status, religious affiliations and other aspects of the population. According to the United Nations, in Costa Rica had an estimated population of people. White and Mestizos make up 83.4% of the population, 7% are black people (including mixed race), 2.4% Amerindians, 0.2% Asians, and 7% other/none. In 2010, just under 3% of the population is of black African descent who are called Afro-Costa Ricans or West Indians and are English-speaking descendants of 19th-century black Jamaican immigrant workers. Another 1% is composed of ethnic Chinese, and less than 1% are Middle Easterners, mainly of Lebanese descent but also Palestinians. The 2011 Census provided the following data: whites and mestizos make up 83.4% of the population, 7% are black people (including mixed race), 2.4% Amerindians, 0.2% Asians, and 7% other/none. There is also a community of North American retirees from the United States and Canada, followed by fairly large numbers of European Union expatriates (esp. Scandinavians and from Germany) come to retire as well, and Australians. Immigration to Costa Rica made up 9% of the population in 2012. This included permanent settlers as well as migrants who were hoping to reach the U.S. In 2015, there were some 420,000 immigrants in Costa Rica and the number of asylum seekers (mostly from Honduras, El Salvador, Guatemala and Nicaragua) rose to more than 110,000. An estimated 10% of the Costa Rican population in 2014 was made up of Nicaraguans. The indigenous population today numbers about 60,000 (just over 1% of the population) with some Miskito and Garifuna (a population of mixed black African and Carib Indian descent) living in the coastal regions. Costa Rica's emigration is the smallest in the Caribbean Basin and is among the smallest in the Americas. By 2015 about just 133,185 (2.77%) of the country's people live in another country as immigrants. The main destination countries are the United States (85,924), Nicaragua (10,772), Panama (7,760), Canada (5,039), Spain (3,339), Mexico (2,464), Germany (1,891), Italy (1,508), Guatemala (1,162) and Venezuela (1,127). In , Costa Rica had a population of . The population is increasing at a rate of 1.5% per year. At current trends the population will increase to 9,158,000 in about 46 years. The population density is 94 people per square km, the third highest in Central America. Approximately 40% lived in rural areas and 60% in urban areas. The rate of urbanization estimated for the period 2005–2015 is 2.74% per annum, one of the highest among developing countries. About 75% of the population live in the upper lands (above 500 meters) where temperature is cooler and milder. The 2011 census counted a population of 4.3 million people distributed among the following groups: 83.6% whites or mestizos, 6.7% black mixed race, 2.4% Native American, 1.1% black or Afro-Caribbean; the census showed 1.1% as Other, 2.9% (141,304 people) as None, and 2.2% (107,196 people) as unspecified. In 2011, there were over 104,000 Native American or indigenous inhabitants, representing 2.4% of the population. Most of them live in secluded reservations, distributed among eight ethnic groups: Quitirrisí (in the Central Valley), Matambú or Chorotega (Guanacaste), Maleku (northern Alajuela), Bribri (southern Atlantic), Cabécar (Cordillera de Talamanca), Guaymí (southern Costa Rica, along the Panamá border), Boruca (southern Costa Rica) and Térraba (southern Costa Rica). The population includes European Costa Ricans (of European ancestry), primarily of Spanish descent, with significant numbers of Italian, German, English, Dutch, French, Irish, Portuguese, and Polish families, as well a sizable Jewish community. The majority of the Afro-Costa Ricans are Creole English-speaking descendants of 19th century black Jamaican immigrant workers. The 2011 census classified 83.6% of the population as white or Mestizo; the latter are persons of combined European and Amerindian descent. The Mulatto segment (mix of white and black) represented 6.7% and indigenous people made up 2.4% of the population. Native and European mixed blood populations are far less than in other Latin American countries. Exceptions are Guanacaste, where almost half the population is visibly mestizo, a legacy of the more pervasive unions between Spanish colonists and Chorotega Amerindians through several generations, and Limón, where the vast majority of the Afro-Costa Rican community lives. According to the United Nations, Costa Rica's literacy rate stands at 95.8%, the fifth highest among American countries. Costa Rica's Education Index in 2006 was 0.882; higher than that of richer countries, such as Singapore and Mexico. Costa Rica's gross enrolment ratio is 73.0%, smaller than that of the neighbors countries of El Salvador and Honduras. All students must complete primary school and secondary school, between 6 and 15 years. Some students drop out because they must work to help support their families. In 2007 there were 536,436 pupils enrolled in 3,771 primary schools and 377,900 students attended public and private secondary schools. Costa Rica's main universities are the University of Costa Rica, in San Pedro and the National University of Costa Rica, in Heredia. Costa Rica also has several small private universities. Costa Rica's emigration is among the smallest in the Caribbean Basin. About 3% of the country's people live in another country as immigrants. The main destination countries are the United States, Spain, Mexico and other Central American countries. In 2005, there were 127,061 Costa Ricans living in another country as immigrants. Remittances were $513,000,000 in 2006 and they represented 2.3% of the country's GDP. Costa Rica's immigration is among the largest in the Caribbean Basin. According to the 2011 census 385,899 residents were born abroad. The vast majority were born in Nicaragua (287,766). Other countries of origin were Colombia (20,514), United States (16,898), Spain (16,482) and Panama (11,250). Outward Remittances were $246,000,000 in 2006. According to the World Bank, about 489,200 migrants lived in the country in 2010 mainly from Nicaragua, Panama, El Salvador, Honduras, Guatemala, and Belize, while 125,306 Costa Ricans live abroad in the United States, Panama, Nicaragua, Spain, Mexico, Canada, Germany, Venezuela, Dominican Republic, and Ecuador. The number of migrants declined in later years but in 2015, there were some 420,000 immigrants in Costa Rica and the number of asylum seekers (mostly from Honduras, El Salvador, Guatemala and Nicaragua) rose to more than 110,000, a fivefold increase from 2012. In 2016, the country was called a "magnet" for migrants from South and Central America and other countries who were hoping to reach the U.S. European Costa Ricans are people from Costa Rica whose ancestry lies within the continent of Europe, most notably Spain. According to DNA studies, around 75% of the population have some level of European ancestry. Percentages of the Costa Rican population by race are known as the national census does have the question of ethnicity included in its form. As for 2012 65.80% of Costa Ricans identify themselves as white/castizo and 13.65% as mestizo, giving around 80% of Caucasian population. This, however, is based in self-identification and not in scientific studies. According to PLoS "Genetics Geographic Patterns of Genome Admixture in Latin American Mestizos" study of 2012, Costa Ricans have 68% of European ancestry, 29% aboriginal and 3% African. According to CIA Factbook, Costa Rica has white or mestizo population of the 83.6%. Cristopher Columbus and crew were the first Europeans ever to set foot on what is now Costa Rica in Columbus last trip when he arrived to Uvita Island (modern day Limón province) in 1502. Costa Rica was part of the Spanish Empire and colonized by Spaniards mostly Castilians, Basque and Sefardi Jews. After the independence large migrations of wealthy Americans, Germans, French and British businessmen came to the country encouraged by the government and followed by their families and employees (many of them technicians and professionals) creating colonies and mixing with the population, especially the high and middle classes. Later, more humble migrations of Italians, Spanish (mostly Catalans) and Arab (mostly Lebanese and Syrians) migrants visit the country escaping economical crisis in their home countries, setting in large, more closed colonies. Polish migrants, mostly Ashkenazi Jews escaping anti-Semitism and nazi persecution in Europe also migrated to the country in large numbers. In 1901 president Ascensión Esquivel Ibarra closes the country to all non-white immigration forbidding the entrance of all Black African, Chinese, Arab, Turkish or Gypsy migration in the country. After the beginning of the Spanish Civil War large migration of Republican refugees also settle in the country, mostly Castilians, Galicians and Asturians, as later Chilean, Mexican and Colombian migrants would leave their countries traveling to Costa Rica escaping from war or dictatorships as Costa Rica is the longest running democracy in Latin America and unlike most of its neighbors had no dictatorship during the 20th century. The following listing is taken from a publication of the Costa Rica 2011 Census: Structure of the population (01.07.2017) (Estimates - the source of data is the national household survey): Source: "UN World Population Prospects" Demographic statistics according to the World Population Review in 2019. Demographic statistics according to the CIA World Factbook, unless otherwise indicated. Roman Catholic 71.8%, Evangelical and Pentecostal 12.3%, other Protestant 2.6%, Jehovah's Witness 0.5%, other 2.4%, none 10.4% (2016 est.) Nearly all Costa Ricans speak Spanish; but many blacks speak a traditional Jamaican dialect of English, also a few of the Natives speak their own language, such as the case of the Ngobes. According to the World Factbook the main religions are: Roman Catholic, 76.3%; Evangelical, 13.7%; Jehovah's Witnesses, 1.3%; other Protestant, 0.7%; other, 4.8%; none, 3.2%. The most recent nationwide survey of religion in Costa Rica, conducted in 2007 by the University of Costa Rica, found that 70.5 percent of the population identify themselves as Roman Catholics (with 44.9 percent practicing, 25.6 percent nonpracticing), 13.8 percent are Evangelical Protestants, 11.3 percent report that they do not have a religion, and 4.3 percent declare that they belong to another religion. Apart from the dominant Catholic religion, there are several other religious groups in the country. Methodist, Lutheran, Episcopal, Baptist, and other Protestant groups have significant membership. The Church of Jesus Christ of Latter-day Saints (LDS Church) claim more than 35,000 members and has a temple in San José that served as a regional worship center for Costa Rica, Panama, Nicaragua, and Honduras. Although they represent less than 1 percent of the population, Jehovah's Witnesses have a strong presence on the Caribbean coast. Seventh-day Adventists operate a university that attracts students from throughout the Caribbean Basin. The Unification Church maintains its continental headquarters for Latin America in San José. Non-Christian religious groups, including followers of Judaism, Islam, Taoism, Hare Krishna, Paganism, Wicca, Scientology, Tenrikyo, and the Bahá'í Faith, claim membership throughout the country, with the majority of worshipers residing in the Central Valley (the area of the capital). While there is no general correlation between religion and ethnicity, indigenous peoples are more likely to practice animism than other religions. Article 75 of the Costa Rican Constitution states that the "Catholic, Apostolic, and Roman Religion is the official religion of the Republic". That same article provides for freedom of religion, and the Government generally respects this right in practice. The US government found no reports of societal abuses or discrimination based on religious belief or practice in 2007.
https://en.wikipedia.org/wiki?curid=5554
Politics of Costa Rica The politics of Costa Rica take place in a framework of a presidential, representative democratic republic, with a multi-party system. Executive power is exercised by the president and their cabinet, and the President of Costa Rica is both the head of state and head of government. Legislative power is vested in the Legislative Assembly. The president and 57 Legislative Assembly deputies are elected for four-year terms. The judiciary operates independent of the executive and the legislature but remains involved in the political process. Costa Rica is a republic with a strong system of constitutional checks and balances. Voting is compulsory in Costa Rica but it is not enforced. The position of governor in the seven provinces was abolished in 1998. There are no provincial legislatures. In 2009, the state monopolies on insurance and telecommunications were opened to private-sector competition. Certain other state agencies enjoy considerable operational independence and autonomy; they include the electrical power company (Instituto Costarricense de Electricidad), the nationalized commercial banks (which are open to competition from private banks), and the social security agency (Caja Costarricense del Seguro Social). Costa Rica has no military but maintains a domestic police force and a Special Forces Unit as part of the Ministry of the President. The 1986 presidential election was won by Óscar Arias of the PLN. During his tenure he experienced some criticism from within his own party for abandoning its traditional social democratic teachings and promoting a neoliberal economic model. He received the Nobel Peace Prize in 1987 for his efforts to end civil wars then raging in several Central American countries. In the February 1998 national election, PUSC candidate Miguel Ángel Rodríguez won the presidency over PLN nominee José Miguel Corrales Bolaños. President Rodriguez assumed office May 8, 1998. The PUSC also obtained 27 seats in the 57-member Legislative Assembly, for a plurality, while the PLN got 23 and five minor parties won seven. Social Christian in philosophy, the PUSC generally favors neoliberalism, conservative fiscal policies, and government reform. President Rodriguez pledged to reduce the country's large internal debt, privatize state-owned utilities, attract additional foreign investment, eliminate social welfare programs, and promote the creation of jobs with decent salaries. The reforms he tried to promote found opposition from several parties, including his own, and he asserted several times the country was "ungovernable". In particular, an attempt by the Legislative Assembly to approve a law that opened up the electricity and telecommunication markets (controlled by a monopoly of the Costa Rican Institute of Electricity - ICE) to market competition, known as the "Combo" law, was met with strong social opposition. The Combo law was supported by both major parties at the time (PLN and PUSC) as well as by President Rodriguez, but the first of three required legislative votes to approve it provoked the largest protest demonstrations the country had seen since 1970. The government quickly resolved to shelve the initiative. President Rodríguez's approval would reach an all-time low, and he was indicted by the Attorney General after leaving office on corruption charges. In September 2000 the Constitutional Court rejected an argument by former president Arias that a 1969 constitutional amendment banning presidential reelection be rescinded. Arias thus remained barred from a second term as president; however, in April 2003–by which time two of the four judges who had voted against the change in 2000 had been replaced–the Court reconsidered the issue and, with the only dissenters being the two anti-reelection judges remaining from 2000, declared the 1969 amendment null and thus opened the way to reelection for former presidents–which in practice meant Arias. In the 2002 national election, a new party founded by former PLN Congressman and government Minister Ottón Solís captured 26% of the vote, forcing a runoff election for the first time in the country's history. Abel Pacheco was elected President, under a national unity platform, but continuing most of the neoliberal and conservative policies of Miguel Ángel Rodríguez. This election was also important because new parties won several seats in Congress, more than ever. The PUSC obtained 19 seats, PLN 17 seats, PAC 14 seats, PML 6 seats and PRC one seat. During 2004, several high-profile corruption scandals shattered the foundations of PUSC. Two former presidents from the party, Miguel Ángel Rodríguez and Rafael Ángel Calderón, were arrested on corruption charges and are currently waiting for the investigation to end and trial to begin. Also involved in scandals has been José María Figueres, former President from PLN and former head of the World Economic Forum. The 2006 national election was expected to be a landslide for former President (1986–1990) and PLN's candidate Óscar Arias, but it turned out to be the closest in modern history. Although polls just a week before the election gave Arias a comfortable lead of at least 12% (and up to 20%), preliminary election results gave him only a .4% lead over rival Ottón Solís and prompted a manual recount of all ballots. After a month-long recount and several appeals from different parties, Arias was declared the official winner with 40.9% of the votes against 39.8% for Solís. When Óscar Arias returned to office, the political debate shifted to the ratification of the Central American Free Trade Agreement (CAFTA). Main supporters of the approval included the President's PLN, which established a coalition with PUSC and ML in Congress to approve the implementation laws in Congress, as well as different business chambers. The main opposition to CAFTA came from PAC, labor unions, environmental organizations and public universities. In April 2007, former PLN Presidential candidate and CAFTA opponent José Miguel Corrales Bolaños won a legal battle at the Supreme Electoral Tribunal, which authorized him to gather over 100,000 signatures to send CAFTA to a referendum and let the people decide the fate of the controversial agreement. As the February 28, 2008 deadline to approve or reject CAFTA loomed, Arias decided to call for the referendum himself, and it took take place on October 7, 2007. CAFTA was approved with 51.5% of voters supporting it, although the election faced criticism due to international, including US, involvement. The Costa Rican general election, 2010 was won by Laura Chinchilla, who had been Vice-President in the previous Arias administration. In 2014, Luis Guillermo Solís, PAC's presidential candidate campaigning on a platform of economic reform and anti-corruption, surprised political observers by winning 30.95% of votes in the first round, while PLN candidate Johnny Araya gained the second most votes with 29.95%. Broad Front's José María Villalta Florez-Estrada won 17% of the votes Soon thereafter, Araya announced that he would cease campaigning, making Solís the favorite. Elections were still be held on April 6, 2014, as required by election law, and Solís won with 77.81 of the votes. According to the BBC, the success of Solís and Villalta is another example of anti-neoliberal politics in Latin America. Executive responsibilities are vested in a president, who is elected directly by the voters, not by the National Assembly as it would be in a parliamentary system. There also are two vice presidents and the president's cabinet composed of his ministers. A constitutional amendment approved in 1969 limits presidents and deputies to one term, although a deputy may run again for an Assembly seat after sitting out a term. The prohibition was officially recognized as unconstitutional in April 2004, allowing Óscar Arias to run for President a second time in the 2006 Costa Rican presidential elections, which he won with approximately a 1% margin. The President of Costa Rica has limited powers, particularly in comparison to other Latin American Presidents. For example, he cannot veto the legislative budget, and thus Congress is sovereign over the year's single most important piece of legislation. On the other hand, he can appoint anyone to his cabinet without approval from Congress. This provides the single most important power versus Congress that any Costa Rican President has. Legislative powers are held by the Legislative Assembly. Legislators, called deputies, are elected to non-consecutive four-year terms by popular, direct vote, using proportional representation in each of the country's seven provinces. Elections were last held in February 2014 and will be held again in February 2018. As a result, there are nine separate political parties serving in the Legislative Assembly, with National Liberation Party holding 18 seats, the Citizens' Action Party holding 13, and Broad Front and the Social Christian Unity Party each holding 8. Other parties hold the remaining seats. The main arm of the judiciary is the Supreme Court of Justice. Twenty-two magistrates are selected for the CSJ for 8-year terms by the Legislative Assembly, and lower courts. Sala IV, also known as the Constitutional Chamber of the Supreme Court, reviews legislation, executive actions, and certain writs for constitutionality. Courts below the Sala IV deal with issues involving legal and criminal disputes. Additionally, the Supreme Electoral Tribunal (TSE for its Spanish initials) is an independent branch of the CSJ, responsible for democratic elections. While the judiciary is independent of the politically-elected executive and legislative branches, it is often responsible for resolving political and legal conflicts. A Comptroller General, Procurator General, and an Ombudsman oversee the government and operate autonomously. These institutions have the right to scrutinize, investigate and prosecute government contracts. In addition, they may impose procedural requirements on most political and governmental agencies. The actions of politicians and political parties are frequently researched by these institutions. On the national level, the president, two vice-presidents and a legislature are elected for a four-year term. The Legislative Assembly has 57 members, elected by proportional representation in each of the country's seven provinces. The electoral process is supervised by an independent Supreme Electoral Tribunal (TSE for its Spanish initials). The TSE is a commission of three principal magistrates and six alternates selected by the Supreme Court of Justice of Costa Rica. All elections are conducted by a secret ballot at local polling stations. On election days, political parties often organize caravans and marches to get supporters to polling stations. In many areas, voting takes on a festive atmosphere with supporters of each party wearing traditional colors and decorating their cars, houses, and livestock with colored ribbons. Because the day of elections is a national holiday, most people have the day off. Currently, there are nine active political parties with representation in the Legislative Assembly of Costa Rica. An additional twelve parties ran, but did not receive enough votes to earn a seat in the assembly, making the total number of active parties in Costa Rica twenty-one. Starting in the 2000s, disagreement about many of the neo-liberal policies promoted by the dominant PLN caused the traditional party system of alliances among a few parties to fracture. Although still a stable country, the shift toward many political parties and away from PUSC and PLN is a recent development. Various elected positions within the country, such as mayors and city council members, are held by many different national and local political parties.
https://en.wikipedia.org/wiki?curid=5555
Economy of Costa Rica The economy of Costa Rica has been very stable for some years now, with continuing growth in the GDP (Gross Domestic Product) and moderate inflation, though with a high unemployment rate: 8.13% in 2018. Costa Rica's economy emerged from recession in 1997 and has shown strong aggregate growth since then. The estimated GDP for 2017 is US$61.5 billion, up significantly from the US$52.6 billion in 2015 while the estimated 2017 per capita (purchasing power parity) is US$12,382. Inflation remained around 4% to 5% per annum for several years up to 2015 but then dropped to 0.7% in 2016; it was expected to rise to a still moderate 2.8% by the end of 2017 In 2017, Costa Rica has one of the highest standards of living in Central America in spite of the high poverty level, which dropped by 1.2% in 2017 to 20.5% thanks to reducing inflation and benefits offered by the government. The estimated unemployment level in 2017 is 8.1%, roughly the same as in 2016. The country has evolved from an economy that once depended solely on agriculture, to one that is more diverse, based on tourism, electronics and medical components exports, medical manufacturing and IT services. Corporate services for foreign companies employ some 3% of the workforce. Of the GDP, 5.5% is generated by agriculture, 18.6% by industry and 75.9% by services (2016). Agriculture employs 12.9% of the labor force, industry 18.57%, services 69.02% (2016) Many foreign companies operate in the various Free-trade zones. In 2015, exports totalled US$12.6 billion while imports totalled US$15 billion for a trade deficit of US$2.39 billion. The growing debt and budget deficit are the country's primary concerns. By August 2017, Costa Rica was having difficulty paying its obligations and the President promised dramatic changes to handle the "liquidity crisis". Other challenges facing Costa Rica in its attempts to increase the economy by foreign investment include a poor infrastructure and a need to improve public sector efficiency. One of the country's major concerns is the level of the public debt, especially as a percentage of the GDP (Gross Domestic Product), increasing from 29.8% in 2011 to 40.8% in 2015 and to 45% in 2016. In a June 2017 report, the International Monetary Fund stated that annual growth was just over 4% with moderate inflation. The report added that "financial system appears sound, and credit growth continues to be consistent with healthy financial deepening and macroeconomic trends. The agency noted that the fiscal deficit remains high and public debt continues to rise rapidly despite the authorities’ deepened consolidation efforts in 2016. Recent advances in fiscal consolidation have been partly reversed and political consensus on a comprehensive fiscal package remains elusive". The IMF also expressed concern about increasing deficits, public debt and the heavy dollarization of bank assets and liabilities, warning that in tighter-than-expected global financial conditions these aspects would "seriously undermine investor confidence". The group also recommended taking steps to reduce pension benefits and increase the amount of contribution by the public and increasing the cost effectiveness of the education system. The country's credit rating was reduced by Moody's Investors Service in early 2017 to Ba2 from Ba1, with a negative outlook on the rating. The agency particularly cited the "rising government debt burden and persistently high fiscal deficit, which was 5.2% of GDP in 2016". Moody's was also concerned about the "lack of political consensus to implement measures to reduce the fiscal deficit [which] will result in further pressure on the government's debt ratios". In late July 2017, the Central Bank estimated the budget deficit at 6.1 percent of the country's GDP. A 2017 study by the Organisation for Economic Co-operation and Development warned that reducing the foreign debt must be a very high priority for the government. Other fiscal reforms were also recommended to moderate the budget deficit. In 2014, President Solís presented a budget with an increase in spending of 19% for 2015, an increase of 0.5% for 2016 and an increase of 12% for 2017. When the 2017 budget was finally proposed, it totaled US$15.9 billion. Debt payments account for one-third of that amount. Of greater concern is the fact that a full 46% of the budget will require financing, a step that will increase the debt owed to foreign entities. In late July 2017, the Central Bank estimated the budget deficit at 6.1 percent of the country's GDP. In early August 2017, President Luis Guillermo Solís admitted that the country was facing a "liquidity crisis", an inability to pay all of its obligations and to guarantee the essential services. To address this issue, he promised that a higher VAT and higher income tax rates were being considered by his government. Such steps are essential, Solís told the nation. "Despite all the public calls and efforts we have made since the start of my administration to contain spending and increase revenues, there is still a gap that we must close with fresh resources," he said. The crisis was occurring in spite of growth, low inflation and continued moderate interest rates, Solís concluded. Solís explained that the Treasury will prioritize payments on the public debt first, then salaries, and then pensions. The subsequent priorities include transfers to institutions "according to their social urgency." All other payments will be made only if funds are available. A 2016 report by the U.S. government report identifies other challenges facing Costa Rica as it works to expand its economy by working with potential foreign investors: Costa Rica's rainfall, and its location in the Central American isthmus, which provides easy access to North and South American markets and direct ocean access to the European and Asian Continents. Costa Rica has two seasons, both of which have their own agricultural resources: the tropical wet and dry seasons. One-fourth of Costa Rica's land is dedicated to national forests, often adjoining beaches, which has made the country a popular destination for affluent retirees and ecotourists. A full 10.27% of the country is protected as national parks while an additional 17% is set aside for reserves, wildlife refuges and protected zones. Costa Rica has over 50 wildlife refuges, 32 major national parks, more than 12 forest reserves and a few biological reserves. Because of ocean access, 23.7% of Costa Ricas people fish and trade their catches to fish companies; this is viewed as "small scale artisanal coastal" fishing and is most common in the Gulf of Nicoya. Costa Rica also charges licensing fees for commercial fishing fleets that are taking tuna, sardines, banga mary, mahi-mahi, red tilapia, shrimp, red snapper, other snappers, shark, marlin and sailfish. In mid 2017, the country was planning to ban large-scale commercial fishing off the southern Pacific Coast in an area nearly a million acres in size. The bill in congress was intended to "protect the extraordinary marine and coastal resources" from "indiscriminate and unsustainable commercial fishing." Sport fishing is an important part of the tourism industry; species include marlin, sailfish, dorado, tarpon, snook, rooster fish, wahoo, tuna, mackerel, snapper and rainbow bass. In terms of the 2012 Environmental Performance Index ranking, Costa Rica is 5th in the world, and first among the Americas. The World Economic Forum's 2017 Travel & Tourism Competitiveness Report ranked Costa Rica as third of 136 countries based on natural resources, the number of World Heritage natural sites, protected areas and species as well as eco tourism. With a $1.92-billion-a-year tourism industry, Costa Rica was the most visited nation in the Central American region, with 2.42 million foreign visitors in 2013. By 2016, 2.6 million tourists visited Costa Rica. The Tourism Board estimates that this sector's spending in the country represented over US$3.4 billion, or about 5.8% of the GDP. The World Travel & Tourism Council's estimates indicate a direct contribution to the 2016 GDP of 5.1% and 110,000 direct jobs in Costa Rica; the total number of jobs indirectly supported by tourism was 271,000. Ecotourism is extremely popular with the many tourists visiting the extensive national parks and protected areas around the country. Costa Rica was a pioneer in this type of tourism and the country is recognized as one of the few with real ecotourism. Other important market segments are adventure, sun and beaches. Most of the tourists come from the U.S. and Canada (46%), and the EU (16%), the prime market travelers in the world, which translates into a relatively high expenditure per tourist of $1000 per trip. In the 2008 Travel and Tourism Competitiveness Index (TTCI), Costa Rica reached the 44th place in the world ranking, being the first among Latin American countries, and second if the Caribbean is included. Just considering the subindex measuring human, cultural, and natural resources, Costa Rica ranks in the 24th place at a worldwide level, and 7th when considering just the natural resources criteria. The TTCI report also notes Costa Rica's main weaknesses, ground transport infrastructure (ranked 113th), and safety and security (ranked 128th). The online travel magazine Travelzoo rated Costa Rica as one of five “Wow Deal Destinations for 2012”. The magazine Travel Weekly named Costa Rica the best destination in Central and South America in 2011. In 2017, the country was nominated in the following categories in the World Travel Awards: Mexico & Central America's Leading Beach Destination, Mexico & Central America's Leading Destination and Mexico & Central America's Leading Tourist Board. Costa Rica's economy was historically based on agriculture, and this has had a large cultural impact through the years. Costa Rica's main cash crop, historically and up to modern times, was Bananas. The coffee crop had been a major export, but decreased in value to the point where it added only 2.5% to the 2013 exports of the country. Agriculture also plays an important part in the country's gross domestic product (GDP). It makes up about 6.5% of Costa Rica’s GDP, and employs 12.9% of the labor force (2016). By comparison, 18.57% work in industry and 69.02 percent in the services sector. Depending on location and altitude, many regions differ in agricultural crops and techniques. The main agricultural exports from the country include: bananas, pineapples (the second highest export, with over 50% share of the world market), other tropical fruits, coffee (much of it grown in the Valle Central or Meseta Central), sugar, rice, palm oil, vegetables, tropical fruits, ornamental plants, maize, and potatoes. Livestock activity consists of cattle, pigs and horses, as well as poultry. Meat and dairy produce are leading exports according to one source, but both were not in the top 10 categories of 2013. The combined export value of forest products and textiles in 2013 did not exceed that of either chemical products or plastics. Mere decades ago, Costa Rica was known principally as a producer of bananas and coffee. Even though bananas, pineapple, sugar, coffee, lumber, wood products and beef are still important exports, in recent times medical instruments, electronics, pharmaceuticals, financial outsourcing, software development, and ecotourism are now the prime exports. High levels of education and fluency in English among its residents make the country an attractive investing location. In 2015 the following were the major export products (US$): Medical Instruments ($2B), Bananas ($1.24B), Tropical Fruits ($1.22B), Integrated Circuits ($841M) and Orthopedic Appliances ($555M). The total exports in 2015 were US$12.6 billion, down from $18.9B in 2010; bananas and medical instruments were the two largest sectors. Total imports in 2015 were $15B, up from $13.8B in 2010; this resulted in a trade deficit. Over the years, Costa Rica successfully attracted important investments by such companies as Intel Corporation, Procter & Gamble, Abbott Laboratories and Baxter Healthcare. Manufacturing and industry's contribution to GDP overtook agriculture over the course of the 1990s, led by foreign investment in Costa Rica's Free Trade Zones (FTZ) where companies benefit from investment and tax incentives. Companies in such zones must export at least 50% of their services. Well over half of that type of investment has come from the U.S. According to the government, the zones supported over 82 thousand direct jobs and 43 thousand indirect jobs in 2015; direct employment grew 5% over 2014. The average wages in the FTZ increased by 7% and were 1.8 times greater than the average for private enterprise work in the rest of the country. Companies with facilities in the America Free Zone in Heredia, for example, include Dell, HP, Bayer, Bosch, DHL, IBM and Okay Industries. In 2006 Intel's microprocessor facility alone was responsible for 20% of Costa Rican exports and 4.9% of the country's GDP. In 2014, Intel announced it would end manufacturing in Costa Rica and lay off 1,500 staff but agreed to maintain at least 1,200 employees. The facility continued as a test and design center with approximately 1,600 remaining staff. In 2017, Intel had 2000 employees in the country, and was operating a facility which assembles, tests and distributes processors and a Global Innovation Center, both in Heredia. The fastest growing aspect of the economy is the provision of corporate services for foreign companies which in 2016 employed approximately 54,000 people in a country with a workforce under 342,000; that was up from 52,400 the previous year. For example, Amazon.com employs some 5,000 people. Many work in the free-trade areas such as Zona Franca America and earn roughly double the national average for service work. This sector generated US$4.6 billion in 2016, nearly as much as tourism. In 2013, the total FDI stock in Costa Rica amounted to about 40 percent of GDP, of which investments from the United States accounted for 64 percent, followed by the United Kingdom and Spain with 6 percent each. Costa Rica's outward foreign direct investment stock is small, at about 3 percent of GDP as of 2011, and mainly concentrated in Central America (about 57 percent of the total outward direct investment stock). Tourism is an important part of the economy, with the number of visitors increasing from 780,000 in 1996, to 1 million in 1999, and to 2.089 million foreign visitors in 2008, allowing the country to earn $2.144-billion in that year. By 2016, 2.6 million tourists visited Costa Rica, spending roughly US$3.4 billion. Tourism directly supported 110,000 jobs and indirectly supported 271,000 in 2016. Costa Rica has not discovered sources of fossil fuels—apart from minor coal deposits—but its mountainous terrain and abundant rainfall have permitted the construction of a dozen hydroelectric power plants, making it self-sufficient in all energy needs, except for refined petroleum. In 2017, Costa Rica was considering the export of electricity to neighbouring countries. Mild climate and trade winds make neither heating nor cooling necessary, particularly in the highland cities and towns where some 90% of the population lives. Renewable energy in Costa Rica is the norm. In 2016, 98.1 per cent of the country's electricity came from green sources: hydro generating stations, geothermal plants, wind turbines, solar panels and biomass plants. Costa Rica's infrastructure has suffered from a lack of maintenance and new investment. The country has an extensive road system of more than 30,000 kilometers, although much of it is in disrepair; this also applies to ports, railways and water delivery systems. According to a 2016 U.S. government report, investment from China which attempted to improve the infrastructure found the "projects stalled by bureaucratic and legal concerns". Most parts of the country are accessible by road. The main highland cities in the country's Central Valley are connected by paved all-weather roads with the Atlantic and Pacific coasts and by the Pan American Highway with Nicaragua and Panama, the neighboring countries to the North and the South. Costa Rica's ports are struggling to keep pace with growing trade. They have insufficient capacity, and their equipment is in poor condition. The railroad didn't function for several years, until recent government effort to reactivate it for city transportation. An August 2016 OECD report provided this summary: "The road network is extensive but of poor quality, railways are in disrepair and only slowly being reactivated after having been shut down in the 1990s, seaports quality and capacity are deficient. Internal transportation overly relies on private road vehicles as the public transport system, especially railways, is inadequate." In a June 2017 interview, President Luis Guillermo Solís said that private sector investment would be required to solve the problems. "Of course Costa Rica’s infrastructure deficit is a challenge that outlasts any one government and I hope that we have created the foundations for future administrations to continue building. I have just enacted a law to facilitate Public Private Partnerships, which are the ideal way to develop projects that are too large for the government to undertake. For example the new airport that we are building to serve the capital city will cost $2 billion, so it will need private-sector involvement. There is also the potential for a ‘dry canal’ linking sea ports on our Atlantic and Caribbean Coasts that could need up to $16 billion of investment." The government hopes to bring foreign investment, technology, and management into the telecommunications and electrical power sectors, which are monopolies of the state. ICE (Instituto Costarricense de Electricidad) has the monopoly on telecommunications, internet and electricity services. Some limited competition is allowed. In 2011, two new private companies began offering cellular phone service and others offer voice communication over internet connections (VOIP) for overseas calls. According to transparency.org, Costa Rica had a reputation as one of the most stable, prosperous, and among the least corrupt in Latin America in 2007. However, in fall 2004, three former Costa Rican presidents, José María Figueres, Miguel Angel Rodríguez, and Rafael Angel Calderon, were investigated on corruption charges related to the issuance of government contracts. After extensive legal proceedings Calderon and Rodriguez were sentenced; however, the inquiry on Figueres was dismissed and he was not charged. More recently, Costa Rica reached 40th place in 2015, with a score of 55 on the Perception of Corruption scale; this is better than the global average. Countries with the lowest perceived corruption rated 90 on the scale. In late May 2017, the country Costa Rica applied to become a member of the OECD Anti-Bribery Convention, to be effective in July 2017. Costa Rica has sought to widen its economic and trade ties, both within and outside the region. Costa Rica signed a bilateral trade agreement with Mexico in 1994, which was later amended to cover a wider range of products. Costa Rica joined other Central American countries, plus the Dominican Republic, in establishing a Trade and Investment Council with the United States in March 1998, which later became the Dominican Republic–Central America Free Trade Agreement. Costa Rica has bilateral free trade agreements with the following countries and blocs which took effect on (see date): There are no significant trade barriers that would affect imports and the country has been lowering its tariffs in accordance with other Central American countries. Costa Rica also is a member of the Cairns Group, an organization of agricultural exporting countries that are seeking access to more markets to increase the exports of agricultural products. Opponents of free agricultural trade have sometimes attempted to block imports of products already grown in Costa Rica, including rice, potatoes, and onions. By 2015, Costa Rica's agricultural exports totalled US$2.7 billion. In 2015, the top export destinations for all types of products were the United States (US$4.29 billion), Guatemala ($587 million), the Netherlands ($537 million), Panama ($535 million) and Nicaragua ($496 million). The top import origins were the United States ($6.06 billion), China ($1.92 billion), Mexico ($1.14 billion), Japan ($410 million) and Guatemala ($409 million). The most significant products imported were Refined Petroleum (8.41% of the total imports) and Automobiles (4.68%). Total imports in 2015 were US$15 billion, somewhat higher than the total exports of a US$12.6 billion, for a negative trade balance of US$2.39 billion. The following table shows the main economic indicators in 1980–2017. GDP: US$61.5 billion (2017 estimate) GDP real growth rate: 4.3% (2017 estimate) GDP per capita: purchasing power parity: $12,382 (2017 estimate) GDP composition by sector: "agriculture:" 5.5% (2016 estimate) Bananas, pineapples, coffee, beef, sugarcane, rice, corn, dairy products, vegetables, timber, fruits and ornamental plants. "industry:" 18.6% (2016 estimate) Electronic components, food processing, textiles and apparel, construction materials, cement, fertilizer. "services:" 75.9% (2016 estimate) Hotels, restaurants, tourist services, banks, call centers and insurance. Government bond ratings: (January 2017) Standard & Poor's: BB-; Moody's: Ba2 Budget deficit: 6.1 percent of the GDP Population below poverty line: 20.5% (2017) Household income or consumption by percentage share: "lowest 10%:" 1.2% "highest 10%:" 39.5% (2009 est.) Inflation rate (consumer prices): 2.6% (2017 estimate) Labor force: 2.295 million (2016) "Note: 15 and older, excluding Nicaraguans living in the country" Labor force by occupation: agriculture 12.9%, industry 18.57%, services 69.02% (2016) Unemployment rate: 8.1% (2017 estimate) Budget: US15.9 billion (2017 proposed) Note: 46% will require financing Industries: microprocessors, food processing, textiles and clothing, construction materials, fertilizer, plastic products Industrial production growth rate: 4.3% (2013) Electricity production: 9.473 billion kWh (2010) Electricity production by source: 98.1% from "green sources" (2016) Agriculture products: bananas, pineapples, other tropical fruits, coffee, palm oil, sugar, corn, rice, beans, potatoes, beef, timber Exports: US$12.6 billion (2015) Major export commodities: Medical Instruments ($2B), Bananas ($1.24B), Tropical Fruits ($1.22B), Integrated Circuits ($841M) and Orthopedic Appliances ($555M). Export partners (2016): United States ($4.29B), Guatemala ($587M), the Netherlands ($537M), Panama ($535M), Nicaragua ($496M) Imports: US $15.1 billion (2015) Major import commodities: Refined Petroleum ($1.26B), Cars ($702M), Packaged Medicaments ($455M), Broadcasting Equipment ($374M) and Computers ($281M). Origin of imports (2016): United States ($6.06B), China ($1.92B), Mexico ($1.14B), Japan ($410M) and Guatemala ($409M). External debt: US$26.2 billion (January 2016) Economic aid – recipient: $107.1 million (1995) Currency: 1 Costa Rican colon (₡) = 100 centimos Exchange rates: Costa Rican colones (₡) per US$1 – 526.46 (March 27, 2015), US$1 – 600 (late May 2017), US$1 – 563 (end of July 2017) Fiscal year: October 1 – September 30
https://en.wikipedia.org/wiki?curid=5556
Transport in Costa Rica There are many modes of transport in Costa Rica but the country's infrastructure has suffered from a lack of maintenance and new investment. There is an extensive road system of more than 30,000 kilometers, although much of it is in disrepair; this also applies to ports, railways and water delivery systems. According to a 2016 U.S. government report, investment from China which attempted to improve the infrastructure found the "projects stalled by bureaucratic and legal concerns". Most parts of the country are accessible by road. The main highland cities in the country's Central Valley are connected by paved all-weather roads with the Atlantic and Pacific coasts and by the Pan American Highway with Nicaragua and Panama, the neighboring countries to the North and the South. Costa Rica's ports are struggling to keep pace with growing trade. They have insufficient capacity, and their equipment is in poor condition. The railroad didn't function for several years, until recent government effort to reactivate it for city transportation. An August 2016 OECD report provided this summary: "The road network is extensive but of poor quality, railways are in disrepair and only slowly being reactivated after having been shut down in the 1990s, seaports quality and capacity are deficient. Internal transportation overly relies on private road vehicles as the public transport system, especially railways, is inadequate." The road system in Costa Rica is not as developed as it might be expected for such a country. However, there are some two-lane trunk roads with restricted access under development. The Ministry of Public Works and Transport (MOPT), along with the National Road Council (Conavi), are the government organizations in charge of national road nomenclature and maintenance. There are three level of nationwide roads: These are trunk roads devised to connect important cities, most of the national roads are connected to the capital city, San José. There are 19 national primary roads, numbered between 1 and 39. These are roads that connect different cities, or primary routes, directly. There are 129 national secondary roads, numbered between 100 and 257. These roads connect main cities to villages or residential areas, there are 175 national tertiary roads, numbered between 301 and 935. , seasonally navigable by small craft In 2016, the government pledged ₡93 million ($166,000) for a new cruise ship terminal for Puerto Limón. "Total": 161
https://en.wikipedia.org/wiki?curid=5558
Public Force of Costa Rica The Public Force of Costa Rica is the country's law enforcement force, which performs policing and border patrol functions. On December 1, 1948, President José Figueres Ferrer of Costa Rica abolished the military of Costa Rica after achieving victory in the civil war that year. In a ceremony in the "Cuartel Bellavista", in the capital San José, Figueres broke a wall with a mallet symbolizing an end to Costa Rica's military spirit. In 1949, the abolition of the military was introduced in Article 12 of the Constitution of Costa Rica. The budget previously dedicated to the military is now dedicated to security, education and culture. Costa Rica maintains Police Guard forces. The museum "Museo Nacional de Costa Rica" was placed in the "Cuartel Bellavista" as a symbol of commitment to culture. In 1986, President Oscar Arias Sánchez declared December 1 as the "Día de la Abolición del Ejército" (Military abolition day) with Law #8115. Unlike its neighbors, Costa Rica has not endured a civil war since 1948. Costa Rica maintains small forces capable of law enforcement, but has no permanent standing army. In 1996, the Ministry of Public Security established the "Fuerza Pública" or Public Force, a gendarmerie which reorganized and eliminated the Civil Guard, Rural Assistance Guard, and Frontier Guards as separate entities. They are now under the Ministry and operate on a geographic command basis performing ground security, law enforcement, counter-narcotics, border patrol, and tourism security functions. The Costa Rica Coast Guard also operates directly under the Ministry. Outside the "Fuerza Pública", there is a small Special Forces Unit, the "Unidad Especial de Intervencion" (UEI) or Special Intervention Unit, an elite commando force which trains with special forces from around the world, but is not part of the main police forces. Instead it is part of the Intelligence and Security Directorate (DIS) which reports directly to the Minister of the Presidency. About 70 members strong, it is organized along military lines, although officially a civilian police unit. Its motto is "God, Fatherland, and Honor." Commissioner of Police Juan José Andrade Morales serves as its current Commissioner General. Handguns, shotguns and submachine guns Assault rifles Machineguns Sniper rifles Grenade launchers
https://en.wikipedia.org/wiki?curid=5559
Foreign relations of Costa Rica Costa Rica is an active member of the international community and, in 1983, claimed it was for neutrality. Due to certain powerful constituencies favoring its methods, it has a weight in world affairs far beyond its size. The country lobbied aggressively for the establishment of the Office of the United Nations High Commissioner for Human Rights and became the first nation to recognize the jurisdiction of the Inter-American Human Rights Court, based in San José. Costa Rica gained election as President of the Group of 77 in the United Nations in 1995. That term ended in 1997 with the South-South Conference held in San Jose. Costa Rica occupied a nonpermanent seat in the Security Council from 1997 to 1999 and exercised a leadership role in confronting crises in the Middle East and Africa, as well as in the former Socialist Federal Republic of Yugoslavia. It is currently a member of the United Nations Commission on Human Rights. On Jan. 1 2008 Costa Rica started its third year term on the Security Council. Costa Rica strongly backed efforts by the United States to implement UN Security Council Resolution 940, which led to the restoration of the democratically elected Government of Haiti in October 1994. Costa Rica was among the first to call for a postponement of the May 22 elections in Peru when international observer missions found electoral machinery not prepared for the vote count. Costa Rica is also a member of the International Criminal Court, without a Bilateral Immunity Agreement of protection for the US-military (as covered under Article 98) In 1987, then President Óscar Arias authored a regional plan that served as the basis for the Esquipulas Peace Agreement and Arias was awarded the 1987 Nobel Peace Prize for his work. Arias also promoted change in the USSR-backed Nicaraguan government of the era. Costa Rica also hosted several rounds of negotiations between the Salvadoran Government and the Farabundo Martí National Liberation Front (FMLN), aiding El Salvador's efforts to emerge from civil war and culminating in that country's 1994 free and fair elections. Costa Rica has been a strong proponent of regional arms-limitation agreements. Former President Miguel Ángel Rodríguez recently proposed the abolition of all Central American militaries and the creation of a regional counternarcotics police force in their stead. With the establishment of democratically-elected governments in all Central American nations by the 1990s, Costa Rica turned its focus from regional conflicts to the pursuit of neoliberal policies on the isthmus. The influence of these policies, along with the US invasion of Panama, was instrumental in drawing Panama into the Central American model of neoliberalism. Costa Rica also participated in the multinational Partnership for Democracy and Development in Central America. Regional political integration has not proven attractive to Costa Rica. The country debated its role in the Central American integration process under former President Calderon. Costa Rica has sought concrete economic ties with its Central American neighbors rather than the establishment of regional political institutions, and it chose not to join the Central American Parliament. Costa Rica has been an active member of the United Nations since its inception at the San Francisco Conference in 1945. Its first ambassador to the United Nations was Fernando Soto Harrison, the Secretary of Governance under President Picado. Costa Rican Christiana Figueres was nominated for post of UN Secretary General on July 2016. http://www.mea.gov.in/Portal/ForeignRelation/Costa_Rica-India_Bilateral-Jan_2013.pdf
https://en.wikipedia.org/wiki?curid=5560
Computational linguistics Computational linguistics is an interdisciplinary field concerned with the statistical or rule-based modeling of natural language from a computational perspective, as well as the study of appropriate computational approaches to linguistic questions. Traditionally, computational linguistics was performed by computer scientists who had specialized in the application of computers to the processing of a natural language. Today, computational linguists often work as members of interdisciplinary teams, which can include regular linguists, experts in the target language, and computer scientists. In general, computational linguistics draws upon the involvement of linguists, computer scientists, experts in artificial intelligence, mathematicians, logicians, philosophers, cognitive scientists, cognitive psychologists, psycholinguists, anthropologists and neuroscientists, among others. Computational linguistics has both theoretical and applied components. Theoretical computational linguistics focuses on issues in theoretical linguistics and cognitive science and applied computational linguistics focuses on the practical outcome of modeling human language use. The Association for Computational Linguistics defines computational linguistics as: Computational linguistics is often grouped within the field of artificial intelligence but was present before the development of artificial intelligence. Computational linguistics originated with efforts in the United States in the 1950s to use computers to automatically translate texts from foreign languages, particularly Russian scientific journals, into English. Since computers can make arithmetic (systematic) calculations much faster and more accurately than humans, it was thought to be only a short matter of time before they could also begin to process language. Computational and quantitative methods are also used historically in the attempted reconstruction of earlier forms of modern languages and sub-grouping modern languages into language families. Earlier methods, such as lexicostatistics and glottochronology, have been proven to be premature and inaccurate. However, recent interdisciplinary studies that borrow concepts from biological studies, especially gene mapping, have proved to produce more sophisticated analytical tools and more reliable results. When machine translation (also known as mechanical translation) failed to yield accurate translations right away, automated processing of human languages was recognized as far more complex than had originally been assumed. Computational linguistics was born as the name of the new field of study devoted to developing algorithms and software for intelligently processing language data. The term "computational linguistics" itself was first coined by David Hays, a founding member of both the Association for Computational Linguistics (ACL) and the International Committee on Computational Linguistics (ICCL). To translate one language into another, it was observed that one had to understand the grammar of both languages, including both morphology (the grammar of word forms) and syntax (the grammar of sentence structure). To understand syntax, one had to also understand the semantics and the lexicon (or 'vocabulary'), and even something of the pragmatics of language use. Thus, what started as an effort to translate between languages evolved into an entire discipline devoted to understanding how to represent and process natural languages using computers. Nowadays research within the scope of computational linguistics is done at computational linguistics departments, computational linguistics laboratories, computer science departments, and linguistics departments. Some research in the field of computational linguistics aims to create working speech or text processing systems while others aim to create a system allowing human-machine interaction. Programs meant for human-machine communication are called conversational agents. Just as computational linguistics can be performed by experts in a variety of fields and through a wide assortment of departments, so too can the research fields broach a diverse range of topics. The following sections discuss some of the literature available across the entire field broken into four main area of discourse: developmental linguistics, structural linguistics, linguistic production, and linguistic comprehension. Language is a cognitive skill that develops throughout the life of an individual. This developmental process has been examined using several techniques, and a computational approach is one of them. Human language development does provide some constraints which make it harder to apply a computational method to understanding it. For instance, during language acquisition, human children are largely only exposed to positive evidence. This means that during the linguistic development of an individual, the only evidence for what is a correct form is provided, and no evidence for what is not correct. This is insufficient information for a simple hypothesis testing procedure for information as complex as language, and so provides certain boundaries for a computational approach to modeling language development and acquisition in an individual. Attempts have been made to model the developmental process of language acquisition in children from a computational angle, leading to both statistical grammars and connectionist models. Work in this realm has also been proposed as a method to explain the evolution of language through history. Using models, it has been shown that languages can be learned with a combination of simple input presented incrementally as the child develops better memory and longer attention span. This was simultaneously posed as a reason for the long developmental period of human children. Both conclusions were drawn because of the strength of the artificial neural network which the project created. The ability of infants to develop language has also been modeled using robots in order to test linguistic theories. Enabled to learn as children might, a model was created based on an affordance model in which mappings between actions, perceptions, and effects were created and linked to spoken words. Crucially, these robots were able to acquire functioning word-to-meaning mappings without needing grammatical structure, vastly simplifying the learning process and shedding light on information which furthers the current understanding of linguistic development. It is important to note that this information could only have been empirically tested using a computational approach. As our understanding of the linguistic development of an individual within a lifetime is continually improved using neural networks and learning robotic systems, it is also important to keep in mind that languages themselves change and develop through time. Computational approaches to understanding this phenomenon have unearthed very interesting information. Using the Price Equation and Pólya urn dynamics, researchers have created a system which not only predicts future linguistic evolution but also gives insight into the evolutionary history of modern-day languages. This modeling effort achieved, through computational linguistics, what would otherwise have been impossible. It is clear that the understanding of linguistic development in humans as well as throughout evolutionary time has been fantastically improved because of advances in computational linguistics. The ability to model and modify systems at will affords science an ethical method of testing hypotheses that would otherwise be intractable. To create better computational models of language, an understanding of language's structure is crucial. To this end, the English language has been meticulously studied using computational approaches to better understand how the language works on a structural level. One of the most important pieces of being able to study linguistic structure is the availability of large linguistic corpora or samples. This grants computational linguists the raw data necessary to run their models and gain a better understanding of the underlying structures present in the vast amount of data which is contained in any single language. One of the most cited English linguistic corpora is the Penn Treebank. Derived from widely-different sources, such as IBM computer manuals and transcribed telephone conversations, this corpus contains over 4.5 million words of American English. This corpus has been primarily annotated using part-of-speech tagging and syntactic bracketing and has yielded substantial empirical observations related to language structure. Theoretical approaches to the structure of languages have also been developed. These works allow computational linguistics to have a framework within which to work out hypotheses that will further the understanding of the language in a myriad of ways. One of the original theoretical theses on the internalization of grammar and structure of language proposed two types of models. In these models, rules or patterns learned increase in strength with the frequency of their encounter. The work also created a question for computational linguists to answer: how does an infant learn a specific and non-normal grammar (Chomsky Normal Form) without learning an overgeneralized version and getting stuck? Theoretical efforts like these set the direction for research to go early in the lifetime of a field of study, and are crucial to the growth of the field. Structural information about languages allows for the discovery and implementation of similarity recognition between pairs of text utterances. For instance, it has recently been proven that based on the structural information present in patterns of human discourse, conceptual recurrence plots can be used to model and visualize trends in data and create reliable measures of similarity between natural textual utterances. This technique is a strong tool for further probing the structure of human discourse. Without the computational approach to this question, the vastly complex information present in discourse data would have remained inaccessible to scientists. Information regarding the structural data of a language is available for English as well as other languages, such as Japanese. Using computational methods, Japanese sentence corpora were analyzed and a pattern of log-normality was found in relation to sentence length. Though the exact cause of this lognormality remains unknown, it is precisely this sort of information which computational linguistics is designed to uncover. This information could lead to further important discoveries regarding the underlying structure of Japanese and could have any number of effects on the understanding of Japanese as a language. Computational linguistics allows for very exciting additions to the scientific knowledge base to happen quickly and with very little room for doubt. In recent days, the structural data of languages is available for several languages of the world other than English language. Computational linguistics work is in progress on Sindhi language because the structure, grammar and domain of Sindhi language is different from the other languages of the World. The computational linguistics models for English language are not suitable for Sindhi language. Viewing this, computational linguistics work on Sindhi language has been started properly by developing methods, algorithms, linguistics tools (https://sindhinlp.com/), machine learning models and deep learning models since 2016 to focus and solve the linguistics problems of Sindhi language. This work could lead to further important discoveries regarding the underlying structure of Sindhi, and could have any number of effects on the understanding of Sindhi as a language. Computational linguistics allows for very exciting additions to the scientific knowledge base to happen quickly and with very little room for doubt. Without a computational approach to the structure of linguistic data, much of the information that is available now would still be hidden under the vastness of data within any single language. Computational linguistics allows scientists to parse huge amounts of data reliably and efficiently, creating the possibility for discoveries unlike any seen in most other approaches. The production of language is equally as complex in the information it provides and the necessary skills which a fluent producer must have. That is to say, comprehension is only half the problem of communication. The other half is how a system produces language, and computational linguistics has made interesting discoveries in this area. In a now famous paper published in 1950 Alan Turing proposed the possibility that machines might one day have the ability to "think". As a thought experiment for what might define the concept of thought in machines, he proposed an "imitation test" in which a human subject has two text-only conversations, one with a fellow human and another with a machine attempting to respond like a human. Turing proposes that if the subject cannot tell the difference between the human and the machine, it may be concluded that the machine is capable of thought. Today this test is known as the Turing test and it remains an influential idea in the area of artificial intelligence. Some projects are still trying to solve the problem which first started computational linguistics off as its field in the first place. However, methods have become more refined, and consequently, the results generated by computational linguists have become more enlightening. To improve computer translation, several models have been compared, including hidden Markov models, smoothing techniques, and the specific refinements of those to apply them to verb translation. The model which was found to produce the most natural translations of German and French words was a refined alignment model with a first-order dependence and a fertility model. They also provide efficient training algorithms for the models presented, which can give other scientists the ability to improve further on their results. This type of work is specific to computational linguistics and has applications that could vastly improve understanding of how language is produced and comprehended by computers. Work has also been done in making computers produce language in a more naturalistic manner. Using linguistic input from humans, algorithms have been constructed which are able to modify a system's style of production based on a factor such as linguistic input from a human, or more abstract factors like politeness or any of the five main dimensions of personality. This work takes a computational approach via parameter estimation models to categorize the vast array of linguistic styles we see across individuals and simplify it for a computer to work in the same way, making human-computer interaction much more natural. Many of the earliest and simplest models of human-computer interaction, such as ELIZA for example, involve a text-based input from the user to generate a response from the computer. By this method, words typed by a user trigger the computer to recognize specific patterns and reply accordingly, through a process known as keyword spotting. Recent technologies have placed more of an emphasis on speech-based interactive systems. These systems, such as Siri of the iOS operating system, operate on a similar pattern-recognizing technique as that of text-based systems, but with the former, the user input is conducted through speech recognition. This branch of linguistics involves the processing of the user's speech as sound waves and the interpreting of the acoustics and language patterns for the computer to recognize the input. Much of the focus of modern computational linguistics is on comprehension. With the proliferation of the internet and the abundance of easily accessible written human language, the ability to create a program capable of understanding human language would have many broad and exciting possibilities, including improved search engines, automated customer service, and online education. Early work in comprehension included applying Bayesian statistics to the task of optical character recognition, as illustrated by Bledsoe and Browing in 1959 in which a large dictionary of possible letters was generated by "learning" from example letters and then the probability that any one of those learned examples matched the new input was combined to make a final decision. Other attempts at applying Bayesian statistics to language analysis included the work of Mosteller and Wallace (1963) in which an analysis of the words used in "The Federalist Papers" was used to attempt to determine their authorship (concluding that Madison most likely authored the majority of the papers). In 1971 Terry Winograd developed an early natural language processing engine capable of interpreting naturally written commands within a simple rule-governed environment. The primary language parsing program in this project was called SHRDLU, which was capable of carrying out a somewhat natural conversation with the user giving it commands, but only within the scope of the toy environment designed for the task. This environment consisted of different shaped and colored blocks, and SHRDLU was capable of interpreting commands such as "Find a block which is taller than the one you are holding and put it into the box." and asking questions such as "I don't understand which pyramid you mean." in response to the user's input. While impressive, this kind of natural language processing has proven much more difficult outside the limited scope of the toy environment. Similarly, a project developed by NASA called LUNAR was designed to provide answers to naturally written questions about the geological analysis of lunar rocks returned by the Apollo missions. These kinds of problems are referred to as question answering. Initial attempts at understanding spoken language were based on work done in the 1960s and 1970s in signal modeling where an unknown signal is analyzed to look for patterns and to make predictions based on its history. An initial and somewhat successful approach to applying this kind of signal modeling to language was achieved with the use of hidden Markov models as detailed by Rabiner in 1989. This approach attempts to determine probabilities for the arbitrary number of models that could be being used in generating speech as well as modeling the probabilities for various words generated from each of these possible models. Similar approaches were employed in early speech recognition attempts starting in the late 70s at IBM using word/part-of-speech pair probabilities. More recently these kinds of statistical approaches have been applied to more difficult tasks such as topic identification using Bayesian parameter estimation to infer topic probabilities in text documents. Modern computational linguistics is often a combination of studies in computer science and programming, math, particularly statistics, language structures, and natural language processing. Combined, these fields most often lead to the development of systems that can recognize speech and perform some task based on that speech. Examples include speech recognition software, such as Apple's Siri feature, spellcheck tools, speech synthesis programs, which are often used to demonstrate pronunciation or help the disabled, and machine translation programs and websites, such as Google Translate. Computational linguistics can be especially helpful in situations involving social media and the Internet. For example, filters in chatrooms or on website searches require computational linguistics. Chat operators often use filters to identify certain words or phrases and deem them inappropriate so that users cannot submit them. Another example of using filters is on websites. Schools use filters so that websites with certain keywords are blocked from children to view. There are also many programs in which parents use Parental controls to put content filters in place. Computational linguists can also develop programs that group and organize content through Social media mining. An example of this is Twitter, in which programs can group tweets by subject or keywords. Computational linguistics is also used for document retrieval and clustering. When doing an online search, documents and websites are retrieved based on the frequency of unique labels related to what was typed into a search engine. For instance, if a person searches "red, large, four-wheeled vehicle," to find pictures of a red truck, the search engine will still find the information desired by matching words such as "four-wheeled" with "car". Computational approaches are also important in the study of change over time. This occurs in several ways: through the computational modeling of language families (see further Quantitative comparative linguistics or phylogenetics), through the modeling of changes in sound and meaning Computational linguistics can be divided into major areas depending upon the medium of the language being processed, whether spoken or textual; and upon the task being performed, whether analyzing language (recognition) or synthesizing language (generation). Speech recognition and speech synthesis deal with how spoken language can be understood or created using computers. Parsing and generation are sub-divisions of computational linguistics dealing respectively with taking language apart and putting it together. Machine translation remains the sub-division of computational linguistics dealing with having computers translate between languages. The possibility of automatic language translation, however, has yet to be realized and remains a notoriously hard branch of computational linguistics. Some of the areas of research that are studied by computational linguistics include: The subject of computational linguistics has had a recurring impact on popular culture:
https://en.wikipedia.org/wiki?curid=5561
Demographics of Ivory Coast This article is about the demographic features of the population of Ivory Coast, including population density, ethnicity, education level, health of the populace, economic status, religious affiliations and other aspects of the population. According to the total population was in , compared to only 2 630 000 in 1950. The proportion of children below the age of 15 in 2010 was 40.9%, 55.3% was between 15 and 65 years of age, while 3.8% was 65 years or older . Structure of the population (DHS 2011-2012) (Males 23 630, Females 25 193 = 48 823) : Population Growth: 1.88% (2016 est.) Registration of vital events is in Ivory Coast not complete. The Population Departement of the United Nations prepared the following estimates. Total Fertility Rate (TFR) (Wanted Fertility Rate) and Crude Birth Rate (CBR): Fertility data as of 2011-2012 (DHS Program): Ivory Coast has more than 60 ethnic groups, usually classified into five principal divisions: Akan (east and center, including Lagoon peoples of the southeast), Krou (southwest), Southern Mandé (west), Northern Mandé (northwest), Sénoufo/Lobi (north center and northeast). The Baoulés, in the Akan division, probably comprise the largest single subgroup with 15%-20% of the population. They are based in the central region around Bouaké and Yamoussoukro. The Bétés in the Krou division, the Sénoufos in the north, and the Malinkés in the northwest and the cities are the next largest groups, with 10%-15% of the national population. Most of the principal divisions have a significant presence in neighboring countries. Of the more than 5 million non-Ivorian Africans living in Ivory Coast, one-third to one-half are from Burkina Faso; the rest are from Ghana, Guinea, Mali, Nigeria, Benin, Senegal, Liberia, and Mauritania. The non-African expatriate community includes roughly 50,000 French (this number may be inaccurate due to the evacuation of roughly 8,000 Frenchmen in November 2004) and possibly 40,000 Lebanese. The number of elementary school-aged children attending classes increased from 22% in 1960 to 67% in 1995. French is official, and there are 81 living indigenous languages, and two that are now extinct (Tonjon and Beti). The Dioula dialect of Bambara is the most widely spoken one. Other language groups include the Gur languages, the Senufo languages, the Kru languages (including the Bété languages, Dida, Nyabwa, Wè, and Western Krahn), and the Kwa languages (Baoulé and Anyin are the most used). The economic development and relative prosperity of Ivory Coast fostered huge demographic shifts during the 20th century. "In 1922, an estimated 100,000 out of 1.6 million (or 6 percent) of people in Côte d'Ivoire were Muslims. By contrast, at independence (in 1960), their share of the population had increased rapidly, and Muslims were moving southward to the cocoa-producing areas and the southern cities. By 1998, [...], Muslims constituted a majority in the north of the country, and approximately 38.6 percent of the total population. This was a significantly larger population than the next largest religious group, Christians, who constituted approximately 29.1 percent of the total." In earlier decades, this shift was mainly due to large-scale immigration from neighboring countries of the interior, that has been going on since colonial times and continued to be promoted during the Houphouet-Boigny era. Since the 1990s, the widening fertility gap between different religious groups has continued to tilt the demographic balance in favor of Muslims although immigration has become less important. The table below shows the number of people born in Ivory Coast who have migrated to OECD countries only (the table only includes communities consisting of at least 1,000 members). Demographic statistics according to the World Population Review in 2019. The following demographic are from the CIA World Factbook unless otherwise indicated. Note: highly pathogenic H5N1 avian influenza has been identified in this country; it poses a negligible risk with extremely rare cases possible among US citizens who have close contact with birds (2009) Noun and adjective: Ivorian (Ivoirian) definition: age 15 and over can read and write (2015 est.)
https://en.wikipedia.org/wiki?curid=5566
Politics of Ivory Coast The Politics of Ivory Coast takes place in a framework of a presidential representative democratic republic, whereby the President of Ivory Coast is both head of state and head of government, and of a multi-party system. Executive power is exercised by the President and the Government. Legislative power is vested in both the government and parliament. The capital since 1983 is Yamoussoukro; however, Abidjan remains the commercial center. Most countries maintain their embassies in Abidjan. A civil war was fought in Ivory Coast between 2002–2004 and a second civil war broke out in March 2011 following contested elections that saw president-elect Alassane Ouattara come into power in 2011 and reelected in 2015. Troops, mostly hailing from the north of the country, mutinied in the early hours of September 19, 2002. They soon after launched attacks in many cities, including Abidjan. By lunchtime, they had control of the north of the country. Their principal claim relates to the definition of who is a citizen of Ivory Coast (and so who can stand for election as president), voting rights and their representation in government in Abidjan. The events in Abidjan shows that it is not a tribal issue, but a crisis of transition from a dictatorship to a democracy, with the clashes inherent in the definition of citizenship. Forces involved in the conflict include: Ivory Coast's 1959 constitution provides for strong presidency within the framework of a separation of powers. The executive is personified in the president, elected for a five-year term. The president is commander in chief of the armed forces, may negotiate and ratify certain treaties, and may submit a bill to a national referendum or to the National Assembly. According to the constitution, the President of the National Assembly assumes the presidency in the event of a vacancy, and he completes the remainder of the deceased president's term. The cabinet is selected by and is responsible to the president. Changes are being proposed to some of these provisions, to extend term of office to 7 years, establish a senate, and make president of the senate interim successor to the president. Laurent Gbagbo took power following a popular overthrow of the interim leader Gen. Robert Guéï who had claimed a dubious victory in presidential elections; Gen. Guéï himself had assumed power on 25 December 1999, following a military coup against the government of former President Henri Konan Bédié. Gbagbo was elected president in 2000 in an election boycotted by many oppositional forces. The president is elected by popular vote for a five-year term. The prime minister is appointed by the president.Alassane Ouattara is currently the president of Ivory Coast. He was reelected in the 2015 Ivorian presidential election. After a new constitution was approved by referendum, it is expected President Alassane Ouattara would appoint a Vice-President before 2020. The President and Vice-President will run on a joint ticket from 2020. They will be both elected for a five-year term, with only one possible reelection. The Vice-President will replace the President in case of death, resignation and any other vacancy, . Parliament of Ivory Coast is a bicameral body composed by the National Assembly and the Senate of Ivory Coast. Prior to November 2016 and the future creation of the Senate, the Parliament of Ivory Coast was only composed of the National Assembly. The National Assembly ("Assemblée Nationale") has 255 members, elected for a five-year term in single-seat constituencies. It passes on legislation typically introduced by the president although it also can introduce legislation. The Senate of Ivory Coast ("Sénat") will have two third of senators indirectly elected and one third appointed by the president-elect, elected five-year term in single-seat constituencies. Ivory Coast is a one party dominant state with the Rally of the Republicans in power. The judicial system culminates in the Supreme Court of Ivory Coast. The High Court of Justice is competent to try government officials for major offenses. The Supreme Court or Court Supreme consists of four chambers: Judicial Chamber for criminal cases, Audit Chamber for financial cases, Constitutional Chamber for judicial review cases, and Administrative Chamber for civil cases; there is no legal limit to the number of members. For administrative purposes, Ivory Coast is divided into 58 departments, each headed by a prefect appointed by the central government. There are 196 communes, each headed by an elected mayor, plus the city of Abidjan with ten mayors. The 58 departments ("départements", singular - "département") are listed in the article Departments of Ivory Coast. ACP, AfDB, ECA, ECOWAS, Entente, FAO, FZ, G-24, G-77, IAEA, IBRD, ICAO, ICCt (signatory), ICRM, IDA, IDB, IFAD, IFC, IFRCS, ILO, IMF, IMO, Interpol, IOC, IOM, ISO, ITU, ITUC, MONUC, NAM, OAU, OIC, OPCW, UN, UNCTAD, UNESCO, UNHCR, UNIDO, UPU, WADB (regional), WAEMU, WCO, WFTU, WHO, WIPO, WMO, WToO, WTrO
https://en.wikipedia.org/wiki?curid=5567
Transport in Ivory Coast Ivory Coast invested remarkably in its transport system. Transport Infrastructures are much more developed than they are other West African countries despite a crisis that restrained their maintenance and development. Since its independence in 1960, Ivory Coast put an emphasis on increasing and modernizing the transport network for human as well as for goods. Major infrastructures of diverse nature were built including railways, roads, waterways, and airports. In spite of the crisis, neighbor countries (Burkina Faso, Mali, Niger, and Guinea) still strongly depend on the Ivorian transport network for importing, exporting, and transiting their immigrants to Ivory Coast. The nation's railway system is part of a 1 260 km long route that links the country to Burkina Faso and Niger. 1 156 km of railroad links Abidjan to Ouagadougou, capital of Burkina Faso. Built during colonial era by the firm Abidjan-Niger (RAN), this railroad freed several landlocked countries among which were ex-High-Volta (Burkina Faso), Niger, and Mali. This railroad, operated by Sitarail, plays a key role as regards to the carriage of the goods (livestock) and the transport of people between Ivory Coast and border countries: 1 million tons of goods have transited in 2006. In 2005, despite the negative impact the crisis had on the sector, benefits engendered by transporting the goods and people via RAN, are estimated respectively at 16 309 et3 837billionCFA. As of 2004, the railway network consisted of a state-controlled 660 km section of a 1,146 km narrow gauge railroad that ran north from Abidjan through Bouaké and Ferkéssédougou to Ouagadougou, Burkina Faso. Ivory Coast road network spreads over 85 000 km consisting of 75 000 unpaved, 65 000 km, and 224 km highways. It provides national and international traffic with neighbor countries. The Trans–West African Coastal Highway provides a paved link to Ghana, Togo, Benin and Nigeria, with paved highways to landlocked Mali and Burkina Faso feeding into the coastal highway. When construction of roads and bridges in Liberia and Sierra Leone is complete, the highway will link to another seven Economic Community of West African States (ECOWAS) nations to the west and north-west. At the national level, vehicles are estimated at 600 000, which includes 75% of used cars (second hand) due to the low purchasing power since the beginning of the economic crisis. 20 000 new cars are registered every year. Although maintenance and renovations works are being carried out since middle-2011, over 80% of the Ivorian network is older than 20 years and therefore damaged. In addition, a significant traffic exists throughout Abidjan, the capital. This traffic is mainly composed of taxi, buses and mini-buses locally referred to as "Gbaka." The country counts with two 4-laned motorways, the first one running from Abidjan to Yamoussoukro for a length of 224 km., and the second joining Abidjan to Grand-Bassam, with a length of 30 km. Both are built with modern technologies and under international standards of security. Landscape view of the Autonomous Port of Abidjan Ivory Coast greatly contributed to developing maritime transport by building two ports on its seaside namely, autonomous port of Abidjan, sometimes referred to as ""lung of Ivorian economy"", and the San-Pedro port. The total traffic in 2005, by adding importation to exportation, was 18 661 784 tons for autonomous port of Abidjan and 1 001 991 tons for San-Pedro. Harbor activity is concentrated at Abidjan (West Africa's largest container port), which has facilities that include a fishing port and equipment for handling containers. The autonomous port of Abidjan cover a 770 hectares area and shelters 60% of the country industries. It is the first tuna fishing port in Africa. It contains 36 conventional berths spread over six kilometers of quays providing a capacity of sixty commercial ships with multiple special docks, a container terminal as well as several specialized and industrial berths. The other major port, the San-Pedro port, operates since 1971 and has two quays covering 18,727 m2 area. Apart from those two major ports, there are also small ports at Sassandra, Aboisso, and Dabou. Ivory Coast has three international airports located in Abidjan, Yamoussoukro, and Bouaké. Fourteen smaller cities also possess regional airports, the most important of which are Daloa, Korhogo, Man, Odiénné et San-pédro. Twenty-seven aerodromes exists and are operated by a public establishment, the Anam (National agency for civil aviation and meteorology), except the activities carried out by the Asecna (Agency for security of air fret in Africa and Madagascar). Since the outbreak of the crisis, only five of these airports are available. These are Abidjan, San-Pédro, Yamoussoukro, Daloa, and Touba. Regarding the International Airport of Abidjan, official statistics from 2005, showed 14 257 commercial movements (departures and arrivals); 745 180 commercial passengers (arrivals, departures, and transit) and 12 552 tons of commercial fret. The Airport of Abidjan covers 90% of the air traffic of Côte d'Ivoire and generate 95% of the overall profits of the sector. The airport of Abidjan is operated by a private company, Aeria, created in association with the Commerce Chamber of Marseilles. Its traffic mainly encompasses European aeronautical companies (Air France, Brussels Airlines) and some African firms (South African Airways, Kenya Airways, Air Sénégal International).
https://en.wikipedia.org/wiki?curid=5570
Armed Forces of the Republic of Ivory Coast The Republican Forces of Côte d'Ivoire (; "FRCI") is the current name of the armed forces of Ivory Coast. The Ivorian military has its roots in the colonial armed forces of French West Africa, which were headquartered in Dakar, Senegal but possessed bases in several distinct military regions. Most Ivorian recruits who joined the colonial army were assigned to Senegalese units during this period. They served with distinction during both world wars, with 20,000 Ivorian soldiers fighting for the French during World War I and another 30,000 during World War II. In 1950, the French government began the process of setting up a specific defence force for the colony, consisting of four infantry companies and a light armoured unit. The Ivory Coast became independent on 7 August 1960. In April 1961, the new government signed the Franco-Ivorian Technical Military Assistance Accord with France, which compelled the latter to assist with the formation of a new national military. It also authorised the continued presence of French troops based in Port-Bouët, and permitted the government to call on French military assistance in the event of external aggression or major internal unrest. By the end of 1962, the fledgling Ivorian armed forces had expanded rapidly into 5,000 soldiers attached to four battalions. Most of the initial recruits were drawn from the defunct colonial military establishment and had served in various French units, particularly the marine regiments. They were armed with old equipment donated by France, including two Max Holste Broussard monoplanes, a single Douglas DC-3 cargo aircraft, fifteen M8 Greyhound armoured cars, and even a SC-497-class submarine chaser. Conscription was instituted, although the large number of volunteers and low manpower requirements ensured it was only applied selectively. Some of the senior positions in the officer corps and Ministry of Defence continued to be held by French nationals. Since the Ivory Coast could ill afford to divert funds from its economic development programmes into the armed forces, and was already dependent on France for its external defence, the military establishment remained quite modest from 1961 to 1974. Defence spending spiralled upwards between 1974 and 1987, and the number of personnel serving with the armed forces increased to 14,920 men. During this period, the air force and navy embarked on a significant modernisation campaign. An international merchant marine training academy was built in Abidjan and trained personnel from several Economic Community of West African States (ECOWAS) governments. In 1997, a collapse in civil-military relations became evident when President Henri Konan Bédié dismissed popular general Robert Guéï on suspicion of disloyalty. Two years later, an army mutiny led by disgruntled recruits and junior officers escalated into a major coup d'état which ousted Bédié and installed Guéï in his place. Guéï subsequently stood for office during a subsequent presidential election, although he attempted to annul the election results when Laurent Gbagbo secured the popular vote. This triggered a civil revolt in Abidjan and two days of street battles between Gbagbo supporters and soldiers loyal to Guéï. Most of the armed forces remained neutral until the third day, when the army's elite units and the gendarmerie announced they would recognise Gbagbo as president of the republic. Guéï conceded defeat, going into exile on October 29, 2000. In September 2002, the Ivory Coast endured a second army mutiny, this time by 750 Muslim soldiers who seized Bouaké, citing religious discrimination and grievances against the predominantly Christian government. The mutineers later took control of most of the northern administrative regions, carrying out a brutal campaign of ethnic cleansing and plunging the country into civil war. For a number of years, troops dispatched by France, ECOWAS, and a United Nations Operation in Côte d'Ivoire (ONUCI) effort enforced a buffer zone between the south and the rebel-held north. President Gbagbo repeatedly demanded France assist him in crushing the rebel forces. France maintained it would not take sides in the civil war, but allowed Ivorian military aircraft to cross the buffer zone and attack rebel positions. In November 2004, an Ivorian pilot targeted a French base during an air strike on Bouaké, killing nine French soldiers. The French retaliated by launching a follow-up operation to destroy the Ivorian Air Force. In March 2011, a rebel coalition, the "Forces Nouvelles de Côte d'Ivoire", launched a renewed offensive on the south with French support, sparking a second civil war. The Ivorian army was quickly overwhelmed, and Gbagbo deposed by the rebels. The "Forces Nouvelles" established a new national military, known as the Republican Forces of Ivory Coast (FRCI). Integration problems arising from the incorporation of various rebel factions into the FRCI, as well as former Gbagbo loyalists, continue to persist. In 2014, some army units launched an abortive mutiny over wage disputes. The crisis ended when the Ivorian political leadership agreed to a new financial settlement with the FRCI. A second mutiny occurred on January 7, 2017, with troops in Bouaké demanding higher salaries and improved living conditions; this resulted in a second financial settlement. The Ivorian army had three infantry battalions, an armoured battalion, an artillery battery, and seven specialist companies in 1993. The effective strength of the army was about 3,000 troops for the first ten years of Ivorian independence, increasing to over 8,000 in the mid-1980s before declining steadily to about 5,500. It has always remained the largest branch of the armed forces. In 1987, the army was responsible for the country's five military regions, each of which was supervised by a colonel. The First Military Region controlled the concentration of forces in and around Abidjan, its principal units there being a rapid intervention battalion (airborne), an infantry battalion, an armored battalion, and an air defense artillery battalion. The Second Military Region was located in Daloa and comprised one infantry battalion. The Third Military Region was headquartered in Bouaké and was home to an artillery, an infantry, and an engineer battalion. The Fourth Military Region maintained only a Territorial Defense Company headquartered in Korhogo The Fifth Military Region was formerly known as the Western Operational Zone, a temporary command created to respond to the security threat caused by the First Liberian Civil War. By 2010, the system of military regions had been abolished. As of July 2011, General Soumaïla Bakayoko is the chief of staff of the army, and colonel-major Gervais Kouakou Kouassi is the Chief of the Gendarmerie. As of October 2011, previously active units around Abidjan reportedly included the: The 2nd Infantry Battalion appears to have been based at Daloa for some time. A 2003 change of command ushered in the 16th commander of the unit, and there are also reports from 2009 and 2011. Reported special forces units include: The Ivorian army has traditionally been equipped with French weapons, most of which were delivered in the 1980s under generous military grants from Paris. During Laurent Ghagbo's administration, large quantities of second-hand Soviet arms were acquired from Angola, Ukraine, and Belarus. After achieving independence from France in 1960, Ivory Coast maintained strong links with France through bilateral defence agreements. French training and operating techniques has been used since the establishment of the air force. The first equipment supplied included three Douglas C-47's and seven MH.1521 Broussard STOL utility aircraft in 1961. The first jet aircraft to enter service in October 1980 were six Alpha Jet CI light attack and advanced training aircraft; six more were ordered, but were subsequently cancelled. However, another was purchased in 1983. The 1979 air force had only transport and liaison aircraft. In 1987, the Library of Congress Country Study said that the Air Force's official name, Ivoirian Air Transport and Liaison Group ("Groupement Aérien de Transport et de Liaison"—GATL), 'reflects an original mission focused more on logistics and transport rather than a combat force.' In 2004, following an air strikes on French peacekeepers by Ivorian forces, the French military destroyed all aircraft in the Air Force of Ivory Coast. Gbagbo had ordered air strikes on Ivorian rebels. On 6 November 2004, at least one Ivorian Sukhoi Su-25 bomber attacked a French peacekeeping position in the rebel town of Bouaké at 1 pm, killing nine French soldiers and wounding 31. An American development worker, reported to have been a missionary, was also killed. The Ivorian government claimed the attack on the French was unintentional, but the French insisted that the attack had been deliberate. Several hours after the attack French President Jacques Chirac ordered the destruction of the Ivorian air force and the seizure of Yamoussoukro airport. The French military performed an overland attack on the airport, destroying two Sukhoi Su-25 ground attack aircraft and three Mi-24 helicopter gunships. Two more military helicopters were destroyed during combat in the skies over Abidjan. France then flew in 300 troops and put three Dassault Mirage F1 jet fighters based in nearby Gabon on standby. Since then, the Air Force of Ivory Coast has been rebuilt. In 2007, Aviation Week & Space Technology reported a total of six aircraft in service: one Antonov An-32 tactical transport, one Cessna 421 Golden Eagle utility aircraft, two Eurocopter SA 365 Dauphin helicopters, one Gulfstream IV VIP aircraft, and one Mil Mi-24 attack helicopter. It is unknown whether any of these aircraft were truly operational. In addition, Deagel.com reported two Mikoyan-Gurevich MiG-23 attack aircraft. Ivory Coast has a brown-water navy whose mission is coastal surveillance and security for the nation's 340-mile coastline. The operational capability of the navy was severely degraded due to the diversion of resources to the army and air force during the civil wars, and it remains incapable of conducting operations beyond the general vicinity of Abidjan. Up to the 1990s the navy had 5 known vessels: Early vessels where second hand: In the late 1970s new vessels were acquired: Beginning in 2014 the navy was delivered the first of three coastal defence vessels. In 2018 it was reported the navy was purchasing more vessels from French builder Raidco Marine (30 inflatable boats and 10 patrol craft), but for use by police and gendarmerie. A mutual defense accord signed with France in April 1961 provides for the stationing of French Armed Forces troops in Ivory Coast. The 43rd Marine Infantry Battalion of the French Army's Troupes de Marine () was based in Port Bouet adjacent to the Abidjan Airport from 1979 and had more than 500 troops assigned until 2011, when it appears to have been disbanded. The French military also maintains a force as part of Operation Licorne. From summer 2011, Operation Licorne, the French force, previously over 5,000 strong, is roughly 700, and consists of Licorne headquarters, Battalion Licorne (BATLIC), seemingly made up of elements of the 2nd Marine Infantry Regiment and the Régiment d'infanterie-chars de marine, and a helicopter detachment. The United Nations has maintained the peacekeeping mission ONUCI in the country since 2004. On 28 February 2011 ONUCI consisted of 7,568 troops, 177 military observers, and numerous international civilians and Police; the mission had received helicopter and infantry reinforcement from UNMIL during the stand-off since the late 2010 elections which had been won by Alassane Ouattara. Since independence, the Ivory Coast has maintained a paramilitary gendarmerie force with a mandate to assist the police with law enforcement duties in the country's rural districts. However, it may also be deployed alongside the army to quell internal unrest. For a number of decades, the size of the Ivorian National Gendarmerie remained consistent at around 4,000 to 5,000 personnel, supervised by a commandant. It underwent a massive expansion following the outbreak of the First Ivorian Civil War, increasing to about 12,000 personnel commanded by a major general. Gendarmes undergo training as cadets at a National Gendarmerie Academy. The National Gendarmerie maintains an investigative branch, the "Brigades de Recherches", which has been accused of various human rights abuses, including extrajudicial killings and unlawful detention.
https://en.wikipedia.org/wiki?curid=5571
Croatia Croatia (, ; , ), officially the Republic of Croatia (, ), is a country in Southeast Europe. It borders Slovenia to the northwest, Hungary to the northeast, Serbia to the east, Bosnia and Herzegovina, and Montenegro to the southeast, sharing a maritime border with Italy. Its capital, Zagreb, forms one of the country's primary subdivisions, along with twenty counties. Croatia has an area of and a population of 4.07 million, most of whom are Roman Catholics. Croatia has been inhabited since the Paleolithic Age. The Croats arrived in the area in the 6th century and organised the territory into two duchies by the 9th century. Croatia was first internationally recognized as an independent state on 7 June 879 during the reign of duke Branimir. Tomislav became the first king by 925, elevating Croatia to the status of a kingdom, which retained its sovereignty for nearly two centuries. During the succession crisis after the Trpimirović dynasty ended, Croatia entered a personal union with Hungary in 1102. In 1527, faced with Ottoman conquest, the Croatian Parliament elected Ferdinand I of Austria to the Croatian throne. In October 1918, in the final days of World War I, the State of Slovenes, Croats and Serbs, independent from Austria-Hungary, was proclaimed in Zagreb, and in December 1918 it was merged into the Kingdom of Serbs, Croats and Slovenes (renamed the Kingdom of Yugoslavia in 1929). Following the Axis invasion of Yugoslavia in April 1941, most of the Croatian territory was incorporated into a Nazi-backed client-state, the Independent State of Croatia. In response, a resistance movement developed. This led to the creation of the Federal State of Croatia, which after the war became a founding member and constituent of the Socialist Federal Republic of Yugoslavia. On 25 June 1991, Croatia declared independence, which came wholly into effect on 8 October of the same year. The Croatian War of Independence was fought successfully for four years following the declaration. A sovereign state, Croatia is a republic governed under a parliamentary system. It is a member of the European Union (EU), the United Nations (UN), the Council of Europe, NATO, the World Trade Organization (WTO), and a founding member of the Union for the Mediterranean. As an active participant in the UN peacekeeping forces, Croatia has contributed troops to the NATO-led mission in Afghanistan and took a non-permanent seat on the UN Security Council for the 2008–2009 term. Since 2000, the Croatian government has constantly invested in infrastructure, especially transport routes and facilities along the Pan-European corridors. Croatia is classified by the World Bank as a high-income country and ranked 46th on the Human Development Index. The economy is dominated by service, industrial sectors and agriculture. Tourism is a significant source of revenue, with Croatia ranked among the top 20 most popular tourist destinations in the world. The state controls a part of the economy, with substantial government expenditure. The European Union is Croatia's most important trading partner. Croatia provides social security, universal health care system, and a tuition-free primary and secondary education, while supporting culture through numerous public institutions and corporate investments in media and publishing. The name of Croatia derives from Medieval Latin "Croātia". Itself a derivation of North-West Slavic "*Xrovat-", by liquid metathesis from Common Slavic period "*"Xorvat"", from proposed Proto-Slavic "*Xъrvátъ" which possibly comes from Old Persian "*xaraxwat-". The word is attested by the Old Iranian toponym "Harahvait-" which is the native name of Arachosia. The origin of the name is uncertain but is thought to be a Gothic or Indo-Aryan term assigned to a Slavic tribe. The oldest preserved record of the Croatian ethnonym "*xъrvatъ" is of the variable stem, attested in the Baška tablet in style "zvъnъmirъ kralъ xrъvatъskъ" ("Zvonimir, Croatian king"). The first attestation of the Latin term is attributed to a charter of Duke Trpimir from the year 852. The original is lost, and just a 1568 copy is preserved, leading to doubts over the authenticity of the claim. The oldest preserved stone inscription is the 9th-century Branimir Inscription found near Benkovac, where Duke Branimir is styled "Dux Cruatorvm". The inscription is not believed to be dated accurately but is likely to be from during the period of 879–892, during Branimir's rule. The area known as Croatia today was inhabited throughout the prehistoric period. Fossils of Neanderthals dating to the middle Palaeolithic period have been unearthed in northern Croatia, with the most famous and the best presented site in Krapina. Remnants of several Neolithic and Chalcolithic cultures were found in all regions of the country. The largest proportion of the sites is in the river valleys of northern Croatia, and the most significant cultures whose presence was discovered include Baden, Starčevo, and Vučedol cultures. The Iron Age left traces of the early Illyrian Hallstatt culture and the Celtic La Tène culture. Much later, the region was settled by Illyrians and Liburnians, while the first Greek colonies were established on the islands of Hvar, Korčula, and Vis. In 9 AD the territory of today's Croatia became part of the Roman Empire. Emperor Diocletian was native to the region, and he had a large palace built in Split to which he retired after his abdication in AD 305. During the 5th century, the last de jure Western Roman Emperor Julius Nepos ruled his small realm from the palace after fleeing Italy to go into exile in 475. The period ends with Avar and Croat invasions in the first half of the 7th century and the destruction of almost all Roman towns. Roman survivors retreated to more favourable sites on the coast, islands and mountains. The city of Dubrovnik was founded by such survivors from Epidaurum. The ethnogenesis of Croats is uncertain and there are several competing theories, Slavic and Iranian being the most frequently put forward. The most widely accepted of these, the Slavic theory, proposes migration of White Croats from the territory of White Croatia during the Migration Period. Conversely, the Iranian theory proposes Iranian origin, based on Tanais Tablets containing Greek inscription of given names "Χορούαθος", "Χοροάθος", and "Χορόαθος" (Khoroúathos, Khoroáthos, and Khoróathos) and their interpretation as anthroponyms of Croatian people. According to the work "De Administrando Imperio" written by the 10th-century Byzantine Emperor Constantine VII, the Croats had arrived in Roman province of Dalmatia in the first half of the 7th century after they defeated the Avars. However, that claim is disputed and competing hypotheses date the event between the 6th and the 9th centuries. Eventually two dukedoms were formed—Duchy of Pannonia and Duchy of Croatia, ruled by Ljudevit and Borna, as attested by chronicles of Einhard starting in 818. The record represents the first document of Croatian realms, vassal states of Francia at the time. The Frankish overlordship ended during the reign of Mislav two decades later. According to Constantine VII Christianization of Croats began in the 7th century, but the claim is disputed and generally Christianization is associated with the 9th century. The first native Croatian ruler recognised by the Pope was Duke Branimir, who received papal recognition from Pope John VIII on 7 June 879. Tomislav was the first king of Croatia, styled as such in a letter of Pope John X in 925. Tomislav defeated Hungarian and Bulgarian invasions, spreading the influence of Croatian kings. The medieval Croatian kingdom reached its peak in the 11th century during the reigns of Petar Krešimir IV (1058–1074) and Dmitar Zvonimir (1075–1089). When Stjepan II died in 1091 ending the Trpimirović dynasty, Dmitar Zvonimir's brother-in-law Ladislaus I of Hungary claimed the Croatian crown. This led to a war and personal union of Croatia and Hungary in 1102 under Coloman. For the next four centuries, the Kingdom of Croatia was ruled by the Sabor (parliament) and a ban (viceroy) appointed by the king. This period saw the rise of influential nobility such as the Frankopan and Šubić families to prominence, and ultimately numerous Bans from the two families. There was increasing threat of Ottoman conquest, and a struggle against the Republic of Venice for control of coastal areas. The Venetians gained control over most of Dalmatia by 1428, with exception of the city-state of Dubrovnik which became independent. Ottoman conquests led to the 1493 Battle of Krbava field and 1526 Battle of Mohács, both ending in decisive Ottoman victories. King Louis II died at Mohács, and in 1527, the Croatian Parliament met in Cetin and chose Ferdinand I of the House of Habsburg as new ruler of Croatia, under the condition that he provide protection to Croatia against the Ottoman Empire while respecting its political rights. Following the decisive Ottoman victories, Croatia was split into civilian and military territories, with the partition formed in 1538. The military territories would become known as the Croatian Military Frontier and were under direct Habsburg control. Ottoman advances in the Croatian territory continued until the 1593 Battle of Sisak, the first decisive Ottoman defeat, and stabilisation of borders. During the Great Turkish War (1683–1698), Slavonia was regained but western Bosnia, which had been part of Croatia before the Ottoman conquest, remained outside Croatian control. The present-day border between the two countries is a remnant of this outcome. Dalmatia, the southern part of the border, was similarly defined by the Fifth and the Seventh Ottoman–Venetian Wars. The Ottoman wars instigated great demographic changes. During the 16th century Croats from western and northern Bosnia, Lika, Krbava, the area between the rivers of Una and Kupa and especially from western Slavonia migrated towards Austria and the present-day Burgenland Croats are direct descendants of these settlers. To replace the fleeing population, the Habsburgs encouraged the populations of Bosnia to provide military service in the Military Frontier. The Croatian Parliament supported King Charles III's Pragmatic Sanction and signed their own Pragmatic Sanction in 1712. Subsequently, the emperor pledged to respect all privileges and political rights of Kingdom of Croatia and Queen Maria Theresa made significant contributions to Croatian matters, such as introducing compulsory education. Between 1797 and 1809 the First French Empire gradually occupied the entire eastern Adriatic coastline and a substantial part of its hinterland, ending the Venetian and the Ragusan republics, establishing the Illyrian Provinces. In response the Royal Navy started the blockade of the Adriatic Sea leading to the Battle of Vis in 1811. The Illyrian Provinces were captured by the Austrians in 1813, and absorbed by the Austrian Empire following the Congress of Vienna in 1815. This led to formation of the Kingdom of Dalmatia and restoration of the Croatian Littoral to the Kingdom of Croatia, now both under the same crown. The 1830s and 1840s saw romantic nationalism inspire the Croatian National Revival, a political and cultural campaign advocating the unity of all South Slavs in the empire. Its primary focus was the establishment of a standard language as a counterweight to Hungarian, along with the promotion of Croatian literature and culture. During the Hungarian Revolution of 1848 Croatia sided with the Austrians, Ban Josip Jelačić helping defeat the Hungarian forces in 1849, and ushering a period of Germanization policy. By the 1860s, failure of the policy became apparent, leading to the Austro-Hungarian Compromise of 1867 and creation of a personal union between the crowns of the Austrian Empire and the Kingdom of Hungary. The treaty left the issue of Croatia's status to Hungary, and the status was resolved by the Croatian–Hungarian Settlement of 1868 when kingdoms of Croatia and Slavonia were united. The Kingdom of Dalmatia remained under de facto Austrian control, while Rijeka retained the status of Corpus separatum introduced in 1779. After Austria-Hungary occupied Bosnia and Herzegovina following the 1878 Treaty of Berlin, the Croatian Military Frontier was abolished and the territory returned to Croatia in 1881, pursuant to provisions of the Croatian-Hungarian settlement. Renewed efforts to reform Austria-Hungary, entailing federalisation with Croatia as a federal unit, were stopped by advent of World War I. On 29 October 1918 the Croatian Parliament ("Sabor") declared independence and decided to join the newly formed State of Slovenes, Croats and Serbs, which in turn entered into union with the Kingdom of Serbia on 4 December 1918 to form the Kingdom of Serbs, Croats, and Slovenes. The Croatian Parliament never ratified a decision to unite with Serbia and Montenegro. The 1921 constitution defining the country as a unitary state and abolition of Croatian Parliament and historical administrative divisions effectively ended Croatian autonomy. The new constitution was opposed by the most widely supported national political party—the Croatian Peasant Party (HSS) led by Stjepan Radić. The political situation deteriorated further as Radić was assassinated in the National Assembly in 1928, leading to the dictatorship of King Alexander in January 1929. The dictatorship formally ended in 1931 when the king imposed a more unitarian constitution, and changed the name of the country to Yugoslavia. The HSS, now led by Vladko Maček, continued to advocate federalisation of Yugoslavia, resulting in the Cvetković–Maček Agreement of August 1939 and the autonomous Banovina of Croatia. The Yugoslav government retained control of defence, internal security, foreign affairs, trade, and transport while other matters were left to the Croatian Sabor and a crown-appointed Ban. In April 1941, Yugoslavia was occupied by Germany and Italy. Following the invasion of the territory, parts of Croatia, Bosnia and Herzegovina, and the region of Syrmia were incorporated into the Independent State of Croatia (NDH), a Nazi-backed puppet state. Parts of Dalmatia were annexed by Italy, and the northern Croatian regions of Baranja and Međimurje by Hungary. The NDH regime was led by Ante Pavelić and ultranationalist Ustaše. NDH was trying to establish such an internal structure that would be consistent with that of the Third Reich and fascist Italy so its authorities introduced racial laws against Jews, Roma and Serbs many of whom were imprisoned in concentration camps. The Ustaše regime murdered around 300,000 Serbs, as part of their genocide campaign. Statistician Vladimir Žerjavić lists 217,000 as victims of terror by the Axis, including the genocide in the NDH. The total number of Serb civilian and military deaths in the territory of the NDH was at around 330,000 to 370,000. Of the almost 40,000 Jews in the NDH, only around 9,000 survived, while most of the 15,000 Roma were killed. At the same time, anti-fascist Croats were targeted by the regime as well. The Yugoslav Royalist and Serbian nationalist Chetniks, a guerilla force that collaborated with the Axis, carried out genocide against Croats and Bosniaks. An estimated 47,000 to 68,000 Croats and Bosniaks died at the hands of the Chetniks, mostly civilians. The total number of Croats who lost their lives during World War II from all causes, based on the studies of Vladimir Žerjavić and Bogoljub Kočović, is estimated to be approximately 200,000. A resistance movement soon emerged. On 22 June 1941, the 1st Sisak Partisan Detachment was formed near Sisak, as the first military unit formed by a resistance movement in occupied Europe. This sparked the beginning of the Yugoslav Partisan movement, a communist multi-ethnic anti-fascist resistance group led by Josip Broz Tito. The movement grew rapidly and at the Tehran Conference in December 1943 the Partisans gained recognition from the Allies. With Allied support in logistics, equipment, training and air power, and with the assistance of Soviet troops taking part in the 1944 Belgrade Offensive, the Partisans gained control of Yugoslavia and the border regions of Italy and Austria by May 1945, during which tens of thousands of members of the NDH armed forces, as well as Croat refugees, were killed by the Yugoslav Partisans. In the following years, ethnic Germans faced persecution in Yugoslavia, and many were interned in camps. The political aspirations of the Partisan movement were reflected in the State Anti-fascist Council for the National Liberation of Croatia, which developed in 1943 as the bearer of Croatian statehood and later transformed into the Parliament of Croatia in 1945, and AVNOJ—its counterpart at the Yugoslav level. After World War II, Croatia became a single-party socialist federal unit of the SFR Yugoslavia, ruled by the Communists, but enjoying a degree of autonomy within the federation. In 1967, Croatian authors and linguists published a Declaration on the Status and Name of the Croatian Standard Language demanding greater autonomy for the Croatian language. The declaration contributed to a national movement seeking greater civil rights and decentralization of the Yugoslav economy, culminating in the Croatian Spring of 1971, suppressed by Yugoslav leadership. Still, the 1974 Yugoslav Constitution gave increased autonomy to federal units, basically fulfilling a goal of the Croatian Spring, and providing a legal basis for independence of the federative constituents. Following the death of Yugoslav President Josip Broz Tito in 1980, the political situation in Yugoslavia deteriorated, with national tension fanned by the 1986 SANU Memorandum and the 1989 coups in Vojvodina, Kosovo and Montenegro. In January 1990, the Communist Party fragmented along national lines, with the Croatian faction demanding a looser federation. In the same year, the first multi-party elections were held in Croatia, with Franjo Tuđman's win raising nationalist tensions further. Some of Serbs in Croatia left Sabor and declared the autonomy of areas that would soon become the unrecognised Republic of Serbian Krajina, intent on achieving independence from Croatia. As tensions rose, Croatia declared independence on 25 June 1991. However, the full implementation of declaration only came into effect on 8 October 1991. In the meantime, tensions escalated into overt war when the Yugoslav People's Army (JNA) and various Serb paramilitary groups attacked Croatia. By the end of 1991, a high-intensity conflict fought along a wide front reduced Croatia to control of only about two-thirds of its territory. The various Serb paramilitary groups then began pursuing a campaign of killing, terror and expulsion against the Croats in the rebel territories, killing thousands of Croat civilians and expelling or displacing as many as 400,000 Croats and other non-Serbs from their homes. On 15 January 1992, Croatia gained diplomatic recognition by the European Economic Community members, and subsequently the United Nations. The war effectively ended in August 1995 with a decisive victory by Croatia; the event is commemorated each year on 5 August as Victory and Homeland Thanksgiving Day and the Day of Croatian Defenders. Following the Croatian victory, about 200,000 Serbs from the self-proclaimed Republic of Serbian Krajina fled from the region and their lands were subsequently settled by Croat refugees from Bosnia and Herzegovina. The remaining occupied areas were restored to Croatia pursuant to the Erdut Agreement of November 1995, with the process concluded in January 1998. Following the end of the war, Croatia faced the challenges of post-war reconstruction, the return of refugees, advancing democratic principles, protection of human rights and general social and economic development. The post-2000 period is characterized by democratization, economic growth and structural and social reforms, as well as problems such as unemployment, corruption and the inefficiency of the public administration. Croatia joined the Partnership for Peace on 25 May 2000 and become a member of the World Trade Organization on 30 November 2000. On 29 October 2001, Croatia signed a Stabilization and Association Agreement with the European Union, submitted a formal application for the EU membership in 2003, was given the status of candidate country in 2004, and began accession negotiations in 2005. In November 2000 and March 2001, the Parliament amended the Constitution changing its bicameral structure back into historic unicameral and reducing the presidential powers. Although Croatia experienced a significant boom in the economy in the early 2000s, the increase of the government debt and the absence of concrete reforms led to a financial crisis in 2008 which forced the government to cut public spending thus provoking a public outcry. On 1 April 2009, Croatia joined NATO. A wave of anti-government protests organized via Facebook took place in early 2011 as general dissatisfaction with political and economic state grew. The majority of Croatian voters voted in favour of country's EU membership at the 2012 referendum. Croatia completed EU accession negotiations in 2011 and joined the European Union on 1 July 2013. Croatia was affected by the European migrant crisis in 2015 when Hungary's closure of its borders with Serbia forced over 700,000 migrants to use Croatia as a transit country on their way to Western Europe. On 22 March 2020, a 5.5 earthquake struck Croatia, with the epicenter located 7 kilometres (4.3 mi) north of Zagreb city center, the nations capital, inflicting heavy structural damage in historic city center, and causing 27 injuries with one fatality. Over 1,900 buildings were reported to have become uninhabitable by the earthquake damage. The earthquake occurred during the coronavirus pandemic and caused problems in enforcement of social distancing measures set out by the Government of Croatia. It occurred during the Croatian Presidency of the Council of the European Union. Croatia is located in Central and Southeast Europe, on the coast of the Adriatic Sea. It borders Hungary to the northeast, Serbia to the east, Bosnia and Herzegovina and Montenegro to the southeast, and Slovenia to the northwest. It lies mostly between latitudes 42° and 47° N and longitudes 13° and 20° E. Part of the territory in the extreme south surrounding Dubrovnik is a practical exclave connected to the rest of the mainland by territorial waters, but separated on land by a short coastline strip belonging to Bosnia and Herzegovina around Neum. The territory covers , consisting of of land and of water. It is the 127th largest country in the world. Elevation ranges from the mountains of the Dinaric Alps with the highest point of the Dinara peak at near the border with Bosnia and Herzegovina in the south to the shore of the Adriatic Sea which makes up its entire southwest border. Insular Croatia consists of over a thousand islands and islets varying in size, 48 of which are permanently inhabited. The largest islands are Cres and Krk, each of them having an area of around . The hilly northern parts of Hrvatsko Zagorje and the flat plains of Slavonia in the east which is part of the Pannonian Basin are traversed by major rivers such as Danube, Drava, Kupa, and Sava. The Danube, Europe's second longest river, runs through the city of Vukovar in the extreme east and forms part of the border with Vojvodina. The central and southern regions near the Adriatic coastline and islands consist of low mountains and forested highlands. Natural resources found in the country in quantities significant enough for production include oil, coal, bauxite, low-grade iron ore, calcium, gypsum, natural asphalt, silica, mica, clays, salt, and hydropower. Karst topography makes up about half of Croatia and is especially prominent in the Dinaric Alps. There are a number of deep caves in Croatia, 49 of which are deeper than , 14 of them deeper than and three deeper than . Croatia's most famous lakes are the Plitvice lakes, a system of 16 lakes with waterfalls connecting them over dolomite and limestone cascades. The lakes are renowned for their distinctive colours, ranging from turquoise to mint green, grey or blue. Croatia can be subdivided between a number of ecoregions because of its climate and geomorphology. The country is consequently one of the richest in Europe in terms of biodiversity. There are four types of biogeographical regions in Croatia—Mediterranean along the coast and in its immediate hinterland, Alpine in most of Lika and Gorski Kotar, Pannonian along Drava and Danube, and Continental in the remaining areas. One of the most significant are karst habitats which include submerged karst, such as Zrmanja and Krka canyons and tufa barriers, as well as underground habitats. The karst geology harbours approximately 7,000 caves and pits, some of which are the habitat of the only known aquatic cave vertebrate—the olm. Forests are also significantly present in the country, as they cover representing 44% of Croatian land surface. Other habitat types include wetlands, grasslands, bogs, fens, scrub habitats, coastal and marine habitats. In terms of phytogeography, Croatia is a part of the Boreal Kingdom and is a part of Illyrian and Central European provinces of the Circumboreal Region and the Adriatic province of the Mediterranean Region. The World Wide Fund for Nature divides Croatia between three ecoregions—Pannonian mixed forests, Dinaric Mountains mixed forests and Illyrian deciduous forests. There are 37,000 known species in Croatia, but their actual number is estimated to be between 50,000 and 100,000. The claim is supported by nearly 400 new taxa of invertebrates discovered in Croatia in the first half of the 2000s alone. There are more than a thousand endemic species, especially in Velebit and Biokovo mountains, Adriatic islands and karst rivers. Legislation protects 1,131 species. The most serious threat to species is loss and degradation of habitats. A further problem is presented by invasive alien species, especially "Caulerpa taxifolia" algae. The invasive algae are regularly monitored and removed to protect the benthic habitat. Indigenous sorts of cultivated plants and breeds of domesticated animals are also numerous. Those include five breeds of horses, five breeds of cattle, eight breeds of sheep, two breeds of pigs, and a poultry breed. Even the indigenous breeds include nine endangered or critically endangered ones. There are 444 protected areas of Croatia, encompassing 9% of the country. Those include eight national parks, two strict reserves, and ten nature parks. The most famous protected area and the oldest national park in Croatia is the Plitvice Lakes National Park, a UNESCO World Heritage Site. Velebit Nature Park is a part of the UNESCO Man and the Biosphere Programme. The strict and special reserves, as well as the national and nature parks, are managed and protected by the central government, while other protected areas are managed by counties. In 2005, the National Ecological Network was set up, as the first step in the preparation of the EU accession and joining of the Natura 2000 network. Most of Croatia has a moderately warm and rainy continental climate as defined by the Köppen climate classification. Mean monthly temperature ranges between in January and in July. The coldest parts of the country are Lika and Gorski Kotar where snowy forested climate is found at elevations above . The warmest areas of Croatia are at the Adriatic coast and especially in its immediate hinterland characterised by the Mediterranean climate, as the temperature highs are moderated by the sea. Consequently, temperature peaks are more pronounced in the continental areas. The lowest temperature of was recorded on 3 February 1919 in Čakovec, and the highest temperature of was recorded on 4 August 1981 in Ploče. Mean annual precipitation ranges between and depending on geographic region and prevailing climate type. The least precipitation is recorded in the outer islands (Biševo, Lastovo, Svetac, Vis) and in the eastern parts of Slavonia. However, in the latter case, it occurs mostly during the growing season. The maximum precipitation levels are observed on the Dinara mountain range and in Gorski kotar. Prevailing winds in the interior are light to moderate northeast or southwest, and in the coastal area, prevailing winds are determined by local area features. Higher wind velocities are more often recorded in cooler months along the coast, generally as the cool northeasterly bura or less frequently as the warm southerly jugo. The sunniest parts of the country are the outer islands, Hvar and Korčula, where more than 2700 hours of sunshine are recorded per year, followed by the middle and southern Adriatic Sea area in general, and northern Adriatic coast, all with more than 2000 hours of sunshine per year. The Republic of Croatia is a unitary state using a parliamentary system of governance. With the collapse of the ruling communist party in SFR Yugoslavia, Croatia organized its first multi-party elections and adopted its present constitution in 1990. It declared independence on 8 October 1991 which led to the break-up of Yugoslavia and countries international recognition by the United Nations in 1992. Under its 1990 Constitution, Croatia operated a semi-presidential system until 2000 when it switched to a parliamentary system. Government powers in Croatia are divided into legislative, executive and judiciary powers. The President of the Republic () is the head of state, directly elected to a five-year term and is limited by the Constitution to a maximum of two terms. In addition to being the commander in chief of the armed forces, the president has the procedural duty of appointing the prime minister with the consent of the parliament, and has some influence on foreign policy. The most recent presidential elections were held on 5 January 2020, when Zoran Milanović became the new President. He took the oath of office on 19 February 2015. The Government is headed by the Prime Minister, who has four deputy prime ministers and 16 ministers in charge of particular sectors of activity. As the executive branch, it is responsible for proposing legislation and a budget, executing the laws, and guiding the foreign and internal policies of the republic. The Government is seated at Banski dvori in Zagreb. Since 19 October 2016, Croatian Prime Minister has been Andrej Plenković. A unicameral parliament ("") holds legislative power. A second chamber, the "House of Counties", set up in 1993 pursuant to the 1990 Constitution, was abolished in 2001. The number of Sabor members can vary from 100 to 160; they are all elected by popular vote to serve four-year terms. The sessions of the Sabor take place from 15 January to 15 July, and from 15 September to 15 December. The two largest political parties in Croatia are the Croatian Democratic Union and the Social Democratic Party of Croatia. Croatia has a civil law legal system in which law arises primarily from written statutes, with judges serving merely as implementers and not creators of law. Its development was largely influenced by German and Austrian legal systems. Croatian law is divided into two principal areas—private and public law. By the time EU accession negotiations were completed on 30 June 2010, Croatian legislation was fully harmonised with the Community acquis. The main law in the county is the Constitution adopted on December 22, 1990. The main national courts are the Constitutional Court, which oversees violations of the Constitution, and the Supreme Court, which is the highest court of appeal. In addition, there are also Administrative, Commercial, County, Misdemeanor, and Municipal courts. Cases falling within judicial jurisdiction are in the first instance decided by a single professional judge, while appeals are deliberated in mixed tribunals of professional judges. Lay magistrates also participate in trials. State's Attorney Office is the judicial body constituted of public prosecutors that is empowered to instigate prosecution of perpetrators of offences. Law enforcement agencies are organised under the authority of the Ministry of the Interior which consist primarily of the national police force. Croatia's security service is the Security and Intelligence Agency (SOA). Croatia was first subdivided into counties in the Middle Ages. The divisions changed over time to reflect losses of territory to Ottoman conquest and subsequent liberation of the same territory, changes of political status of Dalmatia, Dubrovnik, and Istria. The traditional division of the country into counties was abolished in the 1920s when the Kingdom of Serbs, Croats and Slovenes and the subsequent Kingdom of Yugoslavia introduced oblasts and banovinas respectively. Communist-ruled Croatia, as a constituent part of post-World War II Yugoslavia, abolished earlier divisions and introduced municipalities, subdividing Croatia into approximately one hundred municipalities. Counties were reintroduced in 1992 legislation, significantly altered in terms of territory relative to the pre-1920s subdivisions. In 1918, the Transleithanian part of Croatia was divided into eight counties with their seats in Bjelovar, Gospić, Ogulin, Osijek, Požega, Varaždin, Vukovar, and Zagreb, and the 1992 legislation established 14 counties in the same territory. Since the counties were re-established in 1992, Croatia is divided into 20 counties and the capital city of Zagreb, the latter having the authority and legal status of a county and a city at the same time. Borders of the counties changed in some instances since, with the latest revision taking place in 2006. The counties subdivide into 127 cities and 429 municipalities. Nomenclature of Territorial Units for Statistics (NUTS) division of Croatia is performed in several tiers. NUTS 1 level places the entire country in a single unit, while there are three NUTS 2 regions. Those are Northwest Croatia, Central and Eastern (Pannonian) Croatia, and Adriatic Croatia. The latter encompasses all the counties along the Adriatic coast. Northwest Croatia includes Koprivnica-Križevci, Krapina-Zagorje, Međimurje, Varaždin, the city of Zagreb, and Zagreb counties and the Central and Eastern (Pannonian) Croatia includes the remaining areas—Bjelovar-Bilogora, Brod-Posavina, Karlovac, Osijek-Baranja, Požega-Slavonia, Sisak-Moslavina, Virovitica-Podravina, and Vukovar-Syrmia counties. Individual counties and the city of Zagreb also represent NUTS 3 level subdivision units in Croatia. The NUTS Local administrative unit divisions are two-tiered. LAU 1 divisions match the counties and the city of Zagreb in effect making those the same as NUTS 3 units, while LAU 2 subdivisions correspond to the cities and municipalities of Croatia. Croatia has established diplomatic relations with 181 countries. , Croatia maintains a network of 54 embassies, 28 consulates and eight permanent diplomatic missions abroad. Furthermore, there are 52 foreign embassies and 69 consulates in the Republic of Croatia in addition to offices of international organisations such as the European Bank for Reconstruction and Development, International Organization for Migration, Organization for Security and Co-operation in Europe (OSCE), World Bank, World Health Organization (WHO), International Criminal Tribunal for the former Yugoslavia (ICTY), United Nations Development Programme, United Nations High Commissioner for Refugees, and UNICEF. In 2009, the Croatian Ministry of Foreign Affairs and European Integration employed 1,381 personnel and expended 648.2 million kuna (€86.4 million). Stated aims of Croatian foreign policy include enhancing relations with neighbouring countries, developing international co-operation and promotion of the Croatian economy and Croatia itself. Since 2003, Croatian foreign policy has focused on achieving the strategic goal of becoming a member state of the European Union (EU). In December 2011, Croatia completed the EU accession negotiations and signed an EU accession treaty on 9 December 2011. Croatia joined the European Union on 1 July 2013 marking the end of a process started in 2001 by signing of the Stabilisation and Association Agreement and Croatian application for the EU membership in 2003. A recurring obstacle to the negotiations was Croatia's ICTY co-operation record and Slovenian blocking of the negotiations because of Croatia–Slovenia border disputes. The latter should be resolved through an Arbitration Agreement of 4 November 2009, approved by national parliaments and a referendum in Slovenia., but due to the events during arbitration Croatia does not accept results. As of 2019, Croatia has unsolved border issues with all neighbouring former Yugoslav countries (Bosnia and Herzegovina, Montenegro, Serbia, Slovenia). Another strategic Croatian foreign policy goal for the 2000s was NATO membership. Croatia was included in the Partnership for Peace in 2000, invited to NATO membership in 2008 and formally joined the alliance on 1 April 2009. Croatia became a member of the United Nations Security Council for the 2008–2009 term, assuming presidency in December 2008. The country is preparing to join the Schengen Area. The Croatian Armed Forces (CAF) consist of the Air Force, Army, and Navy branches in addition to the Education and Training Command and Support Command. The CAF is headed by the General Staff, which reports to the Defence Minister, who in turn reports to the President of Croatia. According to the constitution, the President is commander-in-chief of the armed forces and in case of immediate threat during wartime he issues orders directly to the General Staff. Following the 1991–95 war defence spending and CAF size have been in constant decline. military spending was an estimated 2.39% of the country's GDP, which placed Croatia 64th in a ranking of all countries. Since 2005 the budget was kept below 2% of GDP, down from the record high of 11.1% in 1994. Traditionally relying on a large number of conscripts, CAF also went through a period of reforms focused on downsizing, restructuring and professionalisation in the years prior to Croatia's accession to NATO in April 2009. According to a presidential decree issued in 2006 the CAF is set to employ 18,100 active duty military personnel, 3,000 civilians and 2,000 voluntary conscripts between the ages of 18 and 30 in peacetime. Compulsory conscription was abolished in January 2008. Until 2008 military service was compulsory for men at age 18 and conscripts served six-month tours of duty, reduced in 2001 from the earlier scheme of nine-month conscription tours. Conscientious objectors could instead opt for an eight-month civilian service. Croatia also has a significant military industry sector which exported around US$120 million worth of military equipment and armament in 2010. Croatian-made weapons and vehicles used by CAF include the standard sidearm HS2000 manufactured by HS Produkt and the M-84D battle tank designed by the Đuro Đaković factory. Uniforms and helmets worn by CAF soldiers are also locally produced and successfully marketed to other countries. Croatia has a high-income economy and International Monetary Fund data projects that Croatian nominal GDP stands at $60,688 billion, or $14,816 per capita for 2018, while purchasing power parity GDP stands at $107.406 billion, or $26,221 per capita. According to Eurostat, Croatian GDP per capita in PPS stood at 65% of the EU average in 2019. Real GDP growth in 2018 was 2,6 per cent. The average net salary of a Croatian worker in October 2019 was 6,496 HRK per month (roughly 873 EUR), and the average gross salary was 8,813 HRK per month (roughly 1,185 EUR). , the unemployment rate dropped to 7.2% from 9.6% in December 2018. The number of unemployed persons was 106.703. Unemployment Rate in Croatia in years 1996–2018 averaged 17.38%, reaching an all-time high of 23.60% in January 2002 and a record low of 8.40% in September 2018. In 2017, economic output was dominated by the service sector which accounted for 70.1% of GDP, followed by the industrial sector with 26.2% and agriculture accounting for 3.7% of GDP. According to 2017 data, 1.9% of the workforce were employed in agriculture, 27.3% by industry and 70.8% in services. The industrial sector is dominated by shipbuilding, food processing, pharmaceuticals, information technology, biochemical and timber industry. In 2018, Croatian exports were valued at 108  billion kuna (€14.61 billion) with 176 billion kuna (€23.82 billion) worth of imports. Croatia's largest trading partner was the rest of the European Union, with top three countries being Germany, Italy and Slovenia. Privatization and the drive toward a market economy had barely begun under the new Croatian Government when war broke out in 1991. As a result of the war, the economic infrastructure sustained massive damage, particularly the revenue-rich tourism industry. From 1989 to 1993, the GDP fell 40.5%. The Croatian state still controls a significant part of the economy, with government expenditures accounting for as much as 40% of GDP. A backlogged judiciary system, combined with inefficient public administration, especially on issues of land ownership and corruption, are particular concerns. In the 2018 Corruption Perceptions Index, published by Transparency International, the country is ranked 60th with a score of 48, where zero denotes "highly corrupt" and 100 "very clean". In June 2013, the national debt stood at 59.5% of the nation's GDP. Tourism dominates the Croatian service sector and accounts for up to 20% of Croatian GDP. Annual tourist industry income for 2017 was estimated at €9.5 billion. Its positive effects are felt throughout the economy of Croatia in terms of increased business volume observed in a retail business, processing industry orders and summer seasonal employment. The industry is considered an export business because it significantly reduces the country's external trade imbalance. Since the end of the Croatian War of Independence, the tourist industry has grown rapidly, recording a fourfold rise in tourist numbers, with more than 11 million tourists each year. The most numerous are tourists from Germany, Slovenia, Austria, Italy, and Poland as well as Croatia itself. Length of a tourist stay in Croatia averaged 4.9 days in 2011. The bulk of the tourist industry is concentrated along the Adriatic Sea coast. Opatija was the first holiday resort. It first became popular in the middle of the 19th century. By the 1890s, it had become one of the most significant European health resorts. Later a number of resorts sprang up along the coast and islands, offering services catering to both mass tourism and various niche markets. The most significant are nautical tourism, as there are numerous marinas with more than 16 thousand berths, cultural tourism relying on the appeal of medieval coastal cities and numerous cultural events taking place during the summer. Inland areas offer agrotourism, mountain resorts, and spas. Zagreb is also a significant tourist destination, rivalling major coastal cities and resorts. Croatia has unpolluted marine areas reflected through numerous nature reserves and 116 Blue Flag beaches. Croatia is ranked as the 23rd most popular tourist destination in the world. About 15% of these visitors, or over one million per year, are involved with naturism, an industry for which Croatia is world-famous. It was also the first European country to develop commercial naturist resorts. The highlight of Croatia's recent infrastructure developments is its rapidly developed motorway network, largely built in the late 1990s and especially in the 2000s (decade). By September 2011, Croatia had completed more than of motorways, connecting Zagreb to most other regions and following various European routes and four Pan-European corridors. The busiest motorways are the A1, connecting Zagreb to Split and the A3, passing east–west through northwest Croatia and Slavonia. A widespread network of state roads in Croatia acts as motorway feeder roads while connecting all major settlements in the country. The high quality and safety levels of the Croatian motorway network were tested and confirmed by several EuroTAP and EuroTest programs. Croatia has an extensive rail network spanning , including of electrified railways and of double track railways. The most significant railways in Croatia are found within the Pan-European transport corridors Vb and X connecting Rijeka to Budapest and Ljubljana to Belgrade, both via Zagreb. All rail services are operated by Croatian Railways. There are international airports in Dubrovnik, Osijek, Pula, Rijeka, Split, Zadar, and Zagreb. The largest and busiest is Franjo Tuđman Airport in Zagreb. , Croatia complies with International Civil Aviation Organization aviation safety standards and the Federal Aviation Administration upgraded it to Category 1 rating. The busiest cargo seaport in Croatia is the Port of Rijeka and the busiest passenger ports are Split and Zadar. In addition to those, a large number of minor ports serve an extensive system of ferries connecting numerous islands and coastal cities in addition to ferry lines to several cities in Italy. The largest river port is Vukovar, located on the Danube, representing the nation's outlet to the Pan-European transport corridor VII. There are of crude oil pipelines in Croatia, connecting the Port of Rijeka oil terminal with refineries in Rijeka and Sisak, as well as several transhipment terminals. The system has a capacity of 20 million tonnes per year. The natural gas transportation system comprises of trunk and regional natural gas pipelines, and more than 300 associated structures, connecting production rigs, the Okoli natural gas storage facility, 27 end-users and 37 distribution systems. Croatian production of energy sources covers 85% of nationwide natural gas demand and 19% of oil demand. In 2008, 47.6% of Croatia's primary energy production structure comprised use of natural gas (47.7%), crude oil (18.0%), fuel wood (8.4%), hydro power (25.4%) and other renewable energy sources (0.5%). In 2009, net total electrical power production in Croatia reached 12,725 GWh and Croatia imported 28.5% of its electric power energy needs. The bulk of Croatian imports are supplied by the Krško Nuclear Power Plant, 50% owned by Hrvatska elektroprivreda, providing 15% of Croatia's electricity. With an estimated population of 4.13 million in 2019, Croatia ranks 127th by population in the world. Its population density stood in 2018 at 72,9 inhabitants per square kilometer, making Croatia one of the more sparsely populated European countries. The overall life expectancy in Croatia at birth was 76.3 years in 2018. The total fertility rate of 1.41 children per mother, is one of the lowest in the world, below the replacement rate of 2.1, it remains considerably below the high of 6.18 children born per woman in 1885. Since 1991, Croatia's death rate has continuously exceeded its birth rate. Croatia subsequently has one of the oldest populations in the world, with the average age of 43.3 years. Since the late 1990s, there has been a positive net migration into Croatia, reaching a level of more than 26,000 net immigrants in 2018. The Croatian Bureau of Statistics forecast that the population may shrink to 3.85 million by 2061, depending on actual birth rate and the level of net migration. The population of Croatia rose steadily from 2.1 million in 1857 until 1991, when it peaked at 4.7 million, with exception of censuses taken in 1921 and 1948, i.e. following two world wars. The natural growth rate of the population is currently negative with the demographic transition completed in the 1970s. In recent years, the Croatian government has been pressured each year to increase permit quotas for foreign workers, reaching an all-time high of 68.100 in 2019. In accordance with its immigration policy, Croatia is trying to entice emigrants to return. The population decrease was also a result of the Croatian War of Independence. During the war, large sections of the population were displaced and emigration increased. In 1991, in predominantly occupied areas, more than 400,000 Croats were either removed from their homes by the rebel Serb forces or fled the violence. During the final days of the war in 1995, about 150−200,000 Serbs fled before the arrival of Croatian forces during the Operation Storm. After the war, the number of displaced persons fell to about 250,000. The Croatian government has taken care of displaced persons by the social security system, and since December 1991 through the Office of Displaced Persons and Refugees. Most of the territories which were abandoned during the Croatian War of Independence were settled by Croat refugees from Bosnia and Herzegovina, mostly from north-western Bosnia, while some of the displaced people returned to their homes. According to the 2013 United Nations report, 17.6% of Croatia's population were foreign-born immigrants. Majority of the inhabitants of Croatia are Croats (90.4%), followed by Serbs (4.4%), Bosniaks (0.73%), Italians (0.42%), Albanians (0.41%, Roma (0.40%), Hungarians (0.33%), Slovenes (0.25%), Czechs (0.22%), Montenegrins (0.11%), Slovaks (0.11%), Macedonians (0.10%), and others (2.12%). Approximately 4 million Croats live abroad. Croatia has no official religion. Freedom of religion is a right defined by the Constitution which also defines all religious communities as equal before the law and separated from the state. According to the 2011 census, 91.36% of Croatians identify as Christian; of these, Roman Catholics make up the largest group, accounting for 86.28% of the population, after which follows Eastern Orthodoxy (4.44%), Protestantism (0.34%) and other Christians (0.30%). The largest religion after Christianity is Islam (1.47%). 4.57% of the population describe themselves as non-religious. In the Eurostat Eurobarometer Poll of 2010, 69% of the population of Croatia responded that "they believe there is a God". In a 2009 Gallup poll, 70% answered yes to the question "Is religion an important part of your daily life?" However, only 24% of the population attends religious services regularly. Croatian is the official language of Croatia and became the 24th official language of the European Union upon its accession in 2013. Minority languages are in official use in local government units where more than a third of the population consists of national minorities or where local legislation defines so. Those languages are Czech, Hungarian, Italian, Ruthenian, Serbian, and Slovak. Besides these, the following languages are also recognised: Albanian, Bosnian, Bulgarian, German, Hebrew, Macedonian, Montenegrin, Polish, Romanian, Romany, Russian, Rusyn, Slovene, Turkish and Ukrainian. According to the 2011 Census, 95.6% of citizens of Croatia declared Croatian as their native language, 1.2% declared Serbian as their native language, while no other language is represented in Croatia by more than 0.5% of native speakers among the population of Croatia. Croatian is a member of the South Slavic languages of Slavic languages group and is written using the Latin alphabet. There are three major dialects spoken on the territory of Croatia, with standard Croatian based on the Shtokavian dialect. The Chakavian and Kajkavian dialects are distinguished by their lexicon, phonology and syntax. Croatian replaced Latin as the official language of the Croatian government in the 19th century. In Yugoslavia, from 1972 to 1989, the language was constitutionally designated as the "Croatian literary language" and the "Croatian or Serbian language". It was the result of the resistance to "Serbo-Croatian" in the form of a Declaration on the Status and Name of the Croatian Literary Language and Croatian Spring. Croatians are protective of their Croatian language from foreign influences and are known for Croatian linguistic purism, as the language was under constant change and threats imposed by previous rulers (i.e. Austrian German, Hungarian, Italian and Turkish words were changed and altered to Slavic-looking or sounding ones). A 2011 survey revealed that 78% of Croatians claim knowledge of at least one foreign language. According to a survey ordered by the European Commission in 2005, 49% of Croatians speak English as the second language, 34% speak German, 14% speak Italian, and 10% speak French. Russian is spoken by 4% each, and 2% of Croatians speak Spanish. However, there are large municipalities that have minority languages that include substantial populations that speak these languages. An odd-majority of Slovenes (59%) have a certain level of knowledge of Croatian. The country is a part of various language-based international associations most notably the European Union Language Association. Literacy in Croatia stands at 99.2 per cent. A worldwide study about the quality of living in different countries published by "Newsweek" in August 2010 ranked the Croatian education system at 22nd, to share the position with Austria. Primary education in Croatia starts at the age of six or seven and consists of eight grades. In 2007 a law was passed to increase free, noncompulsory education until 18 years of age. Compulsory education consists of eight grades of elementary school. Secondary education is provided by gymnasiums and vocational schools. , there are 2,049 elementary schools and 701 schools providing various forms of secondary education. Primary and secondary education are also available in languages of recognized minorities in Croatia, where classes are held in Italian, Czech, German, Hungarian, and Serbian languages. There are 137 elementary and secondary level music and art schools, as well as 120 schools for disabled children and youth and 74 schools for adults. Nationwide leaving exams () were introduced for secondary education students in the school year 2009–2010. It comprises three compulsory subjects (Croatian language, mathematics, and a foreign language) and optional subjects and is a prerequisite for university education. Croatia has eight public universities, the University of Dubrovnik, University of Osijek, University of Pula, University of Rijeka, University of Split, University of Zadar and University of Zagreb, and two private universities, Catholic University of Croatia and Dubrovnik International University. The University of Zadar, the first university in Croatia, was founded in 1396 and remained active until 1807, when other institutions of higher education took over until the foundation of the renewed University of Zadar in 2002. The University of Zagreb, founded in 1669, is the oldest continuously operating university in Southeast Europe. There are also 15 polytechnics, of which two are private, and 30 higher education institutions, of which 27 are private. In total, there are 55 institutions of higher education in Croatia, attended by more than 157 thousand students. There are 205 companies, government or education system institutions and non-profit organisations in Croatia pursuing scientific research and development of technology. Combined, they spent more than 3 billion kuna (€400 million) and employed 10,191 full-time research staff in 2008. Among the scientific institutes operating in Croatia, the largest is the Ruđer Bošković Institute in Zagreb. The Croatian Academy of Sciences and Arts in Zagreb is a learned society promoting language, culture, arts and science from its inception in 1866. Croatia has been the home of many , including Fausto Veranzio, Giovanni Luppis, Slavoljub Eduard Penkala, Franjo Hanaman and Nikola Tesla, as well as , such as Franciscus Patricius, Nikola Nalješković, Nikola Vitov Gučetić, Josip Franjo Domin, Marino Ghetaldi, Roger Joseph Boscovich, Andrija Mohorovičić, Ivan Supek, Ivan Đikić, Miroslav Radman and Marin Soljačić. The Nobel Prize in Chemistry has been awarded to two Croatian laureates, Lavoslav Ružička (1939) and Vladimir Prelog (1975). Croatia has a universal health care system, whose roots can be traced back to the Hungarian-Croatian Parliament Act of 1891, providing a form of mandatory insurance of all factory workers and craftsmen. The population is covered by a basic health insurance plan provided by statute and optional insurance. In 2017, annual healthcare related expenditures reached 22.0 billion kuna (€3.0 billion). Healthcare expenditures comprise only 0.6% of private health insurance and public spending. In 2017, Croatia spent around 6.6% of its GDP on healthcare. In 2015, Croatia ranked 36th in the world in life expectancy with 74.7 years for men and 81.2 years for women, and it had a low infant mortality rate of 3 per 1,000 live births. There are hundreds of healthcare institutions in Croatia, including 79 hospitals and clinics with 23,967 beds. The hospitals and clinics care for more than 700 thousand patients per year and employ 5,205 medical doctors, including 3,929 specialists. There are 6,379 private practice offices, and a total of 41,271 health workers in the country. There are 63 emergency medical service units, responding to more than a million calls. The principal cause of death in 2008 was cardiovascular disease at 43.5% for men and 57.2% for women, followed by tumours, at 29.4% for men and 21.4% for women. In 2009 only 13 Croatians had been infected with HIV/AIDS and six had died from the disease. In 2008 it was estimated by the WHO that 27.4% of Croatians over the age of 15 are smokers. According to 2003 WHO data, 22% of the Croatian adult population is obese. Because of its geographical position, Croatia represents a blend of four different cultural spheres. It has been a crossroads of influences from western culture and the east—ever since the schism between the Western Roman Empire and the Byzantine Empire—and also from Mitteleuropa and Mediterranean culture. The Illyrian movement was the most significant period of national cultural history, as the 19th century proved crucial to the emancipation of the Croatian language and saw unprecedented developments in all fields of art and culture, giving rise to many historical figures. The Ministry of Culture is tasked with preserving the nation's cultural and natural heritage and overseeing its development. Further activities supporting the development of culture are undertaken at the local government level. The UNESCO's World Heritage List includes ten sites in Croatia. The country is also rich with intangible culture and holds 15 of UNESCO's World's intangible culture masterpieces, ranking fourth in the world. A global cultural contribution from Croatia is the necktie, derived from the cravat originally worn by the 17th-century Croatian mercenaries in France. Croatia has 95 professional theatres, 30 professional children's theatres and 52 amateur theatres visited by more than 1.54 million viewers per year. Professional theatres employ 1,195 artists. There are 46 professional orchestras, ensembles, and choirs in the country, attracting an annual attendance of 317 thousand. There are 166 cinemas with attendance exceeding 4.814 million. Croatia has 222 museums, visited by more than 2.7 million people in 2016. Furthermore, there are 1,768 libraries in the country, containing 26.8 million volumes, and 19 state archives. In 2010, 7,348 books and brochures were published, along with 2,676 magazines and 267 newspapers. There are also 135 radio stations and 25 TV stations operating in the country. In the past five years, film production in Croatia produced up to five feature films and 10 to 51 short films, with an additional 76 to 112 TV films. , there are 784 amateur cultural and artistic associations and more than 10 thousand cultural, educational, and artistic events held annually. The book publishing market is dominated by several major publishers and the industry's centrepiece event—Interliber exhibition held annually at Zagreb Fair. Croatia is categorised as having established a very high level of human development in the Human Development Index, with a high degree of equality in HDI achievements between women and men. It promotes disability rights. Recognition of same-sex unions in Croatia has gradually improved over the past decade, culminating in registered civil unions in July 2014, granting same-sex couples equal inheritance rights, tax deductions and limited adoption rights. However, in December 2013 Croatians voted in a constitutional referendum and approved changes to the constitution to define marriage as a union between a man and a woman. Architecture in Croatia reflects influences of bordering nations. Austrian and Hungarian influence is visible in public spaces and buildings in the north and in the central regions, architecture found along coasts of Dalmatia and Istria exhibits Venetian influence. Large squares named after culture heroes, well-groomed parks, and pedestrian-only zones, are features of these orderly towns and cities, especially where large scale Baroque urban planning took place, for instance in Osijek (Tvrđa), Varaždin, and Karlovac. Subsequent influence of the Art Nouveau was reflected in contemporary architecture. Along the coast, the architecture is Mediterranean with a strong Venetian and Renaissance influence in major urban areas exemplified in works of Giorgio da Sebenico and Niccolò Fiorentino such as the Cathedral of St. James in Šibenik. The oldest preserved examples of Croatian architecture are the 9th-century churches, with the largest and the most representative among them being Church of St. Donatus in Zadar. Besides the architecture encompassing the oldest artworks in Croatia, there is a long history of artists in Croatia reaching the Middle Ages. In that period the stone portal of the Trogir Cathedral was made by Radovan, representing the most important monument of Romanesque sculpture from Medieval Croatia. The Renaissance had the greatest impact on the Adriatic Sea coast since the remainder of Croatia was embroiled in the Hundred Years' Croatian–Ottoman War. With the waning of the Ottoman Empire, art flourished during the Baroque and Rococo. The 19th and the 20th centuries brought about affirmation of numerous Croatian artisans, helped by several patrons of the arts such as bishop Josip Juraj Strossmayer. Croatian artists of the period achieving worldwide renown were Vlaho Bukovac and Ivan Meštrović. The Baška tablet, a stone inscribed with the glagolitic alphabet found on the Krk island and dated to 1100, is considered to be the oldest surviving prose in Croatian. The beginning of more vigorous development of Croatian literature is marked by the Renaissance and Marko Marulić. Besides Marulić, Renaissance playwright Marin Držić, Baroque poet Ivan Gundulić, Croatian national revival poet Ivan Mažuranić, novelist, playwright and poet August Šenoa, children's writer Ivana Brlić-Mažuranić, writer and journalist Marija Jurić Zagorka, poet and writer Antun Gustav Matoš, poet Antun Branko Šimić, expressionist and realist writer Miroslav Krleža, poet Tin Ujević and novelist and short story writer Ivo Andrić are often cited as the greatest figures in Croatian literature. In Croatia, the freedom of the press and the freedom of speech are guaranteed by the Constitution. Croatia ranked 64th in the 2019 Press Freedom Index report compiled by Reporters Without Borders which noted that journalists who investigate corruption, organised crime or war crimes face challenges and that the Government was trying to influence the public broadcaster HRT's editorial policies. In its 2019 Freedom in the World report, the Freedom House classified freedoms of press and speech in Croatia as generally free from political interference and manipulation, noting that journalists still face threats and occasional attacks. The state-owned news agency HINA runs a wire service in Croatian and English on politics, economics, society and culture. , there are fourteen nationwide free-to-air DVB-T television channels, with Croatian Radiotelevision (HRT) operating four, Nova TV and RTL Televizija operating two of the channels each, and the remaining three operated by the Croatian Olympic Committee, Kapital Net d.o.o. and Author d.o.o. companies. In addition there are 21 regional or local DVB-T television channels. The HRT is also broadcasting a satellite TV channel. In 2018, there were 147 radio stations and 27 TV stations in Croatia. Cable television and IPTV networks are gaining ground in the country, as the cable TV networks already serve 450 thousand people, 10% of the total population of the country. In 2010, 314 newspapers and 2,678 magazines were published in Croatia. The print media market is dominated by the Croatian-owned Hanza Media and Austrian-owned Styria Media Group who publish their flagship dailies "Jutarnji list", "Večernji list" and "24sata". Other influential newspapers are "Novi list" and "Slobodna Dalmacija". In 2013, "24sata" was the most widely circulated daily newspaper, followed by "Večernji list" and "Jutarnji list". Croatia's film industry is small and heavily subsidised by the government, mainly through grants approved by the Ministry of Culture with films often being co-produced by HRT. Croatian cinema produces between five and ten feature films per year. Pula Film Festival, the national film awards event held annually in Pula, is the most prestigious film event featuring national and international productions. Animafest Zagreb, founded in 1972, is the prestigious annual film festival entirely dedicated to the animated film. The first greatest accomplishment by Croatian filmmakers was achieved by Dušan Vukotić when he won the 1961 Academy Award for Best Animated Short Film for "Ersatz" (). Croatian film producer Branko Lustig won the Academy Awards for Best Picture for "Schindler's List" and "Gladiator". Croatian traditional cuisine varies from one region to another. Dalmatia and Istria draw upon culinary influences of Italian and other Mediterranean cuisines which prominently feature various seafood, cooked vegetables and pasta, as well as condiments such as olive oil and garlic. The continental cuisine is heavily influenced by Austrian, Hungarian, and Turkish culinary styles. In that area, meats, freshwater fish and vegetable dishes are predominant. There are two distinct wine-producing regions in Croatia. The continental region in the northeast of the country, especially Slavonia, is capable of producing premium wines, particularly whites. Along the north coast, Istrian and Krk wines are similar to those produced in neighbouring Italy, while further south in Dalmatia, Mediterranean-style red wines are the norm. Annual production of wine exceeds 140 million litres. Croatia was almost exclusively a wine-consuming country up until the late 18th century when a more massive production and consumption of beer started; the annual consumption of beer in 2008 was 83.3 litres per capita which placed Croatia in 15th place among the world's countries. There are more than 400,000 active sportspeople in Croatia. Out of that number, 277,000 are members of sports associations and nearly 4,000 are members of chess and contract bridge associations. Association football is the most popular sport. The Croatian Football Federation (), with more than 118,000 registered players, is the largest sporting association in the country. The Prva HNL football league attracts the highest average attendance of any professional sports league in the country. In season 2010–11, it attracted 458,746 spectators. Croatian athletes competing at international events since Croatian independence in 1991 won 44 Olympic medals, including 15 gold medals—at the 1996 and 2004 Summer Olympics in handball, 2000 Summer Olympics in weightlifting, 2002 and 2006 Winter Olympics in alpine skiing, 2012 Summer Olympics in discus throw, trap shooting, and water polo, and in 2016 Summer Olympics in shooting, rowing, discus throw, sailing and javelin throw. In addition, Croatian athletes won 16 gold medals at world championships, including four in athletics at the World Championships in Athletics held in 2007, 2009, 2013 and 2017, one in handball at the 2003 World Men's Handball Championship, two in water polo at the 2007 World Aquatics Championships and 2017 World Aquatics Championships, one in rowing at the 2010 World Rowing Championships, six in alpine skiing at the FIS Alpine World Ski Championships held in 2003 and 2005 and two at the World Taekwondo Championships in 2011 and 2007. Croatian athletes also won Davis cup in 2005 and 2018. The Croatian national football team came in third in 1998 and second in the 2018 FIFA World Cup. Croatia hosted several major sport competitions, including the 2009 World Men's Handball Championship, the 2007 World Table Tennis Championships, the 2000 World Rowing Championships, the 1987 Summer Universiade, the 1979 Mediterranean Games and several European Championships. The governing sports authority in the country is the Croatian Olympic Committee (), founded on 10 September 1991 and recognised by the International Olympic Committee since 17 January 1992, in time to permit the Croatian athletes to appear at the 1992 Winter Olympics in Albertville, France representing the newly independent nation for the first time at the Olympic Games.
https://en.wikipedia.org/wiki?curid=5573
History of Croatia Croatia first appeared as two duchies in the 7th century, the Duchy of Croatia and the Duchy of Pannonian Croatia, which were united and elevated into the Kingdom of Croatia which lasted from 925 until 1918. From the 12th century the Kingdom of Croatia entered a Personal Union with the Kingdom of Hungary, it remained a distinct state with its ruler ("Ban") and Sabor, but it elected Royal dynasties from neighboring powers, primarily Hungary, Naples and the Habsburg Monarchy. The period from the 15th to the 17th centuries was marked by bitter struggles with the Ottoman Empire. After being incorporated in Yugoslavia for most of the 20th century, Croatia regained independence in 1991. The area known today as Croatia was inhabited throughout the prehistoric period. Fossils of Neanderthals dating to the middle Palaeolithic period have been unearthed in northern Croatia, with the most famous and the best presented site in Krapina. Remnants of several Neolithic and Chalcolithic cultures were found in all regions of the country. The largest proportion of the sites is in the northern Croatia river valleys, and the most significant cultures whose presence was discovered include Starčevo, Vučedol and Baden cultures. The Iron Age left traces of the early Illyrian Hallstatt culture and the Celtic La Tène culture. Much later, the region was settled by Liburnians and Illyrians, while the first Greek colonies were established on the Vis and Hvar islands. Dalmatia was the northern part of the Illyrian kingdom between the 4th century BC until the Illyrian Wars in the 220s BC and 168 BC when the Roman Republic established its protectorate south of the river Neretva. The area north of the Neretva was slowly incorporated into Roman possession until the province of Illyricum was formally established c. 32–27 BC. The Dalmatia region then became part of the Roman province of Illyricum. Between 6 and 9 AD the Dalmatians raised the last in a series of revolts together with the Pannonians, but it was finally crushed and in 10 AD Illyricum was split into two provinces, Pannonia and Dalmatia. The province of Dalmatia spread inland to cover all of the Dinaric Alps and most of the eastern Adriatic coast. Dalmatia was the birthplace of the Roman Emperor Diocletian, who, upon retirement from Emperor in AD 305, built a large palace near Salona, out of which the city of Split later developed. Historians such as Theodore Mommsen and Bernard Bavant argue that all Dalmatia was fully romanized and Latin speaking by the 4th century. Others, such as Aleksandar Stipčević, argue that the process of romanization was rather selective and involved mostly urban centers but not the countryside, where previous Illyrian socio-political structures were adapted to Roman administration and political structure only in some necessities. Stanko Guldescu argued that the Vlachs, or Morlachs, were Latin speaking and pastoral peoples who lived in the Balkan mountains since pre-Roman times, and they are mentioned in the oldest Croatian chronicles After the Western Roman Empire collapsed in 476, with the beginning of the Migration Period, Julius Nepos shortly ruled his diminished domain from the Diocletian palace after his 476 flight from Italy. The region was then ruled by the Ostrogoths up to 535, when Justinian I added the territory to the Byzantine Empire. Later, the Byzantines formed the Theme of Dalmatia in the same territory. The Roman period ends with Avar and Croat invasions in the 6th and 7th centuries and the destruction of almost all Roman towns. Roman survivors retreated to more favourable sites on the coast, islands and mountains. The city of Ragusa was founded by such survivors from Epidaurum. According to the work "De Administrando Imperio" written by the 10th-century Byzantine Emperor Constantine VII, the Croats had arrived in what is today Croatia, from southern Poland, south of Kraków in the early 7th century, however that claim is disputed and competing hypotheses date the event between the 6th and the 9th centuries. Eventually two dukedoms were formed—Duchy of Pannonia and Duchy of Dalmatia, ruled by Liudewit and Borna, as attested by chronicles of Einhard starting in the year 818. The record represents the first document of Croatian realms, vassal states of Francia at the time. The Frankish overlordship ended during the reign of Mislav two decades later. According to the Constantine VII christianization of Croats began in the 7th century, but the claim is disputed and generally christianization is associated with the 9th century. In 879, under duke Branimir, the duke of Croats, Dalmatian Croatia received papal recognition as a state from Pope John VIII. Tomislav was the first ruler of Croatia who was styled a king in a letter from the Pope John X, dating kingdom of Croatia to year 925. Tomislav defeated Hungarian and Bulgarian invasions, spreading the influence of Croatian kings. The medieval Croatian kingdom reached its peak in the 11th century during the reigns of Petar Krešimir IV (1058–1074) and Dmitar Zvonimir (1075–1089). When Stjepan II died in 1091 ending the Trpimirović dynasty, Ladislaus I of Hungary claimed the Croatian crown. Opposition to the claim led to a war and personal union of Croatia and Hungary in 1102, ruled by Coloman. The consequences of the change to the Hungarian king included the introduction of feudalism and the rise of the native noble families such as Frankopan and Šubić. The later kings sought to restore some of their previously lost influence by giving certain privileges to the towns. For the next four centuries, the Kingdom of Croatia was ruled by the Sabor (parliament) and a Ban (viceroy) appointed by the king. The princes of Bribir from the Šubić family became particularly influential, asserting control over large parts of Dalmatia, Slavonia and Bosnia. Later, however, the Angevins intervened and restored royal power. The period saw rise of native nobility such as the Frankopans and the Šubićs to prominence and ultimately numerous Bans from the two families. Separate coronation as King of Croatia was gradually allowed to fall into abeyance and last crowned king is Charles Robert in 1301 after which Croatia contented herself with a separate diploma inaugurale. The reign of Louis the Great (1342–1382) is considered the golden age of Croatian medieval history. Ladislaus of Naples also sold the whole of Dalmatia to Venice in 1409. The period saw increasing threat of Ottoman conquest and struggle against the Republic of Venice for control of coastal areas. The Venetians gained control over most of Dalmatia by 1428. With exception of the city-state of Dubrovnik which became independent, the rule of Venice on most of Dalmatia will last nearly four centuries (c. 1420 – 1797). In 1490 the estates of Croatia declined to recognize Vladislaus II until he had taken oath to respect their liberties, and insisted upon his erasing from the diploma certain phrases which seemed to reduce Croatia to the rank of a mere province. The dispute was solved in 1492 As the Turkish incursion into Europe started, Croatia once again became a border area. The Croats fought an increasing number of battles and gradually lost increasing swathes of territory to the Ottoman Empire. Ottoman conquests led to the 1493 Battle of Krbava field and 1526 Battle of Mohács, both ending in decisive Ottoman victories. King Louis II died at Mohács, and in November 1526, a Hungarian parliament elected János Szapolyai as the new king of Hungary. In December 1526, another Hungarian parliament elected Ferdinand Habsburg as King of Hungary. On the other side, the Croatian Parliament met in Cetin and chose Ferdinand I of the House of Habsburg as new ruler of Croatia, under the condition that he provide protection to Croatia against the Ottoman Empire while respecting its political rights. A few years later both crown would be again united in Habsburgs hands and the union would be restored. The Ottoman Empire further expanded in the 16th century to include most of Slavonia, western Bosnia (then called Turkish Croatia) and Lika. Ottoman Croatia initially was part of Rumelia Eyalet, and later was parts of Budin Eyalet, Bosnia Eyalet and Kanije Eyalet. Later in the same century, Croatia was so weak that its parliament authorized Ferdinand Habsburg to carve out large areas of Croatia and Slavonia adjacent to the Ottoman Empire for the creation of the Military Frontier ("Vojna Krajina", German: "Militaergrenze") which would be ruled directly from Vienna's military headquarters. The area became rather deserted and was subsequently settled by Serbs, Vlachs, Croats and Germans and others. As a result of their compulsory military service to the Habsburg Empire during conflict with the Ottoman Empire, the population in the Military Frontier was free of serfdom and enjoyed much political autonomy, unlike the population living in the parts ruled by Hungary. After the Bihać fort finally fell in 1592, only small parts of Croatia remained unconquered. The Ottoman army was successfully repelled for the first time on the territory of Croatia following the battle of Sisak in 1593. The lost territory was partially restored, except for large parts of today's Bosnia and Herzegovina, Lika and Slavonia regions of present Croatia. By the 18th century, the Ottoman Empire was driven out of Hungary, and Austria brought the empire under central control. Empress Maria Theresa of Austria was supported by the Croatians in the War of Austrian Succession of 1741–1748 and subsequently made significant contributions to Croatian matters. With the fall of the Venetian Republic in 1797, its possessions in eastern Adriatic became subject to a dispute between France and Austria. The Habsburgs eventually secured them (by 1815) and Dalmatia and Istria became part of the empire, though they were in Cisleithania while Croatia and Slavonia were under Hungary. Croatian romantic nationalism emerged in the mid-19th century to counteract the apparent Germanization and Magyarization of Croatia. The Illyrian movement attracted a number of influential figures from the 1830s on, and produced some important advances in the Croatian language and culture. In the Revolutions of 1848 the Triune Kingdom of Croatia, Slavonia and Dalmatia, driven by fear of Magyar nationalism, supported the Habsburg court against Hungarian revolutionary forces. However, despite the contributions of its ban Jelačić in quenching the Hungarian war of independence, Croatia, not treated any more favourably by Vienna than the Hungarians themselves, lost its domestic autonomy. In 1867 the Dual Monarchy was created; Croatian autonomy was restored in 1868 with the Croatian–Hungarian Settlement which was comparatively favourable for the Croatians, but still problematic because of issues such as the unresolved status of Rijeka. World War I brought an end to the Habsburg Empire. Croatia suffered great loss of life in the war, and while there was some late effort to establish a Croatian section of the Empire, ultimately the military defeat led to the establishment of a separate South Slavic state, which would later be named Yugoslavia. Shortly before the end of the First World War in 1918, the Croatian Parliament severed relations with Austria-Hungary as the Entente armies defeated those of the Habsburgs. "Croatia and Slavonia" became a part of the State of Slovenes, Croats and Serbs composed out of all Southern Slavic territories of the now former Austro-Hungarian Monarchy with a transitional government headed in Zagreb. The creation of Yugoslavia happened in late 1918: after Srijem left Croatia-Slavonia and joined Serbia together with Vojvodina, shortly followed by a referendum to join Bosnia and Herzegovina to Serbia, the People's Council of Slovenes, Croats and Serbs ("Narodno vijeće"), guided by what was by that time a half a century long tradition of pan-Slavism and without sanction of the Croatian sabor, joined the Kingdom of Serbia into the Kingdom of the Serbs, Croats and Slovenes. Although the state inherited much of Austro-Hungary's military arsenal, including the entire fleet, the Kingdom of Italy moved rapidly to annex the state's most western territories, promised to her by the Treaty of London of 1915. An Italian Army eventually took Istria, started to annex the Adriatic islands one by one, and even landed in Zadar. Partial resolution to the so-called Adriatic Question would come in 1920 with the Treaty of Rapallo. The Kingdom underwent a crucial change in 1921 to the dismay of Croatia's largest political party, the Croatian Peasant Party (Hrvatska seljačka stranka). The new constitution abolished the historical/political entities, including Croatia and Slavonia, centralizing authority in the capital of Belgrade. The Croatian Peasant Party boycotted the government of the Serbian People's Radical Party throughout the period, except for a brief interlude between 1925 and 1927, when external Italian expansionism was at hand with her allies, Albania, Hungary, Romania and Bulgaria that threatened Yugoslavia as a whole. In the early 1920s the Yugoslav government of Serbian prime minister Nikola Pasic used police pressure over voters and ethnic minorities, confiscation of opposition pamphlets and other measures of election rigging to keep the opposition, and mainly the Croatian Peasant Party and its allies in minority in Yugoslav parliament. Pasic believed that Yugoslavia should be as centralized as possible, creating in place of distinct regional governments and identities a Greater Serbian national concept of concentrated power in the hands of Belgrade. In the words of a prominent Croatian writer, Shlegel's death became the pretext for terror in all forms. Politics was soon "indistinguishable from gangsterism." Even in this oppressive climate, few rallied to the Ustaša cause and the movement was never able to organise within Croatia. But its leaders did manage to convince the Communist Party that it was a progressive movement. The party's newspaper "Proleter" (December 1932) stated: "[We] salute the Ustaša movement of the peasants of Lika and Dalmatia and fully support them." In 1934, King Aleksandar was assassinated abroad, in Marseille, by a coalition of the Ustaše and a similarly radical movement, the Bulgarian VMORO. The Serbian-Croatian Cvetković-Maček government that came to power, distanced Yugoslavia's former allies of France and the United Kingdom, and moved closer to Fascist Italy and Nazi Germany in the period of 1935–1941. A national Banovina of Croatia was created in 1939 out of the two Banates, as well as parts of the Zeta, Vrbas, Drina and Danube Banates. It had a reconstructed Croatian Parliament which would choose a Croatian Ban and Viceban. This Croatia included a part of Bosnia, most of Herzegovina and the city of Dubrovnik and the surroundings. The Axis occupation of Yugoslavia in 1941 allowed the Croatian radical right Ustaše to come into power, forming the "Independent State of Croatia" ("Nezavisna Država Hrvatska", NDH), led by Ante Pavelić, who assumed the role of "Poglavnik". Following the pattern of other fascist regimes in Europe, the Ustashi enacted racial laws, formed eight concentration camps targeting minority Serbs, Romas and Jewish populations, as well as Croatian and Bosnian Muslim partisans. The biggest concentration camp was Jasenovac in Croatia. The NDH had a program, formulated by Mile Budak, to purge Croatia of Serbs, by "killing one third, expelling the other third and assimilating the remaining third". The main targets for persecution were the Serbs, with approximately 330,000 were killed. The anti-fascist communist-led Partisan movement, based on pan-Yugoslav ideology, emerged in early 1941, under the command of Croatian-born Josip Broz Tito, spreading quickly into many parts of Yugoslavia. The 1st Sisak Partisan Detachment, often hailed as the first armed anti-fascist resistance unit in occupied Europe, was formed in Croatia, in the Brezovica Forest near the town of Sisak. As the movement began to gain popularity, the Partisans gained strength from Croats, Bosniaks, Serbs, Slovenes, and Macedonians who believed in a unified, but federal, Yugoslav state. By 1943, the Partisan resistance movement had gained the upper hand, against the odds, and in 1945, with help from the Soviet Red Army (passing only through small parts such as Vojvodina), expelled the Axis forces and local supporters. The "ZAVNOH", state anti-fascist council of people's liberation of Croatia, functioned since 1942 and formed an interim civil government by 1943. NDH's ministers of War and Internal Security Mladen Lorković and Ante Vokić tried to switch to Allied side. Pavelić was in the beginning supporting them but when he found that he would need to leave his position he imprisoned them in Lepoglava prison where they were executed. Following the defeat of the Independent State of Croatia at the end of the war, a large number of Ustaše, civilians supporting them (ranging from sympathizers, young conscripts, anti-communists, and ordinary serfs who were motivated by Partisan atrocities), Chetniks and anti-Communists attempted to flee in the direction of Austria hoping to surrender to British forces and to be given refuge. They were instead interned by British forces and then returned to the Partisans. Croatia was a Socialist Republic part of a six-part Socialist Federative Republic of Yugoslavia. Under the new communist system, privately owned factories and estates were nationalized, and the economy was based on a type of planned market socialism. The country underwent a rebuilding process, recovered from World War II, went through industrialization and started developing tourism. The country's socialist system also provided free apartments from big companies, which with the workers' self-management investments paid for the living spaces. From 1963, the citizens of Yugoslavia were allowed to travel to almost any country because of the neutral politics. No visas were required to travel to eastern or western countries, or to capitalist or communist nations. Such free travel was unheard of at the time in the Eastern Bloc countries, and in some western countries as well (e.g., Spain or Portugal, both dictatorships at the time). This proved to be very helpful for Croatia's inhabitants who found working in foreign countries more financially rewarding. Upon retirement, a popular plan was to return to live in Croatia (then Yugoslavia) to buy a more expensive property. In Yugoslavia, the people of Croatia were guaranteed free healthcare, free dental care, and secure pensions. The older generation found this very comforting as pensions would sometimes exceed their former paychecks. Free trade and travel within the country also helped Croatian industries that imported and exported throughout all the former republics. Students and military personnel were encouraged to visit other republics to learn more about the country, and all levels of education, especially secondary education and higher education, were free. In reality the housing was inferior with poor heat and plumbing, the medical care often lacking even in availability of antibiotics, schools were propaganda machines and travel was a necessity to provide the country with hard currency. The propagandists, who want people to believe "neutral policies" equalized Serbs and Croats, severely restricted free speech and did not protect citizens from ethnic attacks. Membership in the party was as much a prerequisite for admission to colleges and for government jobs as in the Soviet Union under Joseph Stalin or Nikita Khrushchev. Private sector businesses did not grow as the taxes on private enterprise were often prohibitive. Inexperienced management sometimes ruled policy and controlled decisions by brute force. Strikes were forbidden, owners/managers were not permitted to make changes or decision which would impact their productivity or profit. The economy developed into a type of socialism called "samoupravljanje" (self-management), in which workers controlled socially owned enterprises. This kind of market socialism created significantly better economic conditions than in the Eastern Bloc countries. Croatia went through intensive industrialization in the 1960s and 1970s with industrial output increasing several-fold and with Zagreb surpassing Belgrade for the amount of industry. Factories and other organizations were often named after Partisans who were declared national heroes. This practice also spread to street names, names of parks and buildings, and some more trivial features. Before World War II, Croatia's industry was not significant, with the vast majority of the people employed in agriculture. By 1991 the country was completely transformed into a modern industrialized state. By the same time, the Croatian Adriatic coast had taken shape as an internationally popular tourist destination, all coastal republics (but mostly SR Croatia) profited greatly from this, as tourist numbers reached levels still unsurpassed in modern Croatia. The government brought unprecedented economic and industrial growth, high levels of social security and a very low crime rate. The country completely recovered from WWII and achieved a very high GDP and economic growth rate, significantly higher than those of the present-day republic. The constitution of 1963 balanced the power in the country between the Croats and the Serbs and alleviated imbalance coming from the fact that the Croats were again in a minority position. Trends after 1965 (like the fall of OZNA and UDBA chief Aleksandar Ranković from power in 1966), however, led to the Croatian Spring of 1970–71, when students in Zagreb organized demonstrations to achieve greater civil liberties and greater Croatian autonomy. The regime stifled the public protest and incarcerated the leaders, but this led to the ratification of a new constitution in 1974, giving more rights to the individual republics. At that time, radical Ustaše cells of Croatian émigrés in Western Europe planned and carried out guerilla acts inside Yugoslavia, but they were largely countered. In 1980, after Tito's death, economic, political, and religious difficulties started to mount and the federal government began to crumble. The crisis in Kosovo and, in 1986, the emergence of Slobodan Milošević in Serbia provoked a very negative reaction in Croatia and Slovenia; politicians from both republics feared that his motives would threaten their republics' autonomy. With the climate of change throughout Eastern Europe during the 1980s, the communist hegemony was challenged (at the same time, the Milošević government began to gradually concentrate Yugoslav power in Serbia and calls for free multi-party elections were becoming louder). In June 1989 the Croatian Democratic Union (HDZ) was founded by Croatian nationalist dissidents led by Franjo Tuđman, a former fighter in Tito's Partisan movement and JNA General. At this time Yugoslavia was still a one-party state and open manifestations of Croatian nationalism were dangerous so a new party was founded in an almost conspiratorial manner. It was only on 13 December 1989 that the governing League of Communists of Croatia agreed to legalize opposition political parties and hold free elections in the spring of 1990. On 23 January 1990 at its 14th Congress the Communist League of Yugoslavia voted to remove its monopoly on political power, but the same day effectively ceased to exist as a national party when the League of Communists of Slovenia walked out after Serbia's Slobodan Milošević blocked all their reformist proposals—the League of Communists of Croatia walked out soon after. On 22 April and 7 May 1990, the first free multi-party elections were held in Croatia. Franjo Tuđman's Croatian Democratic Union (HDZ) won by a 42% margin against Ivica Račan's reformed communist Party of Democratic Change (SDP) who won 26%. However, Croatia's first-past-the-post election system enabled Tuđman to form the government relatively independently as the win translated into 205 mandates (out of 351 total). The HDZ's intentions were to secure independence for Croatia, contrary to the wishes of a part of the ethnic Serbs in the republic, and federal politicians in Belgrade. The excessively polarized climate soon escalated into complete estrangement between the two nations and spiralled into sectarian violence. On 25 July 1990, a Serbian Assembly was established in Srb, north of Knin, as the political representation of the Serbian people in Croatia. The Serbian Assembly declared "sovereignty and autonomy of the Serb people in Croatia". Their position was that if Croatia could secede from Yugoslavia, then the Serbs could secede from Croatia. Milan Babić, a dentist from the southern town of Knin, was elected president. The rebel Croatian Serbs established a number of paramilitary militias under the leadership of Milan Martić, the police chief in Knin. On 17 August 1990, the Serbs began what became known as the Log Revolution, where barricades of logs were placed across roads throughout the South as an expression of their secession from Croatia. This effectively cut Croatia in two, separating the coastal region of Dalmatia from the rest of the country. The Croatian government responded to the blockade of roads by sending special police teams in helicopters to the scene, but they were intercepted by SFR Yugoslav Air Force fighter jets and forced to turn back to Zagreb. The Croatian constitution was passed in December 1990 categorizing Serbs as a minority group along with other ethnic groups. Babić's administration announced the creation of a Serbian Autonomous Oblast of Krajina (or "SAO Krajina") on 21 December 1990. Other Serb-dominated communities in eastern Croatia announced that they would also join SAO Krajina and ceased paying taxes to the Zagreb government. On Easter Sunday, 31 March 1991, the first fatal clashes occurred when Croatian police from the Croatian Ministry of the Interior (MUP) entered the Plitvice Lakes national park to expel rebel Serb forces. Serb paramilitaries ambushed a bus carrying Croatian police into the national park on the road north of Korenica, sparking a day-long gun battle between the two sides. During the fighting, two people, one Croat and one Serb policeman, were killed. Twenty other people were injured and twenty-nine Krajina Serb paramilitaries and policemen were taken prisoner by Croatian forces. Among the prisoners was Goran Hadžić, later to become the President of the Republic of Serbian Krajina. On 2 May 1991 the Croatian parliament voted to hold a referendum on independence. On 19 May 1991, on an almost 80% turnout, 93.24% voted for independence. Krajina boycotted the referendum. They held their own referendum a week earlier on 12 May 1991 in the territories they controlled and voted to remain in Yugoslavia which the Croatian government did not recognize as valid. On 25 June 1991 the Croatian Parliament declared independence from Yugoslavia. Slovenia declared independence from Yugoslavia on the same day. The civilian population fled the areas of armed conflict "en masse": generally speaking, hundreds of thousands of Croats moved away from the Bosnian and Serbian border areas. In many places, masses of civilians were forced out by the Yugoslav National Army (JNA), who consisted mostly of conscripts from Serbia and Montenegro, and irregulars from Serbia, in what became known as ethnic cleansing. The border city of Vukovar underwent a three-month siege – the Battle of Vukovar – during which most of the city was destroyed and a majority of the population was forced to flee. The city fell to the Serbian forces on 18 November 1991 and the Vukovar massacre occurred. Subsequent UN-sponsored cease-fires followed, and the warring parties mostly entrenched. The Yugoslav People's Army retreated from Croatia into Bosnia and Herzegovina where a new cycle of tensions were escalating: the Bosnian War was to start. During 1992 and 1993, Croatia also handled an estimated 700,000 refugees from Bosnia, mainly Bosnian Muslims. Armed conflict in Croatia remained intermittent and mostly on a small scale until 1995. In early August, Croatia embarked on Operation Storm. The Croatian attack quickly reconquered most of the territories from the Republic of Serbian Krajina authorities, leading to a mass exodus of the Serbian population. An estimated 90,000–200,000 Serbs fled shortly before, during and after the operation. As a result of this operation, a few months later the Bosnian war ended with the negotiation of the Dayton Agreement. A peaceful integration of the remaining Serbian-controlled territories in eastern Slavonia was completed in 1998 under UN supervision. The majority of the Serbs who fled from the former Krajina have not returned due to fears of ethnic violence, discrimination and property repossession problems, and the Croatian government has yet to achieve the conditions for full reintegration. According to United Nations High Commissioner for Refugees, around 125,000 ethnic Serbs who fled the 1991–1995 conflict are registered as having returned to Croatia, of whom around 55,000 remain permanently. Croatia became a member of the Council of Europe in 1996. The years 1996 and 1997 were a period of post-war recovery and improving economic conditions. However, in 1998 and 1999 Croatia experienced an economic depression, which brought unemployment for thousands of citizens. The remaining part of former "Krajina", areas adjacent to FR Yugoslavia, negotiated a peaceful reintegration process with the Croatian government. The so-called Erdut Agreement made the area a temporary protectorate of the UN Transitional Administration for Eastern Slavonia, Baranja and Western Sirmium. The area was formally re-integrated into Croatia by 1998. Franjo Tuđman's government started to lose popularity as it was criticized (among other things) for its involvement in suspicious privatization deals of the early 1990s as well as a partial international isolation. The country experienced a mild recession in 1998 and 1999. Tuđman died in 1999 and in the early 2000 parliamentary elections, the nationalist Croatian Democratic Union (HDZ) government was replaced by a center-left coalition under the Social Democratic Party of Croatia, with Ivica Račan as prime minister. At the same time, presidential elections were held which were won by a moderate, Stjepan Mesić. The new Račan government amended the constitution, changing the political system from a presidential system to a parliamentary system, transferring most executive presidential powers from the president to the institutions of the parliament and the prime minister. The new government also started several large building projects, including state-sponsored housing, more rebuilding efforts to enable refugee return, and the building of the A1 highway. The country achieved notable economic growth during these years, while the unemployment rate would continue to rise until 2001 when it finally started falling. Croatia became a World Trade Organization (WTO) member in 2000 and started the Accession of Croatia to the European Union in 2003. In late 2003, new parliamentary elections were held and a reformed HDZ party won under the leadership of Ivo Sanader, who became prime minister. European accession was delayed by controversies over the extradition of army generals to the ICTY, including the runaway Ante Gotovina. Sanader was reelected in the closely contested 2007 parliamentary election. Other complications continued to stall the EU negotiating process, most notably Slovenia's blockade of Croatia's EU accession in 2008–2009. In June 2009, Sanader abruptly resigned his post and named Jadranka Kosor in his place. Kosor introduced austerity measures to counter the economic crisis and launched an anti-corruption campaign aimed at public officials. In late 2009, Kosor signed an agreement with Borut Pahor, the premier of Slovenia, that allowed the EU accession to proceed. In the Croatian presidential election, 2009–2010, Ivo Josipović, the candidate of the SDP won a landslide victory. Sanader tried to come back into HDZ in 2010, but was then ejected, and USKOK soon had him arrested on several corruption charges. , his trial is ongoing. In 2011, the accession agreement was concluded, giving Croatia the all-clear to join. The 2011 Croatian parliamentary election was held on 4 December 2011, and the Kukuriku coalition won. Following the ratification of the Treaty of Accession 2011 and the successful 2012 Croatian European Union membership referendum, Croatia joined the EU on 1 July 2013. In the 2014–15 Croatian presidential election, Kolinda Grabar-Kitarović went to become the first Croatian female President. The 2015 Croatian parliamentary election resulted in the victory of the Patriotic Coalition which formed a new government with the Bridge of Independent Lists. However, a vote of no confidence brought down the Cabinet of Tihomir Orešković. After the 2016 Croatian parliamentary election, the Cabinet of Andrej Plenković was formed. General:
https://en.wikipedia.org/wiki?curid=5574
Geography of Croatia The Geography of Croatia is defined by its location—it is described as a part of Central Europe and Southeast Europe, a part of the Balkans and Mitteleuropa. Croatia's territory covers , making it the 127th largest country in the world. Bordered by Bosnia and Herzegovina and Serbia in the east, Slovenia in the west, Hungary in the north and Montenegro and the Adriatic Sea in the south, it lies mostly between latitudes 42° and 47° N and longitudes 13° and 20° E. Croatia's territorial waters encompass in a wide zone, and its internal waters located within the baseline cover an additional . The Pannonian Basin and the Dinaric Alps, along with the Adriatic Basin, represent major geomorphological parts of Croatia. Lowlands make up the bulk of Croatia, with elevations of less than above sea level recorded in 53.42% of the country. Most of the lowlands are found in the northern regions, especially in Slavonia, itself a part of the Pannonian Basin plain. The plains are interspersed with horst and graben structures, believed to have broken the Pliocene Pannonian Sea's surface as islands. The greatest concentration of ground at relatively high elevations is found in the Lika and Gorski Kotar areas in the Dinaric Alps, but high areas are found in all regions of Croatia to some extent. The Dinaric Alps contain the highest mountain in Croatia— Dinara—as well as all other mountains in Croatia higher than . Croatia's Adriatic Sea mainland coast is long, while its 1,246 islands and islets encompass a further of coastline—the most indented coastline in the Mediterranean. Karst topography makes up about half of Croatia and is especially prominent in the Dinaric Alps, as well as throughout the coastal areas and the islands. 62% of Croatia's territory is encompassed by the Black Sea drainage basin. The area includes the largest rivers flowing in the country: the Danube, Sava, Drava, Mur and Kupa. The remainder belongs to the Adriatic Sea drainage basin, where the largest river by far is the Neretva. Most of Croatia has a moderately warm and rainy continental climate as defined by the Köppen climate classification. The mean monthly temperature ranges between and . Croatia has a number of ecoregions because of its climate and geomorphology, and the country is consequently among the most biodiverse in Europe. There are four types of biogeographical regions in Croatia: Mediterranean along the coast and in its immediate hinterland; Alpine in the elevated Lika and Gorski Kotar; Pannonian along the Drava and Danube; and Continental in the remaining areas. There are 444 protected natural areas in Croatia, encompassing 8.5% of the country; there are about 37,000 known species in Croatia, and the total number of species is estimated to be between 50,000 and 100,000. The permanent population of Croatia by the 2011 census reached 4.29 million. The population density was 75.8 inhabitants per square kilometre, and the overall life expectancy in Croatia at birth was 75.7 years. The country is inhabited mostly by Croats (89.6%), while minorities include Serbs (4.5%), and 21 other ethnicities (less than 1% each) recognised by the constitution. Since the counties were re-established in 1992, Croatia is divided into 20 counties and the capital city of Zagreb. The counties subdivide into 127 cities and 429 municipalities. The average urbanisation rate in Croatia stands at 56%, with a growing urban population and shrinking rural population. The largest city and the nation's capital is Zagreb, with an urban population of 686,568 in the city itself and a metropolitan area population of 978,161. The populations of Split and Rijeka exceed 100,000, and five more cities in Croatia have populations over 50,000. Croatia's territory covers , making it the 127th largest country in the world. The physical geography of Croatia is defined by its location—it is described as a part of Southeast Europe Croatia borders Bosnia–Herzegovina (for 1,009.1 km) and Serbia (for 317.6 km) in the east, Slovenia for 667.8 km in the west, Hungary for 355.5 km in the north and Montenegro for 22.6 km and the Adriatic Sea in the south. It lies mostly between latitudes 42° and 47° N and longitudes 13° and 20° E. Part of the extreme south of Croatia is separated from the rest of the mainland by a short coastline strip around Neum belonging to Bosnia–Herzegovina. Croatia's border with Hungary was inherited from Yugoslavia. Much of the border with Hungary follows the Drava River or its former river bed; that part of the border dates from the Middle Ages. The border in Međimurje and Baranya was defined as a border between the Kingdom of Hungary and the Kingdom of Serbs, Croats, and Slovenes, later renamed the Kingdom of Yugoslavia, pursuant to the Treaty of Trianon of 1920. The present outline of the border with Bosnia–Herzegovina and border with Montenegro is largely the result of the Ottoman conquest and subsequent recapture of territories in the Great Turkish War of 1667–1698 formally ending with the Treaty of Karlowitz, as well as the Fifth and Seventh Ottoman–Venetian Wars. This border had minor modifications in 1947 when all borders of the former Yugoslav constituent republics were defined by demarcation commissions implementing the AVNOJ decisions of 1943 and 1945 regarding the federal organisation of Yugoslavia. The commissions also defined Baranya and Međimurje as Croatian territories, and moreover set up the present-day border between Serbia and Croatia in Syrmia and along the Danube River between Ilok and the Drava river's mouth and further north to the Hungarian border; the Ilok/Drava section matched the border between the Kingdom of Croatia-Slavonia and Bács-Bodrog County that existed until 1918 (the end of World War I). Most of the border with Slovenia was also defined by the commissions, matching the northwestern border of the Kingdom of Croatia-Slavonia, and establishing a new section of Croatian border north of the Istrian peninsula according to the ethnic composition of the territory previously belonging to the Kingdom of Italy. Pursuant to the 1947 Treaty of Peace with Italy the islands of Cres, Lastovo and Palagruža and the cities of Zadar and Rijeka and most of Istria went to communist Yugoslavia and Croatia, while carving out the Free Territory of Trieste (FTT) as a city-state. The FTT was partitioned in 1954 as Trieste itself and the area to the north of it were placed under Italian control, and the rest under Yugoslav control. The arrangement was made permanent by the Treaty of Osimo in 1975. The former FTT's Yugoslav part was partitioned between Croatia and Slovenia, largely conforming to the area population's ethnic composition. In the late 19th century, Austria-Hungary established a geodetic network, for which the elevation benchmark was determined by the Adriatic Sea's average level at the Sartorio pier in Trieste. This benchmark was subsequently retained by Austria, adopted by Yugoslavia, and kept by the states that emerged after its dissolution, including Croatia. The geographical extreme points of Croatia are Žabnik in Međimurje County as the northernmost point, Rađevac near Ilok in Vukovar-Syrmia County as the easternmost point, Cape Lako near Bašanija in Istria County as the westernmost point and the islet of Galijula in Palagruža archipelago in Split-Dalmatia County as the southernmost point. On the mainland, Cape Oštra of the Prevlaka peninsula in Dubrovnik-Neretva County is the southernmost point. Italy and Yugoslavia defined their delineation of the continental shelf in the Adriatic Sea in 1968, with an additional agreement on the boundary in the Gulf of Trieste signed in 1975 in accordance with the Treaty of Osimo. All the successor states of former Yugoslavia accepted the agreements. Prior to Yugoslavia's break-up, Albania, Italy and Yugoslavia initially proclaimed territorial waters, subsequently reduced to the international-standard ; all sides adopted baseline systems. Croatia also declared its Ecological and Fisheries Protection Zone (ZERP)—a part of its Exclusive Economic Zone—as extending to the continental shelf boundary. Croatia's territorial waters encompass ; its internal waters located within the baseline cover an additional . Croatia and Slovenia started negotiations to define maritime borders in the Gulf of Piran in 1992 but failed to agree, resulting in a dispute. Both countries also declared their economic zones, which partially overlap. Croatia's application to become an EU member state was initially suspended pending resolution of its border disputes with Slovenia. These were eventually settled with an agreement to accept the decision of an international arbitration commission set up via the UN, enabling Croatia to progress towards EU membership. The dispute has caused no major practical problems in areas other than the EU membership negotiations progress, even before the arbitration agreement. The maritime boundary between Bosnia–Herzegovina and Croatia was formally settled in 1999, but a few issues are still contested—the Klek peninsula and two islets in the border area. The Croatia–Montenegro maritime boundary is disputed in the Bay of Kotor, at the Prevlaka peninsula. The situation was exacerbated by the peninsula's occupation by the Yugoslav People's Army and later by the Serbian-Montenegrin army, which in turn was replaced by a United Nations observer mission that lasted until 2002. Croatia took over the area with an agreement that allowed Montenegrin presence in Croatian waters in the bay, and the dispute has become far less contentious since the independence of Montenegro in 2006. The land border disputes pertain to comparatively small strips of land. The Croatia–Slovenia border disputes are: along the Dragonja River's lower course where Slovenia claims three hamlets on the river's left bank; the Sveta Gera peak of Žumberak where exact territorial claims were never made and appear to be limited to a military barracks on the peak itself; and along the Mura River where Slovenia wants the border to be along the current river bed instead of along a former one and claims a (largely if not completely uninhabited) piece of land near Hotiza. These claims are likewise in the process of being settled by binding arbitration. There are also land border disputes between Croatia and Serbia. The two countries presently control one bank of the present-day river each, but Croatia claims that the border line should follow the cadastral borders between the former municipalities of SR Croatia and SR Serbia along the Danube, as defined by a Yugoslav commission in 1947 (effectively following a former river bed); borders claimed by Croatia also include the Vukovar and Šarengrad islands in the Danube as its territory. There is also a border dispute with Bosnia–Herzegovina, specifically Croatia claims Unčica channel on the right bank of Una as the border at Hrvatska Kostajnica, while Bosnia and Herzegovina claims Una River course as the border there. Most of Croatia is lowlands, with elevations of less than above sea level recorded in 53.42% of the country. Most of the lowlands are found in the country's northern regions, especially in Slavonia, representing a part of the Pannonian Basin. Areas with elevations of above sea level encompass 25.61% of Croatia's territory, and the areas between above sea level cover 17.11% of the country. A further 3.71% of the land is above sea level, and only 0.15% of Croatia's territory is elevated greater than above sea level. The greatest concentration of ground at relatively high elevations is found in the Lika and Gorski Kotar areas in the Dinaric Alps, but such areas are found in all regions of Croatia to some extent. The Pannonian Basin and the Dinaric Alps, along with the Adriatic Basin, represent major geomorphological parts of Croatia. Croatia's Adriatic Sea mainland coast is long, while its 1,246 islands and islets have a further of coastline. The distance between the extreme points of Croatia's coastline is . The number of islands includes all islands, islets, and rocks of all sizes, including ones emerging only at low tide. The largest islands in the Adriatic are Cres and Krk, each covering ; the tallest is Brač, reaching above sea level. The islands include 47 permanently inhabited ones, the most populous among them being Krk and Korčula. The shore is the most indented coastline in the Mediterranean. The majority of the coast is characterised by a karst topography, developed from the Adriatic Carbonate Platform. Karstification there largely began after the final raising of the Dinarides in the Oligocene and Miocene epochs, when carbonate rock was exposed to atmospheric effects such as rain; this extended to below the present sea level, exposed during the Last Glacial Maximum's sea level drop. It is estimated that some karst formations are related to earlier drops of sea level, most notably the Messinian salinity crisis. The eastern coast's largest part consists of carbonate rocks, while flysch rock is significantly represented in the Gulf of Trieste coast, on the Kvarner Gulf coast opposite Krk, and in Dalmatia north of Split. There are comparably small alluvial areas of the Adriatic coast in Croatia—most notably the Neretva river delta. Western Istria is gradually subsiding, having sunk about in the past 2,000 years. In the Middle Adriatic Basin, there is evidence of Permian volcanism in the area of Komiža on the island of Vis, in addition to the volcanic islands of Jabuka and Brusnik. Earthquakes are frequent in the area around the Adriatic Sea, although most are too faint to be felt; an earthquake doing significant damage happens every few decades, with major earthquakes every few centuries. The Dinaric Alps are linked to a Late Jurassic to recent times fold and thrust belt, itself part of the Alpine orogeny, extending southeast from the southern Alps. The Dinaric Alps in Croatia encompass the entire Gorski Kotar and Lika regions, as well as considerable parts of Dalmatia, with their northeastern edge running from Žumberak to the Banovina region, along the Sava River, and their westernmost landforms being Ćićarija and Učka mountains in Istria. The Dinaric Alps contain the highest mountain in Croatia— Dinara—as well as all other mountains in Croatia higher than : Biokovo, Velebit, Plješivica, Velika Kapela, Risnjak, Svilaja and Snježnik. Karst topography makes up about half of Croatia and is especially prominent in the Dinaric Alps. There are numerous caves in Croatia, 49 of which are deeper than , 14 deeper than and 3 deeper than . The longest cave in Croatia, Kita Gaćešina, is at the same time the longest cave in the Dinaric Alps at . The Pannonian Basin took shape through Miocenian thinning and subsidence of crust structures formed during the Late Paleozoic Variscan orogeny. The Paleozoic and Mesozoic structures are visible in Papuk and other Slavonian mountains. The processes also led to the formation of a stratovolcanic chain in the basin 12–17 Mya; intensified subsidence was observed until 5 Mya as well as flood basalts at about 7.5 Mya. The contemporary tectonic uplift of the Carpathian Mountains severed water flow to the Black Sea and the Pannonian Sea formed in the basin. Sediments were transported to the basin from the uplifting Carpathian and Dinaric mountains, with particularly deep fluvial sediments being deposited in the Pleistocene epoch during the Transdanubian Mountains' formation. Ultimately, up to of sediment was deposited in the basin, and the sea eventually drained through the Iron Gate gorge. The results are large plains in eastern Slavonia's Baranya and Syrmia regions, as well as in river valleys, especially along the Sava, Drava and Kupa. The plains are interspersed by horst and graben structures, believed to have broken the Pannonian Sea's surface as islands. The tallest among such landforms are Ivanšćica and Medvednica north of Zagreb—both are also at least partially in Hrvatsko Zagorje—as well as Psunj and Papuk that are the tallest among the Slavonian mountains surrounding Požega. Psunj, Papuk and adjacent Krndija consist mostly of Paleozoic rocks from 300–350 Mya. Požeška gora, adjacent to Psunj, consists of much more recent Neogene rocks, but there are also Upper Cretaceous sediments and igneous rocks forming the main, ridge of the hill; these represent the largest igneous landform in Croatia. A smaller piece of igneous terrain is also present on Papuk, near Voćin. The two, as well as the Moslavačka gora mountains, are possibly remnants of a volcanic arc from the same tectonic plate collision that caused the Dinaric Alps. The largest part of Croatia—62% of its territory—is encompassed by the Black Sea drainage basin. The area includes the largest rivers flowing in the country: the Danube, Sava, Drava, Mura and Kupa. The rest belongs to the Adriatic Sea drainage basin, where the largest river by far is the Neretva. The longest rivers in Croatia are the Sava, Drava, Kupa and a section of the Danube. The longest rivers emptying into the Adriatic Sea are the Cetina and an only section of the Neretva. The largest lakes in Croatia are Lake Vrana located in the northern Dalmatia, Lake Dubrava near Varaždin, Peruća Lake (reservoir) on the Cetina River, Lake Prokljan near Skradin and Lake Varaždin reservoir through which the Drava River flows near Varaždin. Croatia's most famous lakes are the Plitvice lakes, a system of 16 lakes with waterfalls connecting them over dolomite and limestone cascades. The lakes are renowned for their distinctive colours, ranging from turquoise to mint green, grey or blue. Croatia has a remarkable wealth in terms of wetlands. Four of those are included in the Ramsar list of internationally important wetlands: Lonjsko Polje along the Sava and Lonja rivers near Sisak, Kopački Rit at the confluence of the Drava and Danube, the Neretva Delta and Crna Mlaka near Jastrebarsko. Average annual precipitation and evaporation rates are and , respectively. Taking into consideration the overall water balance, the total Croatian water resources amount to per year per capita, including per year per capita from sources inside Croatia. Most of Croatia has a moderately warm and rainy oceanic climate (Cfb) as defined by the Köppen climate classification. Mean monthly temperatures range between (in January) and (in July). The coldest parts of the country are Lika and Gorski Kotar where a snowy forested climate is found at elevations above . The warmest areas of Croatia are at the Adriatic coast and especially in its immediate hinterland, which are characterised by a Mediterranean climate since temperatures are moderated by the sea. Consequently, temperature peaks are more pronounced in the continental areas: the lowest temperature of was recorded on 4 February 1929 in Gospić, and the highest temperature of was recorded on 5 August 1981 in Ploče. The mean annual precipitation is depending on the geographic region and prevailing climate type. The least precipitation is recorded in the outer islands (Vis, Lastovo, Biševo, and Svetac) and in the eastern parts of Slavonia; however, in the latter case it is mostly during the growing season. The most precipitation is observed on the Dinara mountain range and in Gorski Kotar, where some of the highest annual precipitation totals in Europe occur. The prevailing winds in the interior are light to moderate northeast or southwest; in the coastal area, the prevailing winds are determined by local area features. Higher wind velocities are more often recorded in cooler months along the coast, generally as cool northeasterly buras or, less frequently, as warm southerly jugos. The sunniest parts of the country are the outer islands, Hvar and Korčula, where more than 2,700 hours of sunshine are recorded per year, followed by the southern Adriatic Sea area in general, northern Adriatic coast, and Slavonia, all with more than 2,000 hours of sunshine per year. Croatia can be subdivided between a number of ecoregions because of its climate and geomorphology, and the country is consequently one of the richest in Europe in terms of biodiversity. There are four types of biogeographical regions in Croatia: Mediterranean along the coast and in its immediate hinterland, Alpine in most of Lika and Gorski Kotar, Pannonian along the Drava and Danube, and continental in the remaining areas. Among the most significant are karst habitats; these include submerged karst, such as Zrmanja and Krka canyons and tufa barriers, as well as underground habitats. The karst geology has produced approximately 7,000 caves and pits, many of which are inhabited by troglobitic (exclusively cave-dwelling) animals such as the olm, a cave salamander and the only European troglobitic vertebrate. Forests are also significant in the country, as they cover representing 46.8% of Croatia's land surface. The other habitat types include wetlands, grasslands, bogs, fens, scrub habitats, coastal and marine habitats. In terms of phytogeography, Croatia is part of the Boreal Kingdom; specifically, it is part of the Illyrian and Central European provinces of the Circumboreal Region and the Adriatic province of the Mediterranean Region. The World Wide Fund for Nature divides land in Croatia into three ecoregions—Pannonian mixed forests, Dinaric Mountains mixed forests and Illyrian deciduous forests. Biomes in Croatia include temperate broadleaf/mixed forest and Mediterranean forests, woodlands and scrub; all are in the Palearctic realm. Croatia has 38,226 known taxa, 2.8% of which are endemic; the actual number (including undiscovered species) is estimated to be between 50,000 and 100,000. The estimate is supported by nearly 400 new taxa of invertebrates discovered in Croatia in 2000–2005 alone. There are more than a thousand endemic species, especially in the Velebit and Biokovo mountains, Adriatic islands and karst rivers. Legislation protects 1,131 species. Indigenous cultivars of plants and breeds of domesticated animals are also numerous; they include five breeds of horses, five breeds of cattle, eight breeds of sheep, two breeds of pigs and a poultry breed. Even the indigenous breeds include nine endangered or critically endangered ones. There are 444 Croatian protected areas, encompassing 8.5% of the country. These include 8 national parks, 2 strict reserves and 11 nature parks, accounting for 78% of the total protected area. The most famous protected area and the oldest national park in Croatia is the Plitvice Lakes National Park, a UNESCO World Heritage Site. Velebit Nature Park is a part of the UNESCO Man and the Biosphere Programme. The strict and special reserves, as well as the national and nature parks, are managed and protected by the central government, while other protected areas are managed by counties. In 2005, the National Ecological Network was set up as the first step in preparation for EU membership and joining the Natura 2000 network. Habitat destruction represents a threat to biodiversity in Croatia, as developed and agricultural land is expanded into previous natural habitats, while habitat fragmentation occurs as roads are created or expanded. A further threat to biodiversity is the introduction of invasive species, with "Caulerpa racemosa" and "C. taxifolia" identified as especially problematic ones. The invasive algae are monitored and regularly removed to protect the benthic habitat. Agricultural monocultures have also been identified as a threat to biodiversity. The ecological footprint of Croatia's population and industry varies significantly between the country's regions since 50% of the population resides in 26.8% of the nation's territory, with a particularly high impact made by the city of Zagreb and Zagreb County areas—their combined area comprises 6.6% of Croatia's territory while encompassing 25% of the population. The ecological footprint is most notably from the increased development of settlements and the sea coast leading to habitat fragmentation. Between 1998 and 2008, the greatest changes of land use pertained to artificially developed areas, but the scale of development is negligible compared to EU member states. The Croatian Environment Agency (CEA), a public institution established by the Government of Croatia to collect and analyse information on the environment, has identified further ecological problems as well as various degrees of progress in terms of curbing their environmental impact. These problems include inadequate legal landfills as well as the presence of illegal landfills; between 2005 and 2008, 62 authorised and 423 illegal landfills were rehabilitated. In the same period, the number of issued waste management licences doubled, while the annual municipal solid waste volume increased by 23%, reaching per capita. The processes of soil acidification and organic matter degradation are present throughout Croatia, with increasing soil salinity levels in the Neretva river plain and spreading areas of alkali soil in Slavonia. Croatian air pollution levels reflect the drop in industrial production recorded in 1991 at the onset of the Croatian War of Independence—pre-war emission levels were only reached in 1997. The use of desulfurised fuels has led to a 25% reduction of sulphur dioxide emissions between 1997 and 2004, and a further 7.2% drop by 2007. The rise in NOx emissions halted in 2007 and reversed in 2008. The use of unleaded petrol reduced emissions of lead into the atmosphere by 91.5% between 1997 and 2004. Air quality measurements indicate that the air in rural areas is essentially clean, and in urban centres it generally complies with legal requirements. The most significant sources of greenhouse gas (GHG) emissions in Croatia are energy production (72%), industry (13%) and agriculture (11%). The average annual increase of GHG emissions is 3%, remaining within the Kyoto Protocol limits. Between 1990 and 2007, the use of ozone depleting substances was reduced by 92%; their use is expected to be abolished by 2015. Even though Croatia has sufficient water resources at its disposal, these are not uniformly distributed and public water supply network losses remain high—estimated at 44%. Between 2004 and 2008, the number of stations monitoring surface water pollution increased by 20%; the CEA reported 476 cases of water pollution in this period. At the same time organic waste pollution levels decreased slightly, which is attributed to the completion of new sewage treatment plants; their number increased 20%, reaching a total of 101. Nearly all of Croatia's groundwater aquifers are top quality, unlike the available surface water; the latter's quality varies in terms of biochemical oxygen demand and bacteriological water analysis results. As of 2008, 80% of the Croatian population are served by the public water supply system, but only 44% of the population have access to the public sewerage network, with septic systems in use. Adriatic Sea water quality monitoring between 2004 and 2008 indicated very good, oligotrophic conditions along most of the coast, while areas of increased eutrophication were identified in the Bay of Bakar, the Bay of Kaštela, the Port of Šibenik and near Ploče; other areas of localised pollution were identified near the larger coastal cities. In the period between 2004 and 2008, the CEA identified 283 cases of marine pollution (including 128 from vessels), which was a drop of 15% relative to the period encompassed by the previous report, 1997 to August 2005. As of 2006, 46.8% of Croatia was occupied by of forest and shrub, while a further or 40.4% of the land was used for diverse agricultural uses including , or 7.8% of the total, for permanent crops. Bush and grass cover was present on or 8.4% of the territory, inland waters took up or 1.0% and marshes covered or 0.4% of the country. Artificial surfaces (primarily consisting of urban areas, roads, non-agricultural vegetation, sports areas and other recreational facilities) took up or 3.1% of the country's area. The greatest impetus for land use changes is the expansion of settlements and road construction. Because of the Croatian War of Independence, there are numerous leftover minefields in Croatia, largely tracing former front lines. As of 2006, suspected minefields covered . As of 2012, 62% of the remaining minefields are situated in forests, 26% of them are found in agricultural land, and 12% are found in other land; it is expected that mine clearance will be complete by 2019. Croatia is traditionally divided into numerous, often overlapping geographic regions, whose borders are not always clearly defined. The largest and most readily recognisable ones throughout the country are Central Croatia (also described as the Zagreb macro-region), Eastern Croatia (largely corresponding with Slavonia), and Mountainous Croatia (Lika and Gorski Kotar; to the west of Central Croatia). These three comprise the inland or continental part of Croatia. Coastal Croatia consists of a further two regions: Dalmatia or the southern littoral, between the general area of the city of Zadar and the southernmost tip of the country; and the northern littoral located north of Dalmatia, encompassing the Croatian Littoral and Istria. The geographical regions generally do not conform to county boundaries or other administrative divisions, and all of them encompass further, more specific, geographic regions. The demographic features of the Croatian population are known through censuses, normally conducted in ten-year intervals and analysed by various statistical bureaus since the 1850s. The Croatian Bureau of Statistics has performed this task since the 1990s. The latest census in Croatia was performed in April 2011. The permanent population of Croatia at the 2011 census had reached 4.29 million. The population density was 75.8 inhabitants per square kilometre, and the overall life expectancy in Croatia at birth is 75.7 years. The population rose steadily (with the exception of censuses taken following the two world wars) from 2.1 million in 1857 until 1991, when it peaked at 4.7 million. Since 1991, Croatia's death rate has continuously exceeded its birth rate; the natural growth rate of the population is thus currently negative. Croatia is currently in the demographic transition's fourth or fifth stage. In terms of age structure, the population is dominated by the 15‑ to 64‑year‑old segment. The median age of the population is 41.4, and the gender ratio of the total population is 0.93 males per 1 female. Croatia is inhabited mostly by Croats (89.6%), while minorities include Serbs (4.5%) and 21 other ethnicities (less than 1% each) recognised by the Constitution of Croatia. The demographic history of Croatia is marked by significant migrations, including: the Croats' arrival in the area; the growth of the Hungarian and German speaking population after the personal union of Croatia and Hungary; joining of the Habsburg Empire; migrations set off by the Ottoman conquests; and the growth of the Italian-speaking population in Istria and Dalmatia during the Venetian rule there. After Austria-Hungary's collapse, the Hungarian population declined, while the German-speaking population was forced out or fled during the last part of and after World War II, and a similar fate was suffered by the Italian population. The late 19th century and the 20th century were marked by large scale economic migrations abroad. The 1940s and the 1950s in Yugoslavia were marked by internal migrations in Yugoslavia, as well as by urbanisation. The most recent significant migrations came as a result of the Croatian War of Independence when hundreds of thousands were displaced. The Croatian language is Croatia's official language, but the languages of constitutionally-recognised minorities are officially used in some local government units. Croatian is the native language identified by 96% of the population. A 2009 survey revealed that 78% of Croatians claim knowledge of at least one foreign language—most often English. The largest religions of Croatia are Roman Catholicism (86.3%), Orthodox Christianity (4.4%) and Islam (1.5%). Literacy in Croatia stands at 98.1%. The proportion of the population aged 15 and over attaining academic degrees has grown rapidly since 2001, doubling and reaching 16.7% by 2008. An estimated 4.5% of GDP is spent for education. Primary and secondary education are available in Croatian and in the languages of recognised minorities. Croatia has a universal health care system and in 2010, the nation spent 6.9% of its GDP on healthcare. The net monthly income in September 2011 averaged 5,397 kuna ( ). The most significant sources of employment in 2008 were wholesale and retail trade, the manufacturing industry and construction. In October 2011, the unemployment rate was 17.4%. Croatia's median equivalent household income tops the average Purchasing Power Standard of the ten countries which joined the EU in 2004, while trailing the EU average. The 2011 census recorded a total of 1.5 million private households; most owned their own housing. Croatia was first subdivided into counties in the Middle Ages. The divisions changed over time to reflect losses of territory to Ottoman conquest and subsequent liberation of the same territory, in addition to changes in the political status of Dalmatia, Dubrovnik and Istria. The traditional division of the country into counties was abolished in the 1920s, when the Kingdom of Serbs, Croats and Slovenes and the subsequent Kingdom of Yugoslavia introduced oblasts and banovinas respectively. Communist-ruled Croatia, as a constituent part of post-WWII Yugoslavia, abolished earlier divisions and introduced (mostly rural) municipalities, subdividing Croatia into approximately one hundred municipalities. Counties were reintroduced in 1992 by legislation, significantly altered in terms of territory relative to the pre-1920s subdivisions—for instance, in 1918 the Transleithanian part of Croatia was divided into eight counties with their seats in Bjelovar, Gospić, Ogulin, Požega, Vukovar, Varaždin, Osijek and Zagreb, while the 1992 legislation established 14 counties in the same territory. Međimurje County was established in the eponymous region acquired through the 1920 Treaty of Trianon. (The 1990 Croatian Constitution provided for a Chamber of the Counties as part of the government, and for counties themselves without specifying their names or number. However, the counties were not actually re-established until 1992, and the first Chamber of the Counties was elected in 1993.) Since the counties were re-established in 1992, Croatia has been divided into 20 counties and the capital city of Zagreb, the latter having the authority and legal status of a county and a city at the same time (Zagreb County outside the city is administratively separate as of 1997). The county borders have changed in some instances since (for reasons such as historical ties and requests by cities), with the latest revision taking place in 2006. The counties subdivide into 127 cities and 429 municipalities. The EU Nomenclature of Territorial Units for Statistics (NUTS) division of Croatia is performed in several tiers. NUTS 1 level places the entire country in a single unit, while there are three NUTS 2 regions; these are Central and Eastern (Pannonian) Croatia, Northwest Croatia and Adriatic Croatia. The last encompasses all counties along the Adriatic coast. Northwest Croatia includes the city of Zagreb and Krapina-Zagorje, Varaždin, Koprivnica-Križevci, Međimurje and Zagreb counties, and the Central and Eastern (Pannonian) Croatia includes the remaining areas—Bjelovar-Bilogora, Virovitica-Podravina, Požega-Slavonia, Brod-Posavina, Osijek-Baranja, Vukovar-Syrmia, Karlovac and Sisak-Moslavina counties. Individual counties and the city of Zagreb represent NUTS 3 level subdivision units in Croatia. The NUTS Local administrative unit divisions are two-tiered. The LAU 1 divisions match the counties and the city of Zagreb—in effect making these the same as NUTS 3 units—while the LAU 2 subdivisions correspond to the cities and municipalities of Croatia. The average urbanisation rate in Croatia stands at 56%, with a growing urban population and shrinking rural population. The largest city and the nation's capital is Zagreb, with an urban population of 686,568 in the city itself. Zagreb's metropolitan area encompasses 341 additional settlements and, by the year 2001, the population of the area had reached 978,161; approximately 60% of Zagreb County's residents live in Zagreb's metropolitan area, as does about 41% of Croatia's urban population. The cities of Split and Rijeka are the largest settlements on the Croatian Adriatic coast, with each city's population being over 100,000. There are four other Croatian cities exceeding 50,000 people: Osijek, Zadar, Pula and Slavonski Brod; the Zagreb district of Sesvete, which has the status of a standalone settlement but not a city, also has such a large population. A further eleven cities are populated by more than 20,000.
https://en.wikipedia.org/wiki?curid=5575
Demographics of Croatia The demographic characteristics of the population of Croatia are known through censuses, normally conducted in ten-year intervals and analysed by various statistical bureaus since the 1850s. The Croatian Bureau of Statistics has performed this task since the 1990s. The latest census in Croatia was performed in April 2011. The permanent population of Croatia at the 2011 census had reached 4.29 million. The population density is 75.8 inhabitants per square kilometre, and the overall life expectancy in Croatia at birth was 78,2 years in 2018. The population rose steadily (with the exception of censuses taken following the two world wars) from 2.1 million in 1857 until 1991, when it peaked at 4.7 million. Since 1991, Croatia's death rate has continuously exceeded its birth rate; the natural growth rate of the population is negative. Croatia is in the fourth (or fifth) stage of the demographic transition. In terms of age structure, the population is dominated by the 15 to 64 year‑old segment. The median age of the population is 43.4, and the gender ratio of the total population is 0.93 males per 1 female. The country is projected to lose 350,000 citizens by 2045. Croatia is inhabited mostly by Croats (90.4%), while minorities include Serbs (4.36%), and 21 other ethnicities (less than 1% each). The demographic history of Croatia is marked by significant migrations, including the arrival of the Croats in the area growth of Hungarian and German-speaking population since the union of Croatia and Hungary, and joining of the Habsburg Empire, migrations set off by Ottoman conquests and growth of Italian speaking population in Istria and in Dalmatia during Venetian rule there. After the collapse of Austria-Hungary, the Hungarian population declined, while the German-speaking population was forced or compelled to leave after World War II and similar fate was suffered by the Italian population. Late 19th century and the 20th century were marked by large scale economic migrations abroad. The 1940s and the 1950s in Yugoslavia were marked by internal migrations in Yugoslavia, as well as by urbanisation. The most recent significant migrations came as a result of the Croatian War of Independence when hundreds of thousands were displaced. The Croatian language is the official language, but minority languages are officially used in some local government units. Croatian is declared as the native language by 95.60% of the population. A 2009 survey revealed that 78% of Croatians claim knowledge of at least one foreign language—most often English. The main religions of Croatia are Roman Catholic (86.28%), Eastern Orthodoxy (4.44%) and Islam (1.47%). Literacy in Croatia stands at 98.1%. The proportion of the population aged 15 and over attaining academic degrees grew rapidly since 2001, doubling and reaching 16.7% by 2008. An estimated 4.5% of the GDP is spent for education. Primary and secondary education are available in Croatian and in languages of recognised minorities. Croatia has a universal health care system and in 2010, the nation spent 6.9% of its GDP on healthcare. Net monthly income in September 2011 averaged 5,397 kuna ( 729 euro). The most significant sources of employment in 2008 were manufacturing industry, wholesale and retail trade and construction. In January 2020, the unemployment rate was 8.4%. Croatia's median equivalent household income tops average Purchasing Power Standard of the ten countries which joined the EU in 2004, while trailing the EU average. 2011 census recorded a total of 1.5 million private households, which predominantly owned their own housing. Average urbanisation rate in Croatia stands at 56%, with augmentation of urban population and reduction of rural population. With a population of 4,154,213 in 2016 Croatia ranks 126th in the world by population. Its population density is 75.8 inhabitants per square kilometre. The overall life expectancy in Croatia at birth is 78 years. The total fertility rate of 1.50 children per mother is one of the lowest in the world. Since 1991, Croatia's death rate has nearly continuously exceeded its birth rate. The Croatian Bureau of Statistics forecast that the population may even shrink to 3.1 million by 2051, depending on the actual birth rate and the level of net migration. The population of Croatia rose steadily from 2.1 million in 1857 until 1991, when it peaked at 4.7 million, with the exception of censuses taken in 1921 and 1948, i.e. following two world wars. The natural growth rate of the population is negative. Croatia started advancing from the first stage of the demographic transition in the late 18th and early 19th centuries (depending on where in Croatia is being discussed). Croatia is in the fourth or fifth stage of the demographic transition. One explanation for the recent population decrease is the Croatian War of Independence. During the war, large sections of the population were displaced and emigration increased. In 1991, in predominantly Serb areas, more than 400,000 Croats and other non-Serbs were either removed from their homes by the Croatian Serb forces or fled the violence. In 1995, during the final days of the war, more than 120,000 and perhaps as many as 200,000 Serbs fled the country before the arrival of Croatian forces during Operation Storm. Within a decade following the end of the war, only 117,000 Serb refugees returned out of the 300,000 displaced during the entire war. According to 2001 Croatian census there were 201,631 Serbs in Croatia, compared to the census from 1991 when the number was 581,663. Most of Croatia's remaining Serbs never lived in areas occupied in the Croatian War of Independence. Serbs have been only partially re-settled in the regions they previously inhabited, while some of the settlements previously inhabited by Serbs were settled by Croat refugees from Bosnia and Herzegovina, mostly from Republika Srpska. In 2014, there were 39,566 live births in Croatia, comprising 20,374 male and 19,192 female children. Virtually all of those were performed in medical facilities; only 19 births occurred elsewhere. Out of the total number, 32,677 children were born in wedlock or within 300 days after the end of the marriage, and the average age of mothers at the birth of their first child was 28.4 years. General fertility rate, i.e. number of births per 1,000 women aged 15–49 is 42.9, with the age specific rate peaking at 101.0 per million for women aged 25–29. In 2009, 52,414 persons died in Croatia, 48.5% of whom died in medical facilities and 90.0% of whom were receiving medical treatment at the time. Cardiovascular disease and cancer were the primary causes of death in the country, with 26,235 and 13,280 deaths respectively. In the same year, there were 2,986 violent deaths, including 2,121 due to accidents. The latter figure includes 616 deaths in traffic accidents. In 2014, the birth rate was 9.3 per mille, exceeded by the mortality rate of 12.0 per mille. The infant mortality rate was 5.0 per mille in 2014. In terms of age structure, the population of Croatia is dominated by the 15–64 year older segment (68.1%), while the size of the population younger than 15 and older than 64 is relatively small (15.1% and 16.9% respectively). The median age of the population is 41.4. The sex ratio of the population is 1.06 males per 1 female at birth and up to 14 years of age, and 0.99 males per 1 female between the ages of 15 and 64. But at ages over 64 the ratio is 0.64 males per 1 female. The ratio for the total population is 0.93 males per 1 female. In contrast to the shrinking native population, since the late 1990s there has been a positive net migration into Croatia, reaching a level of more than 7,000 net immigrants in 2006. In recent years, the Croatian government has been pressured each year to add 40% to work permit quotas for foreign workers. In accordance with its immigration policy, Croatia is also trying to entice emigrants to return. There were 8,468 immigrants to Croatia in 2009, more than half of them (57.5%) coming from Bosnia and Herzegovina, a sharp decline from the previous year's 14,541. In the same year, there were 9,940 emigrants from the country, 44.8% of them leaving to Serbia. The number of emigrants represents a substantial increase compared to the figure of 7,488 recorded in 2008. In 2009, the net migration to and from abroad peaked in the Sisak-Moslavina County (−1,093 persons) and the city of Zagreb (+830 persons). In 2009, a total of 22,382 marriages were performed in Croatia as well as 5,076 divorces. The 2001 census recorded 1.47 million households in the country. The first modern population census in the country was conducted in 1857, and 15 more have been performed since then. Since 1961 the censuses are conducted in regular ten-year intervals, with the latest one in 2011. The first institution set up in the country specifically for the purposes of maintaining population statistics was the State Statistical Office, founded in 1875. Since its founding, the office changed its name and structure several times and was alternately subordinated to other institutions and independent, until the most recent changes in 1992, when the institution became the Croatian Bureau of Statistics. The 2011 census was performed on 1–28 April 2011, recording situation as of 31 March 2011. The first census results, containing the number of the population by settlement, were published on 29 June 2011, and the final comprehensive set of data was published in December 2012. The 2011 census and processing of the data gathered by the census was expected to cost 171.9 million kuna (23.3 million euro). The 2011 census was performed using new methodology: the permanent population was determined as the enumerated population who lived in the census area for at least 12 months prior to the census, or plans to live in the same area for at least 12 months after the census. This method was also retroactively applied to the 2001 census data. The total fertility rate is the number of children born per woman. It is based on fairly good data for the entire period. Sources: Our World In Data and Gapminder Foundation. Births Deaths Natural increase Croatia is inhabited mostly by Croats (90.42%), while minority groups include Serbs (4.36%), Bosniaks, Hungarians, Italians, Albanians, Slovenes, Germans, Czechs, Romani people and others (5.22%). The Constitution of the Republic of Croatia explicitly identifies 22 minorities. Those are Serbs, Czechs, Slovaks, Italians, Hungarians, Jews, Germans, Austrians, Ukrainians, Ruthenians, Macedonians, Bosniaks, Slovenes, Montenegrins, Russians, Bulgarians, Poles, Romani, Rumanians, Turks, Vlachs and Albanians. The demographic history of Croatia is characterised by significant migrations, starting with the arrival of the Croats in the area. According to the work "De Administrando Imperio" written by the 10th-century Byzantine Emperor Constantine VII, the Croats arrived in the area of modern-day Croatia in the early 7th century. However, that claim is disputed, and competing hypotheses date the event between the 6th and the 9th centuries. Following the establishment of a personal union of Croatia and Hungary in 1102, and the joining of the Habsburg Empire in 1527, the Hungarian and German-speaking population of Croatia began gradually increasing in number. The processes of Magyarization and Germanization varied in intensity but persisted to the 20th century. The Ottoman conquests initiated a westward migration of parts of the Croatian population; the Burgenland Croats are direct descendants of some of those settlers. To replace the fleeing Croats the Habsburgs called on the Orthodox populations of Bosnia and Serbia to provide military service in the Croatian Military Frontier. Serb migration into this region peaked during the Great Serb Migrations of 1690 and 1737–39. Similarly, Venetian Republic rule in Istria and in Dalmatia, following the Fifth and the Seventh Ottoman–Venetian Wars ushered gradual growth of Italian speaking population in those areas. Following the collapse of Austria-Hungary in 1918, the Hungarian population declined, especially in the areas north of the Drava river, where they represented the majority before World War I. The period between 1890 and World War I was marked by large economic emigration from Croatia to the United States, and particularly to the areas of Pittsburgh, Pennsylvania, Cleveland, Ohio, and Chicago, Illinois. Besides the United States, the main destination of the migrants was South America, especially Argentina, Chile, Bolivia and Peru. It is estimated that 500,000 people left Croatia during this period. After World War I, the main focus of emigration shifted to Canada, where about 15,000 people settled before the onset of World War II. During World War II and in the period immediately following the war, there were further significant demographic changes as the German-speaking population, the Volksdeutsche, were either forced or otherwise compelled to leave—reducing their number from the prewar German population of Yugoslavia of 500,000, living in parts of present-day Croatia and Serbia, to the figure of 62,000 recorded in the 1953 census. A similar fate was suffered by the Italian population in Yugoslavia populating parts of present-day Croatia and Slovenia, as 350,000 left for Italy. The 1940s and the 1950s in Yugoslavia were marked by colonisation of settlements where the displaced Germans used to live by people from the mountainous parts of Bosnia and Herzegovina, Serbia and Montenegro, and migrations to larger cities spurred on by the development of industry. In the 1960s and 1970s, another wave of economic migrants left Croatia. They largely moved to Canada, Australia, New Zealand and Western Europe. During this period, 65,000 people left for Canada, and by the mid-1970s there were 150,000 Croats who moved to Australia. Particularly large European emigrant communities of Croats exist in Germany, Austria and Switzerland, which largely stem from the 1960s and 1970s migrations. The most recent significant migrations came as a result of the 1991–1995 Croatian War of Independence. In 1991, more than 400,000 Croats and other non-Serbs were displaced by the Croatian Serb forces or fled the violence in areas with significant Serb populations. During the final days of the war, in 1995, between 120,000 and 200,000 Serbs fled the country following the Operation Storm. Ten years after the war, only a small portion of Serb refugees returned out of the 400,000 displaced during the entire war. Most of the Serbs in Croatia who remained never lived in areas occupied during the Croatian War of Independence. Serbs have been only partially re-settled in the regions they previously inhabited; some of these areas were later settled by Croat refugees from Bosnia and Herzegovina. In addition to demographic losses through significant migrations, the population of Croatia suffered significant losses due to wars and epidemics. In the 20th century alone, there were several such events. The first was World War I, when the loss of the population of Croatia amounted to an estimated 190,000 persons, or about 5.5% of the total population recorded by the 1910 census. The 1918 flu pandemic started to take its toll in Croatia in July 1918, with peaks of the disease occurring in October and November. Available data is scarce, but it is estimated that the pandemic caused at least 15,000–20,000 deaths. During World War II, nearly 195,000 people died in the territory of present-day Croatia. The surrender of the armed forces of the Independent State of Croatia and of the civilians accompanying the troops of the British Army at the end of World War II was followed by the Bleiburg repatriations. A substantial number of people were executed, but the exact number is disputed. The claims range from 12,000–15,000 to as many as 80,000 killed in May 1945. Finally, approximately 20,000 were killed or went missing during the 1991–1995 Croatian War of Independence. The figure pertains only to those persons who would have been recorded by the 1991 census as living in Croatia. Demographic statistics according to the World Population Review. The following demographic statistics are from the CIA World Factbook. definition: age 15 and over can read and write (2015 est.) The Croatian language is the official language of Croatia, and one of 24 official languages of the European Union since 2013. Minority languages are in official use in local government units where more than a third of the population consists of national minorities or where local legislation mandates their use. These languages are Czech, Hungarian, Italian, Ruthenian, Serbian and Slovak. Besides these, the following languages are also recognised: Albanian, Bosnian, Bulgarian, German, Hebrew, Macedonian, Montenegrin, Polish, Romanian, Romany, Russian, Rusyn, Slovenian, Turkish and Ukrainian. According to the 2011 Census, 95.60% of citizens of Croatia declared Croatian as their native language, 1.23% declared Serbian as their native language, while no other language is represented in Croatia by more than 0.5% of native speakers among the population of Croatia. In the region of Dalmatia, each city historically spoke a variant of the Dalmatian language. It developed from Latin like all Romance languages, but became heavily influenced by Venetian and Croatian. The language fell out of use in the region by the 16th century and went extinct when the last speaker died in 1898. Croatian replaced Latin as the official language of the Croatian government in 1847. The Croatian lect is generally viewed as one of the three standard varieties of the Shtokavian dialect of Serbo-Croatian, a South Slavic language. Croatian is written using the Latin alphabet and there are three major dialects spoken on the territory of Croatia, with the Shtokavian idiom used as the literary standard. The Chakavian and Kajkavian dialects are distinguished by their lexicon, phonology, and syntax. From 1961 to 1991, the official language was formally designated as Serbo-Croatian or Croato-Serbian. Even during socialist rule, Croats often referred to their language as Croato-Serbian (instead of Serbo-Croatian) or as Croatian. Croatian and Serbian variants of the language were not officially recognised as separate at the time, but referred to as the west and east version, and had different alphabets: the Latin alphabet and Serbian Cyrillic. Croatians are protective of their language from foreign influences, as the language was under constant change and threats imposed by previous rulers (i.e. Austrian German, Hungarian, Italian and Turkish words were changed and altered to "Slavic" looking/sounding ones). A 2009 survey revealed that 78% of Croatians claim knowledge of at least one foreign language. According to a survey ordered by the European commission in 2005, 49% of Croatians speak English as their second language, 34% speak German, and 14% speak Italian. French and Russian are spoken by 4% each, and 2% of Croatians speak Spanish. A substantial proportion of Slovenes (59%) have a certain level of knowledge of the Croatian language. The main religions of Croatia are Roman Catholicism 86.28%, Eastern Orthodoxy 4.44%, Protestantism 0.34%, other Christianity 0.30%, and Islam 1.47%. In the Eurostat Eurobarometer Poll of 2005, 67% of the population of Croatia responded that "they believe there is a God" and 7% said they do not believe "there is any sort of spirit, God, or life force", while 25% expressed a belief in "some sort of spirit or life force". In a 2009 Gallup poll, 70% answered affirmatively when asked "Is religion an important part of your daily life?" Significantly, a 2008 Gallup survey of the Balkans indicated church and religious organisations as the most trusted institutions in the country. The survey revealed that 62% of the respondents assigned "a lot" or "some" trust to those institutions, ranking them ahead of all types of governmental, international or non-governmental institutions. Public schools allow religious education, in cooperation with religious communities that have agreements with the government, but attendance is not mandatory. The classes are organized widely in public elementary and secondary schools. In 2009, 92% of elementary school pupils and 87% of secondary school students attended the religious education classes. Public holidays in Croatia also include the religious festivals of Epiphany, Easter Monday, Corpus Christi Day, Assumption Day, All Saints' Day, Christmas, and St. Stephen's or Boxing Day. The religious festival public holidays are based on the Catholic liturgical year, but citizens of the Republic of Croatia who celebrate different religious holidays have the right not to work on those dates. This includes Christians who celebrate Christmas on 7 January per the Julian calendar, Muslims on the days of Ramadan Bayram and Kurban Bayram, and Jews on the days of Rosh Hashanah and Yom Kippur. Marriages performed by the religious communities having agreements with the state are officially recognized, eliminating the need to register the marriages in a registrar office. The legal position of religious communities is defined by special legislation, specifically regarding government funding, tax benefits, and religious education in schools. Other matters are left to each religious community to negotiate separately with the government. Registration of the communities is not mandatory, but registered communities become legal persons and enjoy tax and other benefits. The law stipulates that to be eligible for registration, a religious group must have at least 500 believers and be registered as a civil association for 5 years. Religious groups based abroad must submit written permission for registration from their country of origin. Literacy in Croatia is 98.1 percent. The 2001 census reported that 15.7% of the population over the age of 14 has an incomplete elementary education, and 21.9% has only an elementary school education. 42.8% of the population over the age of 14 has a vocational education and 4.9% completed gymnasium. 4.2% of the same population received an undergraduate degree, while 7.5% received an academic degree, and 0.5% received a postgraduate or a doctoral degree. Croatia recorded a substantial growth of the population attaining academic degrees and by 2008, this population segment was estimated to encompass 16.7% of the total population of Croatians 15 and over. A worldwide study about the quality of living in different countries published by "Newsweek" in August 2010 ranked the Croatian education system at 22nd, a position shared with Austria. In 2004, it was estimated that 4.5% of the GDP is spent for education, while schooling expectancy was estimated to 14 years on average. Primary education in Croatia starts at the age of six or seven and consists of eight grades. In 2007 a law was passed to increase free, noncompulsory education until 18 years of age. Compulsory education consists of eight grades of elementary school. Secondary education is provided by gymnasiums and vocational schools. As of 2010, there are 2,131 elementary schools and 713 schools providing various forms of secondary education. Primary and secondary education are also available in languages of recognised minorities in Croatia, where classes are held in Czech, Hungarian, Italian, Serbian and German languages. There are 84 elementary level and 47 secondary level music and art schools, as well as 92 schools for disabled children and youth and 74 schools for adults. Nationwide leaving exams () were introduced for secondary education students in the 2009–2010 school year. It comprises three compulsory subjects (Croatian language, mathematics, and a foreign language) and optional subjects and is a prerequisite for a university education. Croatia has eight public universities, the University of Zagreb, University of Split, University of Rijeka, University of Osijek, University of Zadar, University of Dubrovnik, University of Pula and Dubrovnik International University. The University of Zadar, the first university in Croatia, was founded in 1396 and remained active until 1807, when other institutions of higher education took over. It was reopened in 2002. The University of Zagreb, founded in 1669, is the oldest continuously operating university in Southeast Europe. There are also 11 polytechnics and 23 higher education institutions, of which 19 are private. In total, there are 132 institutions of higher education in Croatia, attended by more than 145 thousand students. There are 205 companies, government or education system institutions and non-profit organizations in Croatia pursuing scientific research and the development of technology. Combined, they spent more than 3 billion kuna (400 million euro) and employed 10,191 full-time research staff in 2008. Among the scientific institutes operating in Croatia, the largest is the Ruđer Bošković Institute in Zagreb. The Croatian Academy of Sciences and Arts in Zagreb is a learned society promoting language, culture, arts and science since its inception in 1866. Scientists from Croatia include inventors and Nobel Prize winners. Croatia has a universal health care system, the roots of which can be traced back to the Hungarian-Croatian Parliament Act of 1891, providing a form of mandatory insurance for all factory workers and craftsmen. The population is covered by a basic health insurance plan provided by statute and optional insurance. In 2014, the annual compulsory healthcare related expenditures reached 21.8 billion kuna (2.9 billion euro). Healthcare expenditures comprise only 0.6% of private health insurance and public spending. In 2010, Croatia spent 6.9% of its GDP on healthcare, representing a decline from approximately 8% estimated in 2008, when 84% of healthcare spending came from public sources. According to the World Health Organization (WHO), Croatia ranks around the 50th in the world in terms of life expectancy. There are hundreds of healthcare institutions in Croatia, including 79 hospitals and clinics with 23,967 beds. The hospitals and clinics care for more than 700 thousand patients per year and employ 5,205 medical doctors, including 3,929 specialists. There are 6,379 private practice offices, and a total of 41,271 health workers in the country. There are 63 emergency medical service units, responding to more than a million calls. The principal cause of death in 2008 was cardiovascular disease at 43.5% for men and 57.2% for women, followed by tumours, at 29.4% for men and 21.4% for women. Other significant causes of death are injuries, poisonings and other external causes (7.7% men/3.9% women), digestive system diseases (5.7% men/3.6% women), respiratory system diseases (5.1% men/3.5% women) and endocrine, nutritional and metabolic diseases (2.1% men/3.0% women). There is no other cause of disease affecting more than 3% of the population. In 2014 only 22 Croatians had been infected with HIV/AIDS and 4 had died from the disease. In 2008 it was estimated by the WHO that 27.4% of Croatians over age of 15 were smokers. According to 2003 WHO data, 22% of the Croatian adult population is obese. Net monthly income in September 2011 averaged 5,397 kuna ( 729 euro), dropping 2.1% relative to the previous month. In the same month, gross monthly income averaged 7,740 kuna ( 1,046 euro), and it includes the net salary along with income tax, retirement pension insurance, healthcare insurance, occupational safety and health insurance and employment promotion tax. The average net monthly income grew compared to 5,311 kuna ( 717 euro) in 2009 or 3,326 kuna ( 449 euro) in 2000. The highest net salaries were paid in financial services sector, and in April 2011 those averaged 10,041 kuna ( 1,356 euro), while the lowest ones, paid in the same month, were in the manufacturing and leather processing industries, averaging at 2,811 kuna ( 380 euro). Since January 2016, the minimum wage in Croatia is 3,120 kuna before tax ( 400 euro). Number of employed persons recorded steady growth between 2000 and 2008 when it peaked, followed by 4% decline in 2009. That year, there were 1.499 million employed persons, with 45% of that number pertaining to women. The total number of employed persons includes 252,000 employed in crafts and freelance professionals and 35,000 employed in agriculture. The most significant sources of employment in 2008 were manufacturing industry and wholesale and retail trade (including motor vehicle repair services) employing 278,640 and 243,640 respectively. Further significant employment sector was construction industry comprising 143,336 jobs that year. In the same year, more than 100,000 were employed in public administration, defence and compulsory social insurance sector as well as in education. Since 2009, negative trends persisted in Croatia with jobs in the industry declined further by 3.5%. Number of unemployed and retired persons combined exceeded number of employed in August 2010, as it fell to 1.474 million. In 2009, labour force consisted of 1.765 million persons out of 3.7 million working age population—aged 15 and over. In October 2011, unemployment rate stood at 17.4%. 7.2% of employed persons hold a second job. In comparison with the member states of the European Union (EU), Croatia's median equivalent household income in terms of the Purchasing Power Standard (PPS) stands at 470, topping average PPS of the ten countries which joined the EU in 2004 (EU10), as well as Romania and Bulgaria, while significantly lagging behind the EU average. Within Croatia, the highest PPS is recorded in Istria County (769), the City of Zagreb (640) and the Primorje-Gorski Kotar County (576). The lowest PPS is observed in the Bjelovar-Bilogora County and the Virovitica-Podravina County (267). 2011 census recorded a total of 1,534,148 private households in Croatia as well as 1,487 other residential communities such as retirement homes, convents etc. At the same time, there were 1,923,522 permanent housing units—houses and apartments. 2001 census recorded 1.66 million permanent housing units, including 196 thousand intermittently occupied and 42 thousand abandoned ones. Average size of a permanently used housing unit is . The intermittently used housing units include 182 thousand vacation houses and 8 thousand houses used during agricultural works. The same census also recorded 25 thousand housing units used for business purposes only. As of 2007, 71% of the households owned their own housing and had no mortgage or other loans to repay related to the housing, while further 9% were repaying loans for their housing. The households vary by type and include single households (13%), couples (15%), single parent households (4%), couples with children (27%) and extended family households (20%). There are approximately 500 homeless persons in Croatia, largely living in Zagreb. Average urbanisation rate in Croatia stands at 56%, with the maximum rate recorded within the territory of the City of Zagreb, where it reached 94.5% and Zagreb metropolitan area comprising the City of Zagreb and the Zagreb County, where it stands at 76.4%. Very significant rate of urbanisation was observed in the second half of the 20th century. 1953 census recorded 57% of population which was active in agriculture, while a census performed in 1991 noted only 9.1% of population active in that field. This points to augmentation of urban population and reduction of rural population.
https://en.wikipedia.org/wiki?curid=5576
Politics of Croatia The politics of Croatia are defined by a parliamentary, representative democratic republic framework, where the Prime Minister of Croatia is the head of government in a multi-party system. Executive power is exercised by the Government and the President of Croatia. Legislative power is vested in the Croatian Parliament (). The Judiciary is independent of the executive and the legislature. The parliament adopted the current Constitution of Croatia on 22 December 1990 and decided to declare independence from Yugoslavia on 25 May 1991. The Constitutional Decision on the Sovereignty and Independence of the Republic of Croatia came into effect on 8 October 1991. The constitution has since been amended several times. The first modern parties in the country developed in the middle of the 19th century, and their agenda and appeal changed, reflecting major social changes, such as the breakup of Austria-Hungary, the Kingdom of Serbs, Croats and Slovenes, dictatorship and social upheavals in the kingdom, World War II, the establishment of Communist rule and the breakup of the SFR Yugoslavia. The President of the Republic () is the head of state and the commander in chief of the Croatian armed forces and is directly elected to serve a five-year term. The government (), the main executive power of Croatia, is headed by the prime minister, who has four deputy prime ministers who serve also as government ministers. Twenty ministers are in charge of particular activities. The executive branch is responsible for proposing legislation and a budget, executing the laws, and guiding the foreign and internal policies. The parliament is a unicameral legislative body. The number of "Sabor" representatives (MPs) ranges from 100 to 160; they are elected by popular vote to serve four-year terms. The powers of the legislature include enactment and amendment of the constitution and laws; adoption of the government budget, declarations of war and peace, defining national boundaries, calling referenda and elections, appointments and relief of officers, supervising the Government of Croatia and other holders of public powers responsible to the Sabor, and granting of amnesties. The Croatian constitution and legislation provides for regular presidential and parliamentary elections, and the election of county prefects (county presidents) and assemblies, and city and municipal mayors and councils. Croatia has a three-tiered, independent judicial system governed by the Constitution of Croatia and national legislation enacted by the Sabor. The Supreme Court () is the highest court of appeal in Croatia, while municipal and county courts are courts of general jurisdiction. Specialised courts in Croatia are: commercial courts and the Superior Commercial Court, misdemeanour courts and the Superior Misdemeanour Court, administrative courts and the Superior Administrative Court. Croatian Constitutional Court () is a court that deals primarily with constitutional law. Its main authority is to rule on whether laws that are challenged are in fact unconstitutional, i.e., whether they conflict with constitutionally established rights and freedoms. The State Attorney's Office represents the state in legal proceedings. Croatia is a unitary democratic parliamentary republic. Following the collapse of the ruling Communist League, Croatia adopted a new constitution in 1990 – which replaced the 1974 constitution adopted by the Socialist Republic of Croatia – and organised its first multi-party elections. While the 1990 constitution remains in force, it has been amended four times since its adoption—in 1997, 2000, 2001 and 2010. Croatia declared independence from Yugoslavia on 8 October 1991, which led to the breakup of Yugoslavia. Croatia's status as a country was internationally recognised by the United Nations in 1992. Under its 1990 constitution, Croatia operated a semi-presidential system until 2000 when it switched to a parliamentary system. Government powers in Croatia are divided into legislative, executive and judiciary powers. The legal system of Croatia is civil law and, along with the institutional framework, is strongly influenced by the legal heritage of Austria-Hungary. By the time EU accession negotiations were completed on 30 June 2010, Croatian legislation was fully harmonised with the Community acquis. Croatia became a member state of the European Union on 1 July 2013. The President of the Republic () is the head of state; he or she is directly elected and serves a five-year term. The president is the commander in chief of the armed forces, has the procedural duty of appointing the prime minister with the consent of the Sabor (Parliament) through a majority vote (majority of all MPs), and has some influence on foreign policy. The most recent presidential election was held on 11 January 2015 and was won by Kolinda Grabar-Kitarović. She took the oath of office on 15 February 2015. The constitution limits holders of the presidential office to a maximum of two terms and prevents the president from being a member of any political party. Consequently, the president-elect withdraws from party membership before inauguration. The government (), the main executive power of Croatia, is headed by the prime minister who has four deputies, who also serve as government ministers. There are 16 other ministers who are appointed by the prime minister with the consent of the Sabor (majority of all MPs); these are in charge of particular sectors of activity. As of 19 October 2016, the Deputy Prime Ministers are Martina Dalić, Davor Ivo Stier, Ivan Kovačić, and Damir Krstičević. Government ministers are from the Croatian Democratic Union (HDZ), and the Bridge of Independent Lists (MOST) with five independent ministers. The executive branch is responsible for proposing legislation and a budget, executing the laws, and guiding the country's foreign and domestic policies. The government's official residence is at Banski dvori. As of 19 October 2016, the prime minister is Andrej Plenković. The Parliament of Croatia () is a unicameral legislative body. A second chamber, the "Chamber of Counties" (), was set up in 1993 pursuant to the 1990 Constitution. The Chamber of Counties was originally composed of three deputies from each of the twenty counties and the city of Zagreb. However, as it had no practical power over the Chamber of Representatives, it was abolished in 2001 and its powers were transferred to the county governments. The number of Sabor representatives can vary from 100 to 160; they are all elected by popular vote and serve four-year terms. 140 members are elected in multi-seat constituencies, up to six members are chosen by proportional representation to represent Croatians living abroad and five members represent ethnic and national communities or minorities. The two largest political parties in Croatia are the Croatian Democratic Union (HDZ) and the Social Democratic Party of Croatia (SDP). The last parliamentary election was held on 11 September 2016 in Croatia and on 10 and 11 September 2016 abroad. The Sabor meets in public sessions in two periods; the first from 15 January to 30 June, and the second from 15 September to 15 December. Extra sessions can be called by the President of the Republic, by the president of the parliament or by the government. The powers of the legislature include enactment and amendment of the constitution, enactment of laws, adoption of the state budget, declarations of war and peace, alteration of the country's boundaries, calling and conducting referenda and elections, appointments and relief of office, supervising the work of the Government of Croatia and other holders of public powers responsible to the Sabor, and granting amnesty. Decisions are made based on a majority vote if more than half of the Chamber is present, except in cases of constitutional issues. The Croatian constitution and legislation provides for regular elections for the office of the President of the Republic, parliamentary, county prefects, county assemblies, city and municipal mayors and city and municipal councils. The President of the Republic is elected to a five-year term by a direct vote of all citizens of Croatia. A majority vote is required to win. A runoff election round is held in cases where no candidate secures the majority in the first round of voting. The presidential elections are regulated by the constitution and dedicated legislation; the latter defines technical details, appeals and similar issues. 140 members of parliament are elected to a four-year term in ten multi-seat constituencies, which are defined on the basis of the existing county borders, with amendments to achieve a uniform number of eligible voters in each constituency to within 5%. Citizens of Croatia living abroad are counted in an eleventh constituency; however, its number of seats was not fixed for the last parliamentary election. It was instead calculated based on numbers of votes cast in the ten constituencies in Croatia and the votes cast in the eleventh constituency. In the 2007 parliamentary election the eleventh constituency elected five MPs. Constitutional changes first applied in the 2011 parliamentary election have abolished this scheme and permanently assigned three MPs to the eleventh constituency. Additionally, eight members of parliament are elected by voters belonging to twenty-two recognised minorities in Croatia: the Serb minority elects three MPs, Hungarians and Italians elect one MP each, Czech and Slovak minorities elect one MP jointly, while all other minorities elect two more MPs to the parliament. The Standard D'Hondt formula is applied to the vote, with a 5% election threshold. The last parliamentary election, held in 2016, elected 151 MPs. The county prefects and city and municipal mayors are elected to four-year terms by majority of votes cast within applicable local government units. A runoff election is held if no candidate achieves a majority in the first round of voting. Members of county, city, and municipal councils are elected to four-year terms through proportional representation; the entire local government unit forms a single constituency. The number of council members is defined by the councils themselves based on applicable legislation. Electoral committees are then tasked with determining whether the national minorities are represented in the council as required by the constitution. If the minorities are not represented, further members, who belong to the minorities and who have not been elected through the proportional representation system, are selected from electoral candidate lists and added to the council. Croatia has a three-tiered, independent judicial system governed by the constitution and national legislation enacted by the Sabor. The Supreme Court () is the highest court of appeal in Croatia; its hearings are open and judgments are made publicly, except in cases where the privacy of the accused is to be protected. Judges are appointed by the National Judicial Council and judicial office is permanent until seventy years of age. The president of the Supreme Court is elected for a four-year term by the Croatian Parliament at the proposal of the President of the Republic. As of 2017, the president of the Supreme Court is Đuro Sessa. The Supreme Court has civil and criminal departments. The lower two levels of the three-tiered judiciary consist of county courts and municipal courts. There are fifteen county courts and sixty-seven municipal courts in the country. There are other specialised courts in Croatia; commercial courts and the Superior Commercial Court, misdemeanour courts that try trivial offences such as traffic violations, the Superior Misdemeanour Court, the Administrative Court and the Croatian Constitutional Court (). The Constitutional Court rules on matters regarding compliance of legislation with the constitution, repeals unconstitutional legislation, reports any breaches of provisions of the constitution to the government and the parliament, declares the speaker of the parliament acting president upon petition from the government in the event the country's president becomes incapacitated, issues consent for commencement of criminal procedures against or arrest of the president, and hears appeals against decisions of the National Judicial Council. The court consists of thirteen judges elected by members of the parliament for an eight-year term. The president of the Constitutional Court is elected by the court judges for a four-year term. As of June 2012, the president of the Constitutional Court is Jasna Omejec. The National Judicial Council () consists of eleven members, specifically seven judges, two university professors of law and two parliament members, nominated and elected by the Parliament for four-year terms, and may serve no more than two terms. It appoints all judges and court presidents, except in case of the Supreme Court. As of January 2015, the president of the National Judicial Council is Ranko Marijan, who is also a Supreme Court judge. The State Attorney's Office represents the state in legal procedures. As of April 2018, Dražen Jelenić is the General State Attorney, and there are twenty-three deputies in the central office and lower-ranking State Attorneys at fifteen county and thirty-three municipal State Attorney's Offices. The General State Attorney is appointed by the parliament. A special State Attorney's Office dedicated to combatting corruption and organised crime, USKOK, was set up in late 2001. Croatia was first subdivided into counties () in the Middle Ages. The divisions changed over time to reflect losses of territory to Ottoman conquest and the subsequent recapture of the same territory, and changes to the political status of Dalmatia, Dubrovnik and Istria. The traditional division of the country into counties was abolished in the 1920s, when the Kingdom of Serbs, Croats and Slovenes and the subsequent Kingdom of Yugoslavia introduced oblasts and banovinas respectively. After 1945 under Communist rule, Croatia, as a constituent part of Yugoslavia, abolished these earlier divisions and introduced municipalities, subdividing Croatia into approximately one hundred municipalities. Counties, significantly altered in terms of territory relative to the pre-1920s subdivisions, were reintroduced in 1992 legislation. In 1918, the Transleithanian part of Croatia was divided into eight counties with their seats in Bjelovar, Gospić, Ogulin, Požega, Vukovar, Varaždin, Osijek and Zagreb; the 1992 legislation established fifteen counties in the same territory. Since the counties were re-established in 1992, Croatia is divided into twenty counties and the capital city of Zagreb, the latter having the authority and legal status of a county and a city at the same time. In some instances, the boundaries of the counties have been changed, with the latest revision taking place in 2006. The counties subdivide into 128 cities and 428 municipalities. The county prefects, city and municipal mayors are elected to four-year terms by a majority of votes cast within applicable local government units. If no candidate achieves a majority in the first round, a runoff election is held. Members of county, city and municipal councils are elected to four-year terms, through proportional representation with the entire local government unit as a single constituency. The number of members of the councils is defined by the councils themselves, based on applicable legislation. Electoral committees are then tasked with determining whether the national ethnic minorities are represented on the council as required by the constitution. Further members who belong to the minorities may be added to the council in no candidate of that minority has been elected through the proportional representation system. Election silence, as in all other types of elections in Croatia, when campaigning is forbidden, is enforced the day before the election and continues until 19:00 hours on the election day when the polling stations close and exit polls may be announced. Eight nationwide local elections have been held in Croatia since 1990, the most recent being the 2017 local elections to elect county prefects and councils, and city and municipal councils and mayors. In 2017, the HDZ-led coalitions won a majority or plurality in fifteen county councils and thirteen county prefect elections. SDP-led coalitions won a majority or plurality in five county councils, including the city of Zagreb council, and the remaining county council election was won by IDS-SDP coalition. The SDP won two county prefect elections, the city of Zagreb mayoral election, the HSS and the HNS won a single county prefect election each. Events of 1848 in Europe and the Austrian Empire brought dramatic changes to Croatian society and politics, provoking the Croatian national revival that strongly influenced and significantly shaped political and social events in Croatia. At the time, the Sabor and Ban Josip Jelačić advocated the severance of ties with the Kingdom of Hungary, emphasising links to other South Slavic lands within the empire. Several prominent Croatian political figures emerged, such as Ante Starčević, Eugen Kvaternik, Franjo Rački and Josip Juraj Strossmayer. A period of neo-absolutism was followed by the Austro-Hungarian Compromise of 1867 and the Croatian–Hungarian Settlement, which granted limited independence to Croatia. This was compounded by Croatian claims of uninterrupted statehood since the early Middle Ages as a basis for a modern state. Two political parties that evolved in the 1860s and contributed significantly to the sentiment were the Party of Rights, led by Starčević and Kvaternik, and the People's Party, led by Janko Drašković, Ivan Kukuljević Sakcinski, Josip Juraj Strossmayer and Ivan Mažuranić. They were opposed by the National Constitutional Party, which was in power for most of the period between the 1860s and the 1918, and advocated closer ties between Croatia and Hungary. Other significant parties formed in the era were the Serb People's Independent Party, which later formed the Croat-Serb Coalition with the Party of Rights and other Croat and Serb parties. The Coalition ruled Croatia between 1903 and 1918. The leaders of the Coalition were Frano Supilo and Svetozar Pribićević. The Croatian Peasant Party (HSS), established in 1904 and led by Stjepan Radić, advocated Croatian autonomy but achieved only moderate gains by 1918. In Dalmatia, the two major parties were the People's Party – a branch of the People's Party active in Croatia-Slavonia – and the Autonomist Party, advocating maintaining autonomy of Dalmatia, opposite to the People's Party demands for unification of Croatia-Slavonia and Dalmatia. The Autonomist Party, most notably led by Antonio Bajamonti, was also linked to Italian irredentism. By 1900, the Party of Rights had made considerable gains in Dalmatia. The Autonomists won the first three elections, but all elections since 1870 were won by the People's Party. In the period 1861–1918 there were seventeen elections in the Kingdom of Croatia-Slavonia and ten in the Kingdom of Dalmatia. After the establishment of the Kingdom of Serbs, Croats and Slovenes, the HSS established itself as the most popular Croatian political party and was very popular despite efforts to ban it. The 1921 constitution defined the kingdom as a unitary state and abolished the historical administrative divisions, which effectively ended Croatian autonomy; the constitution was opposed by HSS. The political situation deteriorated further as Stjepan Radić of the HSS was assassinated in the Yugoslav Parliament in 1928, leading to the dictatorship of King Alexander in January 1929. The HSS, now led by Vladko Maček, continued to advocate the federalisation of Yugoslavia, resulting in the Cvetković–Maček Agreement of August 1939 and the autonomous Banovina of Croatia. The Yugoslav government retained control of defence, internal security, foreign affairs, trade, and transport while other matters were left to the Croatian Sabor and a crown-appointed Ban. This arrangement was soon made obsolete with the beginning of World War II, when the Independent State of Croatia, which banned all political opposition, was established. Since then, the HSS continues to operate abroad. In the 1945 election, the Communists were unopposed because the other parties abstained. Once in power, the Communists introduced a single-party political system, in which the Communist Party of Yugoslavia was the ruling party and the Communist Party of Croatia was its branch. In 1971, the Croatian national movement, which sought greater civil rights and the decentralisation of the Yugoslav economy, culminated in the Croatian Spring, which was suppressed by the Yugoslav leadership. In January 1990, the Communist Party fragmented along national lines; the Croatian faction demanded a looser federation. In 1989, the government of the Socialist Republic of Croatia decided to tolerate political parties in response to growing demands to allow political activities outside the Communist party. The first political party founded in Croatia since the beginning of the Communist rule was the Croatian Social Liberal Party (HSLS), established on 20 May 1989, followed by the Croatian Democratic Union on 17 June 1989. In December 1989, Ivica Račan became the head of the reformed Communist party. At the same time, the party cancelled political trials, released political prisoners and endorsed a multi-party political system. The Civil Organisations Act was formally amended to allow political parties on 11 January 1990, legalising the parties that were already founded. By the time of the first round of the first multi-party elections, held on 22 April 1990, there were 33 registered parties. The most relevant parties and coalitions were the League of Communists of Croatia – Party of Democratic Changes (the renamed Communist party), the Croatian Democratic Union (HDZ), and the Coalition of People's Accord (KNS), which included the HSLS led by Dražen Budiša, and the HSS, which resumed operating in Croatia in December 1989. The runoff election was held on 6 May 1990. The HDZ, led by Franjo Tuđman, won ahead of the reformed Communists and the KNS. The KNS, led by Savka Dabčević-Kučar and Miko Tripalo – who had led the Croatian Spring – soon splintered into individual parties. The HDZ maintained a parliamentary majority until the 2000 parliamentary election, when it was defeated by the Social Democratic Party of Croatia (SDP), led by Račan. Franjo Gregurić, of the HDZ, was appointed prime minister to head a national unity government in July 1991 as the Croatian War of Independence escalated in intensity. His appointment lasted until August 1992. During his term, Croatia's declaration of independence from Yugoslavia took effect on 8 October 1991. The HDZ returned to power in the 2003 parliamentary election, while the SDP remained the largest opposition party. Franjo Tuđman won the presidential elections in 1992 and 1997. During his terms, the Constitution of Croatia, adopted in 1990, provided for a semi-presidential system. After Tuđman's death in 1999, the constitution was amended and much of the presidential powers were transferred to the parliament and the government. Stjepan Mesić won two consecutive terms in 2000 and 2005 on a Croatian People's Party (HNS) ticket. Ivo Josipović, an SDP candidate, won the presidential elections in December 2009 and January 2010. Kolinda Grabar-Kitarović defeated Josipović in the January 2015 election run-off, becoming the first female president of Croatia.
https://en.wikipedia.org/wiki?curid=5577
Economy of Croatia The economy of Croatia is a developing high-income service based economy with the tertiary sector accounting for 60% of total gross domestic product (GDP). After the collapse of socialism, Croatia went through a process of transition to a market-based economy in the 1990s, but its economy suffered badly during the Croatian War of Independence. After the war the economy began to improve, before the financial crisis of 2007–08 the Croatian economy grew at 4-5% annually, incomes doubled, and economic and social opportunities dramatically improved. The Croatian economy is one of the strongest economies in Southeast Europe. Croatia joined the World Trade Organization in 2000, NATO in 2009 and became a member of the European Union on 1 July 2013. Croatian economy was badly affected by the financial crisis which, together with slow progress of economic reforms, resulted in six years of recession and a cumulative decline in GDP of 12.5%. Croatia formally emerged from the recession in the 4th quarter of 2014, and had a continuous GDP growth since. Croatian economy reached pre crisis levels in 2019 but due to Coronavirus disease the economy is set to contract more than -9,0% during first and second quarter of 2020, however growth is set to pick up in third quarter of 2020. The industrial sector with exports of over €10 billion annually is dominated by shipbuilding which accounts for over 10% of exported goods. Food processing and chemical industry also account for significant portions of industrial output and exports. Industrial sector represents 27% of Croatia's total economic output while agriculture represents 6%. Industrial sector is responsible for 25% of Croatia's GDP, with agriculture, forestry and fishing accounting for the remaining 5% of Croatian GDP. Tourism is traditionally a notable source of income, particularly during the summer months, but also more recently during the winter months as well, due to an increase in popularity of snow sports such as skiing. With over 14 million tourists annually, tourism generates revenue in excess of €8 billion. Croatia is ranked among the top 20 most popular tourist destinations in the world, and was voted world's top tourism destination in 2005 by Lonely Planet. Trade plays a major role in Croatian economic output. In 2007 Croatia's exports were valued at US$12.84 billion (24.7 billion including service exports). According to Healy Consultants, trade in Croatia is bolstered by its low trade-weighted average tariff of just 1.2%. Croatia's currency is the kuna, which was implemented in 1994 and has remained stable since. During the 19th century the Kingdom of Croatia had a high ratio of population working in agriculture. Many industrial branches developed in that time, like forestry and wood industry (stave fabrication, the production of potash, lumber mills, shipbuilding). The most profitable one was stave fabrication, the boom of which started in the 1820s with the clearing of the oak forests around Karlovac and Sisak and again in the 1850s with the marshy oak masses along the Sava and Drava rivers. Shipbuilding in Croatia played a huge role in the 1850s Austrian Empire, especially the longe-range sailing boats. Sisak and Vukovar were the centres of river-shipbuilding. Slavonia was also mostly an agricultural land and it was known for its silk production. Agriculture and the breeding of cattle were the most profitable occupations of the inhabitants. It produced corn of all kinds, hemp, flax, tobacco, and great quantities of liquorice. The first steps towards industrialization began in the 1830s and in the following decades the construction of big industrial enterprises took place. During the 2nd half of the 19th and early 20th century there was an upsurge of industry in Croatia, strengthened by the construction of railways and the electric-power production. However, the industrial production was still lower than agricultural production. Regional differences were high. Industrialization was faster in inner Croatia than in other regions, while Dalmatia remained one of the poorest provinces of Austria-Hungary. The slow rate of modernization and rural overpopulation caused extensive emigration, particularly from Dalmatia. According to estimates, roughly 400,000 Croats emigrated from Austria-Hungary between 1880 and 1914. In 1910 8,5% of the population of Croatia-Slavonia lived in urban settlements. In 1918 Croatia became part of the Kingdom of Yugoslavia, which was in the interwar period one of the least developed countries in Europe. Most of its industry was based in Slovenia and Croatia, but further industrial development was modest and centered on textile mills, sawmills, brick yards and food-processing plants. Economy was still traditionally based on agriculture and raising of livestock, with peasants accounting for more than half of Croatia's population. In 1941 the Independent State of Croatia (NDH), a World War II puppet state of Germany and Italy, was established in parts of Axis-occupied Yugoslavia. The economic system of NDH was based on the concept of "Croatian socialism". The main characteristic of the new system was the concept of a planned economy with high levels of state involvement in economic life. The fulfillment of basic economic interests was primarily ensured with measures of repression. All large companies were placed under state control and the property of the regime's national enemies was nationalized. Its currency was the NDH kuna. The Croatian State Bank was the central bank, responsible for issuing currency. As the war progressed the government kept printing more money and its amount in circulation was rapidly increasing, resulting in high inflation rates. After the World War II, the new Communist Party of Yugoslavia resorted to a command economy on the Soviet model of rapid industrial development. By 1948 almost all domestic and foreign-owned capital had been nationalized. The industrialization plan relied on high taxation, fixed prices, war reparations, Soviet credits, and export of food and raw materials. Forced collectivization of agriculture was initiated in 1949. At that time 94% of agricultural land was privately owned, and by 1950 96% was under the control of the social sector. A rapid improvement of food production and the standard of living was expected, but due to bad results the program was abandoned three years later. Throughout the 1950s Croatia experienced rapid urbanization. Decentralization came in 1965 and spurred growth of several sectors including the prosperous tourist industry. SR Croatia was, after SR Slovenia, the second most developed republic in Yugoslavia with a 25% higher GDP per capita than the Yugoslav average. Croatia and Slovenia accounted for nearly half of the total Yugoslav GDP, and this was reflected in the overall standard of living. In the mid-1960s, Yugoslavia lifted emigration restrictions and the number of emigrants increased rapidly. In 1971 224,722 workers from Croatia were employed abroad, mostly in West Germany. Foreign remittances contributed $2 billion annually to the economy by 1990. Profits gained through Croatia's industry were used to develop poor regions in other parts of former Yugoslavia, leading to Croatia contributing much more to the federal Yugoslav economy than it gained in return. This, coupled with austerity programs and hyperinflation in the 1980s, led to discontent in both Croatia and Slovenia which eventually fuelled political movements calling for independence. In the late 1980s and early 1990s, with the collapse of socialism and the beginning of economic transition, Croatia faced considerable economic problems stemming from: At the time Croatia gained independence, its economy (and the whole Yugoslavian economy) was in the middle of recession. Privatization under the new government had barely begun when war broke out in 1991. As a result of the Croatian War of Independence, infrastructure sustained massive damage in the period 1991–92, especially the revenue-rich tourism industry. Privatization in Croatia and transformation from a planned economy to a market economy was thus slow and unsteady, largely as a result of public mistrust when many state-owned companies were sold to politically well-connected at below-market prices. With the end of the war, Croatia's economy recovered moderately, but corruption, cronyism, and a general lack of transparency stymied economic reforms and foreign investment. The privatization of large government-owned companies was practically halted during the war and in the years immediately following the conclusion of peace. As of 2000, roughly 70% of Croatia's major companies were still state-owned, including water, electricity, oil, transportation, telecommunications, and tourism. Early 1990s were characterized with high inflation rates. In 1991 the Croatian dinar was introduced as a transitional currency, but inflation continued to accelerate. The anti-inflationary stabilization steps in 1993 decreased retail price inflation from a monthly rate of 38,7% to 1,4%, and by the end of the year Croatia experienced deflation. In 1994 Croatia introduced the kuna as its currency. As a result of the macro-stabilization programs, the negative growth of GDP during the early 1990s stopped and turned into a positive trend. Post-war reconstruction activity provided another impetus to growth. Consumer spending and private sector investments, both of which were postponed during the war, contributed to the growth in 1995–1997. Croatia began its independence with a relatively low external debt because the debt of Yugoslavia was not shared among its former republics at the beginning. In March 1995 Croatia agreed with the Paris Club of creditor governments and took 28,5% of Yugoslavia's previously non-allocated debt over 14 years. In July 1996 an agreement was reached with the London Club of commercial creditors, when Croatia took 29,5% of Yugoslavia's debt to commercial banks. In 1997 around 60 percent of Croatia's external debt was inherited from former Yugoslavia. At the beginning of 1998 value-added tax was introduced. The central government budget was in surplus in that year, most of which was used to repay foreign debt. Government debt to GDP had fallen from 27,30% to 26,20% at the end of 1998. However, the consumer boom was disrupted in mid 1998, as a result of the bank crisis when 14 banks went bankrupt. Unemployment increased and GDP growth slowed down to 1,9%. The recession that began at the end of 1998 continued through most of 1999, and after a period of expansion GDP in 1999 had a negative growth of -0,9%. In 1999 the government tightened its fiscal policy and revised the budget with a 7% cut in spending. In 1999 the private sector share in GDP reached 60%, which was significantly lower than in other former socialist countries. After several years of successful macroeconomic stabilization policies, low inflation and a stable currency, economists warned that the lack of fiscal changes and the expanding role of the state in economy caused the decline in the late 1990s and were preventing a sustainable economic growth. The new government led by the president of SDP, Ivica Račan, carried out a number of structural reforms after it won the parliamentary elections on 3 January 2000. The country emerged from the recession in the 4th quarter of 1999 and growth picked up in 2000. Due to overall increase in stability, the economic rating of the country improved and interest rates dropped. Economic growth in the 2000s was stimulated by a credit boom led by newly privatized banks, capital investment, especially in road construction, a rebound in tourism and credit-driven consumer spending. Inflation remained tame and the currency, the kuna, stable. In 2000 Croatia generated 5,899 billion kunas in total income from the shipbuilding sector, which employed 13,592 people. Total exports in 2001 amounted to $4,659,286,000, of which 54.7% went to the countries of the EU. Croatia's total imports were $9,043,699,000, 56% of which originated from the EU. Unemployment reached its peak in late 2002, but has since been steadily declining. In 2003, the nation's economy would officially recover to the amount of GDP it had in 1990. In late 2003 the new government led by HDZ took over the office. Unemployment continued falling, powered by growing industrial production and rising GDP, rather than only seasonal changes from tourism. Unemployment reached an all-time low in 2008 when the annual average rate was 8.6%, GDP per capita peaked at $16,158, while public debt as percentage of GDP decreased to 29%. Most economic indicators remained positive in this period except for the external debt as Croatian firms focused more on empowering the economy by taking loans from foreign resources. Between 2003 and 2007, Croatia's private-sector share of GDP increased from 60% to 70%. The Croatian National Bank had to take steps to curb further growth of indebtedness of local banks with foreign banks. The dollar debt figure is quite adversely affected by the EUR/USD ratio—over a third of the increase in debt since 2002 is due to currency value changes. Economic growth has been hurt by the global financial crisis. Immediately after the crisis it seemed that Croatia did not suffer serious consequences like some other countries. However, in 2009, the crisis gained momentum and the decline in GDP growth, at a slower pace, continued during 2010. In 2011 the GDP stagnated as the growth rate was zero. Since the global crisis hit the country, the unemployment rate has been steadily increasing, resulting in the loss of more than 100,000 jobs. While unemployment was 9.6% in late 2007, in January 2014 it peaked at 22.4%. In 2010 Gini coefficient was 0,32. In September 2012, Fitch ratings agency unexpectedly improved Croatia's economic outlook from negative to stable, reaffirming Croatia's current BBB rating. The slow pace of privatization of state-owned businesses and an overreliance on tourism have also been a drag on the economy. Croatia joined the European Union on 1 July 2013 as the 28th member state. The Croatian economy is heavily interdependent on other principal economies of Europe, and any negative trends in these larger EU economies also have a negative impact on Croatia. Italy, Germany and Slovenia are Croatia's most important trade partners. In spite of the rather slow post-recession recovery, in terms of income per capita it is still ahead of some European Union member states such as Bulgaria, and Romania. In terms of average monthly wage, Croatia is ahead of 9 EU members (Czech Republic, Estonia, Slovakia, Latvia, Poland, Hungary, Lithuania, Romania, and Bulgaria). Annual average unemployment rate in 2014 was 17.3% and Croatia has the third highest unemployment rate in the European Union, after Greece (26.5%), and Spain (24.%). Of particular concern is the heavily backlogged judiciary system, combined with inefficient public administration, especially regarding the issues of land ownership and corruption in the public sector. Unemployment is regionally uneven: it is very high in eastern and southern parts of the country, nearing 20% in some areas, while relatively low in the north-west and in larger cities, where it is between 3 and 7%. In 2015 external debt rose by 2.7 billion euros since the end of 2014 and is now around €49.3 billion. During 2015 the Croatian economy started with slow but upward economic growth, which continued during 2016 and conclusive at the end of the year seasonally adjusted was recorded at 3.5%. The better than expected figures during 2016 enabled the Croatian Government and with more tax receipts enabled the repayment of debt as well as narrow the current account deficit during Q3 and Q4 of 2016 This growth in economic output, coupled with the reduction of government debt has made a positive impact on the financial markets with many ratings agencies revising their outlook from negative to stable, which was the first upgrade of Croatia's credit rating since 2007. Due to consecutive months of economic growth and the demand for labour, plus the outflows of residents to other European countries, Croatia had recorded the biggest fall in the number of unemployed during the month of November 2016 from 16.1% to 12.7%. COVID-19 Pandemic has caused more than 400.000 workers to file for minimum net wage of 3,250 HRK given by the Government of Croatia as aid to firms and companies for them not to fire their workers. Public debt is set to rise up to 90% of GDP, while unemployment is set to rise to around 9,3%, due to incoming tourist season (May–October) being unsafe. European Commission has decreased country's GDP growth for 2020 to 2,3%. Tourism is a notable source of income during the summer and a major industry in Croatia. It dominates the Croatian service sector and accounts for up to 20% of Croatian GDP. Annual tourist industry income for 2011 was estimated at €6.61 billion. Its positive effects are felt throughout the economy of Croatia in terms of increased business volume observed in retail business, processing industry orders and summer seasonal employment. The industry is considered an export business, because it significantly reduces the country's external trade imbalance. Since the conclusion of the Croatian War of Independence, the tourist industry has grown rapidly, recording a fourfold rise in tourist numbers, with more than 10 million tourists each year. The most numerous are tourists from Germany, Slovenia, Austria and the Czech Republic as well as Croatia itself. Length of a tourist stay in Croatia averages 4.9 days. The bulk of the tourist industry is concentrated along the Adriatic Sea coast. Opatija was the first holiday resort since the middle of the 19th century. By the 1890s, it became one of the most significant European health resorts. Later a large number of resorts sprang up along the coast and numerous islands, offering services ranging from mass tourism to catering and various niche markets, the most significant being nautical tourism, as there are numerous marinas with more than 16 thousand berths, cultural tourism relying on appeal of medieval coastal cities and numerous cultural events taking place during the summer. Inland areas offer mountain resorts, agrotourism and spas. Zagreb is also a significant tourist destination, rivalling major coastal cities and resorts. Croatia has unpolluted marine areas reflected through numerous nature reserves and 116 Blue Flag beaches. Croatia is ranked as the 18th most popular tourist destination in the world. About 15% of these visitors (over one million per year) are involved with naturism, an industry for which Croatia is world-famous. It was also the first European country to develop commercial naturist resorts. Croatian agricultural sector subsists from exports of blue water fish, which in recent years experienced a tremendous surge in demand, mainly from Japan and South Korea. Croatia is a notable producer of organic foods and much of it is exported to the European Union. Croatian wines, olive oil and lavender are particularly sought after. The highlight of Croatia's recent infrastructure developments is its rapidly developed motorway network, largely built in the late 1990s and especially in the 2000s. By September 2011, Croatia had completed more than of motorways, connecting Zagreb to most other regions and following various European routes and four Pan-European corridors. The busiest motorways are the A1, connecting Zagreb to Split and the A3, passing east–west through northwest Croatia and Slavonia. A widespread network of state roads in Croatia acts as motorway feeder roads while connecting all major settlements in the country. The high quality and safety levels of the Croatian motorway network were tested and confirmed by several EuroTAP and EuroTest programs. Croatia has an extensive rail network spanning , including of electrified railways and of double track railways. The most significant railways in Croatia are found within the Pan-European transport corridors Vb and X connecting Rijeka to Budapest and Ljubljana to Belgrade, both via Zagreb. All rail services are operated by Croatian Railways. There are international airports in Zagreb, Zadar, Split, Dubrovnik, Rijeka, Osijek and Pula. As of January 2011, Croatia complies with International Civil Aviation Organization aviation safety standards and the Federal Aviation Administration upgraded it to Category 1 rating. The busiest cargo seaport in Croatia is the Port of Rijeka and the busiest passenger ports are Split and Zadar. In addition to those, a large number of minor ports serve an extensive system of ferries connecting numerous islands and coastal cities in addition to ferry lines to several cities in Italy. The largest river port is Vukovar, located on the Danube, representing the nation's outlet to the Pan-European transport corridor VII. There are of crude oil pipelines in Croatia, connecting the Port of Rijeka oil terminal with refineries in Rijeka and Sisak, as well as several transhipment terminals. The system has a capacity of 20 million tonnes per year. The natural gas transportation system comprises of trunk and regional natural gas pipelines, and more than 300 associated structures, connecting production rigs, the Okoli natural gas storage facility, 27 end-users and 37 distribution systems. Croatian production of energy sources covers 85% of nationwide natural gas demand and 19% of oil demand. In 2008, 47.6% of Croatia's primary energy production structure comprised use of natural gas (47.7%), crude oil (18.0%), fuel wood (8.4%), hydro power (25.4%) and other renewable energy sources (0.5%). In 2009, net total electrical power production in Croatia reached 12,725 GWh and Croatia imported 28.5% of its electric power energy needs. The bulk of Croatian imports are supplied by the Krško Nuclear Power Plant in Slovenia, 50% owned by Hrvatska elektroprivreda, providing 15% of Croatia's electricity. Electricity: Electricity – production by source: Crude oil: Natural gas: Central bank: Major commercial banks: Overall Budget: "Revenues:" "Expenditures:" Expenditure for 2016 The following table shows the main economic indicators in 1993–2017. "From the CIA World Factbook 2016." GDP – purchasing power parity: $91.28 billion (2015 est.) GDP – real growth rate: 1.6% (2015 est.) GDP – per capita: purchasing power parity - $21,600 (2015 est.) GDP – composition by sector: agriculture: 4.3% industry: 26.2% services: 69.5% (2015 est.) Labor force: 1.677 million (2015 est.) Labor force – by occupation: agriculture 1.9%, industry 27.6%, services 70.4% (2014) Unemployment rate: 17.1% (2015 est.) Population below poverty line: 19.5% (2014) Household income or consumption by percentage share: "lowest 10%:" 3.3% "highest 10%:" 27.5% Distribution of family income – Gini index: 29 (2001) 32 (2010) Inflation rate (consumer prices): -0.5% (2015) Investment (gross fixed): 19,1% of GDP (2014 est.) Budget: "revenues:" $21.3 billion (2015 est.) "expenditures:" $22.86 billion, (2015 est.) Public debt: 86.7% of GDP (2015 est.) Agriculture – products: wheat, corn, sugar beets, sunflower seed, barley, alfalfa, clover, olives, citrus, grapes, soybeans, potatoes; livestock, dairy products Industries: chemicals and plastics, machine tools, fabricated metal, electronics, pig iron and rolled steel products, aluminium, paper, wood products, construction materials, textiles, shipbuilding, petroleum and petroleum refining, food and beverages; tourism Industrial production growth rate: 1.9% (2015 est.) Current account balance: $2.551 billion (2015 est.) Exports: $11.91 billion (2015 est.) Exports – commodities: transport equipment, textiles, chemicals, foodstuffs, fuels Exports – partners: Italy 13.4%, Slovenia 12.5%, Germany 11.4%, Bosnia and Herzegovina 9.9%, Austria 6.6%, Imports: $19.28 billion (2015 est.) Imports – commodities: machinery, transport and electrical equipment, chemicals, fuels and lubricants, foodstuffs Imports – partners: Germany 15.5%, Italy 13.1%, Slovenia 10.7%, Austria 9.2%, Hungary 7.8%, Reserves of foreign exchange and gold: $14.97 billion (31 December 2015 est.) Debt – external: $50.88 billion (31 December 2015 est.) Currency: kuna (HRK) Exchange rates: kuna per US$1 – 6.8583 (2015), 5.7482 (2014), 5.7089 (2013), 5.3735 (2007), 5.8625 (2006), 5.9473 (2005), 6.0358 (2004), 6.7035 (2003), 7.8687 (2002), 8.34 (2001), 8.2766 (2000), 7.112 (1999), 6.362 (1998), 6.157 (1997), 5.434 (1996), 5.230 (1995)
https://en.wikipedia.org/wiki?curid=5578
Transport in Croatia Transport in Croatia relies on several main modes, including transport by car, train, boat and plane. Road transport incorporates a comprehensive network of state, county and local routes augmented by a network of highways for long-distance travelling. Water transport can be divided into sea, based on the ports of Rijeka, Ploče, Split and Zadar, and river transport, based on Sava, Danube and, to a lesser extent, Drava. Croatia has 68 airports, nine of which are international. The country also has several airlines, of which the most notable are Croatia Airlines and Trade Air. Rail transport is fairly developed, with dual track and electrification not very common, although high-speed tilting trains are used on some routes. However, bus still tends to be more common than rail as a mode of inter-city transport. There are international airports in Zagreb, Split, Dubrovnik, Zadar, Pula, Rijeka (on the island of Krk), Osijek, Bol and Mali Lošinj. Currently, the following low cost airlines are flying to Croatia: Jet2, EasyJet, Eurowings, TUIfly, Ryanair, Thomson, flydubai and Wizz Air. Major established companies that fly to Croatia include the domestic Croatia Airlines (member of the Star Alliance), Air France, Aeroflot, Lufthansa, Emirates, Finnair, Austrian Airlines, KLM, Qatar Airways, Swiss International Air Lines, British Airways and Turkish Airlines. Seasonal intercontinental flights also operate from and to Croatia. In January 2011 the U.S. Federal Aviation Administration (FAA) assessed the Government of Croatia's Civil Aviation Authority announcing that Croatia complied with International Civil Aviation Organization (ICAO) aviation safety standards for oversight of Croatia's air carrier opinions. IASA Category 1 rating and that Croatian air carriers are authorized Croatia to the US. Statistics: Overall: 68 airports (2004 estimate) Airports with paved runways (2004 est.): Airports with unpaved runways (2004 est.): Heliports: 3 (2008) Airports: The Croatian railway network is classified into three groups: railways of international, regional and local significance. The most important railway lines follow Pan-European corridors V (branch B) and X, they connect at Zagreb with each other. There are connections to the railway networks of Slovenia, Hungary, Bosnia and Herzegovina and Serbia. The railways need to be modernized because the lion's share of the railway infrastructure in Croatia dates back from the pre-WW2 period and more than half of the core routes were, in fact, built during the Habsburg Monarchy i.e. before the First World War. As a result, most of the routes are not electrified, are only single track traffic around 'bends' and lag significantly behind the West-European standards. Some limited improvements, however, have been made in recent times. A result of this has been an increase in the maximum speed on the railway line on the Pan-European Corridor X, from the Slovenian border via Zagreb, Novska, Vinkovci to the Serbian border (lines M101 through M105). In places the line's speed limit has been raised from to , and plans have been made to raise this further to on certain sections. The railway line of the Pan-European Corridor Vb, from the Hungarian border via Koprivnica, Zagreb, Karlovac to the Port of Rijeka (lines M201, M102 and M202) has a slow section through Gorski Kotar, which is planned to undergo a major overhaul. The railway line of the Pan-European Corridor Vc, from the Hungarian border via Beli Manastir, Osijek, Đakovo or Vinkovci, Slavonski Šamac, to the border of Bosnia and Herzegovina (lines M301, M302 and M303), is also being modernised, the aim is to allow speeds of up to 160 km/h along the entire length. The official rail speed record in Croatia is . Maximum speed reached in regular service is on parts of the Novska–Tovarnik line. Recently Croatian Railways introduced a series of modern tilting trains produced by the German branch of Bombardier Transportation. They usually deployed on the mountainous route between the two largest Croatian cities, Zagreb and Split. Although they can sometimes be found on other routes in the country. Contrary to regular overnight fast trains between Zagreb and Split with scheduled travelling time of circa 8 hours in total, tilting trains on the Zagreb–Split route (lines M202 and M604) offer passengers a faster journey with a journey time of about 6 hours. Croatian Railways aims to revitalize rail traffic through further improvements. Thus, establishing rail as serious competitor to the car, particularly during the busy the summer months. Croatian Railways' plan to build their first high-speed railway service is progressing. Construction of an entirely new line from Karlovac to Rijeka, and reconstruction of the line from the Hungarian border to Karlovac should have originally begun in the fall of 2007. The line is going to be entirely electrified and in order to speed up the journey even more shall be long from end to end, shorter than the existing line. The journey from Zagreb to Rijeka will be cut to one hour compared to 4 hours with the existing track. The line is intended to carry the increasing amount of goods that enter Europe, at the Croatian Port of Rijeka and are then transported to destinations across central and eastern Europe. The plans have originally foreseen 2008-2010 as the date for completion. Due to the world economic crisis, building start date is prolonged to a yet undefined date. The project plans have been drawn, however. Pan-European Corridor X is going to be upgraded to higher speed sometime in the future. Possibly by building a new high capacity connection line from Sisak via Kutina to Lipovljani. The line should allow for speeds of up to 250 km/h. Statistics: Railway length (): From the time of Napoleon and building the Louisiana road, the road transport in Croatia has significantly improved, topping most European countries. Croatian highways are widely regarded as being one of the most modern and safe in Europe. This is because the largest part of the Croatian motorway and expressway system ("autoceste" and "brze ceste", resp.) has been recently constructed (mainly in the 2000s), and further construction is continuing. The motorways in Croatia connect most major Croatian cities and all major seaports. The two longest routes, the A1 and the A3, span the better part of the country and the motorway network connects most major border crossings. Tourism is of major importance for the Croatian economy, and as most tourists come to vacation in Croatia in their own cars, the highways serve to alleviate summer jams. They have also been used as a means of stimulating urgently needed economic growth, and for the sustainable development of this country. Croatia now has a considerable highway density for a country of its size, helping it cope with the consequences of being a transition economy and having suffered in the Croatian War of Independence. , Croatia has a total of of roads. Travel Organiser (2007) The traffic signs adhere to the Vienna Convention on Road Signs and Signals. The general speed limits are: In 2004, a controversial new traffic law was introduced, which provides for drastic safety measures for drunken or dangerous driving: it reduced the blood alcohol limit to 0‰ of alcohol in blood. In 2008 this law was reversed to 0.5‰ of alcohol in blood. Some of the more technical safety measures include that all new Croatian tunnels have modern safety equipment and there are several control centers, which monitor highway traffic. Motorways (, plural ) in Croatia applies to dual carriageway roads with at least two traffic lanes in each driving direction and an emergency lane. Direction road signs at Croatian motorways have green background with white lettering similar to the Swiss Autobahn. The designations of motorways are "A" and the motorway number. , the Croatian motorway network is long. The list of completed motorways is as follows (see individual articles for further construction plans and status): Toll is charged on most Croatian motorways, and exceptions are the A11 motorway, Zagreb bypass and Rijeka bypass, as well as sections adjacent to border crossings (except eastbound A3). Payment in kuna, all major credit cards and euros are accepted at all toll gates. Most motorways are covered by the closed toll collection system, where a driver receives a ticket at the entrance gates and pays at the exit gates according to the number of sections travelled. Open toll collection is used on some bridges and tunnels and short stretches of tolled highway, where drivers immediately pay the toll upon arriving. Various forms of prepaid electronic toll collection systems are in place which allow quicker collection of toll, usually at a discounted rate, as well as use of dedicated toll plaza lanes (for "ENC system" of the electronic toll collection). All heavily traveled routes towards Slovenia, Hungary and Serbia are motorway connections, and almost all parts of Croatia are now easy to reach using motorways. The 479 km long motorway A1 Zagreb - Split - Ploče connects the continental part of Croatia with Mediterranean Dalmatia. It is complete with two lanes in both directions in its full length. The highway winds itself as a panoramic road through the Croatian hills. The highway currently ends near the Ploče sea port, but is planned to continue further on to Dubrovnik. The most important edifice on the A1 is the Sveti Rok Tunnel (5.687 m), which goes through the famous Velebit mountain range barrier. The most modern, and at the same time the longest Croatian tunnel, is also along the A1: the Mala Kapela Tunnel (length 5,780 m), which passes through the Mala Kapela mountain range. The Sveti Ilija Tunnel through Biokovo (4,249 m) connects the Makarska riviera with the A1. Since June 2005 the Istrian Y extends from the Slovenian border in Istria and the A7 connects Slovenia to Rijeka. The A2 connection between Zagreb via Hrvatsko Zagorje to Maribor was completed in May 2007. Motorway A3 extends from Slovenia via the Zagreb bypass through the length of Slavonia to the Serbian border. In Slavonia there is now another motorway, A5, along the European route E73 (European north-south corridor Vc) between the Hungarian border at Beli Manastir, Osijek and the Bosnian border. This route will eventually continue as a modern motorway towards Sarajevo and further on to the Croatian sea port of Ploče. The A4 connection from Zagreb to Varaždin and Čakovec exists since 2003, which was extended to the Hungarian border in 2008 to complete the upgrade of the European route E71. The A6 route between Rijeka and Zagreb was constructed as an expressway in spring 2004, and upgraded to a full motorway in 2008, thus upgrading the Croatian section of the European route E65. Together these two upgrades completed the European north-south corridor Vb. A motorway connection of Zagreb to the important industrial city of Sisak is planned to be constructed as A11. Significant tunnel and bridge construction projects in the south of Dalmatia are planned, such as a long Pelješac bridge, connecting the Pelješac peninsula to the Croatian mainland. By 2008, numerous service areas and petrol stations had been constructed along all Croatian motorways. All Croatian motorways are equipped with enclosed service areas with gas stations and parking. Many areas have restaurants and children's playgrounds. The term "brza cesta" or "expressway" refers to limited-access roads specifically designated as such by legislation and marked with appropriate limited-access road traffic signs. The expressways may comprise two or more traffic lanes, while they normally do not have emergency lanes. "Polu-autocesta" or "semi-highway" refers to a two-lane, undivided road running on one roadway of a motorway while the other is in construction. By legal definition, all semi-highways are expressways. The expressway routes in Croatia usually correspond to a state road (see below) and are marked a "D" followed by a number. The "E" numbers are designations of European routes. Major roads that aren't part of the motorway system are "državne ceste" (state routes). They are marked with the letter D and the road's number. The most traveled state routes in Croatia are: Since the construction of A1 motorway beyond Gorski kotar started, D1 and D8 are much less used. These routes are monitored by Croatian roadside assistance because they connect important locations. Like all state routes outside major cities, they are only two-lane arterials and do not support heavy traffic. All state routes are routinely maintained by Croatian road authorities. The road sign for a state route has a blue background and the route's designation in white. State routes have one, two or three-digit numbers. Secondary routes are known as county roads. They are marked with signs with yellow background and road number. These roads' designations are rarely used, but usually marked on regional maps if these roads are shown. Formally, their designation is the letter Ž and the number. County roads have four-digit numbers. The least known are the so-called local roads. Their designations are never marked on maps or by roadside signs and as such are virtually unknown to public. Their designations consist of the letter L and a five-digit number. In contrast to the fairly underdeveloped rail traffic, buses represent the most-accepted, cheapest and widely used means of public transport. National bus traffic is very well developed and it is very easy to reach even the remotest parts of Croatia by bus. Almost all buses on national routes are air-conditioned and offer pleasant traveling comfort. The Croatian parliament has passed a law that no bus should be older than 12 years - however, this decision is currently frozen because of the high cost for bus operators. In practice, bus fares are collected on the bus while traveling, which is sometimes even cheaper than when paying at the ticket office (there is an additional fee for stored luggage) and sometimes a "ticket office fee". From Croatia, there are many international bus routes to the neighboring countries (Slovenia, Bosnia and Herzegovina, Serbia etc.), as well as to Austria, Germany, Slovenia, Switzerland and to other European countries. International buses correspond to European standards. Zagreb has the largest and most modern bus terminal in Croatia. It is located near the downtown in Trnje district on the Marin Držić Avenue. It sports specially designed waiting areas above the bus stopping and parking area. The Zagreb bus terminal is close to the central train station, Glavni kolodvor and it is easy to reach by tram lines and by car. Croatia has several large seaports. The largest seaport with the deepest channel to a port in the Adriatic is Port of Rijeka on the northern Croatian coast, followed by Ploče in southern Dalmatia. The port of Ploče is of strategic importance for the industries of Bosnia and Herzegovina. The largest Croatian passenger port is Split in Dalmatia, also called "gateway to the islands", followed by Zadar. There are 66 inhabited islands along the Croatian coast which means there is a large number of local ferry connections. Ports and harbors: Merchant marine (): An extensive system of ferries, operated by Jadrolinija, serves Croatia's many islands and links coastal cities. Ferry services to Italian cities of Venice, Ancona, Pescara and Bari from around a dozen of Croatian sea ports, most notably Rovinj, Rijeka, Zadar, Split, Korčula and Dubrovnik, is available on a daily or weekly basis. From April to September the schedule is denser and can include several round trips within a day. Croatia is also on the important Danube waterway, which connects Eastern and Central Europe. The major Danube port is Vukovar. Perennially navigable rivers: Total waterway length (): 785 km There are of crude oil pipelines in Croatia, connecting the Port of Rijeka oil terminal with refineries in Rijeka and Sisak, as well as several transhipment terminals. The system has a capacity of 20 million tonnes per year. The natural gas transportation system comprises of trunk and regional natural gas pipelines, and more than 300 associated structures, connecting production rigs, the Okoli natural gas storage facility, 27 end-users and 37 distribution systems.
https://en.wikipedia.org/wiki?curid=5580
Republic of Croatia Armed Forces The Republic of Croatia Armed Forces ( – OSRH) is the military service of Croatia. The President is the Armed Forces Commander-in-Chief, and exercises administrative powers in times of war by giving orders to the chief of staff, while administration and defence policy execution in peacetime is carried out by the Government through the Ministry of Defence. This unified institution consists of land, sea, and air branches referred to as: The Croatian Armed Forces are charged with protecting the Republic as well as supporting international peacekeeping efforts, when mandated by the NATO, United Nations and/or European Union. The Army has 650 AFVs, around 150 pieces of artillery, 100 MLRSs, around 70 tanks, and 20 SPGs. The Air Force has 12 MiG-21 jet fighters, 10 combat-transport Mi-171 and 16 OH-58 attack helicopters. The Navy has 29 ships, out of which five 60-80 metre fast attack craft are used in offensive capabilities. The total number of active military personnel in the Croatian Armed Forces stands at 14,506 and 6,000 reserves working in various service branches of the armed forces. In May 2016, Armed Forces had 16,019 members, of which 14,506 were active military personnel and 1,513 civil servants. Of the 14,506 active military personnel, 3,183 were officers, 5,389 non-commissioned officers, 5,393 soldiers, 520 military specialists, 337 civil servants and 1,176 other employees. Total available male manpower aged 16–49 numbers 1,035,712, of which 771,323 are technically fit for military service. Male citizens are now no longer subject to compulsory military service since January 1, 2008. However, the last generation of 2007 servicemen was also absolved of compulsory service by an act from then Minister of Defence Berislav Rončević. The Croatian military budget for the past 6–7 years was kept below 2% of GDP, a vast difference from the 1990s when defence expenditure represented a major stake in Croatian budgetary expenditure due to the Croatian War of Independence. For example, 1995 Croatian defence budget stood at 12.4 billion Croatian Kuna or just over 10% of GDP, which was also the highest defence expenditure rate ever. In late 2019, Croatian Government issued revised defence expenditure which will see country increase defence expenditure to meet 2% NATO target by 2020, with 2019 and 2020 defence budgets seeing immediate revisions and increases to meet new spending plan. Defence expenditure in 2024 therefore based on current projections could reach 9.4 billion kuna or around 2% NATO requirement. Defence expenditures in recent years (source Croatian MOD); Although the budget has been decreased from year to year, the Croatian Armed Forces were able to maintain military readiness and to participate in major NATO exercises in Croatia and overseas. This downsizing of the armed forces has allowed for more funds to be allocated to modernisation over the past few years with an average of 1.6 billion kuna spent on modernisation, infrastructure and construction of new facilities. A $3 billion modernisation plan was proposed by the then Prime Minister Ivica Račan of the SDP led government in 2003, with planned modernisation starting in 2006 and ending in 2015. However it has been delayed in part due to the subsequent economic recession, but also due to serious corruption that has cost the Croatian MOD several billion kuna since 2006. A new plan under former Prime Minister Zoran Milanović should define exactly how and what the Croatian armed forces should look like by 2023. A defence white paper was published in 2015 with emphasis placed on modernisation of the Army. The Dr. Franjo Tuđman Military Academy acts as a school of higher learning responsible for training and educating future generations of military personnel. The academy consists of several schools including "Ban Josip Jelačić", "Blago Zadro", "Katarina Zrinska", the Officers Academy, and a school for non commissioned officers. The academy has 300 full-time staff and is the only military academy in Croatia. Each year also 100–120 foreign nationals attend the academy. The Commander-in-Chief of all Croatian armed forces in peace and war is the President of the Republic. The Commander-in-Chief prescribes the organisation of the Croatian Armed Forces at the proposal of the Chief of General Staff, with consent of the Minister of Defence. The Armed Forces consist of peacetime and wartime components. The peacetime component is composed of the active military officers, civil servants and employees in the Croatian Armed Forces, cadets, and conscripts serving a 6-month national service and reservists when on military exercise. The wartime component of the Armed Forces includes all other reservists. The General Staff is part of the Ministry of Defence in charge of commanding, training and use of the Armed Forces. It also has a number of units under its direct command, including the Special Operations Battalion, Honour Guard Battalion and several others. In peace, the Commander-in-Chief exercises his command through the Minister of Defence. In war and in cases where the Minister of Defence is not fulfilling his orders, the Commander-in-Chief exercises his command directly through the General Staff Commander. The Croatian Parliament exercises democratic control over the Armed Forces by adopting defence strategy, defence budget, and defence laws. Special Forces Command ("Zapovjedništvo specijalnih snaga", "ZSS") was established in February 2015, succeeding the Special Operations Battalion (Croatia), in accordance with the Long-term Development Plan of the Croatian Armed Forces in the period 2015–2024. The command staff is composed of the members who served in the special units, guards brigades and reconnaissance units of the Croatian Armed Forces. The main mission of the Special Forces Command is to ensure combat readiness of the special forces in the protection of the territorial integrity, sovereignty and independence of the Republic of Croatia, as well as for the participation in NATO-led operations. Colonel Perica Turalija is the current commanding officer of the command. The Croatian General Staff exercises direct command over the battalion which thus elevates the unit to strategic level deployment for quicker reaction and overall better and faster response to tactical and strategic situations. Also, this means that members of all three branches of the Croatian armed forces can apply for selection. Other special operations units are the Military Intelligence Battalion ("Vojno-obavještajna bojna", "VOB") and Special Military Police Company ("Satnija specijalne vojne policije", "SS VP"). The duties of an Honour Guard are performed by the Honor Guard Battalion ("Počasno zaštitna bojna"), located at Zagreb in the Tuškanac military base. A long term modernisation plan for 2015–2024 has been published outlining overall goals and is available for download (102 pages) at the Ministry of Defence of Croatia website. According to earlier reports from the government, the Croatian Armed Forces are set to receive vitally needed new equipment. Navy plans are still being worked on but present plans call for a moderate expansion of the naval force. As a small country, Croatia has a relatively well developed arms industry that is highly competitive internationally with significant annual arms exports. In 2012, Croatia managed to export nearly €120 million. However it has been reported in The New York Times that Croatia has been arming Syrian rebels with Croatian manufactured arms used during the Homeland War, arms Croatia no longer uses due to their obsolescence. Nevertheless, these arms played a crucial role in some significant rebel gains during 2012. As a result of these arms sales into this volatile region the Croatian government ordered the immediate withdrawal of the Croatian UN Golan Heights contingent to avoid their being targeted in retaliation. In 2013 Croatia exported €143 million worth of arms, however it is not clear if this also includes $36.5 million worth of arms Croatia exported to Jordan for Syrian rebels. Croatia was the top supplier of arms to Syrian rebels in 2013, but much of it through illicit channels without Croatian government approval or knowledge. Most of these arms were exported via Jordan. In 2014 Croatian arms exports reached 1.5 bn HRK (Croatian kuna) or €200 million or $257 million, the majority of exports being to NATO allies and Australia. In late 2014 the Croatian Defence Minister announced a major export deal to Iraq including the State of Kurdistan. This agreement includes the sale of 20,000 VHS Rifles, 150,000 complete sets of uniforms, helmets and associated equipment valued at €100 million. Croatian arms exports are growing steadily at 10–15% year-on-year and were expected to reach 1.75 billion HRK in 2015 or around €230 million, although much of the equipment exported is non-lethal. Croatian firms are well positioned on some major arms tenders in the Middle East, supplying complex military hardware such as the Patria AMV incorporating a newly developed 30mm overhead weapon station (valued at €1.25 million each) and said vehicles valued at €1.75 million. Kuwait, the UAE, and Saudi Arabia have been mentioned as potential customers, although no concrete contracts have been signed so far. Croatian firms are participating in Kuwaiti and UAE tenders for next-generation APC programmes, each valued at billions of euros. Croatian arms exports have grown steadily for the better part of this decade and have reached €325 million per year, placing Croatia in the top 10 arms exporters within NATO, behind the US, Germany, UK, France, Netherlands, Spain, Italy, Poland and Norway. The vast majority of these exports are to NATO partners such as the US, Norway, Australia, Canada, France. Croatia granted €5.75 billion in export licenses in 2016 and 2017; however, only a fraction of this sum has materialised in actual arms exports. On April 1, 2009 Croatia joined NATO and on July 1, 2013 it became the 28th member of the European Union. The Croatian Armed Forces participate in many of the (military) aspects of both organisations as well as actively participating in many United Nations peacekeeping operations worldwide.
https://en.wikipedia.org/wiki?curid=5581
Ancient Egyptian religion Ancient Egyptian religion was a complex system of polytheistic beliefs and rituals that formed an integral part of ancient Egyptian society. It centered on the Egyptians' interactions with many deities believed to be present in, and in control of, the world. Rituals such as prayer and offerings were provided to the gods to gain their favor. Formal religious practice centered on the pharaohs, the rulers of Egypt, believed to possess divine powers by virtue of their positions. They acted as intermediaries between their people and the gods, and were obligated to sustain the gods through rituals and offerings so that they could maintain "Ma'at", the order of the cosmos. The state dedicated enormous resources to religious rituals and to the construction of temples. Individuals could interact with the gods for their own purposes, appealing for help through prayer or compelling the gods to act through magic. These practices were distinct from, but closely linked with, the formal rituals and institutions. The popular religious tradition grew more prominent over the course of Egyptian history as the status of the pharaoh declined. Egyptian belief in the afterlife and the importance of funerary practices is evident in the great efforts made to ensure the survival of their souls after death – via the provision of tombs, grave goods and offerings to preserve the bodies and spirits of the deceased. The religion had its roots in Egypt's prehistory and lasted for more than 3,000 years. The details of religious belief changed over time as the importance of particular gods rose and declined, and their intricate relationships shifted. At various times, certain gods became preeminent over the others, including the sun god Ra, the creator god Amun, and the mother goddess Isis. For a brief period, in the theology promulgated by the pharaoh Akhenaten, a single god, the Aten, replaced the traditional pantheon. Ancient Egyptian religion and mythology left behind many writings and monuments, along with significant influences on ancient and modern cultures. The beliefs and rituals now referred to as "ancient Egyptian religion" were integral within every aspect of Egyptian culture. The Egyptian language possessed no single term corresponding to the modern European concept of religion. Ancient Egyptian religion consisted of a vast and varying set of beliefs and practices, linked by their common focus on the interaction between the world of humans and the world of the divine. The characteristics of the gods who populated the divine realm were inextricably linked to the Egyptians' understanding of the properties of the world in which they lived. The Egyptians believed that the phenomena of nature were divine forces in and of themselves. These deified forces included the elements, animal characteristics, or abstract forces. The Egyptians believed in a pantheon of gods, which were involved in all aspects of nature and human society. Their religious practices were efforts to sustain and placate these phenomena and turn them to human advantage. This polytheistic system was very complex, as some deities were believed to exist in many different manifestations, and some had multiple mythological roles. Conversely, many natural forces, such as the sun, were associated with multiple deities. The diverse pantheon ranged from gods with vital roles in the universe to minor deities or "demons" with very limited or localized functions. It could include gods adopted from foreign cultures, and sometimes humans: deceased pharaohs were believed to be divine, and occasionally, distinguished commoners such as Imhotep also became deified. The depictions of the gods in art were not meant as literal representations of how the gods might appear if they were visible, as the gods' true natures were believed to be mysterious. Instead, these depictions gave recognizable forms to the abstract deities by using symbolic imagery to indicate each god's role in nature. This iconography was not fixed, and many of the gods could be depicted in more than one form. Many gods were associated with particular regions in Egypt where their cults were most important. However, these associations changed over time, and they did not mean that the god associated with a place had originated there. For instance, the god Montu was the original patron of the city of Thebes. Over the course of the Middle Kingdom, however, he was displaced in that role by Amun, who may have arisen elsewhere. The national popularity and importance of individual gods fluctuated in a similar way. Deities had complex interrelationships, which partly reflected the interaction of the forces they represented. The Egyptians often grouped gods together to reflect these relationships. One of the more common combinations was a family triad consisting of a father, mother, and child, who were worshipped together. Some groups had wide-ranging importance. One such group, the Ennead, assembled nine deities into a theological system that was involved in the mythological areas of creation, kingship, and the afterlife. The relationships between deities could also be expressed in the process of syncretism, in which two or more different gods were linked to form a composite deity. This process was a recognition of the presence of one god "in" another when the second god took on a role belonging to the first. These links between deities were fluid, and did not represent the permanent merging of two gods into one; therefore, some gods could develop multiple syncretic connections. Sometimes, syncretism combined deities with very similar characteristics. At other times it joined gods with very different natures, as when Amun, the god of hidden power, was linked with Ra, the god of the sun. The resulting god, Amun-Ra, thus united the power that lay behind all things with the greatest and most visible force in nature. Many deities could be given epithets that seem to indicate that they were greater than any other god, suggesting some kind of unity beyond the multitude of natural forces. This is particularly true of a few gods who, at various points, rose to supreme importance in Egyptian religion. These included the royal patron Horus, the sun god Ra, and the mother goddess Isis. During the New Kingdom (c. 1550–1070 BC) Amun held this position. The theology of the period described in particular detail Amun's presence in and rule over all things, so that he, more than any other deity, embodied the all-encompassing power of the divine. The Egyptian conception of the universe centered on "Ma'at", a word that encompasses several concepts in English, including "truth," "justice," and "order." It was the fixed, eternal order of the universe, both in the cosmos and in human society, and was often personified as a goddess. It had existed since the creation of the world, and without it the world would lose its cohesion. In Egyptian belief, "Ma'at" was constantly under threat from the forces of disorder, so all of society was required to maintain it. On the human level this meant that all members of society should cooperate and coexist; on the cosmic level it meant that all of the forces of nature—the gods—should continue to function in balance. This latter goal was central to Egyptian religion. The Egyptians sought to maintain "Ma'at" in the cosmos by sustaining the gods through offerings and by performing rituals which staved off disorder and perpetuated the cycles of nature. The most important part of the Egyptian view of the cosmos was the conception of time, which was greatly concerned with the maintenance of "Ma'at". Throughout the linear passage of time, a cyclical pattern recurred, in which "Ma'at" was renewed by periodic events which echoed the original creation. Among these events were the annual Nile flood and the succession from one king to another, but the most important was the daily journey of the sun god Ra. When thinking of the shape of the cosmos, the Egyptians saw the earth as a flat expanse of land, personified by the god Geb, over which arched the sky goddess Nut. The two were separated by Shu, the god of air. Beneath the earth lay a parallel underworld and undersky, and beyond the skies lay the infinite expanse of Nu, the chaos that had existed before creation. The Egyptians also believed in a place called the Duat, a mysterious region associated with death and rebirth, that may have lain in the underworld or in the sky. Each day, Ra traveled over the earth across the underside of the sky, and at night he passed through the Duat to be reborn at dawn. In Egyptian belief, this cosmos was inhabited by three types of sentient beings. One was the gods; another was the spirits of deceased humans, who existed in the divine realm and possessed many of the gods' abilities. Living humans were the third category, and the most important among them was the pharaoh, who bridged the human and divine realms. Egyptologists have long debated the degree to which the pharaoh was considered a god. It seems most likely that the Egyptians viewed royal authority itself as a divine force. Therefore, although the Egyptians recognized that the pharaoh was human and subject to human weakness, they simultaneously viewed him as a god, because the divine power of kingship was incarnated in him. He therefore acted as intermediary between Egypt's people and the gods. He was key to upholding "Ma'at", both by maintaining justice and harmony in human society and by sustaining the gods with temples and offerings. For these reasons, he oversaw all state religious activity. However, the pharaoh's real-life influence and prestige could differ from his portrayal in official writings and depictions, and beginning in the late New Kingdom his religious importance declined drastically. The king was also associated with many specific deities. He was identified directly with Horus, who represented kingship itself, and he was seen as the son of Ra, who ruled and regulated nature as the pharaoh ruled and regulated society. By the New Kingdom he was also associated with Amun, the supreme force in the cosmos. Upon his death, the king became fully deified. In this state, he was directly identified with Ra, and was also associated with Osiris, god of death and rebirth and the mythological father of Horus. Many mortuary temples were dedicated to the worship of deceased pharaohs as gods. The Egyptians had elaborate beliefs about death and the afterlife. They believed that humans possessed a "ka", or life-force, which left the body at the point of death. In life, the "ka" received its sustenance from food and drink, so it was believed that, to endure after death, the "ka" must continue to receive offerings of food, whose spiritual essence it could still consume. Each person also had a "ba", the set of spiritual characteristics unique to each individual. Unlike the "ka", the "ba" remained attached to the body after death. Egyptian funeral rituals were intended to release the "ba" from the body so that it could move freely, and to rejoin it with the "ka" so that it could live on as an "akh". However, it was also important that the body of the deceased be preserved, as the Egyptians believed that the "ba" returned to its body each night to receive new life, before emerging in the morning as an "akh". In early times the deceased pharaoh was believed to ascend to the sky and dwell among the stars. Over the course of the Old Kingdom (c. 2686–2181 BC), however, he came to be more closely associated with the daily rebirth of the sun god Ra and with the underworld ruler Osiris as those deities grew more important. In the fully developed afterlife beliefs of the New Kingdom, the soul had to avoid a variety of supernatural dangers in the Duat, before undergoing a final judgment, known as the "Weighing of the Heart", carried out by Osiris and by the Assessors of Maat. In this judgment, the gods compared the actions of the deceased while alive (symbolized by the heart) to Maat, to determine whether he or she had behaved in accordance with Maat. If the deceased was judged worthy, his or her "ka" and "ba" were united into an "akh". Several beliefs coexisted about the "akh"'s destination. Often the dead were said to dwell in the realm of Osiris, a lush and pleasant land in the underworld. The solar vision of the afterlife, in which the deceased soul traveled with Ra on his daily journey, was still primarily associated with royalty, but could extend to other people as well. Over the course of the Middle and New Kingdoms, the notion that the "akh" could also travel in the world of the living, and to some degree magically affect events there, became increasingly prevalent. During the New Kingdom the pharaoh Akhenaten abolished the official worship of other gods in favor of the sun-disk Aten. This is often seen as the first instance of true monotheism in history, although the details of Atenist theology are still unclear and the suggestion that it was monotheistic is disputed. The exclusion of all but one god from worship was a radical departure from Egyptian tradition and some see Akhenaten as a practitioner of monolatry rather than monotheism, as he did not actively deny the existence of other gods; he simply refrained from worshipping any but the Aten. Under Akhenaten's successors Egypt reverted to its traditional religion, and Akhenaten himself came to be reviled as a heretic. While the Egyptians had no unified religious scripture, they produced many religious writings of various types. Together the disparate texts provide an extensive, but still incomplete, understanding of Egyptian religious practices and beliefs. Egyptian myths were metaphorical stories intended to illustrate and explain the gods' actions and roles in nature. The details of the events they recounted could change to convey different symbolic perspectives on the mysterious divine events they described, so many myths exist in different and conflicting versions. Mythical narratives were rarely written in full, and more often texts only contain episodes from or allusions to a larger myth. Knowledge of Egyptian mythology, therefore, is derived mostly from hymns that detail the roles of specific deities, from ritual and magical texts which describe actions related to mythic events, and from funerary texts which mention the roles of many deities in the afterlife. Some information is also provided by allusions in secular texts. Finally, Greeks and Romans such as Plutarch recorded some of the extant myths late in Egyptian history. Among the significant Egyptian myths were the creation myths. According to these stories, the world emerged as a dry space in the primordial ocean of chaos. Because the sun is essential to life on earth, the first rising of Ra marked the moment of this emergence. Different forms of the myth describe the process of creation in various ways: a transformation of the primordial god Atum into the elements that form the world, as the creative speech of the intellectual god Ptah, and as an act of the hidden power of Amun. Regardless of these variations, the act of creation represented the initial establishment of maat and the pattern for the subsequent cycles of time. The most important of all Egyptian myths was the Osiris myth. It tells of the divine ruler Osiris, who was murdered by his jealous brother Set, a god often associated with chaos. Osiris's sister and wife Isis resurrected him so that he could conceive an heir, Horus. Osiris then entered the underworld and became the ruler of the dead. Once grown, Horus fought and defeated Set to become king himself. Set's association with chaos, and the identification of Osiris and Horus as the rightful rulers, provided a rationale for Pharaonic succession and portrayed the Pharaohs as the upholders of order. At the same time, Osiris's death and rebirth were related to the Egyptian agricultural cycle, in which crops grew in the wake of the Nile inundation, and provided a template for the resurrection of human souls after death. Another important mythic motif was the journey of Ra through the Duat each night. In the course of this journey, Ra met with Osiris, who again acted as an agent of regeneration, so that his life was renewed. He also fought each night with Apep, a serpentine god representing chaos. The defeat of Apep and the meeting with Osiris ensured the rising of the sun the next morning, an event that represented rebirth and the victory of order over chaos. The procedures for religious rituals were frequently written on papyri, which were used as instructions for those performing the ritual. These ritual texts were kept mainly in the temple libraries. Temples themselves are also inscribed with such texts, often accompanied by illustrations. Unlike the ritual papyri, these inscriptions were not intended as instructions, but were meant to symbolically perpetuate the rituals even if, in reality, people ceased to perform them. Magical texts likewise describe rituals, although these rituals were part of the spells used for specific goals in everyday life. Despite their mundane purpose, many of these texts also originated in temple libraries and later became disseminated among the general populace. The Egyptians produced numerous prayers and hymns, written in the form of poetry. Hymns and prayers follow a similar structure and are distinguished mainly by the purposes they serve. Hymns were written to praise particular deities. Like ritual texts, they were written on papyri and on temple walls, and they were probably recited as part of the rituals they accompany in temple inscriptions. Most are structured according to a set literary formula, designed to expound on the nature, aspects, and mythological functions of a given deity. They tend to speak more explicitly about fundamental theology than other Egyptian religious writings, and became particularly important in the New Kingdom, a period of particularly active theological discourse. Prayers follow the same general pattern as hymns, but address the relevant god in a more personal way, asking for blessings, help, or forgiveness for wrongdoing. Such prayers are rare before the New Kingdom, indicating that in earlier periods such direct personal interaction with a deity was not believed possible, or at least was less likely to be expressed in writing. They are known mainly from inscriptions on statues and stelae left in sacred sites as votive offerings. Among the most significant and extensively preserved Egyptian writings are funerary texts designed to ensure that deceased souls reached a pleasant afterlife. The earliest of these are the Pyramid Texts. They are a loose collection of hundreds of spells inscribed on the walls of royal pyramids during the Old Kingdom, intended to magically provide pharaohs with the means to join the company of the gods in the afterlife. The spells appear in differing arrangements and combinations, and few of them appear in all of the pyramids. At the end of the Old Kingdom a new body of funerary spells, which included material from the Pyramid Texts, began appearing in tombs, inscribed primarily on coffins. This collection of writings is known as the Coffin Texts, and was not reserved for royalty, but appeared in the tombs of non-royal officials. In the New Kingdom, several new funerary texts emerged, of which the best-known is the Book of the Dead. Unlike the earlier books, it often contains extensive illustrations, or vignettes. The book was copied on papyrus and sold to commoners to be placed in their tombs. The Coffin Texts included sections with detailed descriptions of the underworld and instructions on how to overcome its hazards. In the New Kingdom, this material gave rise to several "books of the netherworld", including the Book of Gates, the Book of Caverns, and the Amduat. Unlike the loose collections of spells, these netherworld books are structured depictions of Ra's passage through the Duat, and by analogy, the journey of the deceased person's soul through the realm of the dead. They were originally restricted to pharaonic tombs, but in the Third Intermediate Period they came to be used more widely. Temples existed from the beginning of Egyptian history, and at the height of the civilization they were present in most of its towns. They included both mortuary temples to serve the spirits of deceased pharaohs and temples dedicated to patron gods, although the distinction was blurred because divinity and kingship were so closely intertwined. The temples were not primarily intended as places for worship by the general populace, and the common people had a complex set of religious practices of their own. Instead, the state-run temples served as houses for the gods, in which physical images which served as their intermediaries were cared for and provided with offerings. This service was believed to be necessary to sustain the gods, so that they could in turn maintain the universe itself. Thus, temples were central to Egyptian society, and vast resources were devoted to their upkeep, including both donations from the monarchy and large estates of their own. Pharaohs often expanded them as part of their obligation to honor the gods, so that many temples grew to enormous size. However, not all gods had temples dedicated to them, as many gods who were important in official theology received only minimal worship, and many household gods were the focus of popular veneration rather than temple ritual. The earliest Egyptian temples were small, impermanent structures, but through the Old and Middle Kingdoms their designs grew more elaborate, and they were increasingly built out of stone. In the New Kingdom, a basic temple layout emerged, which had evolved from common elements in Old and Middle Kingdom temples. With variations, this plan was used for most of the temples built from then on, and most of those that survive today adhere to it. In this standard plan, the temple was built along a central processional way that led through a series of courts and halls to the sanctuary, which held a statue of the temple's god. Access to this most sacred part of the temple was restricted to the pharaoh and the highest-ranking priests. The journey from the temple entrance to the sanctuary was seen as a journey from the human world to the divine realm, a point emphasized by the complex mythological symbolism present in temple architecture. Well beyond the temple building proper was the outermost wall. Between the two lay many subsidiary buildings, including workshops and storage areas to supply the temple's needs, and the library where the temple's sacred writings and mundane records were kept, and which also served as a center of learning on a multitude of subjects. Theoretically it was the duty of the pharaoh to carry out temple rituals, as he was Egypt's official representative to the gods. In reality, ritual duties were almost always carried out by priests. During the Old and Middle Kingdoms, there was no separate class of priests; instead, many government officials served in this capacity for several months out of the year before returning to their secular duties. Only in the New Kingdom did professional priesthood become widespread, although most lower-ranking priests were still part-time. All were still employed by the state, and the pharaoh had final say in their appointments. However, as the wealth of the temples grew, the influence of their priesthoods increased, until it rivaled that of the pharaoh. In the political fragmentation of the Third Intermediate Period (c. 1070–664 BC), the high priests of Amun at Karnak even became the effective rulers of Upper Egypt. The temple staff also included many people other than priests, such as musicians and chanters in temple ceremonies. Outside the temple were artisans and other laborers who helped supply the temple's needs, as well as farmers who worked on temple estates. All were paid with portions of the temple's income. Large temples were therefore very important centers of economic activity, sometimes employing thousands of people. State religious practice included both temple rituals involved in the cult of a deity, and ceremonies related to divine kingship. Among the latter were coronation ceremonies and the sed festival, a ritual renewal of the pharaoh's strength that took place periodically during his reign. There were numerous temple rituals, including rites that took place across the country and rites limited to single temples or to the temples of a single god. Some were performed daily, while others took place annually or on rare occasions. The most common temple ritual was the morning offering ceremony, performed daily in temples across Egypt. In it, a high-ranking priest, or occasionally the pharaoh, washed, anointed, and elaborately dressed the god's statue before presenting it with offerings. Afterward, when the god had consumed the spiritual essence of the offerings, the items themselves were taken to be distributed among the priests. The less frequent temple rituals, or festivals, were still numerous, with dozens occurring every year. These festivals often entailed actions beyond simple offerings to the gods, such as reenactments of particular myths or the symbolic destruction of the forces of disorder. Most of these events were probably celebrated only by the priests and took place only inside the temple. However, the most important temple festivals, like the Opet Festival celebrated at Karnak, usually involved a procession carrying the god's image out of the sanctuary in a model barque to visit other significant sites, such as the temple of a related deity. Commoners gathered to watch the procession and sometimes received portions of the unusually large offerings given to the gods on these occasions. At many sacred sites, the Egyptians worshipped individual animals which they believed to be manifestations of particular deities. These animals were selected based on specific sacred markings which were believed to indicate their fitness for the role. Some of these cult animals retained their positions for the rest of their lives, as with the Apis bull worshipped in Memphis as a manifestation of Ptah. Other animals were selected for much shorter periods. These cults grew more popular in later times, and many temples began raising stocks of such animals from which to choose a new divine manifestation. A separate practice developed in the Twenty-sixth Dynasty, when people began mummifying any member of a particular animal species as an offering to the god whom the species represented. Millions of mummified cats, birds, and other creatures were buried at temples honoring Egyptian deities. Worshippers paid the priests of a particular deity to obtain and mummify an animal associated with that deity, and the mummy was placed in a cemetery near the god's cult center. The Egyptians used oracles to ask the gods for knowledge or guidance. Egyptian oracles are known mainly from the New Kingdom and afterward, though they probably appeared much earlier. People of all classes, including the king, asked questions of oracles, and, especially in the late New Kingdom their answers could be used to settle legal disputes or inform royal decisions. The most common means of consulting an oracle was to pose a question to the divine image while it was being carried in a festival procession, and interpret an answer from the barque's movements. Other methods included interpreting the behavior of cult animals, drawing lots, or consulting statues through which a priest apparently spoke. The means of discerning the god's will gave great influence to the priests who spoke and interpreted the god's message. While the state cults were meant to preserve the stability of the Egyptian world, lay individuals had their own religious practices that related more directly to daily life. This popular religion left less evidence than the official cults, and because this evidence was mostly produced by the wealthiest portion of the Egyptian population, it is uncertain to what degree it reflects the practices of the populace as a whole. Popular religious practice included ceremonies marking important transitions in life. These included birth, because of the danger involved in the process, and naming, because the name was held to be a crucial part of a person's identity. The most important of these ceremonies were those surrounding death, because they ensured the soul's survival beyond it. Other religious practices sought to discern the gods' will or seek their knowledge. These included the interpretation of dreams, which could be seen as messages from the divine realm, and the consultation of oracles. People also sought to affect the gods' behavior to their own benefit through magical rituals. Individual Egyptians also prayed to gods and gave them private offerings. Evidence of this type of personal piety is sparse before the New Kingdom. This is probably due to cultural restrictions on depiction of nonroyal religious activity, which relaxed during the Middle and New Kingdoms. Personal piety became still more prominent in the late New Kingdom, when it was believed that the gods intervened directly in individual lives, punishing wrongdoers and saving the pious from disaster. Official temples were important venues for private prayer and offering, even though their central activities were closed to laypeople. Egyptians frequently donated goods to be offered to the temple deity and objects inscribed with prayers to be placed in temple courts. Often they prayed in person before temple statues or in shrines set aside for their use. Yet in addition to temples, the populace also used separate local chapels, smaller but more accessible than the formal temples. These chapels were very numerous and probably staffed by members of the community. Households, too, often had their own small shrines for offering to gods or deceased relatives. The deities invoked in these situations differed somewhat from those at the center of state cults. Many of the important popular deities, such as the fertility goddess Taweret and the household protector Bes, had no temples of their own. However, many other gods, including Amun and Osiris, were very important in both popular and official religion. Some individuals might be particularly devoted to a single god. Often they favored deities affiliated with their own region, or with their role in life. The god Ptah, for instance, was particularly important in his cult center of Memphis, but as the patron of craftsmen he received the nationwide veneration of many in that occupation. The word "magic" could be used to translate the Egyptian term "heka", which meant, as James P. Allen puts it, "the ability to make things happen by indirect means". Heka was believed to be a natural phenomenon, the force which was used to create the universe and which the gods employed to work their will. Humans could also use it, however, and magical practices were closely intertwined with religion. In fact, even the regular rituals performed in temples were counted as magic. Individuals also frequently employed magical techniques for personal purposes. Although these ends could be harmful to other people, no form of magic was considered inimical in itself. Instead, magic was seen primarily as a way for humans to prevent or overcome negative events. Magic was closely associated with the priesthood. Because temple libraries contained numerous magical texts, great magical knowledge was ascribed to the lector priests who studied these texts. These priests often worked outside their temples, hiring out their magical services to laymen. Other professions also commonly employed magic as part of their work, including doctors, scorpion-charmers, and makers of magical amulets. It is also possible that the peasantry used simple magic for their own purposes, but because this magical knowledge would have been passed down orally, there is limited evidence of it. Language was closely linked with heka, to such a degree that Thoth, the god of writing, was sometimes said to be the inventor of heka. Therefore, magic frequently involved written or spoken incantations, although these were usually accompanied by ritual actions. Often these rituals invoked the power of an appropriate deity to perform the desired action, using the power of heka to compel it to act. Sometimes this entailed casting the practitioner or subject of a ritual in the role of a character in mythology, thus inducing the god to act toward that person as it had in the myth. Rituals also employed sympathetic magic, using objects believed to have a magically significant resemblance to the subject of the rite. The Egyptians also commonly used objects believed to be imbued with heka of their own, such as the magically protective amulets worn in great numbers by ordinary Egyptians. Because it was considered necessary for the survival of the soul, preservation of the body was a central part of Egyptian funerary practices. Originally the Egyptians buried their dead in the desert, where the arid conditions mummified the body naturally. In the Early Dynastic Period, however, they began using tombs for greater protection, and the body was insulated from the desiccating effect of the sand and was subject to natural decay. Thus the Egyptians developed their elaborate embalming practices, in which the corpse was artificially desiccated and wrapped to be placed in its coffin. The quality of the process varied according to cost, however, and those who could not afford it were still buried in desert graves. Once the mummification process was complete, the mummy was carried from the deceased person's house to the tomb in a funeral procession that included his or her friends and relatives, along with a variety of priests. Before the burial, these priests performed several rituals, including the Opening of the mouth ceremony intended to restore the dead person's senses and give him or her the ability to receive offerings. Then the mummy was buried and the tomb sealed. Afterward, relatives or hired priests gave food offerings to the deceased in a nearby mortuary chapel at regular intervals. Over time, families inevitably neglected offerings to long-dead relatives, so most mortuary cults only lasted one or two generations. However, while the cult lasted, the living sometimes wrote letters asking deceased relatives for help, in the belief that the dead could affect the world of the living as the gods did. The first Egyptian tombs were mastabas, rectangular brick structures where kings and nobles were entombed. Each of them contained a subterranean burial chamber and a separate, above ground chapel for mortuary rituals. In the Old Kingdom the mastaba developed into the pyramid, which symbolized the primeval mound of Egyptian myth. Pyramids were reserved for royalty, and were accompanied by large mortuary temples sitting at their base. Middle Kingdom pharaohs continued to build pyramids, but the popularity of mastabas waned. Increasingly, commoners with sufficient means were buried in rock-cut tombs with separate mortuary chapels nearby, an approach which was less vulnerable to tomb robbery. By the beginning of the New Kingdom even the pharaohs were buried in such tombs, and they continued to be used until the decline of the religion itself. Tombs could contain a great variety of other items, including statues of the deceased to serve as substitutes for the body in case it was damaged. Because it was believed that the deceased would have to do work in the afterlife, just as in life, burials often included small models of humans to do work in place of the deceased. The tombs of wealthier individuals could also contain furniture, clothing, and other everyday objects intended for use in the afterlife, along with amulets and other items intended to provide magical protection against the hazards of the spirit world. Further protection was provided by funerary texts included in the burial. The tomb walls also bore artwork, such as images of the deceased eating food that were believed to allow him or her to magically receive sustenance even after the mortuary offerings had ceased. The beginnings of Egyptian religion extend into prehistory, though evidence for them comes only from the sparse and ambiguous archaeological record. Careful burials during the Predynastic period imply that the people of this time believed in some form of an afterlife. At the same time, animals were ritually buried, a practice which may reflect the development of zoomorphic deities like those found in the later religion. The evidence is less clear for gods in human form, and this type of deity may have emerged more slowly than those in animal shape. Each region of Egypt originally had its own patron deity, but it is likely that as these small communities conquered or absorbed each other, the god of the defeated area was either incorporated into the other god's mythology or entirely subsumed by it. This resulted in a complex pantheon in which some deities remained only locally important while others developed more universal significance. The Early Dynastic Period began with the unification of Egypt around 3000 BC. This event transformed Egyptian religion, as some deities rose to national importance and the cult of the divine pharaoh became the central focus of religious activity. Horus was identified with the king, and his cult center in the Upper Egyptian city of Nekhen was among the most important religious sites of the period. Another important center was Abydos, where the early rulers built large funerary complexes. During the Old Kingdom, the priesthoods of the major deities attempted to organize the complicated national pantheon into groups linked by their mythology and worshipped in a single cult center, such as the Ennead of Heliopolis which linked important deities such as Atum, Ra, Osiris, and Set in a single creation myth. Meanwhile, pyramids, accompanied by large mortuary temple complexes, replaced mastabas as the tombs of pharaohs. In contrast with the great size of the pyramid complexes, temples to gods remained comparatively small, suggesting that official religion in this period emphasized the cult of the divine king more than the direct worship of deities. The funerary rituals and architecture of this time greatly influenced the more elaborate temples and rituals used in worshipping the gods in later periods. Early in the Old Kingdom, Ra grew in influence, and his cult center at Heliopolis became the nation's most important religious site. By the Fifth Dynasty, Ra was the most prominent god in Egypt, and had developed the close links with kingship and the afterlife that he retained for the rest of Egyptian history. Around the same time, Osiris became an important afterlife deity. The Pyramid Texts"," first written at this time, reflect the prominence of the solar and Osirian concepts of the afterlife, although they also contain remnants of much older traditions. The texts are an extremely important source for understanding early Egyptian theology. In the 22nd century BC, the Old Kingdom collapsed into the disorder of the First Intermediate Period. Eventually rulers from Thebes reunified the Egyptian nation in the Middle Kingdom (c. 2055–1650 BC). These Theban pharaohs initially promoted their patron god Montu to national importance, but during the Middle Kingdom, he was eclipsed by the rising popularity of Amun. In this new Egyptian state, personal piety grew more important and was expressed more freely in writing, a trend that continued in the New Kingdom. The Middle Kingdom crumbled in the Second Intermediate Period (c. 1650–1550 BC), but the country was again reunited by Theban rulers, who became the first pharaohs of the New Kingdom. Under the new regime, Amun became the supreme state god. He was syncretized with Ra, the long-established patron of kingship and his temple at Karnak in Thebes became Egypt's most important religious center. Amun's elevation was partly due to the great importance of Thebes, but it was also due to the increasingly professional priesthood. Their sophisticated theological discussion produced detailed descriptions of Amun's universal power. Increased contact with outside peoples in this period led to the adoption of many Near Eastern deities into the pantheon. At the same time, the subjugated Nubians absorbed Egyptian religious beliefs, and in particular, adopted Amun as their own. The New Kingdom religious order was disrupted when Akhenaten acceded, and replaced Amun with the Aten as the state god. Eventually, he eliminated the official worship of most other gods and moved Egypt's capital to a new city at Amarna. This part of Egyptian history, the Amarna Period, is named after this. In doing so, Akhenaten claimed unprecedented status: only he could worship the Aten, and the populace directed their worship toward him. The Atenist system lacked well-developed mythology and afterlife beliefs, and the Aten seemed distant and impersonal, so the new order did not appeal to ordinary Egyptians. Thus, many probably continued to worship the traditional gods in private. Nevertheless, the withdrawal of state support for the other deities severely disrupted Egyptian society. Akhenaten's successors restored the traditional religious system, and eventually, they dismantled all Atenist monuments. Before the Amarna period, popular religion had trended toward more personal relationships between worshippers and their gods. Akhenaten's changes had reversed this trend, but once the traditional religion was restored, there was a backlash. The populace began to believe that the gods were much more directly involved in daily life. Amun, the supreme god, was increasingly seen as the final arbiter of human destiny, the true ruler of Egypt. The pharaoh was correspondingly more human and less divine. The importance of oracles as a means of decision-making grew, as did the wealth and influence of the oracles' interpreters, the priesthood. These trends undermined the traditional structure of society and contributed to the breakdown of the New Kingdom. In the 1st millennium BC, Egypt was significantly weaker than in earlier times, and in several periods foreigners seized the country and assumed the position of pharaoh. The importance of the pharaoh continued to decline, and the emphasis on popular piety continued to increase. Animal cults, a characteristically Egyptian form of worship, became increasingly popular in this period, possibly as a response to the uncertainty and foreign influence of the time. Isis grew more popular as a goddess of protection, magic, and personal salvation, and became the most important goddess in Egypt. In the 4th century BC, Egypt became a Hellenistic kingdom under the Ptolemaic dynasty (305–30 BC), which assumed the pharaonic role, maintaining the traditional religion and building or rebuilding many temples. The kingdom's Greek ruling class identified the Egyptian deities with their own. From this cross-cultural syncretism emerged Serapis, a god who combined Osiris and Apis with characteristics of Greek deities, and who became very popular among the Greek population. Nevertheless, for the most part the two belief systems remained separate, and the Egyptian deities remained Egyptian. Ptolemaic-era beliefs changed little after Egypt became a province of the Roman Empire in 30 BC, with the Ptolemaic kings replaced by distant emperors. The cult of Isis appealed even to Greeks and Romans outside Egypt, and in Hellenized form it spread across the empire. In Egypt itself, as the empire weakened, official temples fell into decay, and without their centralizing influence religious practice became fragmented and localized. Meanwhile, Christianity spread across Egypt, and in the third and fourth centuries AD, edicts by Christian emperors and iconoclasm by local Christians eroded traditional beliefs. While it persisted among the populace for some time, Egyptian religion slowly faded away. Egyptian religion produced the temples and tombs which are ancient Egypt's most enduring monuments, but it also influenced other cultures. In pharaonic times many of its symbols, such as the sphinx and winged solar disk, were adopted by other cultures across the Mediterranean and Near East, as were some of its deities, such as Bes. Some of these connections are difficult to trace. The Greek concept of Elysium may have derived from the Egyptian vision of the afterlife. In late antiquity, the Christian conception of Hell was most likely influenced by some of the imagery of the Duat. Egyptian beliefs also influenced or gave rise to several esoteric belief systems developed by Greeks and Romans, who considered Egypt as a source of mystic wisdom. Hermeticism, for instance, derived from the tradition of secret magical knowledge associated with Thoth. Traces of ancient beliefs remained in Egyptian folk traditions into modern times, but its influence on modern societies greatly increased with the French Campaign in Egypt and Syria in 1798 and their seeing the monuments and images. As a result of it, Westerners began to study Egyptian beliefs firsthand, and Egyptian religious motifs were adopted into Western art. Egyptian religion has since had a significant influence in popular culture. Due to continued interest in Egyptian beliefs, in the late 20th century, several new religious groups going under the blanket term of Kemetism have formed based on different reconstructions of ancient Egyptian religion.
https://en.wikipedia.org/wiki?curid=10331
Educational psychology Educational psychology is the branch of psychology concerned with the scientific study of human learning. The study of learning processes, from both cognitive and behavioral perspectives, allows researchers to understand individual differences in intelligence, cognitive development, affect, motivation, self-regulation, and self-concept, as well as their role in learning. The field of educational psychology relies heavily on quantitative methods, including testing and measurement, to enhance educational activities related to instructional design, classroom management, and assessment, which serve to facilitate learning processes in various educational settings across the lifespan. Educational psychology can in part be understood through its relationship with other disciplines. It is informed primarily by psychology, bearing a relationship to that discipline analogous to the relationship between medicine and biology. It is also informed by neuroscience. Educational psychology in turn informs a wide range of specialities within educational studies, including instructional design, educational technology, curriculum development, organizational learning, special education, classroom management, and student motivation. Educational psychology both draws from and contributes to cognitive science and the learning sciences. In universities, departments of educational psychology are usually housed within faculties of education, possibly accounting for the lack of representation of educational psychology content in introductory psychology textbooks. The field of educational psychology involves the study of memory, conceptual processes, and individual differences (via cognitive psychology) in conceptualizing new strategies for learning processes in humans. Educational psychology has been built upon theories of operant conditioning, functionalism, structuralism, constructivism, humanistic psychology, Gestalt psychology, and information processing. Educational psychology has seen rapid growth and development as a profession in the last twenty years. School psychology began with the concept of intelligence testing leading to provisions for special education students, who could not follow the regular classroom curriculum in the early part of the 20th century. However, "school psychology" itself has built a fairly new profession based upon the practices and theories of several psychologists among many different fields. Educational psychologists are working side by side with psychiatrists, social workers, teachers, speech and language therapists, and counselors in attempt to understand the questions being raised when combining behavioral, cognitive, and social psychology in the classroom setting. Educational psychology is a fairly new and growing field of study. Although it can date back as early as the days of Plato and Aristotle, educational psychology was not considered a specific practice. It was unknown that everyday teaching and learning in which individuals had to think about individual differences, assessment, development, the nature of a subject being taught, problem solving, and transfer of learning was the beginning to the field of educational psychology. These topics are important to education and, as a result, they are important in understanding human cognition, learning, and social perception. Educational psychology dates back to the time of Aristotle and Plato. Plato and Aristotle researched individual differences in the field of education, training of the body and the cultivation of psycho-motor skills, the formation of good character, the possibilities and limits of moral education. Some other educational topics they spoke about were the effects of music, poetry, and the other arts on the development of individual, role of teacher, and the relations between teacher and student. Plato saw knowledge acquisition as an innate ability, which evolves through experience and understanding of the world. This conception of human cognition has evolved into a continuing argument of nature vs. nurture in understanding conditioning and learning today. Aristotle observed the phenomenon of "association." His four laws of association included succession, contiguity, similarity, and contrast. His studies examined recall and facilitated learning processes. John Locke is considered one of the most influential philosophers in post-renaissance Europe, a time period which began around the mid-1600s. Locke is considered the "Father of English Psychology". One of Locke's most important works was written in 1690, named "An Essay Concerning Human Understanding". In this essay, he introduced the term "tabula rasa" meaning "blank slate." Locke explained that learning was attained through experience only, and that we are all born without knowledge. He followed by contrasting Plato's theory of innate learning processes. Locke believed the mind was formed by experiences, not innate ideas. Locke introduced this idea as "empiricism," or the understanding that knowledge is only built on knowledge and experience. In the late 1600s, John Locke advanced the hypothesis that people learn primarily from external forces. He believed that the mind was like a blank tablet (tabula rasa), and that successions of simple impressions give rise to complex ideas through association and reflection. Locke is credited with establishing "empiricism" as a criterion for testing the validity of knowledge, thus providing a conceptual framework for later development of experimental methodology in the natural and social sciences. Philosophers of education such as Juan Vives, Johann Pestalozzi, Friedrich Fröbel, and Johann Herbart had examined, classified and judged the methods of education centuries before the beginnings of psychology in the late 1800s. Juan Vives (1493–1540) proposed induction as the method of study and believed in the direct observation and investigation of the study of nature. His studies focused on humanistic learning, which opposed scholasticism and was influenced by a variety of sources including philosophy, psychology, politics, religion, and history. He was one of the first prominent thinkers to emphasize that the location of a school is important to learning. He suggested that a school should be located away from disturbing noises; the air quality should be good and there should be plenty of food for the students and teachers. Vives emphasized the importance of understanding individual differences of the students and suggested practice as an important tool for learning. Vives introduced his educational ideas in his writing, "De anima et vita" in 1538. In this publication, Vives explores moral philosophy as a setting for his educational ideals; with this, he explains that the different parts of the soul (similar to that of Aristotle's ideas) are each responsible for different operations, which function distinctively. The first book covers the different "souls": "The Vegetative Soul;" this is the soul of nutrition, growth, and reproduction, "The Sensitive Soul," which involves the five external senses; "The Cogitative soul," which includes internal senses and cognitive facilities. The second book involves functions of the rational soul: mind, will, and memory. Lastly, the third book explains the analysis of emotions. Johann Pestalozzi (1746–1827), a Swiss educational reformer, emphasized the child rather than the content of the school. Pestalozzi fostered an educational reform backed by the idea that early education was crucial for children, and could be manageable for mothers. Eventually, this experience with early education would lead to a "wholesome person characterized by morality." Pestalozzi has been acknowledged for opening institutions for education, writing books for mother's teaching home education, and elementary books for students, mostly focusing on the kindergarten level. In his later years, he published teaching manuals and methods of teaching. During the time of The Enlightenment, Pestalozzi's ideals introduced "educationalization". This created the bridge between social issues and education by introducing the idea of social issues to be solved through education. Horlacher describes the most prominent example of this during The Enlightenment to be "improving agricultural production methods." Johann Herbart (1776–1841) is considered the father of educational psychology. He believed that learning was influenced by interest in the subject and the teacher. He thought that teachers should consider the students' existing mental sets—what they already know—when presenting new information or material. Herbart came up with what are now known as the formal steps. The 5 steps that teachers should use are: There were three major figures in educational psychology in this period: William James, G. Stanley Hall, and John Dewey. These three men distinguished themselves in general psychology and educational psychology, which overlapped significantly at the end of the 19th century. The period of 1890–1920 is considered the golden era of educational psychology where aspirations of the new discipline rested on the application of the scientific methods of observation and experimentation to educational problems. From 1840 to 1920 37 million people immigrated to the United States. This created an expansion of elementary schools and secondary schools. The increase in immigration also provided educational psychologists the opportunity to use intelligence testing to screen immigrants at Ellis Island. Darwinism influenced the beliefs of the prominent educational psychologists. Even in the earliest years of the discipline, educational psychologists recognized the limitations of this new approach. The pioneering American psychologist William James commented that: James is the father of psychology in America but he also made contributions to educational psychology. In his famous series of lectures "Talks to Teachers on Psychology", published in 1899, James defines education as "the organization of acquired habits of conduct and tendencies to behavior". He states that teachers should "train the pupil to behavior" so that he fits into the social and physical world. Teachers should also realize the importance of habit and instinct. They should present information that is clear and interesting and relate this new information and material to things the student already knows about. He also addresses important issues such as attention, memory, and association of ideas. Alfred Binet published "Mental Fatigue" in 1898, in which he attempted to apply the experimental method to educational psychology. In this experimental method he advocated for two types of experiments, experiments done in the lab and experiments done in the classroom. In 1904 he was appointed the Minister of Public Education. This is when he began to look for a way to distinguish children with developmental disabilities. Binet strongly supported special education programs because he believed that "abnormality" could be cured. The Binet-Simon test was the first intelligence test and was the first to distinguish between "normal children" and those with developmental disabilities. Binet believed that it was important to study individual differences between age groups and children of the same age. He also believed that it was important for teachers to take into account individual students strengths and also the needs of the classroom as a whole when teaching and creating a good learning environment. He also believed that it was important to train teachers in observation so that they would be able to see individual differences among children and adjust the curriculum to the students. Binet also emphasized that practice of material was important. In 1916 Lewis Terman revised the Binet-Simon so that the average score was always 100. The test became known as the Stanford-Binet and was one of the most widely used tests of intelligence. Terman, unlike Binet, was interested in using intelligence test to identify gifted children who had high intelligence. In his longitudinal study of gifted children, who became known as the Termites, Terman found that gifted children become gifted adults. Edward Thorndike (1874–1949) supported the scientific movement in education. He based teaching practices on empirical evidence and measurement. Thorndike developed the theory of instrumental conditioning or the law of effect. The law of effect states that associations are strengthened when it is followed by something pleasing and associations are weakened when followed by something not pleasing. He also found that learning is done a little at a time or in increments, learning is an automatic process and all the principles of learning apply to all mammals. Thorndike's research with Robert Woodworth on the theory of transfer found that learning one subject will only influence your ability to learn another subject if the subjects are similar. This discovery led to less emphasis on learning the classics because they found that studying the classics does not contribute to overall general intelligence. Thorndike was one of the first to say that individual differences in cognitive tasks were due to how many stimulus response patterns a person had rather than a general intellectual ability. He contributed word dictionaries that were scientifically based to determine the words and definitions used. The dictionaries were the first to take into consideration the users maturity level. He also integrated pictures and easier pronunciation guide into each of the definitions. Thorndike contributed arithmetic books based on learning theory. He made all the problems more realistic and relevant to what was being studied, not just to improve the general intelligence. He developed tests that were standardized to measure performance in school related subjects. His biggest contribution to testing was the CAVD intelligence test which used a multidimensional approach to intelligence and the first to use a ratio scale. His later work was on programmed instruction, mastery learning and computer-based learning: John Dewey (1859–1952) had a major influence on the development of progressive education in the United States. He believed that the classroom should prepare children to be good citizens and facilitate creative intelligence. He pushed for the creation of practical classes that could be applied outside of a school setting. He also thought that education should be student-oriented, not subject-oriented. For Dewey, education was a social experience that helped bring together generations of people. He stated that students learn by doing. He believed in an active mind that was able to be educated through observation, problem solving and enquiry. In his 1910 book "How We Think", he emphasizes that material should be provided in a way that is stimulating and interesting to the student since it encourages original thought and problem solving. He also stated that material should be relative to the student's own experience. Jean Piaget (1896–1980) was one of the most powerful researchers in the area of developmental psychology during the 20th century. He developed the theory of cognitive development. The theory stated that intelligence developed in four different stages. The stages are the sensorimotor stage from birth to 2 years old, the preoperational state from 2 years old to 7 years old, the concrete operational stage from 7 years old to 10 years old, and formal operational stage from 12 years old and up. He also believed that learning was constrained to the child's cognitive development. Piaget influenced educational psychology because he was the first to believe that cognitive development was important and something that should be paid attention to in education. Most of the research on Piagetian theory was carried out by American educational psychologists. The number of people receiving a high school and college education increased dramatically from 1920 to 1960. Because very few jobs were available to teens coming out of eighth grade, there was an increase in high school attendance in the 1930s. The progressive movement in the United States took off at this time and led to the idea of progressive education. John Flanagan, an educational psychologist, developed tests for combat trainees and instructions in combat training. In 1954 the work of Kenneth Clark and his wife on the effects of segregation on black and white children was influential in the Supreme Court case Brown v. Board of Education. From the 1960s to present day, educational psychology has switched from a behaviorist perspective to a more cognitive based perspective because of the influence and development of cognitive psychology at this time. Jerome Bruner is notable for integrating Piaget's cognitive approaches into educational psychology. He advocated for discovery learning where teachers create a problem solving environment that allows the student to question, explore and experiment. In his book "The Process of Education" Bruner stated that the structure of the material and the cognitive abilities of the person are important in learning. He emphasized the importance of the subject matter. He also believed that how the subject was structured was important for the student's understanding of the subject and it is the goal of the teacher to structure the subject in a way that was easy for the student to understand. In the early 1960s Bruner went to Africa to teach math and science to school children, which influenced his view as schooling as a cultural institution. Bruner was also influential in the development of MACOS, , which was an educational program that combined anthropology and science. The program explored human evolution and social behavior. He also helped with the development of the head start program. He was interested in the influence of culture on education and looked at the impact of poverty on educational development. Benjamin Bloom (1903–1999) spent over 50 years at the University of Chicago, where he worked in the department of education. He believed that all students can learn. He developed taxonomy of educational objectives. The objectives were divided into three domains: cognitive, affective, and psychomotor. The cognitive domain deals with how we think. It is divided into categories that are on a continuum from easiest to more complex. The categories are knowledge or recall, comprehension, application, analysis, synthesis and evaluation. The affective domain deals with emotions and has 5 categories. The categories are receiving phenomenon, responding to that phenomenon, valuing, organization, and internalizing values. The psychomotor domain deals with the development of motor skills, movement and coordination and has 7 categories, that also goes from simplest to complex. The 7 categories of the psychomotor domain are perception, set, guided response, mechanism, complex overt response, adaptation, and origination. The taxonomy provided broad educational objectives that could be used to help expand the curriculum to match the ideas in the taxonomy. The taxonomy is considered to have a greater influence internationally than in the United States. Internationally, the taxonomy is used in every aspect of education from training of the teachers to the development of testing material. Bloom believed in communicating clear learning goals and promoting an active student. He thought that teachers should provide feedback to the students on their strengths and weaknesses. Bloom also did research on college students and their problem solving processes. He found that they differ in understanding the basis of the problem and the ideas in the problem. He also found that students differ in process of problem solving in their approach and attitude toward the problem. Nathaniel Gage (1917 -2008) is an important figure in educational psychology as his research focused on improving teaching and understanding the processes involved in teaching. He edited the book "Handbook of Research on Teaching" (1963), which helped develop early research in teaching and educational psychology. Gage founded the Stanford Center for Research and Development in Teaching, which contributed research on teaching as well as influencing the education of important educational psychologists. Applied behavior analysis, a research-based science utilizing behavioral principles of operant conditioning, is effective in a range of educational settings. For example, teachers can alter student behavior by systematically rewarding students who follow classroom rules with praise, stars, or tokens exchangeable for sundry items. Despite the demonstrated efficacy of awards in changing behavior, their use in education has been criticized by proponents of self-determination theory, who claim that praise and other rewards undermine intrinsic motivation. There is evidence that tangible rewards decrease intrinsic motivation in specific situations, such as when the student already has a high level of intrinsic motivation to perform the goal behavior. But the results showing detrimental effects are counterbalanced by evidence that, in other situations, such as when rewards are given for attaining a gradually increasing standard of performance, rewards enhance intrinsic motivation. Many effective therapies have been based on the principles of applied behavior analysis, including pivotal response therapy which is used to treat autism spectrum disorders. Among current educational psychologists, the cognitive perspective is more widely held than the behavioral perspective, perhaps because it admits causally related mental constructs such as traits, beliefs, memories, motivations and emotions. Cognitive theories claim that memory structures determine how information is perceived, processed, stored, retrieved and forgotten. Among the memory structures theorized by cognitive psychologists are separate but linked visual and verbal systems described by Allan Paivio's dual coding theory. Educational psychologists have used dual coding theory and cognitive load theory to explain how people learn from multimedia presentations. The spaced learning effect, a cognitive phenomenon strongly supported by psychological research, has broad applicability within education. For example, students have been found to perform better on a test of knowledge about a text passage when a second reading of the passage is delayed rather than immediate (see figure). Educational psychology research has confirmed the applicability to education of other findings from cognitive psychology, such as the benefits of using mnemonics for immediate and delayed retention of information. Problem solving, according to prominent cognitive psychologists, is fundamental to learning. It resides as an important research topic in educational psychology. A student is thought to interpret a problem by assigning it to a schema retrieved from long-term memory. A problem students run into while reading is called "activation." This is when the student's representations of the text are present during working memory. This causes the student to read through the material without absorbing the information and being able to retain it. When working memory is absent from the reader's representations of the working memory they experience something called "deactivation." When deactivation occurs, the student has an understanding of the material and is able to retain information. If deactivation occurs during the first reading, the reader does not need to undergo deactivation in the second reading. The reader will only need to reread to get a "gist" of the text to spark their memory. When the problem is assigned to the wrong schema, the student's attention is subsequently directed away from features of the problem that are inconsistent with the assigned schema. The critical step of finding a mapping between the problem and a pre-existing schema is often cited as supporting the centrality of analogical thinking to problem solving. Each person has an individual profile of characteristics, abilities and challenges that result from predisposition, learning and development. These manifest as individual differences in intelligence, creativity, cognitive style, motivation and the capacity to process information, communicate, and relate to others. The most prevalent disabilities found among school age children are attention deficit hyperactivity disorder (ADHD), learning disability, dyslexia, and speech disorder. Less common disabilities include intellectual disability, hearing impairment, cerebral palsy, epilepsy, and blindness. Although theories of intelligence have been discussed by philosophers since Plato, intelligence testing is an invention of educational psychology, and is coincident with the development of that discipline. Continuing debates about the nature of intelligence revolve on whether intelligence can be characterized by a single factor known as general intelligence, multiple factors (e.g., Gardner's theory of multiple intelligences), or whether it can be measured at all. In practice, standardized instruments such as the Stanford-Binet IQ test and the WISC are widely used in economically developed countries to identify children in need of individualized educational treatment. Children classified as gifted are often provided with accelerated or enriched programs. Children with identified deficits may be provided with enhanced education in specific skills such as phonological awareness. In addition to basic abilities, the individual's personality traits are also important, with people higher in conscientiousness and hope attaining superior academic achievements, even after controlling for intelligence and past performance. Developmental psychology, and especially the psychology of cognitive development, opens a special perspective for educational psychology. This is so because education and the psychology of cognitive development converge on a number of crucial assumptions. First, the psychology of cognitive development defines human cognitive competence at successive phases of development. Education aims to help students acquire knowledge and develop skills which are compatible with their understanding and problem-solving capabilities at different ages. Thus, knowing the students' level on a developmental sequence provides information on the kind and level of knowledge they can assimilate, which, in turn, can be used as a frame for organizing the subject matter to be taught at different school grades. This is the reason why Piaget's theory of cognitive development was so influential for education, especially mathematics and science education. In the same direction, the neo-Piagetian theories of cognitive development suggest that in addition to the concerns above, sequencing of concepts and skills in teaching must take account of the processing and working memory capacities that characterize successive age levels. Second, the psychology of cognitive development involves understanding how cognitive change takes place and recognizing the factors and processes which enable cognitive competence to develop. Education also capitalizes on cognitive change, because the construction of knowledge presupposes effective teaching methods that would move the student from a lower to a higher level of understanding. Mechanisms such as reflection on actual or mental actions vis-à-vis alternative solutions to problems, tagging new concepts or solutions to symbols that help one recall and mentally manipulate them are just a few examples of how mechanisms of cognitive development may be used to facilitate learning. Finally, the psychology of cognitive development is concerned with individual differences in the organization of cognitive processes and abilities, in their rate of change, and in their mechanisms of change. The principles underlying intra- and inter-individual differences could be educationally useful, because knowing how students differ in regard to the various dimensions of cognitive development, such as processing and representational capacity, self-understanding and self-regulation, and the various domains of understanding, such as mathematical, scientific, or verbal abilities, would enable the teacher to cater for the needs of the different students so that no one is left behind. Constructivism is a category of learning theory in which emphasis is placed on the agency and prior "knowing" and experience of the learner, and often on the social and cultural determinants of the learning process. Educational psychologists distinguish individual (or psychological) constructivism, identified with Piaget's theory of cognitive development, from social constructivism. The social constructivist paradigm views the context in which the learning occurs as central to the learning itself. It regards learning as a process of enculturation.  People learn by exposure to the culture of practitioners. They observe and practice the behavior of practitioners and 'pick up relevant jargon, imitate behavior, and gradually start to act in accordance with the norms of the practice'. So, a student learns to become a mathematician through exposure to mathematician using tools to solve mathematical problems. So in order to master a particular domain of knowledge it is not enough for students to be learn the concepts of the domain. They should be exposed to the use of the concepts in authentic activities by the practitioners of the domain. A dominant influence on the social constructivist paradigm is Lev Vygotsky's work on sociocultural learning, describing how interactions with adults, more capable peers, and cognitive tools are internalized to form mental constructs. "Zone of Proximal Development" (ZPD) is a term Vygotsky used to characterize an individual's mental development. He believed the task individuals can do on their own do not give a complete understanding of their mental development. He originally defined the ZPD as “the distance between the actual developmental level as determined by independent problem solving and the level of potential development as determined through problem solving under adult guidance or in collaboration with more capable peers.” He cited a famous example to make his case. Two children in school who originally can solve problems at an eight-year-old developmental level (that is, typical for children who were age 8), might be at different developmental levels. If each child received assistance from an adult, one was able to perform at a nine-year-old level and one was able to perform at a twelve-year-old level. He said “This difference between twelve and eight, or between nine and eight, is what we call "the zone of proximal development."” He further said that the ZPD “defines those functions that have not yet matured but are in the process of maturation, functions that will mature tomorrow but are currently in an embryonic state.” The zone is bracketed by the learner's current ability and the ability they can achieve with the aid of an instructor of some capacity. Vygotsky viewed the ZPD as a better way to explain the relation between children's learning and cognitive development. Prior to the ZPD, the relation between learning and development could be boiled down to the following three major positions: 1) Development always precedes learning (e.g., constructivism): children first need to meet a particular maturation level before learning can occur; 2) Learning and development cannot be separated, but instead occur simultaneously (e.g., behaviorism): essentially, learning is development; and 3) learning and development are separate, but interactive processes (e.g., gestaltism): one process always prepares the other process, and vice versa. Vygotsky rejected these three major theories because he believed that learning should always precede development in the ZPD. According to Vygotsky, through the assistance of a more knowledgeable other, a child is able to learn skills or aspects of a skill that go beyond the child's actual developmental or maturational level. The lower limit of ZPD is the level of skill reached by the child working independently (also referred to as the child's developmental level). The upper limit is the level of potential skill that the child is able to reach with the assistance of a more capable instructor. In this sense, the ZPD provides a prospective view of cognitive development, as opposed to a retrospective view that characterizes development in terms of a child's independent capabilities. The advancement through and attainment of the upper limit of the ZPD is limited by the instructional and scaffolding related capabilities of the more knowledgeable other (MKO). The MKO is typically assumed to be an older, more experienced teacher or parent, but often can be a learner's peer or someone their junior. The MKO need not even be a person, it can be a machine or book, or other source of visual and/or audio input. Elaborating on Vygotsky's theory, Jerome Bruner and other educational psychologists developed the important concept of instructional scaffolding, in which the social or information environment offers supports for learning that are gradually withdrawn as they become internalized. Jean Piaget was interested in how an organism adapts to its environment. Piaget hypothesized that infants are born with a schema operating at birth that he called "reflexes". Piaget identified four stages in cognitive development. The four stages are sensorimotor stage, pre-operational stage, concrete operational stage and formal operational stage. To understand the characteristics of learners in childhood, adolescence, adulthood, and old age, educational psychology develops and applies theories of human development. Often represented as stages through which people pass as they mature, developmental theories describe changes in mental abilities (cognition), social roles, moral reasoning, and beliefs about the nature of knowledge. For example, educational psychologists have conducted research on the instructional applicability of Jean Piaget's theory of development, according to which children mature through four stages of cognitive capability. Piaget hypothesized that children are not capable of abstract logical thought until they are older than about 11 years, and therefore younger children need to be taught using concrete objects and examples. Researchers have found that transitions, such as from concrete to abstract logical thought, do not occur at the same time in all domains. A child may be able to think abstractly about mathematics, but remain limited to concrete thought when reasoning about human relationships. Perhaps Piaget's most enduring contribution is his insight that people actively construct their understanding through a self-regulatory process. Piaget proposed a developmental theory of moral reasoning in which children progress from a naïve understanding of morality based on behavior and outcomes to a more advanced understanding based on intentions. Piaget's views of moral development were elaborated by Kohlberg into a stage theory of moral development. There is evidence that the moral reasoning described in stage theories is not sufficient to account for moral behavior. For example, other factors such as modeling (as described by the social cognitive theory of morality) are required to explain bullying. Rudolf Steiner's model of child development interrelates physical, emotional, cognitive, and moral development in developmental stages similar to those later described by Piaget. Developmental theories are sometimes presented not as shifts between qualitatively different stages, but as gradual increments on separate dimensions. Development of epistemological beliefs (beliefs about knowledge) have been described in terms of gradual changes in people's belief in: certainty and permanence of knowledge, fixedness of ability, and credibility of authorities such as teachers and experts. People develop more sophisticated beliefs about knowledge as they gain in education and maturity. Motivation is an internal state that activates, guides and sustains behavior. Motivation can have several impacting effects on how students learn and how they behave towards subject matter: Educational psychology research on motivation is concerned with the volition or will that students bring to a task, their level of interest and intrinsic motivation, the personally held goals that guide their behavior, and their belief about the causes of their success or failure. As intrinsic motivation deals with activities that act as their own rewards, extrinsic motivation deals with motivations that are brought on by consequences or punishments. A form of attribution theory developed by Bernard Weiner describes how students' beliefs about the causes of academic success or failure affect their emotions and motivations. For example, when students attribute failure to lack of ability, and ability is perceived as uncontrollable, they experience the emotions of shame and embarrassment and consequently decrease effort and show poorer performance. In contrast, when students attribute failure to lack of effort, and effort is perceived as controllable, they experience the emotion of guilt and consequently increase effort and show improved performance. The self-determination theory (SDT) was developed by psychologists Edward Deci and Richard Ryan. SDT focuses on the importance of intrinsic and extrinsic motivation in driving human behavior and posits inherent growth and development tendencies. It emphasizes the degree to which an individual's behavior is self-motivated and self-determined. When applied to the realm of education, the self-determination theory is concerned primarily with promoting in students an interest in learning, a value of education, and a confidence in their own capacities and attributes. Motivational theories also explain how learners' goals affect the way they engage with academic tasks. Those who have "mastery goals" strive to increase their ability and knowledge. Those who have "performance approach goals" strive for high grades and seek opportunities to demonstrate their abilities. Those who have "performance avoidance" goals are driven by fear of failure and avoid situations where their abilities are exposed. Research has found that mastery goals are associated with many positive outcomes such as persistence in the face of failure, preference for challenging tasks, creativity and intrinsic motivation. Performance avoidance goals are associated with negative outcomes such as poor concentration while studying, disorganized studying, less self-regulation, shallow information processing and test anxiety. Performance approach goals are associated with positive outcomes, and some negative outcomes such as an unwillingness to seek help and shallow information processing. Locus of control is a salient factor in the successful academic performance of students. During the 1970s and '80s, Cassandra B. Whyte did significant educational research studying locus of control as related to the academic achievement of students pursuing higher education coursework. Much of her educational research and publications focused upon the theories of Julian B. Rotter in regard to the importance of internal control and successful academic performance. Whyte reported that individuals who perceive and believe that their hard work may lead to more successful academic outcomes, instead of depending on luck or fate, persist and achieve academically at a higher level. Therefore, it is important to provide education and counseling in this regard. Instructional design, the systematic design of materials, activities and interactive environments for learning, is broadly informed by educational psychology theories and research. For example, in defining learning goals or objectives, instructional designers often use a taxonomy of educational objectives created by Benjamin Bloom and colleagues. Bloom also researched mastery learning, an instructional strategy in which learners only advance to a new learning objective after they have mastered its prerequisite objectives. Bloom discovered that a combination of mastery learning with one-to-one tutoring is highly effective, producing learning outcomes far exceeding those normally achieved in classroom instruction. Gagné, another psychologist, had earlier developed an influential method of task analysis in which a terminal learning goal is expanded into a hierarchy of learning objectives connected by prerequisite relationships. The following list of technological resources incorporate computer-aided instruction and intelligence for educational psychologists and their students: Technology is essential to the field of educational psychology, not only for the psychologist themselves as far as testing, organization, and resources, but also for students. Educational Psychologists whom reside in the K- 12 setting focus the majority of their time with Special Education students. It has been found that students with disabilities learning through technology such as iPad applications and videos are more engaged and motivated to learn in the classroom setting. Liu et al. explain that learning-based technology allows for students to be more focused, and learning is more efficient with learning technologies. The authors explain that learning technology also allows for students with social- emotional disabilities to participate in distance learning. Research on classroom management and pedagogy is conducted to guide teaching practice and form a foundation for teacher education programs. The goals of classroom management are to create an environment conducive to learning and to develop students' self-management skills. More specifically, classroom management strives to create positive teacher–student and peer relationships, manage student groups to sustain on-task behavior, and use counseling and other psychological methods to aid students who present persistent psycho-social problems. Introductory educational psychology is a commonly required area of study in most North American teacher education programs. When taught in that context, its content varies, but it typically emphasizes learning theories (especially cognitively oriented ones), issues about motivation, assessment of students' learning, and classroom management. A developing gives more detail about the educational psychology topics that are typically presented in preservice teacher education. In order to become an educational psychologist, students can complete an undergraduate degree in their choice. They then must go to graduate school to study education psychology, counseling psychology, and/ or school counseling. Most students today are also receiving their doctorate degrees in order to hold the "psychologist" title. Educational psychologists work in a variety of settings. Some work in university settings where they carry out research on the cognitive and social processes of human development, learning and education. Educational psychologists may also work as consultants in designing and creating educational materials, classroom programs and online courses. Educational psychologists who work in k–12 school settings (closely related are school psychologists in the US and Canada) are trained at the master's and doctoral levels. In addition to conducting assessments, school psychologists provide services such as academic and behavioral intervention, counseling, teacher consultation, and crisis intervention. However, school psychologists are generally more individual-oriented towards students. Many high school and colleges are increasingly offering educational psychology courses, with some colleges offering it as a general education requirement. Similarly, colleges offer students opportunities to obtain a PhD. in Educational Psychology. Within the UK, students must hold a degree that is accredited by the British Psychological Society (either undergraduate or at Masters level) before applying for a three-year doctoral course that involves further education, placement and a research thesis. Anticipated to grow by 18–26%, employment for psychologists in the United States is expected to grow faster than most occupations in 2014. One in four psychologists are employed in educational settings. In the United States, the median salary for psychologists in primary and secondary schools is US$58,360 as of May 2004. In recent decades, the participation of women as professional researchers in North American educational psychology has risen dramatically. Educational psychology, as much as any other field of psychology heavily relies on a balance of pure observation and quantitative methods in psychology. The study of education generally combines the studies of history, sociology, and ethics with theoretical approaches. Smeyers and Depaepe explain that historically, the study of education and child rearing have been associated with the interests of policymakers and practitioners within the educational field, however, the recent shift to sociology and psychology has opened the door for new findings in education as a social science. Now being its own academic discipline, educational psychology has proven to be helpful for social science researchers. Quantitative research is the backing to most observable phenomena in psychology. This involves observing, creating, and understanding a distribution of data based upon the studies subject matter. Researchers use particular variables to interpret their data distributions from their research and employ statistics as a way of creating data tables and analyzing their data. Psychology has moved from the "common sense" reputations initially posed by Thomas Reid to the methodology approach comparing independent and dependent variables through natural observation, experiments, or combinations of the two. Though results are still, with statistical methods, objectively true based upon significance variables or p- values.
https://en.wikipedia.org/wiki?curid=10332
EFTPOS Electronic funds transfer at point of sale (EFTPOS; ) is an electronic payment system involving electronic funds transfers based on the use of payment cards, such as debit or credit cards, at payment terminals located at points of sale. EFTPOS technology originated in the United States in 1981 and was adopted by other countries. In Australia and New Zealand, it is also the brand name of a specific system used for such payments; these systems are mainly country-specific and do not interconnect. Debit and credit cards are embossed plastic cards complying with ISO/IEC 7810 ID-1 standard. The cards have an embossed bank card number conforming with the ISO/IEC 7812 numbering standard. EFTPOS technology originated in the United States in 1981 and was rolled out in 1982. Initially, a number of nationwide systems were set up, such as "Interlink", which were limited to participating correspondent banking relationships, not being linked to each other. Consumers and merchants were slow to accept it, and there was minimal marketing. As a result, growth and market penetration of EFTPOS was minimal in the US up to the turn of the century. In a short time, other countries adopted the EFTPOS technology, but these systems too were limited to the national borders. Each country adopted various interbank co-operative models. In Australia, in 1984 Westpac was the first major Australian bank to implement an EFTPOS system, at BP petrol stations. The other major banks implemented EFTPOS systems during 1984, initially with petrol stations. The banks' existing debit and credit cards (but only allowed to access debit accounts) were used in the EFTPOS systems. In 1985, the State Bank of Victoria developed the capacity to host connect individual ATMS and helped create the ATM (Financial) Network. Banks started to link their EFTPOS systems to provide access for all customers across all EFTPOS devices. Cards issued by all banks could then be used at all EFTPOS terminals nationally, but debit cards issued in other countries could not. Prior to 1986, the Australian banks organized a widespread uniform credit card, called Bankcard, which had been in existence since 1974. There was a dispute between the banks whether Bankcard (or credit cards in general) should be permitted into the proposed EFTPOS system. At that time several banks were actively promoting MasterCard and Visa credit cards. Store cards and proprietary cards were shut out of the new system. In New Zealand, Bank of New Zealand started issuing EFTPOS debit cards in 1985 with the first merchant terminals being installed in petrol stations. In 1996, mobile EFTPOS arrived, with hotels in Kuala Lumpur installing systems in 1997 and the first example of a pizza delivery in Singapore accepting Visa card via cellular payment in 1998 (a collaboration between Signet, Visa, Citi Bank, and Dynamic Data Systems) beginning the rollout of mobile systems in Asia. By 2004 Cellular based Eftpos infrastructure had really taken off, and by 2010 Cellular Eftpos had become the standard for the Global market. Since 2002 the use of EFTPOS has grown significantly, and it has become the standard payment method, displacing the use of cash. Subsequently, networks facilitating the process of money transfer and payment settlement between the consumer and the merchant grew from a small number of nationwide systems to the majority of payment processing transactions. For EFTPOS, USA based systems allow the use of debit cards or credit cards. In Australia, debit and credit cards are the most common non-cash payment methods at “points of sale” (POS) or via ATMs. Not all merchants provide EFTPOS facilities, but those who wish to accept EFTPOS payments must enter an agreement with one of the seven merchant service providers, which rent an EFTPOS terminal to the merchant. The EFTPOS system in Australia is managed by Eftpos Payments Australia Ltd, which also sets the EFTPOS interchange fee. For credit cards to be accepted by a merchant a separate agreement must be entered into with each credit card company, each of which has its own flexible merchant fee rate. Eftpos machines for merchants are provided by larger banks and specialists such as Live eftpos. The clearing arrangements for EFTPOS are managed by Australian Payments Clearing Association (APCA). The system for ATM and EFTPOS interchanges is called Issuers and Acquirers Community (formerly Consumer Electronic Clearing System; CECS) also called CS3. CECS required authorisations from the Australian Competition and Consumer Commission (ACCC), which was obtained in 2001 and reaffirmed in 2009. ATM and EFTPOS clearances are the made under individual bilateral arrangements between the institutions involved. Australian financial institutions provide their customers with a plastic card, which can be used as a debit card or as an ATM card, and sometimes as a credit card. The card merely provides the means by which a customer's linked bank or other accounts can be accessed using an EFTPOS terminal or ATM. These cards can also be used on some vending machines and other automatic payment mechanisms, such as ticket vending machines. Each Australian bank has given a different name to its debit cards, such as: Some banks offer alternative debit card facilities to their customers using the Visa or MasterCard clearance system. For example, St George Bank offers a Visa Debit Card, as does the National Australia Bank. The main difference with regular debit cards is that these cards can be used outside Australia where the respective credit card is accepted. Those merchants that enter the EFTPOS payment system must accept debit cards issued by any Australian bank, and some also accept various credit cards and other cards. Some merchants set minimum transaction amounts for EFTPOS transactions, which can be different for debit and credit card transactions. Some merchants impose a surcharge on the use of EFTPOS. These can vary between merchants and on the type of card being used, and generally are not imposed on debit card transactions, and widely not on MasterCard and Visa credit card transactions. A feature of a debit card is that an EFTPOS transaction will only be accepted if there is an available credit balance in the bank cheque or savings account linked to the card. Australian debit cards normally cannot be used outside Australia. They can only be used outside Australia if they carry the MasterCard/Maestro/Cirrus or Visa/Plus or other similar logos, in which case the non-Australian transaction will be processed through those transaction systems. Similarly, non-Australian debit and credit cards can only be used at Australian EFTPOS terminals or ATMs if they have these logos or the MasterCard or Visa logos. Diners Club and/or American Express cards will be accepted only if the merchant has an agreement with those card companies, or increasingly if the merchant has modern alternative payment options available for those cards, such as through PayPal. The Discover Card is accepted in Australia as a Diners Club card. In addition, credit card companies issue prepaid cards which act like generic gift cards, which are anonymous and not linked to any bank accounts. These cards are accepted by merchants who accept credit cards and are processed through the EFTPOS terminal in the same way as credit cards. A number of merchants permit customers using a debit card to withdraw cash as part of the EFTPOS transaction. In Australia, this facility (known as debit card cashback in many other countries) is known as "cash out". For the merchant, cash out is a way of reducing their net cash takings, saving on banking of cash. There is no additional cost to the merchant in providing cash out because banks charge a merchant a debit card transaction fee per EFTPOS transaction, and not on the transaction value. Cash out is a facility provided by the merchant, and not the bank, so the merchant can limit or vary how much cash can be withdrawn at a time, or suspend the facility at any time. When available, cash out is convenient for the customer, who can bypass having to visit a bank branch or ATM. Cash out is also cheaper for the customer, since only one bank transaction is involved. For people in some remote areas, cash out may be the only way they can withdraw cash from their personal accounts. However, most merchants who provide the facility set a relatively low limit on cash out, generally $50, and some also charge for the service. Some merchants in Australia only allow cash out with the purchase of goods; other merchants allow cash out whether or not customers buy any goods. Cash out is not available in association with credit card sales because on credit card transactions the merchant is charged a percentage commission based on the transaction value, and also because cash withdrawals are treated differently from purchase transactions by the credit card company. (However, though inconsistent with a merchant's agreement with each credit card company, the merchant may treat a cash withdrawal as part of an ordinary credit card sale.) EFTPOS transactions involving a debit, credit or prepaid card are primarily authenticated via the entry of a personal identification number (PIN) at the point of sale. Historically, these transactions were authenticated by the merchant using the cardholder's signature, as signed on their receipt. However, merchants had become increasingly lax in enforcing this verification, resulting in an increase in fraud. Australian banks have since deployed chip and PIN technology using the global EMV card standard; as of 1 August 2014, Australian merchants no longer accept signatures on transactions by domestic customers at point of sale terminals. As a further security measure, if a user enters an incorrect PIN three times, the card may be locked out of EFTPOS and require reactivation over the phone or at a bank branch. In the case of an ATM, the card will not be returned, and the cardholder will need to visit the branch to retrieve the card, or request a new card to be issued. All debit cards now have a magnetic stripe on which is encoded the card's service codes, consisting of three-digit values. These codes are used to convey instructions to merchant terminals on how a card should be processed. The first digit indicates if a card can be used internationally or is valid for domestic use only. It is also used to signal if the card is chip-enabled. The second digit indicates if the transaction must be sent online for authorization always or if transactions that are below floor limit can take place without authorization. The third digit is used to indicate the preferred card verification method (e.g. PIN) and the environment where the card can be used (e.g. at point of sale only). Merchant terminals are required to recognize and act on service codes or send all transactions for online authorization. In the late 2000s, MasterCard and Visa introduced contactless smart debit cards under the brand names MasterCard PayPass and Visa payWave. These payments are made using either electronic payment networks separate from the regular EFTPOS payment networks, or newer EFTPOS with tap sensors, and is an alternative to the previous swipe or chip systems. These networks are operated by MasterCard and Visa, and not by the banks as is the EFTPOS network, through EFTPOS Payments Australia Limited (ePAL). These cards are based on EMV technology and contain a RFID chip and antenna loop embedded in the plastic of the card. To pay using this system, a customer passes the card within 4 cm of a reader at a merchant checkout. Using this method, for transactions under a specified limit, the customer does not need to authenticate his or her identity by PIN entry or signature, as on a regular EFTPOS machine. For transactions over the above limit, PIN verification is required. The facility is only available for cards branded with the MasterCard PayPass or Visa payWave logos, indicating that they have the system-permitted embedded chip. ANZ has launched ATM solution based on Visa payWave in 2015, where the consumer tap the card on a reader at the ATM and insert a PIN to finalize the cash withdrawals, and not all merchants offer the facility. Bank debit cards and other credit cards do not currently offer a contactless payment facility. ePAL is developing a contactless payment system for debit cards based on EMV technology as well as an extension of debit cards for use for on-line transactions, and a mobile payment system. Using contactless debit cards on tap-and-go terminals routes the transaction through the more expensive credit card system instead of the EFTPOS route, adding to the cost to the merchant, and ultimately the consumer. The name and logo for EFTPOS in Australia were originally owned by the Dr Shiyombo Makasa and were trade marks from 1986 until 1991. The ownership was for convenience and all the banks used the name and logo (commonly called "fat-E") on their cards and advertising. In 1991 dialup EFTPOS was conceived by Key Corp (John Wood) and deployment of dialup commenced in 1993. Until 1993, communications, connections and transactions between banks, ATM banks and EFTPOS devices where conducted via leased lines (a specific power assisted communication line that detects any attempt to tamper with it) but in 1993 mobile wireless EFTPOS was conceived by Dynamic Data Systems (H. Daniel Elbaum). In 1995 Dynamic Data Systems and the banking industry worked together to implement, certify and introduce protocols and standards for cellular networks, and by 1998 the use of mobile EFTPOS began to appear in Australia. In 2006 Commonwealth Bank and MasterCard ran a six-month trial of the contactless smart card system PayPass in Sydney and Wollongong, supplementing the traditional EFTPOS swipe or chip system. The system was rolled out across Australia in 2009; other systems being rolled out are Westpac Bank's MasterCard PayPass and Visa payWave branded cards. In April 2009, a company, “EFTPOS Payments Australia Ltd” (ePal) was formed to manage and promote the EFTPOS system in Australia. ePal regulation commenced in January 2011. The initial members of EFTPOS Payments Australia Ltd were: The current members of EFTPOS Payments Australia Ltd are: In Australia, store cards have been excluded from participation in the EFTPOS and ATM systems. Consequently, several larger store accounts have entered into co-branding arrangements with credit card networks for the store-based accounts to be widely accepted. This was the case with Coles (previously, Coles-Myer) which co-branded with MasterCard, and David Jones which co-branded with American Express. Woolworths organized its credit card called Everyday Rewards (now Woolworths Money) which initially was partnered with credit provider HSBC Bank, but changed on 26 October 2014 to Macquarie Bank. As of June 2018, there were 961,247 EFTPOS terminals in Australia and 30,940 ATMs. Of the terminals, over 60,000 offered cash withdrawals. In 2010, 183 million transactions, worth A$12 billion, were made using Australian EFTPOS terminals per month. In 2011, these figures increased to 750,000 terminals, with 325,000 individual businesses, processing over 2 billion transactions with combined value of approximately $131 billion for the year. The EFT network in Australia is made up of seven proprietary networks in which peers have interchange agreements, making an effective single network. A merchant who wishes to accept EFTPOS payments must enter an agreement with one of the seven merchant service providers, which rent the terminal to the merchant. All the merchant's EFTPOS transactions are processed through one of these gateways. Some of these peers are: Other organisations may have peering agreements with the one or more of the central peers. The network uses the AS 2805 protocol, which is closely related to ISO 8583. EFTPOS is highly popular in New Zealand. The system is operated by two providers, Paymark Limited (formerly Electronic Transaction Services Limited) which processes 75% of all electronic transactions in New Zealand, and EFTPOS New Zealand. Although the term eftpos is popularly used to describe the system, EFTPOS is a trademark of EFTPOS New Zealand the smaller of the two providers. Both providers run an interconnected financial network that allows the processing of not only of debit cards at point of sale terminals but also credit cards and charge cards. The Bank of New Zealand introduced EFTPOS to New Zealand in 1985 through a pilot scheme with petrol stations. In 1989 the system was officially launched and two providers owned by the major banks now run the system. The larger of the two providers, Paymark Limited (formerly Electronic Transaction Services Limited) is owned equally by ASB Bank, Westpac, Bank of New Zealand and ANZ Bank New Zealand (formerly ANZ National Bank). The second is operated by EFTPOS New Zealand which is fully owned by VeriFone Systems, following its sale by ANZ New Zealand in December 2012. 1995 was the first deployment of cellular Eftpos in NZ, by Dynamic Data Systems. During July 2006 the five billionth EFTPOS payment was processed, and at the start of 2012 the 10 billionth transaction was processed. EFTPOS is highly popular in New Zealand, and being used for about 60% of all retail transactions. In 2009, there were 200 EFTPOS transactions per person. Paymark process over 900 million transactions (worth over NZ$48 billion) yearly. More than 75,000 merchants and over 110,000 EFTPOS terminals are connected to Paymark.
https://en.wikipedia.org/wiki?curid=10333
Extermination camp Nazi Germany used six extermination camps (), also called death camps, or killing centers, in Eastern Europe during the Holocaust in World War II to systematically murder about 2.7 million Jews during the Holocaust. Others were murdered at the death camps as well, including Romani people during the Romani genocide. The victims of death camps were primarily killed by gassing, either in permanent installations constructed for this specific purpose, or by means of gas vans. The six extermination camps were Chelmno, Belzec, Sobibor, Treblinka, Majdanek and Auschwitz-Birkenau. Auschwitz and Majdanek death camps also used extreme work under starvation conditions in order to kill their prisoners. The idea of mass extermination with the use of stationary facilities, to which the victims were taken by train, was the result of earlier Nazi experimentation with chemically manufactured poison gas during the secretive Aktion T4 euthanasia programme against hospital patients with mental and physical disabilities. The technology was adapted, expanded, and applied in wartime to unsuspecting victims of many ethnic and national groups; the Jews were the primary target, accounting for over 90 percent of the extermination camp death toll. The genocide of the Jews of Europe was the Third Reich's "Final Solution to the Jewish question". After the invasion of Poland in September 1939, the secret Aktion T4 euthanasia programme – the systematic murder of German, Austrian and Polish hospital patients with mental or physical disabilities – was initiated by the "SS" in order to eliminate "life unworthy of life" (), a Nazi designation for people who had no right to life. In 1941, the experience gained in the secretive killing of these hospital patients led to the creation of extermination camps for the implementation of the Final Solution. By then, the Jews were already confined to new ghettos and interned in Nazi concentration camps along with other targeted groups, including Roma, and the Soviet POWs. The Nazi "Endlösung der Judenfrage" (The Final Solution of the Jewish Question), based on the systematic killing of Europe's Jews by gassing, began during Operation Reinhard, after the onset of the Nazi-Soviet war of June 1941. The adoption of the gassing technology by Nazi Germany was preceded by a wave of hands-on killings carried out by the SS "Einsatzgruppen", who followed the Wehrmacht army during Operation Barbarossa on the Eastern Front. The camps designed specifically for the mass gassings of Jews were established in the months following the Wannsee Conference chaired by Reinhard Heydrich in January 1942 in which the principle was made clear that the Jews of Europe were to be exterminated. Responsibility for the logistics were to be executed by the programme administrator, Adolf Eichmann. On 13 October 1941, the SS and Police Leader Odilo Globocnik stationing in Lublin received an oral order from "Reichsführer-SS" Heinrich Himmler – anticipating the fall of Moscow – to start immediate construction work on the killing centre at Bełżec in the General Government territory of occupied Poland. Notably, the order preceded the Wannsee Conference by three months, but the gassings at Kulmhof north of Łódź using gas vans began already in December, under "Sturmbannführer" Herbert Lange. The camp at Bełżec was operational by March 1942, with leadership brought in from Germany under the guise of Organisation Todt (OT). By mid-1942, two more death camps had been built on Polish lands for Operation Reinhard: Sobibór (ready in May 1942) under the command of "Hauptsturmführer" Franz Stangl, and Treblinka (operational by July 1942) under "Obersturmführer" Irmfried Eberl from T4, the only doctor to have served in such a capacity. Auschwitz concentration camp was fitted with brand new gassing bunkers in March 1942. Majdanek had them built in September. The Nazis distinguished between extermination and concentration camps. The terms "extermination camp" ("Vernichtungslager") and "death camp" ("Todeslager") were interchangeable in the Nazi system, each referring to camps whose primary function was genocide. Six camps meet this definition, though extermination of people happened at every sort of concentration camp or transit camp; the use of the term extermination camp with its exclusive purpose is carried over from Nazi terminology. The six camps were Chelmno extermination camp, Belzec extermination camp, Sobibor extermination camp, Treblinka extermination camp, Majdanek extermination camp and Auschwitz extermination camp (also called Auschwitz-Birkenau). "Todeslagers" were designed specifically for the systematic killing of people delivered en masse by the Holocaust trains. The executioners did not expect the prisoners to survive more than a few hours beyond arrival at Belzec, Sobibór, and Treblinka. The Reinhard extermination camps were under Globocnik's direct command; each of them was run by 20 to 35 men from the "SS-Totenkopfverbände" branch of the "Schutzstaffel", augmented by about one hundred Trawnikis – auxiliaries mostly from Soviet Ukraine, and up to one thousand "Sonderkommando" slave labourers each. The Jewish men, women and children were delivered from the ghettos for "special treatment" in an atmosphere of terror by uniformed police battalions from both Orpo and Schupo. Death camps differed from concentration camps located in Germany proper, such as Bergen-Belsen, Oranienburg, Ravensbrück, and Sachsenhausen, which were prison camps set up prior to World War II for people defined as 'undesirable'. From March 1936, all Nazi concentration camps were managed by the "SS-Totenkopfverbände" (the Skull Units, SS-TV), who operated extermination camps from 1941 as well. An SS anatomist, Dr Johann Kremer, after witnessing the gassing of victims at Birkenau, wrote in his diary on 2 September 1942: "Dante's Inferno seems to me almost a comedy compared to this. They don't call Auschwitz the camp of annihilation for nothing!" The distinction was evident during the Nuremberg trials, when Dieter Wisliceny (a deputy to Adolf Eichmann) was asked to name the "extermination" camps, and he identified Auschwitz and Majdanek as such. Then, when asked, "How do you classify the camps Mauthausen, Dachau, and Buchenwald?", he replied, "They were normal concentration camps, from the point of view of the department of Eichmann." Murders were not limited to these camps. Sites for the “Holocaust by Bullets” are marked on the map of The Holocaust in Occupied Poland by white skulls (without the black background), where people were lined up next to a ravine and shot by soldiers with rifles. Sites included Bronna Góra, Ponary and others. Irrespective of round-ups for extermination camps, the Nazis abducted millions of foreigners for slave labour in other types of camps, which provided perfect cover for the extermination programme. Prisoners represented about a quarter of the total workforce of the Reich, with mortality rates exceeding 75 percent due to starvation, disease, exhaustion, executions, and physical brutality. In the early years of World War II, the Jews were primarily sent to forced labour camps and ghettoised, but from 1942 onward they were deported to the extermination camps under the guise of "resettlement". For political and logistical reasons, the most infamous Nazi German killing factories were built in occupied Poland, where most of the intended victims lived; Poland had the greatest Jewish population in Nazi-controlled Europe. On top of that, the new death camps outside the prewar borders of the Third Reich proper could be kept secret from the German civil populace. During the initial phase of the Final Solution, gas vans producing poisonous exhaust fumes were developed in the occupied Soviet Union (USSR) and at the Chełmno extermination camp in occupied Poland, before being used elsewhere. The killing method was based on experience gained by the SS during the secretive "Aktion T4" programme of involuntary euthanasia. There were two types of death chambers operating during the Holocaust. Unlike at Auschwitz, where the cyanide-based Zyklon-B was used to exterminate trainloads of prisoners under the guise of "relocation", the camps at Treblinka, Bełżec, and Sobibór, built during Operation Reinhard (October 1941 – November 1943), used lethal exhaust fumes produced by large internal combustion engines. The three killing centres of "Einsatz Reinhard" were constructed predominantly for the extermination of Poland's Jews trapped in the Nazi ghettos. At first, the victim's bodies were buried with the use of crawler excavators, but they were later exhumed and incinerated in open-air pyres to hide the evidence of genocide in what became known as Sonderaktion 1005. The six camps considered to be purely for extermination were Chelmno extermination camp, Belzec extermination camp, Sobibor extermination camp, Treblinka extermination camp, Majdanek extermination camp and Auschwitz extermination camp (also called Auschwitz-Birkenau). Whereas the Auschwitz II (Auschwitz–Birkenau) and Majdanek camps were parts of a labor camp complex, the Chełmno and Operation Reinhard death camps (that is, Bełżec, Sobibór, and Treblinka) were built exclusively for the rapid extermination of entire communities of people (primarily Jews) within hours of their arrival. All were constructed near branch lines that linked to the Polish railway system, with staff members transferring between locations. These camps had almost identical design: they were several hundred metres in length and width, and were equipped with only minimal staff housing and support installations not meant for the victims crammed into the railway transports. The Nazis deceived the victims upon their arrival, telling them that they were at a temporary transit stop, and would soon continue to German "Arbeitslagers" (work camps) farther to the east. Selected able-bodied prisoners delivered to the death camps were not immediately killed, but instead were pressed into labor units called "Sonderkommandos" to help with the extermination process by removing corpses from the gas chambers and burning them. At the camps of Operation Reinhard, including Bełżec, Sobibór, and Treblinka, trainloads of prisoners were destined for immediate death in gas chambers designed exclusively for that purpose. The mass killing facilities were developed at about the same time inside the Auschwitz II-Birkenau subcamp of a forced labour complex, and at the Majdanek concentration camp. In most other camps prisoners were selected for slave labor first; they were kept alive on starvation rations and made available to work as required. Auschwitz, Majdanek, and Jasenovac were retrofitted with Zyklon-B gas chambers and crematoria buildings as the time went on, remaining operational until war's end in 1945. The Maly Trostenets extermination camp in the USSR initially operated as a prison camp. It became an extermination camp later in the war with victims undergoing mass shootings. This was supplemented with gassings in a van by exhaust fumes from October 1943. The Sajmište concentration camp operated by the Nazis in Yugoslavia had a gas van stationed for use from March to June 1942. Once the industrial killings were completed, the van was returned to Berlin. After a refit the van was then sent to Maly Trostinets for use at the camp there. The Janowska concentration camp near Lwow (now Lviv) in occupied eastern Poland implemented a selection process. Some prisoners were assigned to work before death. Others were either transported to Bełżec or victims of mass shootings on two slopes in the Piaski sand-hills behind the camp. Heinrich Himmler visited the outskirts of Minsk in 1941 to witness a mass shooting. He was told by the commanding officer there that the shootings were proving psychologically damaging to those being asked to pull the triggers. Thus Himmler knew another method of mass killing was required. After the war, the diary of the Auschwitz Commandant, Rudolf Höss, revealed that psychologically "unable to endure wading through blood any longer", many "Einsatzkommandos" – the killers – either went mad or killed themselves. The Nazis had first used gassing with carbon monoxide cylinders to kill 70,000 disabled people in Germany in what they called a 'euthanasia programme' to disguise that mass murder was taking place. Despite the lethal effects of carbon monoxide, this was seen as unsuitable for use in the East due to the cost of transporting the carbon monoxide in cylinders. Each extermination camp operated differently, yet each had designs for quick and efficient industrialized killing. While Höss was away on an official journey in late August 1941 his deputy, Karl Fritzsch, tested out an idea. At Auschwitz clothes infested with lice were treated with crystallised prussic acid. The crystals were made to order by the IG Farben chemicals company for which the brand name was Zyklon-B. Once released from their container, Zyklon-B crystals in the air released a lethal cyanide gas. Fritzsch tried out the effect of Zyklon B on Soviet POWs, who were locked up in cells in the basement of the bunker for this experiment. Höss on his return was briefed and impressed with the results and this became the camp strategy for extermination as it was also to be at Majdanek. Besides gassing, the camp guards continued killing prisoners via mass shooting, starvation, torture, etc. SS "Obersturmführer" Kurt Gerstein, of the Institute for Hygiene of the Waffen-SS, told a Swedish diplomat during the war of life in a death camp. He recounted that, on 19 August 1942, he arrived at Belzec extermination camp (which was equipped with carbon monoxide gas chambers) and was shown the unloading of 45 train cars filled with 6,700 Jews, many already dead. The rest were marched naked to the gas chambers, where: Auschwitz Camp Commandant Rudolf Höss reported that the first time Zyklon B pellets were used on the Jews, many suspected they were to be killed – despite having been deceived into believing they were to be deloused and then returned to the camp. As a result, the Nazis identified and isolated "difficult individuals" who might alert the prisoners, and removed them from the mass – lest they incite revolt among the deceived majority of prisoners en route to the gas chambers. The "difficult" prisoners were led to a site out of view to be killed off discreetly. A prisoner "Sonderkommando" (Special Detachment) effected in the processes of extermination; they encouraged the Jews to undress without a hint of what was about to happen. They accompanied them into the gas chambers outfitted to appear as shower rooms (with nonworking water nozzles, and tile walls); and remained with the victims until just before the chamber door closed. To psychologically maintain the "calming effect" of the delousing deception, an SS man stood at the door until the end. The "Sonderkommando" talked to the victims about life in the camp to pacify the suspicious ones, and hurried them inside; to that effect, they also assisted the aged and the very young in undressing. To further persuade the prisoners that nothing harmful was happening, the "Sonderkommando" deceived them with small talk about friends or relations who had arrived in earlier transports. Many young mothers hid their infants beneath their piled clothes fearing that the delousing "disinfectant" might harm them. Camp Commandant Höss reported that the "men of the Special Detachment were particularly on the look-out for this", and encouraged the women to take their children into the "shower room". Likewise, the "Sonderkommando" comforted older children who might cry "because of the strangeness of being undressed in this fashion". Yet, not every prisoner was deceived by such psychological tactics; Commandant Höss spoke of Jews "who either guessed, or knew, what awaited them, nevertheless ... [they] found the courage to joke with the children, to encourage them, despite the mortal terror visible in their own eyes". Some women would suddenly "give the most terrible shrieks while undressing, or tear their hair, or scream like maniacs"; the "Sonderkommando" immediately took them away for execution by shooting. In such circumstances, others, meaning to save themselves at the gas chamber's threshold, betrayed the identities and "revealed the addresses of those members of their race still in hiding". Once the door of the filled gas chamber was sealed, pellets of Zyklon B were dropped through special holes in the roof. Regulations required that the Camp Commandant supervise the preparations, the gassing (through a peephole), and the aftermath looting of the corpses. Commandant Höss reported that the gassed victims "showed no signs of convulsion"; the Auschwitz camp physicians attributed that to the "paralyzing effect on the lungs" of the Zyklon-B gas, which killed "before" the victim began suffering convulsions. As a matter of political training, some high-ranked Nazi Party leaders and SS officers were sent to Auschwitz–Birkenau to witness the gassings; Höss reported that, "all were deeply impressed by what they saw ... [yet some] ... who had previously spoken most loudly, about the necessity for this extermination, fell silent once they had actually seen the 'final solution of the Jewish problem'." As the Auschwitz Camp Commandant Rudolf Höss justified the extermination by explaining the need for "the iron determination with which we must carry out Hitler's orders"; yet saw that even "[Adolf] Eichmann, who certainly [was] tough enough, had no wish to change places with me”. After the gassings, the "Sonderkommando" removed the corpses from the gas chambers, then extracted any gold teeth. Initially, the victims were buried in mass graves, but were later cremated during "Sonderaktion 1005" in all camps of Operation Reinhard. The "Sonderkommando" were responsible for burning the corpses in the pits, stoking the fires, draining surplus body fat and turning over the "mountain of burning corpses ... so that the draught might fan the flames" wrote Commandant Höss in his memoir while in the Polish custody. He was impressed by the diligence of prisoners from the so-called Special Detachment who carried out their duties despite their being well aware that they, too, would meet exactly the same fate in the end. At the Lazaret killing station they held the sick so they would never see the gun while being shot. They did it "in such a matter-of-course manner that they might, themselves, have been the exterminators" wrote Höss. He further said that the men ate and smoked "even when engaged in the grisly job of burning corpses which had been lying for some time in mass graves." They occasionally encountered the corpse of a relative, or saw them entering the gas chambers. According to Höss they were obviously shaken by this but "it never led to any incident." He mentioned the case of a "Sonderkommando" who found the body of his wife, yet continued to drag corpses along "as though nothing had happened." At Auschwitz, the corpses were incinerated in crematoria and the ashes either buried, scattered, or dumped in the river. At Sobibór, Treblinka, Bełżec, and Chełmno, the corpses were incinerated on pyres. The efficiency of industrialised killing at Auschwitz-Birkenau led to the construction of three buildings with crematoria designed by specialists from the firm J.A. Topf & Söhne. They burned bodies 24 hours a day, and yet the death rate was at times so high that corpses also needed to be burned in open-air pits. The estimated total number of people who were murdered in the six Nazi extermination camps is 2.7 million, according to the United States Holocaust Memorial Museum. The Nazis attempted to either partially or completely dismantle the extermination camps in order to hide any evidence that people had been murdered there. This was an attempt to conceal not only the extermination process but also the buried remains. As a result of the secretive Sonderaktion 1005, the camps were dismantled by commandos of condemned prisoners, their records were destroyed, and the mass graves were dug up. Some extermination camps that remained uncleared of evidence were liberated by Soviet troops, who followed different standards of documentation and openness than the Western allies did. Nonetheless Majdanek was captured nearly intact due to the rapid advance of the Soviet Red Army during Operation Bagration. In the post-war period the government of the People's Republic of Poland created monuments at the extermination camp sites. These early monuments mentioned no ethnic, religious, or national particulars of the Nazi victims. The extermination camps sites have been accessible to everyone in recent decades. They are popular destinations for visitors from all over the world, especially the most infamous Nazi death camp, Auschwitz near the town of Oświęcim. In the early 1990s, the Jewish Holocaust organisations debated with the Polish Catholic groups about "What religious symbols of martyrdom are appropriate as memorials in a Nazi death camp such as Auschwitz?" The Jews opposed the placement of Christian memorials such as the Auschwitz cross near Auschwitz I where mostly Poles were killed. The Jewish victims of the Holocaust were mostly killed at Auschwitz II Birkenau. The March of the Living is organized in Poland annually since 1988. Marchers come from countries as diverse as Estonia, New Zealand, Panama, and Turkey. Holocaust deniers or negationists are people and organizations who assert that the Holocaust did not occur, or that it did not occur in the historically recognized manner and extent. Holocaust deniers claim that the extermination camps were actually transit camps from which Jews were deported farther east. However, these theories are disproven by surviving German documents, which show that Jews were sent to the camps to be killed. Extermination camp research is difficult because of extensive attempts by the SS and Nazi regime to conceal the existence of the extermination camps. The existence of the extermination camps is firmly established by testimonies of camp survivors and Final Solution perpetrators, material evidence (the remaining camps, etc.), Nazi photographs and films of the killings, and camp administration records. In 2017 a Körber Foundation survey found that 40 percent of 14-year-olds in Germany did not know what Auschwitz was. A 2018 survey organized in the United States by the Claims Conference, United States Holocaust Memorial Museum, and others found that 66% of the American millennials who were surveyed (and 41% of all U.S. adults) did not know what Auschwitz was. In 2019, a survey of 1,100 Canadians found that 49 percent of them could not name any of the Nazi camps which were located in German-occupied Europe.
https://en.wikipedia.org/wiki?curid=10335
Excommunication Excommunication is an institutional act of religious censure used to end or at least regulate the communion of a member of a congregation with other members of the religious institution who are in normal communion with each other. The purpose of the institutional act is to deprive, suspend, or limit membership in a religious community or to restrict certain rights within it, in particular, those of being in communion with other members of the congregation, and of receiving the sacraments. The term is often historically used to refer specifically to excommunications from the Catholic Church, but it is also used more generally to refer to similar types of institutional religious exclusionary practices and shunning among other religious groups. For instance, many Protestant denominations, such as the Lutheran Churches, have similar practices of excusing congregants from church communities, while Jehovah's Witnesses, as well as the Churches of Christ, use the term "disfellowship" to refer to their form of excommunication. The Amish have also been known to excommunicate members that were either seen or known for breaking rules, or questioning the church. The word "excommunication" means putting a specific individual or group out of communion. In some denominations, excommunication includes spiritual condemnation of the member or group. Excommunication may involve banishment, shunning, and shaming, depending on the group, the offense that caused excommunication, or the rules or norms of the religious community. The grave act is often revoked in response to sincere penance, which may be manifested through public recantation, sometimes through the Sacrament of Confession, piety or through mortification of the flesh. Within the Catholic Church, there are differences between the discipline of the majority Latin Church regarding excommunication and that of the Eastern Catholic Churches. Excommunication can be either "latae sententiae" (automatic, incurred at the moment of committing the offense for which canon law imposes that penalty) or "ferendae sententiae" (incurred only when imposed by a legitimate superior or declared as the sentence of an ecclesiastical court). According to Bishop Thomas J. Paprocki, "excommunication does not expel the person from the Catholic Church, but simply forbids the excommunicated person from engaging in certain activities..." These activities are listed in Canon 1331 §1, and prohibit the individual from any ministerial participation in celebrating the sacrifice of the Eucharist or any other ceremonies of worship; celebrating or receiving the sacraments; or exercising any ecclesiastical offices, ministries, or functions. Under current Catholic canon law, excommunicates remain bound by ecclesiastical obligations such as attending Mass, even though they are barred from receiving the Eucharist and from taking an active part in the liturgy (reading, bringing the offerings, etc.). "Excommunicates lose rights, such as the right to the sacraments, but they are still bound to the obligations of the law; their rights are restored when they are reconciled through the remission of the penalty." They are urged to retain a relationship with the Church, as the goal is to encourage them to repent and return to active participation in its life. These are the only effects for those who have incurred a "latae sententiae" excommunication. For instance, a priest may not refuse Communion publicly to those who are under an automatic excommunication, as long as it has not been officially declared to have been incurred by them, even if the priest knows that they have incurred it. On the other hand, if the priest knows that excommunication has been imposed on someone or that an automatic excommunication has been declared (and is no longer merely an undeclared automatic excommunication), he is forbidden to administer Holy Communion to that person. (see canon 915). In the Catholic Church, excommunication is normally resolved by a declaration of repentance, profession of the Creed (if the offense involved heresy) and an Act of Faith, or renewal of obedience (if that was a relevant part of the offending act, i.e., an act of schism) by the excommunicated person and the lifting of the censure (absolution) by a priest or bishop empowered to do this. "The absolution can be in the internal (private) forum only, or also in the external (public) forum, depending on whether scandal would be given if a person were privately absolved and yet publicly considered unrepentant." Since excommunication excludes from reception of the sacraments, absolution from excommunication is required before absolution can be given from the sin that led to the censure. In many cases, the whole process takes place on a single occasion in the privacy of the confessional. For some more serious wrongdoings, absolution from excommunication is reserved to a bishop, another ordinary, or even the pope. These can delegate a priest to act on their behalf. Interdict is a censure similar to excommunication. It too excludes from ministerial functions in public worship and from reception of the sacraments, but not from the exercise of governance. In the Eastern Catholic Churches, excommunication is imposed only by decree, never incurred automatically by "latae sententiae" excommunication. A distinction is made between minor and major excommunication. Those on whom minor excommunication has been imposed are excluded from receiving the Eucharist and can also be excluded from participating in the Divine Liturgy. They can even be excluded from entering a church when divine worship is being celebrated there. The decree of excommunication must indicate the precise effect of the excommunication and, if required, its duration. Those under major excommunication are in addition forbidden to receive not only the Eucharist but also the other sacraments, to administer sacraments or sacramentals, to exercise any ecclesiastical offices, ministries, or functions whatsoever, and any such exercise by them is null and void. They are to be removed from participation in the Divine Liturgy and any public celebrations of divine worship. They are forbidden to make use of any privileges granted to them and cannot be given any dignity, office, ministry, or function in the Church, they cannot receive any pension or emoluments associated with these dignities etc., and they are deprived of the right to vote or to be elected. Minor excommunication is roughly equivalent to the interdict in Western law. The excommunicable offenses can be distinguished Persons belonging to an Eastern Catholic Church are never subject to a "latae sententiae" punishment; this is therefore not explicitly mentioned in the lists below. A person is "latae sententiae" excommunicated or, if an Eastern Catholic, "ferendae sententia" if they: A person may be "ferendae sententiae" excommunicated if he According to the Code of Canon Law of 1917, the excommunications reserved to the Apostolic See were grouped in three categories, those reserved 1. simply, 2. in a special manner, 3. in a most special manner (each solvable by the Pope and by those priests the Pope had delegated the faculty to absolve for precisely that degree); and below the excommunications reserved to the bishop (which is now principally true of every excommunication), there was yet a category of excommunications reserved to no one (i. e., that could be solved by any confessor). The excommunications for desecration of the Blessed Sacrament, physical violence against the Pope, attempted absolution of an accomplice in a sin against the sixth Commandment, and breaking the Seal of the Confessional (no. 1-4 from the "latae sententiae" offences listed above) were reserved to the Apostolic See in most special manner. The excommunications for apostasy, heresy or schism were reserved to the Apostolic see in special manner, though they could be solved by the bishop (though not the general vicar) in his stead (can. 2314 § 2). The possible excommunication of someone not the Confessor who disclosed something under the Seal of the Confessional was reserved to no one; the excommunication for unlawful episcopal consecrations did not then exist (but there was a "latae sententiae" suspension), as neither did the possible excommunication (and certain suspension) of a priest who does have faculties but absolves a penitent he knows to be unrepenting. The other excommunications still in existence were reserved to the bishop as they are now. The following further acts were excommunicable offenses In the Eastern Orthodox churches, excommunication is the exclusion of a member from the Eucharist. It is not expulsion from the churches. This can happen for such reasons as not having confessed within that year; excommunication can also be imposed as part of a penitential period. It is generally done with the goal of restoring the member to full communion. Before an excommunication of significant duration is imposed, the bishop is usually consulted. The Orthodox churches do have a means of expulsion, by pronouncing anathema, but this is reserved only for acts of serious and unrepentant heresy. As an example of this, the Second Council of Constantinople in 553, in its eleventh capitula, declared: "If anyone does not anathematize Arius, Eunomius, Macedonius, Apollinarius, Nestorius, Eutyches and Origen, as well as their heretical books, and also all other heretics who have already been condemned and anathematized by the holy, catholic and apostolic church and by the four holy synods which have already been mentioned, and also all those who have thought or now think in the same way as the aforesaid heretics and who persist in their error even to death: let him be anathema." Although Lutheranism technically has an excommunication process, some denominations and congregations do not use it. In the Smalcald Articles Luther differentiates between the "great" and "small" excommunication. The "small" excommunication is simply barring an individual from the Lord's Supper and "other fellowship in the church." While the "great" excommunication excluded a person from both the church and political communities which he considered to be outside the authority of the church and only for civil leaders. A modern Lutheran practice is laid out in the Lutheran Church-Missouri Synod's 1986 explanation to the Small Catechism, defined beginning at Questions No. 277-284, in "The Office of Keys." They endeavor to follow the process that Jesus laid out in the 18th chapter of the Gospel of Matthew. According to the explanation, excommunication requires: Many Lutheran denominations operate under the premise that the entire congregation (as opposed to the pastor alone) must take appropriate steps for excommunication, and there are not always precise rules, to the point where individual congregations often set out rules for excommunicating laymen (as opposed to clergy). For example, churches may sometimes require that a vote must be taken at Sunday services; some congregations require that this vote be unanimous. In the Church of Sweden and the Church of Denmark, excommunicated individuals are turned out from their parish in front of their congregation. They are not forbidden, however, to attend church and participate in other acts of devotion, although they are to sit in a place appointed by the priest (which was at a distance from others). The Lutheran process, though rarely used, has created unusual situations in recent years due to its somewhat democratic excommunication process. One example was an effort to get serial killer Dennis Rader excommunicated from his denomination (the Evangelical Lutheran Church in America) by individuals who tried to "lobby" Rader's fellow church members into voting for his excommunication. The Church of England does not have any specific canons regarding how or why a member can be excommunicated, although it has a canon according to which ecclesiastical burial may be refused to someone "declared excommunicate for some grievous and notorious crime and no man to testify to his repentance". The punishment of imprisonment for being excommunicated from the Church of England was removed from English law in 1963. The ECUSA is in the Anglican Communion, and shares many canons with the Church of England which would determine its policy on excommunication. In the Reformed Churches, excommunication has generally been seen as the culmination of church discipline, which is one of the three marks of the Church. The Westminster Confession of Faith sees it as the third step after "admonition" and "suspension from the sacrament of the Lord's Supper for a season." Yet, John Calvin argues in his "Institutes of the Christian Religion" that church censures do not "consign those who are excommunicated to perpetual ruin and damnation," but are designed to induce repentance, reconciliation and restoration to communion. Calvin notes, "though ecclesiastical discipline does not allow us to be on familiar and intimate terms with excommunicated persons, still we ought to strive by all possible means to bring them to a better mind, and recover them to the fellowship and unity of the Church." At least one modern Reformed theologian argues that excommunication is not the final step in the disciplinary process. Jay E. Adams argues that in excommunication, the offender is still seen as a brother, but in the final step they become "as the heathen and tax collector" (Matthew 18:17). Adams writes, "Nowhere in the Bible is excommunication (removal from the fellowship of the Lord's Table, according to Adams) equated with what happens in step 5; rather, step 5 is called "removing from the midst, handing over to Satan," and the like." Former Princeton president and theologian, Jonathan Edwards, addresses the notion of excommunication as "removal from the fellowship of the Lord's Table" in his treatise entitled "The Nature and End of Excommunication". Edwards argues that "Particularly, we are forbidden such a degree of associating ourselves with (excommunicants), as there is in making them our guests at our tables, or in being their guests at their tables; as is manifest in the text, where we are commanded to have no company with them, no not to eat". Edwards insists, "That this respects not eating with them at the Lord's supper, but a common eating, is evident by the words, that the eating here forbidden, is one of the lowest degrees of keeping company, which are forbidden. Keep no company with such a one, saith the apostle, no not to eat — as much as to say, no not in so low a degree as to eat with him. But eating with him at the Lord's supper, is the very highest degree of visible Christian communion. Who can suppose that the apostle meant this: Take heed and have no company with a man, no not so much as in the highest degree of communion that you can have? Besides, the apostle mentions this eating as a way of keeping company which, however, they might hold with the heathen. He tells them, not to keep company with fornicators. Then he informs them, he means not with fornicators of this world, that is, the heathens; but, saith he, "if any man that is called a brother be a fornicator, etc. with such a one keep no company, no not to eat." This makes it most apparent, that the apostle doth not mean eating at the Lord's table; for so, they might not keep company with the heathens, any more than with an excommunicated person." In the Methodist Episcopal Church, individuals were able to be excommunicated following "trial before a jury of his peers, and after having had the privilege of an appeal to a higher court." Nevertheless, an excommunication could be lifted after sufficient penance. John Wesley, the founder of the Methodist Churches, excommunicated sixty-four members from the Newcastle Methodist society alone for the following reasons: The Allegheny Wesleyan Methodist Connection, in its 2014 "Discipline", includes "homosexuality, lesbianism, bi-sexuality, bestiality, incest, fornication, adultery, and any attempt to alter one’s gender by surgery", as well as remarriage after divorce among its excommunicable offences. The Evangelical Wesleyan Church, in its 2015 "Discipline", states that "Any member of our church who is accused of neglect of the means of grace or other duties required by the Word of God, the indulgence of sinful tempers, words or actions, the sowing of dissension, or any other violation of the order and discipline of the church, may, after proper labor and admonition, be censured, placed on probation, or expelled by the official board of the circuit of which he is a member. If he request a trial, however, within thirty dates of the final action of the official board, it shall be granted." When believers were baptized and taken into membership of the church by Anabaptists, it was not only done as symbol of cleansing of sin but was also done as a public commitment to identify with Jesus Christ and to conform one's life to the teaching and example of Jesus as understood by the church. Practically, that meant membership in the church entailed a commitment to try to live according to norms of Christian behavior widely held by the Anabaptist tradition. In the ideal, discipline in the Anabaptist tradition requires the church to confront a notoriously erring and unrepentant church member, first directly in a very small circle and, if no resolution is forthcoming, expanding the circle in steps eventually to include the entire church congregation. If the errant member persists without repentance and rejects even the admonition of the congregation, that person is excommunicated or excluded from church membership. Exclusion from the church is recognition by the congregation that this person has separated himself or herself from the church by way of his or her visible and unrepentant sin. This is done ostensibly as a final resort to protect the integrity of the church. When this occurs, the church is expected to continue to pray for the excluded member and to seek to restore him or her to its fellowship. There was originally no "inherent" expectation to shun (completely sever all ties with) an excluded member, however differences regarding this very issue led to early schisms between different Anabaptist leaders and those who followed them. Jakob Ammann, founder of the Amish sect, believed that the shunning of those under the ban should be systematically practiced among the Swiss Anabaptists as it was in the north and as was outlined in the Dordrecht Confession. Ammann's uncompromising zeal regarding this practice was one of the main disputes that led to the schism between the Anabaptist groups that became the Amish and those that eventually would be called Mennonite. Recently more moderate Amish groups have become less strict in their application of excommunication as a discipline. This has led to splits in several communities, an example of which is the Swartzetruber Amish who split from the main body of Old Order Amish because of the latter's practice of lifting the ban from members who later join other churches. In general, the Amish will excommunicate baptized members for failure to abide by their Ordnung (church rules) as it is interpreted by the local Bishop if certain repeat violations of the Ordnung occur. Excommunication among the Old Order Amish results in shunning or "the Meidung", the severity of which depends on many factors, such as the family, the local community as well as the type of Amish. Some Amish communities cease shunning after one year if the person joins another church later on, especially if it is another Mennonite church. At the most severe, other members of the congregation are prohibited almost all contact with an excommunicated member including social and business ties between the excommunicant and the congregation, sometimes even marital contact between the excommunicant and spouse remaining in the congregation or family contact between adult children and parents. In the Mennonite Church excommunication is rare and is carried out only after many attempts at reconciliation and on someone who is flagrantly and repeatedly violating standards of behavior that the church expects. Occasionally excommunication is also carried against those who repeatedly question the church's behavior or who genuinely differ with the church's theology as well, although in almost all cases the dissenter will leave the church before any discipline need be invoked. In either case, the church will attempt reconciliation with the member in private, first one on one and then with a few church leaders. Only if the church's reconciliation attempts are unsuccessful, the congregation formally revokes church membership. Members of the church generally pray for the excluded member. Some regional conferences (the Mennonite counterpart to dioceses of other denominations) of the Mennonite Church have acted to expel member congregations that have openly welcomed non-celibate homosexuals as members. This internal conflict regarding homosexuality has also been an issue for other moderate denominations, such as the American Baptists and Methodists. The practice among Old Order Mennonite congregations is more along the lines of Amish, but perhaps less severe typically. An Old Order member who disobeys the "Ordnung" (church regulations) must meet with the leaders of the church. If a church regulation is broken a second time there is a confession in the church. Those who refuse to confess are excommunicated. However upon later confession, the church member will be reinstated. An excommunicated member is placed under the ban. This person is not banned from eating with their own family. Excommunicated persons can still have business dealings with church members and can maintain marital relations with a marriage partner, who remains a church member. The separatist, communal, and self-contained Hutterites also use excommunication and shunning as form of church discipline. Since Hutterites have communal ownership of goods, the effects of excommunication could impose a hardship upon the excluded member and family leaving them without employment income and material assets such as a home. However, often arrangements are made to provide material benefits to the family leaving the colony such as an automobile and some transition funds for rent, etc. One Hutterite colony in Manitoba (Canada) had a protracted dispute when leaders attempted to force the departure of a group that had been excommunicated but would not leave. About a dozen lawsuits in both Canada and the United States were filed between the various Hutterite factions and colonies concerning excommunication, shunning, the legitimacy of leadership, communal property rights, and fair division of communal property when factions have separated. The Church of Jesus Christ of Latter-day Saints (LDS Church) practices excommunication as a penalty for those who commit serious sins, "i.e.", actions that significantly impair the name or moral influence of the church or pose a threat to other people. According to the church leadership "Handbook", the purposes of church discipline are (1) to save the souls of transgressors, (2) to protect the innocent, and (3) to safeguard the purity, integrity, and good name of the church. The origins of LDS disciplinary procedures and excommunications are traced to a revelation Joseph Smith, Jr. claimed to have received on February 9, 1831, later canonized as Doctrine and Covenants, section 42 and codified in the "Handbook". The LDS Church also practices the lesser sanctions of private counsel and caution, informal probation, formal probation, and disfellowshipment. Disfellowshipping is used for serious sins that do not rise to the level of excommunication. Disfellowshipment denies some privileges but does not include a loss of church membership. Once disfellowshipped, persons may not take the sacrament or enter church temples, nor may they offer public prayers or sermons. Disfellowshipped persons may continue to attend most church functions and are permitted to wear temple garments, pay tithes and offerings, and participate in church classes if their conduct is orderly. Disfellowshipment typically lasts for one year, after which one may be reinstated as a member in good standing. In the more grievous or recalcitrant cases, excommunication becomes a disciplinary option. Excommunication is generally reserved for what are seen as the most serious sins, including committing serious crimes such as murder, child abuse, and incest; committing adultery; involvement in or teaching of polygamy; involvement in homosexual conduct; apostasy; participation in an abortion; teaching false doctrine; or openly criticizing church leaders. A revision to the 2006 and subsequent "Handbook" states that formally joining another church constitutes apostasy and is an excommunicable offense; however, merely attending another church does not constitute apostasy. An excommunication can occur only after a formal disciplinary council. Formerly called a "church court," the councils were renamed to avoid focusing on guilt and instead to emphasize the availability of repentance. The decision to excommunicate a Melchizedek priesthood holder is generally the province of the leadership of a stake. In such a disciplinary council, the stake presidency and stake high council attend. The twelve members of the high council are split in half: one group represents the member in question and is charged with "prevent[ing] insult or injustice"; the other group represents the church as a whole. The member under scrutiny is invited to attend the disciplinary proceedings, but the council can go forward without him. In making a decision, the leaders of the high council consult with the stake presidency, but the decision about which discipline is necessary is the stake president's alone. It is possible to appeal a decision of a stake disciplinary council to the church's First Presidency. For females and for male members not initiated into the Melchizedek priesthood, a ward disciplinary council is held. In such cases, a bishop determines whether excommunication or a lesser sanction is warranted. He does this in consultation with his two counselors, with the bishop making the final determination after prayer. The decision of a ward disciplinary council can be appealed to the stake president. The following list of variables serves as a general set of guidelines for when excommunication or lesser action may be warranted, beginning with those more likely to result in severe sanction: Notices of excommunication may be made public, especially in cases of apostasy, where members could be misled. However, the specific reasons for individual excommunications are typically kept confidential and are seldom made public by church leadership. Those who are excommunicated lose their church membership and the right to partake of the sacrament. Such persons are usually allowed to attend church meetings but participation is limited: they cannot offer public prayers or preach sermons and cannot enter temples. Excommunicated members are also barred from wearing or purchasing temple garments and from paying tithes. Excommunicated members may be re-baptized after a waiting period and sincere repentance, as judged by a series of interviews with church leaders. Some critics have charged that LDS Church leaders have used the threat of excommunication to silence or punish church members and researchers who disagree with established policy and doctrine, who study or discuss controversial subjects, or who may be involved in disputes with local, stake leaders or general authorities; see, e.g., Brian Evenson, a former BYU professor and writer whose fiction came under criticism from BYU officials and LDS Leadership. Another notable case of excommunication from the LDS Church was the "September Six," a group of intellectuals and professors, five of whom were excommunicated and the sixth disfellowshipped. However, church policy dictates that local leaders are responsible for excommunication, without influence from church headquarters. The church thus argues that this policy is evidence against any systematic persecution of scholars or dissenters. Data shows per-capita excommunication rates among the LDS Church have varied dramatically over the years, from a low of about 1 in 6,400 members in the early 1900s to one in 640 by the 1970s, an increase which has been speculatively attributed to "informal guidance from above" in enforcing the growing list of possible transgressions added to "Handbook" editions over time. Jehovah's Witnesses practice a form of excommunication, using the term "disfellowshipping", in cases where a member is believed to have unrepentantly committed one or more of several documented "serious sins". The practice is based on their interpretation of 1 Corinthians 5:11-13 ("quit mixing in company with anyone called a brother that is a fornicator or greedy person or an idolater or a reviler or a drunkard or an extortioner, not even eating with such a man...remove the wicked man from your midst") and 2 John 10 ("never receive him in your home or say a greeting to him"). They interpret these verses to mean that any baptized believer who engages in "gross sins" is to be expelled from the congregation and shunned. When a member confesses to, or is accused of, a "serious sin", a "judicial committee" of at least three elders is formed. This committee investigates the case and determines the magnitude of the sin committed. If the person is deemed guilty of a disfellowshipping offense, the committee then decides, on the basis of the person's attitude and "works befitting repentance" (), whether the person is to be considered repentant. The "works" may include trying to correct the wrong, making apologies to any offended individuals, and compliance with earlier counsel. If deemed guilty but repentant, the person is not disfellowshipped but is formally "reproved" and has "restrictions" imposed, which preclude the individual from various activities such as presenting talks, offering public prayers or making comments at religious meetings. If the person is deemed guilty and unrepentant, he or she will be disfellowshipped. Unless an appeal is made within seven days, the disfellowshipping is made formal by an announcement at the congregation's next Service Meeting. Appeals are granted to determine if procedural errors are felt to have occurred that may have affected the outcome. Disfellowshipping is a severing of friendly relationships between all Jehovah's Witnesses and the disfellowshipped person. Interaction with extended family is typically restricted to a minimum, such as presence at the reading of wills and providing essential care for the elderly. Within a household, typical family contact may continue, but without spiritual fellowship such as family Bible study and religious discussions. Parents of disfellowshipped minors living in the family home may continue to attempt to convince the child about the group's teachings. Jehovah's Witnesses believe that this form of discipline encourages the disfellowshipped individual to conform to biblical standards and prevents the person from influencing other members of the congregation. Along with breaches of the Witnesses' moral code, openly disagreeing with the teachings Jehovah's Witnesses is considered grounds for shunning. These persons are labeled as "apostates" and are described in Watch Tower Society literature as "mentally diseased". Descriptions of "apostates" appearing in the Witnesses literature have been the subject of investigation in the UK to determine if they violate religious hatred laws. Sociologist Andrew Holden claims many Witnesses who would otherwise defect because of disillusionment with the organization and its teachings, remain affiliated out of fear of being shunned and losing contact with friends and family members. Shunning employs what is known "as relational aggression" in psychological literature. When used by church members and member-spouse parents against excommunicant parents it contains elements of what psychologists call "parental alienation". Extreme shunning may cause trauma to the shunned (and to their dependents) similar to what is studied in the psychology of torture. Disassociation is a form of shunning where a member expresses verbally or in writing that they do not wish to be associated with Jehovah's Witnesses, rather than for having committed any specific 'sin'. Elders may also decide that an individual has disassociated, without any formal statement by the individual, by actions such as accepting a blood transfusion, or for joining another religious or military organization. Individuals who are deemed by the elders to have disassociated are given no right of appeal. Each year, congregation elders are instructed to consider meeting with disfellowshipped individuals to determine changed circumstances and encourage them to pursue reinstatement. Reinstatement is not automatic after a certain time period, nor is there a minimum duration; disfellowshipped persons may talk to elders at any time but must apply in writing to be considered for reinstatement into the congregation. Elders consider each case individually, and are instructed to ensure "that sufficient time has passed for the disfellowshipped person to prove that his profession of repentance is genuine." A judicial committee meets with the individual to determine their repentance, and if this is established, the person is reinstated into the congregation and may participate with the congregation in their formal ministry (such as house-to-house preaching), but is prohibited from commenting at meetings or holding any privileges for a period set by the judicial committee. If possible, the same judicial committee members who disfellowshipped the individual are selected for the reinstatement hearing. If the applicant is in a different area, the person will meet with a local judicial committee that will communicate with either the original judicial committee if available or a new one in the original congregation. A Witness who has been formally reproved or reinstated cannot be appointed to any "special privilege of service" for at least one year. Serious sins involving child sex abuse permanently disqualify the sinner from appointment to any congregational "privilege of service", regardless of whether the sinner was convicted of any secular crime. Similarly to many groups having their origins in the 1830s Restoration Movement, Christadelphians call their form of excommunication "disfellowshipping", though they do not practice "shunning". Disfellowshipping can occur for moral reasons, changing beliefs, or (in some ecclesias) for not attending communion (referred to as "the emblems" or "the breaking of bread"). In such cases, the person involved is usually required to discuss the issues. If they do not conform, the church ('meeting' or 'ecclesia') is recommended by the management committee ("Arranging Brethren") to vote on disfellowshipping the person. These procedures were formulated 1863 onwards by early Christadelphians, and then in 1883 codified by Robert Roberts in "A Guide to the Formation and Conduct of Christadelphian Ecclesias" (colloquially "The Ecclesial Guide"). However Christadelphians justify and apply their practice not only from this document but also from passages such as the exclusion in 1Co.5 and recovery in 2Co.2. Christadelphians typically avoid the term "excommunication" which many associate with the Catholic Church; and may feel the word carries implications they do not agree with, such as undue condemnation and punishment, as well as failing to recognise the remedial intention of the measure. In the case of adultery and divorce, the passage of time usually means a member can be restored if he or she wants to be. In the case of ongoing behaviour, cohabitation, homosexual activity, then the terms of the suspension have not been met. The mechanics of "refellowship" follow the reverse of the original process; the individual makes an application to the "ecclesia", and the "Arranging Brethren" give a recommendation to the members who vote. If the "Arranging Brethren" judge that a vote may divide the ecclesia, or personally upset some members, they may seek to find a third party ecclesia which is willing to "refellowship" the member instead. According to the Ecclesial Guide a third party ecclesia may also take the initiative to "refellowship" another meeting's member. However this cannot be done unilaterally, as this would constitute heteronomy over the autonomy of the original ecclesia's members. Among many of the Society of Friends groups (Quakers) one is "read out of meeting" for behaviour inconsistent with the sense of the meeting. In Britain a meeting may record a minute of disunity. However it is the responsibility of each meeting, quarterly meeting, and yearly meeting, to act with respect to their own members. For example, during the Vietnam War many Friends were concerned about Friend Richard Nixon's position on war which seemed at odds with their beliefs; however, it was the responsibility of Nixon's own meeting, the East Whittier Meeting of Whittier, California, to act if indeed that meeting felt the leaning. They did not. In the 17th century, before the founding of abolitionist societies, Friends who too forcefully tried to convince their coreligionists of the evils of slavery were read out of meeting. Benjamin Lay was read out of the Philadelphia Yearly Meeting for this. During the American Revolution over 400 Friends were read out of meeting for their military participation or support. Iglesia ni Cristo practices expulsion of members it deems to have gravely sinned or gone against the teachings and doctrines of the church. The Sanggunian, the church's council, has jurisdiction to expel members from the church. People expelled by the church are referred to as "dismissed" (Tagalog: "tiwalag"). Offenses that may be grounds for expulsion include marrying a non-member, becoming pregnant out of wedlock (unless the couple marries before the child is born) and most especially disagreeing with the church administration. An expelled member can be re-admitted by pledging obedience to the church administration and its rules, values and teachings. Unitarian Universalism, being a liberal religious group and a congregational denomination, has a wide diversity of opinions and sentiments. Nonetheless, Unitarian Universalists have had to deal with disruptive individuals. Congregations which had no policies on disruptive individuals have sometimes found themselves having to create such policies, up to (and including) expulsion. By the late 1990s, several churches were using the West Shore UU Church's policy as a model. If someone is threatening, disruptive, or distracting from the appeal of the church to its membership, a church using this model has three recommended levels of response to the offending individual. While the first level involves dialogue between a committee or clergy member and the offender, the second and third levels involve expulsion, either from the church itself or a church activity. There is no direct equivalent to excommunication in Buddhism. However, in the Theravadan monastic community monks can be expelled from monasteries for heresy or other acts. In addition, the monks have four vows, called the four defeats, which are abstaining from sexual intercourse, stealing, murder, and refraining from lying about spiritual gains (e.g., having special power or ability to perform miracles). If even one is broken, the monk is automatically a layman again and can never become a monk in his or her current life. Most Japanese Buddhist sects hold ecclesiastical authority over their followers and have their own rules for expelling members of the sangha, lay or bishopric. The lay Japanese Buddhist organization Sōka Gakkai was expelled from the Nichiren Shoshu sect in 1991 (1997). Hinduism is too diverse to be seen as a homogenous and monolithic religion, it is often described an unorganised and syncretist religion with a conspicuous absence of any listed doctrines, there are multiple religious institutions (ecclesia is the Christian equivalent) within Hinduism that teach slight variations of Dharma and Karma, hence Hinduism has no concept of excommunication and hence no Hindu may be ousted from the Hindu religion, although a person may easily lose caste status through gramanya for a very wide variety of infringements of caste prohibitions. This may or may not be recoverable. However, some of the modern organised sects within Hinduism may practice something equivalent to excommunication today, by ousting a person from their own sect. In medieval and early-modern times (and sometimes even now) in South Asia, excommunication from one's "caste" ("jati" or "varna") used to be practiced (by the caste-councils) and was often with serious consequences, such as abasement of the person's caste status and even throwing him into the sphere of the untouchables or bhangi. In the 19th century, a Hindu faced excommunication for going abroad, since it was presumed he would be forced to break caste restrictions and, as a result, become polluted. After excommunication, it would depend upon the caste-council whether they would accept any form of repentance (ritual or otherwise) or not. Such current examples of excommunication in Hinduism are often more political or social rather than religious, for example the excommunication of lower castes for refusing to work as scavengers in Tamil Nadu. An earlier example of excommunication in Hinduism is that of Shastri Yagnapurushdas, who voluntarily left and was later expelled from the Vadtal Gadi of the Swaminarayan Sampraday by the then Vadtal acharya in 1906. He went on to form his own institution, "Bochasanwasi Swaminarayan Sanstha" or "BSS" (now BAPS) claiming Gunatitanand Swami was the rightful spiritual successor to Swaminarayan. Excommunication as it exists in Christian faiths does not exist in Islam. The nearest approximation is "takfir", a declaration that an individual or group is "kafir" (or "kuffar" in plural), a non-believer. This does not prevent an individual from taking part in any Islamic rite or ritual, and since the matter of whether a person is "kafir" is a rather subjective matter, a declaration of "takfir" is generally considered null and void if the target refutes it or if the Islamic community in which he or she lives refuses to accept it. "Takfir" has usually been practiced through the courts. More recently, cases have taken place where individuals have been considered kuffar. These decisions followed lawsuits against individuals, mainly in response to their writings that some have viewed as anti-Islamic. The most famous cases are of Salman Rushdie, Nasr Abu Zayd, and Nawal El-Saadawi. The repercussions of such cases have included divorce, since under traditional interpretations of Islamic law, Muslim women are not permitted to marry non-Muslim men. However, "takfir" remains a highly contentious issue in Islam, primarily because there is no universally accepted authority in Islamic law. Indeed, according to classical commentators, the reverse seems to hold true, in that Muhammad reportedly equated the act of declaring someone a "kafir" itself to blasphemy if the accused individual maintained that he was a Muslim. "Herem" is the highest ecclesiastical censure in Judaism. It is the total exclusion of a person from the Jewish community. Except for cases in the Charedi community, "cherem" stopped existing after The Enlightenment, when local Jewish communities lost their political autonomy, and Jews were integrated into the gentile nations in which they lived. A "siruv" order, equivalent to a contempt of court, issued by a Rabbinical court may also limit religious participation. Rabbinical conferences of movements do expel members from time to time, but sometimes choose the lesser penalty of censuring the offending rabbi. Between 2010 and 2015, the Reform Jewish Central Conference of American Rabbis expelled six rabbis, the Orthodox Jewish Rabbinical Council of America expelled three, and the Conservative Jewish Rabbinical Assembly expelled one, suspended three, and caused one to resign without eligibility for reinstatement. While the CCAR and RCA were relatively shy about their reasons for expelling rabbis, the RA was more open about its reasons for kicking rabbis out. Reasons for expulsion from the three conferences include sexual misconduct, failure to comply with ethics investigations, setting up conversion groups without the conference's approval, stealing money from congregations, other financial misconduct, and getting arrested. Judaism, like Unitarian Universalism, tends towards congregationalism, and so decisions to exclude from a community of worship often depend on the congregation. Congregational bylaws sometimes enable the board of a synagogue to ask individuals to leave or not to enter.
https://en.wikipedia.org/wiki?curid=10338
Electrochemical cell An electrochemical cell is a device capable of either generating electrical energy from chemical reactions or using electrical energy to cause chemical reactions. The electrochemical cells which generate an electric current are called voltaic cells or galvanic cells and those that generate chemical reactions, via electrolysis for example, are called electrolytic cells. A common example of a galvanic cell is a standard 1.5 volt cell meant for consumer use. A "battery" consists of one or more cells, connected in parallel, series or series-and-parallel pattern. An electrolytic is an electrochemical cell that drives a non-spontaneous redox reaction through the application of electrical energy. They are often used to decompose chemical compounds, in a process called electrolysis—the Greek word lysis means "to break up". Important examples of electrolysis are the decomposition of water into hydrogen and oxygen, and bauxite into aluminium and other chemicals. Electroplating (e.g. of copper, silver, nickel or chromium) is done using an electrolytic cell. Electrolysis is a technique that uses a direct electric current (DC). An electrolytic cell has three component parts: an electrolyte and two electrodes (a cathode and an anode). The electrolyte is usually a solution of water or other solvents in which ions are dissolved. Molten salts such as sodium chloride are also electrolytes. When driven by an external voltage applied to the electrodes, the ions in the electrolyte are attracted to an electrode with the opposite charge, where charge-transferring (also called faradaic or redox) reactions can take place. Only with an external electrical potential (i.e. voltage) of correct polarity and sufficient magnitude can an electrolytic cell decompose a normally stable, or inert chemical compound in the solution. The electrical energy provided can produce a chemical reaction which would not occur spontaneously otherwise A galvanic cell, or voltaic cell, named after Luigi Galvani, or Alessandro Volta respectively, is an electrochemical cell that derives electrical energy from spontaneous redox reactions taking place within the cell. It generally consists of two different metals connected by a salt bridge, or individual half-cells separated by a porous membrane. Volta was the inventor of the voltaic pile, the first electrical battery. In common usage, the word "battery" has come to include a single galvanic cell, but a battery properly consists of multiple cells. A primary cell is a Galvanic battery that is designed to be used once and discarded, in contrast to a secondary cell (rechargeable battery), which can be recharged with electricity and reused. In general, the electrochemical reaction occurring in the cell is not reversible, rendering the cell unrechargeable. As a primary cell is used, chemical reactions in the battery use up the chemicals that generate the power; when they are gone, the battery stops producing electricity and is useless. In contrast, in a secondary cell, the reaction can be reversed by running a current into the cell with a battery charger to recharge it, regenerating the chemical reactants. Primary cells are made in a range of standard sizes to power small household appliances such as flashlights and portable radios. Primary batteries make up about 90% of the $50 billion battery market, but secondary batteries have been gaining market share. About 15 billion primary batteries are thrown away worldwide every year, virtually all ending up in landfills. Due to the toxic heavy metals and strong acids or alkalis they contain, batteries are hazardous waste. Most municipalities classify them as such and require separate disposal. The energy needed to manufacture a battery is about 50 times greater than the energy it contains. Due to their high pollutant content compared to their small energy content, the primary battery is considered a wasteful, environmentally unfriendly technology. Due mainly to increasing sales of wireless devices and cordless tools which cannot be economically powered by primary batteries and come with integral rechargeable batteries, the secondary battery industry has high growth and has slowly been replacing the primary battery in high end products. A secondary cell, commonly referred to as a "rechargeable battery", is an electrochemical cell that can be run as both a galvanic cell and as an electrolytic cell. This is used as a convenient way to store electricity, when current flows one way the levels of one or more chemicals build up (charging), while it is discharging they reduce and the resulting electromotive force can do work. A "fuel cell" is an electrochemical cell that converts the chemical energy from a fuel into electricity through an electrochemical reaction of hydrogen fuel with oxygen or another oxidizing agent. Fuel cells are different from batteries in requiring a continuous source of fuel and oxygen (usually from air) to sustain the chemical reaction, whereas in a battery the chemical energy comes from chemicals already present in the battery. Fuel cells can produce electricity continuously for as long as fuel and oxygen are supplied. The first fuel cells were invented in 1838. The first commercial use of fuel cells came more than a century later in NASA space programmes to generate power for satellites and space capsules. Since then, fuel cells have been used in many other applications. Fuel cells are used for primary and backup power for commercial, industrial and residential buildings and in remote or inaccessible areas. They are also used to power fuel cell vehicles, including forklifts, automobiles, buses, boats, motorcycles and submarines. There are many types of fuel cells, but they all consist of an anode, a cathode, and an electrolyte that allows positively charged hydrogen ions (protons) to move between the two sides of the fuel cell. At the anode a catalyst causes the fuel to undergo oxidation reactions that generate protons (positively charged hydrogen ions) and electrons. The protons flow from the anode to the cathode through the electrolyte after the reaction. At the same time, electrons are drawn from the anode to the cathode through an external circuit, producing direct current electricity. At the cathode, another catalyst causes hydrogen ions, electrons, and oxygen to react, forming water. Fuel cells are classified by the type of electrolyte they use and by the difference in startup time, which ranges from 1 second for proton exchange membrane fuel cells (PEM fuel cells, or PEMFC) to 10 minutes for solid oxide fuel cells (SOFC). A related technology is flow batteries, in which the fuel can be regenerated by recharging. Individual fuel cells produce relatively small electrical potentials, about 0.7 volts, so cells are "stacked", or placed in series, to create sufficient voltage to meet an application's requirements. In addition to electricity, fuel cells produce water, heat and, depending on the fuel source, very small amounts of nitrogen dioxide and other emissions. The energy efficiency of a fuel cell is generally between 40–60%; however, if waste heat is captured in a cogeneration scheme, efficiencies up to 85% can be obtained. The fuel cell market is growing, and in 2013 Pike Research estimated that the stationary fuel cell market will reach 50 GW by 2020. An electrochemical cell consists of two half-cells. Each "half-cell" consists of an electrode and an electrolyte. The two half-cells may use the same electrolyte, or they may use different electrolytes. The chemical reactions in the cell may involve the electrolyte, the electrodes, or an external substance (as in fuel cells that may use hydrogen gas as a reactant). In a full electrochemical cell, species from one half-cell lose electrons (oxidation) to their electrode while species from the other half-cell gain electrons (reduction) from their electrode. A "salt bridge" (e.g., filter paper soaked in KNO3, NaCl, or some other electrolyte) is often employed to provide ionic contact between two half-cells with different electrolytes, yet prevent the solutions from mixing and causing unwanted side reactions. An alternative to a salt bridge is to allow direct contact (and mixing) between the two half-cells, for example in simple electrolysis of water. As electrons flow from one half-cell to the other through an external circuit, a difference in charge is established. If no ionic contact were provided, this charge difference would quickly prevent the further flow of electrons. A salt bridge allows the flow of negative or positive ions to maintain a steady-state charge distribution between the oxidation and reduction vessels, while keeping the contents otherwise separate. Other devices for achieving separation of solutions are porous pots and gelled solutions. A porous pot is used in the Bunsen cell (right). Each half-cell has a characteristic voltage. Various choices of substances for each half-cell give different potential differences. Each reaction is undergoing an equilibrium reaction between different oxidation states of the ions: When equilibrium is reached, the cell cannot provide further voltage. In the half-cell that is undergoing oxidation, the closer the equilibrium lies to the ion/atom with the more positive oxidation state the more potential this reaction will provide. Likewise, in the reduction reaction, the closer the equilibrium lies to the ion/atom with the more "negative" oxidation state the higher the potential. The cell potential can be predicted through the use of electrode potentials (the voltages of each half-cell). These half-cell potentials are defined relative to the assignment of 0 volts to the standard hydrogen electrode (SHE). (See table of standard electrode potentials). The difference in voltage between electrode potentials gives a prediction for the potential measured. When calculating the difference in voltage, one must first rewrite the half-cell reaction equations to obtain a balanced oxidation-reduction equation. Cell potentials have a possible range of roughly zero to 6 volts. Cells using water-based electrolytes are usually limited to cell potentials less than about 2.5 volts due to high reactivity of the powerful oxidizing and reducing agents with water that is needed to produce a higher voltage. Higher cell potentials are possible with cells using other solvents instead of water. For instance, lithium cells with a voltage of 3 volts are commonly available. The cell potential depends on the concentration of the reactants, as well as their type. As the cell is discharged, the concentration of the reactants decreases and the cell potential also decreases.
https://en.wikipedia.org/wiki?curid=10339
Ecdysis Ecdysis is the moulting of the cuticle in many invertebrates of the clade Ecdysozoa. Since the cuticle of these animals typically forms a largely inelastic exoskeleton, it is shed during growth and a new, larger covering is formed. The remnants of the old, empty exoskeleton are called exuviae. After moulting, an arthropod is described as "teneral", a "callow"; it is "fresh", pale and soft-bodied. Within one or two hours, the cuticle hardens and darkens following a tanning process analogous to the production of leather. During this short phase the animal expands, since growth is otherwise constrained by the rigidity of the exoskeleton. Growth of the limbs and other parts normally covered by hard exoskeleton is achieved by transfer of body fluids from soft parts before the new skin hardens. A spider with a small abdomen may be undernourished but more probably has recently undergone ecdysis. Some arthropods, especially large insects with tracheal respiration, expand their new exoskeleton by swallowing or otherwise taking in air. The maturation of the structure and colouration of the new exoskeleton might take days or weeks in a long-lived insect; this can make it difficult to identify an individual if it has recently undergone ecdysis. Ecdysis allows damaged tissue and missing limbs to be regenerated or substantially re-formed. Complete regeneration may require a series of moults, the stump becoming a little larger with each moult until it is a normal, or near normal, size. The term "ecdysis" comes from (""), "to take off, strip off". In preparation for ecdysis, the arthropod becomes inactive for a period of time, undergoing apolysis or separation of the old exoskeleton from the underlying epidermal cells. For most organisms, the resting period is a stage of preparation during which the secretion of fluid from the moulting glands of the epidermal layer and the loosening of the underpart of the cuticle occur. Once the old cuticle has separated from the epidermis, a digesting fluid is secreted into the space between them. However, this fluid remains inactive until the upper part of the new cuticle has been formed. Then, by crawling movements, the organism pushes forward in the old integumentary shell, which splits down the back allowing the animal to emerge. Often, this initial crack is caused by a combination of movement and increase in blood pressure within the body, forcing an expansion across its exoskeleton, leading to an eventual crack that allows for certain organisms such as spiders to extricate themselves. While the old cuticle is being digested, the new layer is secreted. All cuticular structures are shed at ecdysis, including the inner parts of the exoskeleton, which includes terminal linings of the alimentary tract and of the tracheae if they are present. Each stage of development between moults for insects in the taxon endopterygota is called an instar, or stadium, and each stage between moults of insects in the Exopterygota is called a nymph: there may be up to 15 nymphal stages. Endopterygota tend to have only four or five instars. Endopterygotes have more alternatives to moulting, such as expansion of the cuticle and collapse of air sacs to allow growth of internal organs. The process of moulting in insects begins with the separation of the cuticle from the underlying epidermal cells (apolysis) and ends with the shedding of the old cuticle (ecdysis). In many species it is initiated by an increase in the hormone ecdysone. This hormone causes: After apolysis the insect is known as a pharate. Moulting fluid is then secreted into the exuvial space between the old cuticle and the epidermis, this contains inactive enzymes which are activated only after the new epicuticle is secreted. This prevents the new procuticle from getting digested as it is laid down. The lower regions of the old cuticle, the endocuticle and mesocuticle, are then digested by the enzymes and subsequently absorbed. The exocuticle and epicuticle resist digestion and are hence shed at ecdysis. Spiders generally change their skin for the first time while still inside the egg sac, and the spiderling that emerges broadly resembles the adult. The number of moults varies, both between species and sexes, but generally will be between five times and nine times before the spider reaches maturity. Not surprisingly, since males are generally smaller than females, the males of many species mature faster and do not undergo ecdysis as many times as the females before maturing. Members of the Mygalomorphae are very long-lived, sometimes 20 years or more; they moult annually even after they mature. Spiders stop feeding at some time before moulting, usually for several days. The physiological processes of releasing the old exoskeleton from the tissues beneath typically cause various colour changes, such as darkening. If the old exoskeleton is not too thick it may be possible to see new structures, such as setae, from outside. However, contact between the nerves and the old exoskeleton is maintained until a very late stage in the process. The new, teneral exoskeleton has to accommodate a larger frame than the previous instar, while the spider has had to fit into the previous exoskeleton until it has been shed. This means the spider does not fill out the new exoskeleton completely, so it commonly appears somewhat wrinkled. Most species of spiders hang from silk during the entire process, either dangling from a drop line, or fastening their claws into webbed fibres attached to a suitable base. The discarded, dried exoskeleton typically remains hanging where it was abandoned once the spider has left. To open the old exoskeleton, the spider generally contracts its abdomen (opisthosoma) to supply enough fluid to pump into the prosoma with sufficient pressure to crack it open along its lines of weakness. The carapace lifts off from the front, like a helmet, as its surrounding skin ruptures, but it remains attached at the back. Now the spider works its limbs free and typically winds up dangling by a new thread of silk attached to its own exuviae, which in turn hang from the original silk attachment. At this point the spider is a callow; it is teneral and vulnerable. As it dangles, its exoskeleton hardens and takes shape. The process may take minutes in small spiders, or some hours in the larger Mygalomorphs. Some spiders, such as some "Synema" species, members of the Thomisidae (crab spiders), mate while the female is still callow, during which time she is unable to eat the male. Eurypterids are a group of chelicerates that became extinct in the Late Permian. They underwent ecdysis similarly to extant chelicerates, and most fossils are thought to be of exuviae, rather than cadavers.
https://en.wikipedia.org/wiki?curid=10340
Ebor, New South Wales Ebor is a village on Waterfall Way on the Northern Tablelands in New South Wales, Australia. It is situated about east of Armidale and about a third of the way between Armidale and the coast. Dorrigo to the east is away with the Coffs Coast away along Waterfall Way. In the , Ebor's zone had a population of 166. Although "The Heart of Waterfall Way", Ebor is on the eastern edge of Armidale Regional Council, and close to the border of Clarence Valley Council and Bellingen Shire Council. Until the amalgamation of Guyra and Armidale councils one side of Ebor was under Armidale council and the other under Guyra shire. Likewise, Ebor is close to the borders of 3 state (Northern Tablelands, Oxley and Clarence) and 3 federal electoral boundaries (New England, Cowper and Page). Amenities in the area include a cafe, a combined post office, fuel station and general store, a pub/motel with camp ground, and a NSW DEC primary school. The local sports ground is home of the Ebor Campdraft. There are also Rural Fire Service and National Parks and Wildlife Service depots in the area, but no police or ambulance services based in Ebor. The nearest hospital and 24h emergency department is in Dorrigo. Due to its central position on Waterfall Way, Ebor offers easy access for residents and tourists to Guy Fawkes River National Park, Cathedral Rock National Park, Cunnawarra National Park, New England National Park, part of Oxley Wild Rivers National Park, Nymboi-Binderay National Park and Mount Hyland Nature Reserve. The natural environment of the surrounding district includes some areas which have been cleared for pastoralism and forestry. Nonetheless, the national parks around Ebor have been described as a bush walking "Mecca". The main tourist attraction is the twin Ebor Falls. In 1930 Sydney Smith Jr. wrote that: ""During a recent visit to Ebor I was much impressed with the possibilities of this part of the State as a tourist resort... Around Ebor and Guy Fawkes can be seen some of the most magnificent scenery in this State if not Australia. ...The two falls are scenes of beauty, and in winter time are sometimes frozen, making a beautiful spectacle as they hang in huge icicles. The water from the Ebor eventually finds an outlet in the Clarence River. ...The view, ...as regards expansiveness, ruggedness, and beauty, must compare more than favourably with views of a similar nature in any part of the Commonwealth. It reminded me of the Valley of a Thousand Hills, outside Durban, in South Africa". In 1976, local historian Eric Fahey also wrote: ""I believe the future of Dorrigo will depend largely on tourism. The area has a lot to offer, both in peerless scenery and because of the native fauna which can be seen in large numbers in their natural state."" Wagyu beef specialists Stone Axe have a large holding, "Glen Alvie", on the northern boundary of the village. Stone Axe also acquired "Alfreda" in the nearby locality of Wongwibinda. Black truffles (tuber melanosporum) are grown at the Guy Fawkes Truffle Company outside of Ebor on the Guyra Rd. Trout are another local product. The Dutton Trout Hatchery on Point Lookout Road was established in 1950 and is one of the largest hatcheries in the state. Visitors can see the various stages of trout development prior to their release in the mountain streams. The release of trout into local streams is believed to have led to decline of the endangered Tusked frog. There are two short walks close to the village. One takes walkers through the recreation reserve. This walk follows the Guy Fawkes River upstream for about half of the walk. Some bird life can be seen. The second walk is accessed by crossing the Guy Fawkes River bridge, and following the pedestrian path which winds downstream under the bridge. This path follows the Guy Fawkes River north and meets the national park's Upper and Lower Falls paths. Wallabies, kangaroos, bird life and fire-flies can be seen depending on the season. Platypus have also been sighted in August the pool above the falls. The Bicentennial National Trail (BNT) passes through Ebor which sits on the boundary of sections 7 and 8 of the BNT. The Ebor Falls area is sometimes used for rock climbing, and is described as "holding a rather special place in the History of New England climbing". It is also a location for Highlining. Ebor has a noted problem with speeding vehicles. Both passenger cars and heavy vehicles regularly exceed the posted speed limit of 50 km/h. Traffic noise is also a problem. Waterfall Way has an entry on the Dangerous Roads website. Ebor's Post Office opened on 2 March 1868, closed in 1869 and reopened in 1910. It is currently located at the Ebor petrol station/store having moved from Fusspots Cafe. Ebor has a number of cultural heritage sites, including several Aboriginal meeting places, and massacre sites. "Gwenda Gardens" is an abandoned homestead on the Guyra-Ebor Road. Other sites include: The village of Ebor is at high altitude by Australian standards. It has cold winters with frequent overnight frost and occasional light snow falls. The average rain fall is about .
https://en.wikipedia.org/wiki?curid=10343
Ancient history of Afghanistan Archaeological exploration of the pre-Islamic period of Afghanistan began in Afghanistan in earnest after World War II and proceeded until the late 1970s when the nation was invaded by the Soviet Union. Archaeologists and historians suggest that humans were living in Afghanistan at least 50,000 years ago, and that farming communities of the region were among the earliest in the world. Urbanized culture has existed in the land from between 3000 and 2000 BC. Artifacts typical of the Paleolithic, Mesolithic, Neolithic, Bronze, and Iron ages have been found inside Afghanistan. After the Indus Valley Civilization stretched up to northeast Afghanistan, it was inhabited by the Iranic tribes and controlled by the Medes until about 500 BC when Darius the Great (Darius I) marched with his Persian army to make it part of the Achaemenid Empire. In 330 BC, Alexander the Great of Macedonia invaded the land after defeating Darius III of Persia in the Battle of Gaugamela. Much of Afghanistan became part of the Seleucid Empire followed by the Greco-Bactrian Kingdom. Seleucus I Nicator was defeated by Chandragupta Maurya and gave his daughter in peace treaty. The land was inhabited by various tribes and ruled by many different kingdoms for the next two millenniums. Before the arrival of Islam in the 7th century, there were a number of religions practiced in ancient Afghanistan, including Zoroastrianism, Surya worship, Christianity, Paganism, Hinduism and Buddhism. The Kaffirstan region, in the Hindu Kush, was not converted until the 19th century. Louis Dupree, the University of Pennsylvania, the Smithsonian Institution and others suggest that humans were living in Afghanistan at least 50,000 years ago, and that farming communities of the region were among the earliest in the world. Afghanistan seems in prehistory, as well as in ancient and modern times, to have been connected by culture and trade with the neighbouring regions. Urban civilization, which includes modern-day Afghanistan, North India, and Pakistan, may have begun as early as 3000 to 2000 BC. Archaeological finds indicate the possible beginnings of the Bronze Age, which would ultimately spread throughout the ancient world from Afghanistan. It is also believed that the region had early trade contacts with Mesopotamia. The Indus Valley Civilization (IVC) was a Bronze Age civilization (3300-1300 BCE; mature period 2600-1900 BCE) extending from what today is northwest Pakistan to northwest India and northeast Afghanistan. An Indus Valley site has been found on the Oxus River at Shortugai in northern Afghanistan. Apart from Shortughai is Mundigak, another notable site. There are several smaller IVC colonies to be found in Afghanistan. Between 2000–1200 BC, a branch of Indo-European-speaking tribes known as the Aryans began migrating into the region. This is part of a dispute in regards to the Aryan invasion theory. They appear to have split into Iranic peoples, Nuristani, and Indo-Aryan groups at an early stage, possibly between 1500 and 1000 BC in what is today Afghanistan or much earlier as eastern remnants of the Indo-Aryans drifted much further west as with the Mitanni. The Iranians dominated the modern day plateau, while the Indo-Aryans ultimately headed towards the Indian subcontinent. The Avesta is believed to have been composed possibly as early as 1800 BC and written in ancient Ariana (Aryana), the earliest name of Afghanistan which indicates an early link with today's Iranian tribes to the west, or adjacent regions in Central Asia or northeastern Iran in the 6th century BC. Due to the similarity between early Avestan and Sanskrit (and other related early Indo-European languages such as Latin and Ancient Greek), it is believed that the split between the old Persians and Indo-Aryan tribes had taken place at least by 1000 BC. There are striking similarities between Avestan and Sanskrit, which may support the notion that the split was contemporary with the Indo-Aryans living in Afghanistan at a very early stage. Also, the Avesta itself divides into Old and New sections and neither mention the Medes who are known to have ruled Afghanistan starting around 700 BC. This suggests an early time-frame for the Avesta that has yet to be exactly determined as most academics believe it was written over the course of centuries if not millennia. Much of the archaeological data comes from the Bactria-Margiana Archaeological Complex (BMAC and Indus Valley Civilization) that probably played a key role in early Aryanic civilization in Afghanistan. The Medes, a Western Persian people, arrived from what is today Kurdistan sometime around the 700s BC and came to dominate most of ancient Afghanistan. They were an early tribe that forged the first empire on the present Iranian plateau and were rivals of the Persians whom they initially dominated in the province of Fars to the south. Median domination of parts of far off Afghanistan would last until the Persians challenged and ultimately replaced them from rule. The city of Bactria (which later became Balkh), is believed to have been the home of Zarathustra, who founded the Zoroastrian religion. The Avesta refers to eastern Bactria as being the home of the Zoroastrian faith. Regardless of the debate as to where Zoroaster was from, Zoroastrianism spread to become one of the world's most influential religions and became the main faith of the old Aryan people for centuries. It also remained the official religion of Persia until the defeat of the Sassanian ruler Yazdegerd III—over a thousand years after its founding—by Muslim Arabs. In what is today southern Iran, the Persians emerged to challenge Median supremacy on the Iranian plateau. By 550 BC, the Persians had replaced Median rule with their own dominion and even began to expand past previous Median imperial borders. Both Gandhara and Kamboja Mahajanapadas of the Buddhist texts soon fell a prey to the Achaemenian Dynasty during the reign of Achaemenid, Cyrus the Great (558–530 BC), or in the first year of Darius I, marking the region or of the eastern-most provinces of the empire, located partly in nowadays Afghanistan. According to Pliny's evidence, Cyrus the Great (Cyrus II) had destroyed Kapisa in Capiscene which was a Kamboja city. The former region of Gandhara and Kamboja (upper Indus) had constituted seventh satrapy of the Achaemenid Empire and annually contributed 170 talents of gold dust as a tribute to the Achaemenids. Bactria had a special position in old Afghanistan, being the capital of a vice-kingdom. By the 4th century BC, Persian control of outlying areas and the internal cohesion of the empire had become somewhat tenuous. Although distant provinces like Bactriana had often been restless under Achaemenid rule, Bactrian troops nevertheless fought in the decisive Battle of Gaugamela in 330 BC against the advancing armies of Alexander the Great. The Achaemenids were decisively defeated by Alexander and retreated from his advancing army of Greco-Macedonians and their allies. Darius III, the last Achaemenid ruler, tried to flee to Bactria but was assassinated by a subordinate lord, the Bactrian-born Bessus, who proclaimed himself the new ruler of Persia as Artaxerxes (V). Bessus was unable to mount a successful resistance to the growing military might of Alexander's army so he fled to his native Bactria, where he attempted to rally local tribes to his side but was instead turned over to Alexander who proceeded to have him tortured and executed for having committed regicide. Moving thousands of kilometers eastward from recently subdued Persia, the Macedonian leader Alexander the Great, encountered fierce resistance from the local tribes of Aria (satrapy), Drangiana, Arachosia (South and Eastern Afghanistan, North-West Pakistan) and Bactria (North and Central Afghanistan). Upon Alexander's death in 323 BC, his empire, which had never been politically consolidated, broke apart as his companions began to divide it amongst themselves. Alexander's cavalry commander, Seleucus, took nominal control of the eastern lands and founded the Seleucid dynasty. Under the Seleucids, as under Alexander, Greek colonists and soldiers colonized Bactria, roughly corresponding to modern Afghanistan's borders. However, the majority of Macedonian soldiers of Alexander the Great wanted to leave the east and return home to Greece. Later, Seleucus sought to guard his eastern frontier and moved Ionian Greeks (also known as Yavanas to many local groups) to Bactria in the 3rd century BC. While the Diadochi were warring amongst themselves, the Mauryan Empire was developing in the northern part of the Indian subcontinent. The founder of the empire, Chandragupta Maurya, confronted a Macedonian invasion force led by Seleucus I in 305 BC and following a brief conflict, an agreement was reached as Seleucus ceded Gandhara and Arachosia (centered around ancient Kandahar) and areas south of Bagram (corresponding to the extreme south-east of modern Afghanistan) to the Mauryans. During the 120 years of the Mauryans in southern Afghanistan, Buddhism was introduced and eventually become a major religion alongside Zoroastrianism and local pagan beliefs. The ancient Grand Trunk Road was built linking what is now Kabul to various cities in the Punjab and the Gangetic Plain. Commerce, art, and architecture (seen especially in the construction of stupas) developed during this period. It reached its high point under Emperor Ashoka whose edicts, roads, and rest stops were found throughout the subcontinent. Although the vast majority of them throughout the subcontinent were written in Prakrit, Afghanistan is notable for the inclusion of 2 Greek and Aramaic ones alongside the court language of the Mauryans. Inscriptions made by the Mauryan Emperor Ashoka, a fragment of Edict 13 in Greek, as well as a full Edict, written in both Greek and Aramaic has been discovered in Kandahar. It is said to be written in excellent Classical Greek, using sophisticated philosophical terms. In this Edict, Ashoka uses the word Eusebeia ("Piety") as the Greek translation for the ubiquitous "Dharma" of his other Edicts written in Prakrit: The last ruler in the region was probably Subhagasena (Sophagasenus of Polybius), who, in all probability, belonged to the Ashvaka (q.v.) background. In the middle of the 3rd century BC, an independent, Hellenistic state was declared in Bactria and eventually the control of the Seleucids and Mauryans was overthrown in western and southern Afghanistan. Graeco-Bactrian rule spread until it included a large territory which stretched from Turkmenistan in the west to the Punjab in India in the east by about 170 BC. Graeco-Bactrian rule was eventually defeated by a combination of internecine disputes that plagued Greek and Hellenized rulers to the west, continual conflict with Indian kingdoms, as well as the pressure of two groups of nomadic invaders from Central Asia—the Parthians and Sakas. In the 3rd and 2nd centuries BC, the Parthians, a nomadic Iranian peoples, arrived in Western Asia. While they made large inroads into the modern-day territory of Afghanistan, about 100 years later another Indo-European group from the north—the Kushans (a subgroup of the tribe called the Yuezhi by the Chinese)—entered the region of Afghanistan and established an empire lasting almost four centuries, which would dominate most of the Afghanistan region. The Kushan Empire spread from the Kabul River valley to defeat other Central Asian tribes that had previously conquered parts of the northern central Iranian Plateau once ruled by the Parthians. By the middle of the 1st century BC, the Kushans' base of control became Afghanistan and their empire spanned from the north of the Pamir mountains to the Ganges river valley in India. Early in the 2nd century under Kanishka, the most powerful of the Kushan rulers, the empire reached its greatest geographic and cultural breadth to become a center of literature and art. Kanishka extended Kushan control to the mouth of the Indus River on the Arabian Sea, into Kashmir, and into what is today the Chinese-controlled area north of Tibet. Kanishka was a patron of religion and the arts. It was during his reign that Buddhism, which was promoted in northern India earlier by the Mauryan emperor Ashoka (c. 260 BC–232 BC), reached its zenith in Central Asia. Though the Kushanas supported local Buddhists and Hindus as well as the worship of various local deities. In the 3rd century, Kushan control fragmented into semi-independent kingdoms that became easy targets for conquest by the rising Iranian dynasty, the Sasanians (c. 224–561) which annexed Afghanistan by 300 AD. In these far off eastern-most territories, they established vassal kings as rulers, known as the Kushanshahs. Sasanian control was tenuous at times as numerous challenges from Central Asian tribes led to instability and constant warfare in the region. The disunited Kushan and Sasanian kingdoms were in a poor position to meet the threat several waves of Xionite/Huna invaders from the north from the 4th century onwards. In particular, the Hephthalites (or "Ebodalo"; Bactrian script ηβοδαλο) swept out of Central Asia during the 5th century into Bactria and Iran, overwhelming the last of the Kushan kingdoms. Historians believe that Hephthalite control continued for a century and was marked by constant warfare with the Sassanians to the west who exerted nominal control over the region. By the middle of the 6th century the Hephthalites were defeated in the territories north of the Amu Darya (the Oxus River of antiquity) by another group of Central Asian nomads, the Göktürks, and by the resurgent Sassanians in the lands south of the Amu Darya. It was the ruler of western Göktürks, Sijin (a.k.a. Sinjibu, Silzibul and Yandu Muchu Khan) who led the forces against the Hepthalites who were defeated at the Battle of Chach (Tashkent) and at the Battle of Bukhara. The Shahi dynasties ruled portions of the Kabul Valley (in eastern Afghanistan) and the old province of Gandhara (northern Pakistan and Kashmir) from the decline of the Kushan Empire in the 3rd century to the early 9th century. They are split into two eras the Buddhist-Shahis and the later Hindu-Shahis with the change-over occurring around 870, and ruled up until the Islamic conquest of Afghanistan. When Xuanzang visited the region early in the 7th century, the Kabul region was ruled by a Kshatriya king, who is identified as the "Shahi Khingal", and whose name has been found in an inscription found in Gardez. The Turkic Shahi regency was overthrown and replaced by a Mohyal Shahi dynasty of Brahmins who began the first phase of the Hindu Shahi Hindu Shahi dynasty. These Hindu kings of Kabul and Gandhara may have had links to some ruling families in neighboring Kashmir and other areas to the east. The Shahis were rulers of predominantly Buddhist, Zoroastrian, Hindu and Muslim populations and were thus patrons of numerous faiths, and various artifacts and coins from their rule have been found that display their multicultural domain. In 964 AD, the last Mohyal Shahi was succeeded by the Janjua overlord, Jayapala, of the Panduvanshi dynasty. The last Shahi emperors Jayapala, Anandapala and Tirlochanpala fought the Muslim Ghaznavids of Ghazna and were gradually defeated. Their remaining army were eventually exiled into northern India. Most of the Zoroastrian, Greek, Hellenistic, Buddhist, Hindu and other indigenous cultures were replaced by the coming of Islam and little influence remains in Afghanistan today. Along ancient trade routes, however, stone monuments of the once flourishing Buddhist culture did exist as reminders of the past. The two massive sandstone Buddhas of Bamyan, 35 and 53 meters high, overlooked the ancient route through Bamyan to Balkh and dated from the 3rd and 5th centuries. They survived until 2001, when they were destroyed by the Taliban. In this and other key places in Afghanistan, archaeologists have located frescoes, stucco decorations, statuary, and rare objects from as far away as China, Phoenicia, and Rome, which were crafted as early as the 2nd century and bear witness to the influence of these ancient civilizations upon Afghanistan. One of the early Buddhist schools, the Mahāsāṃghika-Lokottaravāda, were known to be prominent in the area of Bamiyan. The Chinese Buddhist monk Xuanzang visited a Lokottaravāda monastery in the 7th century CE, at Bamiyan, Afghanistan, and this monastery site has since been rediscovered by archaeologists. Birchbark and palm leaf manuscripts of texts in this monastery's collection, including Mahāyāna sūtras, have been discovered at the site, and these are now located in the Schøyen Collection. Some manuscripts are in the Gāndhārī language and Kharoṣṭhī script, while others are in Sanskrit and written in forms of the Gupta script. Manuscripts and fragments that have survived from this monastery's collection include well-known Buddhist texts such as the "Mahāparinirvāṇa Sūtra" (from the "Āgamas"), the "Diamond Sūtra" ("Vajracchedikā Prajñāpāramitā"), the "Medicine Buddha Sūtra", and the "Śrīmālādevī Siṃhanāda Sūtra". In 2010, reports stated that about 42 Buddhist relics have been discovered in the Logar Province of Afghanistan, which is south of Kabul. Some of these items date back to the 2nd century according to Archaeologists. The items included two Buddhist temples (Stupas), Buddha statues, frescos, silver and gold coins and precious beads.
https://en.wikipedia.org/wiki?curid=10344
Gravitational redshift In Einstein's general theory of relativity, the gravitational redshift is the phenomenon that clocks deeper in a gravitational well tick slower when observed from outside the well. More specifically the term refers to the shift of wavelength of a photon to longer wavelength (the red side in an optical spectrum) when observed from a point at a higher gravitational potential. In the latter case the 'clock' is the frequency of the photon and a lower frequency is the same as a longer ("redder") wavelength. The gravitational redshift is a simple consequence of Einstein's equivalence principle (that gravity and acceleration are equivalent) and was found by Einstein eight years before the full theory of relativity. Observing the gravitational redshift in the solar system is one of the classical tests of general relativity. Gravitational redshifts are an important effect in satellite-based navigation systems such as GPS. If the effects of general relativity were not taken into account, such systems would not work at all. Einstein's theory of general relativity incorporates the equivalence principle, which can be stated in various different ways. One such statement is that gravitational effects are locally undetectable for a free-falling observer. Therefore, in a laboratory experiment at the surface of the earth, all gravitational effects should be equivalent to the effects that would have been observed if the laboratory had been accelerating through outer space at "g". One consequence is a gravitational Doppler effect. If a light pulse is emitted at the floor of the laboratory, then a free-falling observer says that by the time it reaches the ceiling, the ceiling has accelerated away from it, and therefore when observed by a detector fixed to the ceiling, it will be observed to have been Doppler shifted toward the red end of the spectrum. This shift, which the free-falling observer considers to be a kinematical Doppler shift, is thought of by the laboratory observer as a gravitational redshift. Such an effect was verified in the 1959 Pound–Rebka experiment. In a case such as this, where the gravitational field is uniform, the change in wavelength is given by where formula_2 is the change in height. Since this prediction arises directly from the equivalence principle, it does not require any of the mathematical apparatus of general relativity, and its verification does not specifically support general relativity over any other theory that incorporates the equivalence principle. When the field is not uniform, the simplest and most useful case to consider is that of a spherically symmetric field. By Birkhoff's theorem, such a field is described in general relativity by the Schwarzschild metric, formula_3, where formula_4 is the clock time of an observer at distance "R" from the center, formula_5 is the time measured by an observer at infinity, formula_6 is the Schwarzschild radius formula_7, "..." represents terms that vanish if the observer is at rest, formula_8 is Newton's gravitational constant, formula_9 the mass of the gravitating body, and formula_10 the speed of light. The result is that frequencies and wavelengths are shifted according to the ratio where This can be related to the redshift parameter conventionally defined as formula_15. In the case where neither the emitter nor the observer is at infinity, the transitivity of Doppler shifts allows us to generalize the result to formula_16. The redshift formula for the frequency formula_17 is formula_18. When formula_19 is small, these results are consistent with the equation given above based on the equivalence principle. For an object compact enough to have an event horizon, the redshift is not defined for photons emitted inside the Schwarzschild radius, both because signals cannot escape from inside the horizon and because an object such as the emitter cannot be stationary inside the horizon, as was assumed above. Therefore, this formula only applies when formula_14 is larger than formula_6. When the photon is emitted at a distance equal to the Schwarzschild radius, the redshift will be "infinitely" large, and it will not escape to "any" finite distance from the Schwarzschild sphere. When the photon is emitted at an infinitely large distance, there is no redshift. In the Newtonian limit, i.e. when formula_14 is sufficiently large compared to the Schwarzschild radius formula_6, the redshift can be approximated as A number of experimenters initially claimed to have identified the effect using astronomical measurements, and the effect was considered to have been finally identified in the spectral lines of the star Sirius B by W.S. Adams in 1925. However, measurements by Adams have been criticized as being too low and these observations are now considered to be measurements of spectra that are unusable because of scattered light from the primary, Sirius A. The first accurate measurement of the gravitational redshift of a white dwarf was done by Popper in 1954, measuring a 21 km/s gravitational redshift of 40 Eridani B. The redshift of Sirius B was finally measured by Greenstein "et al." in 1971, obtaining the value for the gravitational redshift of 89±19 km/s, with more accurate measurements by the Hubble Space Telescope, showing 80.4±4.8 km/s. The effect is now considered to have been definitively verified by the experiments of Pound, Rebka and Snider between 1959 and 1965. The Pound–Rebka experiment of 1959 measured the gravitational redshift in spectral lines using a terrestrial 57Fe gamma source over a vertical height of 22.5 metres. This paper was the first determination of the gravitational redshift which used measurements of the change in wavelength of gamma-ray photons generated with the Mössbauer effect, which generates radiation with a very narrow line width. The accuracy of the gamma-ray measurements was typically 1%. An improved experiment was done by Pound and Snider in 1965, with an accuracy better than the 1% level. A very accurate gravitational redshift experiment was performed in 1976, where a hydrogen maser clock on a rocket was launched to a height of 10,000 km, and its rate compared with an identical clock on the ground. It tested the gravitational redshift to 0.007%. Later tests can be done with the Global Positioning System (GPS), which must account for the gravitational redshift in its timing system, and physicists have analyzed timing data from the GPS to confirm other tests. When the first satellite was launched, it showed the predicted shift of 38 microseconds per day. This rate of the discrepancy is sufficient to substantially impair the function of GPS within hours if not accounted for. An excellent account of the role played by general relativity in the design of GPS can be found in Ashby 2003. James W. Brault, a graduate student of Robert Dicke at Princeton University, measured the gravitational redshift of the sun using optical methods in 1962. In 2011 the group of Radek Wojtak of the Niels Bohr Institute at the University of Copenhagen collected data from 8000 galaxy clusters and found that the light coming from the cluster centers tended to be red-shifted compared to the cluster edges, confirming the energy loss due to gravity. The gravitational weakening of light from high-gravity stars was predicted by John Michell in 1783 and Pierre-Simon Laplace in 1796, using Isaac Newton's concept of light corpuscles (see: emission theory) and who predicted that some stars would have a gravity so strong that light would not be able to escape. The effect of gravity on light was then explored by Johann Georg von Soldner (1801), who calculated the amount of deflection of a light ray by the sun, arriving at the Newtonian answer which is half the value predicted by general relativity. All of this early work assumed that light could slow down and fall, which is inconsistent with the modern understanding of light waves. Once it became accepted that light was an electromagnetic wave, it was clear that the frequency of light should not change from place to place, since waves from a source with a fixed frequency keep the same frequency everywhere. One way around this conclusion would be if time itself were altered—if clocks at different points had different rates. This was precisely Einstein's conclusion in 1911. He considered an accelerating box, and noted that according to the special theory of relativity, the clock rate at the "bottom" of the box (the side away from the direction of acceleration) was slower than the clock rate at the "top" (the side toward the direction of acceleration). Nowadays, this can be easily shown in accelerated coordinates. The metric tensor in units where the speed of light is one is: and for an observer at a constant value of r, the rate at which a clock ticks, R(r), is the square root of the time coefficient, R(r)=r. The acceleration at position r is equal to the curvature of the hyperbola at fixed r, and like the curvature of the nested circles in polar coordinates, it is equal to 1/r. So at a fixed value of g, the fractional rate of change of the clock-rate, the percentage change in the ticking at the top of an accelerating box vs at the bottom, is: The rate is faster at larger values of R, away from the apparent direction of acceleration. The rate is zero at r=0, which is the location of the acceleration horizon. Using the equivalence principle, Einstein concluded that the same thing holds in any gravitational field, that the rate of clocks R at different heights was altered according to the gravitational field g. When g is slowly varying, it gives the fractional rate of change of the ticking rate. If the ticking rate is everywhere almost this same, the fractional rate of change is the same as the absolute rate of change, so that: Since the rate of clocks and the gravitational potential have the same derivative, they are the same up to a constant. The constant is chosen to make the clock rate at infinity equal to 1. Since the gravitational potential is zero at infinity: where the speed of light has been restored to make the gravitational potential dimensionless. The coefficient of the formula_29 in the metric tensor is the square of the clock rate, which for small values of the potential is given by keeping only the linear term: and the full metric tensor is: where again the C's have been restored. This expression is correct in the full theory of general relativity, to lowest order in the gravitational field, and ignoring the variation of the space-space and space-time components of the metric tensor, which only affect fast moving objects. Using this approximation, Einstein reproduced the incorrect Newtonian value for the deflection of light in 1909. But since a light beam is a fast moving object, the space-space components contribute too. After constructing the full theory of general relativity in 1916, Einstein solved for the space-space components in a post-Newtonian approximation and calculated the correct amount of light deflection – double the Newtonian value. Einstein's prediction was confirmed by many experiments, starting with Arthur Eddington's 1919 solar eclipse expedition. The changing rates of clocks allowed Einstein to conclude that light waves change frequency as they move, and the frequency/energy relationship for photons allowed him to see that this was best interpreted as the effect of the gravitational field on the mass–energy of the photon. To calculate the changes in frequency in a nearly static gravitational field, only the time component of the metric tensor is important, and the lowest order approximation is accurate enough for ordinary stars and planets, which are much bigger than their Schwarzschild radius.
https://en.wikipedia.org/wiki?curid=10346
Easter Rising The Easter Rising (), also known as the Easter Rebellion, was an armed insurrection in Ireland during Easter Week, April 1916. The Rising was launched by Irish republicans to end British rule in Ireland and establish an independent Irish Republic while the United Kingdom was fighting the First World War. It was the most significant uprising in Ireland since the rebellion of 1798 and the first armed action of the Irish revolutionary period. Sixteen of the Rising's leaders were executed in May 1916, but the insurrection, the nature of the executions, and subsequent political developments ultimately contributed to an increase in popular support for Irish independence. Organised by a seven-man Military Council of the Irish Republican Brotherhood, the Rising began on Easter Monday, 24 April 1916 and lasted for six days. Members of the Irish Volunteers, led by schoolmaster and Irish language activist Patrick Pearse, joined by the smaller Irish Citizen Army of James Connolly and 200 women of Cumann na mBan, seized strategically important buildings in Dublin and proclaimed an Irish Republic. The British Army brought in thousands of reinforcements as well as artillery and a gunboat. There was street fighting on the routes into the city centre, where the rebels slowed the British advance and inflicted many casualties. Elsewhere in Dublin, the fighting mainly consisted of sniping and long-range gun battles. The main rebel positions were gradually surrounded and bombarded with artillery. There were isolated actions in other parts of Ireland; Volunteer leader Eoin MacNeill had issued a countermand in a bid to halt the Rising, which greatly reduced the number of rebels who mobilised. With much greater numbers and heavier weapons, the British Army suppressed the Rising. Pearse agreed to an unconditional surrender on Saturday 29 April, although sporadic fighting continued briefly. After the surrender the country remained under martial law. About 3,500 people were taken prisoner by the British and 1,800 of them were sent to internment camps or prisons in Britain. Most of the leaders of the Rising were executed following courts-martial. The Rising brought physical force republicanism back to the forefront of Irish politics, which for nearly fifty years had been dominated by constitutional nationalism. Opposition to the British reaction to the Rising contributed to changes in public opinion and the move toward independence, as shown in the December 1918 election which was won by the Sinn Féin party, which convened the First Dáil and declared independence. Of the 485 people killed in the Easter Rising: 54 percent were civilians, 30 percent were British military and police and 16 percent were Irish rebels. More than 2,600 were wounded. Many of the civilians were killed or wounded by British artillery and machine guns or were mistaken for rebels. Others were caught in the crossfire in the crowded city. The shelling and resulting fires left parts of central Dublin in ruins. The Acts of Union 1800 united the Kingdom of Great Britain and the Kingdom of Ireland as the United Kingdom of Great Britain and Ireland, abolishing the Irish Parliament and giving Ireland representation in the British Parliament. From early on, many Irish nationalists opposed the union and the ensuing exploitation and impoverishment of the island, which led to a high level of depopulation. Opposition took various forms: constitutional (the Repeal Association; the Home Rule League), social (disestablishment of the Church of Ireland; the Land League) and revolutionary (Rebellion of 1848; Fenian Rising). The Irish Home Rule movement sought to achieve self-government for Ireland, within the United Kingdom. In 1886, the Irish Parliamentary Party (IPP) under Charles Stewart Parnell succeeded in having the First Home Rule Bill introduced in the British parliament, but it was defeated. The Second Home Rule Bill of 1893 was passed by the House of Commons but rejected by the House of Lords. After the fall of Parnell, younger and more radical nationalists became disillusioned with parliamentary politics and turned toward more extreme forms of separatism. The Gaelic Athletic Association, the Gaelic League and the cultural revival under W. B. Yeats and Augusta, Lady Gregory, together with the new political thinking of Arthur Griffith expressed in his newspaper "Sinn Féin" and organisations such as the National Council and the Sinn Féin League, led many Irish people to identify with the idea of an independent Gaelic Ireland. This was sometimes referred to by the generic term "Sinn Féin". The Third Home Rule Bill was introduced by British Liberal Prime Minister H. H. Asquith in 1912. Irish Unionists, who were overwhelmingly Protestants, opposed it, as they did not want to be ruled by a Catholic-dominated Irish government. Led by Sir Edward Carson and James Craig, they formed the Ulster Volunteers (UVF) in January 1913. In response, Irish nationalists formed a rival paramilitary group, the Irish Volunteers, in November 1913. The Irish Republican Brotherhood (IRB) was a driving force behind the Irish Volunteers and attempted to control it. Its leader was Eoin MacNeill, who was not an IRB member. The Irish Volunteers' stated goal was "to secure and to maintain the rights and liberties common to all the people of Ireland". It included people with a range of political views, and was open to "all able-bodied Irishmen without distinction of creed, politics or social group". Another militant group, the Irish Citizen Army, was formed by trade unionists as a result of the Dublin Lock-out of that year. When the Irish Volunteers smuggled rifles into Dublin, the British Army attempted to stop them and fired into a crowd of civilians. British Army officers then threatened to resign if they were ordered to take action against the UVF. By 1914, Ireland seemed to be on the brink of a civil war. This seemed to be averted in August of that year by the outbreak of the First World War, and Ireland's involvement in it. Nevertheless, on 18 September 1914 the Government of Ireland Act 1914 was enacted and placed on the statute book, but the Suspensory Act was passed at the same time, which deferred Irish Home Rule for one year, with powers for it to be suspended for further periods of six months so long as the war continued. It was widely believed at the time that the war would not last more than a few months. On 14 September 1915 an Order in Council was made under the Suspensory Act to suspend the Government of Ireland Act until 18 March 1916. Another such Order was made on 29 February 1916, suspending the Act for another six months. The Supreme Council of the IRB met on 5 September 1914, just over a month after the British government had declared war on Germany. At this meeting, they decided to stage an uprising before the war ended and to secure help from Germany. Responsibility for the planning of the rising was given to Tom Clarke and Seán MacDermott. The Irish Volunteers—the smaller of the two forces resulting from the September 1914 split over support for the British war effort—set up a "headquarters staff" that included Patrick Pearse as Director of Military Organisation, Joseph Plunkett as Director of Military Operations and Thomas MacDonagh as Director of Training. Éamonn Ceannt was later added as Director of Communications. In May 1915, Clarke and MacDermott established a Military Committee or Military Council within the IRB, consisting of Pearse, Plunkett and Ceannt, to draw up plans for a rising. Clarke and MacDermott joined it shortly after. The Military Council was able to promote its own policies and personnel independently of both the Volunteer Executive and the IRB Executive. Although the Volunteer and IRB leaders were not against a rising in principle, they were of the opinion that it was not opportune at that moment. Volunteer Chief-of-Staff Eoin MacNeill supported a rising only if the British government attempted to suppress the Volunteers or introduce conscription, and if such a rising had some chance of success. IRB President Denis McCullough and prominent IRB member Bulmer Hobson held similar views. The Military Council kept its plans secret, so as to prevent the British authorities learning of the plans, and to thwart those within the organisation who might try to stop the rising. IRB members held officer rank in the Volunteers throughout the country and took their orders from the Military Council, not from MacNeill. Shortly after the outbreak of World War I, Roger Casement and Clan na Gael leader John Devoy met the German ambassador to the United States, Johann Heinrich von Bernstorff, to discuss German backing for an uprising. Casement went to Germany and began negotiations with the German government and military. He persuaded the Germans to announce their support for Irish independence in November 1914. Casement also attempted to recruit an Irish Brigade, made up of Irish prisoners of war, which would be armed and sent to Ireland to join the uprising. However, only 56 men volunteered. Plunkett joined Casement in Germany the following year. Together, Plunkett and Casement presented a plan (the 'Ireland Report') in which a German expeditionary force would land on the west coast of Ireland, while a rising in Dublin diverted the British forces so that the Germans, with the help of local Volunteers, could secure the line of the River Shannon, before advancing on the capital. The German military rejected the plan, but agreed to ship arms and ammunition to the Volunteers. James Connolly—head of the Irish Citizen Army (ICA), a group of armed socialist trade union men and women—was unaware of the IRB's plans, and threatened to start a rebellion on his own if other parties failed to act. If they had done it alone, the IRB and the Volunteers would possibly have come to their aid; however, the IRB leaders met with Connolly in January 1916 and convinced him to join forces with them. They agreed that they would launch a rising together at Easter and made Connolly the sixth member of the Military Council. Thomas MacDonagh would later become the seventh and final member. The death of the old Fenian leader Jeremiah O'Donovan Rossa in New York in August 1915 was an opportunity to mount a spectacular demonstration. His body was sent to Ireland for burial in Glasnevin Cemetery, with the Volunteers in charge of arrangements. Huge crowds lined the route and gathered at the graveside. Pearse made a dramatic funeral oration, a rallying call to republicans, which ended with the words "Ireland unfree shall never be at peace". In early April, Pearse issued orders to the Irish Volunteers for three days of "parades and manoeuvres" beginning on Easter Sunday. He had the authority to do this, as the Volunteers' Director of Organisation. The idea was that IRB members within the organisation would know these were orders to begin the rising, while men such as MacNeill and the British authorities would take it at face value. On 9 April, the German Navy dispatched the SS "Libau" for County Kerry, disguised as the Norwegian ship "Aud". It was loaded with 20,000 rifles, one million rounds of ammunition, and explosives. Casement also left for Ireland aboard the German submarine "U-19". He was disappointed with the level of support offered by the Germans and he intended to stop or at least postpone the rising. On Wednesday 19 April, Alderman Tom Kelly, a Sinn Féin member of Dublin Corporation, read out at a meeting of the Corporation a document purportedly leaked from Dublin Castle, detailing plans by the British authorities to shortly arrest leaders of the Irish Volunteers, Sinn Féin and the Gaelic League, and occupy their premises. Although the British authorities said the "Castle Document" was fake, MacNeill ordered the Volunteers to prepare to resist. Unbeknownst to MacNeill, the document had been forged by the Military Council to persuade moderates of the need for their planned uprising. It was an edited version of a real document outlining British plans in the event of conscription. That same day, the Military Council informed senior Volunteer officers that the rising would begin on Easter Sunday. However, it chose not to inform the rank-and-file, or moderates such as MacNeill, until the last minute. The following day, MacNeill got wind that a rising was about to be launched and threatened to do everything he could to prevent it, short of informing the British. MacNeill was briefly persuaded to go along with some sort of action when Mac Diarmada revealed to him that a German arms shipment was about to land in County Kerry. MacNeill believed that when the British learned of the shipment they would immediately suppress the Volunteers, thus the Volunteers would be justified in taking defensive action, including the planned manoeuvres. The "Aud" and the "U-19" reached the coast of Kerry on Good Friday, 21 April. This was earlier than the Volunteers expected and so none were there to meet the vessels. The Royal Navy had known about the arms shipment and intercepted the "Aud", prompting the captain to scuttle the ship. Furthermore, Casement was captured shortly after he landed at Banna Strand. When MacNeill learned from Vol. Patrick Whelan that the arms shipment had been lost, he reverted to his original position. With the support of other leaders of like mind, notably Bulmer Hobson and The O'Rahilly, he issued a countermand to all Volunteers, cancelling all actions for Sunday. This countermanding order was relayed to Volunteer officers and printed in the Sunday morning newspapers. It succeeded only in delaying the rising for a day, although it greatly reduced the number of Volunteers who turned out. British Naval Intelligence had been aware of the arms shipment, Casement's return, and the Easter date for the rising through radio messages between Germany and its embassy in the United States that were intercepted by the Royal Navy and deciphered in Room 40 of the Admiralty. The information was passed to the Under-Secretary for Ireland, Sir Matthew Nathan, on 17 April, but without revealing its source, and Nathan was doubtful about its accuracy. When news reached Dublin of the capture of the "Aud" and the arrest of Casement, Nathan conferred with the Lord Lieutenant, Lord Wimborne. Nathan proposed to raid Liberty Hall, headquarters of the Citizen Army, and Volunteer properties at Father Matthew Park and at Kimmage, but Wimborne insisted on wholesale arrests of the leaders. It was decided to postpone action until after Easter Monday, and in the meantime Nathan telegraphed the Chief Secretary, Augustine Birrell, in London seeking his approval. By the time Birrell cabled his reply authorising the action, at noon on Monday 24 April 1916, the Rising had already begun. On the morning of Easter Sunday, 23 April, the Military Council met at Liberty Hall to discuss what to do in light of MacNeill's countermanding order. They decided that the Rising would go ahead the following day, Easter Monday, and that the Irish Volunteers and Irish Citizen Army would go into action as the 'Army of the Irish Republic'. They elected Pearse as president of the Irish Republic, and also as Commander-in-Chief of the army; Connolly became Commandant of the Dublin Brigade. Messengers were then sent to all units informing them of the new orders. On the morning of Monday 24 April, about 1,200 members of the Irish Volunteers and Irish Citizen Army mustered at several locations in central Dublin. Among them were members of the all-female Cumann na mBan. Some wore Irish Volunteer and Citizen Army uniforms, while others wore civilian clothes with a yellow Irish Volunteer armband, military hats, and bandoliers. They were armed mostly with rifles (especially 1871 Mausers), but also with shotguns, revolvers, a few Mauser C96 semi-automatic pistols, and grenades. The number of Volunteers who mobilised was much smaller than expected. This was due to MacNeill's countermanding order, and the fact that the new orders had been sent so soon beforehand. However, several hundred Volunteers joined the Rising after it began. Shortly before midday, the rebels began to seize important sites in central Dublin. The rebels' plan was to hold Dublin city centre. This was a large, oval-shaped area bounded by two canals: the Grand to the south and the Royal to the north, with the River Liffey running through the middle. On the southern and western edges of this district were five British Army barracks. Most of the rebels' positions had been chosen to defend against counter-attacks from these barracks. The rebels took the positions with ease. Civilians were evacuated and policemen were ejected or taken prisoner. Windows and doors were barricaded, food and supplies were secured, and first aid posts were set up. Barricades were erected on the streets to hinder British Army movement. A joint force of about 400 Volunteers and Citizen Army gathered at Liberty Hall under the command of Commandant James Connolly. This was the headquarters battalion, and it also included Commander-in-Chief Patrick Pearse, as well as Tom Clarke, Seán MacDermott and Joseph Plunkett. They marched to the General Post Office (GPO) on O'Connell Street, Dublin's main thoroughfare, occupied the building and hoisted two republican flags. Pearse stood outside and read the Proclamation of the Irish Republic. Copies of the Proclamation were also pasted on walls and handed out to bystanders by Volunteers and newsboys. The GPO would be the rebels' headquarters for most of the Rising. Volunteers from the GPO also occupied other buildings on the street, including buildings overlooking O'Connell Bridge. They took over a wireless telegraph station and sent out a radio broadcast in Morse code, announcing that an Irish Republic had been declared. This was the first radio broadcast in Ireland. Elsewhere, some of the headquarters battalion under Michael Mallin occupied St Stephen's Green, where they dug trenches and barricaded the surrounding roads. The 1st battalion, under Edward 'Ned' Daly, occupied the Four Courts and surrounding buildings, while a company under Seán Heuston occupied the Mendicity Institution, across the River Liffey from the Four Courts. The 2nd battalion, under Thomas MacDonagh, occupied Jacob's biscuit factory. The 3rd battalion, under Éamon de Valera, occupied Boland's Mill and surrounding buildings. The 4th battalion, under Éamonn Ceannt, occupied the South Dublin Union and the distillery on Marrowbone Lane. From each of these garrisons, small units of rebels established outposts in the surrounding area. The rebels also attempted to cut transport and communication links. As well as erecting roadblocks, they took control of various bridges and cut telephone and telegraph wires. Westland Row and Harcourt Street railway stations were occupied, though the latter only briefly. The railway line was cut at Fairview and the line was damaged by bombs at Amiens Street, Broadstone, Kingsbridge and Lansdowne Road. Around midday, a small team of Volunteers and Fianna Éireann members swiftly captured the Magazine Fort in the Phoenix Park and disarmed the guards. The goal was to seize weapons and blow up the ammunition store to signal that the Rising had begun. They seized weapons and planted explosives, but the blast was not big enough to be heard across the city. The 23-year-old son of the fort's commander was fatally shot when he ran to raise the alarm. A contingent under Seán Connolly occupied Dublin City Hall and adjacent buildings. They attempted to seize neighbouring Dublin Castle, the heart of British rule in Ireland. As they approached the gate a lone and unarmed police sentry, James O'Brien, attempted to stop them and was shot dead by Connolly. According to some accounts, he was the first casualty of the Rising. The rebels overpowered the soldiers in the guardroom, but failed to press further. The British Army's chief intelligence officer, Major Ivon Price, fired on the rebels while the Under-Secretary for Ireland, Sir Matthew Nathan, helped shut the castle gates. Unbeknownst to the rebels, the Castle was lightly guarded and could have been taken with ease. The rebels instead laid siege to the Castle from City Hall. Fierce fighting erupted there after British reinforcements arrived. The rebels on the roof exchanged fire with soldiers on the street. Seán Connolly was shot dead by a sniper, becoming the first rebel casualty. By the following morning, British forces had re-captured City Hall and taken the rebels prisoner. The rebels did not attempt to take some other key locations, notably Trinity College, in the heart of the city centre and defended by only a handful of armed unionist students. Failure to capture the telephone exchange in Crown Alley left communications in the hands of Government with GPO staff quickly repairing telephone wires that had been cut by the rebels. The failure to occupy strategic locations was attributed to lack of manpower. In at least two incidents, at Jacob's and Stephen's Green, the Volunteers and Citizen Army shot dead civilians trying to attack them or dismantle their barricades. Elsewhere, they hit civilians with their rifle butts to drive them off. The British military were caught totally unprepared by the rebellion and their response of the first day was generally un-coordinated. Two troops of British cavalry were sent to investigate what was happening. They took fire and casualties from rebel forces at the GPO and at the Four Courts. As one troop passed Nelson's Pillar, the rebels opened fire from the GPO, killing three cavalrymen and two horses and fatally wounding a fourth man. The cavalrymen retreated and were withdrawn to barracks. On Mount Street, a group of Volunteer Training Corps men stumbled upon the rebel position and four were killed before they reached Beggars Bush Barracks. The only substantial combat of the first day of the Rising took place at the South Dublin Union where a piquet from the Royal Irish Regiment encountered an outpost of Éamonn Ceannt's force at the northwestern corner of the South Dublin Union. The British troops, after taking some casualties, managed to regroup and launch several assaults on the position before they forced their way inside and the small rebel force in the tin huts at the eastern end of the Union surrendered. However, the Union complex as a whole remained in rebel hands. A nurse in uniform, Margaret Keogh, was shot dead by British soldiers at the Union. She is believed to have been the first civilian killed in the Rising. Three unarmed Dublin Metropolitan Police were shot dead on the first day of the Rising and their Commissioner pulled them off the streets. Partly as a result of the police withdrawal, a wave of looting broke out in the city centre, especially in the area of O'Connell Street (still officially called "Sackville Street" at the time). Lord Wimborne, the Lord Lieutenant, declared martial law on Tuesday evening and handed over civil power to Brigadier-General William Lowe. British forces initially put their efforts into securing the approaches to Dublin Castle and isolating the rebel headquarters, which they believed was in Liberty Hall. The British commander, Lowe, worked slowly, unsure of the size of the force he was up against, and with only 1,269 troops in the city when he arrived from the Curragh Camp in the early hours of Tuesday 25 April. City Hall was taken from the rebel unit that had attacked Dublin Castle on Tuesday morning. In the early hours of Tuesday, 120 British soldiers, with machine-guns, occupied two buildings overlooking St Stephen's Green: the Shelbourne Hotel and United Services Club. At dawn they opened fire on the Citizen Army occupying the green. The rebels returned fire, but were forced to retreat to the Royal College of Surgeons building. They remained there for the rest of the week, exchanging fire with British forces. Fighting erupted along the northern edge of the city centre on Tuesday afternoon. In the northeast, British troops left Amiens Street railway station in an armoured train, to secure and repair a section of damaged tracks. They were attacked by rebels who had taken up position at Annesley Bridge. After a two-hour battle, the British were forced to retreat and several soldiers were captured. At Phibsborough, in the northwest, rebels had occupied buildings and erected barricades at junctions on the North Circular Road. The British summoned 18-pounder field artillery from Athlone and shelled the rebel positions, destroying the barricades. After a fierce firefight, the rebels withdrew. That afternoon Pearse walked out into O'Connell Street with a small escort and stood in front of Nelson's Pillar. As a large crowd gathered, he read out a ',' calling on them to support the Rising. The rebels had failed to take either of Dublin's two main railway stations or either of its ports, at Dublin Port and Kingstown. As a result, during the following week, the British were able to bring in thousands of reinforcements from Britain and from their garrisons at the Curragh and Belfast. By the end of the week, British strength stood at over 16,000 men. Their firepower was provided by field artillery which they positioned on the Northside of the city at Phibsborough and at Trinity College, and by the patrol vessel "Helga", which sailed up the Liffey, having been summoned from the port at Kingstown. On Wednesday, 26 April, the guns at Trinity College and "Helga" shelled Liberty Hall, and the Trinity College guns then began firing at rebel positions, first at Boland's Mill and then in O'Connell Street. Some rebel commanders, particularly James Connolly, did not believe that the British would shell the 'second city' of the British Empire. The principal rebel positions at the GPO, the Four Courts, Jacob's Factory and Boland's Mill saw little action. The British surrounded and bombarded them rather than assault them directly. One Volunteer in the GPO recalled, "we did practically no shooting as there was no target". However, where the insurgents dominated the routes by which the British tried to funnel reinforcements into the city, there was fierce fighting. At 5:25PM Volunteers Eamon Martin, Garry Holohan, Robert Beggs, Sean Cody, Dinny O'Callaghan, Charles Shelley, Peadar Breslin and five others attempted to occupy Broadstone railway station on Church Street, the attack was unsuccessful and Martin was injured. On Wednesday morning, hundreds of British troops encircled the Mendicity Institute, which was occupied by 26 Volunteers under Seán Heuston. British troops advanced on the building, supported by snipers and machine gun fire, but the Volunteers put up stiff resistance. Eventually, the troops got close enough to hurl grenades into the building, some of which the rebels threw back. Exhausted and almost out of ammunition, Heuston's men became the first rebel position to surrender. Heuston had been ordered to hold his position for a few hours, to delay the British, but had held on for three days. Reinforcements were sent to Dublin from Britain, and disembarked at Kingstown on the morning of Wednesday 26 April. Heavy fighting occurred at the rebel-held positions around the Grand Canal as these troops advanced towards Dublin. More than 1,000 Sherwood Foresters were repeatedly caught in a cross-fire trying to cross the canal at Mount Street Bridge. Seventeen Volunteers were able to severely disrupt the British advance, killing or wounding 240 men. Despite there being alternative routes across the canal nearby, General Lowe ordered repeated frontal assaults on the Mount Street position. The British eventually took the position, which had not been reinforced by the nearby rebel garrison at Boland's Mills, on Thursday, but the fighting there inflicted up to two-thirds of their casualties for the entire week for a cost of just four dead Volunteers. It had taken nearly nine hours for the British to advance . On Wednesday Linenhall Barracks on Constitution Hill was burnt down under the orders of Commandant Edward Daly to prevent its reoccupation by the British army. The rebel position at the South Dublin Union (site of the present day St. James's Hospital) and Marrowbone Lane, further west along the canal, also inflicted heavy losses on British troops. The South Dublin Union was a large complex of buildings and there was vicious fighting around and inside the buildings. Cathal Brugha, a rebel officer, distinguished himself in this action and was badly wounded. By the end of the week, the British had taken some of the buildings in the Union, but others remained in rebel hands. British troops also took casualties in unsuccessful frontal assaults on the Marrowbone Lane Distillery. The third major scene of fighting during the week was in the area of North King Street, north of the Four Courts. The rebels had established strong outposts in the area, occupying numerous small buildings and barricading the streets. From Thursday to Saturday, the British made repeated attempts to take the area, in what was some of the fiercest fighting of the Rising. As the troops moved in, the rebels continually opened fire from windows and behind chimneys and barricades. At one point, a platoon led by Major Sheppard made a bayonet charge on one of the barricades, but was cut down by rebel fire. The British employed machine guns and attempted to avoid direct fire by using makeshift armoured trucks, and by mouse-holing through the inside walls of terraced houses to get near the rebel positions. By the time of the rebel headquarters' surrender on Saturday, the South Staffordshire Regiment under Colonel Taylor had advanced only down the street at a cost of 11 dead and 28 wounded. The enraged troops broke into the houses along the street and shot or bayoneted 15 unarmed male civilians whom they accused of being rebel fighters. Elsewhere, at Portobello Barracks, an officer named Bowen Colthurst summarily executed six civilians, including the pacifist nationalist activist, Francis Sheehy-Skeffington. These instances of British troops killing Irish civilians would later be highly controversial in Ireland. The headquarters garrison at the GPO, after days of shelling, was forced to abandon their headquarters when fire caused by the shells spread to the GPO. Connolly had been incapacitated by a bullet wound to the ankle and had passed command on to Pearse. The O'Rahilly was killed in a sortie from the GPO. They tunnelled through the walls of the neighbouring buildings in order to evacuate the Post Office without coming under fire and took up a new position in 16 Moore Street. The young Seán McLoughlin was given military command and planned a break out, but Pearse realised this plan would lead to further loss of civilian life. On Saturday 29 April, from this new headquarters, Pearse issued an order for all companies to surrender. Pearse surrendered unconditionally to Brigadier-General Lowe. The surrender document read: The other posts surrendered only after Pearse's surrender order, carried by nurse Elizabeth O'Farrell, reached them. Sporadic fighting therefore continued until Sunday, when word of the surrender was got to the other rebel garrisons. Command of British forces had passed from Lowe to General John Maxwell, who arrived in Dublin just in time to take the surrender. Maxwell was made temporary military governor of Ireland. Irish Volunteer units mobilised on Easter Sunday in several places outside of Dublin, but because of Eoin MacNeill's countermanding order, most of them returned home without fighting. In addition, because of the interception of the German arms aboard the "Aud", the provincial Volunteer units were very poorly armed. In the south, around 1,200 Volunteers mustered in Cork, under Tomás Mac Curtain, on the Sunday, but they dispersed on Wednesday after receiving nine contradictory orders by dispatch from the Volunteer leadership in Dublin. At their Sheares Street headquarters, some of the Volunteers engaged in a standoff with British forces. Much to the anger of many Volunteers, MacCurtain, under pressure from Catholic clergy, agreed to surrender his men's arms to the British. The only violence in Cork occurred when the RIC attempted to raid the home of the Kent family. The Kent brothers, who were Volunteers, engaged in a three-hour firefight with the RIC. An RIC officer and one of the brothers were killed, while another brother was later executed. In the north, Volunteer companies were mobilised in County Tyrone at Coalisland (including 132 men from Belfast led by IRB President Dennis McCullough) and Carrickmore, under the leadership of Patrick McCartan. They also mobilised at Creeslough, County Donegal under Daniel Kelly and James McNulty. However, in part because of the confusion caused by the countermanding order, the Volunteers in these locations dispersed without fighting. In Fingal (or north County Dublin), about 60 Volunteers mobilised near Swords. They belonged to the 5th Battalion of the Dublin Brigade (also known as the Fingal Battalion), and were led by Thomas Ashe and his second in command, Richard Mulcahy. Unlike the rebels elsewhere, the Fingal Battalion successfully employed guerrilla tactics. They set up camp and Ashe split the battalion into four sections: three would undertake operations while the fourth was kept in reserve, guarding camp and foraging for food. The Volunteers moved against the RIC barracks in Swords, Donabate and Garristown, forcing the RIC to surrender and seizing all the weapons. They also damaged railway lines and cut telegraph wires. The railway line at Blanchardstown was bombed to prevent a troop train reaching Dublin. This derailed a cattle train, which had been sent ahead of the troop train. The only large-scale engagement of the Rising, outside Dublin city, was at Ashbourne. On Friday, about 35 Fingal Volunteers surrounded the Ashbourne RIC barracks and called on it to surrender, but the RIC responded with a volley of gunfire. A firefight followed, and the RIC surrendered after the Volunteers attacked the building with a homemade grenade. Before the surrender could be taken, up to sixty RIC men arrived in a convoy, sparking a five-hour gun battle, in which eight RIC men were killed and 18 wounded. Two Volunteers were also killed and five wounded, and a civilian was fatally shot. The RIC surrendered and were disarmed. Ashe let them go after warning them not to fight against the Irish Republic again. Ashe's men camped at Kilsalaghan near Dublin until they received orders to surrender on Saturday. The Fingal Battalion's tactics during the Rising foreshadowed those of the IRA during the War of Independence that followed. Volunteer contingents also mobilised nearby in counties Meath and Louth, but proved unable to link up with the North Dublin unit until after it had surrendered. In County Louth, Volunteers shot dead an RIC man near the village of Castlebellingham on 24 April, in an incident in which 15 RIC men were also taken prisoner. In County Wexford, 100–200 Volunteers—led by Robert Brennan, Séamus Doyle and Seán Etchingham—took over the town of Enniscorthy on Thursday 27 April until Sunday. Volunteer officer Paul Galligan had cycled 200 km from rebel headquarters in Dublin with orders to mobilise. They blocked all roads into the town and made a brief attack on the RIC barracks, but chose to blockade it rather than attempt to capture it. They flew the tricolour over the Athenaeum building, which they had made their headquarters, and paraded uniformed in the streets. They also occupied Vinegar Hill, where the United Irishmen had made a last stand in the 1798 rebellion. The public largely supported the rebels and many local men offered to join them. By Saturday, up to 1,000 rebels had been mobilised, and a detachment was sent to occupy the nearby village of Ferns. In Wexford, the British assembled a column of 1,000 soldiers (including the Connaught Rangers), two field guns and a 4.7 inch naval gun on a makeshift armoured train. On Sunday, the British sent messengers to Enniscorthy, informing the rebels of Pearse's surrender order. However, the Volunteer officers were skeptical. Two of them were escorted by the British to Arbour Hill Prison, where Pearse confirmed the surrender order. In County Galway, 600–700 Volunteers mobilised on Tuesday under Liam Mellows. His plan was to "bottle up the British garrison and divert the British from concentrating on Dublin". However, his men were poorly armed, with only 25 rifles, 60 revolvers, 300 shotguns and some homemade grenades – many of them only had pikes. Most of the action took place in a rural area to the east of Galway city. They made unsuccessful attacks on the RIC barracks at Clarinbridge and Oranmore, captured several officers, and bombed a bridge and railway line, before taking up position near Athenry. There was also a skirmish between rebels and an RIC mobile patrol at Carnmore crossroads. A constable, Patrick Whelan, was shot dead after he had called to the rebels: "Surrender, boys, I know ye all". On Wednesday, arrived in Galway Bay and shelled the countryside on the northeastern edge of Galway. The rebels retreated southeast to Moyode, an abandoned country house and estate. From here they set up lookout posts and sent out scouting parties. On Friday, landed 200 Royal Marines and began shelling the countryside near the rebel position. The rebels retreated further south to Limepark, another abandoned country house. Deeming the situation to be hopeless, they dispersed on Saturday morning. Many went home and were arrested following the rising, while others, including Mellows, went "on the run". By the time British reinforcements arrived in the west, the rising there had already disintegrated. In County Limerick, 300 Irish Volunteers assembled at Glenquin Castle near Killeedy, but they did not take any military action. In County Clare, Micheal Brennan reputedly marched with 100 Volunteers (from Meelick, Oatfield, and Cratloe) to the River Shannon on Easter Monday to await orders from the Rising leaders in Dublin, and weapons from the expected Casement shipment. However, neither arrived and no actions were taken. The Easter Rising resulted in at least 485 deaths, according to the Glasnevin Trust. Of those killed: More than 2,600 were wounded; including at least 2,200 civilians and rebels, at least 370 British soldiers and 29 policemen. All 16 police fatalities and 22 of the British soldiers killed were Irishmen. About 40 of those killed were children (under 17 years old), four of whom were members of the rebel forces. The number of casualties each day steadily rose, with 55 killed on Monday and 78 killed on Saturday. The British Army suffered their biggest losses in the Battle of Mount Street Bridge on Wednesday, when at least 30 soldiers were killed. The rebels also suffered their biggest losses on that day. The RIC suffered most of their casualties in the Battle of Ashbourne on Friday. The majority of the casualties, both killed and wounded, were civilians. Most of the civilian casualties and most of the casualties overall were caused by the British Army. This was due to the British using artillery, incendiary shells and heavy machine guns in built-up areas, as well as their "inability to discern rebels from civilians". One Royal Irish Regiment officer recalled, "they regarded, not unreasonably, everyone they saw as an enemy, and fired at anything that moved". Many other civilians were killed when caught in the crossfire. Both sides, British and rebel, also shot civilians deliberately on occasion; for not obeying orders (such as to stop at checkpoints), for assaulting or attempting to hinder them, and for looting. There were also instances of British troops killing unarmed civilians out of revenge or frustration: notably in the North King Street Massacre, where fifteen were killed, and at Portobello Barracks, where six were shot. Furthermore, there were incidents of friendly fire. On 29 April, the Royal Dublin Fusiliers under Company Quartermaster Sergeant Robert Flood shot dead two British officers and two Irish civilian employees of the Guinness brewery after he decided they were rebels. Flood was court-martialled for murder but acquitted. According to historian Fearghal McGarry, the rebels attempted to avoid needless bloodshed. Desmond Ryan stated that Volunteers were told "no firing was to take place except under orders or to repel attack". Aside from the engagement at Ashbourne, policemen and unarmed soldiers were not systematically targeted, and a large group of policemen was allowed to stand at Nelson's Pillar throughout Monday. McGarry writes that the Irish Citizen Army "were more ruthless than Volunteers when it came to shooting policemen" and attributes this to the "acrimonious legacy" of the Dublin Lock-out. The vast majority of the Irish casualties were buried in Glasnevin Cemetery in the aftermath of the fighting. British families came to Dublin Castle in May 1916 to reclaim the bodies of British soldiers, and funerals were arranged. Soldiers whose bodies were not claimed were given military funerals in Grangegorman Military Cemetery. General Maxwell quickly signalled his intention "to arrest all dangerous Sinn Feiners", including "those who have taken an active part in the movement although not in the present rebellion", reflecting the popular belief that Sinn Féin, a separatist organisation that was neither militant nor republican, was behind the Rising. A total of 3,430 men and 79 women were arrested, including 425 people for looting. A series of courts-martial began on 2 May, in which 187 people were tried, most of them at Richmond Barracks. The president of the courts-martial was Charles Blackader. Controversially, Maxwell decided that the courts-martial would be held in secret and without a defence, which Crown law officers later ruled to have been illegal. Some of those who conducted the trials had commanded British troops involved in suppressing the Rising, a conflict of interest that the Military Manual prohibited. Only one of those tried by courts-martial was a woman, Constance Markievicz, who was also the only woman to be kept in solitary confinement. Ninety were sentenced to death. Fifteen of those (including all seven signatories of the Proclamation) had their sentences confirmed by Maxwell and fourteen were executed by firing squad at Kilmainham Gaol between 3 and 12 May. Among them was the seriously wounded Connolly, who was shot while tied to a chair because of his shattered ankle. Maxwell stated that only the "ringleaders" and those proven to have committed "coldblooded murder" would be executed. However, the evidence presented was weak, and some of those executed were not leaders and did not kill anyone: Willie Pearse described himself as "a personal attaché to my brother, Patrick Pearse"; John MacBride had not even been aware of the Rising until it began, but had fought against the British in the Boer War fifteen years before; Thomas Kent did not come out at all—he was executed for the killing of a police officer during the raid on his house the week after the Rising. The most prominent leader to escape execution was Éamon de Valera, Commandant of the 3rd Battalion, who did so partly because of his American birth. Most of the executions took place over a ten-day period: As the executions went on, the Irish public grew increasingly hostile towards the British and sympathetic to the rebels. After the first three executions, John Redmond, leader of the moderate Irish Parliamentary Party, said in the British Parliament that the rising "happily, seems to be over. It has been dealt with with firmness, which was not only right, but it was the duty of the Government to so deal with it". However, he urged the Government "not to show undue hardship or severity to the great masses of those who are implicated". As the executions continued, Redmond pleaded with Prime Minister H. H. Asquith to stop them, warning that "if more executions take place in Ireland, the position will become impossible for any constitutional party". Ulster Unionist Party leader Edward Carson expressed similar views. Redmond's deputy, John Dillon, made an impassioned speech in parliament, saying "thousands of people […] who ten days ago were bitterly opposed to the whole of the Sinn Fein movement and to the rebellion, are now becoming infuriated against the Government on account of these executions". He said "it is not murderers who are being executed; it is insurgents who have fought a clean fight, a brave fight, however misguided". Dillon was heckled by English MPs. The British Government itself had also become concerned at the reaction to the executions, and at the way the courts-martial were being carried out. Asquith had warned Maxwell that "a large number of executions would […] sow the seeds of lasting trouble in Ireland". After Connolly's execution, Maxwell bowed to pressure and had the other death sentences commuted to penal servitude. Most of the people arrested were subsequently released, however under Regulation 14B of the Defence of the Realm Act 1914 1,836 men were interned at internment camps and prisons in England and Wales. Many of them, like Arthur Griffith, had little or nothing to do with the Rising. Camps such as Frongoch internment camp became "Universities of Revolution" where future leaders including Michael Collins, Terence McSwiney and J. J. O'Connell began to plan the coming struggle for independence. Sir Roger Casement was tried in London for high treason and hanged at Pentonville Prison on 3 August. After the Rising, claims of atrocities carried out by British troops began to emerge. Although they did not receive as much attention as the executions, they sparked outrage among the Irish public and were raised by Irish MPs in Parliament. One incident was the 'Portobello killings'. On Tuesday 25 April, Dubliner Francis Sheehy-Skeffington, a pacifist nationalist activist, had been arrested by British soldiers. Captain John Bowen-Colthurst then took him with a British raiding party as a hostage and human shield. On Rathmines Road he stopped a boy named James Coade, whom he shot dead. His troops then destroyed a tobacconist's shop with grenades and seized journalists Thomas Dickson and Patrick MacIntyre. The next morning, Colthurst had Skeffington and the two journalists shot by firing squad in Portobello Barracks. The bodies were then buried there. Later that day he shot a Labour Party councillor, Richard O'Carroll. When Major Sir Francis Vane learned of the killings he telephoned his superiors in Dublin Castle, but no action was taken. Vane informed Herbert Kitchener, who told General Maxwell to arrest Colthurst, but Maxwell refused. Colthurst was eventually arrested and court-martialled in June. He was found guilty of murder but insane, and detained for twenty months at Broadmoor. Public and political pressure led to a public inquiry, which reached similar conclusions. Major Vane was discharged "owing to his action in the Skeffington murder case". The other incident was the 'North King Street Massacre'. On the night of 28–29 April, British soldiers of the South Staffordshire Regiment, under Colonel Henry Taylor, had burst into houses on North King Street and killed 15 male civilians whom they accused of being rebels. The soldiers shot or bayoneted the victims, then secretly buried some of them in cellars or back yards after robbing them. The area saw some of the fiercest fighting of the Rising and the British had taken heavy casualties for little gain. General Maxwell attempted to excuse the killings and argued that the rebels were ultimately responsible. He claimed that "the rebels wore no uniform" and that the people of North King Street were rebel sympathisers. Maxwell concluded that such incidents "are absolutely unavoidable in such a business as this" and that "under the circumstance the troops [...] behaved with the greatest restraint". A private brief, prepared for the Prime Minister, said the soldiers "had orders not to take any prisoners" but took it to mean they were to shoot any suspected rebel. The City Coroner's inquest found that soldiers had killed "unarmed and unoffending" residents. The military court of inquiry ruled that no specific soldiers could be held responsible, and no action was taken. These killings, and the British response to them, helped sway Irish public opinion against the British. A Royal Commission was set up to enquire into the causes of the Rising. It began hearings on 18 May under the chairmanship of Lord Hardinge of Penshurst. The Commission heard evidence from Sir Matthew Nathan, Augustine Birrell, Lord Wimborne, Sir Neville Chamberlain (Inspector-General of the Royal Irish Constabulary), General Lovick Friend, Major Ivor Price of Military Intelligence and others. The report, published on 26 June, was critical of the Dublin administration, saying that "Ireland for several years had been administered on the principle that it was safer and more expedient to leave the law in abeyance if collision with any faction of the Irish people could thereby be avoided." Birrell and Nathan had resigned immediately after the Rising. Wimborne had also reluctantly resigned, recalled to London by Lloyd George, but was re-appointed in late 1917. Chamberlain resigned soon after. At first, many Dubliners were bewildered by the outbreak of the Rising. James Stephens, who was in Dublin during the week, thought, "None of these people were prepared for Insurrection. The thing had been sprung on them so suddenly they were unable to take sides." There was great hostility towards the Volunteers in some parts of the city. Historian Keith Jeffery noted that most of the opposition came from people whose relatives were in the British Army and who depended on their army allowances. Those most openly hostile to the Volunteers were the "separation women" (so-called because they were paid "separation money" by the British government), whose husbands and sons were fighting in the British Army in the First World War. There was also hostility from unionists. Supporters of the Irish Parliamentary Party also felt the rebellion was a betrayal of their party. When occupying positions in the South Dublin Union and Jacob's factory, the rebels got involved in physical confrontations with civilians who tried to tear down the rebel barricades and prevent them taking over buildings. The Volunteers shot and clubbed a number of civilians who assaulted them or tried to dismantle their barricades. That the Rising resulted in a great deal of death and destruction, as well as disrupting food supplies, also contributed to the antagonism toward the rebels. After the surrender, the Volunteers were hissed at, pelted with refuse, and denounced as "murderers" and "starvers of the people". Volunteer Robert Holland for example remembered being "subjected to very ugly remarks and cat-calls from the poorer classes" as they marched to surrender. He also reported being abused by people he knew as he was marched through the Kilmainham area into captivity and said the British troops saved them from being manhandled by the crowd. However, some Dubliners expressed support for the rebels. Canadian journalist and writer Frederick Arthur McKenzie wrote that in poorer areas, "there was a vast amount of sympathy with the rebels, particularly after the rebels were defeated". He wrote of crowds cheering a column of rebel prisoners as it passed, with one woman remarking "Shure, we cheer them. Why shouldn't we? Aren't they our own flesh and blood?". At Boland's Mill, the defeated rebels were met with a large crowd, "many weeping and expressing sympathy and sorrow, all of them friendly and kind". Other onlookers were sympathetic but watched in silence. Christopher M. Kennedy notes that "those who sympathised with the rebels would, out of fear for their own safety, keep their opinions to themselves". Áine Ceannt witnessed British soldiers arresting a woman who cheered the captured rebels. An RIC District Inspector's report stated: "Martial law, of course, prevents any expression of it; but a strong undercurrent of disloyalty exists". Thomas Johnson, the Labour leader, thought there was "no sign of sympathy for the rebels, but general admiration for their courage and strategy". The aftermath of the Rising, and in particular the British reaction to it, helped sway a large section of Irish nationalist opinion away from hostility or ambivalence and towards support for the rebels of Easter 1916. Dublin businessman and Quaker James G. Douglas, for example, hitherto a Home Ruler, wrote that his political outlook changed radically during the course of the Rising because of the British military occupation of the city and that he became convinced that parliamentary methods would not be enough to remove the British presence. A meeting called by Count Plunkett on 19 April 1917 led to the formation of a broad political movement under the banner of Sinn Féin which was formalised at the Sinn Féin Ard Fheis of 25 October 1917. The Conscription Crisis of 1918 further intensified public support for Sinn Féin before the general elections to the British Parliament on 14 December 1918, which resulted in a landslide victory for Sinn Féin, winning 73 seats out of 105, whose MPs gathered in Dublin on 21 January 1919 to form Dáil Éireann and adopt the Declaration of Independence. Shortly after the Easter Rising, poet Francis Ledwidge wrote ""O’Connell Street"" and ""Lament for the Poets of 1916"", which both describe his sense of loss and an expression of holding the same "dreams," as the Easter Rising's Irish Republicans. He would also go on to write "lament for Thomas MacDonagh" for his fallen friend and fellow Irish Volunteer. A few months after the Easter Rising, W. B. Yeats commemorated some of the fallen figures of the Irish Republican movement, as well as his torn emotions regarding these events, in the poem "Easter, 1916". Some of the survivors of the Rising went on to become leaders of the independent Irish state. Those who were executed were venerated by many as martyrs; their graves in Dublin's former military prison of Arbour Hill became a national monument and the Proclamation text was taught in schools. An annual commemorative military parade was held each year on Easter Sunday. In 1935, Éamon de Valera unveiled a statue of the mythical Irish hero Cú Chulainn, sculpted by Oliver Sheppard, at the General Post Office as part of the Rising commemorations that year – it is often seen to be an important symbol of martyrdom in remembrance of the 1916 rebels. Memorials to the heroes of the Rising are to be found in other Irish cities, such as Limerick. The parades culminated in a huge national celebration on the 50th anniversary of the Rising in 1966. Medals were issued by the government to survivors who took part in the rising at the event. RTÉ, the Irish national broadcaster, as one of its first major undertakings made a series of commemorative programmes for the 1966 anniversary of the Rising. Roibéárd Ó Faracháin, head of programming said, "While still seeking historical truth, the emphasis will be on homage, on salutation." At the same time, CIÉ, the Republic of Ireland's railway operator, renamed several of its major stations after republicans who played key roles in the Easter Rising. Ireland's first commemorative coin was also issued in 1966 to pay tribute to the Easter Rising. It was valued at 10 shillings, therefore having the highest denomination of any pre-decimal coin issued by the country. The coin featured a bust of Patrick Pearse on the obverse and an image of the statue of Cú Chulainn in the GPO on the reverse. Its edge inscription reads, "Éirí Amach na Cásca 1916", which translates to, "1916 Easter Rising". Due to their 83.5% silver content, many of the coins were melted down shortly after issue. A €2 coin was also issued by Ireland in 2016, featuring the statue of Hibernia above the GPO, to commemorate the Rising's centenary. With the outbreak of the Troubles in Northern Ireland, government, academics and the media began to revise the country's militant past, and particularly the Easter Rising. The coalition government of 1973–77, in particular the Minister for Posts and Telegraphs, Conor Cruise O'Brien, began to promote the view that the violence of 1916 was essentially no different from the violence then taking place in the streets of Belfast and Derry. O'Brien and others asserted that the Rising was doomed to military defeat from the outset, and that it failed to account for the determination of Ulster Unionists to remain in the United Kingdom. Irish republicans continue to venerate the Rising and its leaders with murals in republican areas of Belfast and other towns celebrating the actions of Pearse and his comrades, and annual parades in remembrance of the Rising. The Irish government, however, discontinued its annual parade in Dublin in the early 1970s, and in 1976 it took the unprecedented step of proscribing (under the Offences against the State Act) a 1916 commemoration ceremony at the GPO organised by Sinn Féin and the Republican Commemoration Committee. A Labour Party TD, David Thornley, embarrassed the government (of which Labour was a member) by appearing on the platform at the ceremony, along with Máire Comerford, who had fought in the Rising, and Fiona Plunkett, sister of Joseph Plunkett. With the advent of a Provisional IRA ceasefire and the beginning of what became known as the Peace Process during the 1990s, the government's view of the Rising grew more positive and in 1996 an 80th anniversary commemoration at the Garden of Remembrance in Dublin was attended by the Taoiseach and leader of Fine Gael, John Bruton. In 2005, the Taoiseach, Bertie Ahern, announced the government's intention to resume the military parade past the GPO from Easter 2006, and to form a committee to plan centenary celebrations in 2016. The 90th anniversary was celebrated with a military parade in Dublin on Easter Sunday, 2006, attended by the President of Ireland, the Taoiseach and the Lord Mayor of Dublin. There is now an annual ceremony at Easter attended by relatives of those who fought, by the President, the Taoiseach, ministers, senators and TDs, and by usually large and respectful crowds. The Rising continues to attract debate and analysis. In 2016 "The Enemy Files", a documentary presented by a former British Secretary of State for Defence, Michael Portillo, was shown on RTÉ One and the BBC, ahead of the centenary. Portillo declared that the execution of the 16 leaders of the insurrection could be justified in its context – a military response, against the background of the appalling European war – but that the rebels had set a trap that the British fell into and that every possible response by the British would have been a mistake of some kind. He commented on the role of Patrick Pearse, the martyrdom controversy and the Proclamation's reference to "our gallant [German] allies in Europe". In December 2014 Dublin City Council approved a proposal to create a historical path commemorating the Rising, similar to the Freedom Trail in Boston. Lord Mayor of Dublin Christy Burke announced that the council had committed to building the trail, marking it with a green line or bricks, with brass plates marking the related historic sites such as the Rotunda and the General Post Office. A pedestrian staircase that runs along 53rd Avenue, from 65th Place to 64th Street in west Queens, New York City was named 'Easter Rising Way' in 2016. Sinn Féin leader, Gerry Adams spoke at the naming ceremony. The Easter Rising lasted from Easter Monday 24 April 1916 to Easter Saturday 29 April 1916. Annual commemorations, rather than taking place on 24–29 April, are typically based on the date of Easter, which is a moveable feast. For example, the annual military parade is on Easter Sunday; the date of coming into force of the Republic of Ireland Act 1948 was symbolically chosen as Easter Monday (18 April) 1949. The official programme of centenary events in 2016 climaxed from 25 March (Good Friday) to 2 April (Easter Saturday) with other events earlier and later in the year taking place on the calendrical anniversaries.
https://en.wikipedia.org/wiki?curid=10350
Eschrichtiidae Eschrichtiidae or the gray whales is a family of baleen whale (suborder Mysticeti) with a single extant species, the gray whale ("Eschrichtius robustus"). The family, however, also includes three described fossil genera: "Archaeschrichtius" and "Eschrichtioides" from the Miocene and Pliocene of Italy respectively, and "Gricetoides" from the Pliocene of North Carolina. The names of the extant genus and the family honours Danish zoologist Daniel Eschricht. A number of 18th century authors described the gray whale as "Balaena gibbosa", the "whale with six bosses", apparently based on a brief note by : The gray whale was first described as a distinct species by based on a subfossil found in the brackish Baltic Sea, apparently a specimen from the now extinct north Atlantic population. Lilljeborg, however, identified it as ""Balaenoptera robusta"", a species of rorqual. realized that the rib and scapula of the specimen was different from those of any known rorquals, and therefore erected a new genus for it, "Eschrichtius". Things got increasingly confused as 19th century scientists introduced new species at an alarming rate (e.g. "Eschrichtius pusillus", "E. expansus", "E. priscus", "E. mysticetoides"), often based on fragmentary specimens, and taxonomists started to use several generic and specific names interchangeably and not always correctly (e.g. "Agalephus gobbosus", "Balaenoptera robustus", "Agalephus gibbosus"). Things got even worse in the 1930s when it was finally realised that the extinct Atlantic population was the same species as the extant Pacific population, and the new combination "Eschrichtius gibbosus" was proposed. In his morphological analysis, found that eschrichtiids and Cetotheriidae ("Cetotherium", "Mixocetus" and "Metopocetus") form a monophyletic sister group of Balaenopteridae. A specimen from the Late Pliocene of northern Italy, named ""Cetotherium" gastaldii" by and renamed ""Balaenoptera" gastaldii" by , was identified as a basal eschrichtiid by who recombined it to "Eschrichtioides gastaldii". Fossils of Eschrichtiidae have been found in all major oceanic basins in the Northern Hemisphere, and the family is believed date back to the Late Miocene. Today, gray whales are only present in the northern Pacific, but a population was also present in the northern Atlantic before being driven to extinction by European whalers three centuries ago. Fossil eschrichtiids from before the Holocene are rare compared to other fossil mysticetes. The only Pleistocene fossil from the Pacific referred to "E. eschrichtius" is a partial skeleton and an associated skull from California, estimated to be about 200 thousand years old. However, a late Pliocene fossil from Hokkaido, Japan, referred to "Eschrichtius" sp. is estimated to be and a similar unnamed fossil has been reported from California. In their description of "Archaeschrichtius ruggieroi" from the late Miocene of Italy, argued that eschrichtiids most likely originated in the Mediterranean Basin about and remained there, either permanently or intermittently, at least until the Early Pliocene (5–3 Mya), (but see Messinian salinity crisis.)
https://en.wikipedia.org/wiki?curid=10353
Endothermic process An endothermic process is any process which requires or absorbs thermal energy from its surroundings, usually in the form of heat. It may be a chemical process, such as dissolving ammonium nitrate in water, or a physical process, such as the melting of ice cubes. The term was coined by Marcellin Berthelot from the Greek roots "endo-", derived from the word "endon" (ἔνδον) meaning "within", and the root "therm" (θερμ-), meaning "hot" or "warm" in the sense that a reaction depends on absorbing heat if it is to proceed. The opposite of an endothermic process is an exothermic process, one that releases, "gives out" energy in the form of heat. Thus in each term (endothermic & exothermic) the prefix refers to where heat goes as the reaction occurs, though in reality it only refers to where the energy goes, without necessarily being in the form of heat. All chemical reactions involve both the breaking of existing and the making of new chemical bonds. A reaction to break a bond always requires the input of energy and so such a process is always endothermic. When atoms come together to form new chemical bonds, the electrostatic forces bringing them together leave the bond with a large excess of energy (usually in the form of vibrations and rotations). If that energy is not dissipated, the new bond would quickly break apart again. Instead, the new bond can shed its excess energy - by radiation, by transfer to other motions in the molecule, or to other molecules through collisions - and then become a stable new bond. Shedding this excess energy is the exothermicity that leaves the molecular system. Whether a given overall reaction is exothermic or endothermic is determined by the relative contribution of these bond breaking endothermic steps and new bond stabilizing exothermic steps. The concept is frequently applied in physical sciences to, for example, chemical reactions, where thermal energy (heat) is converted to chemical bond energy. Endothermic (and exothermic) analysis only accounts for the enthalpy change (∆H) of a reaction. The full energy analysis of a reaction is the Gibbs free energy (∆G), which includes an entropy (∆S) and temperature term in addition to the enthalpy. A reaction will be a spontaneous process at a certain temperature if the products have a lower Gibbs free energy (an exergonic reaction) even if the enthalpy of the products is higher. Entropy and enthalpy are different terms, so the change in entropic energy can overcome an opposite change in enthalpic energy and make an endothermic reaction favorable. Eg: Burning of fuel [ Exothermic Reaction ]
https://en.wikipedia.org/wiki?curid=10356
Earle Page Sir Earle Christmas Grafton Page, (8 August 188020 December 1961) was an Australian politician who served as the 11th Prime Minister of Australia, holding office for 19 days after the death of Joseph Lyons in 1939. He was the leader of the Country Party from 1921 to 1939, and was the most influential figure in its early years. He is the third longest-serving Australian parliamentarian of all time. Page was born in Grafton, New South Wales. He entered the University of Sydney at the age of 15, and completed a degree in medicine at the age of 21. After completing his residency at Sydney's Royal Prince Alfred Hospital, he moved back to Grafton and opened a private hospital. He soon became involved in local politics, and in 1915 purchased a part-share in "The Daily Examiner", a local newspaper. He also briefly served as a military surgeon during World War I. Page gained prominence as an advocate of various development schemes for the Northern Rivers region, especially those involving hydroelectricity. He also helped found a movement for New England statehood. In 1919, Page was elected to federal parliament representing the Division of Cowper. He joined the new Country Party the following year as its inaugural whip, and then replaced William McWilliams as party leader in 1921. Page opposed the economic policies of Prime Minister Billy Hughes, and when the Country Party gained the balance of power at the 1922 election, he demanded Hughes' resignation as the price for a coalition with the Nationalist Party. He was subsequently made Treasurer under the new prime minister, Stanley Bruce, serving in that role from 1923 to 1929. He had a significant degree of influence on domestic policy, with Bruce concentrating on international issues. Page returned to cabinet after the 1934 election, when the Country Party entered a new coalition with Joseph Lyons' United Australia Party. He was appointed Minister for Commerce, and concentrated on agricultural issues. When Lyons died in office in April 1939, Page was commissioned as his successor in a caretaker capacity while the UAP elected a new leader, Robert Menzies. Page subsequently refused to serve in Menzies' cabinet, withdrawing the Country Party from the coalition, but this proved unpopular and he resigned the party leadership after a few months. The coalition was eventually reconstituted, and Page served again as Minister for Commerce under Menzies and Arthur Fadden until the government's defeat in October 1941. Page's last major role was as Minister for Health (1949–1956) in the post-war Menzies Government. He retired from cabinet at the age of 76, and died a short time after losing his seat at the 1961 election. Page served in parliament for almost 42 years, and only Menzies lasted longer as the leader of a major Australian political party. He secured his party's independence by refusing overtures to join the Nationalists and the UAP, and the policies that he favoured – decentralisation, agrarianism, and government support of primary industry – have remained the basis of its platform up to the present day. The coalitions that he established and maintained with Bruce and Lyons have served as a model for all subsequent coalition governments. Earle Christmas Grafton Page was born in Grafton, New South Wales, on 8 August 1880. His first middle name, which he disliked, was given to him to carry on the surname of a childless relative, while his second middle name was in honour of his birthplace. Page was the fifth of eleven children born to Charles Page and Mary Johanna Haddon (Annie) Cox. His younger brother Harold was the deputy administrator of the Territory of New Guinea and a Japanese prisoner of war. Page's parents had both lived in Grafton since they were children. His mother was born in Tasmania to an English father and a Scottish mother. His father, born in London, was a successful businessman and a member of the Grafton City Council, serving a single term as mayor in 1908. The family business was a hardware manufacturing firm, which had its origins in a coachbuilding firm established in 1858 by Page's maternal grandfather, Edwin Cox. His other grandfather, James Page, arrived in Grafton in 1855, serving as the town's first schoolmaster and first town clerk. Page began his schooling at Grafton Public School, where he excelled academically. His family could not afford to send him to boarding school, as a result of financial difficulties caused by the banking crisis of 1893. Page consequently had to rely on scholarships to advance his education. He won a bursary to attend Sydney Boys High School in 1895, where he passed the university entrance exams, and the following year – aged 15 – began studying a liberal arts course at the University of Sydney. He was equal top in mathematics in his first year, and was also awarded the lucrative Struth Exhibition for "general proficiency in the arts", which allowed him to switch to medicine and covered his first four years of medical school. His role model was Grafton Smith, who had followed a similar path from Grafton Public School to university. At Sydney Medical School, Page's lecturers included William Haswell (biology), James Hill (biology), Charles Martin (physiology), Anderson Stuart (physiology), and James Wilson (anatomy). He graduated at the top of his class in 1901, with the degrees of Bachelor of Medicine (M.B.) and Master of Surgery (Ch.M.). Page's first professional posting came before he had even been registered as a medical practitioner. Due to a shortage of doctors, he was acting superintendent of the Royal Alexandra Hospital for Children for one month. In 1902, he took up a position as a resident at the Royal Prince Alfred Hospital, serving in a variety of roles including as house surgeon under Robert Scot Skirving. During that time he contracted a near-fatal infection from a postmortem examination. He also met his future wife, nurse Ethel Blunt. Page returned to his home town in 1903, taking over a practice in South Grafton. He and two partners subsequently established a new private hospital, Clarence House Hospital, which opened in 1904 and served both Grafton and the surrounding region. Page was a keen adopter of new technologies. In 1904, he bought what he claimed was "the first Rover car in Australia", which was powered by kerosene. He upgraded to an Itala in 1908, and had the chassis enlarged so it could be used as an ambulance. He also had an x-ray machine installed in his hospital, one of the first in Australia outside a major city. Page developed a reputation for surgical innovation, taking a number of patients from Sydney and even some from interstate. One operation that brought him particular fame was the removal of a patient's diseased lung, a procedure that had only been invented a few years previously. Page became an inaugural Fellow of the Royal Australasian College of Surgeons (FRACS) in 1927, and in 1942 was made an honorary Fellow of the Royal College of Surgeons of England (FRCS). In February 1916, Page enlisted in the Australian Army Medical Corps. He served as chief medical officer aboard the troopship "Ballarat", and was then stationed at an army hospital in Cairo for several months. He was transferred to a hospital in England in July 1916, and concluded his service as a surgical specialist at a casualty clearing station in France. Page returned to Australia in March 1917 and was discharged from the military in July 1917. Although his active involvement in medicine declined as his political career progressed, he was frequently called upon to treat his fellow MPs or parliamentary staff. This was particularly true after the federal government moved to Canberra, as the new capital had only a handful of qualified surgeons. In 1928, for instance, he performed an emergency appendectomy on Parker Moloney. Page's medical career brought him considerable wealth, and he began investing in land. He bought several large farming properties in South-East Queensland, including in Nerang, Kandanga, and the Numinbah Valley; Pages Pinnacle in the Numinbah State Forest is named after him. His entry into public life came about as a result of his passion for hydroelectricity, which he first observed in New Zealand while attending a medical convention in 1910. He believed that it could be applied to the Northern Rivers region, which was still mostly unelectrified outside of the major towns. Page was elected to the South Grafton Municipal Council in 1913, believing his position as an alderman would be useful in his lobbying efforts. However, his overtures to the state government were rebuffed. In 1915, Page was one of the founders of the Northern New South Wales Separation League, which advocated the creation of a new state in the New England region. He toured a number of towns to raise awareness of the new movement, but interest waned as a result of the ongoing war. Later that year, he was part of a syndicate that bought "The Daily Examiner", the local newspaper in Grafton. Page visited a number of hydroelectric sites in North America in 1917, on his way back from military service in France. He was elected mayor of South Grafton in 1918, serving until 1920, and also became the inaugural president of the North Coast Development League. He developed more concrete plans for a hydroelectric project on the Clarence River, and put forward various other development schemes relating to roads, railways, and ports, all of which served to raise his profile in the local district. Page was elected to the House of Representatives at the 1919 federal election, defeating the sitting Nationalist MP, John Thomson in the Division of Cowper. He stood as an independent with the endorsement of the Farmers' and Settlers' Association, and after the election joined the new Country Party, along with 10 other MPs from rural seats. Page continued to advocate for hydroelectricity throughout his political career, and many such projects were built in New South Wales. However, the specific scheme he favoured for the Clarence River was never put in place, only the smaller Nymboida Power Station. Decentralisation also remained a pet project, with Page frequently arguing for New South Wales and Queensland to be divided into smaller states to aid regional development. The movement for New England statehood waned in the 1920s, but re-emerged in the 1950s; a legally binding referendum on the subject was finally held in 1967, after Page's death, but was narrowly defeated in controversial circumstances. Page was elected leader of the Country Party in 1921, replacing William McWilliams. The fledgling Country Party found itself with the balance of power in the House after the 1922 election. The Nationalist government of Billy Hughes lost its majority, and could not govern without Country Party support. It soon became apparent that the price for that support would be a full coalition with the Nationalists. However, the Country Party had been formed partly due to discontent with Hughes' rural policy, and Page's animosity toward Hughes was such that he would not even consider supporting him. Indeed, he would not even begin talks with the Nationalists as long as Hughes remained leader. Bowing to the inevitable, Hughes resigned. Page then began negotiations with Hughes' successor as leader of the Nationalists, Stanley Bruce. His terms were stiff; he wanted his Country Party to have five seats in an 11-man cabinet, including the post of Treasurer and the second rank in the ministry for himself. These demands were unprecedented for a prospective junior coalition partner in a Westminster system, and especially so for such a new party. Nonetheless, Bruce agreed rather than force another election. For all intents and purposes, Page was the first Deputy Prime Minister of Australia (a title that did not officially exist until 1968). Since then, the leader of the Country/National Party has been the second-ranking member in nearly every non-Labor government. On 30 January 1924, as acting prime minister, Page chaired the first meeting of Federal Cabinet ever held in Canberra, at Yarralumla. This was still three years before the opening of Parliament House and Canberra becoming the National Capital. Page continued his professional medical practice. On 22 October 1924, he had to tell his best friend, Thomas Shorten Cole (1870–1957), the news that his wife Mary Ann Crane had just died on the operating table from complications of intestinal or stomach cancer, reputed by their daughter Dorothy May Cole to be "the worst day of his life". He was a strong believer in orthodox finance and conservative policies, except where the welfare of farmers was concerned: then he was happy to see government money spent freely. He was also a "high protectionist": a supporter of high tariff barriers to protect Australian rural industries. The Bruce-Page government was heavily defeated by Labor in 1929 (with Bruce losing his own seat), and Page went into opposition. In 1931, a group of dissident Labor MPs led by Joseph Lyons merged with the Nationalists to form the United Australia Party under Lyons' leadership. Lyons and the UAP won majority government at the 1931 election. Although Lyons was keen to form a coalition with the Country Party, talks broke down, and Lyons opted to govern alone—to date, the last time that the Country/National Party has not had any posts in a non-Labor government. In 1934, however, the UAP suffered an eight-seat swing, forcing Lyons to take the Country Party back into his government in a full-fledged Coalition. Page became Minister for Commerce. He was made a Knight Grand Cross of the Order of St Michael and St George (GCMG) in the New Year's Day Honours of 1938. While nine Australian Prime Ministers were knighted (and Bruce was elevated to the peerage), Page is the only one who was knighted before becoming Prime Minister. When Lyons died suddenly in 1939, the Governor-General Lord Gowrie appointed Page as caretaker Prime Minister pending the UAP choosing a new leader. He held the office for three weeks until the UAP elected former deputy leader Robert Menzies as its new leader, and hence Prime Minister. Page had been close to Lyons, but disliked Menzies, whom he charged publicly with having been disloyal to Lyons. Page contacted Stanley Bruce (now in London as Australian High Commissioner to the UK) and offered to resign his seat if Bruce would return to Australia to seek re-election to the parliament in a by-election for Page's old seat, and then seek election as UAP leader. Bruce said that he would only re-enter the parliament as an independent. When Menzies was elected UAP leader, Page refused to serve under him, and made an extraordinary personal attack on him in the House, accusing him not only of ministerial incompetence but of physical cowardice (for failing to enlist during World War I). His party soon rebelled, though, and Page was deposed as Country Party leader in favour of Archie Cameron. In March 1940, Archie Cameron led the Country Party back into coalition with the UAP. However, he resigned as party leader on 16 October, following the 1940 federal election. Page attempted to regain the party's leadership, but was deadlocked with John McEwen over multiple ballots. As a compromise, the party elected Arthur Fadden as acting leader; he was confirmed in the position a few months later. Page replaced Cameron as Minister for Commerce in the reconstituted ministry. Fadden replaced Menzies as prime minister in August 1941. A few weeks later, cabinet decided to send Page to London as resident minister, with the intention that he would be granted access to the British War Cabinet. While he was en route to England, the Fadden Government lost a confidence motion and was replaced by an ALP minority government. The new prime minister John Curtin nonetheless allowed Page to take up the position, declining his offer to return to Australia. The attack on Pearl Harbor in December changed the dynamic of Anglo-Australian relations, as the War in the Pacific became the primary concern of the Australian government. Page assisted in the creation of the Pacific War Council early the following year. He later recalled Winston Churchill's frustration in war cabinet meetings with Curtin's decision to withdraw troops from the Middle East and North Africa and return them to Australia. He credited himself with helping negate the tensions between the two men, but in February 1942 mistakenly advised Churchill that the Australian government was amenable to diverting the 7th Division to Burma rather than return it directly to Australia. He was heavily rebuked by Curtin and external affairs minister H. V. Evatt for his error. Page wrote to Curtin in April 1942 that since January he had been through "the worst period of acute mental distress of my whole life". His tenure was not regarded as a success, and he was said to have suffered from a lack of experience in diplomacy. Field Marshal Alan Brooke, the Chief of the Imperial General Staff, recalled that in war cabinet meetings he had "the mentality of a greengrocer". Page left London in June 1942 following a severe bout of pneumonia. He had been made a Companion of Honour (CH) before his departure. He returned to Australia in August, travelling via the United States, and quickly turned his attention to planning for post-war reconstruction. Page spent the remaining years of the Curtin and Chifley Governments on the opposition backbench. He served on the Advisory War Council and was a delegate to the constitutional convention in Canberra in late 1942, which included members of all major political parties. However, he was frustrated by the government's failure to offer him any formal role in developing post-war policy, which he believed was due to him given his past work. Page was reappointed Minister for Health after the Coalition won the 1949 federal election, at the age of 69. He was the chief architect of the "National Health Act 1953", which established a national public health scheme based on government subsidies of voluntary private insurance and free medical services for pensioners. He played a key role in securing the support of the medical profession, which had strongly opposed the Chifley Government's attempt to introduce universal health care. Unlike in previous governments, Page had little influence beyond his own policy area and was frustrated by the lack of interest in his ideas for national development. Upon the death of Billy Hughes in October 1952, Page became the Father of the House of Representatives and Father of the Parliament. In 1954, he became the first chancellor of the University of New England, which had become fully autonomous from the University of Sydney. He retired from cabinet at the age of 76, moving to the backbench in January 1956 after the December 1955 election. By the 1961 election, Page was gravely ill, suffering from lung cancer. Although he was too sick to actively campaign, Page refused to even consider retiring from Parliament and soldiered on for his 17th general election. In one of the great upsets of Australian electoral history, he lost his seat to Labor challenger Frank McGuren, whom he had defeated soundly in 1958. Page had gone into the election holding Cowper with what appeared to be an insurmountable 11-point majority, but McGuren managed to win the seat on a swing of 13%. Page had campaigned sporadically before going to Royal Prince Alfred Hospital in Sydney for emergency surgery. He went comatose a few days before the election and never regained consciousness. He died on 20 December, 11 days after the election, without ever knowing that he had been defeated. He was the last former Prime Minister to lose his seat until Tony Abbott lost his seat of Warringah in 2019, though John Howard would lose his seat of Bennelong as a sitting Prime Minister in 2007. Page had represented Cowper for just four days short of 42 years, making him the longest-serving Australian federal parliamentarian who represented the same seat throughout his career. Only Billy Hughes and Philip Ruddock have served in Parliament longer than Page. Page married Ethel Blunt on 18 September 1906. They had met at Royal Prince Alfred Hospital while he was undertaking his medical residency; she was a senior nurse there. Page soon began courting her, and convinced her to become the matron of his new hospital in Grafton. She gave up nursing after their marriage, but was active in politics and community organisations. The couple had five children: Mary (b. 1909), Earle Jr. (b. 1910), Donald (b. 1912), Iven (b. 1914), and Douglas (b. 1916). Their grandchildren include Don Page, who was active in New South Wales state politics, and Geoff Page, a poet. Page was predeceased by his first wife and his oldest son. Earle Jr., a qualified veterinarian, was killed by a lightning strike in January 1933, aged 22. Ethel died in May 1958, aged 82, after a long illness. On 20 July 1959 at St Paul's Cathedral, London, Page married for a second time, wedding his long-serving secretary Jean Thomas (32 years his junior). Stanley Bruce was his best man. The second Lady Page lived for almost 50 years after her husband's death, dying on 20 June 2011; her ashes were interred at Northern Suburbs Crematorium.
https://en.wikipedia.org/wiki?curid=10357
Ephrem the Syrian Ephrem the Syrian (, ; ; ; c. 306 – 373), also known as Saint Ephraem, Ephrem of Edessa, Ephrem Syrus, Ephrem, or Ephraim, was a Syriac Christian deacon and a prolific Syriac-language hymnographer and theologian of the fourth century. Ephrem is especially beloved in the Syriac Orthodox Church, and counted as a Venerable Father (i.e., a sainted Monk) in the Eastern Orthodox Church. His feast day is celebrated on 28 January and on the Saturday of the Venerable Fathers. He was declared a Doctor of the Church in the Roman Catholic Church in 1920. Ephrem wrote a wide variety of hymns, poems, and sermons in verse, as well as prose exegesis. These were works of practical theology for the edification of the Church in troubled times. So popular were his works, that, for centuries after his death, Christian authors wrote hundreds of pseudepigraphal works in his name. He has been called the most significant of all of the fathers of the Syriac-speaking church tradition. Ephrem was born around the year 306 in the city of Nisibis, Mesopotamia (modern Nusaybin, Turkey), then-recently acquired by Rome. Internal evidence from Ephrem's hymnody suggests that both his parents were part of the growing Christian community in the city, although later hagiographers wrote that his father was a pagan priest. Numerous languages were spoken in the Nisibis of Ephrem's day, mostly dialects of Aramaic. The Christian community used the Syriac dialect. The culture included pagan religions, Judaism and early Christian sects. Jacob, the second bishop of Nisibis, was appointed in 308, and Ephrem grew up under his leadership of the community. Jacob of Nisibis is recorded as a signatory at the First Council of Nicea in 325. Ephrem was baptized as a youth and almost certainly became a son of the covenant, an unusual form of syriac proto-monasticism. Jacob appointed Ephrem as a teacher (Syriac "malp̄ānâ", a title that still carries great respect for Syriac Christians). He was ordained as a deacon either at his baptism or later. He began to compose hymns and write biblical commentaries as part of his educational office. In his hymns, he sometimes refers to himself as a "herdsman" (, "‘allānâ"), to his bishop as the "shepherd" (, "rā‘yâ"), and to his community as a 'fold' (, "dayrâ"). Ephrem is popularly credited as the founder of the School of Nisibis, which, in later centuries, was the centre of learning of the Syriac Orthodox Church. In 337, Emperor Constantine I, who had legalised and promoted the practice of Christianity in the Roman Empire, died. Seizing on this opportunity, Shapur II of Persia began a series of attacks into Roman North Mesopotamia. Nisibis was besieged in 338, 346 and 350. During the first siege, Ephrem credits Bishop Jacob as defending the city with his prayers. In the third siege, of 350, Shapur rerouted the River Mygdonius to undermine the walls of Nisibis. The Nisibenes quickly repaired the walls while the Persian elephant cavalry became bogged down in the wet ground. Ephrem celebrated what he saw as the miraculous salvation of the city in a hymn that portrayed Nisibis as being like Noah's Ark, floating to safety on the flood. One important physical link to Ephrem's lifetime is the baptistery of Nisibis. The inscription tells that it was constructed under Bishop Vologeses in 359. In that year, Shapur attacked again. The cities around Nisibis were destroyed one by one, and their citizens killed or deported. Constantius II was unable to respond; the campaign of Julian in 363 ended with his death in battle. His army elected Jovian as the new emperor, and to rescue his army, he was forced to surrender Nisibis to Persia (also in 363) and to permit the expulsion of the entire Christian population. Ephrem, with the others, went first to Amida (Diyarbakır), eventually settling in Edessa (modern Şanlıurfa) in 363. Ephrem, in his late fifties, applied himself to ministry in his new church and seems to have continued his work as a teacher, perhaps in the School of Edessa. Edessa had always been at the heart of the Syriac-speaking world, and the city was full of rival philosophies and religions. Ephrem comments that orthodox Nicene Christians were simply called "Palutians" in Edessa, after a former bishop. Arians, Marcionites, Manichees, Bardaisanites and various gnostic sects proclaimed themselves as the true church. In this confusion, Ephrem wrote a great number of hymns defending Nicene orthodoxy. A later Syriac writer, Jacob of Serugh, wrote that Ephrem rehearsed all-female choirs to sing his hymns set to Syriac folk tunes in the forum of Edessa. After a ten-year residency in Edessa, in his sixties, Ephrem succumbed to the plague as he ministered to its victims. The most reliable date for his death is 9 June 373. Over four hundred hymns composed by Ephrem still exist. Granted that some have been lost, Ephrem's productivity is not in doubt. The church historian Sozomen credits Ephrem with having written over three million lines. Ephrem combines in his writing a threefold heritage: he draws on the models and methods of early Rabbinic Judaism, he engages skillfully with Greek science and philosophy, and he delights in the Mesopotamian/Persian tradition of mystery symbolism. The most important of his works are his lyric, teaching hymns (, "madrāšê"). These hymns are full of rich, poetic imagery drawn from biblical sources, folk tradition, and other religions and philosophies. The madrāšê are written in stanzas of syllabic verse and employ over fifty different metrical schemes. Each madrāšâ had its "qālâ" (), a traditional tune identified by its opening line. All of these qālê are now lost. It seems that Bardaisan and Mani composed madrāšê, and Ephrem felt that the medium was a suitable tool to use against their claims. The madrāšê are gathered into various hymn cycles. Each group has a title — "Carmina Nisibena", "On Faith", "On Paradise", "On Virginity", "Against Heresies" — but some of these titles do not do justice to the entirety of the collection (for instance, only the first half of the "Carmina Nisibena" is about Nisibis). Each madrāšâ usually had a refrain (, "‘ûnîṯâ"), which was repeated after each stanza. Later writers have suggested that the madrāšê were sung by all-women choirs with an accompanying lyre. Particularly influential were his "Hymns Against Heresies". Ephrem used these to warn his flock of the heresies that threatened to divide the early church. He lamented that the faithful were "tossed to and fro and carried around with every wind of doctrine, by the cunning of men, by their craftiness and deceitful wiles." He devised hymns laden with doctrinal details to inoculate right-thinking Christians against heresies such as docetism. The "Hymns Against Heresies" employ colourful metaphors to describe the Incarnation of Christ as fully human and divine. Ephrem asserts that Christ's unity of humanity and divinity represents peace, perfection and salvation; in contrast, docetism and other heresies sought to divide or reduce Christ's nature and, in doing so, rend and devalue Christ's followers with their false teachings. Ephrem also wrote verse homilies (, "mêmrê"). These sermons in poetry are far fewer in number than the madrāšê. The mêmrê were written in a heptosyllabic couplets (pairs of lines of seven syllables each). The third category of Ephrem's writings is his prose work. He wrote a biblical commentary on the Diatessaron (the single gospel harmony of the early Syriac church), the Syriac original of which was found in 1957. His "Commentary on Genesis and Exodus" is an exegesis of Genesis and Exodus. Some fragments exist in Armenian of his commentaries on the Acts of the Apostles and Pauline Epistles. He also wrote refutations against Bardaisan, Mani, Marcion and others. Ephrem is attributed with writing hagiographies such as "The Life of Saint Mary the Harlot", though this credit is called into question. Ephrem wrote exclusively in the Syriac language, which is a dialect of middle Aramaic, but translations of his writings exist in Classical Armenian, Coptic, Old Georgian, Koine Greek and other languages. Some of his works are only extant in translation (particularly in Armenian). Syriac churches still use many of Ephrem's hymns as part of the annual cycle of worship. However, most of these liturgical hymns are edited and conflated versions of the originals. The most complete, critical text of authentic Ephrem was compiled between 1955 and 1979 by Dom Edmund Beck, OSB, as part of the "Corpus Scriptorum Christianorum Orientalium". As Chronologist, St. Ephrem the Syrian has composed the history of the Patriarchs and Kings from the Creation to the Crucifixion of Christ, "The Book of the Cave of Treasure", translated by W. Budge from the Syriac text of the British Museum Mss Add. 25875, published by The Religious Tract Society, 1927. Ephrem's writings contain a rich variety of symbols and metaphors. Christopher Buck gives a summary of analysis of a selection of six key scenarios (the way, robe of glory, sons and daughters of the Covenant, wedding feast, harrowing of hell, Noah’s Ark/Mariner) and six root metaphors (physician, medicine of life, mirror, pearl, Tree of life, paradise). Ephrem's artful meditations on the symbols of Christian faith and his stand against heresy made him a popular source of inspiration throughout the church. This occurred to the extent that there is a huge corpus of Ephrem pseudepigraphy and legendary hagiography. Some of these compositions are in verse, often a version of Ephrem's heptosyllabic couplets. Most of these works are considerably later compositions in Greek. Students of Ephrem often refer to this corpus as having a single author called "Greek Ephrem", or "Ephraem Graecus" (as opposed to the real Ephrem the Syrian). This is not to say that all texts ascribed to Ephrem in Greek are by others, but many are. Although Greek compositions are the main source of pseudepigraphal material, there are also works in Latin, Slavonic and Arabic. There has been very little critical examination of these works, and many are still treasured by some churches as authentic. The best known of these writings is the "Prayer of Saint Ephrem", which is recited at every service during Great Lent and other fasting periods in Eastern Christianity. Soon after Ephrem's death, legendary accounts of his life began to circulate. One of the earlier "modifications" is the statement that Ephrem's father was a pagan priest of Abnil or Abizal. However, internal evidence from his authentic writings suggest that he was raised by Christian parents. The second legend attached to Ephrem is that he was a monk. In Ephrem's day, monasticism was in its infancy in Egypt. He seems to have been a part of the members of the covenant, a close-knit, urban community of Christians that had "covenanted" themselves to service and had refrained from sexual activity. Some of the Syriac terms that Ephrem used to describe his community were later used to describe monastic communities, but the assertion that he was a monk is anachronistic. Later hagiographers often painted a picture of Ephrem as an extreme ascetic, but the internal evidence of his authentic writings show him to have had a very active role, both within his church community and through witness to those outside of it. Ephrem is venerated as an example of monastic discipline in Eastern Christianity. In the Eastern Orthodox scheme of hagiography, Ephrem is counted as a Venerable Father (i.e., a sainted Monk). His feast day is celebrated on 28 January and on the Saturday of the Venerable Fathers (Cheesefare Saturday), which is the Saturday before the beginning of Great Lent. Ephrem is popularly believed to have taken legendary journeys. In one of these he visits Basil of Caesarea. This links the Syrian Ephrem with the Cappadocian Fathers and is an important theological bridge between the spiritual view of the two, who held much in common. Ephrem is also supposed to have visited Saint Pishoy in the monasteries of Scetes in Egypt. As with the legendary visit with Basil, this visit is a theological bridge between the origins of monasticism and its spread throughout the church. On 5 October 1920, Pope Benedict XV proclaimed Ephrem a Doctor of the Church ("Doctor of the Syrians"). This proclamation was made before critical editions of Ephrem's authentic writings were available. The most popular title for Ephrem is "Harp of the Spirit" (Syriac: , "Kenārâ d-Rûḥâ"). He is also referred to as the Deacon of Edessa, the Sun of the Syrians and a Pillar of the Church. His Roman Catholic feast day of 9 June conforms to his date of death. For 48 years (1920–1969), it was on 18 June, and this date is still observed in the Extraordinary Form. Ephrem is honored with a feast day on the liturgical calendar of the Episcopal Church (USA) on June 10.
https://en.wikipedia.org/wiki?curid=10358
Amiga Enhanced Chip Set The Enhanced Chip Set (ECS) is the second generation of the Amiga computer's chipset, offering minor improvements over the original chipset (OCS) design. ECS was introduced in 1990 with the launch of the Amiga 3000. Amigas produced from 1990 onwards featured a mix of OCS and ECS chips, such as later versions of the Amiga 500 and the Commodore CDTV. Other ECS models were the Amiga 500+ in 1991 and lastly the Amiga 600 in 1992. Notable improvements were the "Super Agnus" and the "HiRes Denise" chips. The sound and floppy controller chip, "Paula", remained unchanged from the OCS design. Super Agnus supports 2 MB of Chip RAM, whereas the original Agnus/"Fat Agnus" and subsequent "Fatter Agnus" can address 512 KB and 1 MB, respectively. The ECS Denise chip offers "Productivity" (640×480 non-interlaced) and "SuperHiRes" (1280×200 or 1280×256) display modes (also available in interlaced mode), which are however limited to only 4 on-screen colors. Essentially, a 35 ns pixel mode was added plus the ability to run arbitrary horizontal and vertical scan rates. This made other display modes possible, but only the aforementioned modes were supported originally out of the box. For example, the Linux Amiga framebuffer device driver allows the use of several other display modes. Other improvements were the ability of the blitter to copy regions larger than 1024×1024 pixels in one operation and the ability to display sprites in border regions (outside of any display window where bitplanes are shown). ECS also allows software switching between 60 Hz and 50 Hz video modes. These improvements largely favored application software, which benefited from higher resolution and VGA-like display modes, rather than games. As an incremental update, ECS was intended to be backward compatible with software designed for OCS machines, though some pre-ECS games were found to be incompatible. Additionally, features from the improved Kickstart 2 operating system were used in subsequent software, and since these two technologies largely overlap, some users misjudged the significance of ECS. It is possible to upgrade some OCS machines, such as the Amiga 500, to obtain partial or full ECS functionality by replacing OCS chips with ECS versions. ECS was followed by the third generation AGA chipset with the launch of the Amiga 4000 and Amiga 1200 in 1992.
https://en.wikipedia.org/wiki?curid=10359
European Space Operations Centre The European Space Operations Centre (ESOC) serves as the main mission control centre for the European Space Agency (ESA) and is located in Darmstadt, Germany. ESOC's primary function is the operation of unmanned spacecraft on behalf of ESA and the launch and early orbit phases (LEOP) of ESA and third-party missions. The Centre is also responsible for a range of operations-related activities within ESA and in cooperation with ESA's industry and international partners, including ground systems engineering, software development, flight dynamics and navigation, development of mission control tools and techniques and space debris studies. ESOC's current major activities comprise operating planetary and solar missions, such as Mars Express and the ExoMars Trace Gas Orbiter, astronomy & fundamental physics missions, such as Gaia (spacecraft) and XMM Newton, and Earth observation missions such as CryoSat2 and Swarm (spacecraft). ESOC is responsible for developing, operating and maintaining ESA's ESTRACK network of ground stations. Teams at the Centre are also involved in research and development related to advanced mission control concepts and Space Situational Awareness, and standardisation activities related to frequency management; mission operations; tracking, telemetry and telecommanding; and space debris. ESOC's current missions comprise the following: Planetary and solar missions Astronomy and fundamental physics missions Earth observation missions In addition, the ground segment and mission control teams for several missions are in preparation and training, including: ESOC hosts the control centre for the Agency's European Tracking ESTRACK station network. The core network comprises seven stations in seven countries: Kourou (French Guiana), Cebreros (Spain), Redu (Belgium), Santa Maria (Portugal), Kiruna (Sweden), Malargüe (Argentina) and New Norcia (Australia). Operators are on duty at ESOC 24 hours/day, year round, to conduct tracking passes, uploading telecommands and downloading telemetry and data. In addition to 'pure' mission operations, a number of other activities take place at the Centre, most of which are directly related to ESA's broader space operations activities. The European Space Operations Centre was formally inaugurated in Darmstadt, Germany, on 8 September 1967 by the then-Minister of Research of the Federal Republic of Germany, Gerhard Stoltenberg. Its role was to provide satellite control for the European Space Research Organisation (ESRO), which is today known as its successor organisation, the European Space Agency (ESA). The 90-person ESOC facility was, as it is today, located on the west side of Darmstadt; it employed the staff and resources previously allocated to the European Space Data Centre (ESDAC), which had been established in 1963 to conduct orbit calculations. These were augmented by mission control staff transferred from ESTEC to operate satellites and manage the ESTRACK tracking station network. Within just eight months, ESOC, as part of ESRO, was already operating its first mission, ESRO-2B, a scientific research satellite and the first of many operated from ESOC for ESRO, and later ESA. By July 2012, ESOC had operated over 56 missions spanning science, Earth observation, orbiting observatories, meteorology and space physics. ESOC is located on the west side of the city of Darmstadt, some from the main train station, at Robert-Bosch-Straße 5. In 2011, ESA announced the first phase of the ESOC II modernisation and expansion project valued at €60 million. The new construction will be located across Robert-Bosch-Straße, opposite the current centre. At ESOC, ESA employs approximately 800, comprising some 250 permanent staff and about 550 contractors. Staff from ESOC are routinely dispatched to work at other ESA establishments, ESTRACK stations, the ATV Control Centre (Toulouse), the Columbus Control Centre (Oberpfaffenhofen) and at partner facilities in several countries.
https://en.wikipedia.org/wiki?curid=10361
European Space Agency The European Space Agency (ESA; , ASE; ) is an intergovernmental organisation of 22 member states dedicated to the exploration of space. Established in 1975 and headquartered in Paris, ESA has a worldwide staff of about 2,200 in 2018 and an annual budget of about €6.68 billion (~US$7.43 billion) in 2020. ESA's space flight programme includes human spaceflight (mainly through participation in the International Space Station programme); the launch and operation of uncrewed exploration missions to other planets and the Moon; Earth observation, science and telecommunication; designing launch vehicles; and maintaining a major spaceport, the Guiana Space Centre at Kourou, French Guiana. The main European launch vehicle Ariane 5 is operated through Arianespace with ESA sharing in the costs of launching and further developing this launch vehicle. The agency is also working with NASA to manufacture the Orion Spacecraft service module that will fly on the Space Launch System. The agency's facilities are distributed among the following centres: The European Space Agency Science Programme is a long-term programme of space science and space exploration missions. After World War II, many European scientists left Western Europe in order to work with the United States. Although the 1950s boom made it possible for Western European countries to invest in research and specifically in space-related activities, Western European scientists realised solely national projects would not be able to compete with the two main superpowers. In 1958, only months after the Sputnik shock, Edoardo Amaldi (Italy) and Pierre Auger (France), two prominent members of the Western European scientific community, met to discuss the foundation of a common Western European space agency. The meeting was attended by scientific representatives from eight countries, including Harrie Massey (United Kingdom). The Western European nations decided to have two agencies: one concerned with developing a launch system, ELDO (European Launch Development Organization), and the other the precursor of the European Space Agency, ESRO (European Space Research Organisation). The latter was established on 20 March 1964 by an agreement signed on 14 June 1962. From 1968 to 1972, ESRO launched seven research satellites. ESA in its current form was founded with the ESA Convention in 1975, when ESRO was merged with ELDO. ESA had ten founding member states: Belgium, Denmark, France, West Germany, Italy, the Netherlands, Spain, Sweden, Switzerland, and the United Kingdom. These signed the ESA Convention in 1975 and deposited the instruments of ratification by 1980, when the convention came into force. During this interval the agency functioned in a de facto fashion. ESA launched its first major scientific mission in 1975, Cos-B, a space probe monitoring gamma-ray emissions in the universe, which was first worked on by ESRO. ESA collaborated with NASA on the International Ultraviolet Explorer (IUE), the world's first high-orbit telescope, which was launched in 1978 and operated successfully for 18 years. A number of successful Earth-orbit projects followed, and in 1986 ESA began Giotto, its first deep-space mission, to study the comets Halley and Grigg–Skjellerup. Hipparcos, a star-mapping mission, was launched in 1989 and in the 1990s SOHO, "Ulysses" and the Hubble Space Telescope were all jointly carried out with NASA. Later scientific missions in cooperation with NASA include the "Cassini–Huygens" space probe, to which ESA contributed by building the Titan landing module "Huygens". As the successor of ELDO, ESA has also constructed rockets for scientific and commercial payloads. Ariane 1, launched in 1979, carried mostly commercial payloads into orbit from 1984 onward. The next two versions of the Ariane rocket were intermediate stages in the development of a more advanced launch system, the Ariane 4, which operated between 1988 and 2003 and established ESA as the world leader in commercial space launches in the 1990s. Although the succeeding Ariane 5 experienced a failure on its first flight, it has since firmly established itself within the heavily competitive commercial space launch market with 82 successful launches until 2018. The successor launch vehicle of Ariane 5, the Ariane 6, is under development and is envisioned to enter service in the 2020s. The beginning of the new millennium saw ESA become, along with agencies like NASA, JAXA, ISRO, the CSA and Roscosmos, one of the major participants in scientific space research. Although ESA had relied on co-operation with NASA in previous decades, especially the 1990s, changed circumstances (such as tough legal restrictions on information sharing by the United States military) led to decisions to rely more on itself and on co-operation with Russia. A 2011 press issue thus stated: Notable ESA programs include SMART-1, a probe testing cutting-edge space propulsion technology, the "Mars Express" and "Venus Express" missions, as well as the development of the Ariane 5 rocket and its role in the ISS partnership. ESA maintains its scientific and research projects mainly for astronomy-space missions such as Corot, launched on 27 December 2006, a milestone in the search for exoplanets. On 21 January 2019, ArianeGroup and Arianespace announced a one-year contract with ESA to study and prepare for a mission to mine the Moon for lunar regolith. The treaty establishing the European Space Agency reads: ESA is responsible for setting a unified space and related industrial policy, recommending space objectives to the member states, and integrating national programs like satellite development, into the European program as much as possible. Jean-Jacques Dordain – ESA's Director General (2003–2015) – outlined the European Space Agency's mission in a 2003 interview: ESA describes its work in two overlapping ways: According to the ESA website, the activities are: Every member country must contribute to these programmes: Depending on their individual choices the countries can contribute to the following programmes, listed according to: ESA has formed partnerships with universities. ESA_LAB@ refers to research laboratories at universities. Currently there are ESA_LAB@ By 2015, ESA was an intergovernmental organisation of 22 member states. Member states participate to varying degrees in the mandatory (25% of total expenditures in 2008) and optional space programmes (75% of total expenditures in 2008). The 2008 budget amounted to €3.0 billion the 2009 budget to €3.6 billion. The total budget amounted to about €3.7 billion in 2010, €3.99 billion in 2011, €4.02 billion in 2012, €4.28 billion in 2013, €4.10 billion in 2014 and €4.33 billion in 2015. English and French are the official languages of the ESA. Additionally, official documents are also provided in German and documents regarding the Spacelab are also provided in Italian. If found appropriate, the agency may conduct its correspondence in any language of a member state. The following table lists all the member states and adjunct members, their ESA convention ratification dates, and their contributions in 2020: Currently the only associated member state is Slovenia. Previously associated members were Austria, Norway and Finland, all of which later joined ESA as full members. Since 1 January 1979, Canada has had the special status of a Cooperating State within ESA. By virtue of this accord, the Canadian Space Agency takes part in ESA's deliberative bodies and decision-making and also in ESA's programmes and activities. Canadian firms can bid for and receive contracts to work on programmes. The accord has a provision ensuring a fair industrial return to Canada. The most recent Cooperation Agreement was signed on 15 December 2010 with a term extending to 2020. For 2014, Canada's annual assessed contribution to the ESA general budget was €6,059,449 (CAD$8,559,050). For 2017, Canada has increased its annual contribution to €21,600,000 (CAD$30,000,000). ESA is funded from annual contributions by national governments as well as from an annual contribution by the European Union (EU). The budget of ESA was €5.250 billion in 2016. Every 3–4 years, ESA member states agree on a budget plan for several years at an ESA member states conference. This plan can be amended in future years, however provides the major guideline for ESA for several years. The 2016 budget allocations for major areas of ESA activity are shown in the chart on the right. Countries typically have their own space programmes that differ in how they operate organisationally and financially with ESA. For example, the French space agency CNES has a total budget of €2015 million, of which €755 million is paid as direct financial contribution to ESA. Several space-related projects are joint projects between national space agencies and ESA (e.g. COROT). Also, ESA is not the only European governmental space organisation (for example European Union Satellite Centre). After the decision of the ESA Council of 21/22 March 2001, the procedure for accession of the European states was detailed as described the document titled "The Plan for European Co-operating States (PECS)". Nations that want to become a full member of ESA do so in 3 stages. First a Cooperation Agreement is signed between the country and ESA. In this stage, the country has very limited financial responsibilities. If a country wants to co-operate more fully with ESA, it signs a European Cooperating State (ECS) Agreement. The ECS Agreement makes companies based in the country eligible for participation in ESA procurements. The country can also participate in all ESA programmes, except for the Basic Technology Research Programme. While the financial contribution of the country concerned increases, it is still much lower than that of a full member state. The agreement is normally followed by a Plan For European Cooperating State (or PECS Charter). This is a 5-year programme of basic research and development activities aimed at improving the nation's space industry capacity. At the end of the 5-year period, the country can either begin negotiations to become a full member state or an associated state or sign a new PECS Charter. Many countries, most of which joined the EU in both 2004 and 2007, have started to co-operate with ESA on various levels: During the Ministerial Meeting in December 2014, ESA ministers approved a resolution calling for discussions to begin with Israel, Australia and South Africa on future association agreements. The ministers noted that "concrete cooperation is at an advanced stage" with these nations and that "prospects for mutual benefits are existing". A separate space exploration strategy resolution calls for further co-operation with the United States, Russia and China on "LEO exploration, including a continuation of ISS cooperation and the development of a robust plan for the coordinated use of space transportation vehicles and systems for exploration purposes, participation in robotic missions for the exploration of the Moon, the robotic exploration of Mars, leading to a broad Mars Sample Return mission in which Europe should be involved as a full partner, and human missions beyond LEO in the longer term." The political perspective of the European Union (EU) was to make ESA an agency of the EU by 2014, although this date was not met. The EU is already the largest single donor to ESA's budget and non-ESA EU states are observers at ESA. ESA has a fleet of different launch vehicles in service with which it competes in all sectors of the launch market. ESA's fleet consists of three major rocket designs: Ariane 5, Soyuz-2 and Vega. Rocket launches are carried out by Arianespace, which has 23 shareholders representing the industry that manufactures the Ariane 5 as well as CNES, at ESA's Guiana Space Centre. Because many communication satellites have equatorial orbits, launches from French Guiana are able to take larger payloads into space than from spaceports at higher latitudes. In addition, equatorial launches give spacecraft an extra 'push' of nearly 500 m/s due to the higher rotational velocity of the Earth at the equator compared to near the Earth's poles where rotational velocity approaches zero. The Ariane 5 rocket is ESA's primary launcher. It has been in service since 1997 and replaced Ariane 4. Two different variants are currently in use. The heaviest and most used version, the , delivers two communications satellites of up to 10 tonnes into GTO. It failed during its first test flight in 2002, but has since made 82 consecutive successful flights until a partial failure in January 2018. The other version, , was used to launch the Automated Transfer Vehicle (ATV) to the International Space Station (ISS) and will be used to launch four Galileo navigational satellites at a time. In November 2012, ESA agreed to build an upgraded variant called (Mid-life Evolution) which would increase payload capacity to 11.5 tonnes to GTO and feature a restartable second stage to allow more complex missions. Ariane 5 ME was scheduled to fly in 2018, but the whole project was scrapped in favor of Ariane 6, planned to replace Ariane 5 in the 2020s. ESA's Ariane 1, 2, 3 and 4 launchers (the last of which was ESA's long-time workhorse) have been retired. Soyuz-2 (also called the Soyuz-ST or Soyuz-STK) is a Russian medium payload launcher (ca. 3 metric tons to GTO) which was brought into ESA service in October 2011. ESA entered into a €340 million joint venture with the Russian Federal Space Agency over the use of the Soyuz launcher. Under the agreement, the Russian agency manufactures Soyuz rocket parts for ESA, which are then shipped to French Guiana for assembly. ESA benefits because it gains a medium payload launcher, complementing its fleet while saving on development costs. In addition, the Soyuz rocket—which has been the Russian's space launch workhorse for more than 50 years—is proven technology with a very good safety record. Russia benefits in that it gets access to the Kourou launch site. Due to its proximity to the equator, launching from Kourou rather than Baikonur nearly doubles Soyuz's payload to GTO (3.0 tonnes vs. 1.7 tonnes). Soyuz first launched from Kourou on 21 October 2011, and successfully placed two Galileo satellites into orbit 23,222 kilometres above Earth. Vega is ESA's carrier for small satellites. Developed by seven ESA members led by Italy, it is capable of carrying a payload with a mass of between 300 and 1500 kg to an altitude of 700 km, for low polar orbit. Its maiden launch from Kourou was on 13 February 2012. Vega began full commercial exploitation in December 2015 The rocket has three solid propulsion stages and a liquid propulsion upper stage (the AVUM) for accurate orbital insertion and the ability to place multiple payloads into different orbits. Historically, the Ariane family rockets have been funded primarily "with money contributed by ESA governments seeking to participate in the program rather than through competitive industry bids. This [has meant that] governments commit multiyear funding to the development with the expectation of a roughly 90% return on investment in the form of industrial workshare." ESA is proposing changes to this scheme by moving to competitive bids for the development of the Ariane 6. At the time ESA was formed, its main goals did not encompass human space flight; rather it considered itself to be primarily a scientific research organisation for uncrewed space exploration in contrast to its American and Soviet counterparts. It is therefore not surprising that the first non-Soviet European in space was not an ESA astronaut on a European space craft; it was Czechoslovak Vladimír Remek who in 1978 became the first non-Soviet or American in space (the first man in space being Yuri Gagarin of the Soviet Union) – on a Soviet Soyuz spacecraft, followed by the Pole Mirosław Hermaszewski and East German Sigmund Jähn in the same year. This Soviet co-operation programme, known as Intercosmos, primarily involved the participation of Eastern bloc countries. In 1982, however, Jean-Loup Chrétien became the first non-Communist Bloc astronaut on a flight to the Soviet Salyut 7 space station. Because Chrétien did not officially fly into space as an ESA astronaut, but rather as a member of the French CNES astronaut corps, the German Ulf Merbold is considered the first ESA astronaut to fly into space. He participated in the STS-9 Space Shuttle mission that included the first use of the European-built Spacelab in 1983. STS-9 marked the beginning of an extensive ESA/NASA joint partnership that included dozens of space flights of ESA astronauts in the following years. Some of these missions with Spacelab were fully funded and organizationally and scientifically controlled by ESA (such as two missions by Germany and one by Japan) with European astronauts as full crew members rather than guests on board. Beside paying for Spacelab flights and seats on the shuttles, ESA continued its human space flight co-operation with the Soviet Union and later Russia, including numerous visits to Mir. During the latter half of the 1980s, European human space flights changed from being the exception to routine and therefore, in 1990, the European Astronaut Centre in Cologne, Germany was established. It selects and trains prospective astronauts and is responsible for the co-ordination with international partners, especially with regard to the International Space Station. As of 2006, the ESA astronaut corps officially included twelve members, including nationals from most large European countries except the United Kingdom. In the summer of 2008, ESA started to recruit new astronauts so that final selection would be due in spring 2009. Almost 10,000 people registered as astronaut candidates before registration ended in June 2008. 8,413 fulfilled the initial application criteria. Of the applicants, 918 were chosen to take part in the first stage of psychological testing, which narrowed down the field to 192. After two-stage psychological tests and medical evaluation in early 2009, as well as formal interviews, six new members of the European Astronaut Corps were selected – five men and one woman. The astronauts of the European Space Agency are: In the 1980s, France pressed for an independent European crew launch vehicle. Around 1978 it was decided to pursue a reusable spacecraft model and starting in November 1987 a project to create a mini-shuttle by the name of Hermes was introduced. The craft was comparable to early proposals for the Space Shuttle and consisted of a small reusable spaceship that would carry 3 to 5 astronauts and 3 to 4 metric tons of payload for scientific experiments. With a total maximum weight of 21 metric tons it would have been launched on the Ariane 5 rocket, which was being developed at that time. It was planned solely for use in low Earth orbit space flights. The planning and pre-development phase concluded in 1991; the production phase was never fully implemented because at that time the political landscape had changed significantly. With the fall of the Soviet Union ESA looked forward to co-operation with Russia to build a next-generation space vehicle. Thus the Hermes programme was cancelled in 1995 after about 3 billion dollars had been spent. The Columbus space station programme had a similar fate. In the 21st century, ESA started new programmes in order to create its own crew vehicles, most notable among its various projects and proposals is Hopper, whose prototype by EADS, called Phoenix, has already been tested. While projects such as Hopper are neither concrete nor to be realised within the next decade, other possibilities for human spaceflight in co-operation with the Russian Space Agency have emerged. Following talks with the Russian Space Agency in 2004 and June 2005, a co-operation between ESA and the Russian Space Agency was announced to jointly work on the Russian-designed Kliper, a reusable spacecraft that would be available for space travel beyond LEO (e.g. the moon or even Mars). It was speculated that Europe would finance part of it. A €50 million participation study for Kliper, which was expected to be approved in December 2005, was finally not approved by the ESA member states. The Russian state tender for the project was subsequently cancelled in 2006. In June 2006, ESA member states granted 15 million to the Crew Space Transportation System (CSTS) study, a two-year study to design a spacecraft capable of going beyond Low-Earth orbit based on the current Soyuz design. This project was pursued with Roskosmos instead of the cancelled Kliper proposal. A decision on the actual implementation and construction of the CSTS spacecraft was contemplated for 2008. In mid-2009 EADS Astrium was awarded a €21 million study into designing a crew vehicle based on the European ATV which is believed to now be the basis of the Advanced Crew Transportation System design. In November 2012, ESA decided to join NASA's Orion programme. The ATV would form the basis of a propulsion unit for NASA's new crewed spacecraft. ESA may also seek to work with NASA on Orion's launch system as well in order to secure a seat on the spacecraft for its own astronauts. In September 2014, ESA signed an agreement with Sierra Nevada Corporation for co-operation in Dream Chaser project. Further studies on the Dream Chaser for European Utilization or DC4EU project were funded, including the feasibility of launching a Europeanized Dream Chaser onboard Ariane 5. ESA has signed co-operation agreements with the following states that currently neither plan to integrate as tightly with ESA institutions as Canada, nor envision future membership of ESA: Argentina, Brazil, China, India (for the Chandrayan mission), Russia and Turkey. Additionally, ESA has joint projects with the European Union, NASA of the United States and is participating in the International Space Station together with the United States (NASA), Russia and Japan (JAXA). ESA is not an agency or body of the European Union (EU), and has non-EU countries (Norway, Switzerland, and the United Kingdom) as members. There are however ties between the two, with various agreements in place and being worked on, to define the legal status of ESA with regard to the EU. There are common goals between ESA and the EU. ESA has an EU liaison office in Brussels. On certain projects, the EU and ESA co-operate, such as the upcoming Galileo satellite navigation system. Space policy has since December 2009 been an area for voting in the European Council. Under the European Space Policy of 2007, the EU, ESA and its Member States committed themselves to increasing co-ordination of their activities and programmes and to organising their respective roles relating to space. The Lisbon Treaty of 2009 reinforces the case for space in Europe and strengthens the role of ESA as an R&D space agency. Article 189 of the Treaty gives the EU a mandate to elaborate a European space policy and take related measures, and provides that the EU should establish appropriate relations with ESA. Former Italian astronaut Umberto Guidoni, during his tenure as a Member of the European Parliament from 2004 to 2009, stressed the importance of the European Union as a driving force for space exploration, "since other players are coming up such as India and China it is becoming ever more important that Europeans can have an independent access to space. We have to invest more into space research and technology in order to have an industry capable of competing with other international players." The first EU-ESA International Conference on Human Space Exploration took place in Prague on 22 and 23 October 2009. A road map which would lead to a common vision and strategic planning in the area of space exploration was discussed. Ministers from all 29 EU and ESA members as well as members of parliament were in attendance. ESA has a long history of collaboration with NASA. Since ESA's astronaut corps was formed, the Space Shuttle has been the primary launch vehicle used by ESA's astronauts to get into space through partnership programmes with NASA. In the 1980s and 1990s, the Spacelab programme was an ESA-NASA joint research programme that had ESA develop and manufacture orbital labs for the Space Shuttle for several flights on which ESA participate with astronauts in experiments. In robotic science mission and exploration missions, NASA has been ESA's main partner. "Cassini–Huygens" was a joint NASA-ESA mission, along with the Infrared Space Observatory, INTEGRAL, SOHO, and others. Also, the Hubble Space Telescope is a joint project of NASA and ESA. Future ESA-NASA joint projects include the James Webb Space Telescope and the proposed Laser Interferometer Space Antenna. NASA has committed to provide support to ESA's proposed MarcoPolo-R mission to return an asteroid sample to Earth for further analysis. NASA and ESA will also likely join together for a Mars Sample Return Mission. Since China has started to invest more money into space activities, the Chinese Space Agency has sought international partnerships. ESA is, beside the Russian Space Agency, one of its most important partners. Two space agencies cooperated in the development of the Double Star Mission. In 2017, ESA sent two astronauts to China for two weeks sea survival training with Chinese astronauts in Yantai, Shandong. ESA entered into a major joint venture with Russia in the form of the CSTS, the preparation of French Guiana spaceport for launches of Soyuz-2 rockets and other projects. With India, ESA agreed to send instruments into space aboard the ISRO's Chandrayaan-1 in 2008. ESA is also co-operating with Japan, the most notable current project in collaboration with JAXA is the "BepiColombo" mission to Mercury. Speaking to reporters at an air show near Moscow in August 2011, ESA head Jean-Jacques Dordain said ESA and Russia's Roskosmos space agency would "carry out the first flight to Mars together." With regard to the International Space Station (ISS) ESA is not represented by all of its member states: 10 of the 21 ESA member states currently participate in the project: Belgium, Denmark, France, Germany, Italy, Netherlands, Norway, Spain, Sweden and Switzerland. Austria, Finland and Ireland chose not to participate, because of lack of interest or concerns about the expense of the project. The United Kingdom withdrew from the preliminary agreement because of concerns about the expense of the project. Portugal, Luxembourg, Greece, the Czech Republic, Romania and Poland joined ESA after the agreement had been signed. ESA is taking part in the construction and operation of the ISS with contributions such as Columbus, a science laboratory module that was brought into orbit by NASA's STS-122 Space Shuttle mission and the Cupola observatory module that was completed in July 2005 by Alenia Spazio for ESA. The current estimates for the ISS are approaching €100 billion in total (development, construction and 10 years of maintaining the station) of which ESA has committed to paying €8 billion. About 90% of the costs of ESA's ISS share will be contributed by Germany (41%), France (28%) and Italy (20%). German ESA astronaut Thomas Reiter was the first long-term ISS crew member. ESA has developed the Automated Transfer Vehicle for ISS resupply. Each ATV has a cargo capacity of . The first ATV, "Jules Verne", was launched on 9 March 2008 and on 3 April 2008 successfully docked with the ISS. This manoeuvre, considered a major technical feat, involved using automated systems to allow the ATV to track the ISS, moving at 27,000 km/h, and attach itself with an accuracy of 2 cm. As of 2013, the spacecraft establishing supply links to the ISS are the Russian Progress and Soyuz, European ATV, Japanese Kounotori (HTV), and the USA COTS program vehicles Dragon and Cygnus. European Life and Physical Sciences research on board the International Space Station (ISS) is mainly based on the European Programme for Life and Physical Sciences in Space programme that was initiated in 2001. According to Annex 1, Resolution No. 8 of the "ESA Convention and Council Rules of Procedure", English, French and German may be used in all meetings of the Agency, with interpretation provided into these three languages. All official documents are available in English and French with all documents concerning the ESA Council being available in German as well. The Flag of Europe is the one to be flown in space during missions (for example it was flown by ESA's Andre Kuipers during Delta mission) The Commission is increasingly working together towards common objectives. Some 20 per cent of the funds managed by ESA now originate from the supranational budget of the European Union. In recent years the ties between ESA and the European institutions have been reinforced by the increasing role that space plays in supporting Europe's social, political and economic policies. The legal basis for the EU/ESA co-operation is provided by a Framework Agreement which entered into force in May 2004. According to this agreement, the European Commission and ESA co-ordinate their actions through the Joint Secretariat, a small team of EC's administrators and ESA executive. The Member States of the two organisations meet at ministerial level in the Space Council, which is a concomitant meeting of the EU and ESA Councils, prepared by Member States representatives in the High-level Space Policy Group (HSPG). ESA maintains a liaison office in Brussels to facilitate relations with the European institutions. In May 2007, the 29 European countries expressed their support for the European Space Policy in a resolution of the Space Council, unifying the approach of ESA with those of the European Union and their member states. Prepared jointly by the European Commission and ESA's Director General, the European Space Policy sets out a basic vision and strategy for the space sector and addresses issues such as security and defence, access to space and exploration. Through this resolution, the EU, ESA and their Member States all commit to increasing co-ordination of their activities and programmes and their respective roles relating to space. On 3 August 1984 ESA's Paris headquarters were severely damaged and six people were hurt when a bomb exploded, planted by the far-left armed Action Directe group. On 14 December 2015, hackers from Anonymous breached ESA's subdomains and leaked thousands of login credentials.
https://en.wikipedia.org/wiki?curid=10363
Embouchure Embouchure () or lipping is the use of the lips, facial muscles, tongue, and teeth in playing a wind instrument. This includes shaping the lips to the mouthpiece of a woodwind instrument or the mouthpiece of a brass instrument. The word is of French origin and is related to the root "", 'mouth'. Proper embouchure allows instrumentalists to play their instrument at its full range with a full, clear tone and without strain or damage to their muscles. While performing on a brass instrument, the sound is produced by the player buzzing their lips into a mouthpiece. Pitches are changed in part through altering the amount of muscular contraction in the lip formation. The performer's use of the air, tightening of cheek and jaw muscles, as well as tongue manipulation can affect how the embouchure works. Maintaining an effective embouchure is an essential skill for any brass instrumentalist, but its personal and particular characteristics mean that different pedagogues and researchers have advocated differing, even contradictory, advice on what proper embouchure is and how it should be taught. One point on which there is some agreement is that proper embouchure is not one-size-fits-all: individual differences in dental structure, lip shape and size, jaw shape and the degree of jaw malocclusion, and other anatomical factors will affect whether a particular embouchure technique will be effective or not. In 1962, Philip Farkas hypothesized that the air stream traveling through the lip aperture should be directed straight down the shank of the mouthpiece. He believed that it would be illogical to "violently deflect" the air stream downward at the point of where the air moves past the lips. In this text, Farkas also recommends that the lower jaw be protruded so that the upper and lower teeth are aligned. In 1970, Farkas published a second text which contradicted his earlier writing. Out of 40 subjects, Farkas showed that 39 subjects directed the air downward to varying degrees and one subject directed the air in an upward direction at various degrees. The lower jaw position seen in these photographs show more variation from his earlier text as well. This supports what was written by trombonist and brass pedagogue Donald S. Reinhardt in 1942. In 1972, Reinhardt described and labeled different embouchure patterns according to such characteristics as mouthpiece placement and the general direction of the air stream as it travels past the lips. According to this later text, players who place the mouthpiece higher on the lips, so that more upper lip is inside the mouthpiece, will direct the air downwards to varying degrees while playing. Performers who place the mouthpiece lower, so that more lower lip is inside the mouthpiece, will direct the air to varying degrees in an upward manner. In order for the performer to be successful, the air stream direction and mouthpiece placement need to be personalized based on individual anatomical differences. Lloyd Leno confirmed the existence of both upstream and downstream embouchures. More controversial was Reinhardt's description and recommendations regarding a phenomenon he termed a "pivot". According to Reinhardt, a successful brass embouchure depends on a motion wherein the performer moves both the mouthpiece and lips as a single unit along the teeth in an upward and downward direction. As the performer ascends in pitch, he or she will either move the lips and mouthpiece together slightly up towards the nose or pull them down together slightly towards the chin, and use the opposite motion to descend in pitch. Whether the player uses one general pivot direction or the other, and the degree to which the motion is performed, depends on the performer's anatomical features and stage of development. The placement of the mouthpiece upon the lips doesn't change, but rather the relationship of the rim and lips to the teeth. While the angle of the instrument may change as this motion follows the shape of the teeth and placement of the jaw, contrary to what many brass performers and teachers believe, the angle of the instrument does not actually constitute the motion Reinhardt advised as a pivot. Later research supports Reinhardt's claim that this motion exists and might be advisable for brass performers to adopt. John Froelich describes how mouthpiece pressure towards the lips (vertical forces) and shear pressure (horizontal forces) functioned in three test groups, student trombonists, professional trombonists, and professional symphonic trombonists. Froelich noted that the symphonic trombonists used the least amount of both direct and shear forces and recommends this model be followed. Other research notes that virtually all brass performers rely upon the upward and downward embouchure motion. Other authors and pedagogues remain skeptical about the necessity of this motion, but scientific evidence supporting this view has not been sufficiently developed at this time. Some noted brass pedagogues prefer to instruct the use of the embouchure from a less analytical point of view. Arnold Jacobs, a tubist and well-regarded brass teacher, believed that it was best for the student to focus on his or her use of the air and musical expression to allow the embouchure to develop naturally on its own. Other instructors, such as Carmine Caruso, believed that the brass player's embouchure could best be developed through coordination exercises and drills that bring all the muscles into balance that focus the student's attention on his or her time perception. Still other authors who have differing approaches to embouchure development include Louis Maggio, Jeff Smiley, Jerome Callet. and Clint McLaughlin. Most professional performers, as well as instructors, use a combination called a puckered smile. Farkas told people to blow as if they were trying to cool soup. Raphael Mendez advised saying the letter "M". The skin under the lower lip will be taut with no air pocket. The lips do not overlap nor do they roll in or out. The corners of the mouth are held firmly in place. To play with an extended range one should use a pivot, tongue arch and lip to lip compression. According to Farkas the mouthpiece should have upper lip and lower lip (French horn), lower lip and upper lip (trumpet and cornet), and more latitude for lower brass (trombone, baritone, and tuba). For trumpet, some also advocate upper lip and lower lip. Farkas claimed placement was more important for the instruments with smaller mouthpieces. The lips should not overlap each other, nor should they roll in or out. The mouth corners should be held firm. Farkas speculated that the horn should be held in a downward angle to allow the air stream to go straight into the mouthpiece, although his later text shows that air stream direction actually is either upstream or downstream and is dependent upon the ratio of upper or lower lip inside the mouthpiece, not the horn angle. Farkas advised to moisten the outside of the lips, then form the embouchure and gently place the mouthpiece on it. He also recommended there must be a gap of or so between the teeth so that the air flows freely. Arban and Saint-Jacome were both cornet soloists and authors of well respected and still used method books. Arban stated undogmatically that he believed the mouthpiece should be placed on the top lip. Saint-Jacome to the contrary said dogmatically that the mouthpiece should be placed "two-thirds for the upper and the rest for the under according to all professors and one-third for the upper and two-thirds for the under according to one sole individual, whom I shall not name." The Farkas set is the basis of most lip buzzing embouchures. Mendez did teach lip buzzing by making the student lip buzz for a month before they could play their trumpet and got great results. One can initiate this type of buzz by using the same sensation as spitting seeds, but maintaining a continued flow of air. This technique assists the development of the Farkas approach by preventing the player from using an aperture that is too open. Stevens–Costello embouchure has its origins in the William Costello embouchure and was further developed by Roy Stevens. It uses a slight rolling in of both lips and touching evenly all the way across. It also uses mouthpiece placement of about 40–50% top lip and 50–60% lower lip. The teeth will be about apart and the teeth are parallel or the jaw slightly forward. There is relative mouthpiece pressure to the given air column. One exercise to practice the proper weight to air relationship is the palm exercise where the player holds the horn by laying it on its side in the palm of the hand, not grasping it. The lips are placed on the mouthpiece and the player blows utilizing the weight of the horn in establishing a sound. A puckered embouchure, used by most players, and sometimes used by jazz players for extremely high "screamer" notes. Maggio claimed that the pucker embouchure gives more endurance than some systems. Carlton MacBeth is the main proponent of the pucker embouchure. The Maggio system was established because Louis Maggio had sustained an injury which prevented him from playing. In this system the player cushions the lips by extending them or puckering (like a monkey). This puckering enables the players to overcome physical malformations. It also lets the player play for an extended time in the upper register. The pucker can make it easy to use to open an aperture. Lots of very soft practice can help overcome this. Claude Gordon was a student of Louis Maggio and Herbert L. Clarke and systematized the concepts of these teachers. Claude Gordon made use of pedal tones for embouchure development as did Maggio and Herbert L. Clarke. All three stressed that the mouthpiece should be placed higher on the top lip for a more free vibration of the lips. This embouchure method, advocated by a minority of brass pedagogues such as Jerome Callet, has not yet been sufficiently researched to support the claims that this system is the most effective approach for all brass performers. Advocates of Callet's approach believe that this method was recommended and taught by the great brass instructors of the early 20th Century. Two French trumpet technique books, authored by Jean-Baptiste Arban and Saint-Jacome, were translated into English for use by American players. According to some, due to a misunderstanding arising from differences in pronunciation between French and English, the commonly used brass embouchure in Europe was incorrectly interpreted. Callet attributes this difference in embouchure technique as the reason the great players of the past were able to play at the level of technical virtuosity which they did, although the increased difficulty of contemporary compositions for brass seem to indicate that the level of brass technique achieved by today's performers equals or even exceeds that of most performers from the late 19th and early 20th centuries. Callet's method of brass embouchure consists of the tongue remaining forward and through the teeth at all times. The corners of the mouth always remain relaxed, and only a small amount of air is used. The top and bottom lips curl inward and grip the forward tongue. The tongue will force the teeth, and subsequently the throat, wide open, supposedly resulting in a bigger, more open sound. The forward tongue resists the pressure of the mouthpiece, controls the flow of air for lower and higher notes, and protects the lips and teeth from damage or injury from mouthpiece pressure. Because of the importance of the tongue in this method many refer to this as a "tongue-controlled embouchure". This technique facilitates the use of a smaller mouthpiece and larger bore instruments. It results in improved intonation and stronger harmonically related partials across the player's range. A variety of transverse flute embouchures are employed by professional flautists, though the most natural form is perfectly symmetrical, the corners of the mouth relaxed (i.e. not smiling), the lower lip placed along and at a short distance from the embouchure hole. It must be stressed, however, that achieving a symmetrical, or perfectly centred blowing hole ought not to be an end in itself. Indeed, French flautist Marcel Moyse did not play with a symmetrical embouchure. The end-blown xiao, kaval, shakuhachi and hocchiku flutes demand especially difficult embouchures, sometimes requiring many lessons before any sound can be produced. The embouchure is an important element to tone production. The right embouchure, developed with "time, patience, and intelligent work", will produce a beautiful sound and a correct intonation. The embouchure is produced with the muscles around the lips: principally the orbicularis oris muscle and the depressor anguli oris, whilst avoiding activation of zygomaticus major, which will produce a smile, flattening the top lip against the maxillary (upper jaw) teeth. Beginner flute-players tend to suffer fatigue in these muscles, and notably struggle to use the depressor muscle, which necessarily helps to keep the top lip directing the flow of air across the embouchure hole. These muscles have to be properly warmed up and exercised before practicing. Tone development exercises including long notes and harmonics must be done as part of the warm up every day. Some further adjustments to the embouchure are necessary when moving from the transverse orchestral flute to the piccolo. With the piccolo, it becomes necessary to place the near side of the embouchure hole slightly higher on the lower lip, i.e. above the lip margin, and greater muscle tone from the lip muscles is needed to keep the stream/pressure of air directed across the smaller embouchure hole, particularly when playing in higher piccolo registers. With the woodwinds, aside from the flute, piccolo, and recorder, the sound is generated by a reed and not with the lips. The embouchure is therefore based on sealing the area around the reed and mouthpiece. This serves to prevent air from escaping while simultaneously supporting the reed, allowing it to vibrate, and constrict the reed preventing it from vibrating too much. With woodwinds, it is important to ensure that the mouthpiece is not placed too far into the mouth, which would result in too much vibration (no control), often creating a sound an octave (or harmonic twelfth for the clarinet) above the intended note. If the mouthpiece is not placed far enough into the mouth, no sound will be generated, as the reed will not vibrate. The standard embouchures for single reed woodwinds like the clarinet and saxophone are variants of the "single lip embouchure", formed by resting the reed upon the bottom lip, which rests on the teeth and is supported by the chin muscles and the buccinator muscles on the sides of the mouth. The top teeth rest on top of the mouthpiece. The manner in which the lower lip rests against the teeth differs between clarinet and saxophone embouchures. In clarinet playing, the lower lip is rolled over the teeth and corners of the mouth are drawn back, which has the effect of drawing the upper lip around the mouthpiece to create a seal due to the angle at which the mouthpiece rests in the mouth. With the saxophone embouchure, the lower lip rests against, but not over, the teeth as in pronouncing the letter "V" and the corners of the lip are drawn in (similar to a drawstring bag). With the less common double-lip embouchure, the top lip is placed under (around) the top teeth. In both instances, the position of the tongue in the mouth plays a vital role in focusing and accelerating the air stream blown by the player. This results in a more mature and full sound, rich in overtones. The double reed woodwinds, the oboe and bassoon, have no mouthpiece. Instead the reed is two pieces of cane extending from a metal tube (oboe – staple) or placed on a bocal (bassoon, English horn). The reed is placed directly on the lips and then played like the double-lip embouchure described above. Compared to the single reed woodwinds, the reed is very small and subtle changes in the embouchure can have a dramatic effect on tuning, tone and pitch control.
https://en.wikipedia.org/wiki?curid=10365
The Elephant 6 Recording Company The Elephant 6 Recording Company (or simply Elephant 6) is a collective of American musicians that spawned many notable independent bands of the 1990s, including the Apples in Stereo, the Olivia Tremor Control, Neutral Milk Hotel, Beulah, Elf Power, of Montreal, The Minders and Circulatory System. The collective was founded in Denver, Colorado, by childhood friends Robert Schneider, Bill Doss, Will Cullen Hart and Jeff Mangum, along with founding Apples in Stereo members Jim McIntyre and Hilarie Sidney. Schneider, Doss, Hart and Mangum grew up making music and sharing cassette tapes in Ruston, Louisiana, while attending high school. They started several bands and pet projects: Doss and Hart with the Olivia Tremor Control (then called Synthetic Flying Machine), Mangum with Neutral Milk Hotel, and Schneider with the Apples in Stereo. They shared an admiration for the music of the 1960s, especially the Beach Boys, considering the unfinished "Smile" project the era's "Holy Grail". Wishing to emulate the Beach Boys' Brother Records, Schneider created the Elephant 6 record label when he moved to Denver in late 1991 to attend the University of Colorado at Boulder. According to Schneider, Hart coined the name "Elephant 6" and Schneider added "Recording Company." In Denver, Schneider and new friends founded the Apples, which eventually became the Apples in Stereo. Recorded at the newly built Pet Sounds Studio, and released in June 1993, The Apples' "Tidal Wave" 7" EP was the first Elephant 6 product. Hart designed the Art Nouveau-inspired Elephant 6 logo for the EP's label. Doss had moved to Athens, Georgia, where he joined Hart and Mangum in Synthetic Flying Machine, which became the Olivia Tremor Control. They released "California Demise" as their first recording and Elephant 6's second. Afterward, Elephant 6's base moved from Denver to Athens. Several Elephant 6 projects began to find commercial success in the late 1990s, including Beulah, the Minders, Elf Power, Dressy Bessy, the Music Tapes, and of Montreal, as well as the founding bands. Most of the bands subsequently signed with major record labels; E6 as an entity slowly deteriorated until the collective called it quits in 2002 due to recording difficulties and lack of organization. The collective's bands all moved on to various labels and projects of their own. Many band members are still friends and even tour together under various guises. Many live together on the Orange Twin Conservation Community in Athens. The term "Elephant 6" has since come to refer to a broad range of bands and spin-off projects that the record label spawned. In 2007, The Apples in Stereo featured the Elephant 6 logo on their album "New Magnetic Wonder", announcing "The Elephant 6 Recording Company reopens our doors and windows, and invites the world: join together with your friends and make something special, something meaningful, something to remember when you are old." This marked the first major release in five years to bear the Elephant 6 logo. In 2008, the Elephant 6 logo was also used when Julian Koster released his long-awaited "Music Tapes For Clouds and Tornadoes" under the name the Music Tapes. He also assembled The Elephant 6 Holiday Surprise Tour, which featured contributions from Will Cullen Hart and Bill Doss of Olivia Tremor Control, Scott Spillane of the Gerbils, Andrew Reiger and Laura Carter of Elf Power, Theo Hilton of Nana Grizol, John Fernandes and Eric Harris of OTC, The Music Tapes and Circulatory System, Robbie Cucchiaro of The Music Tapes, Charlie Johnston and Suzanne Allison of the 63 Crayons, Nesey Gallons, Jeff Mangum, and Peter Erchick. The show took place in the UK, curated by Jeff Mangum of Neutral Milk Hotel, as part of the All Tomorrow's Parties festival in March 2012 in Minehead, England. In recent years, Robert Schneider has explored a number of experimental music projects, such as the Teletron mind-controlled synthesizer and non-Pythagorean scale of his own invention. Since Elephant 6's inception, its members showed strong interest and participated in experimental music, particularly the Olivia Tremor Control and Von Hemmling. In recent years Will Cullen Hart has also collaborated with bands such as Athens Tango Project on a cover version of "I Have Been Floated", to which he contributed vocals. In 2017, Faster Circuits released the album "Apophenia", which features the Elephant 6 logo. The Elephant 6 Recording company has begun reissuing material on vinyl, one of the first reissues being the Gerbils' "Are You Sleepy?" in 2019. In 2019, a documentary directed by Chad Stockfleth, "A Future History Of: The Elephant 6 Recording Co", was released on VHS tape.
https://en.wikipedia.org/wiki?curid=10368
Evangelicalism Evangelicalism (), evangelical Christianity, or evangelical Protestantism, is a worldwide, trans-denominational movement within Protestant Christianity that maintains the belief that the essence of the Gospel consists of the doctrine of salvation by grace alone, solely through faith in Jesus's atonement. Evangelicals believe in the centrality of the conversion or "born again" experience in receiving salvation, in the authority of the Bible as God's revelation to humanity, and in spreading the Christian message. The movement has long had a presence in the Anglosphere before spreading further afield in the 19th, 20th and early 21st centuries. Its origins are usually traced to 1738, with various theological streams contributing to its foundation, including Pietism, Puritanism, Presbyterianism and the Moravian Church (in particular its bishop Nicolaus Zinzendorf and his community at Herrnhut). Preeminently, John Wesley and other early Methodists were at the root of sparking this new movement during the First Great Awakening. Today, evangelicals are found across many Protestant branches, as well as in various denominations not subsumed to a specific branch. Among leaders and major figures of the evangelical Protestant movement were John Wesley, George Whitefield, Jonathan Edwards, Billy Graham, Bill Bright, Harold Ockenga, John Stott and Martyn Lloyd-Jones. The movement gained great momentum during the 18th and 19th centuries with the Great Awakenings in Great Britain and the United States. In 2016, there were an estimated 619 million evangelicals in the world, meaning that one in four Christians would be classified as evangelical. The United States has the largest concentration of evangelicals in the world. American evangelicals are a quarter of that nation's population and its single largest religious group. As a trans-denominational coalition, evangelicals can be found in nearly every Protestant denomination and tradition, particularly within the Reformed, Baptist, Wesleyan, Pentecostal and charismatic churches. The word "evangelical" has its etymological roots in the Greek word for "gospel" or "good news": "euangelion", from "eu" "good", "angel"- the stem of, among other words, "angelos" "messenger, angel", and the neuter suffix -"ion". By the English Middle Ages, the term had expanded semantically to include not only the message, but also the New Testament which contained the message as well as more specifically the Gospels, which portray the life, death, and resurrection of Jesus. The first published use of "evangelical" in English was in 1531, when William Tyndale wrote "He exhorteth them to proceed constantly in the evangelical truth." One year later, Thomas More wrote the earliest recorded use in reference to a theological distinction when he spoke of "Tyndale [and] his evangelical brother Barns". During the Reformation, Protestant theologians embraced the term as referring to "gospel truth". Martin Luther referred to the "evangelische Kirche" ("evangelical church") to distinguish Protestants from Catholics in the Catholic Church. Into the 21st century, "evangelical" has continued in use as a synonym for (mainline) Protestant in continental Europe, and elsewhere. This usage is reflected in the names of Protestant denominations, such as the Evangelical Church in Germany (a union of Lutheran and Reformed churches) and the Evangelical Lutheran Church in America. In the English-speaking world, "evangelical" was commonly applied to describe the series of revival movements that occurred in Britain and North America during the late 18th and early 19th centuries. Christian historian David W. Bebbington writes that, "Although 'evangelical', with a lower-case initial, is occasionally used to mean 'of the gospel', the term 'Evangelical', with a capital letter, is applied to any aspect of the movement beginning in the 1730s." According to the "Oxford English Dictionary", "evangelicalism" was first used in 1831. In 1812, the term "evangelicalism" appeared in "The History of Lynn" by William Richards. In the summer of 1811 the term "evangelicalists" was used in "The Sin and Danger of Schism" by Rev. Dr. Andrew Burnaby, Archdeacon of Leicester. The term may also be used outside any religious context to characterize a generic missionary, reforming, or redeeming impulse or purpose. For example, "The Times Literary Supplement" refers to "the rise and fall of evangelical fervor within the Socialist movement". One influential definition of evangelicalism has been proposed by historian David Bebbington. Bebbington notes four distinctive aspects of evangelical faith: conversionism, biblicism, crucicentrism, and activism, noting, "Together they form a quadrilateral of priorities that is the basis of Evangelicalism." Conversionism, or belief in the necessity of being "born again", has been a constant theme of evangelicalism since its beginnings. To evangelicals, the central message of the gospel is justification by faith in Christ and repentance, or turning away, from sin. Conversion differentiates the Christian from the non-Christian, and the change in life it leads to is marked by both a rejection of sin and a corresponding personal holiness of life. A conversion experience can be emotional, including grief and sorrow for sin followed by great relief at receiving forgiveness. The stress on conversion differentiates evangelicalism from other forms of Protestantism by the associated belief that an assurance will accompany conversion. Among evangelicals, individuals have testified to both sudden and gradual conversions. Biblicism is reverence for the Bible and high regard for biblical authority. All evangelicals believe in biblical inspiration, though they disagree over how this inspiration should be defined. Many evangelicals believe in biblical inerrancy, while other evangelicals believe in biblical infallibility. Crucicentrism is the centrality that evangelicals give to the Atonement, the saving death and the resurrection of Jesus, that offers forgiveness of sins and new life. This is understood most commonly in terms of a substitutionary atonement, in which Christ died as a substitute for sinful humanity by taking on himself the guilt and punishment for sin. Activism describes the tendency toward active expression and sharing of the gospel in diverse ways that include preaching and social action. This aspect of evangelicalism continues to be seen today in the proliferation of evangelical voluntary religious groups and parachurch organizations. The word "church" has several meanings among evangelicals. It can refer to the universal church (the body of Christ) including all Christians everywhere. It can also refer to the local church, which is the visible representation of the invisible church. It is responsible for teaching and administering the sacraments or ordinances (baptism and the Lord's Supper, but some evangelicals also count footwashing as an ordinance as well). Many evangelical traditions adhere to the doctrine of the believers' Church, which teaches that one becomes a member of the Church by the new birth and profession of faith. This originated in the Radical Reformation with Anabaptists but is held by denominations that practice believer's baptism. Evangelicals in the Anglican, Methodist and Reformed traditions practice infant baptism as one's initiation into the community of faith and the New Testament counterpart to circumcision, while also stressing the necessity of personal conversion later in life for salvation. Some evangelical denominations operate according to episcopal polity or presbyterian polity. However, the most common form of church government within Evangelicalism is congregational polity. This is especially common among non-denominational evangelical churches. Many churches are members of Evangelical Christian denominations and adhere to a common confession of faith and regulations. Common ministries within evangelical congregations are pastor, elder, deacon, evangelist and worship leader. The ministry of bishop with a function of supervision over churches on a regional or national scale is present in all the Evangelical Christian denominations, even if the titles president of the council or general overseer are mainly used for this function. The term bishop is explicitly used in certain denominations. Some evangelical denominations are members of the World Evangelical Alliance and its 129 national alliances. Some evangelical denominations officially authorize the ordination of women in churches. The female ministry is justified by the fact that Mary Magdalene was chosen by Jesus to announce his resurrection to the apostles. The first Baptist woman who was consecrated pastor is the American Clarissa Danforth in the denomination Free Will Baptist in 1815. In 1882, in the National Baptist Convention, USA. In the Assemblies of God of the United States, since 1927. In 1961, in the Progressive National Baptist Convention. In 1975, in The Foursquare Church. For evangelicals, there are three interrelated meanings to the term "worship". It can refer to living a "God-pleasing and God-focused way of life", specific actions of praise to God, and a public worship service. Diversity characterizes evangelical worship practices. Liturgical, contemporary, charismatic and seeker-sensitive worship styles can all be found among evangelical churches. Overall, evangelicals tend to be more flexible and experimental with worship practices than mainline Protestant churches. It is usually run by a Christian pastor. A service is often divided into several parts, including congregational singing, a sermon, intercessory prayer, and other ministry. During worship there is usually a nursery for babies. Children and young people receive an adapted education, Sunday school, in a separate room. Places of worship are usually called "churches". In some megachurches, the building is called "campus". The architecture of places of worship is mainly characterized by its sobriety. The latin cross is one of the only spiritual symbols that can usually be seen on the building of an evangelical church and that identifies the place's belonging. In the United States, the Solomon Awards recognize excellence in the design and renovation of evangelical church buildings at an annual ceremony. Some services take place in theaters, schools or multipurpose rooms, rented for Sunday only. Because of their understanding of the second of the Ten Commandments, evangelicals do not have religious material representations such as statues, icons, or paintings in their places of worship. There is usually a baptistery on the stage of the auditorium (also called sanctuary) or in a separate room, for the baptisms by immersion. In some countries of the world which apply sharia or communism, government authorizations for worship are complex for Evangelical Christians. Because of persecution of Christians, Evangelical house churches have thus developed. For example, there is the Evangelical house churches in China movement. The meetings thus take place in private houses, in secret and in "illegality". The main Christian feasts celebrated by the Evangelicals are Christmas, Pentecost (by a majority of Evangelical denominations) and Easter for all believers. In matters of sexuality, several evangelical churches promote the virginity pledge among young Evangelical Christians, who are invited to commit themselves during a public ceremony at sexual abstinence until Christian marriage. This pledge is often symbolized by a purity ring. In evangelical churches, young adults and unmarried couples are encouraged to marry early in order to live a sexuality according to the will of God. A 2009 American study of the National Campaign to Prevent Teen and Unplanned Pregnancy reported that 80% of young unmarried evangelicals had had relationships and that 42% were in a relationship with sex, when surveyed. The majority of evangelical Christian churches are against voluntary termination of pregnancy and support adoption agencies and social support agencies for young mothers. Masturbation is seen as being prohibited by some evangelical pastors because of the sexual thoughts that may accompany it. In United States and Nigeria, other evangelical pastors believe that masturbation can be beneficial for the body and that it is a gift from God to avoid fornication, especially for singles. Some evangelical churches speak only of sexual abstinence and do not speak of sexuality in marriage. Other evangelical churches in United States and Switzerland speak of satisfying sexuality as a gift from God and a component of a Christian marriage harmonious, in messages during worship services or conferences. Many evangelical books and websites are specialized on the subject. The perceptions of homosexuality in the Evangelical Churches are varied. They range from liberal to conservative, through moderate. It is in the fundamentalist and conservative positions, that there are anti-gay activists on TV or radio who claim that homosexuality is the cause of many social problems, such as terrorism. Some churches have a moderate position. Although they do not approve homosexual practices, they show sympathy and respect for homosexuals. Some evangelical denominations have adopted neutral positions, leaving the choice to local churches to decide for same-sex marriage. For a majority of evangelical Christians, a belief in biblical inerrancy ensures that the miracles described in the Bible are still relevant and may be present in the life of the believer. Healings, academic or professional successes, the birth of a child after several attempts, the end of an addiction, etc., would be tangible examples of God's intervention with the faith and prayer, by the Holy Spirit. In the 1980s, the neo-charismatic movement re-emphasized miracles and faith healing. In certain churches, a special place is thus reserved for faith healings with laying on of hands during worship services or for evangelization campaigns. Faith healing or divine healing is considered to be an inheritance of Jesus acquired by his death and resurrection. In terms of science and the origin of the earth and human life, some evangelicals support creationism. For example, Answers in Genesis, founded in Australia in 1986, is an evangelical organization which defends this thesis. In 2007, she [who?] founded the Creation Museum in Petersburg, in Kentucky and in 2016 the Ark Encounter in Williamstown. Since the end of the 20th century, literalist creationism has been abandoned by some evangelicals in favor of intelligent design. For example, the think tank Discovery Institute, established in 1991 in Seattle, defends this thesis. Other evangelicals believe in evolutionary creation—the notion that God used the process of evolution to create life; a Christian organization that espouses this view is the BioLogos Foundation. As a trans-denominational movement, evangelicalism occurs in nearly every Protestant denomination and tradition. The Reformed, Baptist, Wesleyan, Pentecostal, Churches of Christ, Plymouth Brethren, charismatic Protestant, and nondenominational Protestant traditions have all had strong influence within contemporary evangelicalism. Some Anabaptist denominations (such as the Brethren Church) are evangelical, and some Lutherans self-identify as evangelicals. There are also evangelical Anglicans. In the early 20th century, evangelical influence declined within mainline Protestantism and Christian fundamentalism developed as a distinct religious movement. Between 1950 and 2000 a mainstream evangelical consensus developed that sought to be more inclusive and more culturally relevant than fundamentalism while maintaining conservative Protestant teaching. According to Brian Stanley, professor of world Christianity, this new postwar consensus is termed "neo-evangelicalism", the "new evangelicalism", or simply "evangelicalism" in the United States, while in Great Britain and in other English-speaking countries, it is commonly termed "conservative evangelicalism". Over the years, less-conservative evangelicals have challenged this mainstream consensus to varying degrees. Such movements have been classified by a variety of labels, such as progressive, open, post-conservative, and post-evangelical. Some commentators have complained that Evangelicalism as a movement is too broad and its definition too vague to be of any practical value. Theologian Donald Dayton has called for a "moratorium" on use of the term. Historian D. G. Hart has also argued that "evangelicalism needs to be relinquished as a religious identity because it does not exist". Fundamentalism regards biblical inerrancy, the virgin birth of Jesus, penal substitutionary atonement, the literal resurrection of Christ, and the Second Coming of Christ as fundamental Christian doctrines. Fundamentalism arose among evangelicals in the 1920s to combat modernist or liberal theology in mainline Protestant churches. Failing to reform the mainline churches, fundamentalists separated from them and established their own churches, refusing to participate in ecumenical organizations such as the National Council of Churches (founded in 1950). They also made separatism (rigid separation from non-fundamentalist churches and their culture) a true test of faith. According to historian George Marsden, most fundamentalists are Baptists and dispensationalist. Mainstream evangelicalism is historically divided between two main orientations: confessionalism and revivalism. These two streams have been critical of each other. Confessional evangelicals have been suspicious of unguarded religious experience, while revivalist evangelicals have been critical of overly intellectual teaching that (they suspect) stifles vibrant spirituality. In an effort to broaden their appeal, many contemporary evangelical congregations intentionally avoid identifying with any single form of evangelicalism. These "generic evangelicals" are usually theologically and socially conservative, but their churches often present themselves as nondenominational (or, if a denominational member, strongly de-emphasizing its ties to such, such as a church name which excludes the denominational name) within the broader evangelical movement. In the words of Albert Mohler, president of the Southern Baptist Theological Seminary, confessional evangelicalism refers to "that movement of Christian believers who seek a constant convictional continuity with the theological formulas of the Protestant Reformation". While approving of the evangelical distinctions proposed by Bebbington, confessional evangelicals believe that authentic evangelicalism requires more concrete definition in order to protect the movement from theological liberalism and from heresy. According to confessional evangelicals, subscription to the ecumenical creeds and to the Reformation-era confessions of faith (such as the confessions of the Reformed churches) provides such protection. Confessional evangelicals are represented by conservative Presbyterian churches (emphasizing the Westminster Confession), certain Baptist churches that emphasize historic Baptist confessions such as the Second London Confession, evangelical Anglicans who emphasize the Thirty-Nine Articles (such as in the Anglican Diocese of Sydney, Australia), Methodist churches that adhere to the Articles of Religion, and some confessional Lutherans with pietistic convictions. The emphasis on historic Protestant orthodoxy among confessional evangelicals stands in direct contrast to an anti-creedal outlook that has exerted its own influence on evangelicalism, particularly among churches strongly affected by revivalism and by pietism. Revivalist evangelicals are represented by some quarters of Methodism, the Wesleyan Holiness churches, the Pentecostal/charismatic churches, some Anabaptist churches, and some Baptists and Presbyterians. Revivalist evangelicals tend to place greater emphasis on religious experience than their confessional counterparts. Evangelicals dissatisfied with the movement's conservative mainstream have been variously described as progressive evangelicals, post-conservative evangelicals, Open Evangelicals and post-evangelicals. Progressive evangelicals, also known as the evangelical left, share theological or social views with other progressive Christians while also identifying with evangelicalism. Progressive evangelicals commonly advocate for women's equality, pacifism and social justice. As described by Baptist theologian Roger E. Olson, post-conservative evangelicalism is a theological school of thought that adheres to the four marks of evangelicalism, while being less rigid and more inclusive of other Christians. According to Olson, post-conservatives believe that doctrinal truth is secondary to spiritual experience shaped by Scripture. Post-conservative evangelicals seek greater dialogue with other Christian traditions and support the development of a multicultural evangelical theology that incorporates the voices of women, racial minorities, and Christians in the developing world. Some post-conservative evangelicals also support open theism and the possibility of near universal salvation. The term "Open Evangelical" refers to a particular Christian school of thought or churchmanship, primarily in Great Britain (especially in the Church of England). Open evangelicals describe their position as combining a traditional evangelical emphasis on the nature of scriptural authority, the teaching of the ecumenical creeds and other traditional doctrinal teachings, with an approach towards culture and other theological points-of-view which tends to be more inclusive than that taken by other evangelicals. Some open evangelicals aim to take a middle position between conservative and charismatic evangelicals, while others would combine conservative theological emphases with more liberal social positions. British author Dave Tomlinson coined the phrase "post-evangelical" to describe a movement comprising various trends of dissatisfaction among evangelicals. Others use the term with comparable intent, often to distinguish evangelicals in the so-called emerging church movement from post-evangelicals and anti-evangelicals. Tomlinson argues that "linguistically, the distinction "[between evangelical and post-evangelical]" resembles the one that sociologists make between the modern and postmodern eras". Evangelicalism emerged in the 18th century, first in Britain and its North American colonies. Nevertheless, there were earlier developments within the larger Protestant world that preceded and influenced the later evangelical revivals. According to religion scholar Randall Balmer, Evangelicalism resulted "from the confluence of Pietism, Presbyterianism, and the vestiges of Puritanism. Evangelicalism picked up the peculiar characteristics from each strain – warmhearted spirituality from the Pietists (for instance), doctrinal precisionism from the Presbyterians, and individualistic introspection from the Puritans". Historian Mark Noll adds to this list High Church Anglicanism, which contributed to Evangelicalism a legacy of "rigorous spirituality and innovative organization". During the 17th century, Pietism emerged in Europe as a movement for the revival of piety and devotion within the Lutheran church. As a protest against "cold orthodoxy" or against an overly formal and rational Christianity, Pietists advocated for an experiential religion that stressed high moral standards both for clergy and for lay people. The movement included both Christians who remained in the liturgical, state churches as well as separatist groups who rejected the use of baptismal fonts, altars, pulpits, and confessionals. As Pietism spread, the movement's ideals and aspirations influenced and were absorbed by evangelicals. The Presbyterian heritage not only gave Evangelicalism a commitment to Protestant orthodoxy but also contributed a revival tradition that stretched back to the 1620s in Scotland and northern Ireland. Central to this tradition was the communion season, which normally occurred in the summer months. For Presbyterians, celebrations of Holy Communion were infrequent but popular events preceded by several Sundays of preparatory preaching and accompanied with preaching, singing, and prayers. Puritanism combined Calvinism with a doctrine that conversion was a prerequisite for church membership and with an emphasis on the study of Scripture by lay people. It took root in the colonies of New England, where the Congregational church became an established religion. There the Half-Way Covenant of 1662 allowed parents who had not testified to a conversion experience to have their children baptized, while reserving Holy Communion for converted church members alone. By the 18th century Puritanism was in decline and many ministers expressed alarm at the loss of religious piety. This concern over declining religious commitment led many people to support evangelical revival. High-Church Anglicanism also exerted influence on early Evangelicalism. High Churchmen were distinguished by their desire to adhere to primitive Christianity. This desire included imitating the faith and ascetic practices of early Christians as well as regularly partaking of Holy Communion. High Churchmen were also enthusiastic organizers of voluntary religious societies. Two of the most prominent were the Society for Promoting Christian Knowledge (founded in London in 1698), which distributed Bibles and other literature and built schools, and the Society for the Propagation of the Gospel in Foreign Parts, which was founded in England in 1701 to facilitate missionary work in British colonies (especially among colonists in North America). Samuel and Susanna Wesley, the parents of John and Charles Wesley (born 1703 and 1707 respectively), were both devoted advocates of High-Church ideas. In the 1730s, Evangelicalism emerged as a distinct phenomenon out of religious revivals that began in Britain and New England. While religious revivals had occurred within Protestant churches in the past, the evangelical revivals that marked the 18th century were more intense and radical. Evangelical revivalism imbued ordinary men and women with a confidence and enthusiasm for sharing the gospel and converting others outside of the control of established churches, a key discontinuity with the Protestantism of the previous era. It was developments in the doctrine of assurance that differentiated Evangelicalism from what went before. Bebbington says, "The dynamism of the Evangelical movement was possible only because its adherents were assured in their faith." He goes on: The first local revival occurred in Northampton, Massachusetts, under the leadership of Congregationalist minister Jonathan Edwards. In the fall of 1734, Edwards preached a sermon series on "Justification By Faith Alone", and the community's response was extraordinary. Signs of religious commitment among the laity increased, especially among the town's young people. The revival ultimately spread to 25 communities in western Massachusetts and central Connecticut until it began to wane by the spring of 1735. Edwards was heavily influenced by Pietism, so much so that one historian has stressed his "American Pietism". One practice clearly copied from European Pietists was the use of small groups divided by age and gender, which met in private homes to conserve and promote the fruits of revival. At the same time, students at Yale University (at that time Yale College) in New Haven, Connecticut, were also experiencing revival. Among them was Aaron Burr, Sr., who would become a prominent Presbyterian minister and future president of Princeton University. In New Jersey, Gilbert Tennent, another Presbyterian minister, was preaching the evangelical message and urging the Presbyterian Church to stress the necessity of converted ministers. The spring of 1735 also marked important events in England and Wales. Howell Harris, a Welsh schoolteacher, had a conversion experience on May 25 during a communion service. He described receiving assurance of God's grace after a period of fasting, self-examination, and despair over his sins. Sometime later, Daniel Rowland, the Anglican curate of Llangeitho, Wales, experienced conversion as well. Both men began preaching the evangelical message to large audiences, becoming leaders of the Welsh Methodist revival. At about the same time that Harris experienced conversion in Wales, George Whitefield was converted at Oxford University after his own prolonged spiritual crisis. Whitefield later remarked, "About this time God was pleased to enlighten my soul, and bring me into the knowledge of His free grace, and the necessity of being justified in His sight by "faith only"". Whitefield's fellow Holy Club member and spiritual mentor, Charles Wesley, reported an evangelical conversion in 1738. In the same week, Charles' brother and future founder of Methodism, John Wesley was also converted after a long period of inward struggle. During this spiritual crisis, John Wesley was directly influenced by Pietism. Two years before his conversion, Wesley had traveled to the newly established colony of Georgia as a missionary for the Society for Promoting Christian Knowledge. He shared his voyage with a group of Moravian Brethren led by August Gottlieb Spangenberg. The Moravians' faith and piety deeply impressed Wesley, especially their belief that it was a normal part of Christian life to have an assurance of one's salvation. Wesley recounted the following exchange with Spangenberg on February 7, 1736: Wesley finally received the assurance he had been searching for at a meeting of a religious society in London. While listening to a reading from Martin Luther's preface to the Epistle to the Romans, Wesley felt spiritually transformed: Pietism continued to influence Wesley, who had translated 33 Pietist hymns from German to English. Numerous German Pietist hymns became part of the English Evangelical repertoire. By 1737, Whitefield had become a national celebrity in England where his preaching drew large crowds, especially in London where the Fetter Lane Society had become a center of evangelical activity. Whitfield joined forces with Edwards to "fan the flame of revival" in the Thirteen Colonies in 1739–40. Soon the First Great Awakening stirred Protestants throughout America. Evangelical preachers emphasized personal salvation and piety more than ritual and tradition. Pamphlets and printed sermons crisscrossed the Atlantic, encouraging the revivalists. The Awakening resulted from powerful preaching that gave listeners a sense of deep personal revelation of their need of salvation by Jesus Christ. Pulling away from ritual and ceremony, the Great Awakening made Christianity intensely personal to the average person by fostering a deep sense of spiritual conviction and redemption, and by encouraging introspection and a commitment to a new standard of personal morality. It reached people who were already church members. It changed their rituals, their piety and their self-awareness. To the evangelical imperatives of Reformation Protestantism, 18th century American Christians added emphases on divine outpourings of the Holy Spirit and conversions that implanted within new believers an intense love for God. Revivals encapsulated those hallmarks and forwarded the newly created Evangelicalism into the early republic. By the 1790s, the Evangelical party in the Church of England remained a small minority but were not without influence. John Newton and Joseph Milner were influential evangelical clerics. Evangelical clergy networked together through societies such as the Eclectic Society in London and the Elland Society in Yorkshire. The Old Dissenter denominations (the Baptists, Congregationalists and Quakers) were falling under evangelical influence, with the Baptists most affected and Quakers the least. Evangelical ministers dissatisfied with both Anglicanism and Methodism often chose to work within these churches. In the 1790s, all of these evangelical groups, including the Anglicans, were Calvinist in orientation. Methodism (the "New Dissent") was the most visible expression of evangelicalism by the end of the 18th century. The Wesleyan Methodists boasted around 70,000 members throughout the British Isles, in addition to the Calvinistic Methodists in Wales and the Countess of Huntingdon's Connexion, which was organized under George Whitefield's influence. The Wesleyan Methodists, however, were still nominally affiliated with the Church of England and would not completely separate until 1795, four years after Wesley's death. The Wesleyan Methodist Church's Arminianism distinguished it from the other evangelical groups. At the same time, evangelicals were an important faction within the Presbyterian Church of Scotland. Influential ministers included John Erskine, Henry Wellwood Moncrieff and Stevenson Macgill. The church's General Assembly, however, was controlled by the Moderate Party, and evangelicals were involved in the First and Second Secessions from the national church during the 18th century. The start of the 19th century saw an increase in missionary work and many of the major missionary societies were founded around this time (see Timeline of Christian missions). Both the Evangelical and high church movements sponsored missionaries. The Second Great Awakening (which actually began in 1790) was primarily an American revivalist movement and resulted in substantial growth of the Methodist and Baptist churches. Charles Grandison Finney was an important preacher of this period. In Britain in addition to stressing the traditional Wesleyan combination of "Bible, cross, conversion, and activism", the revivalist movement sought a universal appeal, hoping to include rich and poor, urban and rural, and men and women. Special efforts were made to attract children and to generate literature to spread the revivalist message. "Christian conscience" was used by the British Evangelical movement to promote social activism. Evangelicals believed activism in government and the social sphere was an essential method in reaching the goal of eliminating sin in a world drenched in wickedness. The Evangelicals in the Clapham Sect included figures such as William Wilberforce who successfully campaigned for the abolition of slavery. In the late 19th century, the revivalist Holiness movement, based on the doctrine of "entire sanctification", took a more extreme form in rural America and Canada, where it ultimately broke away from institutional Methodism. In urban Britain the Holiness message was less exclusive and censorious. John Nelson Darby of the Plymouth Brethren was a 19th-century Irish Anglican minister who devised modern dispensationalism, an innovative Protestant theological interpretation of the Bible that was incorporated in the development of modern Evangelicalism. Cyrus Scofield further promoted the influence of dispensationalism through the explanatory notes to his Scofield Reference Bible. According to scholar Mark S. Sweetnam, who takes a cultural studies perspective, dispensationalism can be defined in terms of its Evangelicalism, its insistence on the literal interpretation of Scripture, its recognition of stages in God's dealings with humanity, its expectation of the imminent return of Christ to rapture His saints, and its focus on both apocalypticism and premillennialism. During the 19th century, the megachurches, churches with more than 2,000 people, began to develop. The first evangelical megachurch, the Metropolitan Tabernacle with a 6000-seat auditorium, was inaugurated in 1861 in London by Charles Spurgeon. Dwight L. Moody founded the Illinois Street Church in Chicago. An advanced theological perspective came from the Princeton theologians from the 1850s to the 1920s, such as Charles Hodge, Archibald Alexander and B.B. Warfield. After 1910 the Fundamentalist movement dominated Evangelicalism in the early part of the 20th century; the Fundamentalists rejected liberal theology and emphasized the inerrancy of the Scriptures. Following the 1904–1905 Welsh revival, the Azusa Street Revival in 1906 began the spread of Pentecostalism in North America. The 20th century also marked by the emergence of the televangelism. Aimee Semple McPherson, who founded the megachurch "Angelus Temple" in Los Angeles, used radio in the 1920s to reach a wider audience. After the Scopes trial in 1925, "Christian Century" wrote of "Vanishing Fundamentalism." In 1929 Princeton University, once the bastion of conservative theology, added several modernists to its faculty, resulting in the departure of J. Gresham Machen and a split in the Presbyterian Church in the United States of America. Evangelicalism began to reassert itself in the second half of the 1930s. One factor was the advent of the radio as a means of mass communication. When [Charles E. Fuller] began his "Old Fashioned Revival Hour" on October 3, 1937, he sought to avoid the contentious issues that had caused fundamentalists to be characterized as narrow. One hundred forty-seven representatives from thirty-four denominations met from April 7 through 9, 1942, in St. Louis, Missouri, for a "National Conference for United Action among Evangelicals." The next year six hundred representatives in Chicago established the National Association of Evangelicals (NAE) with Harold Ockenga as its first president. The NAE was partly a reaction to the founding of the American Council of Christian Churches (ACCC) under the leadership of the fundamentalist Carl McIntire. The ACCC in turn had been founded to counter the influence of the Federal Council of Churches (later merged into the National Council of Churches), which fundamentalists saw as increasingly embracing modernism in its ecumenism. Those who established the NAE had come to view the name fundamentalist as "an embarrassment instead of a badge of honor." Evangelical revivalist radio preachers organized themselves in the National Religious Broadcasters in 1944 in order to regulate their activity. With the founding of the NAE, American Protestantism was divided into three large groups—the fundamentalists, the modernists, and the new evangelicals, who sought to position themselves between the other two. In 1947 Harold Ockenga coined the term "neo-evangelicalism" to identify a movement distinct from fundamentalism. The neo-evangelicals had three broad characteristics that distinguished them from the conservative fundamentalism of the ACCC: Each of these characteristics took concrete shape by the mid-1950s. In 1947 Carl F. H. Henry's book "The Uneasy Conscience of Fundamentalism" called on evangelicals to engage in addressing social concerns: In the same year Fuller Theological Seminary was established with Ockenga as its president and Henry as the head of its theology department. The strongest impetus, however, was the development of the work of Billy Graham. Graham had begun his career with the support of McIntire and fellow conservatives Bob Jones Sr. and John R. Rice. However, in broadening the reach of his London crusade of 1954, he accepted the support of denominations that those men disapproved of. When he went even further in his 1957 New York crusade, conservatives strongly condemned him and withdrew their support. According to William Martin: A fourth development—the founding of "Christianity Today" ("CT") with Henry as its first editor—was strategic in giving neo-evangelicals a platform to promote their views and in positioning them between the fundamentalists and modernists. In a letter to Harold Lindsell, Graham said that "CT" would: The post-war period also saw growth of the ecumenical movement and the founding of the World Council of Churches, which the Evangelical community generally regarded with suspicion. In the United Kingdom, John Stott (1921–2011) and Martyn Lloyd-Jones (1899–1981) emerged as key leaders in Evangelical Christianity. The charismatic movemen] began in the 1960s and resulted in the introduction of Pentecostal theology and practice into many mainline denominations. New charismatic groups such as the Association of Vineyard Churches and Newfrontiers trace their roots to this period (see also British New Church Movement). The closing years of the 20th century saw controversial postmodern influences entering some parts of Evangelicalism, particularly with the emerging church movement. Also controversial is the relationship between spiritualism and contemporary military metaphors and practices animating many branches of Christianity but especially relevant in the sphere of Evangelicalism. Spiritual warfare is the latest iteration in a long-standing partnership between religious organization and militarization, two spheres that are rarely considered together, although aggressive forms of prayer have long been used to further the aims of expanding Evangelical influence. Major moments of increased political militarization have occurred concurrently with the growth of prominence of militaristic imagery in evangelical communities. This paradigmatic language, paired with an increasing reliance on sociological and academic research to bolster militarized sensibility, serves to illustrate the violent ethos that effectively underscores militarized forms of evangelical prayer. In Nigeria, evangelical megachurches, such as Redeemed Christian Church of God and Living Faith Church Worldwide, have built autonomous cities with houses, supermarkets, banks, universities, and power plants. Evangelical Christian film production societies were founded in the early 2000s, such as Sherwood Pictures and Pure Flix . The growth of evangelical churches continues with the construction of new places of worship or enlargements in various regions of the world. According to a 2011 Pew Forum study on global Christianity, 285,480,000 or 13.1 percent of all Christians are Evangelicals. These figures do not include the Evangelical movements Pentecostalism and Charismatic movement. The study states that the category "Evangelicals" should not be considered as a separate category of "Pentecostal and Charismatic" categories, since some believers consider themselves in both movements where their church is affiliated with an Evangelical association. In 2015, the World Evangelical Alliance is "a network of churches in 129 nations that have each formed an Evangelical alliance and over 100 international organizations joining together to give a world-wide identity, voice, and platform to more than 600 million Evangelical Christians". The Alliance was formed in 1951 by Evangelicals from 21 countries. It has worked to support its members to work together globally. According to Sébastien Fath of CNRS, in 2016, there are 619 million Evangelicals in the world, one in four Christians. In 2017, about 630 million, an increase of 11 million, including Pentecostals. Operation World estimates the number of Evangelicals at 545.9 million, which makes for 7.9 percent of the world's population. From 1960 to 2000, the global growth of the number of reported Evangelicals grew three times the world's population rate, and twice that of Islam. According to Operation World, the Evangelical population's current annual growth rate is 2.6 percent, still more than twice the world's population growth rate. In the 21st century, there are Evangelical churches active in Sudan, Angola, Mozambique, Zimbabwe, Malawi, Rwanda, Uganda, Ghana, Kenya, Zambia, South Africa, and Nigeria. They have grown especially since independence came in the 1960s, the strongest movements are based on Pentecostal-charismatic beliefs. There is a wide range of theology and organizations, including some sponsored by European missionaries and others that have emerged from African culture such as the Apostolic and Zionist Churches which enlist 40% of black South Africans, and their Aladura counterparts in western Africa. In Nigeria the Evangelical Church Winning All (formerly "Evangelical Church of West Africa") is the largest church organization with five thousand congregations and over three million members. It sponsors two seminaries and eight Bible colleges, and 1600 missionaries who serve in Nigeria and other countries with the Evangelical Missionary Society (EMS). There have been serious confrontations since 1999 between Muslims and Evangelical Christians standing in opposition to the expansion of Sharia law in northern Nigeria. The confrontation has radicalized and politicized the Christians. Violence has been escalating. In Kenya, mainstream Evangelical denominations have taken the lead in promoting political activism and backers, with the smaller Evangelical sects of less importance. Daniel arap Moi was president 1978 to 2002 and claimed to be an Evangelical; he proved intolerant of dissent or pluralism or decentralization of power. The Berlin Missionary Society (BMS) was one of four German Protestant mission societies active in South Africa before 1914. It emerged from the German tradition of Pietism after 1815 and sent its first missionaries to South Africa in 1834. There were few positive reports in the early years, but it was especially active 1859–1914. It was especially strong in the Boer republics. The World War cut off contact with Germany, but the missions continued at a reduced pace. After 1945 the missionaries had to deal with decolonization across Africa and especially with the apartheid government. At all times the BMS emphasized spiritual inwardness, and values such as morality, hard work and self-discipline. It proved unable to speak and act decisively against injustice and racial discrimination and was disbanded in 1972. Since 1974, young professionals have been the active proselytizers of Evangelicalism in the cities of Malawi. In Mozambique, Evangelical Protestant Christianity emerged around 1900 from black migrants whose converted previously in South Africa. They were assisted by European missionaries, but, as industrial workers, they paid for their own churches and proselytizing. They prepared southern Mozambique for the spread of Evangelical Protestantism. During its time as a colonial power in Mozambique, the Catholic Portuguese government tried to counter the spread of Evangelical Protestantism. The East African Revival was a renewal movement within Evangelical churches in East Africa during the late 1920s and 1930s that began at a Church Missionary Society mission station in the Belgian territory of Ruanda-Urundi in 1929, and spread to: Uganda, Tanzania and Kenya during the 1930s and 1940s contributing to the significant growth of the church in East Africa through the 1970s and had a visible influence on Western missionaries who were observer-participants of the movement. In modern Latin America, the term "Evangelical" is often simply a synonym for "Protestant". Protestantism in Brazil largely originated with German immigrants and British and American missionaries in the 19th century, following up on efforts that began in the 1820s. In the late nineteenth century, while the vast majority of Brazilians were nominal Catholics, the nation was underserved by priests, and for large numbers their religion was only nominal. The Catholic Church in Brazil was de-established in 1890, and responded by increasing the number of dioceses and the efficiency of its clergy. Many Protestants came from a large German immigrant community, but they were seldom engaged in proselytism and grew mostly by natural increase. Methodists were active along with Presbyterians and Baptists. The Scottish missionary Dr. Robert Reid Kalley, with support from the Free Church of Scotland, moved to Brazil in 1855, founding the first Evangelical church among the Portuguese-speaking population there in 1856. It was organized according to the Congregational policy as the Igreja Evangélica Fluminense; it became the mother church of Congregationalism in Brazil. The Seventh-day Adventists arrived in 1894, and the YMCA was organized in 1896. The missionaries promoted schools colleges and seminaries, including a liberal arts college in São Paulo, later known as Mackenzie, and an agricultural school in Lavras. The Presbyterian schools in particular later became the nucleus of the governmental system. In 1887 Protestants in Rio de Janeiro formed a hospital. The missionaries largely reached a working-class audience, as the Brazilian upper-class was wedded either to Catholicism or to secularism. By 1914, Protestant churches founded by American missionaries had 47,000 communicants, served by 282 missionaries. In general, these missionaries were more successful than they had been in Mexico, Argentina or elsewhere in Latin America. There were 700,000 Protestants by 1930, and increasingly they were in charge of their own affairs. In 1930, the Methodist Church of Brazil became independent of the missionary societies and elected its own bishop. Protestants were largely from a working-class, but their religious networks help speed their upward social mobility. Protestants accounted for fewer than 5% of the population until the 1960s, but grew exponentially by proselytizing and by 2000 made up over 15% of Brazilians affiliated with a church. Pentecostals and charismatic groups account for the vast majority of this expansion. Pentecostal missionaries arrived early in the 20th century. Pentecostal conversions surged during the 1950s and 1960s, when native Brazilians began founding autonomous churches. The most influential included Brasil Para o Cristo (Brazil for Christ), founded in 1955 by Manoel de Mello. With an emphasis on personal salvation, on God's healing power, and on strict moral codes these groups have developed broad appeal, particularly among the booming urban migrant communities. In Brazil, since the mid-1990s, groups committed to uniting black identity, antiracism, and Evangelical theology have rapidly proliferated. Pentecostalism arrived in Brazil with Swedish and American missionaries in 1911. it grew rapidly, but endured numerous schisms and splits. In some areas the Evangelical Assemblies of God churches have taken a leadership role in politics since the 1960s. They claimed major credit for the election of Fernando Collor de Mello as president of Brazil in 1990. According to the 2000 Census, 15.4% of the Brazilian population was Protestant. A recent research conducted by the Datafolha institute shows that 25% of Brazilians are Protestants, of which 19% are followers of Pentecostal denominations. The 2010 Census found out that 22.2% were Protestant at that date. Protestant denominations saw a rapid growth in their number of followers since the last decades of the 20th century. They are politically and socially conservative, and emphasize that God's favor translates into business success. The rich and the poor remained traditional Catholics, while most Evangelical Protestants were in the new lower-middle class–known as the "C class" (in a A–E classification system). Chesnut argues that Pentecostalism has become "one of the principal organizations of the poor," for these churches provide the sort of social network that teach members the skills they need to thrive in a rapidly developing meritocratic society. One large Evangelical church that originated from Brasil is the Universal Church of the Kingdom of God (IURD), a neo‐Pentecostal denomination begun in 1977. It now has a presence in many countries, and claims millions of members worldwide. Protestants remained a small portion of the population until the late-twentieth century, when various Protestant groups experienced a demographic boom that coincided with the increasing violence of the Guatemalan Civil War. Two former Guatemalan heads of state, General Efraín Ríos Montt and Jorge Serrano Elías have been practicing Evangelical Protestants, as is Guatemala's current President, Jimmy Morales. General Montt, an Evangelical from the Pentecostal tradition, came to power through a coup. He escalated the war against leftist guerrilla insurgents as a holy war against atheistic "forces of evil". Protestant missionary activity in Asia was most successful in Korea. American Presbyterians and Methodists arrived in the 1880s and were well received. Between 1910 and 1945, when Korea was a Japanese colony, Christianity became in part an expression of nationalism in opposition to Japan's efforts to enforce the Japanese language and the Shinto religion. In 1914, out of 16 million people, there were 86,000 Protestants and 79,000 Catholics; by 1934, the numbers were 168,000 and 147,000. Presbyterian missionaries were especially successful. Since the Korean War (1950–53), many Korean Christians have migrated to the U.S., while those who remained behind have risen sharply in social and economic status. Most Korean Protestant churches in the 21st century emphasize their Evangelical heritage. Korean Protestantism is characterized by theological conservatism coupled with an emotional revivalistic style. Most churches sponsor revival meetings once or twice a year. Missionary work is a high priority, with 13,000 men and women serving in missions across the world, putting Korea in second place just behind the US. Sukman argues that since 1945, Protestantism has been widely seen by Koreans as the religion of the middle class, youth, intellectuals, urbanites, and modernists. It has been a powerful force supporting South Korea's pursuit of modernity and emulation of the United States, and opposition to the old Japanese colonialism and to the authoritarianism of North Korea. South Korea has been referred as an "evangelical superpower" for being the home to some of the largest and most dynamic Christian churches in the world; South Korea is also second to the U.S. in the number of missionaries sent abroad. According to 2015 South Korean census, 9.7 million or 19.7% of the population described themselves as Protestants, many of whom belong to Presbyterian churches shaped by Evangelicalism. According to the 2010 census, 2.68 percent of Filipinos are Evangelicals. The Philippine Council of Evangelical Churches (PCEC), an organization of more than seventy Evangelical and Mainline Protestant churches, and more than 210 para-church organizations in the Philippines, counts more than 11 million members as of 2011. In 2019, it was reported that Evangelicalism in France is growing and a new Evangelical church is built every 10 days and now counts 700,000 followers across France. John Wesley (1703–1791) was an Anglican cleric and theologian who, with his brother Charles Wesley (1707–1788) and fellow cleric George Whitefield (1714 – 1770), founded Methodism. After 1791 the movement became independent of the Anglican Church as the "Methodist Connection". It became a force in its own right, especially among the working class. The "Clapham Sect" was a group of Church of England evangelicals and social reformers based in Clapham, London; they were active 1780s–1840s). John Newton (1725–1807) was the founder. They are described by the historian Stephen Tomkins as "a network of friends and families in England, with William Wilberforce as its centre of gravity, who were powerfully bound together by their shared moral and spiritual values, by their religious mission and social activism, by their love for each other, and by marriage". Evangelicalism was a major force in the Anglican Church from about 1800 to the 1860s. By 1848 when an evangelical John Bird Sumner became Archbishop of Canterbury, between a quarter and a third of all Anglican clergy were linked to the movement, which by then had diversified greatly in its goals and they were no longer considered an organized faction. In the 21st century there are an estimated 2 million Evangelicals in the UK. According to research performed by the Evangelical Alliance in 2013, 87% of UK evangelicals attend Sunday morning church services every week and 63% attend weekly or fortnightly small groups. An earlier survey conducted in 2012 found that 92% of evangelicals agree it is a Christian's duty to help those in poverty and 45% attend a church which has a fund or scheme that helps people in immediate need, and 42% go to a church that supports or runs a foodbank. 63% believe in tithing, and so give around 10% of their income to their church, Christian organizations and various charities 83% of UK evangelicals believe that the Bible has supreme authority in guiding their beliefs, views and behavior and 52% read or listen to the Bible daily. The Evangelical Alliance, formed in 1846, was the first ecumenical evangelical body in the world and works to unite evangelicals, helping them listen to, and be heard by, the government, media and society. By the late 19th to early 20th century, most American Protestants were Evangelicals. A bitter divide had arisen between the more liberal-modernist mainline denominations and the fundamentalist denominations, the latter typically consisting of Evangelicals. Key issues included the truth of the Bible—literal or figurative, and teaching of evolution in the schools. During and after World War II, Evangelicals became increasingly organized. There was a great expansion of Evangelical activity within the United States, "a revival of revivalism." Youth for Christ was formed; it later became the base for Billy Graham's revivals. The National Association of Evangelicals formed in 1942 as a counterpoise to the mainline Federal Council of Churches. In 1942–43, the Old-Fashioned Revival Hour had a record-setting national radio audience. With this organization, though, fundamentalist groups separated from Evangelicals. According to a Pew Forum on Religion and Public Life study, Evangelicals can be broadly divided into three camps: traditionalist, centrist, and modernist. A 2004 Pew survey identified that while 70.4% of Americans call themselves "Christian," Evangelicals only make up 26.3 percent of the population, while Catholics make up 22 percent and mainline Protestants make up 16 percent. Evangelicals have been socially active throughout US history, a tradition dating back to the abolitionist movement of the Antebellum period and the prohibition movement. As a group, evangelicals are most often associated with the Christian right. However, a large number of black self-labeled Evangelicals, and a small proportion of liberal white self-labeled Evangelicals, gravitate towards the Christian left. Recurrent themes within American Evangelical discourse include abortion, the creation–evolution controversy, secularism, and the notion of the United States as a Christian nation. Today, evangelicals are conservative leaning, primarily Republican voting, individuals who faithfully attend church service. Evangelicals’ religious faith and biblical interpretations are often used as a tool to justify their political views. Evangelicals’ political views more so than their religious beliefs often guide their positions on social issues such as: abortion, immigration, gay marriage, gender roles, affirmative action, race, criminal justice, the death penalty, and the privatization of public education. At the beginning of the 20th century, evangelical churches developed the importance of social justice and Christian humanitarian aid. The majority of evangelical Christian humanitarian organizations were founded in the second half of the 20th century. Among the most important are World Vision International (1950), Samaritan's Purse (1970), Mercy Ships (1978), Prison Fellowship International (1979), International Justice Mission (1997). A particularly controversial doctrine in the Evangelical Churches is that of the prosperity theology, which spread in the 1970s and 1980s in the United States, mainly through televangelism. This doctrine is centered on the teaching of Christian faith as a means to enrich oneself financially and materially, through a "positive confession" and a contribution to Christian ministries. Promises of divine healing and prosperity are guaranteed in exchange for certain amounts of donations. Fidelity in the tithe would allow one to avoid the curses of God, the attacks of the devil and poverty. The offerings and the tithe occupies a lot of time in the worship services. Often associated with the tithe mandatory, this doctrine is sometimes compared to a religious business. It is criticized by pastors and church unions, such as the National Council of Evangelicals of France, in France. Since the 1970s, various financial scandals of embezzlement have been reported in churches and evangelical organizations. The Evangelical Council for Financial Accountability was founded in 1979 to strengthen financial integrity in evangelical organizations and churches that voluntarily wish to be members and to undergo annual accounting audits. In some evangelical churches, pricing of prayer against promises of healing has been found. Some pastors and evangelists have been charged with claiming false healings. Some churches in United States or Nigeria forbid the faithful from vaccination or medical treatment. In 2019, in Mbandjock in Cameroon, three deaths have been linked to this position in a revival church. This position is not representative of all evangelical churches, as the document indicates "The Miraculous Healing" published in 2015 by the National Council of Evangelicals of France, which mentions that medicine is one of the gifts of God made to humans. Churches and certain evangelical humanitarian organizations are also involved in medical health programs. Some churches and evangelical organizations have been criticized by victims of rape and domestic violence for their silent handling of cases of abuse by pastors or members. Failure to report abuses to the police appears to be prevalent in churches which are not members of evangelical Christian denomination, or affiliated to denominations which attach great importance to a large autonomy of the churches and which thus have a low external accountability relationship. The evangelical organization GRACE was founded in 2004 by the Baptist professor Boz Tchividjian to help churches combat sexual abuses, psychological abuses and physical abuses in Christian organizations. The fact that evangelicals do evangelism and speak about their faith in public is often criticized by the media and associated with proselytism. According to the evangelicals, freedom of religion and freedom of expression allow them to talk about their faith like anything else. Christian films made by American evangelical production companies are also regularly associated with proselytism. According to Sarah-Jane Murray, screenwriting teacher at the US Film and Christian Television Commission United, Christian films are works of art, not proselytism. For Hubert de Kerangat, communications manager at Saje distribution, distributor of these American Christian films in France, if Christian films are "proselytes", all films are "proselytes", since each film transmits a message, whether the viewer is free to approve or not.
https://en.wikipedia.org/wiki?curid=10370
Euphonium The euphonium is a medium-sized, 3 or 4-valve, often compensating, conical-bore, tenor-voiced brass instrument that derives its name from the Ancient Greek word "euphōnos", meaning "well-sounding" or "sweet-voiced" ( "eu" means "well" or "good" and "phōnē" means "sound", hence "of good sound"). The euphonium is a valved instrument. Nearly all current models have piston valves, though some models with rotary valves do exist. The euphonium may be played in bass clef as a non-transposing instrument or in treble clef as a transposing instrument. In British brass bands, it is typically treated as a treble-clef instrument, while in American band music, parts may be written in either treble clef or bass clef, or both. The euphonium is in the family of brass instruments, more particularly low-brass instruments with many relatives. It is extremely similar to a baritone horn. The difference is that the bore size of the baritone horn is typically smaller than that of the euphonium, and the baritone is a primarily cylindrical bore, whereas the euphonium is predominantly conical bore. It is controversial whether this is sufficient to make them two different instruments. In the trombone family large and small bore trombones are both called trombones, while the cylindrical trumpet and the conical flugelhorn are given different names. As with the trumpet and flugelhorn, the two instruments are easily doubled by one player, with some modification of breath and embouchure, since the two have identical range and essentially identical fingering. The cylindrical baritone offers a brighter sound and the conical euphonium offers a more mellow sound. The "American baritone", featuring three valves on the front of the instrument and a curved, forward-pointing bell, was dominant in American school bands throughout most of the 20th century, its weight, shape, and configuration conforming to the needs of the marching band. While this instrument is a conical-cylindrical bore hybrid, somewhere between the classic baritone horn and euphonium, it was almost universally labelled a "baritone" by both band directors and composers, thus contributing to the confusion of terminology in the United States. Several late 19th century music catalogs (such as Pepper and Lyon & Healy) sold a euphonium-like instrument called the "B bass" (to distinguish it from the E and BB bass). In these catalog drawings, the B Bass had thicker tubing than the baritone; both had three valves. Along the same lines, drum and bugle corps introduced the "Bass-baritone", and distinguished it from the baritone. The thicker tubing of the three-valve B bass allowed for production of strong false-tones, providing chromatic access to the pedal register. Ferdinand Sommer's original name for the instrument was the "euphonion". It is sometimes called the tenor tuba in B, although this can also refer to other varieties of tuba. Names in other languages, as included in scores, can be ambiguous as well. They include French "basse", "saxhorn basse", and "tuba basse"; German "Baryton", "Tenorbass", and "Tenorbasshorn"; Italian "baritono", "bombardino", "eufonio", and "flicorno basso". The most common German name, "Baryton", may have influenced Americans to adopt the name "baritone" for the instrument, due to the influx of German musicians to the United States in the nineteenth century. As a baritone-voiced brass instrument, the euphonium traces its ancestry to the ophicleide and ultimately back to the serpent. The search for a satisfactory foundational wind instrument that could support massed sound above its pitch took many years. While the serpent was used for over two centuries dating back to the late Renaissance, it was notoriously difficult to control its pitch and tone quality due to its disproportionately small open finger holes. The ophicleide, which was used in bands and orchestras for a few decades in the early to mid-19th century, used a system of keys and was an improvement over the serpent but was still unreliable, especially in the high register. With the invention of the piston valve system 1818, the construction of brass instruments with an even sound and facility of playing in all registers became possible. The euphonium is said to have been invented, as a "wide-bore, valved bugle of baritone range", by Ferdinand Sommer of Weimar in 1843, though Carl Moritz in 1838 and Adolphe Sax in 1843 have also been credited. While Sax's family of saxhorns were invented at about the same time and the bass saxhorn is very similar to a euphonium, there are also differences—such as the bass saxhorn being narrower throughout the length of the instrument. The "British-style" compensating euphonium was developed by David Blaikley in 1874, and has been in use in Britain with the basic construction little changed since then. Modern day euphonium makers have been working to further enhance the construction of the euphonium. Companies such as Adams and Besson have been leading the way in perfecting the instrument. Adams euphoniums have developed an adjustable lead pipe receiver which allows players to change the timbre of the instrument to whatever they find preferable. Besson has also been credited with the adjustable main tuning slide trigger, which allows players more flexibility with intonation. The euphonium, like the tenor trombone, is pitched in concert B. For a valved brass instrument like the euphonium, this means that when no valves are in use the instrument will produce partials of the B harmonic series. It is generally orchestrated as a non-transposing instrument like the trombone, written at concert pitch in the bass clef with higher passages in the tenor clef. Treble clef euphonium parts transposing down a major ninth are included in much concert band music: in the British-style brass band tradition, euphonium music is always written this way. In continental European band music, parts for the euphonium may be written in the bass clef as a B transposing instrument sounding a major second lower than written. Professional models have three top-action valves, played with the first three fingers of the right hand, plus a "compensating" fourth valve, generally found midway down the right side of the instrument, played with the left index finger; such an instrument is shown at the top of this page. Beginner models often have only the three top-action valves, while some intermediate "student" models may have a fourth top-action valve, played with the fourth finger of the right hand. Compensating systems are expensive to build, and there is in general a substantial difference in price between compensating and non-compensating models. For a thorough discussion of the valves and the compensation system, see the article on brass instruments. The euphonium has an extensive range, comfortably from E2 to about E4 for intermediate players (using scientific pitch notation). In professional hands this may extend from B0 to as high as B5. The lowest notes obtainable depend on the valve set-up of the instrument. All instruments are chromatic down to E2, but four-valved instruments extend that down to at least C2. Non-compensating four-valved instruments suffer from intonation problems from E2 down to C2 and cannot produce the low B1; compensating instruments do not have such intonation problems and can play the low B1. From B1 down lies the "pedal range", i.e., the fundamentals of the instrument's harmonic series. They are easily produced on the euphonium as compared to other brass instruments, and the extent of the range depends on the make of the instrument in exactly the same way as just described. Thus, on a compensating four-valved instrument, the lowest note possible is B0, sometimes called double pedal B, which is six ledger lines below the bass clef. As with the other conical-bore instruments, the cornet, flugelhorn, horn, and tuba, the euphonium's tubing (excepting the tubing in the valve section, which is necessarily cylindrical) gradually increases in diameter throughout its length, resulting in a softer, gentler tone compared to cylindrical-bore instruments such as the trumpet, trombone, sudrophone, and baritone horn. While a truly characteristic euphonium sound is rather hard to define precisely, most players would agree that an ideal sound is dark, rich, warm, and velvety, with virtually no hardness to it. This also has to do with the different models preferred by British and American players. Though the euphonium's fingerings are no different from those of the trumpet or tuba, beginning euphoniumists will likely experience significant problems with intonation, response and range compared to other beginning brass players. In addition, it is very difficult for students, even of high-school age, to develop the rich sound characteristic of the euphonium, due partly to the instrument models used in schools and partly to the lack of awareness of good euphonium sound models. The compensating euphonium is common among professionals. It utilizes a three-plus-one-valve system with three upright valves and one side valve. The compensating valve system uses extra tubing, usually coming off of the back of the three upright valves, in order to achieve proper intonation in the lower range of the instrument. This range being from E2 down to B1. Not all four-valve and three-plus-one-valve euphoniums are compensating. Only those designed with extra tubing are compensating. There were, at one time, three-valve compensating euphoniums available. This configuration utilized extra tubing, just as the three-plus-one compensating models did, in order to bring the notes C2 and B1 in tune. This three-valve compensating configuration is still available in British style baritone horns, usually on professional models. A creation unique to the United States was the double-bell euphonium, featuring a second smaller bell in addition to the main one; the player could switch bells for certain passages or even for individual notes by use of an additional valve, operated with the left hand. Ostensibly, the smaller bell was intended to emulate the sound of a trombone (it was cylindrical-bore) and was possibly intended for performance situations in which trombones were not available. The extent to which the difference in sound and timbre was apparent to the listener, however, is up for debate. "Harry Whittier" of the Patrick S. Gilmore band introduced the instrument in 1888, and it was used widely in both school and service bands for several decades. "Harold Brasch" (see "List of important players" below) brought the British-style compensating euphonium to the United States c. 1939, but the double-belled euphonium may have remained in common use even into the 1950s and 1960s. In any case, they have become rare (they were last in Conn's advertisements in the 1940s, and King's catalog in the 1960s), and are generally unknown to younger players. They are chiefly known now through their mention in the song "Seventy-Six Trombones" from the musical "The Music Man" by Meredith Willson. A marching version of the euphonium may be found in a marching band, though it is often replaced by its smaller, easier-to-carry cousin, the marching baritone (which has a similar bell and valve configuration to a trumpet). Marching euphoniums are used by marching bands in schools, and in Drum and Bugle Corps, and some corps (such as the Blue Devils and Phantom Regiment) march all-euphonium sections rather than only marching Baritone or a mix of both. Depending on the manufacturer, the weight of these instruments can be straining to the average marcher and require great strength to hold during practices and performances, leading to nerve problems in the right pinky, a callus on the left hand, and possibly back and arm problems. Marching euphoniums and marching baritones commonly have 3 valves, opposed to the regular euphonium having 4. Another form of the marching euphonium is the convertible euphonium. Recently widely produced, the horn resembles a convertible tuba, being able to change from a concert upright to a marching forward bell on either the left or right shoulder. These are mainly produced by Jupiter or Yamaha, but other less expensive versions can be found. The five-valve euphonium (noncompensating) is an extremely rare variation of the euphonium manufactured in the late 19th and early 20th centuries by Britain's Besson musical instrument company and Highams of Manchester Musical Instrument Company. Higham and Besson's "Clearbore" five-valve euphonium was economical but not widely used. The Besson five-valve euphonium featured the standard three piston valves horizontally on top, but had an additional two piston valves off to the side. The standard euphonium has eight possible fingering and non-fingering positions by which sound is produced. The Besson and the Highams "clearbore" model rare fourth and fifth extra "side" valves change the possible fingering and non-fingering positions from eight to thirty-two. The term 'five-valve euphonium' does not refer to variations of the double bell euphonium made by various brass instrument companies during the same time period. Some of the double-bell euphoniums had five valves, with the fifth valve either on top with the other four, or by itself off to the side, but the double-bell fifth valve was used for switching the sound to the second smaller trombone-sized bell, and not for changing the fingering pitch of the instrument. Also, Musical Instruments manufactures several euphoniums with five vertical rotary valves today, but this is an unrelated recent development. Euphonium has historically been exclusively a "band" instrument (rather than an "orchestra" or "jazz" instrument), whether of the "wind" or "brass" variety, where it is frequently featured as a solo instrument. It has been called the "king of band instruments", or the "cello of the band", because of its similarity in timbre and ensemble role to the stringed instrument. Euphoniums are widely used in many marches (such as those by John Philip Sousa), and in brass band music of the British tradition. Other ensembles including euphonium are the tuba-euphonium quartet or larger tuba-euphonium ensemble; the brass quintet, where it can supply the tenor voice, though the trombone is much more common in this role; and other mixed brass ensemble. Such ensembles are almost non-existent: they are more likely to be semi-professional or amateur in nature. Most of the United States Armed Forces service bands include a tuba-euphonium quartet made up of players from the band that occasionally performs in its own right. The euphonium is not traditionally an orchestral instrument and has not been common in symphony orchestras. However, there are a few works from the late Romantic period, in which composers wrote a part for "tenor tuba", all of which are played on the euphonium: most notably Gustav Holst's "Planets Suite" and "Ein Heldenleben" by Richard Strauss. In addition, the euphonium is sometimes used in older orchestral works as a replacement for its predecessors, such as the ophicleide, or, less correctly, the bass trumpet or the Wagner tuba, both of which are significantly different instruments, and still in use today. While the euphonium was not historically part of the standard jazz big band or combo, the instrument's technical facility and large range make it well-suited to a jazz solo role, and a jazz euphonium niche has been carved out over the last 40 or so years, largely starting with the pioneer Rich Matteson (see "List of important players" below). The euphonium can also double on a trombone part in a jazz combo. Jazz euphoniums are most likely to be found in tuba-euphonium groups, though modern funk or rock bands occasionally feature a brass player doubling on euphonium, and this trend is growing. Due to this dearth of performance opportunities, aspiring euphonium players in the United States are in a rather inconvenient position when seeking future employment. Often, college players must either obtain a graduate degree and go on to teach at the college level, or audition for one of the major or regional military service bands. Because these bands are relatively few in number and the number of euphonium positions in the bands is small (2–4 in most service bands), job openings do not occur very often and when they do are highly competitive; before the current slate of openings in four separate bands, the last opening for a euphonium player in an American service band was in May 2004. A career strictly as a solo performer, unaffiliated with any university or performing ensemble, is a very rare sight, but some performers, such as Riki McDonnell, have managed to do it. In Britain, Australia and New Zealand the strongest euphonium players are most likely to find a position in a brass band, but even though they often play at world-class levels, the members of the top brass bands are in most cases unpaid amateurs. There are hundreds, if not thousands, of brass bands in Britain (and the Commonwealth) ranging in standard from world class to local bands. Even The Salvation Army has strong ties to the brass band world, as this was a common and practical musical genre in the 1800s. Almost all brass bands in Britain perform regularly, particularly during the summer months. A large number of bands also enter contests against other brass bands of a similar standard. Each band requires two euphoniums (principal and second) and consequently there are considerable opportunities for euphonium players. Due to limited vocational opportunities, there are a considerable number of relatively serious, quasi-professional avocational euphonium players participating in many higher-caliber unpaid ensembles. Unlike a generation or two ago, many colleges with music programs now offer students the opportunity to major in euphonium. However, due to the smaller numbers of euphonium students at most schools, it is possible, and even likely, that they will study with a professor whose major instrument is not the euphonium. Most often tubas and euphoniums will be combined into a studio taught by one professor, and at small schools they may be grouped with trombones and/or French horns as well, taught by one low brass professor. Universities will usually require professors in this situation to have a high level of proficiency on all the instruments they teach, and some of the best college euphonium studios are taught by non-euphonium players. German Ferdinand Sommer, if one discounts the claims of Moritz and Sax each of whose horns also approached a euphonium in nature, in addition to being credited with inventing the euphonium as the Sommerhorn in 1843, as a soloist on the horn, qualifies as the first euphonium player to significantly advance and alter the understanding of the instrument. The "euphonium repertoire" consists of solo literature and orchestral, or, more commonly, concert band parts written for the euphonium. Since its invention in 1843, the euphonium has always had an important role in ensembles, but solo literature was slow to appear, consisting of only a handful of lighter solos until the 1960s. Since then, however, the breadth and depth of the solo euphonium repertoire has increased dramatically. In the current age, there has been a huge number of new commissions and repertoire development and promotion through Steven Mead's World of the Euphonium Series and the Beyond the Horizon series from Euphonium.com. There has also been a vast number of new commissions by more and more players and a proliferation of large scale Consortium Commissions that are occurring including current ones in 2008 and 2009 organized by Brian Meixner (Libby Larson), Adam Frey (The Euphonium Foundation Consortium), and Jason Ham (David Gillingham). Upon its invention, it was clear that the euphonium had, compared to its predecessors the serpent and ophicleide, a wide range and had a consistently rich, pleasing sound throughout that range. It was flexible both in tone quality and intonation and could blend well with a variety of ensembles, gaining it immediate popularity with composers and conductors as the principal tenor-voices solo instrument in brass band settings, especially in Britain. It is no surprise, then, that when British composers – some of the same ones who were writing for brass bands – began to write serious, original music for the concert band in the early 20th century, they used the euphonium in a very similar role. When American composers also began writing for the concert band as its own artistic medium in the 1930s and 1940s, they continued the British brass and concert band tradition of using the euphonium as the principal tenor-voiced solo. This is not to say that composers, then and now, valued the euphonium only for its lyrical capabilities. Indeed, examination of a large body of concert band literature reveals that the euphonium functions as a "jack of all trades." Though the euphonium was, as previously noted, embraced from its earliest days by composers and arrangers in band settings, orchestral composers have, by and large, not taken advantage of this capability. There are, nevertheless, several orchestral works, a few of which are standard repertoire, in which composers have called for instruments, such as the Wagner tuba, for which euphonium is commonly substituted in the present. In contrast to the long-standing practice of extensive euphonium use in wind bands and orchestras, there was, until approximately forty years ago, literally no body of solo literature written specifically for the euphonium, and euphonium players were forced to borrow the literature of other instruments. Fortunately, given the instrument's multifaceted capabilities discussed above, solos for many different instruments are easily adaptable to performance on the euphonium. The earliest surviving solo composition written specifically for euphonium or one of its saxhorn cousins is the "Concerto per Flicorno Basso" (1872) by Amilcare Ponchielli. For almost a century after this, the euphonium solo repertoire consisted of only a dozen or so virtuosic pieces, mostly light in character. However, in the 1960s and 1970s, American composers began to write the first of the "new school" of serious, artistic solo works specifically for euphonium. Since then, there has been a virtual explosion of solo repertoire for the euphonium. In a mere four decades, the solo literature has expanded from virtually zero to thousands of pieces. More and more composers have become aware of the tremendous soloistic capabilities of the euphonium, and have constantly "pushed the envelope" with new literature in terms of tessitura, endurance, technical demands, and extended techniques. Finally, the euphonium has, thanks to a handful of enterprising individuals, begun to make inroads in jazz, pop and other non-concert performance settings.
https://en.wikipedia.org/wiki?curid=10371
Entire function In complex analysis, an entire function, also called an integral function, is a complex-valued function that is holomorphic at all finite points over the whole complex plane. Typical examples of entire functions are polynomials and the exponential function, and any finite sums, products and compositions of these, such as the trigonometric functions sine and cosine and their hyperbolic counterparts sinh and cosh, as well as derivatives and integrals of entire functions such as the error function. If an entire function "f"("z") has a root at "w", then "f"("z")/("z−w"), taking the limit value at "w", is an entire function. On the other hand, neither the natural logarithm nor the square root is an entire function, nor can they be continued analytically to an entire function. A transcendental entire function is an entire function that is not a polynomial. Every entire function "f"("z") can be represented as a power series that converges everywhere in the complex plane, hence uniformly on compact sets. The radius of convergence is infinite, which implies that or Any power series satisfying this criterion will represent an entire function. If (and only if) the coefficients of the power series are all real then the function evidently takes real values for real arguments, and the value of the function at the complex conjugate of "z" will be the complex conjugate of the value at "z". Such functions are sometimes called self-conjugate (the conjugate function, formula_4, being given by formula_5 If the real part of an entire function is known in a neighborhood of a point then both the real and imaginary parts are known for the whole complex plane, up to an imaginary constant. For instance, if the real part is known in a neighborhood of zero, then we can find the coefficients for "n" > 0 from the following derivatives with respect to a real variable "r": (Likewise, if the imaginary part is known in a neighborhood then the function is determined up to a real constant.) In fact, if the real part is known just on an arc of a circle, then the function is determined up to an imaginary constant. (For instance, if the real part is known on part of the unit circle, then it is known on the whole unit circle by analytic extension, and then the coefficients of the infinite series are determined from the coefficients of the Fourier series for the real part on the unit circle.) Note however that an entire function is not determined by its real part on all curves. In particular, if the real part is given on any curve in the complex plane where the real part of some other entire function is zero, then any multiple of that function can be added to the function we are trying to determine. For example, if the curve where the real part is known is the real line, then we can add "i" times any self-conjugate function. If the curve forms a loop, then it is determined by the real part of the function on the loop since the only functions whose real part is zero on the curve are those that are everywhere equal to some imaginary number. The Weierstrass factorization theorem asserts that any entire function can be represented by a product involving its zeroes (or "roots"). The entire functions on the complex plane form an integral domain (in fact a Prüfer domain). They also form a commutative unital associative algebra over the complex numbers. Liouville's theorem states that any bounded entire function must be constant. Liouville's theorem may be used to elegantly prove the fundamental theorem of algebra. As a consequence of Liouville's theorem, any function that is entire on the whole Riemann sphere (complex plane "and" the point at infinity) is constant. Thus any non-constant entire function must have a singularity at the complex point at infinity, either a pole for a polynomial or an essential singularity for a transcendental entire function. Specifically, by the Casorati–Weierstrass theorem, for any transcendental entire function "f" and any complex "w" there is a sequence formula_7 such that Picard's little theorem is a much stronger result: any non-constant entire function takes on every complex number as value, possibly with a single exception. When an exception exists, it is called a lacunary value of the function. The possibility of a lacunary value is illustrated by the exponential function, which never takes on the value 0. One can take a suitable branch of the logarithm of an entire function that never hits 0, so that this will also be an entire function (according to the Weierstrass factorization theorem). The logarithm hits every complex number except possibly one number, which implies that the first function will hit any value other than 0 an infinite number of times. Similarly, a non-constant, entire function that does not hit a particular value will hit every other value an infinite number of times. Liouville's theorem is a special case of the following statement: Entire functions may grow as fast as any increasing function: for any increasing function "g": [0,∞) → [0,∞) there exists an entire function "f" such that "f"("x") > "g"(|"x"|) for all real "x". Such a function "f" may be easily found of the form: for a constant "c" and a strictly increasing sequence of positive integers "nk". Any such sequence defines an entire function "f"("z"), and if the powers are chosen appropriately we may satisfy the inequality "f"("x") > "g"(|"x"|) for all real "x". (For instance, it certainly holds if one chooses "c" := "g"(2) and, for any integer formula_16 one chooses an even exponent formula_17 such that formula_18). The order (at infinity) of an entire function formula_19 is defined using the limit superior as: where "Br" is the disk of radius "r" and formula_21 denotes the supremum norm of formula_19 on "Br". The order is a non-negative real number or infinity (except when formula_23 for all "z"). In other words, the order of formula_19 is the infimum of all "m" such that: The example of formula_26 shows that this does not mean "f"("z") = O(exp(|"z"|"m")) if formula_19 is of order "m". If formula_28 one can also define the type: If the order is 1 and the type is "σ", the function is said to be "of exponential type "σ"". If it is of order less than 1 it is said to be of exponential type 0. If then the order and type can be found by the formulas Let formula_32 denote the "n"th derivative of "f", then we may restate these formulas in terms of the derivatives at any arbitrary point "z"0: The type may be infinite, as in the case of the reciprocal gamma function, or zero (see example below under #Order 1). Here are some examples of functions of various orders: For arbitrary positive numbers "ρ" and "σ" one can construct an example of an entire function of order "ρ" and type "σ" using: where where Entire functions of finite order have Hadamard's canonical representation: where formula_43 are those roots of formula_44 that are not zero (formula_45), formula_46 is the order of the zero of formula_44 at formula_48 (the case formula_49 being taken to mean formula_50), formula_51 a polynomial (whose degree we shall call formula_52), and formula_53 is the smallest non-negative integer such that the series converges. The non-negative integer formula_55 is called the genus of the entire function formula_44. If the order ρ is not an integer, then formula_57 is the integer part of formula_58. If the order is a positive integer, then there are two possibilities: formula_59 or formula_60. For example, formula_61 and formula_62 are entire functions of genus "1". According to J. E. Littlewood, the Weierstrass sigma function is a 'typical' entire function. This statement can be made precise in the theory of random entire functions: the asymptotic behavior of almost all entire functions is similar to that of the sigma function. Other examples include the Fresnel integrals, the Jacobi theta function, and the reciprocal Gamma function. The exponential function and the error function are special cases of the Mittag-Leffler function. According to the fundamental theorem of Paley and Wiener, Fourier transforms of functions (or distributions) with bounded support are entire functions of order "1" and finite type. Other examples are solutions of linear differential equations with polynomial coefficients. If the coefficient at the highest derivative is constant, then all solutions of such equations are entire functions. For example, the exponential function, sine, cosine, Airy functions and Parabolic cylinder functions arise in this way. The class of entire functions is closed with respect to compositions. This makes it possible to study dynamics of entire functions. An entire function of the square root of a complex number is entire if the original function is even, for example formula_63. If a sequence of polynomials all of whose roots are real converges in a neighborhood of the origin to a limit which is not identically equal to zero, then this limit is an entire function. Such entire functions form the Laguerre–Pólya class, which can also be characterized in terms of the Hadamard product, namely, "f" belongs to this class if and only if in the Hadamard representation all "zn" are real, "p" ≤ 1, and "P"("z") = "a" + "bz" + "cz"2, where "b" and "c" are real, and "c" ≤ 0. For example, the sequence of polynomials converges, as "n" increases, to exp(−("z"−"d")2). The polynomials have all real roots, and converge to cos("z"). The polynomials also converge to cos("z"), showing the buildup of the Hadamard product for cosine.
https://en.wikipedia.org/wiki?curid=10372
Essay An essay is, generally, a piece of writing that gives the author's own argument — but the definition is vague, overlapping with those of a paper, an article, a pamphlet, and a short story. Essays have traditionally been sub-classified as formal and informal. Formal essays are characterized by "serious purpose, dignity, logical organization, length," whereas the informal essay is characterized by "the personal element (self-revelation, individual tastes and experiences, confidential manner), humor, graceful style, rambling structure, unconventionality or novelty of theme," etc. Essays are commonly used as literary criticism, political manifestos, learned arguments, observations of daily life, recollections, and reflections of the author. Almost all modern essays are written in prose, but works in verse have been dubbed essays (e.g., Alexander Pope's "An Essay on Criticism" and "An Essay on Man"). While brevity usually defines an essay, voluminous works like John Locke's "An Essay Concerning Human Understanding" and Thomas Malthus's "An Essay on the Principle of Population" are counterexamples. In some countries (e.g., the United States and Canada), essays have become a major part of formal education. Secondary students are taught structured essay formats to improve their writing skills; admission essays are often used by universities in selecting applicants, and in the humanities and social sciences essays are often used as a way of assessing the performance of students during final exams. The concept of an "essay" has been extended to other media beyond writing. A film essay is a movie that often incorporates documentary filmmaking styles and focuses more on the evolution of a theme or idea. A photographic essay covers a topic with a linked series of photographs that may have accompanying text or captions. The word "essay" derives from the French infinitive , "to try" or "to attempt". In English "essay" first meant "a trial" or "an attempt", and this is still an alternative meaning. The Frenchman Michel de Montaigne (1533–1592) was the first author to describe his work as essays; he used the term to characterize these as "attempts" to put his thoughts into writing. Subsequently, "essay" has been defined in a variety of ways. One definition is a "prose composition with a focused subject of discussion" or a "long, systematic discourse". It is difficult to define the genre into which essays fall. Aldous Huxley, a leading essayist, gives guidance on the subject. He notes that "the essay is a literary device for saying almost everything about almost anything", and adds that "by tradition, almost by definition, the essay is a short piece". Furthermore, Huxley argues that "essays belong to a literary species whose extreme variability can be studied most effectively within a three-poled frame of reference". These three poles (or worlds in which the essay may exist) are: Huxley adds that the most satisfying essays "...make the best not of one, not of two, but of all the three worlds in which it is possible for the essay to exist." Montaigne's "attempts" grew out of his commonplacing. Inspired in particular by the works of Plutarch, a translation of whose "Œuvres Morales" ("Moral works") into French had just been published by Jacques Amyot, Montaigne began to compose his essays in 1572; the first edition, entitled "Essais", was published in two volumes in 1580. For the rest of his life, he continued revising previously published essays and composing new ones. A third volume was published posthumously; together, their over 100 examples are widely regarded as the predecessor of the modern essay. While Montaigne's philosophy was admired and copied in France, none of his most immediate disciples tried to write essays. But Montaigne, who liked to fancy that his family (the Eyquem line) was of English extraction, had spoken of the English people as his "cousins", and he was early read in England, notably by Francis Bacon. Bacon's essays, published in book form in 1597 (only five years after the death of Montaigne, containing the first ten of his essays), 1612, and 1625, were the first works in English that described themselves as "essays". Ben Jonson first used the word "essayist" in 1609, according to the "Oxford English Dictionary". Other English essayists included Sir William Cornwallis, who published essays in 1600 and 1617 that were popular at the time, Robert Burton (1577–1641) and Sir Thomas Browne (1605–1682). In Italy, Baldassare Castiglione wrote about courtly manners in his essay "Il Cortigiano". In the 17th century, the Spanish Jesuit Baltasar Gracián wrote about the theme of wisdom. In England, during the Age of Enlightenment, essays were a favored tool of polemicists who aimed at convincing readers of their position; they also featured heavily in the rise of periodical literature, as seen in the works of Joseph Addison, Richard Steele and Samuel Johnson. Addison and Steele used the journal "Tatler" (founded in 1709 by Steele) and its successors as storehouses of their work, and they became the most celebrated eighteenth-century essayists in England. Johnson's essays appear during the 1750s in various similar publications. As a result of the focus on journals, the term also acquired a meaning synonymous with "article", although the content may not the strict definition. On the other hand, Locke's "An Essay Concerning Human Understanding" is not an essay at all, or cluster of essays, in the technical sense, but still it refers to the experimental and tentative nature of the inquiry which the philosopher was undertaking. In the 18th and 19th centuries, Edmund Burke and Samuel Taylor Coleridge wrote essays for the general public. The early 19th century, in particular, saw a proliferation of great essayists in English—William Hazlitt, Charles Lamb, Leigh Hunt and Thomas de Quincey all penned numerous essays on diverse subjects, reviving the earlier graceful style. Later in the century, Robert Louis Stevenson also raised the form's literary level. In the 20th century, a number of essayists, such as T.S. Eliot, tried to explain the new movements in art and culture by using essays. Virginia Woolf, Edmund Wilson, and Charles du Bos wrote literary criticism essays. In France, several writers produced longer works with the title of that were not true examples of the form. However, by the mid-19th century, the "Causeries du lundi", newspaper columns by the critic Sainte-Beuve, are literary essays in the original sense. Other French writers followed suit, including Théophile Gautier, Anatole France, Jules Lemaître and Émile Faguet. As with the novel, essays existed in Japan several centuries before they developed in Europe with a genre of essays known as "zuihitsu" — loosely connected essays and fragmented ideas. Zuihitsu have existed since almost the beginnings of Japanese literature. Many of the most noted early works of Japanese literature are in this genre. Notable examples include "The Pillow Book" (c. 1000), by court lady Sei Shōnagon, and "Tsurezuregusa" (1330), by particularly renowned Japanese Buddhist monk Yoshida Kenkō. Kenkō described his short writings similarly to Montaigne, referring to them as "nonsensical thoughts" written in "idle hours". Another noteworthy difference from Europe is that women have traditionally written in Japan, though the more formal, Chinese-influenced writings of male writers were more prized at the time. This section describes the different forms and styles of essay writing. These are used by an array of authors, including university students and professional essayists. The defining features of a "cause and effect" essay are causal chains that connect from a cause to an effect, careful language, and chronological or emphatic order. A writer using this rhetorical method must consider the subject, determine the purpose, consider the audience, think critically about different causes or consequences, consider a thesis statement, arrange the parts, consider the language, and decide on a conclusion. Classification is the categorization of objects into a larger whole while division is the breaking of a larger whole into smaller parts. Compare and contrast essays are characterized by a basis for comparison, points of comparison, and analogies. It is grouped by the object (chunking) or by point (sequential). The comparison highlights the similarities between two or more similar objects while contrasting highlights the differences between two or more objects. When writing a compare/contrast essay, writers need to determine their purpose, consider their audience, consider the basis and points of comparison, consider their thesis statement, arrange and develop the comparison, and reach a conclusion. Compare and contrast is arranged emphatically. Expository essay is used to inform, describe or explain a topic, using important facts and teaching reader about the topic. Mostly written in third-person, using "it", "he", "she", "they". Expository essay uses formal language to discuss someone or something. Examples of expository essays are: a medical or biological condition, social or technological process, life or character of a famous person. Writing of expository essay often consists of following next steps: organizing thoughts (brainstorming), researching a topic, developing a thesis statement, writing the introduction, writing the body of essay, writing the conclusion. Expository essays are often assigned as a part of SAT and other standardized testings or as a homework for high school and college students. Descriptive writing is characterized by sensory details, which appeal to the physical senses, and details that appeal to a reader's emotional, physical, or intellectual sensibilities. Determining the purpose, considering the audience, creating a dominant impression, using descriptive language, and organizing the description are the rhetorical choices to consider when using a description. A description is usually arranged spatially but can also be chronological or emphatic. The focus of a description is the scene. Description uses tools such as denotative language, connotative language, figurative language, metaphor, and simile to arrive at a dominant impression. One university essay guide states that "descriptive writing says what happened or what another author has discussed; it provides an account of the topic". Lyric essays are an important form of descriptive essays. In the dialectic form of the essay, which is commonly used in philosophy, the writer makes a thesis and argument, then objects to their own argument (with a counterargument), but then counters the counterargument with a final and novel argument. This form benefits from presenting a broader perspective while countering a possible flaw that some may present. This type is sometimes called an ethics paper. An exemplification essay is characterized by a generalization and relevant, representative, and believable examples including anecdotes. Writers need to consider their subject, determine their purpose, consider their audience, decide on specific examples, and arrange all the parts together when writing an exemplification essay. An essayist writes a "familiar essay" if speaking to a single reader, writing about both themselves, and about particular subjects. Anne Fadiman notes that "the genre's heyday was the early nineteenth century," and that its greatest exponent was Charles Lamb. She also suggests that while critical essays have more brain than the heart, and personal essays have more heart than brain, familiar essays have equal measures of both. A history essay sometimes referred to as a thesis essay describes an argument or claim about one or more historical events and supports that claim with evidence, arguments, and references. The text makes it clear to the reader why the argument or claim is as such. A narrative uses tools such as flashbacks, flash-forwards, and transitions that often build to a climax. The focus of a narrative is the plot. When creating a narrative, authors must determine their purpose, consider their audience, establish their point of view, use dialogue, and organize the narrative. A narrative is usually arranged chronologically. An argumentative essay is a critical piece of writing, aimed at presenting objective analysis of the subject matter, narrowed down to a single topic. The main idea of all the criticism is to provide an opinion either of positive or negative implication. As such, a critical essay requires research and analysis, strong internal logic and sharp structure. Its structure normally builds around introduction with a topic's relevance and a thesis statement, body paragraphs with arguments linking back to the main thesis, and conclusion. In addition, an argumentative essay may include a refutation section where conflicting ideas are acknowledged, described, and criticized. Each argument of argumentative essay should be supported with sufficient evidence, relevant to the point. A process essay is used for an explanation of making or breaking something. Often, it is written in chronological order or numerical order to show step-by-step processes. It has all the qualities of a technical document with the only difference is that it is often written in descriptive mood, while a technical document is mostly in imperative mood. An economic essay can start with a thesis, or it can start with a theme. It can take a narrative course and a descriptive course. It can even become an argumentative essay if the author feels the need. After the introduction, the author has to do his/her best to expose the economic matter at hand, to analyze it, evaluate it, and draw a conclusion. If the essay takes more of a narrative form then the author has to expose each aspect of the economic puzzle in a way that makes it clear and understandable for the reader A "reflective essay" is an analytical piece of writing in which the writer describes a real or imaginary scene, event, interaction, passing thought, memory, or form — adding a personal reflection on the meaning of the topic in the author's life. Thus, the focus is not merely descriptive. The writer doesn't just describe the situation, but revisits the scene with more detail and emotion to examine what went well, or reveal a need for additional learning — and may relate what transpired to the rest of the author's life. The logical progression and organizational structure of an essay can take many forms. Understanding how the movement of thought is managed through an essay has a profound impact on its overall cogency and ability to impress. A number of alternative logical structures for essays have been visualized as diagrams, making them easy to implement or adapt in the construction of an argument. In countries like the United States and the United Kingdom, essays have become a major part of a formal education in the form of free response questions. Secondary students in these countries are taught structured essay formats to improve their writing skills, and essays are often used by universities in these countries in selecting applicants ("see" admissions essay). In both secondary and tertiary education, essays are used to judge the mastery and comprehension of the material. Students are asked to explain, comment on, or assess a topic of study in the form of an essay. In some courses, university students must complete one or more essays over several weeks or months. In addition, in fields such as the humanities and social sciences, mid-term and end of term examinations often require students to write a short essay in two or three hours. In these countries, so-called academic essays, also called "papers", are usually more formal than literary ones. They may still allow the presentation of the writer's own views, but this is done in a logical and factual manner, with the use of the first person often discouraged. Longer academic essays (often with a word limit of between 2,000 and 5,000 words) are often more discursive. They sometimes begin with a short summary analysis of what has previously been written on a topic, which is often called a literature review. Longer essays may also contain an introductory page that defines words and phrases of the essay's topic. Most academic institutions require that all substantial facts, quotations, and other supporting material in an essay be referenced in a bibliography or works cited page at the end of the text. This scholarly convention helps others (whether teachers or fellow scholars) to understand the basis of facts and quotations the author uses to support the essay's argument and helps readers evaluate to what extent the argument is supported by evidence, and to evaluate the quality of that evidence. The academic essay tests the student's ability to present their thoughts in an organized way and is designed to test their intellectual capabilities. One of the challenges facing universities is that in some cases, students may submit essays purchased from an essay mill (or "paper mill") as their own work. An "essay mill" is a ghostwriting service that sells pre-written essays to university and college students. Since plagiarism is a form of academic dishonesty or academic fraud, universities and colleges may investigate papers they suspect are from an essay mill by using plagiarism detection software, which compares essays against a database of known mill essays and by orally testing students on the contents of their papers. Essays often appear in magazines, especially magazines with an intellectual bent, such as "The Atlantic" and "Harpers". Magazine and newspaper essays use many of the essay types described in the section on forms and styles (e.g., descriptive essays, narrative essays, etc.). Some newspapers also print essays in the op-ed section. Employment essays detailing experience in a certain occupational field are required when applying for some jobs, especially government jobs in the United States. Essays known as Knowledge Skills and Executive Core Qualifications are required when applying to certain US federal government positions. A KSA, or "Knowledge, Skills, and Abilities," is a series of narrative statements that are required when applying to Federal government job openings in the United States. KSAs are used along with resumes to determine who the best applicants are when several candidates qualify for a job. The knowledge, skills, and abilities necessary for the successful performance of a position are contained on each job vacancy announcement. KSAs are brief and focused essays about one's career and educational background that presumably qualify one to perform the duties of the position being applied for. An Executive Core Qualification, or ECQ, is a narrative statement that is required when applying to Senior Executive Service positions within the US Federal government. Like the KSAs, ECQs are used along with resumes to determine who the best applicants are when several candidates qualify for a job. The Office of Personnel Management has established five executive core qualifications that all applicants seeking to enter the Senior Executive Service must demonstrate. A film essay (or "cinematic essay") consists of the evolution of a theme or an idea rather than a plot per se, or the film literally being a cinematic accompaniment to a narrator reading an essay. From another perspective, an essay film could be defined as a documentary film visual basis combined with a form of commentary that contains elements of self-portrait (rather than autobiography), where the signature (rather than the life story) of the filmmaker is apparent. The cinematic essay often blends documentary, fiction, and experimental film making using tones and editing styles. The genre is not well-defined but might include propaganda works of early Soviet parliamentarians like Dziga Vertov, present-day filmmakers including Chris Marker, Michael Moore ("Roger & Me", "Bowling for Columbine" and "Fahrenheit 9/11"), Errol Morris ("The Thin Blue Line"), Morgan Spurlock ("Supersize Me") and Agnès Varda. Jean-Luc Godard describes his recent work as "film-essays". Two filmmakers whose work was the antecedent to the cinematic essay include Georges Méliès and Bertolt Brecht. Méliès made a short film ("The Coronation of Edward VII" (1902)) about the 1902 coronation of King Edward VII, which mixes actual footage with shots of a recreation of the event. Brecht was a playwright who experimented with film and incorporated film projections into some of his plays. Orson Welles made an essay film in his own pioneering style, released in 1974, called "F for Fake", which dealt specifically with art forger Elmyr de Hory and with the themes of deception, "fakery," and authenticity in general. These are often published online on video hosting services. David Winks Gray's article "The essay film in action" states that the "essay film became an identifiable form of filmmaking in the 1950s and '60s". He states that since that time, essay films have tended to be "on the margins" of the filmmaking the world. Essay films have a "peculiar searching, questioning tone ... between documentary and fiction" but without "fitting comfortably" into either genre. Gray notes that just like written essays, essay films "tend to marry the personal voice of a guiding narrator (often the director) with a wide swath of other voices". The University of Wisconsin Cinematheque website echoes some of Gray's comments; it calls a film essay an "intimate and allusive" genre that "catches filmmakers in a pensive mood, ruminating on the margins between fiction and documentary" in a manner that is "refreshingly inventive, playful, and idiosyncratic". In the realm of music, composer Samuel Barber wrote a set of "Essays for Orchestra," relying on the form and content of the music to guide the listener's ear, rather than any extra-musical plot or story. A photographic essay strives to cover a topic with a linked series of photographs. Photo essays range from purely photographic works to photographs with captions or small notes to full-text essays with a few or many accompanying photographs. Photo essays can be sequential in nature, intended to be viewed in a particular order — or they may consist of non-ordered photographs viewed all at once or in an order that the viewer chooses. All photo essays are collections of photographs, but not all collections of photographs are photo essays. Photo essays often address a certain issue or attempt to capture the character of places and events. In the visual arts, an essay is a preliminary drawing or sketch that forms a basis for a final painting or sculpture, made as a test of the work's composition (this meaning of the term, like several of those following, comes from the word "essay"'s meaning of "attempt" or "trial").
https://en.wikipedia.org/wiki?curid=10374
Error detection and correction In information theory and coding theory with applications in computer science and telecommunication, error detection and correction or error control are techniques that enable reliable delivery of digital data over unreliable communication channels. Many communication channels are subject to channel noise, and thus errors may be introduced during transmission from the source to a receiver. Error detection techniques allow detecting such errors, while error correction enables reconstruction of the original data in many cases. "Error detection " is the detection of errors caused by noise or other impairments during transmission from the transmitter to the receiver. "Error correction" is the detection of errors and reconstruction of the original, error-free data. The modern development of error correction codes is credited to Richard Hamming in 1947. A description of Hamming's code appeared in Claude Shannon's "A Mathematical Theory of Communication" and was quickly generalized by Marcel J. E. Golay. All error-detection and correction schemes add some redundancy (i.e., some extra data) to a message, which receivers can use to check consistency of the delivered message, and to recover data that has been determined to be corrupted. Error-detection and correction schemes can be either systematic or non-systematic. In a systematic scheme, the transmitter sends the original data, and attaches a fixed number of "check bits" (or "parity data"), which are derived from the data bits by some deterministic algorithm. If only error detection is required, a receiver can simply apply the same algorithm to the received data bits and compare its output with the received check bits; if the values do not match, an error has occurred at some point during the transmission. In a system that uses a non-systematic code, the original message is transformed into an encoded message carrying the same information and that has at least as many bits as the original message. Good error control performance requires the scheme to be selected based on the characteristics of the communication channel. Common channel models include memoryless models where errors occur randomly and with a certain probability, and dynamic models where errors occur primarily in bursts. Consequently, error-detecting and correcting codes can be generally distinguished between "random-error-detecting/correcting" and "burst-error-detecting/correcting". Some codes can also be suitable for a mixture of random errors and burst errors. If the channel characteristics cannot be determined, or are highly variable, an error-detection scheme may be combined with a system for retransmissions of erroneous data. This is known as automatic repeat request (ARQ), and is most notably used in the Internet. An alternate approach for error control is hybrid automatic repeat request (HARQ), which is a combination of ARQ and error-correction coding. There are three major types of error correction. Automatic Repeat reQuest (ARQ) is an error control method for data transmission that makes use of error-detection codes, acknowledgment and/or negative acknowledgment messages, and timeouts to achieve reliable data transmission. An "acknowledgment" is a message sent by the receiver to indicate that it has correctly received a data frame. Usually, when the transmitter does not receive the acknowledgment before the timeout occurs (i.e., within a reasonable amount of time after sending the data frame), it retransmits the frame until it is either correctly received or the error persists beyond a predetermined number of retransmissions. Three types of ARQ protocols are Stop-and-wait ARQ, Go-Back-N ARQ, and Selective Repeat ARQ. ARQ is appropriate if the communication channel has varying or unknown capacity, such as is the case on the Internet. However, ARQ requires the availability of a back channel, results in possibly increased latency due to retransmissions, and requires the maintenance of buffers and timers for retransmissions, which in the case of network congestion can put a strain on the server and overall network capacity. For example, ARQ is used on shortwave radio data links in the form of ARQ-E, or combined with multiplexing as ARQ-M. Forward error correction (FEC) is a process of adding redundant data such as an error-correcting code (ECC) to a message so that it can be recovered by a receiver even when a number of errors (up to the capability of the code being used) were introduced, either during the process of transmission, or on storage. Since the receiver does not have to ask the sender for retransmission of the data, a backchannel is not required in forward error correction, and it is therefore suitable for simplex communication such as broadcasting. Error-correcting codes are frequently used in lower-layer communication, as well as for reliable storage in media such as CDs, DVDs, hard disks, and RAM. Error-correcting codes are usually distinguished between convolutional codes and block codes: Shannon's theorem is an important theorem in forward error correction, and describes the maximum information rate at which reliable communication is possible over a channel that has a certain error probability or signal-to-noise ratio (SNR). This strict upper limit is expressed in terms of the channel capacity. More specifically, the theorem says that there exist codes such that with increasing encoding length the probability of error on a discrete memoryless channel can be made arbitrarily small, provided that the code rate is smaller than the channel capacity. The code rate is defined as the fraction "k/n" of "k" source symbols and "n" encoded symbols. The actual maximum code rate allowed depends on the error-correcting code used, and may be lower. This is because Shannon's proof was only of existential nature, and did not show how to construct codes which are both optimal and have efficient encoding and decoding algorithms. Hybrid ARQ is a combination of ARQ and forward error correction. There are two basic approaches: The latter approach is particularly attractive on an erasure channel when using a rateless erasure code. Error detection is most commonly realized using a suitable hash function (or specifically, a checksum, cyclic redundancy check or other algorithm). A hash function adds a fixed-length "tag" to a message, which enables receivers to verify the delivered message by recomputing the tag and comparing it with the one provided. There exists a vast variety of different hash function designs. However, some are of particularly widespread use because of either their simplicity or their suitability for detecting certain kinds of errors (e.g., the cyclic redundancy check's performance in detecting burst errors). A random-error-correcting code based on minimum distance coding can provide a strict guarantee on the number of detectable errors, but it may not protect against a preimage attack. A repetition code is a coding scheme that repeats the bits across a channel to achieve error-free communication. Given a stream of data to be transmitted, the data are divided into blocks of bits. Each block is transmitted some predetermined number of times. For example, to send the bit pattern "1011", the four-bit block can be repeated three times, thus producing "1011 1011 1011". If this twelve-bit pattern was received as "1010 1011 1011" – where the first block is unlike the other two – an error has occurred. A repetition code is very inefficient, and can be susceptible to problems if the error occurs in exactly the same place for each group (e.g., "1010 1010 1010" in the previous example would be detected as correct). The advantage of repetition codes is that they are extremely simple, and are in fact used in some transmissions of numbers stations. A "parity bit" is a bit that is added to a group of source bits to ensure that the number of set bits (i.e., bits with value 1) in the outcome is even or odd. It is a very simple scheme that can be used to detect single or any other odd number (i.e., three, five, etc.) of errors in the output. An even number of flipped bits will make the parity bit appear correct even though the data is erroneous. Extensions and variations on the parity bit mechanism are longitudinal redundancy checks, transverse redundancy checks, and similar bit-grouping techniques. A "checksum" of a message is a modular arithmetic sum of message code words of a fixed word length (e.g., byte values). The sum may be negated by means of a ones'-complement operation prior to transmission to detect unintentional all-zero messages. Checksum schemes include parity bits, check digits, and longitudinal redundancy checks. Some checksum schemes, such as the Damm algorithm, the Luhn algorithm, and the Verhoeff algorithm, are specifically designed to detect errors commonly introduced by humans in writing down or remembering identification numbers. A "cyclic redundancy check" (CRC) is a non-secure hash function designed to detect accidental changes to digital data in computer networks. It is not suitable for detecting maliciously introduced errors. It is characterized by specification of a "generator polynomial", which is used as the divisor in a polynomial long division over a finite field, taking the input data as the dividend. The remainder becomes the result. A CRC has properties that make it well suited for detecting burst errors. CRCs are particularly easy to implement in hardware and are therefore commonly used in computer networks and storage devices such as hard disk drives. The parity bit can be seen as a special-case 1-bit CRC. The output of a "cryptographic hash function", also known as a "message digest", can provide strong assurances about data integrity, whether changes of the data are accidental (e.g., due to transmission errors) or maliciously introduced. Any modification to the data will likely be detected through a mismatching hash value. Furthermore, given some hash value, it is typically infeasible to find some input data (other than the one given) that will yield the same hash value. If an attacker can change not only the message but also the hash value, then a "keyed hash" or message authentication code (MAC) can be used for additional security. Without knowing the key, it is not possible for the attacker to easily or conveniently calculate the correct keyed hash value for a modified message. Any error-correcting code can be used for error detection. A code with "minimum Hamming distance", "d", can detect up to "d" − 1 errors in a code word. Using minimum-distance-based error-correcting codes for error detection can be suitable if a strict limit on the minimum number of errors to be detected is desired. Codes with minimum Hamming distance "d" = 2 are degenerate cases of error-correcting codes, and can be used to detect single errors. The parity bit is an example of a single-error-detecting code. Applications that require low latency (such as telephone conversations) cannot use automatic repeat request (ARQ); they must use forward error correction (FEC). By the time an ARQ system discovers an error and re-transmits it, the re-sent data will arrive too late to be usable. Applications where the transmitter immediately forgets the information as soon as it is sent (such as most television cameras) cannot use ARQ; they must use FEC because when an error occurs, the original data is no longer available. Applications that use ARQ must have a return channel; applications having no return channel cannot use ARQ. Applications that require extremely low error rates (such as digital money transfers) must use ARQ due to the possibility of uncorrectable errors with FEC. Reliability and inspection engineering also make use of the theory of error-correcting codes. In a typical TCP/IP stack, error control is performed at multiple levels: Development of error-correction codes was tightly coupled with the history of deep-space missions due to the extreme dilution of signal power over interplanetary distances, and the limited power availability aboard space probes. Whereas early missions sent their data uncoded, starting from 1968, digital error correction was implemented in the form of (sub-optimally decoded) convolutional codes and Reed–Muller codes. The Reed–Muller code was well suited to the noise the spacecraft was subject to (approximately matching a bell curve), and was implemented at the Mariner spacecraft for missions between 1969 and 1977. The Voyager 1 and Voyager 2 missions, which started in 1977, were designed to deliver color imaging amongst scientific information of Jupiter and Saturn. This resulted in increased coding requirements, and thus, the spacecraft were supported by (optimally Viterbi-decoded) convolutional codes that could be concatenated with an outer Golay (24,12,8) code. The Voyager 2 craft additionally supported an implementation of a Reed–Solomon code: the concatenated Reed–Solomon–Viterbi (RSV) code allowed for very powerful error correction, and enabled the spacecraft's extended journey to Uranus and Neptune. Both crafts used V2 RSV coding due to ECC system upgrades after 1989. The CCSDS currently recommends usage of error correction codes with performance similar to the Voyager 2 RSV code as a minimum. Concatenated codes are increasingly falling out of favor with space missions, and are replaced by more powerful codes such as Turbo codes or LDPC codes. The different kinds of deep space and orbital missions that are conducted suggest that trying to find a "one size fits all" error correction system will be an ongoing problem for some time to come. For missions close to Earth, the nature of the noise in the communication channel is different from that which a spacecraft on an interplanetary mission experiences. Additionally, as a spacecraft increases its distance from Earth, the problem of correcting for noise gets bigger. The demand for satellite transponder bandwidth continues to grow, fueled by the desire to deliver television (including new channels and high-definition television) and IP data. Transponder availability and bandwidth constraints have limited this growth, because transponder capacity is determined by the selected modulation scheme and forward error correction (FEC) rate. Overview Error detection and correction codes are often used to improve the reliability of data storage media. A "parity track" was present on the first magnetic tape data storage in 1951. The "Optimal Rectangular Code" used in group coded recording tapes not only detects but also corrects single-bit errors. Some file formats, particularly archive formats, include a checksum (most often CRC32) to detect corruption and truncation and can employ redundancy and/or parity files to recover portions of corrupted data. Reed Solomon codes are used in compact discs to correct errors caused by scratches. Modern hard drives use CRC codes to detect and Reed–Solomon codes to correct minor errors in sector reads, and to recover data from sectors that have "gone bad" and store that data in the spare sectors. RAID systems use a variety of error correction techniques to correct errors when a hard drive completely fails. Filesystems such as ZFS or Btrfs, as well as some RAID implementations, support data scrubbing and resilvering, which allows bad blocks to be detected and (hopefully) recovered before they are used. The recovered data may be re-written to exactly the same physical location, to spare blocks elsewhere on the same piece of hardware, or the data may be rewritten onto replacement hardware. DRAM memory may provide stronger protection against soft errors by relying on error correcting codes. Such error-correcting memory, known as "ECC" or "EDAC-protected" memory, is particularly desirable for mission-critical applications, such as scientific computing, financial, medical, etc. as well as deep-space applications due to the increased radiation in space. Error-correcting memory controllers traditionally use Hamming codes, although some use triple modular redundancy. Interleaving allows distributing the effect of a single cosmic ray potentially upsetting multiple physically neighboring bits across multiple words by associating neighboring bits to different words. As long as a single event upset (SEU) does not exceed the error threshold (e.g., a single error) in any particular word between accesses, it can be corrected (e.g., by a single-bit error correcting code), and the illusion of an error-free memory system may be maintained. In addition to hardware providing features required for ECC memory to operate, operating systems usually contain related reporting facilities that are used to provide notifications when soft errors are transparently recovered. An increasing rate of soft errors might indicate that a DIMM module needs replacing, and such feedback information would not be easily available without the related reporting capabilities. One example is the Linux kernel's "EDAC" subsystem (previously known as "Bluesmoke"), which collects the data from error-checking-enabled components inside a computer system; beside collecting and reporting back the events related to ECC memory, it also supports other checksumming errors, including those detected on the PCI bus. A few systems also support memory scrubbing.
https://en.wikipedia.org/wiki?curid=10375
Euclidean domain In mathematics, more specifically in ring theory, a Euclidean domain (also called a Euclidean ring) is an integral domain that can be endowed with a Euclidean function which allows a suitable generalization of the Euclidean division of the integers. This generalized Euclidean algorithm can be put to many of the same uses as Euclid's original algorithm in the ring of integers: in any Euclidean domain, one can apply the Euclidean algorithm to compute the greatest common divisor of any two elements. In particular, the greatest common divisor of any two elements exists and can be written as a linear combination of them (Bézout's identity). Also every ideal in a Euclidean domain is principal, which implies a suitable generalization of the fundamental theorem of arithmetic: every Euclidean domain is a unique factorization domain. It is important to compare the class of Euclidean domains with the larger class of principal ideal domains (PIDs). An arbitrary PID has much the same "structural properties" of a Euclidean domain (or, indeed, even of the ring of integers), but when an explicit algorithm for Euclidean division is known, one may use Euclidean algorithm and extended Euclidean algorithm to compute greatest common divisors and Bézout's identity. In particular, the existence of efficient algorithms for Euclidean division of integers and of polynomials in one variable over a field is of basic importance in computer algebra. So, given an integral domain , it is often very useful to know that has a Euclidean function: in particular, this implies that is a PID. However, if there is no "obvious" Euclidean function, then determining whether is a PID is generally a much easier problem than determining whether it is a Euclidean domain. Euclidean domains appear in the following chain of class inclusions: Let be an integral domain. A Euclidean function on is a function from } to the non-negative integers satisfying the following fundamental division-with-remainder property: A Euclidean domain is an integral domain which can be endowed with at least one Euclidean function. It is important to note that a particular Euclidean function is "not" part of the structure of a Euclidean domain; in general, a Euclidean domain will admit many different Euclidean functions. Most algebra texts require a Euclidean function to have the following additional property: However, one can show that (EF1) alone suffices to define a Euclidean domain, since any domain which can be endowed with a function satisfying (EF1) can also be endowed with a proper Euclidean function; that is, a function satisfying (EF1) and (EF2). Indeed, for }, one can define as follows: In words, one may define to be the minimum value attained by on the set of all non-zero elements of the principal ideal generated by . A multiplicative Euclidean function is a Euclidean function such that and is never zero. It follows that . More generally, if and only if is a unit. Many authors use other terms in place of "Euclidean function", such as "degree function", "valuation function", "gauge function" or "norm function". Some authors also require the domain of the Euclidean function to be the entire ring ; however this does not essentially affect the definition, since (EF1) does not involve the value of . The definition is sometimes generalized by allowing the Euclidean function to take its values in any well-ordered set; this weakening does not affect the most important implications of the Euclidean property. The property (EF1) can be restated as follows: for any principal ideal of with nonzero generator , all nonzero classes of the quotient ring have a representative with . Since the possible values of are well-ordered, this property can be established by proving that for any with minimal value of in its class. Note that, for a Euclidean function that is so established, there need not exist an effective method to determine and in (EF1). Examples of Euclidean domains include: Examples of domains that are "not" Euclidean domains include: Let "R" be a domain and "f" a Euclidean function on "R". Then: However, in many finite extensions of Q with trivial class group, the ring of integers is Euclidean (not necessarily with respect to the absolute value of the field norm; see below). Assuming the extended Riemann hypothesis, if "K" is a finite extension of Q and the ring of integers of "K" is a PID with an infinite number of units, then the ring of integers is Euclidean. In particular this applies to the case of totally real quadratic number fields with trivial class group. In addition (and without assuming ERH), if the field "K" is a Galois extension of Q, has trivial class group and unit rank strictly greater than three, then the ring of integers is Euclidean. An immediate corollary of this is that if the number field is Galois over Q, its class group is trivial and the extension has degree greater than 8 then the ring of integers is necessarily Euclidean. Algebraic number fields "K" come with a canonical norm function on them: the absolute value of the field norm "N" that takes an algebraic element "α" to the product of all the conjugates of "α". This norm maps the ring of integers of a number field "K", say "O""K", to the nonnegative rational integers, so it is a candidate to be a Euclidean norm on this ring. If this norm satisfies the axioms of a Euclidean function then the number field "K" is called "norm-Euclidean" or simply "Euclidean". Strictly speaking it is the ring of integers that is Euclidean since fields are trivially Euclidean domains, but the terminology is standard. If a field is not norm-Euclidean then that does not mean the ring of integers is not Euclidean, just that the field norm does not satisfy the axioms of a Euclidean function. In fact, the rings of integers of number fields may be divided in several classes: The norm-Euclidean quadratic fields have been fully classified; they are formula_8, where "d" takes the values Every Euclidean imaginary quadratic field is norm-Euclidean and is one of the five first fields in the preceding list.
https://en.wikipedia.org/wiki?curid=10376
Euclidean algorithm In mathematics, the Euclidean algorithm, or Euclid's algorithm, is an efficient method for computing the greatest common divisor (GCD) of two integers (numbers), the largest number that divides them both without a remainder. It is named after the ancient Greek mathematician Euclid, who first described it in his "Elements" (c. 300 BC). It is an example of an "algorithm", a step-by-step procedure for performing a calculation according to well-defined rules, and is one of the oldest algorithms in common use. It can be used to reduce fractions to their simplest form, and is a part of many other number-theoretic and cryptographic calculations. The Euclidean algorithm is based on the principle that the greatest common divisor of two numbers does not change if the larger number is replaced by its difference with the smaller number. For example, 21 is the GCD of 252 and 105 (as 252 = 21 × 12 and 105 = 21 × 5), and the same number 21 is also the GCD of 105 and 252 − 105 = 147. Since this replacement reduces the larger of the two numbers, repeating this process gives successively smaller pairs of numbers until the two numbers become equal. When that occurs, they are the GCD of the original two numbers. By reversing the steps, the GCD can be expressed as a sum of the two original numbers each multiplied by a positive or negative integer, e.g., 21 = 5 × 105 + (−2) × 252. The fact that the GCD can always be expressed in this way is known as Bézout's identity. The version of the Euclidean algorithm described above (and by Euclid) can take many subtraction steps to find the GCD when one of the given numbers is much bigger than the other. A more efficient version of the algorithm shortcuts these steps, instead replacing the larger of the two numbers by its remainder when divided by the smaller of the two (with this version, the algorithm stops when reaching a zero remainder). With this improvement, the algorithm never requires more steps than five times the number of digits (base 10) of the smaller integer. This was proven by Gabriel Lamé in 1844, and marks the beginning of computational complexity theory. Additional methods for improving the algorithm's efficiency were developed in the 20th century. The Euclidean algorithm has many theoretical and practical applications. It is used for reducing fractions to their simplest form and for performing division in modular arithmetic. Computations using this algorithm form part of the cryptographic protocols that are used to secure internet communications, and in methods for breaking these cryptosystems by factoring large composite numbers. The Euclidean algorithm may be used to solve Diophantine equations, such as finding numbers that satisfy multiple congruences according to the Chinese remainder theorem, to construct continued fractions, and to find accurate rational approximations to real numbers. Finally, it can be used as a basic tool for proving theorems in number theory such as Lagrange's four-square theorem and the uniqueness of prime factorizations. The original algorithm was described only for natural numbers and geometric lengths (real numbers), but the algorithm was generalized in the 19th century to other types of numbers, such as Gaussian integers and polynomials of one variable. This led to modern abstract algebraic notions such as Euclidean domains. The Euclidean algorithm calculates the greatest common divisor (GCD) of two natural numbers "a" and "b". The greatest common divisor "g" is the largest natural number that divides both "a" and "b" without leaving a remainder. Synonyms for the GCD include the "greatest common factor" (GCF), the "highest common factor" (HCF), the "highest common divisor" (HCD), and the "greatest common measure" (GCM). The greatest common divisor is often written as gcd("a", "b") or, more simply, as ("a", "b"), although the latter notation is ambiguous, also used for concepts such as an ideal in the ring of integers, which is closely related to GCD. If gcd("a", "b") = 1, then "a" and "b" are said to be coprime (or relatively prime). This property does not imply that "a" or "b" are themselves prime numbers. For example, neither 6 nor 35 is a prime number, since they both have two prime factors: 6 = 2 × 3 and 35 = 5 × 7. Nevertheless, 6 and 35 are coprime. No natural number other than 1 divides both 6 and 35, since they have no prime factors in common. Let "g" = gcd("a", "b"). Since "a" and "b" are both multiples of "g", they can be written "a" = "mg" and "b" = "ng", and there is no larger number "G" > "g" for which this is true. The natural numbers "m" and "n" must be coprime, since any common factor could be factored out of "m" and "n" to make "g" greater. Thus, any other number "c" that divides both "a" and "b" must also divide "g". The greatest common divisor "g" of "a" and "b" is the unique (positive) common divisor of "a" and "b" that is divisible by any other common divisor "c". The GCD can be visualized as follows. Consider a rectangular area "a" by "b", and any common divisor "c" that divides both "a" and "b" exactly. The sides of the rectangle can be divided into segments of length "c", which divides the rectangle into a grid of squares of side length "c". The greatest common divisor "g" is the largest value of "c" for which this is possible. For illustration, a 24-by-60 rectangular area can be divided into a grid of: 1-by-1 squares, 2-by-2 squares, 3-by-3 squares, 4-by-4 squares, 6-by-6 squares or 12-by-12 squares. Therefore, 12 is the greatest common divisor of 24 and 60. A 24-by-60 rectangular area can be divided into a grid of 12-by-12 squares, with two squares along one edge (24/12 = 2) and five squares along the other (60/12 = 5). The GCD of two numbers "a" and "b" is the product of the prime factors shared by the two numbers, where a same prime factor can be used multiple times, but only as long as the product of these factors divides both "a" and "b". For example, since 1386 can be factored into 2 × 3 × 3 × 7 × 11, and 3213 can be factored into 3 × 3 × 3 × 7 × 17, the greatest common divisor of 1386 and 3213 equals 63 = 3 × 3 × 7, the product of their shared prime factors. If two numbers have no prime factors in common, their greatest common divisor is 1 (obtained here as an instance of the empty product), in other words they are coprime. A key advantage of the Euclidean algorithm is that it can find the GCD efficiently without having to compute the prime factors. Factorization of large integers is believed to be a computationally very difficult problem, and the security of many widely used cryptographic protocols is based upon its infeasibility. Another definition of the GCD is helpful in advanced mathematics, particularly ring theory. The greatest common divisor "g"  of two nonzero numbers "a" and "b" is also their smallest positive integral linear combination, that is, the smallest positive number of the form "ua" + "vb" where "u" and "v" are integers. The set of all integral linear combinations of "a" and "b" is actually the same as the set of all multiples of "g" ("mg", where "m" is an integer). In modern mathematical language, the ideal generated by "a" and "b" is the ideal generated by "g" alone (an ideal generated by a single element is called a principal ideal, and all ideals of the integers are principal ideals). Some properties of the GCD are in fact easier to see with this description, for instance the fact that any common divisor of "a" and "b" also divides the GCD (it divides both terms of "ua" + "vb"). The equivalence of this GCD definition with the other definitions is described below. The GCD of three or more numbers equals the product of the prime factors common to all the numbers, but it can also be calculated by repeatedly taking the GCDs of pairs of numbers. For example, Thus, Euclid's algorithm, which computes the GCD of two integers, suffices to calculate the GCD of arbitrarily many integers. The Euclidean algorithm proceeds in a series of steps such that the output of each step is used as an input for the next one. Let "k" be an integer that counts the steps of the algorithm, starting with zero. Thus, the initial step corresponds to "k" = 0, the next step corresponds to "k" = 1, and so on. Each step begins with two nonnegative remainders "r""k"−1 and "r""k"−2. Since the algorithm ensures that the remainders decrease steadily with every step, "r""k"−1 is less than its predecessor "r""k"−2. The goal of the "k"th step is to find a quotient "q""k" and remainder "r""k" that satisfy the equation and that have 0 ≤ "r""k" "k"−1. In other words, multiples of the smaller number "r""k"−1 are subtracted from the larger number "r""k"−2 until the remainder "r""k" is smaller than "r""k"−1. In the initial step ("k" = 0), the remainders "r"−2 and "r"−1 equal "a" and "b", the numbers for which the GCD is sought. In the next step ("k" = 1), the remainders equal "b" and the remainder "r"0 of the initial step, and so on. Thus, the algorithm can be written as a sequence of equations If "a" is smaller than "b", the first step of the algorithm swaps the numbers. For example, if "a" 0 equals zero, and the remainder "r"0 is "a". Thus, "r""k" is smaller than its predecessor "r""k"−1 for all "k" ≥ 0. Since the remainders decrease with every step but can never be negative, a remainder "r""N" must eventually equal zero, at which point the algorithm stops. The final nonzero remainder "r""N"−1 is the greatest common divisor of "a" and "b". The number "N" cannot be infinite because there are only a finite number of nonnegative integers between the initial remainder "r"0 and zero. The validity of the Euclidean algorithm can be proven by a two-step argument. In the first step, the final nonzero remainder "r""N"−1 is shown to divide both "a" and "b". Since it is a common divisor, it must be less than or equal to the greatest common divisor "g". In the second step, it is shown that any common divisor of "a" and "b", including "g", must divide "r""N"−1; therefore, "g" must be less than or equal to "r""N"−1. These two conclusions are inconsistent unless "r""N"−1 = "g". To demonstrate that "r""N"−1 divides both "a" and "b" (the first step), "r""N"−1 divides its predecessor "r""N"−2 since the final remainder "r""N" is zero. "r""N"−1 also divides its next predecessor "r""N"−3 because it divides both terms on the right-hand side of the equation. Iterating the same argument, "r""N"−1 divides all the preceding remainders, including "a" and "b". None of the preceding remainders "r""N"−2, "r""N"−3, etc. divide "a" and "b", since they leave a remainder. Since "r""N"−1 is a common divisor of "a" and "b", "r""N"−1 ≤ "g". In the second step, any natural number "c" that divides both "a" and "b" (in other words, any common divisor of "a" and "b") divides the remainders "r""k". By definition, "a" and "b" can be written as multiples of "c" : "a" = "mc" and "b" = "nc", where "m" and "n" are natural numbers. Therefore, "c" divides the initial remainder "r"0, since "r"0 = "a" − "q"0"b" = "mc" − "q"0"nc" = ("m" − "q"0"n")"c". An analogous argument shows that "c" also divides the subsequent remainders "r"1, "r"2, etc. Therefore, the greatest common divisor "g" must divide "r""N"−1, which implies that "g" ≤ "r""N"−1. Since the first part of the argument showed the reverse ("r""N"−1 ≤ "g"), it follows that "g" = "r""N"−1. Thus, "g" is the greatest common divisor of all the succeeding pairs: For illustration, the Euclidean algorithm can be used to find the greatest common divisor of "a" = 1071 and "b" = 462. To begin, multiples of 462 are subtracted from 1071 until the remainder is less than 462. Two such multiples can be subtracted ("q"0 = 2), leaving a remainder of 147: Then multiples of 147 are subtracted from 462 until the remainder is less than 147. Three multiples can be subtracted ("q"1 = 3), leaving a remainder of 21: Then multiples of 21 are subtracted from 147 until the remainder is less than 21. Seven multiples can be subtracted ("q"2 = 7), leaving no remainder: Since the last remainder is zero, the algorithm ends with 21 as the greatest common divisor of 1071 and 462. This agrees with the gcd(1071, 462) found by prime factorization . In tabular form, the steps are: The Euclidean algorithm can be visualized in terms of the tiling analogy given above for the greatest common divisor. Assume that we wish to cover an "a"-by-"b" rectangle with square tiles exactly, where "a" is the larger of the two numbers. We first attempt to tile the rectangle using "b"-by-"b" square tiles; however, this leaves an "r"0-by-"b" residual rectangle untiled, where "r"0 0-by-"r"0 square tiles. This leaves a second residual rectangle "r"1-by-"r"0, which we attempt to tile using "r"1-by-"r"1 square tiles, and so on. The sequence ends when there is no residual rectangle, i.e., when the square tiles cover the previous residual rectangle exactly. The length of the sides of the smallest square tile is the GCD of the dimensions of the original rectangle. For example, the smallest square tile in the adjacent figure is 21-by-21 (shown in red), and 21 is the GCD of 1071 and 462, the dimensions of the original rectangle (shown in green). At every step "k", the Euclidean algorithm computes a quotient "q""k" and remainder "r""k" from two numbers "r""k"−1 and "r""k"−2 where the "r""k" is non-negative and is strictly less than the absolute value of "r""k"−1. The theorem which underlies the definition of the Euclidean division ensures that such a quotient and remainder always exist and are unique. In Euclid's original version of the algorithm, the quotient and remainder are found by repeated subtraction; that is, "r""k"−1 is subtracted from "r""k"−2 repeatedly until the remainder "r""k" is smaller than "r""k"−1. After that "r""k" and "r""k"−1 are exchanged and the process is iterated. Euclidean division reduces all the steps between two exchanges into a single step, which is thus more efficient. Moreover, the quotients are not needed, thus one may replace Euclidean division by the modulo operation, which gives only the remainder. Thus the iteration of the Euclidean algorithm becomes simply Implementations of the algorithm may be expressed in pseudocode. For example, the division-based version may be programmed as At the beginning of the "k"th iteration, the variable "b" holds the latest remainder "r""k"−1, whereas the variable "a" holds its predecessor, "r""k"−2. The step "b" := "a" mod "b" is equivalent to the above recursion formula "r""k" ≡ "r""k"−2 mod "r""k"−1. The temporary variable "t" holds the value of "r""k"−1 while the next remainder "r""k" is being calculated. At the end of the loop iteration, the variable "b" holds the remainder "r""k", whereas the variable "a" holds its predecessor, "r""k"−1. In the subtraction-based version which was Euclid's original version, the remainder calculation (codice_1) is replaced by repeated subtraction. Contrary to the division-based version, which works with arbitrary integers as input, the subtraction-based version supposes that the input consists of positive integers and stops when "a" = "b": The variables "a" and "b" alternate holding the previous remainders "r""k"−1 and "r""k"−2. Assume that "a" is larger than "b" at the beginning of an iteration; then "a" equals "r""k"−2, since "r""k"−2 > "r""k"−1. During the loop iteration, "a" is reduced by multiples of the previous remainder "b" until "a" is smaller than "b". Then "a" is the next remainder "r""k". Then "b" is reduced by multiples of "a" until it is again smaller than "a", giving the next remainder "r""k"+1, and so on. The recursive version is based on the equality of the GCDs of successive remainders and the stopping condition gcd("r""N"−1, 0) = "r""N"−1. For illustration, the gcd(1071, 462) is calculated from the equivalent gcd(462, 1071 mod 462) = gcd(462, 147). The latter GCD is calculated from the gcd(147, 462 mod 147) = gcd(147, 21), which in turn is calculated from the gcd(21, 147 mod 21) = gcd(21, 0) = 21. In another version of Euclid's algorithm, the quotient at each step is increased by one if the resulting negative remainder is smaller in magnitude than the typical positive remainder. Previously, the equation assumed that . However, an alternative negative remainder can be computed: if or if . If is replaced by when , then one gets a variant of Euclidean algorithm such that at each step. Leopold Kronecker has shown that this version requires the least number of steps of any version of Euclid's algorithm. More generally, it has been proven that, for every input numbers "a" and "b", the number of steps is minimal if and only if is chosen in order that formula_3 where formula_4 is the golden ratio. The Euclidean algorithm is one of the oldest algorithms in common use. It appears in Euclid's "Elements" (c. 300 BC), specifically in Book 7 (Propositions 1–2) and Book 10 (Propositions 2–3). In Book 7, the algorithm is formulated for integers, whereas in Book 10, it is formulated for lengths of line segments. (In modern usage, one would say it was formulated there for real numbers. But lengths, areas, and volumes, represented as real numbers in modern usage, are not measured in the same units and there is no natural unit of length, area, or volume; the concept of real numbers was unknown at that time.) The latter algorithm is geometrical. The GCD of two lengths "a" and "b" corresponds to the greatest length "g" that measures "a" and "b" evenly; in other words, the lengths "a" and "b" are both integer multiples of the length "g". The algorithm was probably not discovered by Euclid, who compiled results from earlier mathematicians in his "Elements". The mathematician and historian B. L. van der Waerden suggests that Book VII derives from a textbook on number theory written by mathematicians in the school of Pythagoras. The algorithm was probably known by Eudoxus of Cnidus (about 375 BC). The algorithm may even pre-date Eudoxus, judging from the use of the technical term ἀνθυφαίρεσις ("anthyphairesis", reciprocal subtraction) in works by Euclid and Aristotle. Centuries later, Euclid's algorithm was discovered independently both in India and in China, primarily to solve Diophantine equations that arose in astronomy and making accurate calendars. In the late 5th century, the Indian mathematician and astronomer Aryabhata described the algorithm as the "pulverizer", perhaps because of its effectiveness in solving Diophantine equations. Although a special case of the Chinese remainder theorem had already been described in the Chinese book "Sunzi Suanjing", the general solution was published by Qin Jiushao in his 1247 book "Shushu Jiuzhang" (數書九章 "Mathematical Treatise in Nine Sections"). The Euclidean algorithm was first described "numerically" and popularized in Europe in the second edition of Bachet's "Problèmes plaisants et délectables" ("Pleasant and enjoyable problems", 1624). In Europe, it was likewise used to solve Diophantine equations and in developing continued fractions. The extended Euclidean algorithm was published by the English mathematician Nicholas Saunderson, who attributed it to Roger Cotes as a method for computing continued fractions efficiently. In the 19th century, the Euclidean algorithm led to the development of new number systems, such as Gaussian integers and Eisenstein integers. In 1815, Carl Gauss used the Euclidean algorithm to demonstrate unique factorization of Gaussian integers, although his work was first published in 1832. Gauss mentioned the algorithm in his "Disquisitiones Arithmeticae" (published 1801), but only as a method for continued fractions. Peter Gustav Lejeune Dirichlet seems to have been the first to describe the Euclidean algorithm as the basis for much of number theory. Lejeune Dirichlet noted that many results of number theory, such as unique factorization, would hold true for any other system of numbers to which the Euclidean algorithm could be applied. Lejeune Dirichlet's lectures on number theory were edited and extended by Richard Dedekind, who used Euclid's algorithm to study algebraic integers, a new general type of number. For example, Dedekind was the first to prove Fermat's two-square theorem using the unique factorization of Gaussian integers. Dedekind also defined the concept of a Euclidean domain, a number system in which a generalized version of the Euclidean algorithm can be defined (as described below). In the closing decades of the 19th century, the Euclidean algorithm gradually became eclipsed by Dedekind's more general theory of ideals. Other applications of Euclid's algorithm were developed in the 19th century. In 1829, Charles Sturm showed that the algorithm was useful in the Sturm chain method for counting the real roots of polynomials in any given interval. The Euclidean algorithm was the first integer relation algorithm, which is a method for finding integer relations between commensurate real numbers. Several novel integer relation algorithms have been developed, such as the algorithm of Helaman Ferguson and R.W. Forcade (1979) and the LLL algorithm. In 1969, Cole and Davie developed a two-player game based on the Euclidean algorithm, called "The Game of Euclid", which has an optimal strategy. The players begin with two piles of "a" and "b" stones. The players take turns removing "m" multiples of the smaller pile from the larger. Thus, if the two piles consist of "x" and "y" stones, where "x" is larger than "y", the next player can reduce the larger pile from "x" stones to "x" − "my" stones, as long as the latter is a nonnegative integer. The winner is the first player to reduce one pile to zero stones. Bézout's identity states that the greatest common divisor "g" of two integers "a" and "b" can be represented as a linear sum of the original two numbers "a" and "b". In other words, it is always possible to find integers "s" and "t" such that "g" = "sa" + "tb". The integers "s" and "t" can be calculated from the quotients "q"0, "q"1, etc. by reversing the order of equations in Euclid's algorithm. Beginning with the next-to-last equation, "g" can be expressed in terms of the quotient "q""N"−1 and the two preceding remainders, "r""N"−2 and "r""N"−3: Those two remainders can be likewise expressed in terms of their quotients and preceding remainders, Substituting these formulae for "r""N"−2 and "r""N"−3 into the first equation yields "g" as a linear sum of the remainders "r""N"−4 and "r""N"−5. The process of substituting remainders by formulae involving their predecessors can be continued until the original numbers "a" and "b" are reached: After all the remainders "r"0, "r"1, etc. have been substituted, the final equation expresses "g" as a linear sum of "a" and "b": "g" = "sa" + "tb". Bézout's identity, and therefore the previous algorithm, can both be generalized to the context of Euclidean domains. Bézout's identity provides yet another definition of the greatest common divisor "g" of two numbers "a" and "b". Consider the set of all numbers "ua" + "vb", where "u" and "v" are any two integers. Since "a" and "b" are both divisible by "g", every number in the set is divisible by "g". In other words, every number of the set is an integer multiple of "g". This is true for every common divisor of "a" and "b". However, unlike other common divisors, the greatest common divisor is a member of the set; by Bézout's identity, choosing "u" = "s" and "v" = "t" gives "g". A smaller common divisor cannot be a member of the set, since every member of the set must be divisible by "g". Conversely, any multiple "m" of "g" can be obtained by choosing "u" = "ms" and "v" = "mt", where "s" and "t" are the integers of Bézout's identity. This may be seen by multiplying Bézout's identity by "m", Therefore, the set of all numbers "ua" + "vb" is equivalent to the set of multiples "m" of "g". In other words, the set of all possible sums of integer multiples of two numbers ("a" and "b") is equivalent to the set of multiples of gcd("a", "b"). The GCD is said to be the generator of the ideal of "a" and "b". This GCD definition led to the modern abstract algebraic concepts of a principal ideal (an ideal generated by a single element) and a principal ideal domain (a domain in which every ideal is a principal ideal). Certain problems can be solved using this result. For example, consider two measuring cups of volume "a" and "b". By adding/subtracting "u" multiples of the first cup and "v" multiples of the second cup, any volume "ua" + "vb" can be measured out. These volumes are all multiples of "g" = gcd("a", "b"). The integers "s" and "t" of Bézout's identity can be computed efficiently using the extended Euclidean algorithm. This extension adds two recursive equations to Euclid's algorithm with the starting values Using this recursion, Bézout's integers "s" and "t" are given by "s" = "s""N" and "t" = "t""N", where "N+1" is the step on which the algorithm terminates with "r""N+1" = 0. The validity of this approach can be shown by induction. Assume that the recursion formula is correct up to step "k" − 1 of the algorithm; in other words, assume that for all "j" less than "k". The "k"th step of the algorithm gives the equation Since the recursion formula has been assumed to be correct for "r""k"−2 and "r""k"−1, they may be expressed in terms of the corresponding "s" and "t" variables Rearranging this equation yields the recursion formula for step "k", as required The integers "s" and "t" can also be found using an equivalent matrix method. The sequence of equations of Euclid's algorithm can be written as a product of 2-by-2 quotient matrices multiplying a two-dimensional remainder vector Let M represent the product of all the quotient matrices This simplifies the Euclidean algorithm to the form To express "g" as a linear sum of "a" and "b", both sides of this equation can be multiplied by the inverse of the matrix M. The determinant of M equals (−1)"N"+1, since it equals the product of the determinants of the quotient matrices, each of which is negative one. Since the determinant of M is never zero, the vector of the final remainders can be solved using the inverse of M Since the top equation gives the two integers of Bézout's identity are "s" = (−1)"N"+1"m"22 and "t" = (−1)"N""m"12. The matrix method is as efficient as the equivalent recursion, with two multiplications and two additions per step of the Euclidean algorithm. Bézout's identity is essential to many applications of Euclid's algorithm, such as demonstrating the unique factorization of numbers into prime factors. To illustrate this, suppose that a number "L" can be written as a product of two factors "u" and "v", that is, "L" = "uv". If another number "w" also divides "L" but is coprime with "u", then "w" must divide "v", by the following argument: If the greatest common divisor of "u" and "w" is 1, then integers "s" and "t" can be found such that by Bézout's identity. Multiplying both sides by "v" gives the relation Since "w" divides both terms on the right-hand side, it must also divide the left-hand side, "v". This result is known as Euclid's lemma. Specifically, if a prime number divides "L", then it must divide at least one factor of "L". Conversely, if a number "w" is coprime to each of a series of numbers "a"1, "a"2, ..., "a""n", then "w" is also coprime to their product, "a"1 × "a"2 × ... × "a""n". Euclid's lemma suffices to prove that every number has a unique factorization into prime numbers. To see this, assume the contrary, that there are two independent factorizations of "L" into "m" and "n" prime factors, respectively Since each prime "p" divides "L" by assumption, it must also divide one of the "q" factors; since each "q" is prime as well, it must be that "p" = "q". Iteratively dividing by the "p" factors shows that each "p" has an equal counterpart "q"; the two prime factorizations are identical except for their order. The unique factorization of numbers into primes has many applications in mathematical proofs, as shown below. Diophantine equations are equations in which the solutions are restricted to integers; they are named after the 3rd-century Alexandrian mathematician Diophantus. A typical "linear" Diophantine equation seeks integers "x" and "y" such that where "a", "b" and "c" are given integers. This can be written as an equation for "x" in modular arithmetic: Let "g" be the greatest common divisor of "a" and "b". Both terms in "ax" + "by" are divisible by "g"; therefore, "c" must also be divisible by "g", or the equation has no solutions. By dividing both sides by "c"/"g", the equation can be reduced to Bezout's identity where "s" and "t" can be found by the extended Euclidean algorithm. This provides one solution to the Diophantine equation, "x"1 = "s" ("c"/"g") and "y"1 = "t" ("c"/"g"). In general, a linear Diophantine equation has no solutions, or an infinite number of solutions. To find the latter, consider two solutions, ("x"1, "y"1) and ("x"2, "y"2), where or equivalently Therefore, the smallest difference between two "x" solutions is "b"/"g", whereas the smallest difference between two "y" solutions is "a"/"g". Thus, the solutions may be expressed as By allowing "u" to vary over all possible integers, an infinite family of solutions can be generated from a single solution ("x"1, "y"1). If the solutions are required to be "positive" integers ("x" > 0, "y" > 0), only a finite number of solutions may be possible. This restriction on the acceptable solutions allows some systems of Diophantine equations with more unknowns than equations to have a finite number of solutions; this is impossible for a system of linear equations when the solutions can be any real number (see Underdetermined system). A finite field is a set of numbers with four generalized operations. The operations are called addition, subtraction, multiplication and division and have their usual properties, such as commutativity, associativity and distributivity. An example of a finite field is the set of 13 numbers {0, 1, 2, ..., 12} using modular arithmetic. In this field, the results of any mathematical operation (addition, subtraction, multiplication, or division) is reduced modulo 13; that is, multiples of 13 are added or subtracted until the result is brought within the range 0–12. For example, the result of 5 × 7 = 35 mod 13 = 9. Such finite fields can be defined for any prime "p"; using more sophisticated definitions, they can also be defined for any power "m" of a prime "p" "m". Finite fields are often called Galois fields, and are abbreviated as GF("p") or GF("p" "m"). In such a field with "m" numbers, every nonzero element "a" has a unique modular multiplicative inverse, "a"−1 such that This inverse can be found by solving the congruence equation "ax" ≡ 1 mod "m", or the equivalent linear Diophantine equation This equation can be solved by the Euclidean algorithm, as described above. Finding multiplicative inverses is an essential step in the RSA algorithm, which is widely used in electronic commerce; specifically, the equation determines the integer used to decrypt the message. Although the RSA algorithm uses rings rather than fields, the Euclidean algorithm can still be used to find a multiplicative inverse where one exists. The Euclidean algorithm also has other applications in error-correcting codes; for example, it can be used as an alternative to the Berlekamp–Massey algorithm for decoding BCH and Reed–Solomon codes, which are based on Galois fields. Euclid's algorithm can also be used to solve multiple linear Diophantine equations. Such equations arise in the Chinese remainder theorem, which describes a novel method to represent an integer "x". Instead of representing an integer by its digits, it may be represented by its remainders "x""i" modulo a set of "N" coprime numbers "m""i": The goal is to determine "x" from its "N" remainders "x""i". The solution is to combine the multiple equations into a single linear Diophantine equation with a much larger modulus "M" that is the product of all the individual moduli "m""i", and define "M""i" as Thus, each "M""i" is the product of all the moduli "except" "m""i". The solution depends on finding "N" new numbers "h""i" such that With these numbers "h""i", any integer "x" can be reconstructed from its remainders "x""i" by the equation Since these numbers "h""i" are the multiplicative inverses of the "M""i", they may be found using Euclid's algorithm as described in the previous subsection. The Euclidean algorithm can be used to arrange the set of all positive rational numbers into an infinite binary search tree, called the Stern–Brocot tree. The number 1 (expressed as a fraction 1/1) is placed at the root of the tree, and the location of any other number "a"/"b" can be found by computing gcd("a","b") using the original form of the Euclidean algorithm, in which each step replaces the larger of the two given numbers by its difference with the smaller number (not its remainder), stopping when two equal numbers are reached. A step of the Euclidean algorithm that replaces the first of the two numbers corresponds to a step in the tree from a node to its right child, and a step that replaces the second of the two numbers corresponds to a step in the tree from a node to its left child. The sequence of steps constructed in this way does not depend on whether "a"/"b" is given in lowest terms, and forms a path from the root to a node containing the number "a"/"b". This fact can be used to prove that each positive rational number appears exactly once in this tree. For example, 3/4 can be found by starting at the root, going to the left once, then to the right twice: The Euclidean algorithm has almost the same relationship to another binary tree on the rational numbers called the Calkin–Wilf tree. The difference is that the path is reversed: instead of producing a path from the root of the tree to a target, it produces a path from the target to the root. The Euclidean algorithm has a close relationship with continued fractions. The sequence of equations can be written in the form The last term on the right-hand side always equals the inverse of the left-hand side of the next equation. Thus, the first two equations may be combined to form The third equation may be used to substitute the denominator term "r"1/"r"0, yielding The final ratio of remainders "r""k"/"r""k"−1 can always be replaced using the next equation in the series, up to the final equation. The result is a continued fraction In the worked example above, the gcd(1071, 462) was calculated, and the quotients "q""k" were 2, 3 and 7, respectively. Therefore, the fraction 1071/462 may be written as can be confirmed by calculation. Calculating a greatest common divisor is an essential step in several integer factorization algorithms, such as Pollard's rho algorithm, Shor's algorithm, Dixon's factorization method and the Lenstra elliptic curve factorization. The Euclidean algorithm may be used to find this GCD efficiently. Continued fraction factorization uses continued fractions, which are determined using Euclid's algorithm. The computational efficiency of Euclid's algorithm has been studied thoroughly. This efficiency can be described by the number of division steps the algorithm requires, multiplied by the computational expense of each step. The first known analysis of Euclid's algorithm is due to A. A. L. Reynaud in 1811, who showed that the number of division steps on input ("u", "v") is bounded by "v"; later he improved this to "v"/2  + 2. Later, in 1841, P. J. E. Finck showed that the number of division steps is at most 2 log2 "v" + 1, and hence Euclid's algorithm runs in time polynomial in the size of the input. Émile Léger, in 1837, studied the worst case, which is when the inputs are consecutive Fibonacci numbers. Finck's analysis was refined by Gabriel Lamé in 1844, who showed that the number of steps required for completion is never more than five times the number "h" of base-10 digits of the smaller number "b". In the uniform cost model (suitable for analyzing the complexity of gcd calculation on numbers that fit into a single machine word), each step of the algorithm takes constant time, and Lamé's analysis implies that the total running time is also "O"("h"). However, in a model of computation suitable for computation with larger numbers, the computational expense of a single remainder computation in the algorithm can be as large as "O"("h"2). In this case the total time for all of the steps of the algorithm can be analyzed using a telescoping series, showing that it is also "O"("h"2). Modern algorithmic techniques based on the Schönhage–Strassen algorithm for fast integer multiplication can be used to speed this up, leading to quasilinear algorithms for the GCD. The number of steps to calculate the GCD of two natural numbers, "a" and "b", may be denoted by "T"("a", "b"). If "g" is the GCD of "a" and "b", then "a" = "mg" and "b" = "ng" for two coprime numbers "m" and "n". Then as may be seen by dividing all the steps in the Euclidean algorithm by "g". By the same argument, the number of steps remains the same if "a" and "b" are multiplied by a common factor "w": "T"("a", "b") = "T"("wa", "wb"). Therefore, the number of steps "T" may vary dramatically between neighboring pairs of numbers, such as T("a", "b") and T("a", "b" + 1), depending on the size of the two GCDs. The recursive nature of the Euclidean algorithm gives another equation where "T"("x", 0) = 0 by assumption. If the Euclidean algorithm requires "N" steps for a pair of natural numbers "a" > "b" > 0, the smallest values of "a" and "b" for which this is true are the Fibonacci numbers "F""N"+2 and "F""N"+1, respectively. More precisely, if the Euclidean algorithm requires "N" steps for the pair "a" > "b", then one has "a" ≥ "F""N"+2 and "b" ≥ "F""N"+1. This can be shown by induction. If "N" = 1, "b" divides "a" with no remainder; the smallest natural numbers for which this is true is "b" = 1 and "a" = 2, which are "F"2 and "F"3, respectively. Now assume that the result holds for all values of "N" up to "M" − 1. The first step of the "M"-step algorithm is "a" = "q"0"b" + "r"0, and the Euclidean algorithm requires "M" − 1 steps for the pair "b" > "r"0. By induction hypothesis, one has "b" ≥ "F""M"+1 and "r"0 ≥ "F""M". Therefore, "a" = "q"0"b" + "r"0 ≥ "b" + "r"0 ≥ "F""M"+1 + "F""M" = "F""M"+2, which is the desired inequality. This proof, published by Gabriel Lamé in 1844, represents the beginning of computational complexity theory, and also the first practical application of the Fibonacci numbers. This result suffices to show that the number of steps in Euclid's algorithm can never be more than five times the number of its digits (base 10). For if the algorithm requires "N" steps, then "b" is greater than or equal to "F""N"+1 which in turn is greater than or equal to "φ""N"−1, where "φ" is the golden ratio. Since "b" ≥ "φ""N"−1, then "N" − 1 ≤ log"φ""b". Since log10"φ" > 1/5, ("N" − 1)/5 10"φ" log"φ""b" = log10"b". Thus, "N" ≤ 5 log10"b". Thus, the Euclidean algorithm always needs less than "O"("h") divisions, where "h" is the number of digits in the smaller number "b". The average number of steps taken by the Euclidean algorithm has been defined in three different ways. The first definition is the average time "T"("a") required to calculate the GCD of a given number "a" and a smaller natural number "b" chosen with equal probability from the integers 0 to "a" − 1 with the residual error being of order "a"−(1/6) + "ε", where "ε" is infinitesimal. The constant "C" ("Porter's Constant") in this formula equals where "γ" is the Euler–Mascheroni constant and ζ' is the derivative of the Riemann zeta function. The leading coefficient (12/π2) ln 2 was determined by two independent methods. Since the first average can be calculated from the tau average by summing over the divisors "d" of "a" it can be approximated by the formula where Λ("d") is the Mangoldt function. A third average "Y"("n") is defined as the mean number of steps required when both "a" and "b" are chosen randomly (with uniform distribution) from 1 to "n" Substituting the approximate formula for "T"("a") into this equation yields an estimate for "Y"("n") In each step "k" of the Euclidean algorithm, the quotient "q""k" and remainder "r""k" are computed for a given pair of integers "r""k"−2 and "r""k"−1 The computational expense per step is associated chiefly with finding "q""k", since the remainder "r""k" can be calculated quickly from "r""k"−2, "r""k"−1, and "q""k" The computational expense of dividing "h"-bit numbers scales as "O"("h"("ℓ"+1)), where "ℓ" is the length of the quotient. For comparison, Euclid's original subtraction-based algorithm can be much slower. A single integer division is equivalent to the quotient "q" number of subtractions. If the ratio of "a" and "b" is very large, the quotient is large and many subtractions will be required. On the other hand, it has been shown that the quotients are very likely to be small integers. The probability of a given quotient "q" is approximately ln|"u"/("u" − 1)| where "u" = ("q" + 1)2. For illustration, the probability of a quotient of 1, 2, 3, or 4 is roughly 41.5%, 17.0%, 9.3%, and 5.9%, respectively. Since the operation of subtraction is faster than division, particularly for large numbers, the subtraction-based Euclid's algorithm is competitive with the division-based version. This is exploited in the binary version of Euclid's algorithm. Combining the estimated number of steps with the estimated computational expense per step shows that the Euclid's algorithm grows quadratically ("h"2) with the average number of digits "h" in the initial two numbers "a" and "b". Let "h"0, "h"1, ..., "h""N"−1 represent the number of digits in the successive remainders "r"0, "r"1, ..., "r""N"−1. Since the number of steps "N" grows linearly with "h", the running time is bounded by Euclid's algorithm is widely used in practice, especially for small numbers, due to its simplicity. For comparison, the efficiency of alternatives to Euclid's algorithm may be determined. One inefficient approach to finding the GCD of two natural numbers "a" and "b" is to calculate all their common divisors; the GCD is then the largest common divisor. The common divisors can be found by dividing both numbers by successive integers from 2 to the smaller number "b". The number of steps of this approach grows linearly with "b", or exponentially in the number of digits. Another inefficient approach is to find the prime factors of one or both numbers. As noted above, the GCD equals the product of the prime factors shared by the two numbers "a" and "b". Present methods for prime factorization are also inefficient; many modern cryptography systems even rely on that inefficiency. The binary GCD algorithm is an efficient alternative that substitutes division with faster operations by exploiting the binary representation used by computers. However, this alternative also scales like "O"("h"²). It is generally faster than the Euclidean algorithm on real computers, even though it scales in the same way. Additional efficiency can be gleaned by examining only the leading digits of the two numbers "a" and "b". The binary algorithm can be extended to other bases ("k"-ary algorithms), with up to fivefold increases in speed. Lehmer's GCD algorithm uses the same general principle as the binary algorithm to speed up GCD computations in arbitrary bases. A recursive approach for very large integers (with more than 25,000 digits) leads to quasilinear integer GCD algorithms, such as those of Schönhage, and Stehlé and Zimmermann. These algorithms exploit the 2×2 matrix form of the Euclidean algorithm given above. These quasilinear methods generally scale as Although the Euclidean algorithm is used to find the greatest common divisor of two natural numbers (positive integers), it may be generalized to the real numbers, and to other mathematical objects, such as polynomials, quadratic integers and Hurwitz quaternions. In the latter cases, the Euclidean algorithm is used to demonstrate the crucial property of unique factorization, i.e., that such numbers can be factored uniquely into irreducible elements, the counterparts of prime numbers. Unique factorization is essential to many proofs of number theory. Euclid's algorithm can be applied to real numbers, as described by Euclid in Book 10 of his "Elements". The goal of the algorithm is to identify a real number such that two given real numbers, and , are integer multiples of it: and , where and are integers. This identification is equivalent to finding an integer relation among the real numbers and ; that is, it determines integers and such that . Euclid uses this algorithm to treat the question of incommensurable lengths. The real-number Euclidean algorithm differs from its integer counterpart in two respects. First, the remainders are real numbers, although the quotients are integers as before. Second, the algorithm is not guaranteed to end in a finite number of steps. If it does, the fraction is a rational number, i.e., the ratio of two integers and can be written as a finite continued fraction . If the algorithm does not stop, the fraction is an irrational number and can be described by an infinite continued fraction . Examples of infinite continued fractions are the golden ratio and the square root of two, . The algorithm is unlikely to stop, since almost all ratios of two real numbers are irrational. An infinite continued fraction may be truncated at a step to yield an approximation to that improves as is increased. The approximation is described by convergents ; the numerator and denominators are coprime and obey the recurrence relation where and are the initial values of the recursion. The convergent is the best rational number approximation to with denominator : Polynomials in a single variable "x" can be added, multiplied and factored into irreducible polynomials, which are the analogs of the prime numbers for integers. The greatest common divisor polynomial of two polynomials and is defined as the product of their shared irreducible polynomials, which can be identified using the Euclidean algorithm. The basic procedure is similar to that for integers. At each step , a quotient polynomial and a remainder polynomial are identified to satisfy the recursive equation where and . Each quotient polynomial is chosen such that each remainder is either zero or has a degree that is smaller than the degree of its predecessor: . Since the degree is a nonnegative integer, and since it decreases with every step, the Euclidean algorithm concludes in a finite number of steps. The last nonzero remainder is the greatest common divisor of the original two polynomials, and . For example, consider the following two quartic polynomials, which each factor into two quadratic polynomials Dividing by yields a remainder . In the next step, is divided by yielding a remainder . Finally, dividing by yields a zero remainder, indicating that is the greatest common divisor polynomial of and , consistent with their factorization. Many of the applications described above for integers carry over to polynomials. The Euclidean algorithm can be used to solve linear Diophantine equations and Chinese remainder problems for polynomials; continued fractions of polynomials can also be defined. The polynomial Euclidean algorithm has other applications, such as Sturm chains, a method for counting the zeros of a polynomial that lie inside a given real interval. This in turn has applications in several areas, such as the Routh–Hurwitz stability criterion in control theory. Finally, the coefficients of the polynomials need not be drawn from integers, real numbers or even the complex numbers. For example, the coefficients may be drawn from a general field, such as the finite fields described above. The corresponding conclusions about the Euclidean algorithm and its applications hold even for such polynomials. The Gaussian integers are complex numbers of the form , where and are ordinary integers and is the square root of negative one. By defining an analog of the Euclidean algorithm, Gaussian integers can be shown to be uniquely factorizable, by the argument above. This unique factorization is helpful in many applications, such as deriving all Pythagorean triples or proving Fermat's theorem on sums of two squares. In general, the Euclidean algorithm is convenient in such applications, but not essential; for example, the theorems can often be proven by other arguments. The Euclidean algorithm developed for two Gaussian integers and is nearly the same as that for ordinary integers, but differs in two respects. As before, the task at each step is to identify a quotient and a remainder such that where , where , and where every remainder is strictly smaller than its predecessor: . The first difference is that the quotients and remainders are themselves Gaussian integers, and thus are complex numbers. The quotients are generally found by rounding the real and complex parts of the exact ratio (such as the complex number ) to the nearest integers. The second difference lies in the necessity of defining how one complex remainder can be "smaller" than another. To do this, a norm function is defined, which converts every Gaussian integer into an ordinary integer. After each step of the Euclidean algorithm, the norm of the remainder is smaller than the norm of the preceding remainder, . Since the norm is a nonnegative integer and decreases with every step, the Euclidean algorithm for Gaussian integers ends in a finite number of steps. The final nonzero remainder is , the Gaussian integer of largest norm that divides both and ; it is unique up to multiplication by a unit, or . Many of the other applications of the Euclidean algorithm carry over to Gaussian integers. For example, it can be used to solve linear Diophantine equations and Chinese remainder problems for Gaussian integers; continued fractions of Gaussian integers can also be defined. A set of elements under two binary operations, denoted as addition and multiplication, is called a Euclidean domain if it forms a commutative ring and, roughly speaking, if a generalized Euclidean algorithm can be performed on them. The two operations of such a ring need not be the addition and multiplication of ordinary arithmetic; rather, they can be more general, such as the operations of a mathematical group or monoid. Nevertheless, these general operations should respect many of the laws governing ordinary arithmetic, such as commutativity, associativity and distributivity. The generalized Euclidean algorithm requires a "Euclidean function", i.e., a mapping from into the set of nonnegative integers such that, for any two nonzero elements and in , there exist and in such that and . Examples of such mappings are the absolute value for integers, the degree for univariate polynomials, and the norm for Gaussian integers above. The basic principle is that each step of the algorithm reduces "f" inexorably; hence, if can be reduced only a finite number of times, the algorithm must stop in a finite number of steps. This principle relies on the well-ordering property of the non-negative integers, which asserts that every non-empty set of non-negative integers has a smallest member. The fundamental theorem of arithmetic applies to any Euclidean domain: Any number from a Euclidean domain can be factored uniquely into irreducible elements. Any Euclidean domain is a unique factorization domain (UFD), although the converse is not true. The Euclidean domains and the UFD's are subclasses of the GCD domains, domains in which a greatest common divisor of two numbers always exists. In other words, a greatest common divisor may exist (for all pairs of elements in a domain), although it may not be possible to find it using a Euclidean algorithm. A Euclidean domain is always a principal ideal domain (PID), an integral domain in which every ideal is a principal ideal. Again, the converse is not true: not every PID is a Euclidean domain. The unique factorization of Euclidean domains is useful in many applications. For example, the unique factorization of the Gaussian integers is convenient in deriving formulae for all Pythagorean triples and in proving Fermat's theorem on sums of two squares. Unique factorization was also a key element in an attempted proof of Fermat's Last Theorem published in 1847 by Gabriel Lamé, the same mathematician who analyzed the efficiency of Euclid's algorithm, based on a suggestion of Joseph Liouville. Lamé's approach required the unique factorization of numbers of the form , where and are integers, and is an th root of 1, that is, . Although this approach succeeds for some values of (such as , the Eisenstein integers), in general such numbers do factor uniquely. This failure of unique factorization in some cyclotomic fields led Ernst Kummer to the concept of ideal numbers and, later, Richard Dedekind to ideals. The quadratic integer rings are helpful to illustrate Euclidean domains. Quadratic integers are generalizations of the Gaussian integers in which the imaginary unit "i" is replaced by a number . Thus, they have the form , where and are integers and has one of two forms, depending on a parameter . If does not equal a multiple of four plus one, then If, however, "D" does equal a multiple of four plus one, then If the function corresponds to a norm function, such as that used to order the Gaussian integers above, then the domain is known as "norm-Euclidean". The norm-Euclidean rings of quadratic integers are exactly those where is one of the values −11, −7, −3, −2, −1, 2, 3, 5, 6, 7, 11, 13, 17, 19, 21, 29, 33, 37, 41, 57, or 73. The cases and yield the Gaussian integers and Eisenstein integers, respectively. If is allowed to be any Euclidean function, then the list of possible values of for which the domain is Euclidean is not yet known. The first example of a Euclidean domain that was not norm-Euclidean (with ) was published in 1994. In 1973, Weinberger proved that a quadratic integer ring with is Euclidean if, and only if, it is a principal ideal domain, provided that the generalized Riemann hypothesis holds. The Euclidean algorithm may be applied to some noncommutative rings such as the set of Hurwitz quaternions. Let and represent two elements from such a ring. They have a common right divisor if and for some choice of and in the ring. Similarly, they have a common left divisor if and for some choice of and in the ring. Since multiplication is not commutative, there are two versions of the Euclidean algorithm, one for right divisors and one for left divisors. Choosing the right divisors, the first step in finding the by the Euclidean algorithm can be written where represents the quotient and the remainder. This equation shows that any common right divisor of and is likewise a common divisor of the remainder . The analogous equation for the left divisors would be With either choice, the process is repeated as above until the greatest common right or left divisor is identified. As in the Euclidean domain, the "size" of the remainder must be strictly smaller than , and there must be only a finite number of possible sizes for , so that the algorithm is guaranteed to terminate. Most of the results for the GCD carry over to noncommutative numbers. For example, Bézout's identity states that the right can be expressed as a linear combination of and . In other words, there are numbers and such that The analogous identity for the left GCD is nearly the same: Bézout's identity can be used to solve Diophantine equations. For instance, one of the standard proofs of Lagrange's four-square theorem, that every positive integer can be represented as a sum of four squares, is based on quaternion GCDs in this way.
https://en.wikipedia.org/wiki?curid=10377
European Centre for Medium-Range Weather Forecasts The European Centre for Medium-Range Weather Forecasts (ECMWF) is an independent intergovernmental organisation supported by most of the nations of Europe and is based at Shinfield Park, Reading, United Kingdom. It operates one of the largest supercomputer complexes in Europe and the world's largest archive of numerical weather prediction data. ECMWF was established in 1975, in recognition of the need to pool the scientific and technical resources of Europe's meteorological services and institutions for the production of weather forecasts for medium-range timescales (up to approximately two weeks) and of the economic and social benefits expected from it. It comprises 22 European countries: It also has co-operation agreements with other states: Bulgaria, Czech Republic, Estonia, Hungary, Israel, Latvia, Lithuania, North Macedonia, Montenegro, Morocco, Romania and Slovakia. The Centre employs about 350 staff, mostly appointed from across the member states and co-operating states. In 2017, the Centre's member states accepted an offer from the Italian Government to move ECMWF's data centre to Bologna, Italy. The new site, a former tobacco factory, would be redesigned by the architecture firm gmp. In May 2020, the French Environment Ministry suggested Toulouse as the location for the centre as ECMWF plans to move some of its teams, to European Union territory. The move has been directly attributed to Brexit. Several other ECMWF Member States have expressed an interest in hosting this new facility, including Italy, Austria, Germany and Spain. ECMWF aims to provide accurate medium-range global weather forecasts out to 15 days and seasonal forecasts out to 12 months. Its products are provided to the national weather services of its member states and co-operating states as a complement to their national short-range and climatological activities, and those national states use ECMWF's products for their own national duties, in particular to give early warning of potentially damaging severe weather. ECMWF's core mission is to: To deliver this core mission, the Centre provides: The Centre develops and operates global atmospheric models and data assimilation systems for the dynamics, thermodynamics and composition of the Earth's atmosphere and for interacting parts of the Earth-system. It uses numerical weather prediction methods to prepare forecasts and their initial conditions, and it contributes to monitoring the relevant parts of the Earth system. Numerical weather prediction (NWP) requires input of meteorological data, collected by satellites and earth observation systems such as automatic and manned stations, aircraft, ships and weather balloons. Assimilation of this data is used to produce an initial state of a computer model of the atmosphere, from which an atmospheric model is used to forecast the weather. These forecasts are typically: Over the past three decades ECMWF's wide-ranging programme of research has played a major role in developing such assimilation and modelling systems. This improves the accuracy and reliability of weather forecasting by about a day per decade, so that a seven-day forecast now (2015) is as accurate as a three-day forecast was four decades ago (1975). ECMWF's monthly and seasonal forecasts provide early predictions of events such as heat waves, cold spells and droughts, as well as their impacts on sectors such as agriculture, energy and health. Since ECMWF runs a wave model, there are also predictions of coastal waves and storm surges in European waters which can be used to provide warnings. Forecasts of severe weather events allow appropriate mitigating action to be taken and contingency plans to be put into place by the authorities and the public. The increased time gained by issuing accurate warnings can save lives, for instance by evacuating people from a storm surge area. Authorities and businesses can plan to maintain services around threats such as high winds, floods or snow. In October 2012 the ECMWF model suggested seven days in advance that Hurricane Sandy was likely to make landfall on the East Coast of the United States. It also predicted the intensity and track of the November 2012 nor'easter, which impacted the east coast a week after Sandy. ECMWF's Extreme Forecast Index (EFI) was developed as a tool to identify where the EPS (Ensemble Prediction System) forecast distribution differs substantially from that of the model climate. It contains information regarding variability of weather parameters, in location and time and can highlight an abnormality of a weather situation without having to define specific space- and time-dependent thresholds. ECMWF, through its partnerships with EUMETSAT, ESA, the EU and the European Science community has established a leading position for Europe in the exploitation of satellite data for operational numerical weather prediction, and for operational seasonal forecasting with coupled atmosphere-ocean-land models. The increasing amount of satellite data and the development of more sophisticated ways of extraction information from that data have made a major contribution to improving the accuracy and utility of NWP forecasts. ECMWF continuously endeavours to improve the use of satellite observations for NWP. ECMWF makes significant contributions to support research on climate variability, pioneering an approach known as reanalysis. This involves feeding weather observations collected over decades into a NWP system to recreate past atmospheric, sea- and land-surface conditions over specific time periods to obtain a clearer picture of how the climate has changed. Reanalysis provides a four-dimensional picture of the atmosphere and effectively allows monitoring of the variability and change of global climate, thereby contributing also to the understanding and attribution of climate change. To date, and with support from Europe's National Meteorological Services and the European Commission, ECMWF has conducted several major reanalyses of the global atmosphere: the first ECMWF re-analysis (ERA-15) project generated reanalyses from December 1978 to February 1994; the ERA-40 project generated reanalyses from September 1957 to August 2002. The ERA-Interim reanalysis covered the period from 1979 onwards. A reanalysis product (ERA5) with higher spatial resolution (31 km) was released by ECMWF in 2019 as part of the Copernicus Climate Change Service. ECMWF's operational forecasts are produced from its "Integrated Forecast System" (sometimes informally known in the United States as the "European model") which is run every twelve hours and forecasts out to ten days. It includes both a "deterministic forecast" mode and an ensemble. The deterministic forecast is a single model run that is relatively high in resolution as well as in computational expense. The ensemble is relatively low (about half that of the deterministic) in resolution (and in computational expense), so less accurate. But it is run 51 times in parallel, from slightly different initial conditions to give a spread of likelihood over the range of the forecast. The Centre currently serves as the Entrusted Entity responsible for delivery of two of the Services of the EU's Copernicus Programme. The two services are the Copernicus Atmosphere Monitoring Service (CAMS) and the Copernicus Climate Change Service (C3S).
https://en.wikipedia.org/wiki?curid=10378
European Broadcasting Union The European Broadcasting Union (EBU; ; ) is an alliance of public service media organisations, established on 12 February 1950. The organisation is made up of 116 member organisations in 56 countries, and 34 associate members from a further 21 countries. It is best known for producing the Eurovision Song Contest. It also hosted debates between candidates for the European Commission presidency for the 2014 and 2019 parliamentary elections but is unrelated to the European Union itself. EBU members are public service media (PSM) broadcasters whose output is made, financed and controlled by the public, for the public. PSM broadcasters are often established by law but are non-partisan, independent and run for the benefit of society as a whole. EBU members come from as far north as Iceland and as far south as Egypt, from Ireland in the west and Azerbaijan in the east, and almost every nation from geographical Europe in between. Associate Members are from countries and territories beyond Europe, such as Canada, Japan, Mexico, India and China. Associate members from the United States include ABC, CBS, NBC, CPB, NPR, APM and the only individual station, Chicago-based classical music radio WFMT. Membership is for media organisations whose countries are within the European Broadcasting Area (EBA), as defined by the International Telecommunication Union, or who are members of the Council of Europe. Members benefit from: The EBU's highest-profile production is the Eurovision Song Contest. The EBU also organises the Eurovision Dance Contest, the Junior Eurovision Song Contest, the Eurovision Young Dancers competition, and other competitions which are modeled along similar lines. Radio collaborations include Euroclassic Notturno – an overnight classical music stream, produced by BBC Radio 3 and broadcast in the United Kingdom as "Through the Night" – and special theme days, such as the annual Christmas music relays from around Europe. The EBU is a member of the International Music Council. Most EBU broadcasters have group deals to carry major sporting events including the FIFA World Cup and the inaugural European Championships. Another annually recurring event which is broadcast across Europe through the EBU is the Vienna New Year's Concert. Eurovision Media Services is the business arm of the EBU and provides first-class media services for many media organisations and sports federations around the world. The theme music played before EBU broadcasts are Marc-Antoine Charpentier's "Prelude to Te Deum". It is well known to Europeans as it is played before and after the Eurovision Song Contest and other important events. The EBU was a successor to the International Broadcasting Union (IBU) that was founded in 1925 and had its administrative headquarters in Geneva and technical office in Brussels. It fostered programming exchanges between members and mediated technical disputes between members that were mostly concerned with frequency and interference issues. It was in effect taken over by Nazi Germany during the Second World War and when the conflict ended in the eyes of the Allies it was a compromised organisation that they could not trust. In the spring of 1946, representatives of the Soviet radio committee proposed forming a new organisation; however, at the same time preparations were being made for an inter-governmental “European Broadcasting Conference” (EBC) in Copenhagen in 1948 to draw up a new plan for frequency use in the European Broadcasting Area (EBA). It was considered necessary to have an organisation that could implement the “Copenhagen Wavelength Plan” but there was disagreement among broadcasters and particularly a fear expressed by the BBC that a new association might be dominated by the USSR and its proposal to give each of its constituent states one vote. France proposed that it would have four votes with the inclusion of its North African colonies. The United Kingdom felt it would have little influence with just one vote. On 27 June 1946, the alternative International Broadcasting Organisation (IBO) was founded with 26 members and without British participation. The following day the IBU met in General Assembly and an attempt was made to dissolve it but failed; though 18 of its 28 members left to join the IBO. For a period of time in the late 1940s both the IBU and IBO vied for the role of organising frequencies but Britain decided to be in involved in neither. The BBC attempted but failed to find suitable working arrangements with them. However, for practical purposes, the IBO rented the IBU technical centre in Brussels and employed its staff. The BBC then proposed a new solution based on the IBO changing its constitution so there will be only one member per ITU country, thus ensuring a Western majority over the USSR and its satellite states. In August 1949 a meeting took place in Stresa, Italy but it resulted in disagreement between delegates on how to resolve the problems. One proposal was for the European Broadcasting Area to be replaced by one that would exclude Eastern Europe, the Levant and North Africa. After Stresa, a consensus emerged among the Western Europeans to form a new organisation and the BBC proposed it be based in London. Meetings in Paris on 31 October and 1 November 1949 sealed the fate of the IBU and IBO, but it was decided not to allow West Germany to be a founder of the new organisation. On 13 February 1950 the European Broadcasting Union had its first meeting with 23 members from the ITU defined European Broadcasting Area at the Imperial Hotel in Torquay, England, United Kingdom. The first president was Ian Jacob of the BBC who remained at the helm for 10 years while its operation was largely dominated by the BBC due to its financial, technical and staff input. The most important difference between the EBU and its predecessors was that EBU membership was for broadcasters and not governments. Early delegates said EBU meetings were cordial and professional and very different from the abrupt tone of its predecessors. West Germany was admitted in 1951 and a working relationship forged with the USSR's Organisation for International Radio and TV (OIRT) which existed in parallel with the EBU until its merger in 1993. In 1967, the first concert in the International Concert Season of the European Broadcasting Union was broadcast from the Queen Elizabeth Hall in London. The objective of the EBU's technical activities is simply to assist EBU Members (see below) in this period of unprecedented technological changes. This includes the provision of technical information to Members via conferences and workshops, as well as in written form (such as the EBU Technical Review, and the EBU tech-"i" magazine). The EBU also encourages active collaboration between its Members on the basis that they can freely share their knowledge and experience, thus achieving considerably more than individual Members could achieve by themselves. Much of this collaboration is achieved through Project Groups which study specific technical issues of common interest: for example, EBU Members have long been preparing for the revision of the 1961 Stockholm Plan. The EBU places great emphasis on the use of open standards. Widespread use of open standards (such as MPEG-2, DAB, DVB, etc.) ensures interoperability between products from different vendors, as well as facilitating the exchange of programme material between EBU Members and promoting "horizontal markets" for the benefit of all consumers. EBU Members and the EBU Technical Department have long played an important role in the development of many systems used in radio and television broadcasting, such as: The EBU has also actively encouraged the development and implementation of: On 11 June 2013, the Greek government shut down the state broadcaster ERT, at short notice, citing government spending concerns related to the Euro crisis. In response, the European Broadcasting Union set up a makeshift studio on the same day, near the former ERT offices in Athens, in order to continue providing EBU members with the news-gathering and broadcast relay services which had formerly been provided by ERT. The EBU put out a statement expressing its "profound dismay" at the shutdown, urged the Greek Prime Minister "to use all his powers to immediately reverse this decision" and offered the "advice, assistance and expertise necessary for ERT to be preserved". Starting on 4 May 2014, the New Hellenic Radio, Internet and Television broadcaster began nationwide transmissions, taking over ERT's vacant active membership slot in the EBU. On 11 June 2015, two years after ERT's closure, NERIT SA renamed as ERT SA which reopened with a comprehensive program in all radio stations (with nineteen regional, two world-range and five pan-Hellenic range radio stations) and four TV channels ERT1, ERT2, ERT3 and ERT HD (now ERT Sports HD). The Member list , comprises the following 72 broadcasting companies from 56 countries. Any group or organisation from an International Telecommunication Union (ITU) member country, which provides a radio or television service outside of the European Broadcasting Area, are permitted to submit applications to the EBU for Associate Membership. Countries which have this status also pay an annual fee to maintain this status. It was also noted by the EBU that any country that is granted Associate Member status does not include any access into the Eurovision events (with the exceptions of Australia, who have participated in the Eurovision Song Contest and the Junior Eurovision Song Contest since 2015, Canada in Eurovision Young Dancers between 1987 and 1989 and Kazakhstan, who have participated in Junior Eurovision since 2018). The list of Associate Members of EBU comprised the following 34 broadcasting companies from 21 countries . Any groups or organisations from a country with International Telecommunication Union (ITU) membership, which does not qualify for either the EBU's Active or Associate memberships, but still provide a broadcasting activity for the EBU, are granted a unique Approved Participants membership, which lasts approximately five years. An application for this status may be submitted to the EBU at any given time, providing an annual fee is paid. The following seven EBU broadcast members had status as Approved Participants in May 2016. The EBU in co-operation with the respective host broadcaster organises competitions and events in which its members can participate if they wish to do so. These include: The Eurovision Song Contest () is an annual international song competition between EBU Members, that was first held in Lugano, Switzerland, on 24 May 1956. Seven countries participated – each submitting two songs, for a total of 14. This was the only contest in which more than one song per country was performed: since 1957 all contests have allowed one entry per country. The 1956 contest was won by the host nation, Switzerland. The most recent host city was Tel Aviv, Israel, where the Netherlands won the competition. Let the Peoples Sing is a biennial choir competition, the participants of which are chosen from radio recordings entered by EBU radio members. The final, encompassing three categories and around ten choirs, is offered as a live broadcast to all EBU members. The overall winner is awarded the "Silver Rose Bowl". Jeux Sans Frontières (, or Games Without Borders) was a Europe-wide television game show. In its original conception, it was broadcast from 1965 to 1999 under the auspices of the EBU. The original series run ended in 1982 but was revived in 1988 with a different complexion of nations and was hosted by smaller broadcasters. Eurovision Young Musicians is a competition for European musicians that are between the ages of 12 and 21 years old. It is organised by the EBU and is a member of EMCY. The first competition was held in Manchester, the United Kingdom on 11 May 1982. The televised competition is held every two years, with some countries holding national heats. Since its foundation in 1982, the Eurovision Young Musicians competition has become one of the most important music competitions on an international level. The Eurovision Young Dancers is a biennial dance showcase broadcast on television throughout Europe. The first competition was held in Reggio Emilia, Italy on 16 June 1985. It uses a format similar to the Eurovision Song Contest, every country that is a member of the EBU has had the opportunity to send a dance act to compete for the title of "Eurovision Young Dancer". The competition is for solo dancers and all contestants must be between the ages of 16 and 21 years and not professionally engaged. Euroclassic Notturno is a six-hour sequence of classical music recordings assembled by BBC Radio from material supplied by members of the EBU and streamed back to those broadcasters by satellite for use in their overnight classical-music schedules. The recordings used are taken not from commercial CDs but from earlier (usually live) radio broadcasts. Junior Eurovision Song Contest (), is an annual international song competition, that was first held in Copenhagen, Denmark, on 15 November 2003. Sixteen countries participated – each submitting one song, for a total of 16 entries. The 2003 Contest was won by Croatia and the current winner is Poland. The Eurovision Dance Contest (not to be confused with the "Eurovision Young Dancers Competition") was an international dancing competition that was held for the first time in London, the United Kingdom on 1 September 2007. The competition was repeated in 2008 when it was held in Glasgow, United Kingdom, but has not been held since. The Eurovision Magic Circus Show was an entertainment show organised by the EBU, which took place in 2010, 2011 and 2012 in Geneva. Children aged between 7-14 representing 8 countries within the EBU membership area, performed a variety of circus acts at the "Geneva Christmas Circus" (). The main show was also accompanied by the "Magic Circus Show Orchestra". The inaugural Eurovision Choir featuring non-professional choirs selected by EBU Members, took place on 22 July 2017 in Riga, hosted by the Latvian broadcaster Latvijas Televīzija (LTV). 9 countries took part in the first edition. Carmen Manet from Slovenia was the first winner. The European Sports Championships is a multi-sport event involving some of the leading sports in Europe. The European Governing Bodies for athletics, aquatics, cycling, rowing, golf, gymnastics and triathlon, will coordinate their individual championships as part of the first edition in the summer of 2018, hosted by the cities of Berlin (already chosen as the host for the 2018 European Athletics Championships) and Glasgow (already chosen as the host for the 2018 European Aquatics Championships, and which will now also host the events of the other sports).
https://en.wikipedia.org/wiki?curid=10380
Electrothermal-chemical technology Electrothermal-chemical (ETC) technology is an attempt to increase accuracy and muzzle energy of future tank, artillery, and close-in weapon system guns by improving the predictability and rate of expansion of propellants inside the barrel. An electrothermal-chemical gun uses a plasma cartridge to ignite and control the ammunition's propellant, using electrical energy to trigger the process. ETC increases the performance of conventional solid propellants, reduces the effect of temperature on propellant expansion and allows for more advanced, higher density propellants to be used. The technology has been under development since the mid-1980s and at present is actively being researched in the United States by the Army Research Laboratory, Sandia National Laboratories and defense industry contractors, including FMC Corporation, General Dynamics Land Systems, Olin Ordnance, and Soreq Nuclear Research Center. It is possible that electrothermal-chemical gun propulsion will be an integral part of US Army's future combat system and those of other countries such as Germany and the United Kingdom. Electrothermal-chemical technology is part of a broad research and development program that encompasses all electric gun technology, such as railguns and coil guns. The constant battle between armour and armor-piercing round has led to continuous development of the main battle tank design. The evolution of American anti-tank weapons can be traced back to requirements to combat Soviet tanks. In the late 1980s, it was thought that the protection level of the Future Soviet Tank (FST) could exceed 700 mm of rolled homogeneous armour equivalence at its maximum thickness, which was effectively immune against the contemporary M829 armour piercing fin stabilized discarding sabot. In the 1980s the most immediate method available to NATO to counter Soviet advances in armour technology was the adoption of a 140 mm main gun, but this required a redesigned turret that could incorporate the larger breech and ammunition, and it also required some sort of automatic loader. Although the 140 mm gun was considered a real interim solution it was decided after the fall of the Soviet Union that the increase in muzzle energy it provided was not worth the increase in weight. Resources were therefore spent on research into other programs that could provide the needed muzzle energy. One of the most successful alternative technologies remains electrothermal-chemical ignition. Most proposed advances in gun technology are based on the assumption that the solid propellant as a stand-alone propulsion system is no longer capable of delivering the required muzzle energy. This requirement has been underscored by the appearance of the Russian T-90 main battle tank. Even the elongation of current gun tubes, such as the new German 120 mm L/55, which was introduced by Rheinmetall is considered only an interim solution as it does not offer the required increase in muzzle velocity. Even advanced kinetic energy ammunition such as the United States' M829A3 is considered only an interim solution against future threats. To that extent the solid propellant is considered to have reached the end of its usefulness, although it will remain the principal propulsion method for at least the next decade until newer technologies mature. To improve on the capabilities of a solid propellant weapon the electrothermal-chemical gun may see production as early as 2016. ETC technology offers a medium-risk upgrade and is developed to the point that further improvements are so minor that it can be considered mature. The lightweight American 120 mm XM291 came close to achieving 17 MJ of muzzle energy, which is the lower-end muzzle energy spectrum for a 140 mm gun. However, the success of the XM291 does not imply the success of ETC technology as there are key parts of the propulsion system that are not yet understood or fully developed, such as the plasma ignition process. Nevertheless, there is substantial existing evidence that ETC technology is viable and worth the money to continue development. Furthermore, it can be integrated into current gun systems. An electrothermal-chemical gun uses a plasma cartridge to ignite and control the ammunition's propellant, using electrical energy as a catalyst to begin the process. Originally researched by Dr. Jon Parmentola for the U.S. Army, it has grown into a very plausible successor to a standard solid propellant tank gun. Since the beginning of research the United States has funded the XM291 gun project with USD 4,000,000, basic research with USD 300,000, and applied research with USD 600,000. Since then it has been proven to work, although efficiency to the level required has not yet been accomplished. ETC increases the performance of conventional solid propellants, reduces the effect of temperature on propellant expansion and allows for more advanced, higher density propellants to be used. It will also reduce pressure placed on the barrel in comparison to alternative technologies that offer the same muzzle energy given the fact that it helps spread the propellant's gas much more smoothly during ignition. Currently, there are two principal methods of plasma initiation: the flashboard large area emitter (FLARE) and the triple coaxial plasma igniter (TCPI). Flashboards run in several parallel strings to provide a large area of plasma or ultraviolet radiation and uses the breakdown and vaporization of gaps of diamonds to produce the required plasma. These parallel strings are mounted in tubes and oriented to have their gaps azimuthal to the tube's axis. It discharges by using high pressure air to move air out of the way. FLARE initiators can ignite propellants through the release of plasma, or even through the use of ultraviolet heat radiation. The absorption length of a solid propellant is sufficient to be ignited by radiation from a plasma source. However, FLARE has most likely not reached optimal design requirements and further understanding of FLARE and how it works is completely necessary to ensure the evolution of the technology. If FLARE provided the XM291 gun project with the sufficient radiative heat to ignite the propellant to achieve a muzzle energy of 17 MJ one could only imagine the possibilities with a fully developed FLARE plasma igniter. Current areas of study include how plasma will affect the propellant through radiation, the deliverance of mechanical energy and heat directly and by driving gas flow. Despite these daunting tasks FLARE has been seen as the most plausible igniter for future application on ETC guns. A coaxial igniter consists of a fully insulated conductor, covered by four strips of aluminum foil. All of this is further insulated in a tube about 1.6 cm in diameter that is perforated with small holes. The idea is to use an electrical flow through the conductor and then exploding the flow into vapor and then breaking it down into plasma. Consequently, the plasma escapes through the constant perforations throughout the insulating tube and initiates the surrounding propellant. A TCPI igniter is fitted in individual propellant cases for each piece of ammunition. However, TCPI is no longer considered a viable method of propellant ignition because it may damage the fins and does not deliver energy as efficiently as a FLARE igniter. The XM291 is the best existing example of a working electrothermal-chemical gun. It was an alternative technology to the heavier caliber 140 mm gun by using the dual-caliber approach. It uses a breech that is large enough to accept 140 mm ammunition and be mounted with both a 120 mm barrel and a 135 mm or 140 mm barrel. The XM291 also mounts a larger gun tube and a larger ignition chamber than the existing M256 L/44 main gun. Through the application of electrothermal-chemical technology the XM291 has been able to achieve muzzle energy outputs that equate that to a low-level 140 mm gun, while achieving muzzle velocities greater than those of the larger 140 mm gun. Although the XM291 does not mean that ETC technology is viable it does offer an example that it is possible. ETC is also a more viable option than other alternatives by definition. ETC requires much less energy input from outside sources, like a battery, than a railgun or a coilgun would. Tests have shown that energy output by the propellant is higher than energy input from outside sources on ETC guns. In comparison, a railgun currently cannot achieve a higher muzzle velocity than the amount of energy input. Even at 50% efficiency a rail gun launching a projectile with a kinetic energy of 20 MJ would require an energy input into the rails of 40 MJ, and 50% efficiency has not yet been achieved. To put this into perspective, a rail gun launching at 9 MJ of energy would need roughly 32 MJ worth of energy from capacitors. Current advances in energy storage allow for energy densities as high as 2.5 MJ/dm³, which means that a battery delivering 32 MJ of energy would require a volume of 12.8 dm³ per shot; this is not a viable volume for use in a modern main battle tank, especially one designed to be lighter than existing models. There has even been discussion about eliminating the necessity for an outside electrical source in ETC ignition by initiating the plasma cartridge through a small explosive force. Furthermore, ETC technology is not only applicable to solid propellants. To increase muzzle velocity even further electrothermal-chemical ignition can work with liquid propellants, although this would require further research into plasma ignition. ETC technology is also compatible with existing projects to reduce the amount of recoil delivered to the vehicle while firing. Understandably, recoil of a gun firing a projectile at 17 MJ or more will increase directly with the increase in muzzle energy in accordance to Newton's third law of motion and successful implementation of recoil reduction mechanisms will be vital to the installation of an ETC powered gun in an existing vehicle design. For example, OTO Melara's new lightweight 120 mm L/45 gun has achieved a recoil force of 25 t by using a longer recoil mechanism (550 mm) and a pepperpot muzzle brake. Reduction in recoil can also be achieved through mass attenuation of the thermal sleeve. The ability of ETC technology to be applied to existing gun designs means that for future gun upgrades there's no longer the necessity to redesign the turret to include a larger breech or caliber gun barrel. Several countries have already determined that ETC technology is viable for the future and have funded indigenous projects considerably. These include the United States, Germany and the United Kingdom, amongst others. The United States' XM360, which was planned to equip the Future Combat Systems Mounted Combat System light tank and may be the M1 Abrams' next gun upgrade, is reportedly based on the XM291 and may include ETC technology, or portions of ETC technology. Tests of this gun have been performed using "precision ignition" technology, which may refer to ETC ignition.
https://en.wikipedia.org/wiki?curid=10382
Boeing E-3 Sentry The Boeing E-3 Sentry, commonly known as AWACS, is an American airborne early warning and control (AEW&C) aircraft developed by Boeing. Derived from the Boeing 707, it provides all-weather surveillance, command, control, and communications, and is used by the United States Air Force, NATO, Royal Air Force, French Air Force, and Royal Saudi Air Force. The E-3 is distinguished by the distinctive rotating radar dome (rotodome) above the fuselage. Production ended in 1992 after 68 aircraft had been built. In the mid-1960s, the U.S. Air Force (USAF) was seeking an aircraft to replace its piston-engined Lockheed EC-121 Warning Star, which had been in service for over a decade. After issuing preliminary development contracts to three companies, the USAF picked Boeing to construct two airframes to test Westinghouse Electric and Hughes's competing radars. Both radars used pulse-Doppler technology, with Westinghouse's design emerging as the contract winner. Testing on the first production E-3 began in October 1975. The first USAF E-3 was delivered in March 1977, and during the next seven years, a total of 34 aircraft were manufactured. NATO, as a single identity, also had 18 aircraft manufactured, basing them in Germany. The E-3 was also sold to the United Kingdom (seven) and France (four) and Saudi Arabia (five, plus eight E-3-derived tanker aircraft). In 1991, when the last aircraft had been delivered, E-3s participated in the Persian Gulf War, playing a crucial role of directing coalition aircraft against Iraqi forces. Throughout the aircraft's service life, numerous upgrades were performed to enhance its capabilities. In 1996, Westinghouse Electric's Defense & Electronic Systems division was acquired by Northrop Corporation, before being renamed Northrop Grumman Mission Systems, which currently supports the E-3's radar. In 1963, the USAF asked for proposals for an Airborne Warning and Control System (AWACS) to replace its EC-121 Warning Stars, which had served in the airborne early warning role for over a decade. The new aircraft would take advantage of improvements in radar technology and in computer aided radar data analysis and data reduction. These developments allowed airborne radars to "look down", detect the movement of low-flying aircraft (see Look-down/shoot-down), and discriminate, even over land, target aircraft's movements—previously this had been impossible, due to the inability to discriminate an aircraft's track from ground clutter. Contracts were issued to Boeing, Douglas, and Lockheed, the latter being eliminated in July 1966. In 1967, a parallel program was put into place to develop the radar, with Westinghouse Electric Corporation and Hughes Aircraft being asked to compete in producing the radar system. In 1968, it was referred to as Overland Radar Technology (ORT) during development tests on the modified EC-121Q. The Westinghouse radar antenna was going to be used by whichever company won the radar competition, since Westinghouse had pioneered in the design of high-power RF phase-shifters, which are used to both focus the RF into a pencil beam, and scan electronically for altitude determination. Boeing initially proposed a purpose-built aircraft, but tests indicated it would not outperform the already-operational 707, so the latter was chosen instead. To increase endurance, this design was to be powered by eight General Electric TF34s. It would carry its radar in a rotating dome mounted at the top of a forward-swept tail, above the fuselage. Boeing was selected ahead of McDonnell Douglas's DC-8-based proposal in July 1970. Initial orders were placed for two aircraft, designated EC-137D as test beds to evaluate the two competing radars. As the test-beds did not need the same 14-hour endurance demanded of the production aircraft, the EC-137s retained the Pratt & Whitney JT3D commercial engines, and a later reduction in the endurance requirement led to retaining the normal engines in production. The first EC-137 made its maiden flight on 9 February 1972, with the fly-off between the two radars taking place from March to July of that year. Favorable test results led to the selection of Westinghouse's radar for the production aircraft. Hughes's radar was initially thought to be a certain winner, simply because much of its design was also going into the new F-15 Eagle's radar program. The Westinghouse radar used a pipelined fast Fourier transform (FFT) to digitally resolve 128 Doppler frequencies, while Hughes's radars used analog filters based on the design for the F-15 fighter. Westinghouse's engineering team won this competition by using a programmable 18-bit computer whose software could be modified before each mission. This computer was the AN/AYK-8 design from the B-57G program, and designated AYK-8-EP1 for its much expanded memory. This radar also multiplexed a beyond-the-horizon (BTH) pulse mode that could complement the pulse-Doppler radar mode. This proved to be beneficial especially when the BTH mode is used to detect ships at sea when the radar beam is directed below the horizon. Approval was given on 26 January 1973 for full-scale development of the AWACS system. To allow further development of the aircraft's systems, orders were placed for three preproduction aircraft, the first of which performed its maiden flight in February 1975. To save costs, the endurance requirements were relaxed, allowing the new aircraft to retain the four JT3D (U.S. military designation TF33) engines. IBM and Hazeltine were selected to develop the mission computer and display system. The IBM computer was designated 4PI, and the software was written in JOVIAL. A Semi-Automatic Ground Environment (SAGE) or back-up interceptor control (BUIC) operator would immediately be at home with the track displays and tabular displays, but differences in symbology would create compatibility problems in tactical ground radar systems in Iceland, mainland Europe, and South Korea over Link-11 (TADIL-A). Modifications to the Boeing 707 for the E-3 Sentry included a rotating radar dome (rotodome), uprated hydraulics from 241 to 345 bar (3500–5000 PSI) to drive the rotodome, single-point ground refueling, air refueling, and a bail-out tunnel or chute. The original design had two (one forward, and one aft), but the aft bail-out chute was deleted to cut mounting costs. Engineering, test and evaluation began on the first E-3 Sentry in October 1975. Between 1977 and 1992, a total of 68 E-3s were built. Because the Boeing 707 is no longer in production, the E-3 mission package has been fitted into the Boeing E-767 for the Japan Air Self Defense Forces. The E-10 MC2A was intended to replace USAF E-3s—along with the RC-135 and the E-8, but the E-10 program was canceled by the Department of Defense. The USAF is now performing a series of incremental improvements, mainly to avionics, to bring the E-3 up to current standards of performance. Boeing is flight-testing its Block 40/45 E-3s. This modified E-3 contains upgrades of the mission crew and air battle management sections, as well as significantly upgraded electronic equipment. Another program that the Air Force is considering is the "Avionics Modernization Program" (AMP). AMP would equip the E-3s with glass cockpits. The Air Force also wants modified E-3s with jet engines that are more reliable than the original ones, and also with at least 19% higher fuel efficiencies. New turbofan engines would give these E-3s longer ranges, longer time-on-station, and a shorter critical runway length. If the modification is carried out, the E-3s could take off with full fuel loads using runways only long, and also at higher ambient temperatures and lower barometric pressures, such as from bases in mountainous areas. The E-8 Joint STARS briefly fitted a test aircraft with the new Pratt & Whitney JT8D-219 turbofans, stated as having one-half the cost of the competing engine, the CFM56, the Air Force is again studying the possibility of replacing the E-3's original turbofan engines with more-efficient ones. NATO intends to extend the operational status of its AWACS until 2035. To comply, fourteen AWACS aircraft will be significantly modified in the "Final Lifetime Extension Program" (FLEP). Most FLEP modifications will be implemented in the communications and operational systems area e.g. expansion of data capacity, expansion of bandwidth for satellite communications, new encryption equipment, new "have quick" radios, upgraded mission computing hard- and software and new operator consoles. The supporting groundsystems (mission training center and mission planning and evaluation system) will also be upgraded to the latest configuration. NATO Airborne Early Warning & Control Program Management Agency (NAPMA) is the preparing and executing authority for the FLEP which will be implemented from 2019–2026. To minimize impact on the operational capacity the NAEW&C force commander will be informed in advance. FLEP will be combined with the standard planned higher echelon technical mainenance. The E-3 Sentry's airframe is a modified Boeing 707-320B Advanced model. USAF and NATO E-3s have an unrefueled range of or 8 hours of flying. The newer E-3 versions bought by France, Saudi Arabia, and the UK are equipped with newer CFM56-2 turbofan engines, and these can fly for about 11 hours or more than . The Sentry's range and on-station time can be increased through air-to-air refueling and the crews can work in shifts by the use of an on-board crew rest and meals area. The aircraft are equipped with one toilet in the rear, and one behind the cockpit. The Saudi E-3 was delivered with an additional toilet in the rear. When deployed, the E-3 monitors an assigned area of the battlefield and provides information for commanders of air operations to gain and maintain control of the battle; while as an air defense asset, E-3s can detect, identify, and track airborne enemy forces far from the boundaries of the U.S. or NATO countries and can direct interceptor aircraft to these targets. In support of air-to-ground operations, the E-3 can provide direct information needed for interdiction, reconnaissance, airlift, and close-air support for friendly ground forces. The unpressurized rotodome is in diameter, thick at the center, and is held above the fuselage by 2 struts. It is tilted down at the front to reduce its aerodynamic drag, which lessens its detrimental effect on take-offs and endurance (which is corrected electronically by both the radar and secondary surveillance radar antenna phase shifters). The rotodome uses bleed air, outside cooling doors, and fluorocarbon based cold plate cooling to maintain the electronic and mechanical equipment temperatures. The hydraulically rotated antenna system permits the Westinghouse Corporation's and AN/APY-2 passive electronically scanned array radar system to provide surveillance from the Earth's surface up into the stratosphere, over land or water. Other major subsystems in the E-3 Sentry are navigation, communications, and computers. 14 consoles display computer-processed data in graphic and tabular format on screens. Its operators perform surveillance, identification, weapons control, battle management and communications functions. The E-3's radar and computer subsystems gather and present broad and detailed battlefield information including position and tracking information of enemy aircraft and vessels, and of course location and status of all friendly assets. Data may be forwarded in real-time to any major command and control center in rear areas or aboard ships. In times of crisis, data may also be forwarded to the National Command Authority in the U.S. via RC-135 or naval aircraft carrier task forces. Electrical generators mounted in each of the E-3's 4r engines provide 1 megawatt of electrical power required by the E-3's radars and electronics. Its pulse-Doppler radar (PD) has a range of more than 250 mi (400 km) for low-flying targets at its operating altitude, and the pulse (BTH) radar has a range of approximately 400 mi (650 km) for aircraft flying at medium to high altitudes. The radar, combined with a secondary surveillance radar (SSR) and electronic support measures (ESM), provides a look down capability, to detect, identify, and track low-flying aircraft, while eliminating ground clutter (radar) returns. Starting in 1987, USAF E-3s were upgraded under the "Block 30/35 Modification Program" to enhance the E-3's capabilities. On 30 October 2001, the final airframe to be upgraded under this program was rolled out. Several major enhancements were made, firstly the installation of ESM and an electronic surveillance capability, for both active and passive means of detection. Also, Joint Tactical Information Distribution System (JTIDS) was installed, which provides rapid and secure communication for transmitting information, including target positions and identification data, to other friendly platforms. Global Positioning System (GPS) capability was also added. Onboard computers were also overhauled to accommodate JTIDS, Link-16, the new ESM systems and to provide for future enhancements. The Radar System Improvement Program (RSIP) was a joint US/NATO development program. RSIP enhances the operational capability of the E-3 radars' electronic countermeasures, and dramatically improve the system's reliability, maintainability, and availability. Essentially, this program replaced the older transistor-transistor logic (TTL) and emitter-coupled logic (MECL) electronic components, long-since out of production, with off-the-shelf digital computers that utilised a High-level programming language instead of assembly language. Significant improvement came from adding pulse compression to the pulse-Doppler mode, replacing the old 8-bit FFT with 24-bit FFTs, and the 12-bit A/D (Sign + 12-bits) with a 15-bit A/D (Sign + 15-bits). These hardware and software modifications improve the E-3 radars' performance, providing enhanced detection with an emphasis towards low radar cross-section (RCS) targets. The RAF had also joined the USAF in adding RSIP to upgrade the E-3's radars. The retrofitting of the E-3 squadrons was completed in December 2000. Along with the RSIP upgrade was installation of the Global Positioning System/Inertial Navigation Systems which dramatically improve positioning accuracy. In 2002, Boeing was awarded a contract to add RSIP to the small French AWACS squadron. Installation was completed in 2006. In the 1990s all NATO AWACS aircraft have been upgraded by the "Radar System Improvement System" (RSIP), a US/NATO joint venture in the area of hard and software modifications of the existing radarsystem. This resulted in an aircraft with 2 large sensorpods alongside the fuselage and a 'bulb' under the aircraft nose containing the electronic warfare equipment. After several years of planning in 2000 NATO decided to bring the E-3's on "Mid Term Program" (MTP) standard. Again this implied technical upgrades and a total multi-sensor-systems integration. MTP ended in 2008. In 2009, the USAF, in cooperation with NATO, entered into a major flight deck avionics modernization program in order to maintain compliance with worldwide airspace mandates. The program, called DRAGON (for DMS Replacement of Avionics for Global Operation and Navigation), was awarded in 2010 – the prime was Boeing and the subcontractor was Rockwell Collins. Drawing on their Flight2 Flight Management System (FMS), almost all the existing avionics were replaced with new, more modern digital equipment from Rockwell Collins. The major upgraded capabilities include a Digital Audio Distribution System, Mode-5/ADS-B transponder, Inmarsat & VDL datalinks, and a Terrain Avoidance and Warning System (TAWS). The navigation system is capable of RNP-0.3, STARs and LNAV/VNAV approaches, and SIDs. The centerpiece flight deck hardware consists of five 6x8 color graphics displays (2 PFDs, 2 IMFDs, and an EICAS), two color CDUs, and a Thales ISD standby. DRAGON lays the foundation for subsequent upgrades including GPS M-Code, Iridium ATC, and Autopilot. USAF DRAGON Production begins in 2018. In March 1977, the 552nd Airborne Warning and Control Wing (now the 552d Air Control Wing) at Tinker AFB, Oklahoma received the first E-3 aircraft, flown and commanded by Major James R. Sterk. The 34th and last USAF Sentry was delivered in June 1984. In March 1996, the USAF activated the 513th Air Control Group (513 ACG), an ACC-gained Air Force Reserve Command (AFRC) AWACS unit under the Reserve Associate Program. Collocated with the 552 ACW at Tinker AFB, the 513 ACG which performs similar duties on active duty E-3 aircraft shared with the 552 ACW. The USAF has a total of thirty-one E-3s in active service. Twenty-seven are stationed at Tinker AFB and belong to the Air Combat Command (ACC). Four are assigned to the Pacific Air Forces (PACAF) and stationed at Kadena AB, Okinawa and Elmendorf AFB, Alaska. One aircraft (TS-3) was assigned to Boeing for testing and development (retired/scrapped June 2012). In 1977, Iran placed an order for ten E-3s, however this order was cancelled following the Iranian Revolution. NATO acquired 18 E-3As and support equipment for a NATO air defense force. Since all aircraft must be registered with a specific country, the decision was made to register the 18 NATO Sentries with Luxembourg, a NATO member that previously did not have any air force. The first NATO E-3 was delivered in January 1982. The eighteen E-3s were operated by Number 1, 2 and 3 Squadrons of NATO's E-3 Component, based at NATO Air Base Geilenkirchen. Presently, 16 NATO E-3As are in the inventory, since one E-3 was lost in a crash and one was retired from service in 2015. NATO, UK, French and USAF AWACS played an important role in the air campaign against Serbia and Montenegro in the former republic of Yugoslavia. From March – June 1999 the aircraft have been deployed in operation Allied Force directing allied strike and air defence aircraft to and from their targets. Over 1,000 aircraft operating from bases in Germany and Italy took part in the air campaign which was intended to destroy Yugoslav air defenses and high-value targets such as the bridges across the Danube river, factories, powerstations, telecommunications facilities, military installations and the Avala TV Tower. The UK and France procured their E-3 aircraft through a joint project. However, while France operate their E-3F aircraft independently of NATO, UK E-3Ds formed the E-3D Component of the NATO Airborne Early Warning and Control Force (NAEWCF), receiving much of their tasking directly from NATO. However, RAF E-3Ds remain UK manned and capable of independent, national tasking outside of NATO command lines. This has been done on numerous occasions throughout the E-3D's service life, notably when the UK committed their E-3Ds to operations over Afghanistan in 2001 and Iraq in 2003. The UK fleet has slowly been reduced from 7 since 2011 with plans for replacement by five E-7 Wedgetails from 2023. France operates four aircraft, all fitted with the newer CFM56-2 engines. The British requirement came about following the cancellation of the British Aerospace Nimrod AEW3 project to replace the Avro Shackleton AEW2 during the 1980s. The UK E-3 order was placed in February 1987, with deliveries starting in 1990. The other operator of the type, delivered between June 1986 and September 1987, is Saudi Arabia which operates five aircraft, all fitted with CFM56-2 engines. This particular sale was hotly contested between the Reagan administration and opponents of the sale, mainly the Israeli lobby in the United States and their allies in Congress. E-3 Sentry aircraft were among the first to deploy during Operation Desert Shield, where they immediately established as an around-the-clock radar screen to defend against Iraqi forces. During Operation Desert Storm, E-3s flew 379 missions and logged 5,052 hours of on-station time. The data collection capability of the E-3 radar and computer subsystems allowed an entire air war to be recorded for the first time. In addition to providing senior leadership with time-critical information on the actions of enemy forces, E-3 controllers assisted in 38 of the 41 air-to-air kills recorded during the conflict. NATO E-3s joined their USAF colleagues for joint air defense as part of Operation Eagle Assist in the wake of the September 11, 2001 terrorist attacks on the World Trade Center towers and the Pentagon. NATO and RAF E-3s participated in the international military operation in Libya. On 27 January 2015, the RAF deployed an E-3D Sentry to Cyprus in support of U.S.-led coalition airstrikes against Islamic State militants in Iraq and Syria. The Sentry joins RAF Panavia Tornado, MQ-9 Reaper, and AirTanker Voyager aircraft performing or supporting almost daily strikes against militants. On 23 June 2015, the first of the original 18 NATO E-3A AWACS aircraft to retire, arrived at Davis-Monthan AFB near Tucson, AZ. The aircraft, LX-N 90449, will be placed in parts reclamation storage where critical parts will be removed by NATO technicians to support their remaining fleet of 16 Boeing E-3A aircraft. It had accumulated 22,206 flight hours between 19 August 1983 and 13 May 2015 and operated out of twenty-one different countries in support of NATO activities. The aircraft was due in mid-July 2015 for a six-year cycle Depot Level Maintenance (DLM) inspection which would have been very costly. Without the inspection, the aircraft would no longer be allowed to fly. The so-called "449 Retirement Project" will result in reclamation of critical parts with a value of upwards of $40,000,000. Some of the parts to be removed are no longer on the market or have become very expensive. On 18 November 2015, an E-3G was deployed to the Middle East to begin immediately flying combat missions in support of Operation Inherent Resolve against ISIL, marking the first combat deployment of the upgraded AWACS Block 40/45. The $2.7 billion development effort started in 2003, with the first five aircraft achieving initial operational capability (IOC) in July 2015. The Block 40/45 upgrade is the most extensive the E-3 has undergone, replacing its 1970s computer technology with an early 2000s standard and including a deployable ground system that receives, processes, and disseminates data. The Air Force plans to convert 24 AWACS to E-3G standard, while retiring seven from the fleet to avoid upgrade costs and harvest out-of-production components. E-3s have been involved in three hull-loss accidents, and one radar antenna was destroyed during RSIP development (see photo under Avionics). Notes Citations Bibliography
https://en.wikipedia.org/wiki?curid=10384
Northrop Grumman E-8 Joint STARS The Northrop Grumman E-8 Joint Surveillance Target Attack Radar System (Joint STARS) is a United States Air Force airborne ground surveillance, battle management and command and control aircraft. It tracks ground vehicles and some aircraft, collects imagery, and relays tactical pictures to ground and air theater commanders. The aircraft is operated by both active duty U.S. Air Force and Air National Guard units and also carries specially trained U.S. Army personnel as additional flight crew. Joint STARS evolved from separate United States Army and Air Force programs to develop technology to detect, locate and attack enemy armor at ranges beyond the forward area of troops. In 1982, the programs were merged and the U.S. Air Force became the lead agent. The concept and sensor technology for the E-8 was developed and tested on the Tacit Blue experimental aircraft. The prime contract was awarded to Grumman Aerospace Corporation in September 1985 for two E-8A development systems. In late 2005, Northrop Grumman was awarded a contract for upgrading engines and other systems. Pratt & Whitney, in a joint venture with Seven Q Seven (SQS), will produce and deliver JT8D-219 engines for the E-8s. Their greater efficiency will allow the Joint STARS to spend more time on station, take off from a wider range of runways, climb faster, fly higher all with a much reduced cost per flying hour. In December 2008, an E-8C test aircraft took its first flight with the new engines. In 2009, the company began engine replacement and additional upgrade efforts. However, the re-engining funding was temporarily halted in 2009 as the Air Force began to consider other options for performing the JSTARS mission. The Air Force began an analysis of alternatives (AOA) in March 2010 for its next generation ground moving target indication (GMTI) radar aircraft fleet. The study was completed in March 2012 and recommended buying a new business jet-based ISR aircraft, such as an Air Force version of the Navy P-8 Poseidon, and the RQ-4B Global Hawk Block 40. However, at a Senate Armed Services Committee meeting on 20 March 2012, the Air Force said it cannot afford a new ISR platform. The E-8 is considered viable in the near and medium terms. As of October 2012, a test plane has had a Hewlett-Packard central computer installed, with work to begin on the rest in 2013. Before the AOA started, Northrop Grumman received funds for two batches of new engines. One set of engines has flown on a JSTARS test plane, and the other set is in storage. The Air Force does not plan to replace the engines of the 16-plane fleet due to the fiscal environment. The company wants to replace the aircraft's data link, but the Air Force will not, due to cost and because they can still receive data through satellite links. Northrop also wants to upgrade its communications with Force XXI Battle Command Brigade and Below because of the Army's shift towards the system. The Air Force says JSTARS is in a phase of capability improvements and is expected to remain in operation through 2030. On 23 January 2014, the USAF revealed a plan for the acquisition of a new business jet-class replacement for the E-8C JSTARS. The program is called JSTARS Recap and plans for the aircraft to reach initial operating capability (IOC) by 2022. The airframe must be more efficient, and separate contracts will be awarded for developing the aircraft, airborne sensor, battle management command and control (BMC2) system, and communications subsystem. JSTARS Recap is currently unfunded and the Air Force FY 2014 budget did not include requests. The program may be launched in FY 2015. On 8 April 2014, the Air Force held an industry day for companies interested in competing for JSTARS Recap; attendees included Boeing, Bombardier Aerospace, and Gulfstream Aerospace. Air Force procurement documents call for a replacement for the Boeing 707-based E-8C as a "business jet class" airframe that is "significantly smaller and more efficient." Current pre-decisional requirements are for an aircraft with a 10-13 person crew with a radar array. Though smaller than the crew and radar size of the E-8C, it could be challenging to meet those demands in a typical business jet and could require a relatively large platform. The staffing and sensor requirements are comparable to the cancelled Northrop Grumman E-10 MC2A, which was originally planned as the E-8's replacement. The Air Force plans to award a contract at the end of FY 2016, a relatively quick pace partly to avoid budget redistributions to other programs. Replacing the E-8C avoids nearly $11 billion in operations and sustainment costs needed to keep the fleet relevant and airworthy. The aircraft is to fly at 38,000 ft for eight hours. Program managers are interested in integrating an FAA-certified flight deck, aerial refueling capability, and potentially full motion video and the joint range extension applications protocol to transmit data to joint agencies at further distances. Another potential feature could be beyond-line-of-sight communications with unmanned aerial vehicles like the RQ-4 Global Hawk. Gulfstream confirmed in late May 2014 that they would offer their Gulfstream G650 for the Air Force's JSTARS replacement. Their bidding strategy is to team with a defense contractor to serve as the integrator. Bombardier is considering offering the Global 6000, in use with the Royal Air Force as the Raytheon Sentinel and the USAF as the E-11A airborne communications relay. Aircraft selection may be based on whether the service wants a large airframe to carry heavy payloads, or a smaller aircraft that would be more nimble and operate from shorter runways. Companies that attended the industry week in April that are contenders for providing electrical systems include Harris Corporation, Rockwell Collins, Lockheed Martin, L-3 Communications, Raytheon, DRS Technologies, and BAE Systems. Boeing plans to offer a solution based on their Boeing 737-700 commercial jetliner airframe; the 737-800 configuration is already in military service with the U.S. Navy as the P-8 Poseidon for maritime surveillance, and would be favored if the Air Force chooses a larger platform. The decision on airframe size may be based on whether the Air Force thinks it can have processing capabilities off-board or if it wants to keep everything on the physical platform. Northrop Grumman has also announced their intention to compete for the future of Joint STARS, although they have not confirmed what airframe they will use. The company has a Gulfstream G550 test aircraft that has been integrated with Joint STARS capabilities and has flown for more than 500 hours. The test aircraft's existence was announced in 2014. Lockheed Martin has teamed with Raytheon and L-3 Communications to offer a JSTARS replacement, but will not decide which platform to use until the Air Force decides if it wants a converted airliner or business jet-sized class aircraft. On 7 August 2015, the Air Force issued contracts to Boeing, Lockheed Martin, and Northrop Grumman for a one-year pre-engineering and manufacturing development effort to mature and test competing designs ahead of a downselect in late 2017. While the E-8C will begin retirement in 2019, an EMD contract will be awarded for two test aircraft, followed by low-rate production of three aircraft for initial operational capability in late 2023, with the remaining 12 aircraft purchased through 2024. On 2 March 2017, Northrop Grumman announced the submittal of its proposal to the U.S. Air Force for the Joint STARS Recapitalization program. During the fiscal 2019 budget rollout briefing it was announced that the Air Force will not move forward with an E-8C replacement aircraft. Funding for the JSTARS recapitalization program will instead be diverted to pay for development of an advanced battle management system. The E-8C is an aircraft modified from the Boeing 707-300 series commercial airliner. The E-8 carries specialized radar, communications, operations and control subsystems. The most prominent external feature is the 40 ft (12 m) canoe-shaped radome under the forward fuselage that houses the 24 ft (7.3 m) APY-7 passive electronically scanned array Side looking airborne radar antenna. The E-8C can respond quickly and effectively to support worldwide military contingency operations. It is a jam-resistant system capable of operating while experiencing heavy electronic countermeasures. The E-8C can fly a mission profile for 9 hours without refueling. Its range and on-station time can be substantially increased through in-flight refueling. The AN/APY-7 radar can operate in wide area surveillance, ground moving target indicator (GMTI), fixed target indicator (FTI) target classification, and synthetic aperture radar (SAR) modes. To pick up moving targets, the Doppler radar looks at the Doppler frequency shift of the returned signal. It can look from a long range, which the military refers to as a high standoff capability. The antenna can be tilted to either side of the aircraft for a 120-degree field of view covering nearly 50,000 km² (19,305 mile²) and can simultaneously track 600 targets at more than 250 km (152 miles). The GMTI modes cannot pick up objects that are too small, insufficiently dense, or stationary. Data processing allows the APY-7 to differentiate between armored vehicles (tracked tanks) and trucks, allowing targeting personnel to better select the appropriate ordnance for various targets. The system's SAR modes can produce images of stationary objects. Objects with many angles (for example, the interior of a pick-up bed) will give a much better radar signature, or specular return. In addition to being able to detect, locate and track large numbers of ground vehicles, the radar has a limited capability to detect helicopters, rotating antennas and low, slow-moving fixed-wing aircraft. The radar and computer subsystems on the E-8C can gather and display broad and detailed battlefield information. Data is collected as events occur. This includes position and tracking information on enemy and friendly ground forces. The information is relayed in near-real time to the US Army's common ground stations via the secure jam-resistant surveillance and control data link (SCDL) and to other ground C4I nodes beyond line-of-sight via ultra high frequency satellite communications. Other major E-8C prime mission equipment are the communications/datalink (COMM/DLX) and operations and control (O&C)subsystems. Eighteen operator workstations display computer-processed data in graphic and tabular format on video screens. Operators and technicians perform battle management, surveillance, weapons, intelligence, communications and maintenance functions. Northrop Grumman has tested the installation of a MS-177 camera on an E-8C to provide real time visual target confirmation. In missions from peacekeeping operations to major theater war, the E-8C can provide targeting data and intelligence for attack aviation, naval surface fire, field artillery and friendly maneuver forces. The information helps air and land commanders to control the battlespace. The E-8's ground-moving radar can tell approximate number of vehicles, location, speed, and direction of travel. It cannot identify exactly what type of vehicle a target is, tell what equipment it has, or discern whether it is friendly, hostile, or a bystander, so commanders often crosscheck the JSTARS data against other sources. In the Army, JSTARS data is analyzed in and disseminated from a Ground Station Module (GSM). The two E-8A development aircraft were deployed in 1991 to participate in Operation Desert Storm under the direction of Albert J. Verderosa, even though they were still in development. The joint program accurately tracked mobile Iraqi forces, including tanks and Scud missiles. Crews flew developmental aircraft on 49 combat sorties, accumulating more than 500 combat hours and a 100% mission effectiveness rate. These Joint STARS developmental aircraft also participated in Operation Joint Endeavor, a NATO peacekeeping mission, in December 1995. While flying in friendly air space, the test-bed E-8A and pre-production E-8C aircraft monitored ground movements to confirm compliance with the Dayton Peace Accords agreements. Crews flew 95 consecutive operational sorties and more than 1,000 flight hours with a 98% mission effectiveness rate. The 93d Air Control Wing, which activated 29 January 1996, accepted its first aircraft, 11 June 1996, and deployed in support of Operation Joint Endeavor in October. The provisional 93d Air Expeditionary Group monitored treaty compliance while NATO rotated troops through Bosnia and Herzegovina. The first production E-8C and a pre-production E-8C flew 36 operational sorties and more than 470 flight hours with a 100% effectiveness rate. The wing declared initial operational capability 18 December 1997 after receiving the second production aircraft. Operation Allied Force saw Joint STARS in action again from February to June 1999 accumulating more than 1,000 flight hours and a 94.5% mission-effectiveness rate in support of the U.S. lead Kosovo War. On 1 October 2002, the 93d Air Control Wing (93 ACW) was "blended" with the 116th Bomb Wing in a ceremony at Robins Air Force Base, Georgia. The 116 BW was an Air National Guard wing equipped with the B-1B Lancer bomber at Robins AFB. As a result of a USAF reorganization of the B-1B force, all B-1Bs were assigned to active duty wings, resulting in the 116 BW lacking a current mission. Extensive efforts by the Georgia's governor and congressional delegation led to the resulting "blending", with the newly created wing designated as the 116th Air Control Wing (116 ACW). The 93 ACW was inactivated the same day. The 116 ACW constituted the first fully blended wing of active duty and Air National Guard airmen. The 116 ACW has been heavily involved in both Operation Enduring Freedom and Operation Iraqi Freedom, earning high marks for operational effectiveness and recently completing 10,000 combat hours. The wing took delivery of the 17th and final E-8C on 23 March 2005. The E-8C Joint STARS routinely supports various taskings of the Combined Force Command Korea during the North Korean winter exercise cycle and for the United Nations enforcing resolutions on Iraq. The twelfth production aircraft, outfitted with an upgraded operations and control subsystem, was delivered to the USAF on 5 November 2001. On 13 March 2009, a Joint STARS aircraft was damaged beyond economical repair when a test plug was left on a fuel tank vent, subsequently causing the fuel tank to rupture during in-flight refueling. There were no casualties but the aircraft sustained $25 million in damage. On 3 September 2009, Loren B. Thompson of the Lexington Institute raised the question of why most of the JSTARS fleet was sitting idle instead of being used to track insurgents in Afghanistan. Thompson states that the JSTARS' radar has an inherent capacity to find what the Army calls 'dismounted' targets—insurgents walking around or placing roadside bombs. Thompson's neutrality has been questioned by some since Lexington Institute has been heavily funded by defense contractors, including Northrop Grumman. Recent trials of JSTARS in Afghanistan are destined to develop tactics, techniques and procedures in tracking dismounted, moving groups of Taliban. On 28 November 2010, amidst escalating danger of war breaking out between North and South Korea, the South Korean government requested the U.S. to implement JSTARS in order to monitor and track North Korean military movements near the DMZ. On 17 January 2011, Northrop Grumman's E-8C Joint Surveillance Target Attack Radar System (Joint STARS) test bed aircraft recently completed the second of two deployments to Naval Air Station Point Mugu, California, in support of the U.S. Navy Joint Surface Warfare Joint Capability Technology Demonstration to test its Network-Enabled Weapon (NEW) architecture. The Joint STARS aircraft executed three Operational Utility Assessment flights and demonstrated its ability to guide anti-ship weapons against surface combatants at a variety of standoff distances in the NEW architecture. The Joint STARS aircraft served as the network command-and-control node, as well as a node for transmitting in-flight target message updates to an AGM-154 C-1 Joint Standoff Weapon carried by U.S. Navy F/A-18 Hornets using its advanced long range tracking and targeting capability. From 2001 to January 2011 the Joint STARS fleet has flown over 63,000 hours in 5,200 combat missions in support of Operations Iraqi Freedom, Enduring Freedom and New Dawn. On 1 October 2011, the "blended" wing construct of the 116th Air Control Wing (116 ACW), combining Air National Guard and Regular Air Force personnel in a single unit was discontinued. On this date, the 461st Air Control Wing (461 ACW) was established at Robins AFB as the Air Force's sole active duty E-8 JSTARS wing while the 116 ACW reverted to a traditional Air National Guard wing within the Georgia Air National Guard. Both units share the same E-8 aircraft and will often fly with mixed crews, but now function as separate units. Air Combat Command - 1991–present
https://en.wikipedia.org/wiki?curid=10385
Eric Cheney Eric Cheney (5 January 1924 – 30 December 2001) was an English motorcycle racer, designer and independent constructor. He was known as one of the best motorcycle frame designers of his era, concentrating mainly in the off-road competition aftermarket. Cheney attended a school in Winchester based on Lancasterian principles before joining the Royal Navy at the age of 18, where he served on wartime Arctic convoys and in motor torpedo boats and gained experience of engineering and working on high performance engines. Cheney also worked on the development of remote controlled submarines for the Royal Navy. After World War II, Cheney joined the motorcycle dealers Archers of Aldershot as a mechanic. Cheney began racing motocross and became one Britain's best riders, along with his travelling companion Les Archer, who went on to become European champion. He had ten successful years on the Continental circuit but a prolonged illness due to an infection contracted while racing in Algeria ended his riding career. He moved into bike preparation and designs for motorcycle chassis and suspension systems. Cheney had no formal training as a motorcycle designer yet was able to create original and high-performance motorcycle chassis designs working in a simple workshop that was essentially a domestic garage. His approach has been described as "like a medieval engineer" as in an age of computer-aided design and significant resources for research and development teams, he worked entirely by intuition. Eric relied on his long personal experience of international off-road competition riding and would prepare his initial designs for a new motorcycle frame in chalk on the wall of his workshop. Experimenting with different lines until he was satisfied, Cheney would then form the steel tubing using his chalk drawings as a guide. Only when he had built a working prototype motorcycle would he start work on a final jig for mass production. He was once quoted as saying "I know when it's right and it screams at me when it's wrong." In the late 1960s, the British motorcycle industry was unable to support a national team to compete in the International Six Days Trial so, Cheney hand built a limited number of ISDT Cheney-Triumphs using his own design of twin down-tube frame with a specially tuned Triumph 5TA engine. Fitted with tapered conical hubs, special motocross forks and large alloy fuel tanks, a Cheney Triumph was first used in the 1968 British Trophy Team. In 1970 and 1971 three 504cc Cheney Triumphs were used by the British team in the ISDT, in which Cheney won a manufacturer's prize. Replicas were built, but production was short-lived due to a shortage of engines. Cheney's most noted successes were in the Grand Prix road racing championships, with Phil Read using his chassis in tandem with a Yamaha engine to win the 1971 250cc world championship. His designs were the last British ones to win a Grand Prix. He never worked for any of the major manufacturers but maintained a productive relationship with BSA in its heyday. After the demise of BSA in 1972, Cheney joined with former BSA factory rider John Banks to develop and campaign a highly successful BSA powered motocross bike. Some of Cheney's motorcycle designs are now famous in their own right, such as the competition BSA Gold Stars of Jerry Scott and Keith Hickman and the John Banks replica which used a BSA B50 engine specially tuned by Cheney. He also built some racing frames for Suzuki Grand Prix motorcycles in 1968 and, it has been suggested that Suzuki engineers incorporated features of Cheney's designs, such as magnesium hubs and lower fork legs into production road going motorcycles. Cheney's company was originally known as Eric Cheney Designs, then changed to Inter-Moto, now known as Cheney Racing. American actor Steve McQueen, an experienced off-road rider who represented the United States in the ISDT bought a number of Cheney's motorcycles at full price because he considered them better than other makes.
https://en.wikipedia.org/wiki?curid=10388
Econometrics Econometrics is the application of statistical methods to economic data in order to give empirical content to economic relationships. More precisely, it is "the quantitative analysis of actual economic phenomena based on the concurrent development of theory and observation, related by appropriate methods of inference". An introductory economics textbook describes econometrics as allowing economists "to sift through mountains of data to extract simple relationships". The first known use of the term "econometrics" (in cognate form) was by Polish economist Paweł Ciompa in 1910. Jan Tinbergen is considered by many to be one of the founding fathers of econometrics. Ragnar Frisch is credited with coining the term in the sense in which it is used today. A basic tool for econometrics is the multiple linear regression model. "Econometric theory" uses statistical theory and mathematical statistics to evaluate and develop econometric methods. Econometricians try to find estimators that have desirable statistical properties including unbiasedness, efficiency, and consistency. "Applied econometrics" uses theoretical econometrics and real-world data for assessing economic theories, developing econometric models, analysing economic history, and forecasting. A basic tool for econometrics is the multiple linear regression model. In modern econometrics, other statistical tools are frequently used, but linear regression is still the most frequently used starting point for an analysis. Estimating a linear regression on two variables can be visualised as fitting a line through data points representing paired values of the independent and dependent variables. For example, consider Okun's law, which relates GDP growth to the unemployment rate. This relationship is represented in a linear regression where the change in unemployment rate (formula_1) is a function of an intercept (formula_2), a given value of GDP growth multiplied by a slope coefficient formula_3 and an error term, formula_4: The unknown parameters formula_2 and formula_3 can be estimated. Here formula_3 is estimated to be −1.77 and formula_2 is estimated to be 0.83. This means that if GDP growth increased by one percentage point, the unemployment rate would be predicted to drop by 1.77 points. The model could then be tested for statistical significance as to whether an increase in growth is associated with a decrease in the unemployment, as hypothesized. If the estimate of formula_3 were not significantly different from 0, the test would fail to find evidence that changes in the growth rate and unemployment rate were related. The variance in a prediction of the dependent variable (unemployment) as a function of the independent variable (GDP growth) is given in polynomial least squares. Econometric theory uses statistical theory and mathematical statistics to evaluate and develop econometric methods. Econometricians try to find estimators that have desirable statistical properties including unbiasedness, efficiency, and consistency. An estimator is unbiased if its expected value is the true value of the parameter; it is consistent if it converges to the true value as the sample size gets larger, and it is efficient if the estimator has lower standard error than other unbiased estimators for a given sample size. Ordinary least squares (OLS) is often used for estimation since it provides the BLUE or "best linear unbiased estimator" (where "best" means most efficient, unbiased estimator) given the Gauss-Markov assumptions. When these assumptions are violated or other statistical properties are desired, other estimation techniques such as maximum likelihood estimation, generalized method of moments, or generalized least squares are used. Estimators that incorporate prior beliefs are advocated by those who favour Bayesian statistics over traditional, classical or "frequentist" approaches. "Applied econometrics" uses theoretical econometrics and real-world data for assessing economic theories, developing econometric models, analysing economic history, and forecasting. Econometrics may use standard statistical models to study economic questions, but most often they are with observational data, rather than in controlled experiments. In this, the design of observational studies in econometrics is similar to the design of studies in other observational disciplines, such as astronomy, epidemiology, sociology and political science. Analysis of data from an observational study is guided by the study protocol, although exploratory data analysis may be useful for generating new hypotheses. Economics often analyses systems of equations and inequalities, such as supply and demand hypothesized to be in equilibrium. Consequently, the field of econometrics has developed methods for identification and estimation of simultaneous equations models. These methods are analogous to methods used in other areas of science, such as the field of system identification in systems analysis and control theory. Such methods may allow researchers to estimate models and investigate their empirical consequences, without directly manipulating the system. One of the fundamental statistical methods used by econometricians is regression analysis. Regression methods are important in econometrics because economists typically cannot use controlled experiments. Econometricians often seek illuminating natural experiments in the absence of evidence from controlled experiments. Observational data may be subject to omitted-variable bias and a list of other problems that must be addressed using causal analysis of simultaneous-equation models. In addition to natural experiments, quasi-experimental methods have been used increasingly commonly by econometricians since the 1980s, in order to credibly identify causal effects. A simple example of a relationship in econometrics from the field of labour economics is: This example assumes that the natural logarithm of a person's wage is a linear function of the number of years of education that person has acquired. The parameter formula_12 measures the increase in the natural log of the wage attributable to one more year of education. The term formula_4 is a random variable representing all other factors that may have direct influence on wage. The econometric goal is to estimate the parameters, formula_14 under specific assumptions about the random variable formula_4. For example, if formula_4 is uncorrelated with years of education, then the equation can be estimated with ordinary least squares. If the researcher could randomly assign people to different levels of education, the data set thus generated would allow estimation of the effect of changes in years of education on wages. In reality, those experiments cannot be conducted. Instead, the econometrician observes the years of education of and the wages paid to people who differ along many dimensions. Given this kind of data, the estimated coefficient on Years of Education in the equation above reflects both the effect of education on wages and the effect of other variables on wages, if those other variables were correlated with education. For example, people born in certain places may have higher wages and higher levels of education. Unless the econometrician controls for place of birth in the above equation, the effect of birthplace on wages may be falsely attributed to the effect of education on wages. The most obvious way to control for birthplace is to include a measure of the effect of birthplace in the equation above. Exclusion of birthplace, together with the assumption that formula_17 is uncorrelated with education produces a misspecified model. Another technique is to include in the equation additional set of measured covariates which are not instrumental variables, yet render formula_12 identifiable. An overview of econometric methods used to study this problem were provided by Card (1999). The main journals that publish work in econometrics are "Econometrica", the "Journal of Econometrics", "The Review of Economics and Statistics", "Econometric Theory", the "Journal of Applied Econometrics", "Econometric Reviews", "The Econometrics Journal", "Applied Econometrics and International Development", and the "Journal of Business & Economic Statistics". Like other forms of statistical analysis, badly specified econometric models may show a spurious relationship where two variables are correlated but causally unrelated. In a study of the use of econometrics in major economics journals, McCloskey concluded that some economists report p-values (following the Fisherian tradition of tests of significance of point null-hypotheses) and neglect concerns of type II errors; some economists fail to report estimates of the size of effects (apart from statistical significance) and to discuss their economic importance. She also argues that some economists also fail to use economic reasoning for model selection, especially for deciding which variables to include in a regression. In some cases, economic variables cannot be experimentally manipulated as treatments randomly assigned to subjects. In such cases, economists rely on observational studies, often using data sets with many strongly associated covariates, resulting in enormous numbers of models with similar explanatory ability but different covariates and regression estimates. Regarding the plurality of models compatible with observational data-sets, Edward Leamer urged that "professionals ... properly withhold belief until an inference can be shown to be adequately insensitive to the choice of assumptions".
https://en.wikipedia.org/wiki?curid=10390
Emacs Lisp Emacs Lisp is a dialect of the Lisp programming language used as a scripting language by Emacs (a text editor family most commonly associated with GNU Emacs and XEmacs). It is used for implementing most of the editing functionality built into Emacs, the remainder being written in C, as is the Lisp interpreter. Emacs Lisp is also termed Elisp, although there is also an older, unrelated Lisp dialect with that name. Users of Emacs commonly write Emacs Lisp code to customize and extend Emacs. Other options include the "Customize" feature that's been in GNU Emacs since version 20. Itself written in Emacs Lisp, Customize provides a set of preferences pages allowing the user to set options and preview their effect in the running Emacs session. When the user saves their changes, Customize simply writes the necessary Emacs Lisp code to the user's config file, which can be set to a special file that only Customize uses, to avoid the possibility of altering the user's own file. Emacs Lisp can also function as a scripting language, much like the Unix Bourne shell or Perl, by calling Emacs in "batch mode". In this way it may be called from the command line or via an executable file, and its editing functions, such as buffers and movement commands are available to the program just as in the normal mode. No user interface is presented when Emacs is started in batch mode; it simply executes the passed-in script and exits, displaying any output from the script. Emacs Lisp is most closely related to Maclisp, with some later influence from Common Lisp. It supports imperative and functional programming methods. Richard Stallman chose Lisp as the extension language for his rewrite of Emacs (the original used Text Editor and Corrector (TECO) as its extension language) because of its powerful features, including the ability to treat functions as data. Although the Common Lisp standard had yet to be formulated, Scheme existed at the time Stallman was rewriting Gosling Emacs into GNU Emacs. He chose not to use it because of its comparatively poor performance on workstations (as opposed to the minicomputers that were Emacs' traditional home), and he wanted to develop a dialect which he thought would be more easily optimized. The Lisp dialect used in Emacs differs substantially from the more modern Common Lisp and Scheme dialects used for applications programming. A prominent characteristic of Emacs Lisp is in its use of dynamic rather than lexical scope by default (see below). That is, a function may reference local variables in the scope it is called from, but not in the scope where it was defined. To understand the logic behind Emacs Lisp, it is important to remember that there is an emphasis on providing data structures and features specific to making a versatile text editor over implementing a general-purpose programming language. For example, Emacs Lisp cannot easily read a file a line at a time—the entire file must be read into an Emacs buffer. However, Emacs Lisp provides many features for navigating and modifying buffer text at a sentence, paragraph, or higher syntactic level as defined by modes. Here follows a simple example of an Emacs extension written in Emacs Lisp. In Emacs, the editing area can be split into separate areas called "windows", each displaying a different "buffer". A buffer is a region of text loaded into Emacs' memory (possibly from a file) which can be saved into a text document. Users can press the default C-x 2 key binding to open a new window. This runs the Emacs Lisp function codice_1. Normally, when the new window appears, it displays the same buffer as the previous one. Suppose we wish to make it display the next available buffer. In order to do this, the user writes the following Emacs Lisp code, in either an existing Emacs Lisp source file or an empty Emacs buffer: The first statement, codice_2, defines a new function, codice_3, which calls codice_1 (the old window-splitting function), then tells the new window to display another (new) buffer. The second statement, codice_5 re-binds the key sequence "C-x 2" to the new function. This can also be written using the feature called "advice", which allows the user to create wrappers around existing functions instead of defining their own. This has the advantage of not requiring keybindings to be changed and working wherever the original function is called, as well as being simpler to write but the disadvantage of making debugging more complicated. For this reason, "advice" is not allowed in the source code of GNU Emacs, but if a user wishes, the advice feature can be used in their code to reimplement the above code as follows: (defadvice split-window-below This instructs codice_1 to execute the user-supplied code whenever it is called, before executing the rest of the function. Advice can also be specified to execute after the original function, around it—literally wrapping the original, or to conditionally execute the original function based on the results of the advice. Emacs 24.4 replaces this codice_7 mechanism with codice_8, which is claimed to be more flexible and simpler. The advice above could be reimplemented using the new system as: These changes take effect as soon as the code is evaluated. It is not necessary to recompile, restart Emacs, or even rehash a configuration file. If the code is saved into an Emacs init file, then Emacs will load the extension the next time it starts. Otherwise, the changes must be reevaluated manually when Emacs is restarted. Emacs Lisp code is stored in filesystems as plain text files, by convention with the filename suffix ".el". The user's init file is an exception, often appearing as ".emacs" despite being evaluated as any Emacs Lisp code. Recent versions of Emacs ("recent" in a 40-year-old program meaning roughly any version released since the mid-1990s) will also load ~/.emacs.el and ~/.emacs.d/init.el. Additionally, users may specify any file to load as a config file on the command line, or explicitly state that no config file is to be loaded. When the files are loaded, an interpreter component of the Emacs program reads and parses the functions and variables, storing them in memory. They are then available to other editing functions, and to user commands. Functions and variables can be freely modified and redefined without restarting the editor or reloading the config file. In order to save time and memory space, much of the functionality of Emacs loads only when required. Each set of optional features shipped with Emacs is implemented by a collection of Emacs code called a package or library. For example, there is a library for highlighting keywords in program source code, and a library for playing the game of Tetris. Each library is implemented using one or more Emacs Lisp source files. Libraries can define one or more "major modes" to activate and control their function. Emacs developers write certain functions in C. These are "primitives", also termed "built-in functions" or "subrs". Although primitives can be called from Lisp code, they can only be modified by editing the C source files and recompiling. In GNU Emacs, primitives are not available as external libraries; they are part of the Emacs executable. In XEmacs, runtime loading of such primitives is possible, using the operating system's support for dynamic linking. Functions may be written as primitives because they need access to external data and libraries not otherwise available from Emacs Lisp, or because they are called often enough that the comparative speed of C versus Emacs Lisp makes a worthwhile difference. However, because errors in C code can easily lead to segmentation violations or to more subtle bugs, which crash the editor, and because writing C code that interacts correctly with the Emacs Lisp garbage collector is error-prone, the number of functions implemented as primitives is kept to a necessary minimum. "Byte-compiling" can make Emacs Lisp code execute faster. Emacs contains a compiler which can translate Emacs Lisp source files into a special representation termed bytecode. Emacs Lisp bytecode files have the filename suffix ".elc". Compared to source files, bytecode files load faster, occupy less space on the disk, use less memory when loaded, and run faster. Bytecode still runs more slowly than primitives, but functions loaded as bytecode can be easily modified and re-loaded. In addition, bytecode files are platform-independent. The standard Emacs Lisp code distributed with Emacs is loaded as bytecode, although the matching source files are usually provided for the user's reference as well. User-supplied extensions are typically not byte-compiled, as they are neither as large nor as computationally intensive. Notably, the "cl" package implements a fairly large subset of Common Lisp. Emacs Lisp (unlike some other Lisp implementations) does not do tail-call optimization. Without this, tail recursions can eventually lead to stack overflow. The apel library aids in writing portable Emacs Lisp code, with the help of the polysylabi platform bridge. Emacs Lisp is a Lisp-2 meaning that it has a function namespace which is separate from the namespace it uses for other variables. Like MacLisp, Emacs Lisp uses dynamic scope, offering static (or lexical) as an option starting from version 24. It can be activated by setting the file local variable codice_9. In dynamic scoping, if a programmer declares a variable within the scope of a function, it is available to subroutines called from within that function. Originally, this was intended as an optimization; lexical scoping was still uncommon and of uncertain performance. "I asked RMS when he was implementing emacs lisp why it was dynamically scoped and his exact reply was that lexical scope was too inefficient." Dynamic scoping was also meant to provide greater flexibility for user customizations. However, dynamic scoping has several disadvantages. Firstly, it can easily lead to bugs in large programs, due to unintended interactions between variables in different functions. Secondly, accessing variables under dynamic scoping is generally slower than under lexical scoping. Also, the codice_10 macro in the "cl" package does provide effective lexical scope to Emacs Lisp programmers, but while "cl" is widely used, codice_10 is rarely used.
https://en.wikipedia.org/wiki?curid=10392
Edward Bulwer-Lytton Edward George Earle Lytton Bulwer-Lytton, 1st Baron Lytton, PC (25 May 1803 – 18 January 1873) was an English writer and politician. He served as a Whig member of Parliament from 1831 to 1841 and a Conservative from 1851 to 1866. He was Secretary of State for the Colonies from June 1858 to June 1859, choosing Richard Clement Moody as founder of British Columbia. He declined the Crown of Greece in 1862 after King Otto abdicated. He was created Baron Lytton of Knebworth in 1866. His marriage to the writer Rosina Bulwer Lytton broke down. Her detention in an insane asylum provoked a public outcry. Bulwer-Lytton's works sold and paid him well. He coined the phrases "the great unwashed", "pursuit of the almighty dollar", "the pen is mightier than the sword", and "dweller on the threshold", and the opening phrase "It was a dark and stormy night." Yet his standing declined and he is little read today. The sardonic Bulwer-Lytton Fiction Contest, held annually since 1982, claims to seek the "opening sentence of the worst of all possible novels". Bulwer was born on 25 May 1803 to General William Earle Bulwer of Heydon Hall and Wood Dalling, Norfolk and Elizabeth Barbara Lytton, daughter of Richard Warburton Lytton of Knebworth House, Hertfordshire. He had two older brothers, William Earle Lytton Bulwer (1799–1877) and Henry (1801–1872), later Lord Dalling and Bulwer. His father died and his mother moved to London when he was four years old. When he was 15, a tutor named Wallington, who tutored him at Ealing, encouraged him to publish an immature work: "Ishmael and Other Poems". Around this time, Bulwer fell in love, but the woman's father induced her to marry another man. She died about the time that Bulwer went to Cambridge and he stated that her loss affected all his subsequent life. In 1822 Bulwer-Lytton entered Trinity College, Cambridge, where he met John Auldjo, but soon moved to Trinity Hall. In 1825 he won the Chancellor's Gold Medal for English verse. In the following year he took his BA degree and printed for private circulation a small volume of poems, "Weeds and Wild Flowers". He purchased an army commission in 1826, but sold it in 1829 without serving. In August 1827, he married Rosina Doyle Wheeler (1802–1882), a noted Irish beauty, but against the wishes of his mother, who withdrew his allowance, forcing him to work for a living. They had two children, Lady Emily Elizabeth Bulwer-Lytton (1828–1848), and (Edward) Robert Lytton Bulwer-Lytton, 1st Earl of Lytton (1831–1891) who became Governor-General and Viceroy of British India (1876–1880). His writing and political work strained their marriage and his infidelity embittered Rosina. In 1833 they separated acrimoniously and in 1836 the separation became legal. Three years later, Rosina published "Cheveley, or the Man of Honour" (1839), a near-libellous fiction satirising her husband's alleged hypocrisy. In June 1858, when her husband was standing as parliamentary candidate for Hertfordshire, she denounced him at the hustings. He retaliated by threatening her publishers, withholding her allowance and denying her access to their children. Finally he had her committed to a mental asylum, but she was released a few weeks later after a public outcry. This she chronicled in a memoir, "A Blighted Life" (1880). She continued attacking her husband's character for several years. The death of Bulwer's mother in 1843 meant his "exhaustion of toil and study had been completed by great anxiety and grief," and by "about the January of 1844, I was thoroughly shattered." In his mother's room at Knebworth House, which he inherited, he "had inscribed above the mantelpiece a request that future generations preserve the room as his beloved mother had used it." It remains hardly changed to this day. On 20 February 1844, in accordance with his mother's will, he changed his surname from Bulwer to Bulwer-Lytton and assumed the arms of Lytton by royal licence. His widowed mother had done the same in 1811. His brothers remained plain "Bulwer". By chance Bulwer-Lytton encountered a copy of "Captain Claridge's work on the "Water Cure", as practised by Priessnitz, at Graefenberg", and "making allowances for certain exaggerations therein", pondered the option of travelling to Graefenberg, but preferred to find something closer to home, with access to his own doctors in case of failure: "I who scarcely lived through a day without leech or potion!". After reading a pamphlet by Doctor James Wilson, who operated a hydropathic establishment with James Manby Gully at Malvern, he stayed there for "some nine or ten weeks", after which he "continued the system some seven weeks longer under Doctor Weiss, at Petersham", then again at "Doctor Schmidt's magnificent hydropathic establishment at Boppart" (at the former Marienberg Convent at Boppard), after developing a cold and fever upon his return home. When Otto, King of Greece abdicated in 1862, Bulwer-Lytton was offered the Greek Crown, but declined. The English Rosicrucian society, founded in 1867 by Robert Wentworth Little, claimed Bulwer-Lytton as their "Grand Patron", but he wrote to the society complaining that he was "extremely surprised" by their use of the title, as he had "never sanctioned such." Nevertheless, a number of esoteric groups have continued to claim Bulwer-Lytton as their own, chiefly because some of his writings – such as the 1842 book "Zanoni" – have included Rosicrucian and other esoteric notions. According to the Fulham Football Club, he once resided in the original Craven Cottage, today the site of their stadium. Bulwer-Lytton had long suffered from a disease of the ear, and for the last two or three years of his life lived in Torquay nursing his health. After an operation to cure deafness, an abscess formed in the ear and burst; he endured intense pain for a week and died at 2 am on 18 January 1873, just short of his 70th birthday. The cause of death was unclear but it was thought the infection had affected his brain and caused a fit. Rosina outlived him by nine years. Against his wishes, Bulwer-Lytton was honoured with a burial in Westminster Abbey. His unfinished history "Athens: Its Rise and Fall" was published posthumously. Bulwer began his political career as a follower of Jeremy Bentham. In 1831 he was elected member for St Ives, Huntingdonshire, after which he was returned for Lincoln in 1832, and sat in Parliament for that city for nine years. He spoke in favour of the Reform Bill and took the lead in securing the reduction, after vainly essaying the repeal, of the newspaper stamp duties. His influence was perhaps most keenly felt when, on the Whigs' dismissal from office in 1834, he issued a pamphlet entitled "A Letter to a Late Cabinet Minister on the Crisis". Lord Melbourne, then Prime Minister, offered him a lordship of the Admiralty, which he declined as likely to interfere with his activity as an author. Bulwer was created a baronet, of Knebworth House in the County of Hertford, in the Baronetage of the United Kingdom, in 1838. In 1841, he left Parliament and spent much of his time in travel. He did not return to politics until 1852, when having differed from Lord John Russell over the Corn Laws, he stood for Hertfordshire as a Conservative. Bulwer-Lytton held that seat until 1866, when he was raised to the peerage as Baron Lytton of Knebworth in the County of Hertford. In 1858 he entered Lord Derby's government as Secretary of State for the Colonies, thus serving alongside his old friend Disraeli. He was comparatively inactive in the House of Lords. When news of the Fraser Canyon Gold Rush reached London, Bulwer-Lytton, as Secretary of State for the Colonies, requested that the War Office recommend a field officer, "a man of good judgement possessing a knowledge of mankind", to lead a Corps of 150 (later increased to 172) Royal Engineers, who had been selected for their "superior discipline and intelligence". The War Office chose Richard Clement Moody, and Lord Lytton, who described Moody as his "distinguished friend", accepted the nomination in view of Moody's military record, his success as Governor of the Falkland Islands, and the distinguished record of his father, Colonel Thomas Moody, Knight at the Colonial Office. Moody was charged to establish British order and transform the newly established Colony of British Columbia (1858–66) into the British Empire's "bulwark in the farthest west" and "found a second England on the shores of the Pacific." Lytton desired to send to the colony "representatives of the best of British culture, not just a police force": he sought men who possessed "courtesy, high breeding and urbane knowledge of the world," and decided to send Moody, whom the Government considered to be the archetypal "English gentleman and British Officer." at the head of the Royal Engineers, Columbia Detachment, to whom he wrote an impassioned letter. The former HBC Fort Dallas at Camchin, the confluence of the Thompson and Fraser Rivers, was renamed in his honour by Governor Sir James Douglas in 1858 as Lytton, British Columbia. Bulwer-Lytton's literary career began in 1820 with the publication of a book of poems and spanned much of the 19th century. He wrote in a variety of genres, including historical fiction, mystery, romance, the occult and science fiction. He financed his extravagant way of life with a varied and prolific literary output, sometimes publishing anonymously. Bulwer-Lytton published "Falkland" in 1827, a novel which was only a moderate success. But "Pelham" brought him public acclaim in 1828 and established his reputation as a wit and dandy. Its intricate plot and humorous, intimate portrayal of pre-Victorian dandyism kept gossips busy trying to associate public figures with characters in the book. "Pelham" resembled Benjamin Disraeli's first novel "Vivian Grey" (1827). The character of the villainous Richard Crawford in "The Disowned", also published in 1828, borrowed much from that of banker and forger Henry Fauntleroy, who was hanged in London in 1824 before a crowd of some 100,000. Bulwer-Lytton admired Disraeli's father Isaac D'Israeli, himself a noted author. They began corresponding in the late 1820s and met for the first time in March 1830, when Isaac D'Israeli dined at Bulwer-Lytton's house. Also present that evening were Charles Pelham Villiers and Alexander Cockburn. The young Villiers had a long parliamentary career, while Cockburn became Lord Chief Justice of England in 1859. Bulwer-Lytton reached his height of popularity with the publication of "Godolphin" (1833). This was followed by "The Pilgrims of the Rhine" (1834), "The Last Days of Pompeii" (1834), "Rienzi, Last of the Roman Tribunes" (1835), "Leila; or, The Siege of Granada" (1838), and "Harold, the Last of the Saxons" (1848). "The Last Days of Pompeii" was inspired by Karl Briullov's painting "The Last Day of Pompeii", which Bulwer-Lytton saw in Milan. His "New Timon" lampooned Tennyson, who responded in kind. Bulwer-Lytton also wrote the horror story "The Haunted and the Haunters" or "The House and the Brain" (1859). Another novel with a supernatural theme was "A Strange Story" (1862), which was an influence on Bram Stoker's "Dracula". Bulwer-Lytton wrote many other works, including "Vril: The Power of the Coming Race" (1871) which drew heavily on his interest in the occult and contributed to the early growth of the science fiction genre. Its story of a subterranean race waiting to reclaim the surface of the Earth is an early science fiction theme. The book popularised the Hollow Earth theory and may have inspired Nazi mysticism. His term "vril" lent its name to Bovril meat extract. "Vril" has been adopted by theosophists and occultists since the 1870s and became closely associated with the ideas of an esoteric neo-Nazism after 1945. His play "Money" (1840) was first produced at the Theatre Royal, Haymarket, London, on 8 December 1840. The first American production was at the Old Park Theater in New York on 1 February 1841. Subsequent productions include the Prince of Wales's Theatre's in 1872 and as the inaugural play at the new California Theatre (San Francisco) in 1869. Among Bulwer-Lytton's lesser-known contributions to literature was that he convinced Charles Dickens to revise the ending of "Great Expectations" to make it more palatable to the reading public, as in the original version of the novel, Pip and Estella do not get together. Bulwer-Lytton's most famous quotation is "The pen is mightier than the sword" from his play "Richelieu": beneath the rule of men entirely great, the pen is mightier than the sword He popularized the phrase "pursuit of the almighty dollar" from his novel "The Coming Race", and he is credited with "the great unwashed", using this disparaging term in his 1830 novel "Paul Clifford": He is certainly a man who bathes and "lives cleanly", (two especial charges preferred against him by Messrs. the Great Unwashed). The writers of theosophy were among those influenced by Bulwer-Lytton's work. Annie Besant and especially Helena Blavatsky incorporated his thoughts and ideas, particularly from "The Last Days of Pompeii", "Vril, the Power of the Coming Race" and "Zanoni" in her own books. Bulwer-Lytton's name lives on in the annual Bulwer-Lytton Fiction Contest, in which contestants think up terrible openings for imaginary novels, inspired by the first line of his 1830 novel "Paul Clifford": It was a dark and stormy night; the rain fell in torrentsexcept at occasional intervals, when it was checked by a violent gust of wind which swept up the streets (for it is in London that our scene lies), rattling along the housetops, and fiercely agitating the scanty flame of the lamps that struggled against the darkness. Entrants in the contest seek to capture the rapid changes in point of view, the florid language, and the atmosphere of the full sentence. The opening was popularized by the "Peanuts" comic strip, in which Snoopy's sessions on the typewriter usually began with It was a dark and stormy night. The same words also form the first sentence of Madeleine L'Engle's Newbery Medal-winning novel "A Wrinkle in Time." Similar wording appears in Edgar Allan Poe's 1831 short story, "The Bargain Lost", although not at the very beginning. It reads: It was a dark and stormy night. The rain fell in cataracts; and drowsy citizens started, from dreams of the deluge, to gaze upon the boisterous sea, which foamed and bellowed for admittance into the proud towers and marble palaces. Who would have thought of passions so fierce in that calm water that slumbers all day long? At a slight alabaster stand, trembling beneath the ponderous tomes which it supported, sat the hero of our story. Several of Bulwer-Lytton's novels were made into operas. One of them, "Rienzi, der Letzte der Tribunen" (1842) by Richard Wagner, eventually became more famous than the novel. "Leonora" (1846) by William Henry Fry, the first European-styled "grand" opera composed in the United States, is based on Bulwer-Lytton's play "The Lady of Lyons", as is Frederic Cowen's first opera "Pauline" (1876). Verdi rival Errico Petrella's most successful opera, "Jone" (1858), was based upon Bulwer-Lytton's "The Last Days of Pompeii", and was performed all over the world until the First World War. "Harold, the Last of the Saxons" (1848) was the source for Verdi's opera "Aroldo" in 1857. In addition to his political and literary work, Bulwer-Lytton became the editor of the "New Monthly" in 1831, but he resigned the following year. In 1841, he started the "Monthly Chronicle", a semi-scientific magazine. During his career he wrote poetry, prose, and stage plays; his last novel was "Kenelm Chillingly", which was in course of publication in "Blackwood’s Magazine" at the time of his death in 1873. Bulwer-Lytton's works of fiction and non-fiction were translated in his day and since then into many languages, including Serbian (by Laza Kostic), German, Russian, Norwegian, Swedish, French, Finnish, and Spanish. In 1879, his "Ernest Maltravers" was the first complete novel from the West to be translated into Japanese. In Queensland, Australia the Brisbane suburb of Lytton is to be found on Bulwer Island which today is home to the Port of Brisbane. Also in Queensland on Moreton Island (Moorgumpin) is located another settlement by the name of Bulwer. The township of Lytton, Quebec (today part of Montcerf-Lytton) was named after him as was Lytton, British Columbia, and Lytton, Iowa. Lytton Road in Gisborne, New Zealand was named after the novelist. Later a state secondary school, Lytton High School, was founded in the road. Bulwer-Lytton was portrayed by the actor Brett Usher in the 1978 television serial "Disraeli". Bulwer-Lytton ebooks Other links
https://en.wikipedia.org/wiki?curid=10393
History of Esperanto L. L. Zamenhof developed Esperanto in the 1870s and 80s and published the first publication about it, "Unua Libro", in 1887. The number of Esperanto speakers has grown gradually since then, although it has not had much support from governments and international organizations and has sometimes been outlawed or otherwise suppressed. Around 1880, while in Moscow and approximately simultaneously with working on Esperanto, Zamenhof made an aborted attempt to standardize Yiddish, based on his native Bialystok (Northeastern) dialect, as a unifying language for the Jews of the Russian Empire. He even used a Latin alphabet, with the letters "ć, h́, ś, ź" (the same as in early drafts of Esperanto, later "ĉ, ĥ, ŝ, ĵ") and "ě" for schwa. However, he concluded there was no future for such a project, and abandoned it, dedicating himself to Esperanto as a unifying language for all humankind. Paul Wexler proposed that Esperanto was not an arbitrary pastiche of major European languages but a Latinate relexification of Yiddish, a native language of its founder. This model is generally unsupported by mainstream linguists. Zamenhof would later say that he had dreamed of a world language since he was a child. At first he considered a revival of Latin, but after learning it in school he decided it was too complicated to be a common means of international communication. When he learned English, he realised that verb conjugations were unnecessary, and that grammatical systems could be much simpler than he had expected. He still had the problem of memorising a large vocabulary, until he noticed two Russian signs labelled "Швейцарская" ("švejtsarskaja," a porter's lodge – from "швейцар" "švejtsar," a porter) and "Кондитерская" ("konditerskaja," a confectioner's shop – from "кондитер" "konditer," a confectioner). He then realised that a judicious use of affixes could greatly decrease the number of root words needed for communication. He chose to take his vocabulary from Romance and Germanic, the languages that were most widely taught in schools around the world and would therefore be recognisable to the largest number of people. Zamenhof taught an early version of the language to his high-school classmates. Then, for several years, he worked on translations and poetry to refine his creation. In 1895 he wrote, "I worked for six years perfecting and testing the language, even though it had seemed to me in 1878 that it was already completely ready." When he was ready to publish, the Czarist censors would not allow it. Stymied, he spent his time in translating works such as the Bible and Shakespeare. This enforced delay led to continued improvement. In July 1887 he published his "Unua Libro" ("First Book"), a basic introduction to the language. This was essentially the language spoken today. "Unua Libro" was published in 1887. At first the movement grew most in the Russian empire and eastern Europe, but soon spread to western Europe and beyond: to Argentina in 1889; to Canada in 1901; to Algeria, Chile, Japan, Mexico, and Peru in 1903; to Tunisia in 1904; and to Australia, the United States, Guinea, Indochina, New Zealand, Tonkin, and Uruguay in 1905. In its first years Esperanto was used mainly in publications by Zamenhof and early adopters like Antoni Grabowski, in extensive correspondence (mostly now lost), in the magazine "La Esperantisto", published from 1889 to 1895 and only occasionally in personal encounters. In 1894, under pressure from Wilhelm Trompeter, the publisher of the magazine "La Esperantisto", and some other leading users, Zamenhof reluctantly put forward a radical reform to be voted on by readers. He proposed the reduction of the alphabet to 22 letters (by eliminating the accented letters and most of their sounds), the change of the plural to "-i", the use of a positional accusative instead of the ending "-n", the removal of the distinction between adjectives and adverbs, the reduction of the number of participles from six to two, and the replacement of the table of correlatives with more Latinate words or phrases. These reforms were overwhelmingly rejected, but some were picked up in subsequent reforms (such as Ido) and criticisms of the language. In the following decade Esperanto spread into western Europe, especially France. By 1905 there were already 27 magazines being published (Auld 1988). A small international conference was held in 1904, leading to the first world congress in August 1905 in Boulogne-sur-Mer, France. There were 688 Esperanto speakers present from 20 nationalities. At this congress, Zamenhof officially resigned his leadership of the Esperanto movement, as he did not want personal prejudice against himself (or anti-Semitism) to hinder the progress of the language. He proposed a declaration on founding principles of the Esperanto movement, which the attendees of the congress endorsed. World congresses have been held every year since 1905, except during the two World Wars. The autonomous territory of Neutral Moresnet, between Belgium and Germany, had a sizable proportion of Esperanto-speakers among its small and multiethnic population. There was a proposal to make Esperanto its official language. In the early 1920s, a great opportunity seemed to arise for Esperanto when the Iranian delegation to the League of Nations proposed that it be adopted for use in international relations, following a report by Nitobe Inazō, an official delegate of League of Nations during the 13th World Congress of Esperanto in Prague. Ten delegates accepted the proposal with only one voice against, the French delegate, Gabriel Hanotaux, who employed France's Council veto privilege to squash all League attempts at recognition of Esperanto, starting on the first vote on December 18, 1920 and continuing through the next three years. Hanotaux did not like how the French language was losing its position as the international language and saw Esperanto as a threat. However, two years later the League recommended that its member states include Esperanto in their educational curricula. The French retaliated by banning all instruction in Esperanto in French schools and universities. The French Ministry Of Instruction said that acceptance of Esperanto would mean that "French and English would perish and that the literary standard of the world would be debased". Nonetheless, many people see the 1920s as the heyday of the Esperanto movement. In 1941, the Soviet Union started performing mass arrests, deportations, and killings of many Esperantists and their relatives for fear of an anti-nationalistic movement, but it was interrupted by the Nazi invasion. Hitler wrote in "Mein Kampf" that Esperanto was created as a universal language to unite the Jewish diaspora. The creation of a Jew-free National German Esperanto League was not enough to placate the Nazis. The teaching of Esperanto was not allowed in German prisoner-of-war camps during World War II. Esperantists sometimes were able to get around the ban by convincing guards that they were teaching Italian, the language of Germany's closest ally. In the early years of the Soviet Union, Esperanto was given a measure of government support, and an officially recognized Soviet Esperanto Association came into being. However, in 1937, Stalin reversed this policy and the use of Esperanto was effectively banned until 1956. While Esperanto itself was not enough cause for execution, its use was extended among Jews or trade unionists and encouraged contacts with foreigners. Fascist Italy, on the other hand, made some efforts of promoting tourism in Italy through Esperanto leaflets and appreciated the similarities of Italian and Esperanto. Portugal's right-wing governments cracked down on the language from 1936 until the Carnation Revolution of 1974. After the Spanish Civil War, Francoist Spain cracked down on the Anarchists and Catalan nationalists among whom the speaking of Esperanto had been quite widespread; but in the 1950s, the Esperanto movement was tolerated again, with Francisco Franco accepting the honorary patronage of the Madrid World Esperanto Congress. The Cold War, especially in the 1950s and 1960s, put a damper on the Esperanto movement as well, as there were fears on both sides that Esperanto could be used for enemy propaganda. However, the language experienced something of a renaissance in the 1970s and spread to new parts of the world, such as its veritable explosion in popularity in Iran in 1975. By 1991 there were enough African Esperantists to warrant a pan-African congress. The language continues to spread, although it is not officially recognised by any country, and is part of the state educational curriculum of only a few. The Declaration of Boulogne (1905) limited changes to Esperanto. That declaration stated, among other things, that the basis of the language should remain the "Fundamento de Esperanto" ("Foundation of Esperanto", a group of early works by Zamenhof), which is to be binding forever: nobody has the right to make changes to it. The declaration also permits new concepts to be expressed as the speaker sees fit, but it recommends doing so in accordance with the original style. Many Esperantists believe this declaration stabilising the language is a major reason why the Esperanto speaker community grew beyond the levels attained by other constructed languages and has developed a flourishing culture. Other constructed languages have been hindered from developing a stable speaking community by continual tinkering. Also, many developers of constructed languages have been possessive of their creation and have worked to prevent others from contributing to the language. One such ultimately disastrous case was Schleyer's Volapük. In contrast, Zamenhof declared that "Esperanto belongs to the Esperantists", and moved to the background once the language was published, allowing others to share in the early development of the language. The grammatical description in the earliest books was somewhat vague, so a consensus on usage (influenced by Zamenhof's answers to some questions) developed over time within boundaries set by the initial outline (Auld 1988). Even before the Declaration of Boulogne, the language was remarkably stable; only one set of lexical changes were made in the first year after publication, namely changing "when", "then", "never", "sometimes", "always" from "kian", "tian", "nenian", "ian", "ĉian" to "kiam", "tiam", "neniam" etc., to avoid confusion with the accusative forms of "kia" "what sort of", "tia" "that sort of", etc. Thus Esperanto achieved a stability of structure and grammar similar to that which natural languages enjoy by virtue of their native speakers and established bodies of literature. One could learn Esperanto without having it move from underfoot. Changes could and did occur in the language, but only by acquiring widespread popular support; there was no central authority making arbitrary changes, as happened with Volapük and some other languages. Modern Esperanto usage may in fact depart from that originally described in the "Fundamento", though the differences are largely semantic (involving changed meaning of words) rather than grammatical or phonological. The translation given for "I like this one", in the sample phrases in the main Esperanto article, offers a significant example. According to the "Fundamento", "Mi ŝatas ĉi tiun" would in fact have meant "I esteem this one". The traditional usage is "Tiu ĉi plaĉas al mi" (literally, "this one is pleasing to me"), which reflects the phrasing of most European languages (French "celui-ci me plaît", Spanish "éste me gusta", Russian "это мне нравится" [eto mnye nravitsya], German "Das gefällt mir", Italian "mi piace"). However, the original "Ĉi tiu plaĉas al mi" continues to be commonly used. For later changes to the language, see Modern evolution of Esperanto. Esperanto has not fragmented into regional dialects through natural language use. This may be because it is the language of daily communication for only a small minority of its speakers. However at least three other factors work against dialects, namely the centripetal force of the Fundamento, the unifying influence of the Plena Vortaro and its successors, which exemplified usage from the works of Zamenhof and leading writers, and the transnational ambitions of the speech community itself. Slang and jargon have developed to some extent, but such features interfere with universal communication – the whole point of Esperanto – and so have generally been avoided. However, in the early twentieth century numerous reform projects were proposed. Almost all of these Esperantidos were stillborn, but the very first, Ido ("Offspring"), had significant success for several years. Ido was proposed by the Delegation for the Adoption of an International Auxiliary Language in Paris in October 1907. Its main reforms were in bringing the alphabet, semantics, and some grammatical features into closer alignment with the Romance languages, as well as removal of adjectival agreement and the accusative case except when necessary. At first, a number of leading Esperantists put their support behind the Ido project, but the movement stagnated and declined, first with the accidental death of one of its main proponents and later as people proposed further changes, and the number of current speakers is estimated at between 250 and 5000. However, Ido has proven to be a rich source of Esperanto vocabulary. Some more focused reform projects, affecting only a particular feature of the language, have gained a few adherents. One of these is riism, which modifies the language to incorporate non-sexist language and gender-neutral pronouns. However, most of these projects are specific to individual nationalities (riism from English speakers, for example), and the only changes that have gained acceptance in the Esperanto community have been the minor and gradual bottom-up reforms discussed in the last section. Esperanto is credited with influencing or inspiring several later competing language projects, such as Occidental (1922) and Novial (1928). These always lagged far behind Esperanto in their popularity. By contrast, Interlingua (1951) has greatly surpassed Ido in terms of popularity. It shows little or no Esperanto influence, however.
https://en.wikipedia.org/wiki?curid=10400
Esperanto grammar Esperanto is a constructed language. It is designed to have a highly regular grammar, and as such is considered an easy language to learn. Each part of speech has a unique suffix: nouns end with "‑o"; adjectives with "‑a"; present‑tense indicative verbs with "‑as", and so on. It is possible to communicate effectively with a vocabulary of 400 to 500 "meaning words", though more exist in the language. The original vocabulary of Esperanto had around 900 meaning words, but was quickly expanded. Reference grammars include the "" () by Kálmán Kalocsay and Gaston Waringhien, and the "Plena Manlibro de Esperanta Gramatiko" () by Bertilo Wennergren. Esperanto has an agglutinative morphology, no grammatical gender, and simple verbal and nominal inflections. Verbal suffixes indicate four moods, of which the indicative has three tenses, and are derived for several aspects, but do not agree with the grammatical person or number of their subjects. Nouns and adjectives have two cases, nominative/oblique and accusative/allative, and two numbers, singular and plural; the adjectival form of personal pronouns behaves like a genitive case. Adjectives generally agree with nouns in case and number. In addition to indicating direct objects, the accusative/allative case is used with nouns, adjectives and adverbs for showing the destination of a motion, or for replacing certain prepositions; the nominative/oblique is used in all other situations. The case system allows for a flexible word order that reflects information flow and other pragmatic concerns, as in Russian, Greek, and Latin. Esperanto uses the Latin alphabet with six additional lettersĉ, ĝ, ĥ, ĵ, ŝ, and ŭand does not use the letters q, w, x, or y. Zamenhof suggested Italian as a model for Esperanto pronunciation. Esperanto has a single definite article, "la", which is invariable. It is similar to English "the". "La" is used: The article is also used for inalienable possession of body parts and kin terms, where English would use a possessive adjective: The article "la", like the demonstrative adjective "tiu" (this, that), occurs at the beginning of the noun phrase. There is no grammatically required indefinite article: "homo" means either "human being" or "a human being", depending on the context, and similarly the plural "homoj" means "human beings" or "some human beings". The words "iu" and "unu" (or their plurals "iuj" and "unuj") may be used somewhat like indefinite articles, but they're closer in meaning to "some" and "a certain" than to English "a". This use of "unu" corresponds to English "a" when the "a" indicates a specific individual. The suffixes "‑o", "‑a", "‑e", and "‑i" indicate that a word is a noun, adjective, adverb, and infinitive verb, respectively. Many new words can be derived simply by changing these suffixes. Derivations from the word "vidi" (to see) are "vida" (visual), "vide" (visually), and "vido" (vision). Each root word has an inherent part of speech: nominal, adjectival, verbal, or adverbial. These must be memorized explicitly and affect the use of the part-of-speech suffixes. With an adjectival or verbal root, the nominal suffix "‑o" indicates an abstraction: "parolo" (an act of speech, one's word) from the verbal root "paroli" (to speak); "belo" (beauty) from the adjectival root "bela" (beautiful); whereas with a noun, the nominal suffix simply indicates the noun. Nominal or verbal roots may likewise be modified with the adjectival suffix "‑a": "reĝa" (royal), from the nominal root "reĝo" (a king); "parola" (spoken). The various verbal endings mean "to be [__]" when added to an adjectival root: "beli" (to be beautiful); and with a nominal root they mean "to act as" the noun, "to use" the noun, etc., depending on the semantics of the root: "reĝi" (to reign). There are relatively few adverbial roots, so most words ending in "-e" are derived: "bele" (beautifully). Often with a nominal or verbal root, the English equivalent is a prepositional phrase: "parole" (by speech, orally); "vide" (by sight, visually); "reĝe" (like a king, royally). The meanings of part-of-speech affixes depend on the inherent part of speech of the root they are applied to. For example, "brosi" (to brush) is based on a nominal root (and therefore listed in modern dictionaries under the entry "broso)," whereas "kombi" (to comb) is based on a verbal root (and therefore listed under "kombi)." Change the suffix to "-o," and the similar meanings of "brosi" and "kombi" diverge: "broso" is a brush, the name of an instrument, whereas "kombo" is a combing, the name of an action. That is, changing verbal "kombi" (to comb) to a noun simply creates the name for the action; for the name of the tool, the suffix "-ilo" is used, which derives words for instruments from verbal roots: "kombilo" (a comb). On the other hand, changing the nominal root "broso" (a brush) to a verb gives the action associated with that noun, "brosi" (to brush). For the name of the action, the suffix "-ado" will change a derived verb back to a noun: "brosado" (a brushing). Similarly, an abstraction of a nominal root (changing it to an adjective and then back to a noun) requires the suffix "-eco," as in "infaneco" (childhood), but an abstraction of an adjectival or verbal root merely requires the nominal "-o: belo" (beauty). Nevertheless, redundantly affixed forms such as "beleco" are acceptable and widely used. A limited number of basic adverbs do not end with "-e," but with an undefined part-of-speech ending "-aŭ". Not all words ending in "-aŭ" are adverbs, and most of the adverbs that end in "-aŭ" have other functions, such as "hodiaŭ" "today" [noun or adverb] or "ankoraŭ" "yet, still" [conjunction or adverb]. About a dozen other adverbs are bare roots, such as "nun" "now", "tro" "too, too much", not counting the adverbs among the correlatives. (See special Esperanto adverbs.) Other parts of speech occur as bare roots, without special suffixes. These are the prepositions "(al" "to"), conjunctions "(kaj" "and"), interjections "(ho" "oh"), numerals "(du" "two"), and pronouns "(mi" "I"—The final "-i" found on pronouns is not a suffix, but part of the root). There are also several grammatical "particles" that fit neatly into no category, and which must generally precede the words they modify, such as "ne" (not), "ankaŭ" (also), "nur" (only), "eĉ" (even). A suffix "-j" following the noun or adjective suffixes "-o" or "-a" makes a word plural. Without this suffix, a countable noun is understood to be singular. Direct objects take an accusative case suffix "-n," which goes after any plural suffix. (The resulting sequence "-ojn" rhymes with English "coin," and "-ajn" rhymes with "fine.)" Adjectives agree with nouns. That is, they are plural if the nouns that they modify are plural, and accusative if the nouns that they modify are accusative. Compare "bona tago; bonaj tagoj; bonan tagon; bonajn tagojn" (good day/days). This requirement allows for free word orders of adjective-noun and noun-adjective, even when two noun phrases are adjacent in subject–object–verb or verb–subject–object clauses: Agreement clarifies the syntax in other ways also. Adjectives take the plural suffix when they modify more than one noun, even if those nouns are all singular: A predicative adjective does not take the accusative case suffix even when the noun that it modifies does: There are three types of pronouns in Esperanto: personal "(vi" "you"), demonstrative "(tio" "that", "iu" "someone"), and relative/interrogative "(kio" "what"). According to the fifth rule of the "Fundamento de Esperanto": The Esperanto personal pronoun system is similar to that of English, but with the addition of a reflexive pronoun. Personal pronouns take the accusative suffix "-n" as nouns do: "min" (me), "lin" (him), "ŝin" (her). Possessive adjectives are formed with the adjectival suffix "-a: mia" (my), "ĝia" (its), "nia" (our). These agree with their noun like any other adjective: "ni salutis liajn amikojn" (we greeted his friends). Esperanto does not have separate forms for the possessive pronouns; this sense is generally (though not always) indicated with the definite article: "la mia" (mine). The reflexive pronoun is used, in non-subject phrases only, to refer back to the subject, usually only in the third and indefinite persons: The indefinite pronoun is used when making general statements, and is often used where English would have the subject "it" with a passive verb, With "impersonal" verbs where there is actually no item or being that is doing an action, no pronoun is used: The rain is falling by itself, therefore the subject pronoun is omitted. "Ĝi" ("it"; third person neutral) is mostly used with items that have physical bodies. Zamenhof proposed that this pronoun can also be used as an epicene (gender-neutral) third-person singular pronoun, meaning for use when the gender of an individual is unknown or for when the speaker simply doesn't wish to clarify the gender. However, this proposal is not generally accepted. In popular usage it's usually only used when referring to children: When speaking of adults or people in general, in popular usage it is much more common for the demonstrative adjective and pronoun "tiu" ("that thing or person that is already known to the listener") to be used in such situations. This mirrors languages such as Japanese, but it's not a method that can always be used. For example, in the sentence the word "tiu" would be understood as referring to someone other than the person speaking (like English pronouns "this" or "that" but also referring to people), and so cannot be used in place of "ĝi", "li" or "ŝi." The demonstrative and relative pronouns form part of the correlative system, and are described in that article. The pronouns are the forms ending in "-o" (simple pronouns) and "-u" (adjectival pronouns). Their accusative case is formed in "-n," but the genitive case ends in "-es," which is the same for singular and plural and does not take accusative marking. Compare the nominative phases "lia domo" (his house) and "ties domo" (that one's house, those ones' house) with the plural "liaj domoj (his houses) and "ties domoj (that one's houses, those ones' houses), and with the accusative genitive "lian domon" and "ties domon." "Ri" and "ria" are unofficial pronouns, that some Esperanto speakers use to refer to a single person regardless of their gender identity. Some Esperanto speakers use it only to refer to a single person whose gender identity is unknown or non-binary or to refer to unspecified person, while other Esperanto speakers use it also to refer to a person whose gender identity is purely female or purely male. E. g. "Ĉiu amas ordinare personon, kiu estas simila al ri" ("Everybody usually loves a person who is similar to them"); "Alekso kaj ria amiko estas ambaŭ neduumaj" ("Alex and their friend are both non-binary"). The use of the Esperanto pronoun "ri" is (almost) the same as the use of the English singular "they" and the Swedish pronoun "hen". Although Esperanto word order is fairly free, prepositions must come at the beginning of a noun phrase. Whereas in languages such as German, prepositions may require that a noun be in various cases (accusative, dative, "and so on)," in Esperanto all prepositions govern the nominative: "por Johano" (for John). The only exception is when there are two or more prepositions and one is "replaced" by the accusative. Prepositions should be used with a definite meaning. When no one preposition is clearly correct, the indefinite preposition "je" should be used: Alternatively, the accusative may be used without a preposition: Note that although "la trian" (the third) is in the accusative, "de majo" (of May) is still a prepositional phrase, and so the noun "majo" remains in the nominative case. A frequent use of the accusative is in place of "al" (to) to indicate the direction or goal of motion (allative construction). It is especially common when there would otherwise be a double preposition: The accusative/allative may stand in for other prepositions also, especially when they have vague meanings that add not much to the clause. Adverbs, with or without the case suffix, are frequently used instead of prepositional phrases: Both "por" and "pro" often translate English 'for'. However, they distinguish "for a goal" (looking forward in time, or causing: "por") and "for a cause" (looking back in time, or being caused by: "pro"): To vote "por" your friend means to cast a ballot with their name on it, whereas to vote "pro" your friend would mean to vote in their place or as they asked you to. The preposition most distinct from English usage is perhaps "de", which corresponds to English "of, from, off," and "(done) by": However, English "of" corresponds to several Esperanto prepositions also: "de, el" (out of, made of), and "da" (quantity of, unity of form and contents): The last of these, "da", is semantically Slavic and is difficult for Western Europeans, to the extent that even many Esperanto dictionaries and grammars define it incorrectly. Occasionally a new preposition is coined. Because a bare root may indicate a preposition or interjection, removing the grammatical suffix from another part of speech can be used to derive a preposition or interjection. For example, from "fari" (to do, to make) we get the preposition "far" (done by). All verbal inflection is regular. There are three tenses, all of which are in the indicative mood. The other moods are the infinitive, conditional, and jussive. No aspectual distinctions are required by the grammar, but derivational expressions of Aktionsart are common. Verbs do not change form according to their subject. "I am, we are," and "he is" are simply "mi estas, ni estas," and "li estas," respectively. Impersonal subjects are not used: "pluvas" (it is raining); "estas muso en la domo" (there's a mouse in the house). Most verbs are inherently transitive or intransitive. As with the inherent part of speech of a root, this is not apparent from the shape of the verb and must simply be memorized. Transitivity is changed with the suffixes "-ig-" (the transitivizer/causative) and "-iĝ-" (the intransitivizer/middle voice): The tenses have characteristic vowels. "A" indicates the present tense, "i" the past, and "o" the future. The verbal forms may be illustrated with the root "esper-" (hope): A verb can be made emphatic with the particle "ja" (indeed): "mi ja esperas" (I do hope), "mi ja esperis" (I did hope). As in English, Esperanto present tense may be used for generic statements such as "birds fly" ("la birdoj flugas"). The Esperanto future is a true tense, used whenever future time is meant. For example, in English "(I'll give it to you) when I see you" the verb "see" is in the present tense despite the time being in the future; in Esperanto, future tense is required: "(Mi donos ĝin al vi) kiam mi vidos vin". Esperanto tense is relative. This differs from English absolute tense, where the tense is past, present, or future of the moment of speaking: In Esperanto, the tense of a subordinate verb is instead anterior or posterior to the time of the main verb. For example, "John said that he would go" is in Esperanto "Johano diris, ke li iros" (lit., "John said that he will go"); this does not mean that he will go at some point in the future from now (as "John said that he will go" means in English), but that at the time he said this, his going was still in the future. The conditional mood is used for such expressions as "se mi povus, mi irus" (if I could, I would go) and "se mi estus vi, mi irus" (if I were you, I'd go). The jussive mood, is used for wishing and requesting, and serves as the imperative and subjunctive: The verb "esti" (to be) is both the copula ("X is Y") and the existential ("there is") verb. As a copula linking two noun phrases, it causes neither to take the accusative case. Therefore, unlike the situation with other verbs, word order with "esti" can be semantically important: compare "hundoj estas personoj" (dogs are people) and "personoj estas hundoj" (people are dogs). Participles are verbal derivatives. In Esperanto, there are six forms: times The participles represent aspect by retaining the vowel of the related verbal tense: i, a, o. In addition to carrying aspect, participles are the principal means of representing voice, with either nt or t following the vowel (see next section). The basic principle of the participles may be illustrated with the verb "fali" (to fall). Picture a cartoon character running off a cliff. Before the character falls, they are "falonta" (about to fall). As they drop, they are "falanta" (falling). After they hit the ground, they are "falinta" (fallen). Active and passive pairs can be illustrated with the transitive verb "haki" (to chop). Picture a woodsman approaching a tree with an axe, intending to chop it down. He is "hakonta" (about to chop) and the tree is "hakota" (about to be chopped). While swinging the axe, he is "hakanta" (chopping) and the tree "hakata" (being chopped). After the tree has fallen, he is "hakinta" (having chopped) and the tree "hakita" (chopped). Adjectival participles agree with nouns in number and case, just as other adjectives do: Compound tenses are formed with the adjectival participles plus "esti" (to be) as the auxiliary verb. The participle reflects aspect and voice, while the verb carries tense: These are not used as often as their English equivalents. For "I "am" go"ing" to the store", you would normally use the simple present "mi iras" in Esperanto. The tense and mood of "esti" can be changed in these compound tenses: Although such periphrastic constructions are familiar to speakers of most European languages, the option of contracting ["esti" + adjective] into a verb is often seen for adjectival participles: The active synthetic forms are: Infinitive and jussive forms are also found. There is a parallel passive paradigm. A few of these forms, notably "-intus" and "-atas", entered common usage, but most of them are very rare because they are difficult to understand. Participles may be turned into adverbs or nouns by replacing the adjectival suffix "-a" with "-e" or "-o." This means that, in Esperanto, some nouns may be inflected for tense. A nominal participle indicates "one who participates" in the action specified by the verbal root. For example, "esperinto" is a "hoper" (past tense), or "one who had been hoping." Adverbial participles are used with subjectless clauses: Occasionally, the participle paradigm will be extended to include conditional participles, with the vowel "u (-unt-, -ut-)." If, for example, in our tree-chopping example, the woodsman found that the tree had been spiked and so couldn't be cut down after all, he would be "hakunta" and the tree "hakuta" (he, the one "who would chop", and the tree, the one that "would be chopped"). This can also be illustrated with the verb "prezidi" (to preside). Just after the recount of the 2000 United States presidential election: The tense-neutral word "prezidento" is officially a separate root, not a derivative of the verb "prezidi." A statement is made negative by using "ne" or one of the negative "(neni-)" correlatives. Ordinarily, only one negative word is allowed per clause: Two negatives within a clause cancel each other out, with the result being a positive sentence. The word "ne" comes before the word it negates: The latter will frequently be reordered as "ne tion mi skribis" depending on the flow of information. "Wh" questions are asked with one of the interrogative/relative "(ki-") correlatives. They are commonly placed at the beginning of the sentence, but different word orders are allowed for stress: Yes/no questions are marked with the conjunction "ĉu" (whether): Such questions can be answered "jes" (yes) or "ne" (no) in the European fashion of aligning with the polarity of the answer, or "ĝuste" (correct) or "malĝuste" (incorrect) in the Japanese fashion of aligning with the polarity of the question: Note that Esperanto questions may have the same word order as statements. Basic Esperanto conjunctions are "kaj" (both/and), "kun" (with),"aŭ" (either/or), "nek" (neither/nor), "se" (if), "ĉu" (whether/or), "sed" (but), "anstataŭ" (instead of), "krom" (besides, except for), "kiel" (like, as), "ke" (that). Like prepositions, they precede the phrase or clause they modify: However, unlike prepositions, they allow the accusative case, as in the following example from Don Harlow: Interjections may be derived from bare affixes or roots: "ek!" (get going!), from the perfective prefix; "um" (um, er), from the indefinite/undefined suffix; "fek!" (shit!), from "feki" (to defecate). Esperanto derivational morphology uses a large number of lexical and grammatical affixes (prefixes and suffixes). These, along with compounding, decrease the memory load of the language, as they allow for the expansion of a relatively small number of basic roots into a large vocabulary. For example, the Esperanto root "vid-" (see) regularly corresponds to several dozen English words: "see (saw, seen), sight, blind, vision, visual, visible, nonvisual, invisible, unsightly, glance, view, vista, panorama, observant" etc., though there are also separate Esperanto roots for a couple of these concepts. The cardinal numerals are: These are grammatically numerals, not nouns, and as such do not take the accusative case suffix. However, "unu" (and only "unu") is sometimes used adjectivally or demonstratively, meaning "a certain", and in such cases it may take the plural affix "-j," just as the demonstrative pronoun "tiu" does: In such use "unu" is irregular in that it doesn't take the accusative affix "-n" in the singular, but does in the plural: but Additionally, when counting off, the final "u" of "unu" may be dropped, as if it were a part-of-speech suffix: At numbers beyond the thousands, the international roots "miliono" (million) and "miliardo" (milliard) are used. Beyond this there are two systems: A "billion" in most English-speaking countries is different from a "billion" in most other countries (109 "vs." 1012 respectively; that is, a thousand million "vs." a million million). The international root "biliono" is likewise ambiguous in Esperanto, and is deprecated for this reason. An unambiguous system based on adding the Esperanto suffix "-iliono" to numerals is generally used instead, sometimes supplemented by a second suffix "-iliardo:" Note that these forms are grammatically nouns, not numerals, and therefore cannot modify a noun directly: "mil homojn" (a thousand people [accusative]) but "milionon da homoj" (a million people [accusative]). Numerals are written together as one word when their values are multiplied, and separately when their values are added ("dudek" 20, "dek du" 12, "dudek du" 22). Ordinals are formed with the adjectival suffix "-a," quantities with the nominal suffix "-o," multiples with "-obl-," fractions with "‑on‑", collectives with "‑op‑", and repetitions with the root "‑foj‑". The particle "po" is used to mark distributive numbers, that is, the idea of distributing a certain number of items to each member of a group. Consequently, the logogram @ is not used (except in email addresses, of course): Note that particle "po" forms a phrase with the numeral "tri" and is not a preposition for the noun phrase "tri pomojn," so it does not prevent a grammatical object from taking the accusative case. Comparisons are made with the adverbial correlatives "tiel ... kiel" (as ... as), the adverbial roots "pli" (more) and "plej" (most), the antonym prefix "mal-," and the preposition "ol" (than): Implied comparisons are made with "tre" (very) and "tro" (too [much]). Phrases like "The more people, the smaller the portions" and "All the better!" are translated using "ju" and "des" in place of "the": Esperanto has a fairly flexible word order. However, word order does play a role in Esperanto grammar, even if a much lesser role than it does in English. For example, the negative particle "ne" generally comes before the element being negated; negating the verb has the effect of negating the entire clause (or rather, there is ambiguity between negating the verb alone and negating the clause): However, when the entire clause is negated, the "ne" may be left till last: The last order reflects a typical topic–comment (or theme–rheme) order: Known information, the topic under discussion, is introduced first, and what one has to say about it follows. (I went not: As for my going, there was none.) For example, yet another order, "ne iris mi", would suggest that the possibility of not having gone was under discussion, and "mi" is given as an example of one who did not go. Compare: Within a noun phrase, either the order "adjective–noun" or "noun–adjective" may occur, though the former is somewhat more common. Less flexibility occurs with numerals and demonstratives, with "numeral–noun" and "demonstrative–noun" being the norm, as in English. Adjective–noun order is much freer. With simple adjectives, adjective–noun order predominates, especially if the noun is long or complex. However, a long or complex adjective typically comes after the noun, in some cases parallel to structures in English, as in the second example below: Adjectives also normally occur after correlative nouns. Again, this is one of the situations where adjectives come after nouns in English: Changing the word order here can change the meaning, at least with the correlative "nenio" 'nothing': With multiple words in a phrase, the order is typically demonstrative/pronoun–numeral–(adjective/noun): However, the article "la" precedes a noun phrase: In prepositional phrases, the preposition is "required" to come at the front of the noun phrase (that is, even before the article "la"), though it is commonly replaced by turning the noun into an adverb: Constituent order "within" a clause is generally free, apart from copular clauses. The default order is subject–verb–object, though any order may occur, with subject and object distinguished by case, and other constituents distinguished by prepositions: The expectation of a topic–comment (theme–rheme) order apply here, so the context will influence word order: in "la katon ĉasis la hundo", the cat is the topic of the conversation, and the dog is the news; in "la hundo la katon ĉasis", the dog is the topic of the conversation, and it is the action of chasing that is the news; and in "ĉasis la hundo la katon", the action of chasing is already the topic of discussion. Context is required to tell whether means the dog chased a cat which was in the garden, or there, in the garden, the dog chased the cat. These may be disambiguated with and Of course, if it chases the cat in"to" the garden, the case of 'garden' would change: Within copulative clauses, however, there are restrictions. Copulas are words such as "esti" 'be', "iĝi" 'become', "resti" 'remain', and "ŝajni" 'seem', for which neither noun phrase takes the accusative case. In such cases only two orders are generally found: noun-copula-predicate and, much less commonly, predicate-copula-noun. Generally, if a characteristic of the noun is being described, the choice between the two orders is not important: However, "la vento sovaĝa estas" is unclear, at least in writing, as it could be interpreted as 'the wild wind is', leaving the reader to ask, 'is what?'. In the sentence above, "la hundo ĉasis la katon, kiu estis en la ĝardeno" 'the dog chased the cat, which was in the garden', the relative pronoun "kiu" 'which' is restricted to a position "after" the noun 'cat'. In general, relative clauses and attributive prepositional phrases follow the noun they modify. Attributive prepositional phrases, which are dependent on nouns, include genitives ("la libro de Johano" 'John's book') as well as "la kato en la ĝardeno" 'the cat in the garden' in the example above. Their order cannot be reversed: neither "*la de Johano libro" nor "*la en la ĝardeno kato" is possible. This behavior is more restrictive than prepositional phrases which are dependent on verbs, and which can be moved around: both "ĉasis en la ĝardeno" and "en la ĝardeno ĉasis" are acceptable for 'chased in the garden'. Relative clauses are similar, in that they are attributive and are subject to the same word-order constraint, except that rather than being linked by a preposition, the two elements are linked by a relative pronoun such as "kiu" 'which': Note that the noun and its adjacent relative pronoun do not agree in case. Rather, their cases depend on their relationships with their respective verbs. However, they do agree in number: Other word orders are possible, as long as the relative pronoun remains adjacent to the noun it depends on: Coordinate clauses allow flexible word order, but tend to be iconic. For example, in the inference is that the cat fled after the dog started to chase it, not that the dog chased a cat which was already fleeing. For the latter reading, the clause order would be reversed: This distinction is lost in subordinate clauses such as the relative clauses in the previous section: In written English, a comma disambiguates the two readings, but both typically have a comma in Esperanto. Non-relative subordinate clauses are similarly restricted. They follow the conjunction "ke" 'that', as in, Esperanto's vocabulary, syntax, and semantics derive predominantly from Indo-European national languages. Roots are typically Romance or Germanic in origin. The semantics shows a significant Slavic influence. It is often claimed that there are elements of the grammar which are not found in these language families. Frequently mentioned is Esperanto's agglutinative morphology based on invariant morphemes, and the subsequent lack of ablaut (internal inflection of its roots), which Zamenhof himself thought would prove alien to non-Indo-European language speakers. Ablaut is an element of all the source languages; an English example is "song sing sang sung." However, the majority of words in all Indo-European languages inflect without ablaut, as "cat, cats" and "walk, walked" do in English. (This is the so-called strong–weak dichotomy.) Historically, many Indo-European languages have expanded the range of their 'weak' inflections, and Esperanto has merely taken this development closer to its logical conclusion, with the only remaining ablaut being frozen in a few sets of semantically related roots such as "pli", "plej", "plu" (more, most, further), "tre", "tro" (very, too much), and in the verbal morphemes "‑as", "‑anta", "‑ata"; "‑is", "‑inta", "‑ita"; "‑os", "‑onta", "‑ota"; and "‑us". Other features often cited as being nonstandard for an Indo-European language, such as the dedicated suffixes for different parts of speech, or the "-o" suffix for singular nouns, actually do occur in Indo‑European languages such as Russian. More pertinent is the accusative plural in "-jn". Esperanto is superficially similar to the non‑Indo‑European Hungarian and Turkish languages—that is, it is similar in its mechanics, but not in use. None of these proposed "non-European" elements of the original Esperanto proposal were actually taken from non-European or non-Indo-European languages, and any similarities with those languages are accidental. East Asian languages may have had some influence on the development of Esperanto grammar after its creation. The principally cited candidate is the replacement of predicate adjectives with verbs, such as "la ĉielo bluas" (the sky is blue) for "la ĉielo estas blua" and "mia filino belu!" (may my daughter be beautiful!) for the "mia filino estu bela!" mentioned above. The Pater noster, from the first Esperanto publication in 1887, illustrates many of the grammatical points presented above: The morphologically complex words (see Esperanto word formation) are:
https://en.wikipedia.org/wiki?curid=10402
Esperanto culture Esperanto culture refers to the shared cultural experience of the Esperantujo, or Esperanto-speaking community. Despite being a constructed language, Esperanto has a history dating back to the late 19th century, and shared cultural social mores have developed among its speakers. Some of these can be traced back to the initial ideas of the language's creator, Ludwig Zamenhof, including the theory that a global second language would foster international communication. Others have developed over time, as the language has allowed different national and linguistic cultures to blend together. Esperanto culture also includes art, literature, and music, as well as international celebrations and cultural exchanges such as the Pasporta Servo. Native Esperanto speakers are people who have acquired Esperanto as one of their native languages. As of 1996, there were 350 or so attested cases of families with native Esperanto speakers. Estimates from associations indicate that there are currently around 1,000 Esperanto-speaking families, involving perhaps 2,000 children. In all known cases, speakers are natively bilingual, or multilingual, raised in both Esperanto and either the local national language or the native language of their parents. In all but a handful of cases, it was the father who used Esperanto with the child. In the majority of such families, the parents had the same native language, though in many the parents had different native languages, and only Esperanto in common. Esperanto speakers create a makeshift café (whether in a rented space or someone's home), using Esperanto coins or voucher-like items as well as real money to pay for food and drink. Live music, poetry reading, or literature reading are usual activities. This custom arose in 1995 in order to contrast with the more usual custom of after-convention partying at a bar. Esperanto has had an influence on certain religious traditions (Oomoto, Bahá'í Faith, etc., see Esperanto and religion). While some Esperantists subscribe to these beliefs, they are not necessarily common, and are certainly not required nor encouraged by any Esperanto groups. Books that are translated to Esperanto are not usually internationally famous books, because everyone can already read those in another language that they know. For example, Natsume Sōseki's "Kokoro" does not exist but several Japanese crime novels, and several Icelandic novels, that have never been translated to English (or any other language) have been translated to Esperanto. One reason for this is that people are actually translating their favourite stories instead of the famous stories, and another is that it's simply cheaper and easier to get the rights to translate a small-time book compared to a famous one. The first Harry Potter book, for example, was translated and the translator enquired about how to purchase translating rights so the book could be published, but J.K. Rowling refused to allow it to be published in Esperanto (despite Harry Potter being one of the most-translated books in the whole world). In lieu of physical books, the translation now exists as a free download on the internet. As Esperanto speakers were persecuted throughout WWII (Esperanto was banned in Germany and the Soviet Union), and even today might be mocked by journalists and the average person, oppression and acceptance have become strong themes in Esperanto writing and conversation. There are over 25,000 Esperanto books (originals and translations) as well as over a hundred regularly distributed Esperanto magazines. This is despite that Esperanto has only existed for around 100 years. In comparison, the entire literature of Iceland (a country created in the 900s, and with a population of around 320,000 people) totals fewer than 50,000 books. Many speakers travel the world using the Pasporta Servo which is a free couchsurfing and homestay service combined, meaning that their trips are possible because they don't have to pay for lodging while at the same time they stay with people who speak a language they know fluently. Many people have commented that this is also a useful tool for actual immigrants, as Esperanto speakers are normally much friendlier and more willing to see the immigrant as a "human" compared to the normal natives of the country. For example, it's suggested that an English speaker in Japan should make friends with Esperanto speakers instead, because the Esperanto speakers won't "simply use the friendship as a way to get free English lessons". Esperanto was originally a language that one had to learn entirely through books, and even today most people live apart from each other and converse through the internet, so writing and reading are a big part of Esperanto culture. Most people have created or translated some sort of written work whether fiction or nonfiction, published or available to read online for free. Penpals have been popular since Esperanto's earliest days, as Esperanto was originally advertised as a language where you could "send a letter with a message, short list of grammar rules and a dictionary to a complete stranger, and they'll be able to look up the words and write a coherent reply back". Many people did indeed do this in order to recruit more Esperanto speakers. At the time, in the early 1900s, there was no major world language that could be used "anywhere" and it was difficult to get accurate information about foreign countries. On top of that, things like stamp collecting were popular hobbies for children. In the modern day, most Esperanto speakers talk to each other through the internet — which is just the modern version of a penpal. "Monato" ("month") is a general news magazine "like a genuinely international "Time" or "Newsweek"", written by local correspondents. A magazine for the blind, "Aŭroro", has been published since 1920 and in general, Esperanto hosts the largest Braille publications in the world — starting in the early 1900s Esperanto was taught in schools of the blind in Europe, and that is where the trend started. "Esperanto" is the magazine used by the World Esperanto Association to inform its members about everything happening in the Esperanto community. There are many more magazines created by individual Esperanto clubs from towns in places such as from Japan and China. Many people wear their country's traditional clothing to Esperanto conventions, whether or not they would ever wear it in their own country. Swedish people, for example, who usually never wear their traditional clothing in their own country, may still wear traditional clothing for any meeting involving Esperanto speakers. Every year, the World Congress of Esperanto (), which is held in different countries around the world according to year (though it mostly takes place in Europe). Each convention draws in an average of 1500–3000 attendees, and the best-attended conferences are those held in Central or Eastern Europe (generally meaning Poland, Hungary etc.), as Esperanto is an option for fulfilling mandatory foreign-language requirements in Hungarian schools, and the creator of Esperanto came from Poland (see statistics at World Congress of Esperanto). Esperanto music is usually done in the traditional style of a person's country, but "international" music (American pop music, rap music etc.) also exists. Many famous songs are translated to Esperanto as well, for example "La vie en rose" and "En el frente de Gandesa" (the links are to the Esperanto versions of the songs on YouTube). There are currently radio broadcasts from China Radio International, Melbourne Ethnic Community Radio, Radio Habana Cuba, Radio Audizioni Italiane (Rai), Radio Polonia, Radio F.R.E.I. and Radio Vatican. Many more people have personal podcasts and vlogs. In 1964, Jacques-Louis Mahé produced the first full-length feature film in Esperanto, entitled "Angoroj". This was followed in 1965 by the first American Esperanto-production: "Incubus", starring William Shatner. Incubus is, however, seen as a common, joke way of first introducing a person to Esperanto as none of the actors even knew how to pronounce Esperanto in the first place, the dialogue is strange and bad due to the scriptwriter not getting a second opinion before the filming was done, and the plot is confusing in general. Internacia Televido, an internet television channel, began broadcasting in November 2005. Australia is the hotspot of much of the organization behind Esperanto television. Several short films have been produced, and at times plays have been recorded "for television". , the Esperanto-language Wikipedia 14 films and 3 short films. In 2011, Academy Award-nominated director Sam Green ("The Weather Underground"), released a new documentary about Esperanto titled The Universal Language (La Universala Lingvo.) This 30-minute film traces the history of Esperanto. It's known for having extremely good camera quality and filming sense, as well as being a good "absolute introduction" to what Esperanto is, but is criticized for being too short. Many more films, cartoons and documentaries that aren't Esperanto originals are simply subtitled in Esperanto and put up on YouTube. Some fan-dubs exist, especially of Disney songs and short scenes, and the quality mirrors that of what would come from any small country or a country "unused to technology" (such as Greenland, Kazakhstan, etc.). As Esperanto speakers are from all over the world, and families whose children speak Esperanto natively usually have parents from two vastly different countries, recipes incorporating elements from different countries are naturally born. Traditional foods are also enjoyed in settings where a native wouldn't normally mix or eat them. One cookbook is "Internacie kuiri" “Cooking Internationally” by Maria Becker-Meisberger, published by FEL (Flemish Esperanto League), Antwerp 1989, . Another is "Manĝoj el sanigaj plantoj" “Meals from Healthy Vegetable Dishes” by Zlata Nanić, published by BIO-ZRNO, Zagreb 2002, . Some Esperanto periodicals, such as "MONATO" include recipes from time to time. On December 15 (L. L. Zamenhof's birthday), Esperanto speakers around the world celebrate Zamenhof Day, sometimes called Book Day. It's a common goal to have a book written in Esperanto published on or by that day, as Zamenhof was a strong advocate of the idea that in order to spread Esperanto around the world, its speakers need to create a large body of literature. The poem "La Espero" is the Esperanto anthem, and most Esperanto speakers know it by heart. It's often sung at conventions. Whether or not one enjoys the lyrics, the song is in general something that ties all Esperanto speakers together — as it's been around since Esperanto's early days, is known at least in tune to every Esperanto speaker, and is a general tradition.
https://en.wikipedia.org/wiki?curid=10404
Emotion Emotions are biological states associated with the nervous system brought on by neurophysiological changes variously associated with thoughts, feelings, behavioural responses, and a degree of pleasure or displeasure. There is currently no scientific consensus on a definition. Emotions are often intertwined with mood, temperament, personality, disposition, creativity and motivation. Research on emotion has increased significantly over the past two decades with many fields contributing including psychology, neuroscience, affective neuroscience, endocrinology, medicine, history, sociology of emotions, and computer science. The numerous theories that attempt to explain the origin, neurobiology, experience, and function of emotions have only fostered more intense research on this topic. Current areas of research in the concept of emotion include the development of materials that stimulate and elicit emotion. In addition, PET scans and fMRI scans help study the affective picture processes in the brain. From a purely mechanistic perspective, "Emotions can be defined as a positive or negative experience that is associated with a particular pattern of physiological activity." Emotions produce different physiological, behavioral and cognitive changes. The original role of emotions was to motivate adaptive behaviors that in the past would have contributed to the passing on of genes through survival, reproduction, and kin selection. In some theories, cognition is an important aspect of emotion. For those who act primarily on emotions, they may assume that they are not thinking, but mental processes involving cognition are still essential, particularly in the interpretation of events. For example, the realization of our believing that we are in a dangerous situation and the subsequent arousal of our body's nervous system (rapid heartbeat and breathing, sweating, muscle tension) is integral to the experience of our feeling afraid. Other theories, however, claim that emotion is separate from and can precede cognition. Consciously experiencing an emotion is exhibiting a mental representation of that emotion from a past or hypothetical experience, which is linked back to a content state of pleasure or displeasure. The content states are established by verbal explanations of experiences, describing an internal state. Emotions are complex. According to some theories, they are states of feeling that result in physical and psychological changes that influence our behavior. The physiology of emotion is closely linked to arousal of the nervous system with various states and strengths of arousal relating, apparently, to particular emotions. Emotion is also linked to behavioral tendency. Extroverted people are more likely to be social and express their emotions, while introverted people are more likely to be more socially withdrawn and conceal their emotions. Emotion is often the driving force behind motivation, positive or negative. According to other theories, emotions are not causal forces but simply syndromes of components, which might include motivation, feeling, behavior, and physiological changes, but no one of these components is the emotion. Nor is the emotion an entity that causes these components. Emotions involve different components, such as subjective experience, cognitive processes, expressive behavior, psychophysiological changes, and instrumental behavior. At one time, academics attempted to identify the emotion with one of the components: William James with a subjective experience, behaviorists with instrumental behavior, psychophysiologists with physiological changes, and so on. More recently, emotion is said to consist of all the components. The different components of emotion are categorized somewhat differently depending on the academic discipline. In psychology and philosophy, emotion typically includes a subjective, conscious experience characterized primarily by psychophysiological expressions, biological reactions, and mental states. A similar multicomponential description of emotion is found in sociology. For example, Peggy Thoits described emotions as involving physiological components, cultural or emotional labels (anger, surprise, etc.), expressive body actions, and the appraisal of situations and contexts. The word "emotion" dates back to 1579, when it was adapted from the French word "émouvoir", which means "to stir up". The term emotion was introduced into academic discussion as a catch-all term to passions, sentiments and affections. The word emotion was coined in the early 1800s by Thomas Brown and it is around the 1830s that the modern concept of emotion first emerged for English Language. "No one felt emotions before about 1830. Instead they felt other things - "passions", "accidents of the soul", "moral sentiments" - and explained them very differently from how we understand emotions today." Some cross-cultural studies indicate that the categorization of "emotion" and classification of basic emotions such as "anger" and "sadness" are not universal and that the boundaries and domains of these concepts are categorized differently by all cultures. However, others argue that there are some universal bases of emotions (see Section 6.1). In psychiatry and psychology, an inability to express or perceive emotion is sometimes referred to as alexithymia. The Oxford Dictionaries definition of emotion is "A strong feeling deriving from one's circumstances, mood, or relationships with others." Emotions are responses to significant internal and external events. Emotions can be occurrences (e.g., panic) or dispositions (e.g., hostility), and short-lived (e.g., anger) or long-lived (e.g., grief). Psychotherapist Michael C. Graham describes all emotions as existing on a continuum of intensity. Thus fear might range from mild concern to terror or shame might range from simple embarrassment to toxic shame. Emotions have been described as consisting of a coordinated set of responses, which may include verbal, physiological, behavioral, and neural mechanisms. Emotions have been categorized, with some relationships existing between emotions and some direct opposites existing. Graham differentiates emotions as functional or dysfunctional and argues all functional emotions have benefits. In some uses of the word, emotions are intense feelings that are directed at someone or something. On the other hand, emotion can be used to refer to states that are mild (as in annoyed or content) and to states that are not directed at anything (as in anxiety and depression). One line of research looks at the meaning of the word emotion in everyday language and finds that this usage is rather different from that in academic discourse. In practical terms, Joseph LeDoux has defined emotions as the result of a cognitive and conscious process which occurs in response to a body system response to a trigger. According to Scherer's Component Process Model (CPM) of emotion, there are five crucial elements of emotion. From the component process perspective, emotional experience requires that all of these processes become coordinated and synchronized for a short period of time, driven by appraisal processes. Although the inclusion of cognitive appraisal as one of the elements is slightly controversial, since some theorists make the assumption that emotion and cognition are separate but interacting systems, the CPM provides a sequence of events that effectively describes the coordination involved during an emotional episode. Emotion can be differentiated from a number of similar constructs within the field of affective neuroscience: One view is that emotions facilitate adaptive responses to environmental challenges. Emotions have been described as a result of evolution because they provided good solutions to ancient and recurring problems that faced our ancestors. Emotions can function as a way to communicate what's important to us, such as values and ethics. However some emotions, such as some forms of anxiety, are sometimes regarded as part of a mental illness and thus possibly of negative value. A distinction can be made between emotional episodes and emotional dispositions. Emotional dispositions are also comparable to character traits, where someone may be said to be generally disposed to experience certain emotions. For example, an irritable person is generally disposed to feel irritation more easily or quickly than others do. Finally, some theorists place emotions within a more general category of "affective states" where affective states can also include emotion-related phenomena such as pleasure and pain, motivational states (for example, hunger or curiosity), moods, dispositions and traits. For more than 40 years, Paul Ekman has supported the view that emotions are discrete, measurable, and physiologically distinct. Ekman's most influential work revolved around the finding that certain emotions appeared to be universally recognized, even in cultures that were preliterate and could not have learned associations for facial expressions through media. Another classic study found that when participants contorted their facial muscles into distinct facial expressions (for example, disgust), they reported subjective and physiological experiences that matched the distinct facial expressions. Ekman's facial-expression research examined six basic emotions: anger, disgust, fear, happiness, sadness and surprise. Later in his career, Ekman theorized that other universal emotions may exist beyond these six. In light of this, recent cross-cultural studies led by Daniel Cordaro and Dacher Keltner, both former students of Ekman, extended the list of universal emotions. In addition to the original six, these studies provided evidence for amusement, awe, contentment, desire, embarrassment, pain, relief, and sympathy in both facial and vocal expressions. They also found evidence for boredom, confusion, interest, pride, and shame facial expressions, as well as contempt, interest, relief, and triumph vocal expressions. Robert Plutchik agreed with Ekman's biologically driven perspective but developed the "wheel of emotions", suggesting eight primary emotions grouped on a positive or negative basis: joy versus sadness; anger versus fear; trust versus disgust; and surprise versus anticipation. Some basic emotions can be modified to form complex emotions. The complex emotions could arise from cultural conditioning or association combined with the basic emotions. Alternatively, similar to the way primary colors combine, "primary emotions" could blend to form the full spectrum of human emotional experience. For example, interpersonal anger and disgust could blend to form contempt. Relationships exist between basic emotions, resulting in positive or negative influences. Psychologists have used methods such as factor analysis to attempt to map emotion-related responses onto a more limited number of dimensions. Such methods attempt to boil emotions down to underlying dimensions that capture the similarities and differences between experiences. Often, the first two dimensions uncovered by factor analysis are valence (how negative or positive the experience feels) and arousal (how energized or enervated the experience feels). These two dimensions can be depicted on a 2D coordinate map. This two-dimensional map has been theorized to capture one important component of emotion called core affect. Core affect is not theorized to be the only component to emotion, but to give the emotion its hedonic and felt energy. In stoic theories it was seen as a hindrance to reason and therefore a hindrance to virtue. Aristotle believed that emotions were an essential component of virtue. In the Aristotelian view all emotions (called passions) corresponded to appetites or capacities. During the Middle Ages, the Aristotelian view was adopted and further developed by scholasticism and Thomas Aquinas in particular. In Chinese antiquity, excessive emotion was believed to cause damage to "qi", which in turn, damages the vital organs. The four humours theory made popular by Hippocrates contributed to the study of emotion in the same way that it did for medicine. In the early 11th century, Avicenna theorized about the influence of emotions on health and behaviors, suggesting the need to manage emotions. Early modern views on emotion are developed in the works of philosophers such as René Descartes, Niccolò Machiavelli, Baruch Spinoza, Thomas Hobbes and David Hume. In the 19th century emotions were considered adaptive and were studied more frequently from an empiricist psychiatric perspective. Christian perspective on emotion presupposes a theistic origin to humanity. God who created humans gave humans the ability to feel emotion and interact emotionally. Biblical content expresses that God is a person who feels and expresses emotion. Though a somatic view would place the locus of emotions in the physical body, Christian theory of emotions would view the body more as a platform for the sensing and expression of emotions. Therefore emotions themselves arise from the person, or that which is "imago-dei" or image of God in humans. In Christian thought, emotions have the potential to be controlled through reasoned reflection. That reasoned reflection also mimics God who made mind. The purpose of emotions in human life are therefore summarized in God's call to enjoy Him and creation, humans are to enjoy emotions and benefit from them and use them to energize behavior. Perspectives on emotions from evolutionary theory were initiated during the mid-late 19th century with Charles Darwin's 1872 book "The Expression of the Emotions in Man and Animals". Surprisingly, Darwin argued that emotions served no evolved purpose for humans, neither in communication, nor in aiding survival. Darwin largely argued that emotions evolved via the inheritance of acquired characters. He pioneered various methods for studying non-verbal expressions, from which he concluded that some expressions had cross-cultural universality. Darwin also detailed homologous expressions of emotions that occur in animals. This led the way for animal research on emotions and the eventual determination of the neural underpinnings of emotion. More contemporary views along the evolutionary psychology spectrum posit that both basic emotions and social emotions evolved to motivate (social) behaviors that were adaptive in the ancestral environment. Emotion is an essential part of any human decision-making and planning, and the famous distinction made between reason and emotion is not as clear as it seems. Paul D. MacLean claims that emotion competes with even more instinctive responses, on one hand, and the more abstract reasoning, on the other hand. The increased potential in neuroimaging has also allowed investigation into evolutionarily ancient parts of the brain. Important neurological advances were derived from these perspectives in the 1990s by Joseph E. LeDoux and António Damásio. Research on social emotion also focuses on the physical displays of emotion including body language of animals and humans (see affect display). For example, spite seems to work against the individual but it can establish an individual's reputation as someone to be feared. Shame and pride can motivate behaviors that help one maintain one's standing in a community, and self-esteem is one's estimate of one's status. Somatic theories of emotion claim that bodily responses, rather than cognitive interpretations, are essential to emotions. The first modern version of such theories came from William James in the 1880s. The theory lost favor in the 20th century, but has regained popularity more recently due largely to theorists such as John Cacioppo, António Damásio, Joseph E. LeDoux and Robert Zajonc who are able to appeal to neurological evidence. In his 1884 article William James argued that feelings and emotions were "secondary" to physiological phenomena. In his theory, James proposed that the perception of what he called an "exciting fact" directly led to a physiological response, known as "emotion." To account for different types of emotional experiences, James proposed that stimuli trigger activity in the autonomic nervous system, which in turn produces an emotional experience in the brain. The Danish psychologist Carl Lange also proposed a similar theory at around the same time, and therefore this theory became known as the James–Lange theory. As James wrote, "the perception of bodily changes, as they occur, "is" the emotion." James further claims that "we feel sad because we cry, angry because we strike, afraid because we tremble, and either we cry, strike, or tremble because we are sorry, angry, or fearful, as the case may be." An example of this theory in action would be as follows: An emotion-evoking stimulus (snake) triggers a pattern of physiological response (increased heart rate, faster breathing, etc.), which is interpreted as a particular emotion (fear). This theory is supported by experiments in which by manipulating the bodily state induces a desired emotional state. Some people may believe that emotions give rise to emotion-specific actions, for example, "I'm crying because I'm sad," or "I ran away because I was scared." The issue with the James–Lange theory is that of causation (bodily states causing emotions and being "a priori"), not that of the bodily influences on emotional experience (which can be argued and is still quite prevalent today in biofeedback studies and embodiment theory). Although mostly abandoned in its original form, Tim Dalgleish argues that most contemporary neuroscientists have embraced the components of the James-Lange theory of emotions. Walter Bradford Cannon agreed that physiological responses played a crucial role in emotions, but did not believe that physiological responses alone could explain subjective emotional experiences. He argued that physiological responses were too slow and often imperceptible and this could not account for the relatively rapid and intense subjective awareness of emotion. He also believed that the richness, variety, and temporal course of emotional experiences could not stem from physiological reactions, that reflected fairly undifferentiated fight or flight responses. An example of this theory in action is as follows: An emotion-evoking event (snake) triggers simultaneously both a physiological response and a conscious experience of an emotion. Phillip Bard contributed to the theory with his work on animals. Bard found that sensory, motor, and physiological information all had to pass through the diencephalon (particularly the thalamus), before being subjected to any further processing. Therefore, Cannon also argued that it was not anatomically possible for sensory events to trigger a physiological response prior to triggering conscious awareness and emotional stimuli had to trigger both physiological and experiential aspects of emotion simultaneously. Stanley Schachter formulated his theory on the earlier work of a Spanish physician, Gregorio Marañón, who injected patients with epinephrine and subsequently asked them how they felt. Marañón found that most of these patients felt something but in the absence of an actual emotion-evoking stimulus, the patients were unable to interpret their physiological arousal as an experienced emotion. Schachter did agree that physiological reactions played a big role in emotions. He suggested that physiological reactions contributed to emotional experience by facilitating a focused cognitive appraisal of a given physiologically arousing event and that this appraisal was what defined the subjective emotional experience. Emotions were thus a result of two-stage process: general physiological arousal, and experience of emotion. For example, the physiological arousal, heart pounding, in a response to an evoking stimulus, the sight of a bear in the kitchen. The brain then quickly scans the area, to explain the pounding, and notices the bear. Consequently, the brain interprets the pounding heart as being the result of fearing the bear. With his student, Jerome Singer, Schachter demonstrated that subjects can have different emotional reactions despite being placed into the same physiological state with an injection of epinephrine. Subjects were observed to express either anger or amusement depending on whether another person in the situation (a confederate) displayed that emotion. Hence, the combination of the appraisal of the situation (cognitive) and the participants' reception of adrenaline or a placebo together determined the response. This experiment has been criticized in Jesse Prinz's (2004) "Gut Reactions". With the two-factor theory now incorporating cognition, several theories began to argue that cognitive activity in the form of judgments, evaluations, or thoughts were entirely necessary for an emotion to occur. One of the main proponents of this view was Richard Lazarus who argued that emotions must have some cognitive intentionality. The cognitive activity involved in the interpretation of an emotional context may be conscious or unconscious and may or may not take the form of conceptual processing. Lazarus' theory is very influential; emotion is a disturbance that occurs in the following order: For example: Jenny sees a snake. Lazarus stressed that the quality and intensity of emotions are controlled through cognitive processes. These processes underline coping strategies that form the emotional reaction by altering the relationship between the person and the environment. George Mandler provided an extensive theoretical and empirical discussion of emotion as influenced by cognition, consciousness, and the autonomic nervous system in two books ("Mind and Emotion", 1975, and "Mind and Body: Psychology of Emotion and Stress", 1984) There are some theories on emotions arguing that cognitive activity in the form of judgments, evaluations, or thoughts are necessary in order for an emotion to occur. A prominent philosophical exponent is Robert C. Solomon (for example, "The Passions, Emotions and the Meaning of Life", 1993). Solomon claims that emotions are judgments. He has put forward a more nuanced view which responds to what he has called the ‘standard objection’ to cognitivism, the idea that a judgment that something is fearsome can occur with or without emotion, so judgment cannot be identified with emotion. The theory proposed by Nico Frijda where appraisal leads to action tendencies is another example. It has also been suggested that emotions (affect heuristics, feelings and gut-feeling reactions) are often used as shortcuts to process information and influence behavior. The affect infusion model (AIM) is a theoretical model developed by Joseph Forgas in the early 1990s that attempts to explain how emotion and mood interact with one's ability to process information. Perceptual theory Theories dealing with perception either use one or multiples perceptions in order to find an emotion. A recent hybrid of the somatic and cognitive theories of emotion is the perceptual theory. This theory is neo-Jamesian in arguing that bodily responses are central to emotions, yet it emphasizes the meaningfulness of emotions or the idea that emotions are about something, as is recognized by cognitive theories. The novel claim of this theory is that conceptually-based cognition is unnecessary for such meaning. Rather the bodily changes themselves "perceive" the meaningful content of the emotion because of being causally triggered by certain situations. In this respect, emotions are held to be analogous to faculties such as vision or touch, which provide information about the relation between the subject and the world in various ways. A sophisticated defense of this view is found in philosopher Jesse Prinz's book "Gut Reactions", and psychologist James Laird's book "Feelings". Affective events theory Affective events theory is a communication-based theory developed by Howard M. Weiss and Russell Cropanzano (1996), that looks at the causes, structures, and consequences of emotional experience (especially in work contexts). This theory suggests that emotions are influenced and caused by events which in turn influence attitudes and behaviors. This theoretical frame also emphasizes "time" in that human beings experience what they call emotion episodes –\ a "series of emotional states extended over time and organized around an underlying theme." This theory has been utilized by numerous researchers to better understand emotion from a communicative lens, and was reviewed further by Howard M. Weiss and Daniel J. Beal in their article, "Reflections on Affective Events Theory", published in "Research on Emotion in Organizations" in 2005. A situated perspective on emotion, developed by Paul E. Griffiths and Andrea Scarantino, emphasizes the importance of external factors in the development and communication of emotion, drawing upon the situationism approach in psychology. This theory is markedly different from both cognitivist and neo-Jamesian theories of emotion, both of which see emotion as a purely internal process, with the environment only acting as a stimulus to the emotion. In contrast, a situationist perspective on emotion views emotion as the product of an organism investigating its environment, and observing the responses of other organisms. Emotion stimulates the evolution of social relationships, acting as a signal to mediate the behavior of other organisms. In some contexts, the expression of emotion (both voluntary and involuntary) could be seen as strategic moves in the transactions between different organisms. The situated perspective on emotion states that conceptual thought is not an inherent part of emotion, since emotion is an action-oriented form of skillful engagement with the world. Griffiths and Scarantino suggested that this perspective on emotion could be helpful in understanding phobias, as well as the emotions of infants and animals. Emotions can motivate social interactions and relationships and therefore are directly related with basic physiology, particularly with the stress systems. This is important because emotions are related to the anti-stress complex, with an oxytocin-attachment system, which plays a major role in bonding. Emotional phenotype temperaments affect social connectedness and fitness in complex social systems. These characteristics are shared with other species and taxa and are due to the effects of genes and their continuous transmission. Information that is encoded in the DNA sequences provides the blueprint for assembling proteins that make up our cells. Zygotes require genetic information from their parental germ cells, and at every speciation event, heritable traits that have enabled its ancestor to survive and reproduce successfully are passed down along with new traits that could be potentially beneficial to the offspring. In the five million years since the lineages leading to modern humans and chimpanzees split, only about 1.2% of their genetic material has been modified. This suggests that everything that separates us from chimpanzees must be encoded in that very small amount of DNA, including our behaviors. Students that study animal behaviors have only identified intraspecific examples of gene-dependent behavioral phenotypes. In voles (Microtus spp.) minor genetic differences have been identified in a vasopressin receptor gene that corresponds to major species differences in social organization and the mating system. Another potential example with behavioral differences is the FOCP2 gene, which is involved in neural circuitry handling speech and language. Its present form in humans differed from that of the chimpanzees by only a few mutations and has been present for about 200,000 years, coinciding with the beginning of modern humans. Speech, language, and social organization are all part of the basis for emotions. Based on discoveries made through neural mapping of the limbic system, the neurobiological explanation of human emotion is that emotion is a pleasant or unpleasant mental state organized in the limbic system of the mammalian brain. If distinguished from reactive responses of reptiles, emotions would then be mammalian elaborations of general vertebrate arousal patterns, in which neurochemicals (for example, dopamine, noradrenaline, and serotonin) step-up or step-down the brain's activity level, as visible in body movements, gestures and postures. Emotions can likely be mediated by pheromones (see fear). For example, the emotion of love is proposed to be the expression of paleocircuits of the mammalian brain (specifically, modules of the cingulate gyrus) which facilitate the care, feeding, and grooming of offspring. Paleocircuits are neural platforms for bodily expression configured before the advent of cortical circuits for speech. They consist of pre-configured pathways or networks of nerve cells in the forebrain, brain stem and spinal cord. The motor centers of reptiles react to sensory cues of vision, sound, touch, chemical, gravity, and motion with pre-set body movements and programmed postures. With the arrival of night-active mammals, smell replaced vision as the dominant sense, and a different way of responding arose from the olfactory sense, which is proposed to have developed into mammalian emotion and emotional memory. The mammalian brain invested heavily in olfaction to succeed at night as reptiles slept – one explanation for why olfactory lobes in mammalian brains are proportionally larger than in the reptiles. These odor pathways gradually formed the neural blueprint for what was later to become our limbic brain. Emotions are thought to be related to certain activities in brain areas that direct our attention, motivate our behavior, and determine the significance of what is going on around us. Pioneering work by Paul Broca (1878), James Papez (1937), and Paul D. MacLean (1952) suggested that emotion is related to a group of structures in the center of the brain called the limbic system, which includes the hypothalamus, cingulate cortex, hippocampi, and other structures. More recent research has shown that some of these limbic structures are not as directly related to emotion as others are while some non-limbic structures have been found to be of greater emotional relevance. There is ample evidence that the left prefrontal cortex is activated by stimuli that cause positive approach. If attractive stimuli can selectively activate a region of the brain, then logically the converse should hold, that selective activation of that region of the brain should cause a stimulus to be judged more positively. This was demonstrated for moderately attractive visual stimuli and replicated and extended to include negative stimuli. Two neurobiological models of emotion in the prefrontal cortex made opposing predictions. The valence model predicted that anger, a negative emotion, would activate the right prefrontal cortex. The direction model predicted that anger, an approach emotion, would activate the left prefrontal cortex. The second model was supported. This still left open the question of whether the opposite of approach in the prefrontal cortex is better described as moving away (direction model), as unmoving but with strength and resistance (movement model), or as unmoving with passive yielding (action tendency model). Support for the action tendency model (passivity related to right prefrontal activity) comes from research on shyness and research on behavioral inhibition. Research that tested the competing hypotheses generated by all four models also supported the action tendency model. Another neurological approach proposed by Bud Craig in 2003 distinguishes two classes of emotion: "classical" emotions such as love, anger and fear that are evoked by environmental stimuli, and "homeostatic emotions" – attention-demanding feelings evoked by body states, such as pain, hunger and fatigue, that motivate behavior (withdrawal, eating or resting in these examples) aimed at maintaining the body's internal milieu at its ideal state. Derek Denton calls the latter "primordial emotions" and defines them as "the subjective element of the instincts, which are the genetically programmed behavior patterns which contrive homeostasis. They include thirst, hunger for air, hunger for food, pain and hunger for specific minerals etc. There are two constituents of a primordial emotion--the specific sensation which when severe may be imperious, and the compelling intention for gratification by a consummatory act." Joseph LeDoux differentiates between the human's defence system, which has evolved over time, and emotions such as fear and anxiety. He has said that the amygdala may release hormones due to a trigger (such as an innate reaction to seeing a snake), but ""then we elaborate it through cognitive and conscious processes."" Lisa Feldman Barrett highlights differences in emotions between different cultures, and says that emotions (such as anxiety) ""are not triggered; you create them. They emerge as a combination of the physical properties of your body, a flexible brain that wires itself to whatever environment it develops in, and your culture and upbringing, which provide that environment."" She has termed this approach the theory of constructed emotion. Many different disciplines have produced work on the emotions. Human sciences study the role of emotions in mental processes, disorders, and neural mechanisms. In psychiatry, emotions are examined as part of the discipline's study and treatment of mental disorders in humans. Nursing studies emotions as part of its approach to the provision of holistic health care to humans. Psychology examines emotions from a scientific perspective by treating them as mental processes and behavior and they explore the underlying physiological and neurological processes. In neuroscience sub-fields such as social neuroscience and affective neuroscience, scientists study the neural mechanisms of emotion by combining neuroscience with the psychological study of personality, emotion, and mood. In linguistics, the expression of emotion may change to the meaning of sounds. In education, the role of emotions in relation to learning is examined. Social sciences often examine emotion for the role that it plays in human culture and social interactions. In sociology, emotions are examined for the role they play in human society, social patterns and interactions, and culture. In anthropology, the study of humanity, scholars use ethnography to undertake contextual analyses and cross-cultural comparisons of a range of human activities. Some anthropology studies examine the role of emotions in human activities. In the field of communication sciences, critical organizational scholars have examined the role of emotions in organizations, from the perspectives of managers, employees, and even customers. A focus on emotions in organizations can be credited to Arlie Russell Hochschild's concept of emotional labor. The University of Queensland hosts EmoNet, an e-mail distribution list representing a network of academics that facilitates scholarly discussion of all matters relating to the study of emotion in organizational settings. The list was established in January 1997 and has over 700 members from across the globe. In economics, the social science that studies the production, distribution, and consumption of goods and services, emotions are analyzed in some sub-fields of microeconomics, in order to assess the role of emotions on purchase decision-making and risk perception. In criminology, a social science approach to the study of crime, scholars often draw on behavioral sciences, sociology, and psychology; emotions are examined in criminology issues such as anomie theory and studies of "toughness," aggressive behavior, and hooliganism. In law, which underpins civil obedience, politics, economics and society, evidence about people's emotions is often raised in tort law claims for compensation and in criminal law prosecutions against alleged lawbreakers (as evidence of the defendant's state of mind during trials, sentencing, and parole hearings). In political science, emotions are examined in a number of sub-fields, such as the analysis of voter decision-making. In philosophy, emotions are studied in sub-fields such as ethics, the philosophy of art (for example, sensory–emotional values, and matters of taste and sentimentality), and the philosophy of music (see also Music and emotion). In history, scholars examine documents and other sources to interpret and analyze past activities; speculation on the emotional state of the authors of historical documents is one of the tools of interpretation. In literature and film-making, the expression of emotion is the cornerstone of genres such as drama, melodrama, and romance. In communication studies, scholars study the role that emotion plays in the dissemination of ideas and messages. Emotion is also studied in non-human animals in ethology, a branch of zoology which focuses on the scientific study of animal behavior. Ethology is a combination of laboratory and field science, with strong ties to ecology and evolution. Ethologists often study one type of behavior (for example, aggression) in a number of unrelated animals. The history of emotions has become an increasingly popular topic recently, with some scholars arguing that it is an essential category of analysis, not unlike class, race, or gender. Historians, like other social scientists, assume that emotions, feelings and their expressions are regulated in different ways by both different cultures and different historical times, and the constructivist school of history claims even that some sentiments and meta-emotions, for example Schadenfreude, are learnt and not only regulated by culture. Historians of emotion trace and analyse the changing norms and rules of feeling, while examining emotional regimes, codes, and lexicons from social, cultural, or political history perspectives. Others focus on the history of medicine, science, or psychology. What somebody can and may feel (and show) in a given situation, towards certain people or things, depends on social norms and rules; thus historically variable and open to change. Several research centers have opened in the past few years in Germany, England, Spain, Sweden, and Australia. Furthermore, research in historical trauma suggests that some traumatic emotions can be passed on from parents to offspring to second and even third generation, presented as examples of transgenerational trauma. A common way in which emotions are conceptualized in sociology is in terms of the multidimensional characteristics including cultural or emotional labels (for example, anger, pride, fear, happiness), physiological changes (for example, increased perspiration, changes in pulse rate), expressive facial and body movements (for example, smiling, frowning, baring teeth), and appraisals of situational cues. One comprehensive theory of emotional arousal in humans has been developed by Jonathan Turner (2007: 2009). Two of the key eliciting factors for the arousal of emotions within this theory are expectations states and sanctions. When people enter a situation or encounter with certain expectations for how the encounter should unfold, they will experience different emotions depending on the extent to which expectations for Self, other and situation are met or not met. People can also provide positive or negative sanctions directed at Self or other which also trigger different emotional experiences in individuals. Turner analyzed a wide range of emotion theories across different fields of research including sociology, psychology, evolutionary science, and neuroscience. Based on this analysis, he identified four emotions that all researchers consider being founded on human neurology including assertive-anger, aversion-fear, satisfaction-happiness, and disappointment-sadness. These four categories are called primary emotions and there is some agreement amongst researchers that these primary emotions become combined to produce more elaborate and complex emotional experiences. These more elaborate emotions are called first-order elaborations in Turner's theory and they include sentiments such as pride, triumph, and awe. Emotions can also be experienced at different levels of intensity so that feelings of concern are a low-intensity variation of the primary emotion aversion-fear whereas depression is a higher intensity variant. Attempts are frequently made to regulate emotion according to the conventions of the society and the situation based on many (sometimes conflicting) demands and expectations which originate from various entities. The expression of anger is in many cultures discouraged in girls and women to a greater extent than in boys and men (the notion being that an angry man has a valid complaint that needs to be rectified, while an angry women is hysterical or oversensitive, and her anger is somehow invalid), while the expression of sadness or fear is discouraged in boys and men relative to girls and women (attitudes implicit in phrases like "man up" or "don't be a sissy"). Expectations attached to social roles, such as "acting as man" and not as a woman, and the accompanying "feeling rules" contribute to the differences in expression of certain emotions. Some cultures encourage or discourage happiness, sadness, or jealousy, and the free expression of the emotion of disgust is considered socially unacceptable in most cultures. Some social institutions are seen as based on certain emotion, such as love in the case of contemporary institution of marriage. In advertising, such as health campaigns and political messages, emotional appeals are commonly found. Recent examples include no-smoking health campaigns and political campaigns emphasizing the fear of terrorism. Sociological attention to emotion has varied over time. Émile Durkheim (1915/1965) wrote about the collective effervescence or emotional energy that was experienced by members of totemic rituals in Australian aborigine society. He explained how the heightened state of emotional energy achieved during totemic rituals transported individuals above themselves giving them the sense that they were in the presence of a higher power, a force, that was embedded in the sacred objects that were worshipped. These feelings of exaltation, he argued, ultimately lead people to believe that there were forces that governed sacred objects. In the 1990s, sociologists focused on different aspects of specific emotions and how these emotions were socially relevant. For Cooley (1992), pride and shame were the most important emotions that drive people to take various social actions. During every encounter, he proposed that we monitor ourselves through the "looking glass" that the gestures and reactions of others provide. Depending on these reactions, we either experience pride or shame and this results in particular paths of action. Retzinger (1991) conducted studies of married couples who experienced cycles of rage and shame. Drawing predominantly on Goffman and Cooley's work, Scheff (1990) developed a micro sociological theory of the social bond. The formation or disruption of social bonds is dependent on the emotions that people experience during interactions. Subsequent to these developments, Randall Collins (2004) formulated his interaction ritual theory by drawing on Durkheim's work on totemic rituals that was extended by Goffman (1964/2013; 1967) into everyday focused encounters. Based on interaction ritual theory, we experience different levels or intensities of emotional energy during face-to-face interactions. Emotional energy is considered to be a feeling of confidence to take action and a boldness that one experiences when they are charged up from the collective effervescence generated during group gatherings that reach high levels of intensity. There is a growing body of research applying the sociology of emotion to understanding the learning experiences of students during classroom interactions with teachers and other students (for example, Milne & Otieno, 2007; Olitsky, 2007; Tobin, et al., 2013; Zembylas, 2002). These studies show that learning subjects like science can be understood in terms of classroom interaction rituals that generate emotional energy and collective states of emotional arousal like emotional climate. Apart from interaction ritual traditions of the sociology of emotion, other approaches have been classed into one of six other categories: This list provides a general overview of different traditions in the sociology of emotion that sometimes conceptualise emotion in different ways and at other times in complementary ways. Many of these different approaches were synthesized by Turner (2007) in his sociological theory of human emotions in an attempt to produce one comprehensive sociological account that draws on developments from many of the above traditions. Emotion regulation refers to the cognitive and behavioral strategies people use to influence their own emotional experience. For example, a behavioral strategy in which one avoids a situation to avoid unwanted emotions (trying not to think about the situation, doing distracting activities, etc.). Depending on the particular school's general emphasis on either cognitive components of emotion, physical energy discharging, or on symbolic movement and facial expression components of emotion different schools of psychotherapy approach the regulation of emotion differently. Cognitively oriented schools approach them via their cognitive components, such as rational emotive behavior therapy. Yet others approach emotions via symbolic movement and facial expression components (like in contemporary Gestalt therapy). Research on emotions reveals the strong presence of cross-cultural differences in emotional reactions and that emotional reactions are likely to be culture-specific. In strategic settings, cross-cultural research on emotions is required for understanding the psychological situation of a given population or specific actors. This implies the need to comprehend the current emotional state, mental disposition or other behavioral motivation of a target audience located in a different culture, basically founded on its national political, social, economic, and psychological peculiarities but also subject to the influence of circumstances and events. There are many cultural variations in emotions. Trnka et al. (2018) proposed framework which conceptually distinguishes five main components of cultural complexity relating to emotions: "1) emotion language, 2) conceptual knowledge about emotions, 3) emotion-related values, 4) feelings rules, i.e. norms for subjective experience, and 5) display rules, i.e. norms for emotional expression." In the 2000s, research in computer science, engineering, psychology and neuroscience has been aimed at developing devices that recognize human affect display and model emotions. In computer science, affective computing is a branch of the study and development of artificial intelligence that deals with the design of systems and devices that can recognize, interpret, and process human emotions. It is an interdisciplinary field spanning computer sciences, psychology, and cognitive science. While the origins of the field may be traced as far back as to early philosophical enquiries into emotion, the more modern branch of computer science originated with Rosalind Picard's 1995 paper on affective computing. Detecting emotional information begins with passive sensors which capture data about the user's physical state or behavior without interpreting the input. The data gathered is analogous to the cues humans use to perceive emotions in others. Another area within affective computing is the design of computational devices proposed to exhibit either innate emotional capabilities or that are capable of convincingly simulating emotions. Emotional speech processing recognizes the user's emotional state by analyzing speech patterns. The detection and processing of facial expression or body gestures is achieved through detectors and sensors. In the late 19th century, the most influential theorists were William James (1842–1910) and Carl Lange (1834–1900). James was an American psychologist and philosopher who wrote about educational psychology, psychology of religious experience/mysticism, and the philosophy of pragmatism. Lange was a Danish physician and psychologist. Working independently, they developed the James–Lange theory, a hypothesis on the origin and nature of emotions. The theory states that within human beings, as a response to experiences in the world, the autonomic nervous system creates physiological events such as muscular tension, a rise in heart rate, perspiration, and dryness of the mouth. Emotions, then, are feelings which come about as a result of these physiological changes, rather than being their cause. Silvan Tomkins (1911–1991) developed the affect theory and script theory. The affect theory introduced the concept of basic emotions, and was based on the idea that the dominance of the emotion, which he called the affected system, was the motivating force in human life. Some of the most influential deceased theorists on emotion from the 20th century include Magda B. Arnold (1903–2002), an American psychologist who developed the appraisal theory of emotions; Richard Lazarus (1922–2002), an American psychologist who specialized in emotion and stress, especially in relation to cognition; Herbert A. Simon (1916–2001), who included emotions into decision making and artificial intelligence; Robert Plutchik (1928–2006), an American psychologist who developed a psychoevolutionary theory of emotion; Robert Zajonc (1923–2008) a Polish–American social psychologist who specialized in social and cognitive processes such as social facilitation; Robert C. Solomon (1942–2007), an American philosopher who contributed to the theories on the philosophy of emotions with books such as "What Is An Emotion?: Classic and Contemporary Readings" (2003); Peter Goldie (1946–2011), a British philosopher who specialized in ethics, aesthetics, emotion, mood and character; Nico Frijda (1927–2015), a Dutch psychologist who advanced the theory that human emotions serve to promote a tendency to undertake actions that are appropriate in the circumstances, detailed in his book "The Emotions" (1986); Jaak Panksepp (1943-2017), an Estonian-born American psychologist, psychobiologist, neuroscientist and pioneer in affective neuroscience. Influential theorists who are still active include the following psychologists, neurologists, philosophers, and sociologists:
https://en.wikipedia.org/wiki?curid=10406
Epictetus Epictetus (; , "Epíktētos"; 50 135 AD) was a Greek Stoic philosopher. He was born a slave at Hierapolis, Phrygia (present day Pamukkale, Turkey) and lived in Rome until his banishment, when he went to Nicopolis in northwestern Greece for the rest of his life. His teachings were written down and published by his pupil Arrian in his "Discourses" and "Enchiridion". Epictetus taught that philosophy is a way of life and not just a theoretical discipline. To Epictetus, all external events are beyond our control; we should accept calmly and dispassionately whatever happens. However, individuals are responsible for their own actions, which they can examine and control through rigorous self-discipline. Epictetus was born AD 50, presumably at Hierapolis, Phrygia. The name his parents gave him is unknown; the word "epíktētos" (ἐπίκτητος) in Greek simply means "gained" or "acquired"; the Greek philosopher Plato, in his "Laws", uses the term as property that is "added to one's hereditary property". He spent his youth as a slave in Rome to Epaphroditos, a wealthy freedman and secretary to Nero. Early in life, Epictetus acquired a passion for philosophy and, with the permission of his wealthy owner, he studied Stoic philosophy under Musonius Rufus, which allowed him to rise in respectability as he grew more educated. Somehow, he became crippled. Origen stated that his leg was deliberately broken by his master. Simplicius stated that he had been lame from childhood. Epictetus obtained his freedom sometime after the death of Nero in 68 A.D., and he began to teach philosophy in Rome. About 93 A.D. Emperor Domitian banished all philosophers from the city, and Epictetus went to Nicopolis in Epirus, Greece, where he founded a philosophical school. His most famous pupil, Arrian, studied under him when a young man (ca. AD 108) and claimed to have written the famous "Discourses" from his lecture notes, which he argued should be considered comparable to the Socratic literature. Arrian describes Epictetus as being a powerful speaker who could "induce his listener to feel just what Epictetus wanted him to feel." Many eminent figures sought conversations with him. Emperor Hadrian was friendly with him, and may have listened to him speak at his school in Nicopolis. He lived a life of great simplicity, with few possessions. He lived alone for a long time, but in his old age he adopted a friend's child who otherwise would have been left to die, and raised him with the aid of a woman. It is unclear whether Epictetus and she were married. He died sometime around 135 A.D. After his death, according to Lucian, his oil lamp was purchased by an admirer for 3,000 drachmae. No writings by Epictetus are known. His discourses were transcribed and compiled by his pupil Arrian (author of the "Anabasis Alexandri"). The main work is "The Discourses", four books of which have been preserved (out of the original eight). Arrian also compiled a popular digest, entitled the "Enchiridion", or "Handbook." In a preface to the "Discourses" that is addressed to Lucius Gellius, Arrian states that "whatever I heard him say I used to write down, word for word, as best I could, endeavouring to preserve it as a memorial, for my own future use, of his way of thinking and the frankness of his speech." Epictetus maintains that the foundation of all philosophy is self-knowledge, that is, the conviction of our ignorance and gullibility ought to be the first subject of our study. Logic provides valid reasoning and certainty in judgment, but it is subordinate to practical needs. The first and most necessary part of philosophy concerns the application of doctrine, for example, that people should not lie. The second concerns reasons, e.g. why people should not lie. While the third, lastly, examines and establishes the reasons. This is the logical part, which finds reasons, shows what is a reason, and that a given reason is a correct one. This last part is necessary, but only on account of the second, which again is rendered necessary by the first. Both the "Discourses" and the "Enchiridion" begin by distinguishing between those things in our power ("prohairetic" things) and those things not in our power ("aprohairetic" things). That alone is in our power, which is our own work; and in this class are our opinions, impulses, desires, and aversions. What, on the contrary, is not in our power, are our bodies, possessions, glory, and power. Any delusion on this point leads to the greatest errors, misfortunes, and troubles, and to the slavery of the soul. We have no power over external things, and the good that ought to be the object of our earnest pursuit, is to be found only within ourselves. The determination between what is good and what is not good is made by the capacity for choice ("prohairesis"). Prohairesis allows us to act, and gives us the kind of freedom that only rational animals have. It is determined by our reason, which of all our faculties, sees and tests itself and everything else. It is the correct use of the impressions ("phantasia") that bombard the mind that is in our power: Practice then from the start to say to every harsh impression, "You are an impression, and not at all the thing you appear to be." Then examine it and test it by these rules you have, and firstly, and chiefly, by this: whether the impression has to do with the things that are up to us, or those that are not; and if it has to do with the things that are not up to us, be ready to reply, "It is nothing to me." We will not be troubled at any loss, but will say to ourselves on such an occasion: "I have lost nothing that belongs to me; it was not something of mine that was torn from me, but something that was not in my power has left me." Nothing beyond the use of our opinion is properly ours. Every possession rests on opinion. What is to cry and to weep? An opinion. What is misfortune, or a quarrel, or a complaint? All these things are opinions; opinions founded on the delusion that what is not subject to our own choice can be either good or evil, which it cannot. By rejecting these opinions, and seeking good and evil in the power of choice alone, we may confidently achieve peace of mind in every condition of life. Reason alone is good, the irrational is evil, and the irrational is intolerable to the rational. The good person should labour chiefly on their own reason; to perfect this is in our power. To repel evil opinions by the good is the noble contest in which humans should engage; it is not an easy task, but it promises true freedom, peace of mind ("ataraxia"), and a divine command over the emotions ("apatheia"). We should especially be on our guard against the opinion of pleasure because of its apparent sweetness and charms. The first object of philosophy, therefore, is to purify the mind. Epictetus teaches that the preconceptions ("prolepsis") of good and evil are common to all. Good alone is profitable and to be desired, and evil is hurtful and to be avoided. Different opinions arise only from the application of these preconceptions to particular cases, and it is then that the darkness of ignorance, which blindly maintains the correctness of its own opinion, must be dispelled. People entertain different and conflicting opinions of good, and in their judgment of a particular good, people frequently contradict themselves. Philosophy should provide a standard for good and evil. This process is greatly facilitated because the mind and the works of the mind are alone in our power, whereas all external things that aid life are beyond our control. The essence of divinity is goodness; we have all good that could be given to us. The deities too gave us the soul and reason, which is not measured by breadth or depth, but by knowledge and sentiments, and by which we attain to greatness, and may equal even with the deities. We should, therefore, cultivate the mind with special care. If we wish for nothing, but what God wills, we shall be truly free, and all will come to pass with us according to our desire; and we shall be as little subject to restraint as Zeus himself. Every individual is connected with the rest of the world, and the universe is fashioned for universal harmony. Wise people, therefore, will pursue, not merely their own will, but also will be subject to the rightful order of the world. We should conduct ourselves through life fulfilling all our duties as children, siblings, parents, and citizens. For our country or friends we ought to be ready to undergo or perform the greatest difficulties. The good person, if able to foresee the future, would peacefully and contentedly help to bring about their own sickness, maiming, and even death, knowing that this is the correct order of the universe. We have all a certain part to play in the world, and we have done enough when we have performed what our nature allows. In the exercise of our powers, we may become aware of the destiny we are intended to fulfill. We are like travellers at an inn, or guests at a stranger's table; whatever is offered we take with thankfulness, and sometimes, when the turn comes, we may refuse; in the former case we are a worthy guest of the deities, and in the latter we appear as a sharer in their power. Anyone who finds life intolerable is free to quit it, but we should not abandon our appointed role without sufficient reason. The Stoic sage will never find life intolerable and will complain of no one, neither deity nor human. Those who go wrong we should pardon and treat with compassion, since it is from ignorance that they err, being as it were, blind. It is only our opinions and principles that can render us unhappy, and it is only the ignorant person who finds fault with another. Every desire degrades us, and renders us slaves of what we desire. We ought not to forget the transitory character of all external advantages, even in the midst of our enjoyment of them; but always to bear in mind that they are not our own, and that therefore, they do not properly belong to us. Thus prepared, we shall never be carried away by opinions. The final entry of the "Enchiridion," or "Handbook," begins: "Upon all occasions we ought to have these maxims ready at hand": Conduct me, Zeus, and thou, Destiny, Wherever thy decree has fixed my lot. I follow willingly; and, did I not, Wicked and wretched would I follow still. (Diogenes Laërtius quoting Cleanthes; quoted also by Seneca, Epistle 107.)" Whoe'er yields properly to Fate is deemed Wise among men, and knows the laws of Heaven. Crito, if it thus pleases the gods, thus let it be. Anytus and Meletus may indeed kill me, but they cannot harm me. The philosophy of Epictetus was an influence on the Roman Emperor Marcus Aurelius (AD 121 to AD 180) whose reign was marked by wars with the resurgent Parthia in western Asia and against the Germanic tribes in Europe. Aurelius quotes from Epictetus repeatedly in his own work, "Meditations", written during his campaigns in central Europe. In the sixth century, the Neoplatonist philosopher Simplicius, who was persecuted for his pagan beliefs during the reign of Justinian, wrote an extant commentary on the "Enchiridion". At the end of the commentary Simplicius wrote: "Nor does my writing this commentary prove beneficial to others only, for I myself have already found great advantage from it, by the agreeable diversion it has given me, in a season of trouble and public calamity." When Bernard Stiegler was imprisoned for five years for armed robbery in France, he assembled an "ensemble of disciplines," which he called (in reference to Epictetus) his "melete". This ensemble amounted to a practice of reading and writing that Stiegler derived from the writings of Epictetus. This led to his transformation, and upon being released from incarceration he became a professional philosopher. Stiegler tells the story of this transformation in his book, "Acting Out". The philosophy of Epictetus plays a key role in the 1998 novel by Tom Wolfe, "A Man in Full". This was in part the outcome of discussions Wolfe had with James Stockdale (see below). The character Conrad, who through a series of mishaps finds himself in jail, and accidentally gets a copy of the Enchiridion of Epictetus, the Stoic’s manual, finds a philosophy that strengthens him to endure the brutality of the jail environment. He experiences Joseph Campbell's 'hero's journey' call to action and becomes a strong, honorable, undefeatable protagonist. The importance of Epictetus' Stoicism for Stockdale, its role in "A Man in Full", and its significance in Ridley Scott's "Gladiator" is discussed by William O. Stephens in "The Rebirth of Stoicism?". Mohun Biswas, in the novel "A House for Mr Biswas" (1961), by V.S. Naipaul, is pleased to think himself a follower of Epictetus and Marcus Aurelius; the irony is that he never actually behaves as a Stoic. "Everything has two handles, the one by which it may be carried, the other by which it cannot" is the theme of "Disturbances in the Field" (1983), by Lynne Sharon Schwartz. Lydia, the central character, turns often to "The Golden Sayings of Epictetus"—the latter being a modern selection from the writings of Epictetus, compiled and translated by Hastings Crossley. A line from the "Enchiridion" is used as a title quotation in "The Life and Opinions of Tristram Shandy, Gentleman" by Laurence Sterne, which translates to, "Not things, but opinions about things, trouble men." The quotation alludes to a theme of the novel about how the suffering of many of its characters (above all Walter Shandy) is the result of the opinions and assumptions they make about reality. This is similar to Shakespeare's "There is nothing either good or bad, but thinking makes it so." (Hamlet: Act 2, Scene 2), and John Milton's "The mind is its own place, and in itself can make a heaven of hell, a hell of heaven." Epictetus is mentioned in "A Portrait of the Artist as a Young Man" by James Joyce: in the fifth chapter of the novel the protagonist Stephen Daedalus discusses Epictetus's famous lamp with a dean of his college. Epictetus also is mentioned briefly in "Franny and Zooey" by J. D. Salinger, and is referred to by Theodore Dreiser in his novel "Sister Carrie". Both the longevity of Epictetus's life and his philosophy are alluded to in John Berryman's poem, "Of Suicide." Epictetus is referred to, but not mentioned by name, in Matthew Arnold's sonnet "To a Friend". Arnold provides three historical personalities as his inspiration and support in difficult times (Epictetus is preceded by Homer and succeeded by Sophocles): Much he, whose friendship I not long since won, That halting slave, who in Nicopolis Taught Arrian, when Vespasian's brutal son Cleared Rome of what most shamed him. In the Chapter XXX of François Rabelais' "Pantagruel" (c.1532), Pantagruel's tutor Epistemon had his head cut off after a battle. After he had his head reattached and was brought back to life, he recounts his experience of the damned in hell: The philosophy of Epictetus is well known in the U.S. military through the writings and example of James Stockdale, the 1992 vice presidential candidate of Ross Perot and a fighter pilot who was shot down while serving in the Vietnam War. He was introduced to the works of Epictetus while at Stanford University. In "Courage under Fire: Testing Epictetus's Doctrines in a Laboratory of Human Behavior" (1993), Stockdale credits Epictetus with helping him endure his seven and a half years in captivity, which included torture and four years in solitary confinement. When he was shot down, he reportedly said to himself "I'm leaving the world of technology and entering the world of Epictetus!" as he bailed out. Quoting Epictetus, Stockdale concludes the book with: The emotions of grief, pity, and even affection are well-known disturbers of the soul. Grief is the most offensive; Epictetus considered the suffering of grief an act of evil. It is a willful act, going against the will of God to have all men share happiness. Psychologist Albert Ellis, the founder of Rational Emotive Behavior Therapy, credited Epictetus with providing a foundation for his system of psychotherapy. Kiyozawa Manshi, a controversial reformer within the Higashi Honganji branch of Jodo Shinshu Buddhism cited Epictetus as one of the three major influences on his spiritual development and thought. Epictetus' philosophy is an influence on the acting method introduced by David Mamet and William H. Macy, known as Practical Aesthetics. The main book that describes the method, "The Practical Handbook for the Actor", lists the "Enchiridion" in the bibliography. Note 10, Origen's Contra Celcus, Book vii, add that the episode is in chapter LIII, with a secondary mention of the episode in chapter LIV.
https://en.wikipedia.org/wiki?curid=10407
Edward Lear Edward Lear (12 May 1812, Holloway – 29 January 1888, Sanremo) was an English artist, illustrator, musician, author and poet, now known mostly for his literary nonsense in poetry and prose and especially his limericks, a form he popularised. His principal areas of work as an artist were threefold: as a draughtsman employed to illustrate birds and animals; making coloured drawings during his journeys, which he reworked later, sometimes as plates for his travel books; as a (minor) illustrator of Alfred, Lord Tennyson's poems. As an author, he is known principally for his popular nonsense collections of poems, songs, short stories, botanical drawings, recipes and alphabets. He also composed and published twelve musical settings of Tennyson's poetry. Lear was born into a middle-class family at Holloway, North London, the penultimate of 21 children (and youngest to survive) of Ann Clark Skerrett and Jeremiah Lear, a stockbroker formerly working for the family sugar refining business. He was raised by his eldest sister, also named Ann, 21 years his senior. Jeremiah Lear ended up defaulting to the London Stock Exchange in the economic upheaval following the Napoleonic Wars; owing to the family's now more limited finances, Lear and his sister were required to leave the family home, Bowmans Lodge, and live together when he was aged four. Ann doted on Edward and continued to act as a mother for him until her death, when he was almost 50 years of age. Lear suffered from lifelong health afflictions. From the age of six he suffered frequent "grand mal" epileptic seizures, and bronchitis, asthma, and during later life, partial blindness. Lear experienced his first seizure at a fair near Highgate with his father. The event scared and embarrassed him. Lear felt lifelong guilt and shame for his epileptic condition. His adult diaries indicate that he always sensed the onset of a seizure in time to remove himself from public view. When Lear was about seven years old he began to show signs of depression, possibly due to the instability of his childhood. He suffered from periods of severe melancholia which he referred to as "the Morbids." Lear was already drawing "for bread and cheese" by the time he was aged 16 and soon developed into a serious "ornithological draughtsman" employed by the Zoological Society and then from 1832 to 1836 by the Earl of Derby, who kept a private menagerie at his estate, Knowsley Hall. He was the first major bird artist to draw birds from real live birds, instead of skins. Lear's first publication, published when he was 19 years old, was "Illustrations of the Family of Psittacidae, or Parrots" in 1830. One of the greatest ornithological artists of his era, he taught Elizabeth Gould whilst also contributing to John Gould's works and was compared by some to the naturalist John James Audubon. His eyesight deteriorated too much to work with such precision on the fine drawings and etchings of plates used in lithography, thus he turned to landscape painting and travel. Among other travels, he visited Greece and Egypt during 1848–49, and toured India during 1873–75, including a brief detour to Ceylon. While travelling he produced large quantities of coloured wash drawings in a distinctive style, which he converted later in his studio into oil and watercolour paintings, as well as prints for his books. His landscape style often shows views with strong sunlight, with intense contrasts of colour. Between 1878 and 1883 Lear spent his summers on Monte Generoso, a mountain on the border between the Swiss canton of Ticino and the Italian region of Lombardy. His oil painting "The Plains of Lombardy from Monte Generoso" is in the Ashmolean Museum in Oxford. Throughout his life, he continued to paint seriously. He had a lifelong ambition to illustrate Tennyson's poems; near the end of his life, a volume with a small number of illustrations was published. In 1842 Lear began a journey into the Italian peninsula, travelling through the Lazio, Rome, Abruzzo, Molise, Apulia, Basilicata, Calabria, and Sicily. In personal notes, together with drawings, Lear gathered his impressions on the Italian way of life, folk traditions, and the beauty of the ancient monuments. Of particular interest to Lear was the Abruzzo, which he visited in 1843, through the Marsica (Celano, Avezzano, Alba Fucens, Trasacco) and the plateau of Cinque Miglia (Castel di Sangro and Alfedena), by an old sheep track of the shepherds. Lear drew a sketch of the medieval village of Albe with Mount Sirente, and described the medieval village of Celano, with the castle of Piccolomini dominating the vast plain of Lago Fucino, which was drained a few years later to promote agricultural development. At Castel di Sangro, Lear described the winter stillness of the mountains and the beautiful basilica. Lear primarily played the piano, but he also played the accordion, flute, and small guitar. He composed music for many Romantic and Victorian poems, but was known mostly for his many musical settings of Tennyson's poetry. He published four settings in 1853, five in 1859, and three in 1860. Lear's were the only musical settings that Tennyson approved of. Lear also composed music for many of his nonsense songs, including "The Owl and the Pussy-cat," but only two of the scores have survived, the music for "The Courtship of the Yonghy-Bonghy-Bò" and "The Pelican Chorus". While he never played professionally, he did perform his own nonsense songs and his settings of others' poetry at countless social gatherings, sometimes adding his own lyrics (as with the song "The Nervous Family"), and sometimes replacing serious lyrics with nursery rhymes. Lear's most fervent and painful friendship was with Franklin Lushington. He met the young barrister in Malta in 1849 and then toured southern Greece with him. Lear developed an infatuation for him that Lushington did not wholly reciprocate. Although they remained friends for almost forty years, until Lear's death, the disparity of their feelings constantly tormented Lear. Indeed, Lear's attempts at male companionship were not always successful; the very intensity of Lear's affections may have doomed these relationships. The closest he came to marriage was two proposals, both to the same woman 46 years his junior, which were not accepted. For companions, he relied instead on friends and correspondents, and especially, during later life, on his Albanian Souliote chef, Giorgis, a faithful friend and, as Lear complained, a thoroughly unsatisfactory chef. Another trusted companion in San Remo was his cat, Foss, who died in 1886 and was buried with some ceremony in a garden at Villa Tennyson. Lear travelled widely throughout his life and eventually settled in San Remo, on his beloved Mediterranean coast, in the 1870s, at a villa he named "Villa Tennyson." Lear was known to introduce himself with a long pseudonym: "Mr Abebika kratoponoko Prizzikalo Kattefello Ablegorabalus Ableborinto phashyph" or "Chakonoton the Cozovex Dossi Fossi Sini Tomentilla Coronilla Polentilla Battledore & Shuttlecock Derry down Derry Dumps" which he based on "Aldiborontiphoskyphorniostikos". After a long decline in his health, Lear died at his villa in 1888 of heart disease, from which he had suffered since at least 1870. Lear's funeral was said to be a sad, lonely affair by the wife of Dr. Hassall, Lear's physician, none of Lear's many lifelong friends being able to attend. Lear is buried in the Cemetery Foce in San Remo. On his headstone are inscribed these lines about Mount Tomohrit (in Albania) from Tennyson's poem "To E.L. [Edward Lear], On His Travels in Greece": The centenary of his death was marked in Britain with a set of Royal Mail stamps in 1988 and an exhibition at the Royal Academy. Lear's birthplace area is now marked with a plaque at Bowman's Mews, Islington, in London, and his bicentenary during 2012 was celebrated with a variety of events, exhibitions and lectures in venues across the world including an International Owl and Pussycat Day on his birth anniversary. In 1846 Lear published "," a volume of limericks which went through three editions and helped popularize the form and the genre of literary nonsense. In 1871 he published "Nonsense Songs, Stories, Botany and Alphabets", which included his most famous nonsense song, "The Owl and the Pussycat," which he wrote for the children of his patron Edward Stanley, 13th Earl of Derby. Many other works followed. Lear's nonsense books were quite popular during his lifetime, but a rumour developed that "Edward Lear" was merely a pseudonym, and the books' true author was the man to whom Lear had dedicated the works, his patron the Earl of Derby. Promoters of this rumour offered as evidence that both men were named Edward, and that "Lear" is an anagram of "Earl." Lear's nonsense works are distinguished by a facility of verbal invention and a poet's delight in the sounds of words, both real and imaginary. A stuffed rhinoceros becomes a "diaphanous doorscraper." A "blue Boss-Woss" plunges into "a perpendicular, spicular, orbicular, quadrangular, circular depth of soft mud." His heroes are Quangle-Wangles, Pobbles, and Jumblies. One of his most famous verbal inventions, the phrase "runcible spoon," occurs in the closing lines of "The Owl and the Pussycat," and is now found in many English dictionaries: Though famous for his neologisms, Lear used a number of other devices in his works in order to defy reader expectations. For example, "Cold Are the Crabs" conforms to the sonnet tradition until its dramatically foreshortened last line. Today, limericks are invariably typeset as five lines. Lear's limericks, however, were published in a variety of formats; it appears that Lear wrote them in manuscript in as many lines as there was room for beneath the picture. For the first three editions, most are typeset as, respectively, two, five, and three lines. The cover of one edition bears an entire limerick typeset in two lines: In Lear's limericks, the first and last lines usually end with the same word rather than rhyming. For the most part they are truly nonsensical and devoid of any punch line or point. They are completely free of the bawdiness with which the verse form is now associated. A typical thematic element is the presence of a callous and critical "they." An example of a typical Lear limerick: Lear's self-description in verse, "How Pleasant to know Mr. Lear," ends with this stanza, a reference to his own mortality: Five of Lear's limericks from the Book of Nonsense, in the 1946 Italian translation by Carlo Izzo, were set to music for choir a cappella by Goffredo Petrassi in 1952. Edward Lear has been played in radio dramas by Andrew Sachs in "The Need for Nonsense" by Julia Blackburn (BBC Radio 4, 9 February 2009) and by Derek Jacobi in "By the Coast of Coromandel" by Lavinia Murray (BBC Radio 4, 21 December 2011).
https://en.wikipedia.org/wiki?curid=10408
Eve Arden Eve Arden (born Eunice Mary Quedens, April 30, 1908 – November 12, 1990) was an American film, radio, stage, and television actress, and comedian. She performed in leading and supporting roles for nearly six decades. Beginning her film career in 1929 and on Broadway in the early 1930s, Arden's first major role was in the RKO Radio Pictures drama "Stage Door" (1937) opposite Katharine Hepburn, followed by roles in the comedies "Having Wonderful Time" (1938) and the Marx Brothers' "At the Circus" (1939). She received an Academy Award nomination for Best Supporting Actress for her role in "Mildred Pierce" (1945). Later in her career, Arden played the sardonic but engaging high school teacher in the eponymous "Our Miss Brooks", for which she won the first Primetime Emmy Award for Outstanding Lead Actress in a Drama Series. She also played the school principal in the musicals "Grease" (1978) and "Grease 2" (1982). Eve Arden was born Eunice Mary Quedens on April 30, 1908, in Mill Valley, California, to Charles Peter Quedens (son of Charles Henry Augustus and Meta (née Dierks) Quedens) and Lucille (née Frank) Quedens, (daughter of Bernard and Louisa (née Mertens) Frank) both of German descent. Lucille, a milliner, divorced Charles over his gambling, and went into business for herself. Although not Roman Catholic, young Eunice was sent to a Dominican convent school in San Rafael, California. She then attended Tamalpais High School, a public high school in Mill Valley, until age 16. After leaving school, she joined the stock theater company of Henry "Terry" Duffy. She made her film debut under her real name in the backstage musical "Song of Love" (1929), as a wisecracking, homewrecking showgirl who becomes a rival to the film's star, singer Belle Baker. The film was one of Columbia Pictures' earliest successes. In 1933, she relocated to New York City, where she had supporting parts in multiple Broadway stage productions. In 1934, she was cast in the "Ziegfeld Follies" revue, the first role where she was credited as Eve Arden. When she was told to adopt a stage name for the show, Arden looked at her cosmetics and "stole my first name from Evening in Paris, and the second from Elizabeth Arden". Between 1934 and 1941, she appeared in Broadway productions of "Parade", "Very Warm for May", "Two for the Show", and "Let's Face It!". Arden's film career began in earnest in 1937 when she signed a contract with RKO Radio Pictures and appeared in the films "Oh Doctor" and "Stage Door". Her "Stage Door" portrayal of a fast-talking, witty supporting character gained Arden considerable notice and was a template for many of Arden's future roles. In 1938, she played a supporting part in the comedy "Having Wonderful Time", starring Ginger Rogers and Lucille Ball. This was followed by roles in the crime film "The Forgotten Woman" (1939), and the Marx Brothers comedy "At the Circus" (1939), a role that required her to perform acrobatics. In 1940, she appeared opposite Clark Gable in "Comrade X", followed by the drama "Manpower" (1941) opposite Marlene Dietrich. She also had a supporting part in the Red Skelton comedy "Whistling in the Dark" (1941) and the romantic comedy "Obliging Young Lady" (1942). Her many memorable screen roles include a supporting role as Joan Crawford's wise-cracking friend in "Mildred Pierce" (1945), for which she received an Academy Award nomination as Best Supporting Actress; and as James Stewart's wistful secretary in Otto Preminger's mystery "Anatomy of a Murder" (1959) (which also featured her husband, Brooks West). In 1946, exhibitors voted her the sixth-most promising "star of tomorrow". Arden became familiar to a new generation of filmgoers when she played Principal McGee in "Grease" (1978) and "Grease 2 (1982)". Arden was known for her deadpan comedic delivery. Arden's ability with witty scripts made her a natural talent for radio. She was a regular on Danny Kaye's short-lived but memorably zany comedy-variety show in 1946, which also featured swing bandleader Harry James and gravel-voiced character actor-comedian Lionel Stander. Kaye's show lasted one season, but Arden's comic talent led to her best-known role, that of Madison High School English teacher Connie Brooks in "Our Miss Brooks". Arden portrayed the character on radio from 1948 to 1957, in a television version of the program from 1952 to 1956, and in a 1956 feature film. Her character clashed with the school's principal, Osgood Conklin (played by Gale Gordon) and nursed an unrequited crush on fellow teacher Philip Boynton (played originally by future film star Jeff Chandler; and later on radio and TV by Robert Rockwell). Except for Chandler, the entire radio cast of Arden, Gordon, Richard Crenna (Walter Denton), Robert Rockwell (Mr. Philip Boynton), Gloria McMillan (Harriet Conklin) and Jane Morgan (landlady Margaret Davis) played the same roles on TV. Arden's portrayal of Miss Brooks was so popular that she was made an honorary member of the National Education Association, received a 1952 award from the Teachers College of Connecticut's Alumni Association "for humanizing the American teacher", and even received teaching job offers. Her wisecracking, deadpan character ultimately became her public persona as a comedienne. She won a listeners' poll by "Radio Mirror" magazine as the top-ranking comedienne of 1948–1949, receiving her award at the end of an "Our Miss Brooks" broadcast that March. "I'm certainly going to try in the coming months to merit the honor you've bestowed upon me, because I understand that if I win this (award) two years in a row, I get to keep Mr. Boynton," she joked. She was also a hit with the critics: A winter 1949 poll of newspaper and magazine radio editors by "Motion Picture Daily" named her the year's best radio comedienne. Arden had a very brief guest appearance in a 1955 "I Love Lucy" episode titled "L.A. at Last", where she played herself. While awaiting their food at the Brown Derby, Lucy Ricardo (Lucille Ball) and Ethel Mertz (Vivian Vance) argue over whether a certain portrait on a nearby wall is Shelley Winters or Judy Holliday. Lucy urges Ethel to ask a lady occupying the next booth, who turns and replies, "Neither. That's Eve Arden." As Ethel realizes she just spoke to Arden herself, Arden passes Lucy and Ethel's table to leave the restaurant while the pair gawk. Desilu Productions, jointly owned by Desi Arnaz and Ball during their marriage, was the production company for the "Our Miss Brooks" television show, filmed during the same years as "I Love Lucy". Ball and Arden met when they costarred in the film "Stage Door" in 1937. Ball, according to numerous radio historians, suggested Arden for "Our Miss Brooks" after Shirley Booth auditioned for but failed to land the role and Ball—committed at the time to "My Favorite Husband"—could not. Arden tried another series in the fall of 1957, "The Eve Arden Show", but it was canceled in spring of 1958 after 26 episodes. In 1966, she played Nurse Kelton in an episode of "Bewitched". She later costarred with Kaye Ballard as her neighbor and in-law, Eve Hubbard, in the 1967–1969 situation comedy "The Mothers-in-Law", produced by Arnaz after the dissolution of Desilu Productions. In her later career, Arden made appearances on such television shows as "Bewitched", "Alice", "Maude, Hart to Hart", and "Falcon Crest". In 1985, she appeared as the wicked stepmother in the "Faerie Tale Theatre" production of "Cinderella". Arden was one of many actresses to take on the title roles in "Hello, Dolly!" and "Auntie Mame" in the 1960s; in 1967, she won the Sarah Siddons Award for her work in Chicago theatre. Arden was cast in 1983 as the leading lady in what was to be her Broadway comeback in "Moose Murders", but she wisely withdrew and was replaced with the much younger Holland Taylor after one preview performance, citing "artistic differences". The show went on to open and close on the same night, becoming known as one of the most legendary flops in Broadway history. Arden published an autobiography, "The Three Phases of Eve", in 1985. In addition to her Academy Award nomination, Arden has two stars on the Hollywood Walk of Fame: Radio and Television (see List of stars on the Hollywood Walk of Fame for addresses). She was inducted into the National Radio Hall of Fame in 1995. Arden was married to Ned Bergen from 1939 to 1947, and to actor Brooks West (1916-1984), from 1952 until his death in 1984 from a heart ailment, aged 67. West and she had four children; all but the youngest were adopted. All four survived their parents. On November 12, 1990, Arden died from cardiac arrest and arteriosclerotic heart disease, aged 82, at her home, according to her death certificate. She is interred in the Westwood Village Memorial Park Cemetery, Westwood, Los Angeles, California.
https://en.wikipedia.org/wiki?curid=10409
Elementary function In mathematics, an elementary function is a function of a single variable composed of particular simple functions. Elementary functions are typically defined as a sum, product, and/or composition of finitely many polynomials, rational functions, trigonometric and exponential functions, and their inverse functions (including arcsin, log, "x"1/"n"). Elementary functions were introduced by Joseph Liouville in a series of papers from 1833 to 1841. An algebraic treatment of elementary functions was started by Joseph Fels Ritt in the 1930s. The elementary functions (of ) include: Some elementary functions, such as roots, logarithms, or inverse trigonometric functions, are not entire functions and may be multivalued. Examples of elementary functions include: The last function is equal to formula_13, the inverse cosine, in the entire complex plane. All monomials, polynomials and rational functions are elementary. Also, the absolute value function, for real formula_2, is also elementary as it can be expressed as the composition of a power and root of formula_2: formula_16. An example of a function that is "not" elementary is the error function a fact that may not be immediately obvious, but can be proven using the Risch algorithm. It follows directly from the definition that the set of elementary functions is closed under arithmetic operations and composition. The elementary functions are closed under differentiation. They are not closed under limits and infinite sums. Importantly, the elementary functions are closed under integration, as shown by Liouville's theorem, see Nonelementary integral. The Liouvillian functions are defined as the elementary functions and, recursively, the integrals of the Liouvillian functions. The mathematical definition of an elementary function, or a function in elementary form, is considered in the context of differential algebra. A differential algebra is an algebra with the extra operation of derivation (algebraic version of differentiation). Using the derivation operation new equations can be written and their solutions used in extensions of the algebra. By starting with the field of rational functions, two special types of transcendental extensions (the logarithm and the exponential) can be added to the field building a tower containing elementary functions. A differential field "F" is a field "F"0 (rational functions over the rationals Q for example) together with a derivation map "u" → ∂"u". (Here ∂"u" is a new function. Sometimes the notation "u"′ is used.) The derivation captures the properties of differentiation, so that for any two elements of the base field, the derivation is linear and satisfies the Leibniz product rule An element "h" is a constant if "∂h = 0". If the base field is over the rationals, care must be taken when extending the field to add the needed transcendental constants. A function "u" of a differential extension "F"["u"] of a differential field "F" is an elementary function over "F" if the function "u" (see also Liouville's theorem)
https://en.wikipedia.org/wiki?curid=10412
Enchiridion of Epictetus The Enchiridion or Handbook of Epictetus (, "Enkheirídion Epiktḗtou") is a short manual of Stoic ethical advice compiled by Arrian, a 2nd-century disciple of the Greek philosopher Epictetus. Although the content is mostly derived from the "Discourses of Epictetus", it is not a summary of the "Discourses" but rather a compilation of practical precepts. Eschewing metaphysics, Arrian focuses his attention on Epictetus's work applying philosophy to daily life. The book is thus a manual to show the way to achieve mental freedom and happiness in all circumstances. The "Enchiridion" was well-known in the ancient world, and in the medieval period it was specially adapted for use in Greek-speaking monasteries. In the 15th-century it was translated into Latin, and then, with the advent of printing, into multiple European languages. It reached the height of popularity in the 17th century, in parallel with the Neostoicism movement. The word "encheiridion" () is an adjective meaning "in the hand" or "ready to hand". The word sometimes meant a handy sword, or dagger, but coupled with the word "book" ("biblion", ) it means a handy book or hand-book. Epictetus in the "Discourses" often speaks of principles which his pupils should have "ready to hand" (). Common English translations of the title are Manual or Handbook. The work consists of fifty-three short chapters typically consisting of a paragraph or two. It was compiled some time in the early 2nd-century. The 6th-century philosopher Simplicius, in his "Commentary" on the work, refers to a letter written by Arrian which prefaced the text. In this letter Arrian stated that the "Enchiridion" was selected from the "Discourses of Epictetus" according to what he considered to be most useful, most necessary, and most adapted to move people's minds. Around half of the material in the "Enchiridion" has been shown to have been derived from the surviving four books of "Discourses" but variously modified. Other parts are presumed to be derived from the lost "Discourses". Some chapters appear to be reformulations of ideas which appear throughout the "Discourses". There are some puzzles concerning the inclusion of two chapters. Chapter 29 is practically word for word identical with "Discourse" iii. 15. Since it was omitted in one of the early Christian editions ("Par"), and not commented on by Simplicius, it may not have been in the original edition. Chapter 33 consists of a list of moral instructions, which are "not obviously related to Epictetus' normal Stoic framework." The current division of the work into fifty-three chapters was first adopted by Johann Schweighäuser in his 1798 edition; earlier editions tended to divide the text into more chapters (especially splitting chapter 33). Gerard Boter in his 1999 critical edition keeps Schweighäuser's fifty-three chapters but splits chapters 5, 14, 19, and 48 into two parts. The "Enchiridion" appears to be a loosely-structured selection of maxims. In his 6th-century "Commentary", Simplicius divided the text into four distinct sections suggesting a graded approach to philosophy: Chapter 29, which was probably absent from the text used by Simplicius, is a one-page "Discourse" which compares the training needed to become a Stoic with the rigorous approach needed to become an Olympic victor. The "Enchiridion" begins with the statement that "Of things, some depend upon ourselves, others do not depend upon ourselves." So it starts with announcing that the business and concern of the real self is with matters subject to its own control, uninfluenced by external chance or change. Epictetus makes a sharp distinction between our own internal world of mental benefits and harms, and the external world beyond our control. Freedom is to wish for nothing which is not up to ourselves. When we are tried by misfortune we should never let our suffering overwhelm our sense of inward mastery and freedom. A constant vigilance is required, and one should never relax attention to one's reason, for it is judgements, not things, which disturb people. Reason is the decisive principle in everything. Thus we must exercise our power of assent over impressions, and wish for nothing nor avoid anything that is up to other people. To a large extent the "Enchiridion" suppresses many of the more amiable aspects of Epictetus which can be found in the "Discourses", but this reflects the nature of the compilation. Unlike the "Discourses" which seeks to encourage the student through argument and logic, the "Enchiridion" largely consists of a set of rules to follow. The work is built on the conception that the wise person, by the aid of philosophy, may reap benefit from every experience in life. With proper training the student can flourish in adverse situations as well as favourable ones. The human spirit has capacities as yet undeveloped, but which it is for our good to develop. Thus the book is a manual on how to make progress towards what is necessary and sufficient for happiness. Epictetus makes a vivid use of imagery, and analogies include life depicted as: a ship's voyage (Ch. 7), an inn (Ch. 11), a banquet (Chs. 15, 36), and acting in a play (Ch. 17, 37). He takes many examples from everyday life, including: a broken jug (Ch. 3), a trip to the baths (Chs. 4, 43), his own lameness (Ch. 9), the loss of a child (Ch. 11), and the price of lettuce (Ch. 25). For many centuries, the "Enchiridion" maintained its authority both with Pagans and Christians. Simplicius of Cilicia wrote a commentary upon it in the 6th century, and in the Byzantine era Christian writers wrote paraphrases of it. Over one hundred manuscripts of the "Enchiridion" survive. The oldest extant manuscripts of the authentic "Enchiridion" date from the 14th century, but the oldest Christianised ones date from the 10th and 11th centuries, perhaps indicating the Byzantine world's preference for the Christian versions. The "Enchiridion" was first translated into Latin by Niccolò Perotti in 1450, and then by Angelo Poliziano in 1479. The first printed edition ("editio princeps") was Poliziano's Latin translation published in 1497. The original Greek was first published (somewhat abbreviated) with Simplicius's "Commentary" in 1528. The edition published by Johann Schweighäuser in 1798 was the major edition for the next two-hundred years. A critical edition was produced by Gerard Boter in 1999. The separate editions and translations of the "Enchiridion" are very many. The "Enchiridion" reached its height of popularity in the period 1550–1750. It was translated into most European languages, and there were multiple translations in English, French, and German. The first English translation was by James Sandford in 1567 (a translation of a French version) and this was followed by a translation (from the Greek) by John Healey in 1610. The "Enchiridion" was even partly translated into Chinese by the Jesuit missionary Matteo Ricci. The popularity of the work was assisted by the Neostoicism movement initiated by Justus Lipsius in the 16th century. Another Neostoic, Guillaume du Vair, translated the book into French in 1586 and popularised it in his "La Philosophie morale des Stoiques". In the English-speaking world it was particularly well-known in the 17th-century: at that time it was the "Enchiridion" rather than the "Discourses" which was usually read. It was among the books John Harvard bequeathed to the newly-founded Harvard College in 1638. The work, being written in a clear distinct style, made it accessible to readers with no formal training in philosophy, and there was a wide readership among women in England. The "Enchiridion" was a common school text in Scotland during the Scottish Enlightenment—Adam Smith had a 1670 edition in his library, acquired as a schoolboy. At the end of the 18th-century the "Enchiridion" is attested in the personal libraries of Benjamin Franklin and Thomas Jefferson. In the 6th-century the Neoplatonist philosopher Simplicius wrote a huge commentary on the "Enchiridion", which is more than ten times the bulk of the original text. Chapter after chapter of the "Enchiridion" is dissected, discussed, and its lessons drawn out with a certain laboriousness. Simplicius' commentary offers a distinctly Platonist vision of the world, one which is often at odds with the Stoic content of the "Enchiridion". Sometimes Simplicius exceeds the scope of a commentary; thus his commentary on "Enchiridion" 27 (Simplicius ch. 35) becomes a refutation of Manichaeism. The "Commentary" enjoyed its own period of popularity in the 17th and 18th centuries. An English translation by George Stanhope in 1694 ran through four editions in the early 1700s. Edward Gibbon remarked in his "Decline and Fall of the Roman Empire" that Simplicius' "Commentary" on Epictetus "is preserved in the library of nations, as a classic book" unlike the commentaries on Aristotle "which have passed away with the fashion of the times." The "Enchiridion" was adapted three different times by Greek Christian writers. The oldest manuscript, "Paraphrasis Christiana" ("Par"), dates to the 10th century. Another manuscript, falsely ascribed to Nilus ("Nil"), dates to the 11th century. A third manuscript, Vaticanus gr. 2231 ("Vat"), dates to the 14th century. It is not known when the original versions of these manuscripts were first made. These guides served as a rule and guide for monastic life. The most obvious changes are in the use of proper names: thus the name Socrates is sometimes changed to Paul. All three texts follow the "Enchiridion" quite closely, although the "Par" manuscript is more heavily modified: adding or omitting words, abridging or expanding passages, and occasionally inventing new passages. In the 17th century the German monk Matthias Mittner did something similar, compiling a guide on mental tranquillity for the Carthusian Order by taking the first thirty-five of his fifty precepts from the "Enchiridion". a. Gerard Boter in his 1999 critical edition catalogues 59 extant manuscripts of the "Encheiridion" proper, and another 27 manuscripts of Simplicius' "Commentary" which contain the "Encheiridion" as "lemmata" (headings). He also lists 37 Christianised manuscripts, (24 "Par", 12 "Nil", 1 "Vat"). Cf.
https://en.wikipedia.org/wiki?curid=10413
Celestines The Celestines were a Roman Catholic monastic order, a branch of the Benedictines, founded in 1244. At the foundation of the new rule, they were called Hermits of St Damiano, or Moronites (or Murronites), and did not assume the appellation of Celestines until after the election of their founder, Peter of Morone (Pietro Murrone), to the Papacy as Celestine V. They used the post-nominal initials O.S.B. Cel. The order was absorbed by Order of the Most Holy Annunciation from 1778 by order of Pius VI in 1776. In 1810 the last Celestines were transferred. The fame of the holy life and the austerities practised by Pietro Morone in his solitude on the Mountain of Majella, near Sulmona, attracted many visitors, several of whom were moved to remain and share his mode of life. They built a small convent on the spot inhabited by the holy hermit, which became too small for the accommodation of those who came to share their life of privations. Peter of Morone (later Pope Celestine V), their founder, built a number of other small oratories in that neighborhood. About the year 1254, Peter of Morone gave the order a rule formulated in accordance with his own practices. In 1264 the new institution was approved as a branch of the Benedictines by Urban IV; however, the next pope Pope Gregory X had commanded that all orders founded since the prior Lateran Council should not be further multiplied. Hearing a rumor that the order was to be suppressed, the reclusive Peter traveled to Lyon, where the Pope was holding a council. There he persuaded Gregory to approve his new order, making it a branch of the Benedictines and following the rule of Saint Benedict, but adding to it additional severities and privations. Gregory took it under the Papal protection, assured to it the possession of all property it might acquire, and endowed it with exemption from the authority of the ordinary. Nothing more was needed to ensure the rapid spread of the new association and Peter the hermit of Morone lived to see himself "Superior-General" to thirty-six monasteries and more than six hundred monks. As soon as he had seen his new order thus consolidated he gave up the government of it to a certain Robert, and retired once again to an even more remote site to devote himself to solitary penance and prayer. Shortly afterwards, in a chapter of the order held in 1293, the original monastery of Majella being judged to be too desolate and exposed to too rigorous a climate, it was decided that the Abbey of the Holy Spirit at Monte Morrone, located in Sulmona, should be the headquarters of the order and the residence of the General-Superior, where it continued for centuries. The next year Peter of Morrone, despite his reluctance, was elected Pope by the name of Celestine V. From there on, the order he had founded took the name of Celestines. During his short reign as Pope, the former hermit confirmed the rule of the order, which he had himself composed, and conferred on the society a variety of special graces and privileges. In the only creation of cardinals promoted by him, among the twelve raised to the purple, there were two monks of his order. He also visited personally the Benedictine monastery on Monte Cassino, where he persuaded the monks to accept his more rigorous rule. He sent fifty monks of his order to introduce it, who remained there, however, for only a few months. After the death of the founder the order was favoured and privileged by Benedict XI, and rapidly spread through Italy, Germany, Flanders, and France, where they were received by Philip the Fair in 1300. The administration of the order was carried on somewhat after the pattern of Cluny, that is all monasteries were subject to the Abbey of the Holy Ghost at Sulmona, and these dependent houses were divided into provinces. The Celestines had ninety-six houses in Italy, twenty-one in France, and a few in Germany. Subsequently, the French Celestines, with the consent of the Italian superiors of the order, and of Pope Martin V in 1427, obtained the privilege of making new constitutions for themselves, which they did in the 17th century in a series of regulations accepted by the provincial chapter in 1667. At that time the French congregation of the order was composed of twenty-one monasteries, the head of which was that of Paris, and was governed by a Provincial with the authority of General. Paul V was a notable benefactor of the order. The order became extinct in the eighteenth century. According to their special constitutions the Celestines were bound to say matins in the choir at two o'clock in the morning, and always to abstain from eating meat, save in illness. The distinct rules of their order with regard to fasting are numerous, but not more severe than those of similar congregations, though much more so than is required by the old Benedictine rule. In reading their minute directions for divers degrees of abstinence on various days, it is impossible to avoid being struck by the conviction that the great object of the framers of these rules was the general purpose of ensuring an ascetic mode of life. The Celestines wore a white woollen cassock bound with a linen band, and a leathern girdle of the same colour, with a scapular unattached to the body of the dress, and a black hood. It was not permitted to them to wear any shirt save of serge. Their dress in short was very like that of the Cistercians. But it is a tradition in the order that in the time of the founder they wore a coarse brown cloth. The church and monastery of San Pietro in Montorio originally belonged to the Celestines in Rome; but they were turned out of it by Sixtus IV to make way for Franciscans, receiving from the Pope in exchange the Church of St Eusebius of Vercelli with the adjacent mansion for a monastery.
https://en.wikipedia.org/wiki?curid=6537
Cessna The Cessna Aircraft Company () was an American general aviation aircraft manufacturing corporation headquartered in Wichita, Kansas. Cessna produced small, piston-powered aircraft, as well as business jets. For many years the company was one of the highest-volume producers of general aviation aircraft in the world. The company was founded in 1927. It was purchased by General Dynamics in 1985, then by Textron, Inc., in 1992. In March 2014, when Textron purchased the Beechcraft and Hawker Aircraft businesses, Cessna ceased operations as a subsidiary company and joined the others as one of the three distinct brands produced by Textron Aviation. Clyde Cessna, a farmer in Rago, Kansas, built his own aircraft and flew it in June 1911. He was the first person to do so between the Mississippi River and the Rocky Mountains. Cessna started his wood-and-fabric aircraft ventures in Enid, Oklahoma, testing many of his early planes on the salt flats. When bankers in Enid refused to lend him more money to build his planes, he moved to Wichita. Cessna Aircraft was formed when Clyde Cessna and Victor Roos became partners in the Cessna-Roos Aircraft Company in 1927. Roos resigned just one month into the partnership, selling back his interest to Cessna. Shortly afterward, Roos's name was dropped from the company name. The Cessna DC-6 earned certification on the same day as the stock market crash of 1929, October 29, 1929. In 1932, the Cessna Aircraft Company closed its doors due to the Great Depression. However, the Cessna CR-3 custom racer made its first flight in 1933. The plane won the 1933 American Air Race in Chicago and later set a new world speed record for engines smaller than 500 cubic inches by averaging . Cessna's nephews, brothers Dwane and Dwight Wallace, bought the company from Cessna in 1934. They reopened it and began the process of building it into what would become a global success. The Cessna C-37 was introduced in 1937 as Cessna's first seaplane when equipped with Edo floats. In 1940, Cessna received their largest order to date, when they signed a contract with the U.S. Army for 33 specially equipped Cessna T-50s. Later in 1940, the Royal Canadian Air Force placed an order for 180 T-50s. Cessna returned to commercial production in 1946, after the revocation of wartime production restrictions (L-48), with the release of the Model 120 and Model 140. The approach was to introduce a new line of all-metal aircraft that used production tools, dies and jigs rather than the hand-built process tube-and-fabric construction used before the war. The Model 140 was named by the US Flight Instructors Association as the "Outstanding Plane of the Year", in 1948. Cessna's first helicopter, the Cessna CH-1, received FAA type certification in 1955. Cessna introduced the Cessna 172 in 1956. It became the most produced airplane in history. In 1960, Cessna affiliated itself with Reims Aviation of Reims, France. In 1963, Cessna produced its 50,000th airplane, a Cessna 172. Cessna's first business jet, the Cessna Citation I, performed its maiden flight on September 15, 1969. Cessna produced its 100,000th single-engine airplane in 1975. In 1985, Cessna ceased to be an independent company. It was purchased by General Dynamics Corporation and became a wholly owned subsidiary. Production of the Cessna Caravan began. General Dynamics in turn sold Cessna to Textron in 1992. Late in 2007, Cessna purchased the bankrupt Columbia Aircraft company for US$26.4M and would continue production of the Columbia 350 and 400 as the Cessna 350 and Cessna 400 at the Columbia factory in Bend, Oregon. However, production of both aircraft had ended by 2018. On November 27, 2007, Cessna announced the then-new Cessna 162 would be built in the People's Republic of China by Shenyang Aircraft Corporation, which is a subsidiary of the China Aviation Industry Corporation I (AVIC I), a Chinese government-owned consortium of aircraft manufacturers. Cessna reported that the decision was made to save money and also that the company had no more plant capacity in the United States at the time. Cessna received much negative feedback for this decision, with complaints centering on the recent quality problems with Chinese production of other consumer products, China's human rights record, exporting of jobs and China's less than friendly political relationship with the United States. The customer backlash surprised Cessna and resulted in a company public relations campaign. In early 2009, the company attracted further criticism for continuing plans to build the 162 in China while laying off large numbers of workers in the United States. In the end, the Cessna 162 was not a commercial success and only a small number were delivered before production was cancelled. The company's business suffered notably during the late-2000s recession, laying off more than half its workforce between January 2009 and September 2010. On November 4, 2008, Cessna's parent company, Textron, indicated that Citation production would be reduced from the original 2009 target of 535 "due to continued softening in the global economic environment" and that this would result in an undetermined number of lay-offs at Cessna. On November 8, 2008, at the Aircraft Owners and Pilots Association (AOPA) Expo, CEO Jack Pelton indicated that sales of Cessna aircraft to individual buyers had fallen, but piston and turboprop sales to businesses had not. "While the economic slowdown has created a difficult business environment, we are encouraged by brisk activity from new and existing propeller fleet operators placing almost 200 orders for 2009 production aircraft," Pelton stated. Beginning in January 2009, a total of 665 jobs were cut at Cessna's Wichita and Bend, Oregon plants. The Cessna factory at Independence, Kansas, which builds the Cessna piston-engined aircraft and the Cessna Mustang, did not see any layoffs, but one third of the workforce at the former Columbia Aircraft facility in Bend was laid off. This included 165 of the 460 employees who built the Cessna 350 and 400. The remaining 500 jobs were eliminated at the main Cessna Wichita plant. In January 2009, the company laid off an additional 2,000 employees, bringing the total to 4,600. The job cuts included 120 at the Bend, Oregon, facility reducing the plant that built the Cessna 350 and 400 to fewer than half the number of workers that it had when Cessna bought it. Other cuts included 200 at the Independence, Kansas, plant that builds the single-engined Cessnas and the Mustang, reducing that facility to 1,300 workers. On April 29, 2009, the company suspended the Citation Columbus program and closed the Bend, Oregon, facility. The Columbus program was finally cancelled in early July 2009. The company reported, "Upon additional analysis of the business jet market related to this product offering, we decided to formally cancel further development of the Citation Columbus". With the 350 and 400 production moving to Kansas, the company indicated that it would lay off 1,600 more workers, including the remaining 150 employees at the Bend plant and up to 700 workers from the Columbus program. In early June 2009, Cessna laid off an additional 700 salaried employees, bringing the total number of lay-offs to 7,600, which was more than half the company's workers at the time. The company closed its three Columbus, Georgia, manufacturing facilities between June 2010 and December 2011. The closures included the new facility that was opened in August 2008 at a cost of US$25M, plus the McCauley Propeller Systems plant. These closures resulted in total job losses of 600 in Georgia. Some of the work was relocated to Cessna's Independence, Kansas, or Mexican facilities. Cessna's parent company, Textron, posted a loss of US$8M in the first quarter of 2010, largely driven by continuing low sales at Cessna, which were down 44%. Half of Cessna's workforce remained laid-off and CEO Jack Pelton stated that he expected the recovery to be long and slow. In September 2010, a further 700 employees were laid off, bringing the total to 8,000 jobs lost. CEO Jack Pelton indicated this round of layoffs was due to a "stalled [and] lackluster economy" and noted that while the number of orders cancelled for jets had been decreasing new orders had not met expectations. Pelton added "our strategy is to defend and protect our current markets while investing in products and services to secure our future, but we can do this only if we succeed in restructuring our processes and reducing our costs." On May 2, 2011, CEO Jack J. Pelton retired. The new CEO, Scott A. Ernest, started on May 31, 2011. Ernest joined Textron after 29 years at General Electric, where he had most recently served as vice president and general manager, global supply chain for GE Aviation. Ernest previously worked for Textron CEO Scott Donnelly when both worked at General Electric. In September 2011, the Federal Aviation Administration (FAA) proposed a US$2.4 million fine against the company for its failure to follow quality assurance requirements while producing fiberglass components at its plant in Chihuahua, Mexico. Excess humidity meant that the parts did not cure correctly and quality assurance did not detect the problems. The failure to follow procedures resulted in the delamination in flight of a section of one Cessna 400's wing skin from the spar while the aircraft was being flown by an FAA test pilot. The aircraft was landed safely. The FAA also discovered 82 other aircraft parts that had been incorrectly made and not detected by the company's quality assurance. The investigation resulted in an emergency Airworthiness Directive that affected 13 Cessna 400s. Since March 2012, Cessna has been pursuing building business jets in China as part of a joint venture with Aviation Industry Corporation of China (AVIC). The company stated that it intends to eventually build all aircraft models in China, saying "The agreements together pave the way for a range of business jets, utility single-engine turboprops and single-engine piston aircraft to be manufactured and certified in China." In late April 2012, the company added 150 workers in Wichita as a result of anticipated increased demand for aircraft production. Overall, they have cut more than 6000 jobs in the Wichita plant since 2009. In March 2014, Cessna ceased operations as a company and instead became a brand of Textron Aviation. During the 1950s and 1960s, Cessna's marketing department followed the lead of Detroit automakers and came up with many unique marketing terms in an effort to differentiate its product line from their competitions'. Other manufacturers and the aviation press widely ridiculed and spoofed many of the marketing terms, but Cessna built and sold more aircraft than any other manufacturer during the boom years of the 1960s and 1970s. Generally, the names of Cessna models do not follow a theme, but there is logic to the numbering: the 100 series are the light singles, the 200s are the heftier, the 300s are light to medium twins, the 400s have “wide oval” cabin-class accommodation and the 500s are jets. Many Cessna models have names starting with C for the sake of alliteration (e.g. Citation, Crusader, Chancellor). Cessna marketing terminology includes: As of July 2018, Textron Aviation is producing the following Cessna models:
https://en.wikipedia.org/wiki?curid=6539
Czesław Miłosz Czesław Miłosz (, also , ; 30 June 1911 – 14 August 2004) was a Polish-American poet, prose writer, translator, and diplomat. Regarded as one of the great poets of the 20th century, he won the 1980 Nobel Prize in Literature. In its citation, the Swedish Academy called Miłosz a writer who "voices man's exposed condition in a world of severe conflicts". Miłosz survived the German occupation of Warsaw during World War II and became a cultural attaché for the Polish government during the postwar period. When communist authorities threatened his safety, he defected to France and ultimately chose exile in the United States, where he became a professor at the University of California, Berkeley. His poetry—particularly about his wartime experience—and his appraisal of Stalinism in a prose book, "The Captive Mind", brought him renown as a leading "émigré" artist and intellectual. Throughout his life and work, Miłosz tackled questions of morality, politics, history, and faith. As a translator, he introduced Western works to a Polish audience, and as a scholar and editor, he championed a greater awareness of Slavic literature in the West. Faith played a role in his work as he explored his Catholicism and personal experience. Miłosz died in Kraków, Poland, in 2004. He is interred in Skałka, a church known in Poland as a place of honor for distinguished Poles. Czesław Miłosz was born on 30 June 1911, in the village of Šeteniai (), Kovno Governorate, Russian Empire (now Kėdainiai district, Kaunas County, Lithuania). He was the son of Aleksander Miłosz (1883–1959), a Polish civil engineer, and his wife, Weronika (née Kunat; 1887–1945). Miłosz was born into a prominent family. On his mother's side, his grandfather was Zygmunt Kunat, a descendant of a Polish family that traced its lineage to the 13th century and owned an estate in Krasnogruda (in present-day Poland). Having studied agriculture in Warsaw, Zygmunt settled in Šeteniai after marrying Miłosz's grandmother, Jozefa, a descendant of the noble Syruć family, which was of Lithuanian origin. One of her ancestors, Szymon Syruć, had been personal secretary to Stanisław I, King of Poland and Grand Duke of Lithuania. Miłosz's paternal grandfather, Artur Miłosz, was also from a noble family and fought in the 1863 January Uprising for Polish independence. Miłosz's grandmother, Stanisława, was a doctor's daughter from Riga, Latvia, and a member of the German/Polish von Mohl family. The Miłosz estate was in Serbiny, a name that Miłosz's biographer Andrzej Franaszek has suggested could indicate Serbian origin; it is possible the Miłosz family originated in Serbia and settled in present-day Lithuania after being expelled from Germany centuries earlier. Miłosz's father was born and educated in Riga. Miłosz's mother was born in Šeteniai and educated in Kraków. Despite this noble lineage, Miłosz's childhood on his maternal grandfather's estate in Šeteniai lacked the trappings of wealth or the customs of the upper class. He memorialized his childhood in a 1955 novel, "The Issa Valley", and a 1959 memoir, "Native Realm." In these works, he described the influence of his Catholic grandmother, Jozefa, his burgeoning love for literature, and his early awareness, as a member of the Polish gentry in Lithuania, of the role of class in society. Miłosz's early years were marked by upheaval. When his father was hired to work on infrastructure projects in Siberia, he and his mother traveled to be with him. After World War I broke out in 1914, Miłosz's father was conscripted into the Russian army, tasked with engineering roads and bridges for troop movements. Miłosz and his mother were sheltered in Wilno when the German army captured it in 1915. Afterward, they once again joined Miłosz's father, following him as the front moved further into Russia, where, in 1917, Miłosz's brother, Andrzej, was born. Finally, after moving through Estonia and Latvia, the family returned to Šeteniai in 1918. But the Polish–Soviet War broke out in 1919, during which Miłosz's father was involved in a failed attempt to incorporate the newly independent Lithuania into the Second Polish Republic, resulting in his expulsion from Lithuania and the family's move to Wilno, which had become part of Poland after the Polish–Lithuanian War of 1920. The Polish-Soviet War continued, forcing the family to move again. At one point during the conflict, Polish soldiers fired at Miłosz and his mother, an episode he recounted in "Native Realm." The family returned to Wilno when the war ended in 1921. Despite the interruptions of wartime wanderings, Miłosz proved to be an exceptional student with a facility for languages. He ultimately learned Polish, Lithuanian, Russian, English, French, and Hebrew. After graduation from Sigismund Augustus Gymnasium in Wilno, he entered Stefan Batory University in 1929 as a law student. While at university, Miłosz joined a student group called The Intellectuals' Club and a student poetry group called Żagary, along with the young poets Jerzy Zagórski, Teodor Bujnicki, Aleksander Rymkiewicz, Jerzy Putrament, and Józef Maśliński. His first published poems appeared in the university's student magazine in 1930. In 1931, he visited Paris, where he first met his distant cousin, Oscar Milosz, a French-language poet of Lithuanian descent who had become a Swedenborgian. Oscar became a mentor and inspiration. Returning to Wilno, Miłosz's early awareness of class difference and sympathy for those less fortunate than himself inspired his defense of Jewish students at the university who were being harassed by an anti-Semitic mob. Stepping between the mob and the Jewish students, Miłosz fended off attacks. One student was killed when a rock was thrown at his head. Miłosz's first volume of poetry, "A Poem on Frozen Time", was published in Polish in 1933. In the same year, he publicly read his poetry at an anti-racist "Poetry of Protest" event in Wilno, occasioned by Hitler's rise to power in Germany. In 1934, he graduated with a law degree, and the poetry group Żagary disbanded. Miłosz relocated to Paris on a scholarship to study for one year and write articles for a newspaper back in Wilno. In Paris, he frequently met with his cousin Oscar. By 1936, he had returned to Wilno, where he worked on literary programs at Radio Wilno. His second poetry collection, "Three Winters", was published that same year, eliciting from one critic a comparison to Adam Mickiewicz. After only one year at Radio Wilno, Miłosz was dismissed due to an accusation that he was a left-wing sympathizer: as a student, he had adopted socialist views from which, by then, he had publicly distanced himself, and he and his boss, Tadeusz Byrski, had produced programming that included performances by Jews and Byelorussians, which angered right-wing nationalists. After Byrski made a trip to the Soviet Union, an anonymous complaint was lodged with the management of Radio Wilno that the station housed a communist cell, and Byrski and Miłosz were dismissed. In summer 1937, Miłosz moved to Warsaw, where he found work at Polish Radio and met his future wife, Janina (née Dłuska; 1909–1986), who was at the time married to another man. Miłosz was in Warsaw when it was bombarded as part of the German invasion of Poland in September 1939. Along with colleagues from Polish Radio, he escaped the city, making his way to Lwów. But when he learned that Janina had remained in Warsaw with her parents, he looked for a way back. The Soviet invasion of Poland thwarted his plans, and, to avoid the incoming Red Army, he fled to Bucharest. There he obtained a Lithuanian identity document and Soviet visa that allowed him to travel by train to Kiev and then Wilno. After the Red Army invaded Lithuania, he procured fake documents that he used to enter the part of German-occupied Poland the Germans had dubbed the "General Government". It was a difficult journey, mostly on foot, that ended in summer 1940. Finally back in Warsaw, he reunited with Janina. Like many Poles at the time, to evade notice by German authorities, Miłosz participated in underground activities. For example, with higher education officially forbidden to Poles, he attended underground lectures by Władysław Tatarkiewicz, the Polish philosopher and historian of philosophy and aesthetics. He translated Shakespeare's "As You Like It" and T. S. Eliot's "The Waste Land" into Polish. Along with his friend the novelist Jerzy Andrzejewski, he also arranged for the publication of his third volume of poetry, "Poems", under a pseudonym in September 1940. The pseudonym was "Jan Syruć" and the title page said the volume had been published by a fictional press in Lwów in 1939; in fact, it may have been the first clandestine book published in occupied Warsaw. In 1942, Miłosz arranged for the publication of an anthology of Polish poets, "Invincible Song: Polish Poetry of War Time", by an underground press. Miłosz's riskiest underground wartime activity was aiding Jews in Warsaw, which he did through an underground socialist organization called Freedom. His brother, Andrzej, was also active in helping Jews in Nazi-occupied Poland; in 1943, he transported the Polish Jew Seweryn Tross and his wife from Wilno to Warsaw. Miłosz took in the Trosses, found them a hiding place, and supported them financially. The Trosses ultimately died during the Warsaw Uprising. Miłosz helped at least three other Jews in similar ways: Felicja Wołkomińska and her brother and sister. Despite his willingness to engage in underground activity and vehement opposition to the Nazis, Miłosz did not join the Polish Home Army. In later years, he explained that this was partly out of an instinct for self-preservation and partly because he saw its leadership as right-wing and dictatorial. He also did not participate in the planning or execution of the Warsaw Uprising. According to Irena Grudzińska-Gross, he saw the uprising as a "doomed military effort" and lacked the "patriotic elation" for it. He called the uprising "a blameworthy, lightheaded enterprise", but later criticized the Red Army for failing to support it when it had the opportunity to do so. As German troops began torching Warsaw buildings in August 1944, Miłosz was captured and held in a prisoner transit camp; he was later rescued by a Catholic nun—a stranger to him—who pleaded with the Germans on his behalf. Once freed, he and Janina escaped the city, ultimately settling in a village outside Kraków, where they were staying when the Red Army swept through Poland in January 1945, after Warsaw had been largely destroyed. In the preface to his 1953 book "The Captive Mind", Miłosz wrote, "I do not regret those years in Warsaw, which was, I believe, the most agonizing spot in the whole of terrorized Europe. Had I then chosen emigration, my life would certainly have followed a very different course. But my knowledge of the crimes which Europe has witnessed in the twentieth century would be less direct, less concrete than it is". Immediately after the war, Miłosz published his fourth poetry collection, "Rescue"; it focused on his wartime experiences and contains some of his most critically praised work, including the 20-poem cycle "The World," composed like a primer for naïve schoolchildren, and the cycle "Voices of Poor People". The volume also contains some of his most frequently anthologized poems, including "A Song on the End of the World", "Campo Dei Fiori", and "A Poor Christian Looks at the Ghetto". From 1945 to 1951, Miłosz served as a cultural attaché for the newly formed People's Republic of Poland. It was in this capacity that he first met Jane Zielonko, the future translator of "The Captive Mind", with whom he had a brief relationship. He moved from New York City to Washington, D.C., and finally to Paris, organizing and promoting Polish cultural occasions such as musical concerts, art exhibitions, and literary and cinematic events. Although he was a representative of Poland, which had become a Soviet satellite country behind the Iron Curtain, he was not a member of any communist party. In "The Captive Mind", he explained his reasons for accepting the role:My mother tongue, work in my mother tongue, is for me the most important thing in life. And my country, where what I wrote could be printed and could reach the public, lay within the Eastern Empire. My aim and purpose was to keep alive freedom of thought in my own special field; I sought in full knowledge and conscience to subordinate my conduct to the fulfillment of that aim. I served abroad because I was thus relieved from direct pressure and, in the material which I sent to my publishers, could be bolder than my colleagues at home. I did not want to become an émigré and so give up all chance of taking a hand in what was going on in my own country.Miłosz did not publish a book while a representative of the Polish government. Instead, he wrote articles for various Polish periodicals introducing readers to American writers like Eliot, William Faulkner, Ernest Hemingway, Norman Mailer, Robert Lowell, and W. H. Auden. He also translated into Polish Shakespeare's "Othello" and the work of Walt Whitman, Carl Sandburg, Pablo Neruda, and others. In 1947, Miłosz's son, Anthony, was born in Washington, D.C. In 1948, Miłosz arranged for the Polish government to fund a Department of Polish Studies at Columbia University. Named for Adam Mickiewicz, the department featured lectures by Manfred Kridl, Miłosz's friend who was then on the faculty of Smith College, and produced a scholarly book about Mickiewicz. Mickiewicz's granddaughter wrote a letter to Dwight D. Eisenhower, then the president of Columbia University, to express her approval, but the Polish American Congress, an influential group of Polish émigrés, denounced the arrangement in a letter to Eisenhower that they shared with the press, which alleged a communist infiltration at Columbia. Students picketed and called for boycotts. One faculty member resigned in protest. Despite the controversy, the department was established, the lectures took place, and the book was produced, but the department was discontinued in 1954 when funding from Poland ceased. In 1949, Miłosz visited Poland for the first time since joining its diplomatic corps and was appalled by the conditions he saw, including an atmosphere of pervasive fear of the government. After returning to the U.S., he began to look for a way to leave his post, even soliciting advice from Albert Einstein, whom he met in the course of his duties. As the Polish government, influenced by Josef Stalin, became more oppressive, his superiors began to view Miłosz as a threat: he was outspoken in his reports to Warsaw and met with people not approved by his superiors. Consequently, his superiors called him "an individual who ideologically is totally alien". Toward the end of 1950, when Janina was pregnant with their second child, Miłosz was recalled to Warsaw, where in December 1950 his passport was confiscated, ostensibly until it could be determined that he did not plan to defect. After intervention by Poland's foreign minister, Zygmunt Modzelewski, Miłosz's passport was returned. Realizing that he was in danger if he remained in Poland, Miłosz left for Paris in January 1951. Upon arriving in Paris, Miłosz went into hiding, aided by the staff of the Polish émigré magazine "Kultura." With his wife and son still in the United States, he applied to enter the U.S. and was denied. At the time, the U.S. was in the grip of McCarthyism, and influential Polish émigrés had convinced American officials that Miłosz was a communist. Unable to leave France, Miłosz was not present for the birth of his second son, John Peter, in Washington, D.C., in 1951. With the United States closed to him, Miłosz requested—and was granted—political asylum in France. After three months in hiding, he announced his defection at a press conference and in a "Kultura" article, "No", that explained his refusal to live in Poland or continue working for the Polish regime. He was the first artist of note from a communist country to make public his reasons for breaking ties with his government. His case attracted attention in Poland, where his work was banned and he was attacked in the press, and in the West, where prominent individuals voiced criticism and support. For example, the future Nobel laureate Pablo Neruda, then a supporter of the Soviet Union, attacked him in a communist newspaper as "The Man Who Ran Away". On the other hand, Albert Camus, another future Nobel laureate, visited Miłosz and offered his support. Another supporter during this period was the Swiss philosopher Jeanne Hersch, with whom Miłosz had a brief romantic affair. Miłosz was finally reunited with his family in 1953, when Janina and the children joined him in France. That same year saw the publication of "The Captive Mind", a nonfiction work that uses case studies to dissect the methods and consequences of Soviet communism, which at the time had prominent admirers in the West. The book brought Miłosz his first readership in the United States, where it was credited by some on the political left (such as Susan Sontag) with helping to change perceptions about communism. The German philosopher Karl Jaspers described it as a "significant historical document". It became a staple of political science courses and is considered a classic work in the study of totalitarianism. Miłosz's years in France were productive. In addition to "The Captive Mind", he published two poetry collections ("Daylight" (1954) and "A Treatise on Poetry" (1957)), two novels ("The Seizure of Power" (1955) and "The Issa Valley" (1955)), and a memoir ("Native Realm" (1959)). All were published in Polish by an émigré press in Paris. Andrzej Franaszek has called "A Treatise on Poetry" Miłosz's magnum opus, while the scholar Helen Vendler compared it to "The Waste Land", a work "so powerful that it bursts the bounds in which it was written—the bounds of language, geography, epoch". A long poem divided into four sections, "A Treatise on Poetry" surveys Polish history, recounts Miłosz's experience of war, and explores the relationship between art and history. In 1956, Miłosz and Janina were married. In 1960, Miłosz was offered a position as a visiting lecturer at the University of California at Berkeley. With this offer, and with the climate of McCarthyism abated, he was able to move to the United States. He proved to be an adept and popular teacher, and was offered tenure after only two months. The rarity of this, and the degree to which he had impressed his colleagues, are underscored by the fact that Miłosz lacked a PhD and teaching experience. Yet his deep learning was obvious, and after years of working administrative jobs that he found stifling, he told friends that he was in his element in a classroom. With stable employment as a tenured professor of Slavic languages and literatures, Miłosz was able to secure American citizenship and purchase a home in Berkeley. Miłosz began to publish scholarly articles in English and Polish on a variety of authors, including Fyodor Dostoevsky. But despite his successful transition to the U.S., he described his early years at Berkeley as frustrating, as he was isolated from friends and viewed as a political figure rather than a great poet. (In fact, some of his Berkeley faculty colleagues, unaware of his creative output, expressed astonishment when he won the Nobel Prize.) His poetry was not available in English, and he was not able to publish in Poland. As part of an effort to introduce American readers to his poetry, as well as to his fellow Polish poets' work, Miłosz conceived and edited the anthology "Postwar Polish Poetry", which was published in English in 1965. American poets like W.S. Merwin, and American scholars like Clare Cavanagh, have credited it with a profound impact. It was many English-language readers' first exposure to Miłosz's poetry, as well as that of Polish poets like Wisława Szymborska, Zbigniew Herbert, and Tadeusz Różewicz. (In the same year, Miłosz's poetry also appeared in the first issue of "Modern Poetry in Translation," an English-language journal founded by prominent literary figures Ted Hughes and Daniel Weissbort. The issue also featured Miroslav Holub, Yehuda Amichai, Ivan Lalić, Vasko Popa, Zbigniew Herbert, and Andrei Voznesensky.) In 1969, Miłosz's textbook "The History of Polish Literature" was published in English. He followed this with a volume of his own work, "Selected Poems" (1973), some of which he translated into English himself. At the same time, Miłosz continued to publish in Polish with an émigré press in Paris. His poetry collections from this period include "King Popiel and Other Poems" (1962), "Bobo’s Metamorphosis" (1965), "City Without a Name" (1969), and "From the Rising of the Sun" (1974). During Miłosz's time at Berkeley, the campus became a hotbed of student protest, notably as the home of the Free Speech Movement, which has been credited with helping to "define a generation of student activism" across the United States. Miłosz's relationship to student protesters was sometimes antagonistic: he called them "spoiled children of the bourgeoisie" and their political zeal naïve. At one campus event in 1970, he mocked protesters who claimed to be demonstrating for peace and love: "Talk to me about love when they come into your cell one morning, line you all up, and say 'You and you, step forward—it’s your time to die—unless any of your friends loves you so much he wants to take your place!'" Comments like these were in keeping with his stance toward American counterculture of the 1960s in general. For example, in 1968, when Miłosz was listed as a signatory of an open letter of protest written by poet and counterculture figure Allen Ginsberg and published in "The New York Review of Books", Miłosz responded by calling the letter "dangerous nonsense" and insisting that he had not signed it. After 18 years, Miłosz retired from teaching in 1978. To mark the occasion, he was awarded a "Berkeley Citation", the University of California's equivalent of an honorary doctorate. But when his wife, Janina, fell ill and required expensive medical treatment, Miłosz returned to teaching seminars. On 9 October 1980, the Swedish Academy announced that Miłosz had won the Nobel Prize in Literature. The award catapulted him to global fame. On the day the prize was announced, Miłosz held a brief press conference and then left to teach a class on Dostoevsky. In his Nobel lecture, Miłosz described his view of the role of the poet, lamented the tragedies of the 20th century, and paid tribute to his cousin Oscar. Many Poles became aware of Miłosz for the first time when he won the Nobel Prize. After a 30-year ban in Poland, his writing was finally published there in limited selections. He was also able to visit Poland for the first time since fleeing in 1951 and was greeted by crowds with a hero's welcome. He met with leading Polish figures like Lech Wałęsa and Pope John Paul II. At the same time, his early work, until then only available in Polish, began to be translated into English and many other languages. In 1981, Miłosz was appointed the Norton Professor of Poetry at Harvard University, where he was invited to deliver the Charles Eliot Norton Lectures. He used the opportunity, as he had before becoming a Nobel laureate, to draw attention to writers who had been unjustly imprisoned or persecuted. The lectures were published as "The Witness of Poetry" (1983). Miłosz continued to publish work in Polish through his longtime publisher in Paris, including the poetry collections "Hymn of the Pearl" (1981), "Bells in Winter" (1984) and "Unattainable Earth" (1986), and the essay collection "Beginning with My Streets" (1986). In 1986, Miłosz's wife, Janina, died. In 1988, Miłosz's "Collected Poems" appeared in English; it was the first of several attempts to collect all his poetry into a single volume. After the fall of communism in Poland, he split his time between Berkeley and Kraków, and he began to publish his writing in Polish with a publisher based in Kraków. When Lithuania broke free from the Soviet Union in 1991, Miłosz visited for the first time since 1939. In 2000, he moved to Kraków. In 1992, Miłosz married Carol Thigpen, an academic at Emory University in Atlanta, Georgia. They remained married until her death in 2002. His work from the 1990s includes the poetry collections "Facing the River" (1994) and "Roadside Dog" (1997), and the collection of short prose "Miłosz’s ABC’s" (1997). Miłosz's last stand-alone volumes of poetry were "This" (2000), and "The Second Space" (2002). Uncollected poems written afterward appeared in English in "New and Selected Poems" (2004) and, posthumously, in "Selected and Last Poems" (2011). Czesław Miłosz died on 14 August 2004, at his Kraków home, aged 93. He was given a state funeral at the historic Mariacki Church in Kraków. Polish Prime Minister Marek Belka attended, as did the former president of Poland, Lech Wałęsa. Thousands of people lined the streets to witness his coffin moved by military escort to his final resting place at Skałka Roman Catholic Church, where he was one of the last to be commemorated. In front of that church, the poets Seamus Heaney, Adam Zagajewski, and Robert Hass read Miłosz's poem "In Szetejnie" in Polish, French, English, Russian, Lithuanian, and Hebrew—all the languages Miłosz knew. Media from around the world covered the funeral. Protesters threatened to disrupt the proceedings on the grounds that Miłosz was anti-Polish, anti-Catholic, and had signed a petition supporting gay and lesbian freedom of speech and assembly. Pope John Paul II, along with Miłosz's confessor, issued public messages confirming that Miłosz had received the sacraments, which quelled the protest. Miłosz's brother, Andrzej Miłosz (1917–2002), was a Polish journalist, translator, and documentary film producer. His work included Polish documentaries about his brother. Miłosz's son, Anthony, is a composer and software designer. He studied linguistics, anthropology, and chemistry at the University of California at Berkeley, and neuroscience at the University of California Medical Center in San Francisco. In addition to releasing recordings of his own compositions, he has translated some of his father's poems into English. In addition to the Nobel Prize in Literature, Miłosz received the following awards: Miłosz was named a distinguished visiting professor or fellow at many institutions, including the University of Michigan and University of Oklahoma, where he was a Puterbaugh Fellow in 1999. He was an elected member of both the American Academy of Arts and Sciences and the American Academy of Arts and Letters. He received honorary doctorates from Harvard University, the University of Michigan, the University of California at Berkeley, Jagiellonian University, Catholic University of Lublin, and Vytautas Magnus University in Lithuania. The last institution also has an academic center named for Miłosz. In 1992, Miłosz was made an honorary citizen of Lithuania, where his birthplace was made into a museum and conference center. In 1993, he was made an honorary citizen of Kraków. His books also received awards. His first, "A Poem on Frozen Time", won an award from the Union of Polish Writers in Wilno. "The Seizure of Power" received the Prix Littéraire Européen (European Literary Prize). The collection "Roadside Dog" received a Nike Award in Poland. In 1989, Miłosz was named one of the "Righteous Among the Nations" at Israel's Yad Vashem memorial to the Holocaust, in recognition of his efforts to save Jews in Warsaw during World War II. Miłosz has also been honored posthumously. The Polish Parliament declared 2011, the centennial of his birth, the "Year of Miłosz". It was marked by conferences and tributes throughout Poland, as well as in New York City, at Yale University, and at the Dublin Writers Festival, among many other locations. The same year, he was featured on a Lithuanian postage stamp. Streets are named for him near Paris, Vilnius, and in the Polish cities of Kraków, Poznań, Gdańsk, Białystok, and Wrocław. In Gdańsk there is a Czesław Miłosz Square. In 2013, a primary school in Vilnius was named for Miłosz, joining schools in Mierzecice, Poland, and Schaumburg, Illinois, that bear his name. In 1978, the Russian-American poet Joseph Brodsky called Miłosz "one of the great poets of our time; perhaps the greatest". Miłosz has been cited as an influence by numerous writers—contemporaries and succeeding generations. For example, scholars have written about Miłosz's influence on the writing of Seamus Heaney, and Clare Cavanagh has identified the following poets as having benefited from Miłosz's influence: Robert Pinsky, Edward Hirsch, Rosanna Warren, Robert Hass, Charles Simic, Mary Karr, Carolyn Forché, Mark Strand, Ted Hughes, Joseph Brodsky, and Derek Walcott. By being smuggled into Poland, Miłosz's writing was a source of inspiration to the anti-communist Solidarity movement there in the early 1980s. Lines from his poem "You Who Wronged" are inscribed on the Monument to the Fallen Shipyard Workers of 1970 in Gdańsk, where Solidarity originated. Of the effect of Miłosz's edited volume "Postwar Polish Poetry" on English-language poets, Merwin wrote, "Miłosz’s book had been a talisman and had made most of the literary bickering among the various ideological encampments, then most audible in the poetic doctrines in English, seem frivolous and silly". Similarly, the British poet and scholar Donald Davie argued that, for many English-language writers, Miłosz's work encouraged an expansion of poetry to include multiple viewpoints and an engagement with subjects of intellectual and historical importance: "I have suggested, going for support to the writings of Miłosz, that no concerned and ambitious poet of the present day, aware of the enormities of twentieth-century history, can for long remain content with the privileged irresponsibility allowed to, or imposed on, the lyric poet". Miłosz's writing continues to be the subject of academic study, conferences, and cultural events. His papers, including manuscripts, correspondence, and other materials, are housed at the Beinecke Rare Book and Manuscript Library at Yale University. Miłosz's birth in a time and place of shifting borders and overlapping cultures, and his later naturalization as an American citizen, have led to competing claims about his nationality. Although his family identified as Polish and Polish was his primary language, and although he frequently spoke of Poland as his country, he also publicly identified himself as one of the last citizens of the multi-ethnic Grand Duchy of Lithuania. Writing in a Polish newspaper in 2000, he claimed, "I was born in the very center of Lithuania and so have a greater right than my great forebear, Mickiewicz, to write 'O Lithuania, my country.'" But in his Nobel lecture, he said, "My family in the 16th century already spoke Polish, just as many families in Finland spoke Swedish and in Ireland English, so I am a Polish, not a Lithuanian, poet". Public statements such as these, and numerous others, inspired discussion about his nationality, including a claim that he was "arguably the greatest spokesman and representative of a Lithuania that, in Miłosz’s mind, was bigger than its present incarnation". Others have viewed Miłosz as an American author, hosting exhibitions about him from that perspective and including his work in anthologies of American poetry. In "The New York Review of Books" in 1981, the critic John Bayley wrote, "nationality is not a thing [Miłosz] can take seriously; it would be hard to imagine a greater writer more emancipated from even its most subtle pretensions". Echoing this notion, the scholar and diplomat Piotr Wilczek argued that, even when he was greeted as a national hero in Poland, Miłosz "made a distinct effort to remain a universal thinker". Speaking at a ceremony to celebrate his birth centenary in 2011, Lithuanian President Dalia Grybauskaitė stressed that Miłosz's works "unite the Lithuanian and Polish people and reveal how close and how fruitful the ties between our people can be". Though raised Catholic, Miłosz as a young man came to adopt a "scientific, atheistic position mostly", though he later returned to the Catholic faith. He translated parts of the Bible into Polish, and allusions to Catholicism pervade his poetry, culminating in a long 2001 poem, "A Theological Treatise". For some critics, Miłosz's belief that literature should provide spiritual fortification was outdated: Franaszek suggests that Miłosz's belief was evidence of a "beautiful naïveté", while David Orr, citing Miłosz's dismissal of "poetry which does not save nations or people", accused him of "pompous nonsense". Miłosz expressed some criticism of both Catholicism and Poland (a majority-Catholic country), causing furor in some quarters when it was announced that he would be interred in Kraków's historic Skałka church. Cynthia Haven writes that, to some readers, Miłosz's embrace of Catholicism can seem surprising and complicates the understanding of him and his work. Miłosz's body of work comprised multiple literary genres: poetry, fiction (particularly the novel), autobiography, scholarship, personal essay, and lectures. His letters are also of interest to scholars and lay readers; for example, his correspondence with writers such as Jerzy Andrzejewski, Witold Gombrowicz, and Thomas Merton have been published. At the outset of his career, Miłosz was known as a "catastrophist" poet—a label critics applied to him and other poets from the Żagary poetry group to describe their use of surreal imagery and formal inventiveness in reaction to a Europe beset by extremist ideologies and war. While Miłosz evolved away from the apocalyptic view of catastrophist poetry, he continued to pursue formal inventiveness throughout his career. As a result, his poetry demonstrates a wide-ranging mastery of form, from long or epic poems (e.g., "A Treatise on Poetry") to poems of just two lines (e.g., "On the Death of a Poet" from the collection "This"), and from prose poems and free verse to classic forms such as the ode or elegy. Some of his poems use rhyme, but many do not. In numerous cases, Miłosz used form to illuminate meaning in his poetry; for example, by juxtaposing variable stanzas to accentuate ideas or voices that challenge each other. Miłosz's work is known for its complexity; according to the scholars Leonard Nathan and Arthur Quinn, Miłosz "prided himself on being an esoteric writer accessible to a mere handful of readers". Nevertheless, some common themes are readily apparent throughout his body of work. The poet, critic, and frequent Miłosz translator Robert Hass has described Miłosz as "a poet of great inclusiveness", with a fidelity to capturing life in all of its sensuousness and multiplicities. According to Hass, Miłosz's poems can be viewed as "dwelling in contradiction", where one idea or voice is presented only to be immediately challenged or changed. According to Donald Davie, this allowance for contradictory voices—a shift from the solo lyric voice to a chorus—is among the most important aspects of Miłosz's work. The poetic chorus is deployed not just to highlight the complexity of the modern world but also to search for morality, another of Miłosz's recurrent themes. Nathan and Quinn write, "Miłosz’s work is devoted to unmasking man’s fundamental duality; he wants to make his readers admit the contradictory nature of their own experience" because doing so "forces us to assert our preferences as preferences". That is, it forces readers to make conscious choices, which is the arena of morality. At times, Miłosz's exploration of morality was explicit and concrete, such as when, in "The Captive Mind", he ponders the right way to respond to three Lithuanian women who were forcibly moved to a Russian communal farm and wrote to him for help, or when, in the poems "Campo Dei Fiori" and "A Poor Christian Looks at the Ghetto", he addresses survivor's guilt and the morality of writing about another's suffering. Miłosz's exploration of morality takes place in the context of history, and confrontation with history is another of his major themes. Vendler wrote, "for Miłosz, the person is irrevocably a person in history, and the interchange between external event and the individual life is the matrix of poetry". Having experienced both Nazism and Stalinism, Miłosz was particularly concerned with the notion of "historical necessity", which, in the 20th century, was used to justify human suffering on a previously unheard-of scale. Yet Miłosz did not reject the concept entirely. Nathan and Quinn summarize Miłosz's appraisal of historical necessity as it appears in his essay collection "Views from San Francisco Bay": "Some species rise, others fall, as do human families, nations, and whole civilizations. There may well be an internal logic to these transformations, a logic that when viewed from sufficient distance has its own elegance, harmony, and grace. Our reason tempts us to be enthralled by this superhuman splendor; but when so enthralled we find it difficult to remember, except perhaps as an element in an abstract calculus, the millions of individuals, the millions upon millions, who unwillingly paid for this splendor with pain and blood". Miłosz's willingness to accept a form of logic in history points to another recurrent aspect of his writing: his capacity for wonder, amazement, and, ultimately, faith—not always religious faith, but "faith in the objective reality of a world to be known by the human mind but not constituted by that mind". At other times, Miłosz was more explicitly religious in his work. According to scholar and translator Michael Parker, "crucial to any understanding of Miłosz’s work is his complex relationship to Catholicism". His writing is filled with allusions to Christian figures, symbols, and theological ideas, though Miłosz was closer to Gnosticism, or what he called Manichaeism, in his personal beliefs, viewing the universe as ruled by an evil whose influence human beings must try to escape. From this perspective, "he can at once admit that the world is ruled by necessity, by evil, and yet still find hope and sustenance in the beauty of the world. History reveals the pointlessness of human striving, the instability of human things; but time also is the moving image of eternity". According to Hass, this viewpoint left Miłosz "with the task of those heretical Christians…to suffer time, to contemplate being, and to live in the hope of the redemption of the world". Miłosz had numerous literary and intellectual influences, although scholars of his work—and Miłosz himself, in his writings—have identified the following as significant: Oscar Miłosz (who inspired Miłosz's interest in the metaphysical) and, through him, Emanuel Swedenborg; Lev Shestov; Simone Weil (whose work Miłosz translated into Polish); Dostoevsky; William Blake (whose concept of "Ulro" Miłosz borrowed for his book "The Land of Ulro"), and Eliot.
https://en.wikipedia.org/wiki?curid=6542
Carnivore A carnivore , meaning "meat eater" (Latin, "caro", genitive "carnis", meaning "meat" or "flesh" and "vorare" meaning "to devour"), is an animal whose food and energy requirements derive solely from animal tissue or meat, whether through hunting or scavenging. Animals that depend solely on animal flesh for their nutrient requirements are called obligate carnivores while those that also consume non-animal food are called facultative carnivores. Omnivores also consume both animal and non-animal food, and, apart from the more general definition, there is no clearly defined ratio of plant to animal material that would distinguish a facultative carnivore from an omnivore. A carnivore at the top of the food chain, not preyed upon by other animals, is termed an apex predator. "Carnivore" also may refer to the mammalian order Carnivora, but this is somewhat misleading: many, but not all, Carnivora are meat eaters, and even fewer are true obligate carnivores (see below). For example, while the Arctic polar bear eats meat almost exclusively (more than 90% of its diet is meat), most species of bears are omnivorous, and the giant panda is exclusively herbivorous. There are also many carnivorous species that are not members of Carnivora. The correct term for mammals in this group is "carnivoran". Besides, some mammals, especially the cetaceans, are highly carnivorous yet are not true Carnivorans. Outside the animal kingdom, there are several genera containing carnivorous plants (predominantly insectivores) and several phyla containing carnivorous fungi (preying mostly on microscopic invertebrates such as nematodes, amoebae and springtails). Carnivores are sometimes characterized by their type of prey. For example, animals that eat mainly insects and similar invertebrates are called insectivores, while those that eat mainly fish are called piscivores. The first tetrapods, or land-dwelling vertebrates, were piscivorous amphibians known as labyrinthodonts. They gave rise to insectivorous vertebrates and, later, to predators of other tetrapods. Carnivores may alternatively be classified according to the percentage of meat in their diet. The diet of a hypercarnivore consists of more than 70% meat, that of a mesocarnivore 30–70%, and that of a hypocarnivore less than 30%, with the balance consisting of non-animal foods such as fruits, other plant material, or fungi. Obligate or "true" carnivores are those whose diet requires nutrients found only in animal flesh. While obligate carnivores might be able to ingest small amounts of plant matter, they lack the necessary physiology required to fully digest it. In fact, some obligate carnivorous mammals will only ingest vegetation to use as an emetic, to self-induce vomiting of the vegetation along with the other food it had ingested that upset its stomach. Obligate carnivores include the axolotl, which consumes mainly worms and larvae in its environment, but if necessary will consume algae, as well as all felids (including the domestic cat) which require a diet of primarily animal flesh and organs. Specifically, cats have high protein requirements and their metabolisms appear unable to synthesize essential nutrients such as retinol, arginine, taurine, and arachidonic acid; thus, in nature, they must consume flesh to supply these nutrients. Characteristics commonly associated with carnivores include strength, speed, and keen senses for hunting, as well as teeth and claws for capturing and tearing prey. However, some carnivores do not hunt and are scavengers, lacking the physical characteristics to bring down prey; in addition, most hunting carnivores will scavenge when the opportunity arises. Carnivores have comparatively short digestive systems, as they are not required to break down the tough cellulose found in plants. Many hunting animals have evolved eyes facing forward, enabling depth perception. This is almost universal among mammalian predators, while most reptile and amphibian predators have eyes facing sideways. "Predation" (the eating of one living creature by another for nutrition) predates the rise of commonly recognized carnivores by hundreds of millions (perhaps billions) of years. The earliest predators were microbial organisms, which engulfed or grazed on others. Because the fossil record is poor, these first predators could date back anywhere between 1 and over 2.7 Gya (billion years ago). The rise of eukaryotic cells at around 2.7 Gya, the rise of multicellular organisms at about 2 Gya, and the rise of mobile predators (around 600 Mya – 2 Gya, probably around 1 Gya) have all been attributed to early predatory behavior, and many very early remains show evidence of boreholes or other markings attributed to small predator species. Among more familiar species, the first vertebrate carnivores were fish, and then amphibians that moved on to land. Early tetrapods were large amphibious piscivores. Some scientists assert that "Dimetrodon" "was the first terrestrial vertebrate to develop the curved, serrated teeth that enable a predator to eat prey much larger than itself." While amphibians continued to feed on fish and later insects, reptiles began exploring two new food types: tetrapods (carnivory) and then plants (herbivory). Carnivory was a natural transition from insectivory for medium and large tetrapods, requiring minimal adaptation; in contrast, a complex set of adaptations was necessary for feeding on highly fibrous plant materials. In the Mesozoic, some theropod dinosaurs such as "Tyrannosaurus rex" were probably obligate carnivores. Though the theropods were the larger carnivores, several carnivorous mammal groups were already present. Most notable are the gobiconodontids, the triconodontid "Jugulator", the deltatheroideans and "Cimolestes". Many of these, such as "Repenomamus", "Jugulator" and "Cimolestes", were among the largest mammals in their faunal assemblages, capable of attacking dinosaurs. In the early-to-mid-Cenozoic, the dominant predator forms were mammals: hyaenodonts, oxyaenids, entelodonts, ptolemaiidans, arctocyonids and mesonychians, representing a great diversity of eutherian carnivores in the northern continents and Africa. In South America, sparassodonts were dominant, while Australia saw the presence of several marsupial predators, such as the dasyuromorphs and thylacoleonids. From the Miocene to the present, the dominant carnivorous mammals have been carnivoramorphs. Most carnivorous mammals, from dogs to "Deltatheridium", share several dental adaptations, such as carnassialiforme teeth, long canines and even similar tooth replacement patterns. Most aberrant are thylacoleonids, with a diprodontan dentition completely unlike that of any other mammal; and eutriconodonts like gobioconodontids and "Jugulator", with a three-cusp anatomy which nevertheless functioned similarly to carnassials.
https://en.wikipedia.org/wiki?curid=6543
Celts The Celts (, see pronunciation of "Celt" for different usages) are a collection of Indo-European peoples of Europe identified by their use of the Celtic languages and other cultural similarities. The history of pre-Celtic Europe and the exact relationship between ethnic, linguistic and cultural factors in the Celtic world remains uncertain and controversial. The exact geographic spread of the ancient Celts is disputed; in particular, the ways in which the Iron Age inhabitants of Great Britain and Ireland should be regarded as Celts have become a subject of controversy. According to one theory, the common root of the Celtic languages, the Proto-Celtic language, arose in the Late Bronze Age Urnfield culture of Central Europe, which flourished from around 1200 BC. According to another theory proposed in the 19th century, the first people to adopt cultural characteristics regarded as Celtic were the people of the Iron Age Hallstatt culture in central Europe (c. 800–450 BC), named for the rich grave finds in Hallstatt, Austria. It is thus that this area is sometimes called the "Celtic homeland". By or during the later La Tène period (c. 450 BC to the Roman conquest), named after the La Tène site in Switzerland, this Celtic culture was supposed to have expanded by trans-cultural diffusion or migration to the British Isles (Insular Celts), France and the Low Countries (Gauls), Bohemia, Poland and much of Central Europe, the Iberian Peninsula (Celtiberians, Celtici, Lusitanians and Gallaeci) and northern Italy (Golasecca culture and Cisalpine Gauls) and, following the Celtic settlement of Eastern Europe beginning in 279 BC, as far east as central Anatolia (Galatians) in modern-day Turkey. The earliest undisputed direct examples of a Celtic language are the Lepontic inscriptions beginning in the 6th century BC. Continental Celtic languages are attested almost exclusively through inscriptions and place-names. Insular Celtic languages are attested beginning around the 4th century in Ogham inscriptions, although they were clearly being spoken much earlier. Celtic literary tradition begins with Old Irish texts around the 8th century AD. Coherent texts of Early Irish literature, such as the "Táin Bó Cúailnge" ("Cattle Raid of Cooley"), survive in 12th-century recensions. By the mid-1st millennium, with the expansion of the Roman Empire and migrating Germanic tribes, Celtic culture and Insular Celtic languages had become restricted to Ireland, the western and northern parts of Great Britain (Wales, Scotland, and Cornwall), the Isle of Man, and Brittany. Between the 5th and 8th centuries, the Celtic-speaking communities in these Atlantic regions emerged as a reasonably cohesive cultural entity. They had a common linguistic, religious and artistic heritage that distinguished them from the culture of the surrounding polities. By the 6th century, however, the Continental Celtic languages were no longer in wide use. Insular Celtic culture diversified into that of the Gaels (Irish, Scottish and Manx) and the Celtic Britons (Welsh, Cornish, and Bretons) of the medieval and modern periods. A modern Celtic identity was constructed as part of the Romanticist Celtic Revival in Great Britain, Ireland, and other European territories, such as Portugal and Spanish Galicia. Today, Irish, Scottish Gaelic, Welsh, and Breton are still spoken in parts of their historical territories, and Cornish and Manx are undergoing a revival. The first recorded use of the name of Celts – as ("Keltoi") in Greek – to refer to an ethnic group was by Hecataeus of Miletus, the Greek geographer, in 517 BC, when writing about a people living near Massilia (modern Marseille). In the fifth century BC, Herodotus referred to "Keltoi" living around the head of the Danube and also in the far west of Europe. The etymology of the term "Keltoi" is unclear. Possible roots include Indo-European *"kʲel" 'to hide' (present also in Old Irish "ceilid"), IE *"kʲel" 'to heat' or *"kel" 'to impel'. Several authors have supposed it to be Celtic in origin, while others view it as a name coined by Greeks. Linguist Patrizia De Bernardo Stempel falls in the latter group, and suggests the meaning "the tall ones". In the 1st century BC, Julius Caesar reported that the people known to the Romans as Gauls () called themselves Celts, which suggests that even if the name "Keltoi" was bestowed by the Greeks, it had been adopted to some extent as a collective name by the tribes of Gaul. The geographer Strabo, writing about Gaul towards the end of the first century BC, refers to the "race which is now called both Gallic and Galatic," though he also uses the term Celtica as a synonym for Gaul, which is separated from Iberia by the Pyrenees. Yet he reports Celtic peoples in Iberia, and also uses the ethnic names Celtiberi and Celtici for peoples there, as distinct from Lusitani and Iberi. Pliny the Elder cited the use of Celtici in Lusitania as a tribal surname, which epigraphic findings have confirmed. Latin Gallus (pl. "Galli") might stem from a Celtic ethnic or tribal name originally, perhaps one borrowed into Latin during the Celtic expansions into Italy during the early fifth century BC. Its root may be the Proto-Celtic "*galno", meaning "power, strength", hence Old Irish "gal" "boldness, ferocity" and Welsh "gallu" "to be able, power". The tribal names of Gallaeci and the Greek Γαλάται ("Galatai", Latinized "Galatae"; see the region Galatia in Anatolia) most probably have the same origin. The suffix "-atai" might be an Ancient Greek inflection. Classical writers did not apply the terms ("Keltoi") or "Celtae" to the inhabitants of Britain or Ireland, which has led to some scholars preferring not to use the term for the Iron Age inhabitants of those islands. Celt is a modern English word, first attested in 1707, in the writing of Edward Lhuyd, whose work, along with that of other late 17th-century scholars, brought academic attention to the languages and history of the early Celtic inhabitants of Great Britain. The English form Gaul (first recorded in the 17th century) and Gaulish come from the French "Gaule" and "Gaulois", a borrowing from Frankish "*Walholant", "Roman land" (see Gaul: Name), the root of which is Proto-Germanic "*walha-", "foreigner, Roman, Celt", whence the English word Welsh (Old English "wælisċ" < *"walhiska-"), South German "", meaning "Celtic speaker", "French speaker" or "Italian speaker" in different contexts, and Old Norse "valskr", pl. "valir", "Gaulish, French"). Proto-Germanic "*walha" is derived ultimately from the name of the Volcae, a Celtic tribe who lived first in the south of Germany and in central Europe and then migrated to Gaul. This means that English Gaul, despite its superficial similarity, is not actually derived from Latin "Gallia" (which should have produced "**Jaille" in French), though it does refer to the same ancient region. Celtic refers to a family of languages and, more generally, means "of the Celts" or "in the style of the Celts". Several archaeological cultures are considered Celtic in nature, based on unique sets of artefacts. The link between language and artefact is aided by the presence of inscriptions. The relatively modern idea of an identifiable Celtic cultural identity or "Celticity" generally focuses on similarities among languages, works of art, and classical texts, and sometimes also among material artefacts, social organisation, homeland and mythology. Earlier theories held that these similarities suggest a common racial origin for the various Celtic peoples, but more recent theories hold that they reflect a common cultural and language heritage more than a genetic one. Celtic cultures seem to have been widely diverse, with the use of a Celtic language being the main thing they had in common. Today, the term Celtic generally refers to the languages and respective cultures of Ireland, Scotland, Wales, Cornwall, the Isle of Man, and Brittany, also known as the Celtic nations. These are the regions where four Celtic languages are still spoken to some extent as mother tongues. The four are Irish Gaelic, Scottish Gaelic, Welsh, and Breton; plus two recent revivals, Cornish (one of the Brittonic languages) and Manx (one of the Goidelic languages). There are also attempts to reconstruct Cumbric, a Brittonic language from North West England and South West Scotland. Celtic regions of Continental Europe are those whose residents claim a Celtic heritage, but where no Celtic language has survived; these areas include the western Iberian Peninsula, i.e. Portugal and north-central Spain (Galicia, Asturias, Cantabria, Castile and León, Extremadura). Continental Celts are the Celtic-speaking people of mainland Europe and Insular Celts are the Celtic-speaking peoples of the British and Irish islands and their descendants. The Celts of Brittany derive their language from migrating insular Celts, mainly from Wales and Cornwall, and so are grouped accordingly. The Celtic languages form a branch of the larger Indo-European family. By the time speakers of Celtic languages entered history around 400 BC, they were already split into several language groups, and spread over much of Western continental Europe, the Iberian Peninsula, Ireland and Britain. The Greek historian Ephorus of Cyme in Asia Minor, writing in the 4th century BC, believed that the Celts came from the islands off the mouth of the Rhine and were "driven from their homes by the frequency of wars and the violent rising of the sea". Some scholars think that the Urnfield culture of western Middle Europe represents an origin for the Celts as a distinct cultural branch of the Indo-European family. This culture was preeminent in central Europe during the late Bronze Age, from circa 1200 BC until 700 BC, itself following the Unetice and Tumulus cultures. The Urnfield period saw a dramatic increase in population in the region, probably due to innovations in technology and agriculture. The spread of iron-working led to the development of the Hallstatt culture directly from the Urnfield (c. 700 to 500 BC). Proto-Celtic, the latest common ancestor of all known Celtic languages, is considered by this school of thought to have been spoken at the time of the late Urnfield or early Hallstatt cultures, in the early 1st millennium BC. The spread of the Celtic languages to Iberia, Ireland and Britain would have occurred during the first half of the 1st millennium BC, the earliest chariot burials in Britain dating to c. 500 BC. Other scholars see Celtic languages as covering Britain and Ireland, and parts of the Continent, long before any evidence of "Celtic" culture is found in archaeology. Over the centuries the language(s) developed into the separate Celtiberian, Goidelic and Brittonic languages. The Hallstatt culture was succeeded by the La Tène culture of central Europe, which was overrun by the Roman Empire, though traces of La Tène style are still to be seen in Gallo-Roman artefacts. In Britain and Ireland La Tène style in art survived precariously to re-emerge in Insular art. Early Irish literature casts light on the flavour and tradition of the heroic warrior elites who dominated Celtic societies. Celtic river-names are found in great numbers around the upper reaches of the Danube and Rhine, which led many Celtic scholars to place the ethnogenesis of the Celts in this area. Diodorus Siculus and Strabo both suggest that the heartland of the people they called Celts was in southern France. The former says that the Gauls were to the north of the Celts, but that the Romans referred to both as Gauls (in linguistic terms the Gauls were certainly Celts). Before the discoveries at Hallstatt and La Tène, it was generally considered that the Celtic heartland was southern France, see Encyclopædia Britannica for 1813. Myles Dillon and Nora Kershaw Chadwick accepted that "the Celtic settlement of the British Isles" might have to be dated to the Bell Beaker culture concluding that "There is no reason why so early a date for the coming of the Celts should be impossible". Martín Almagro Gorbea proposed the origins of the Celts could be traced back to the 3rd millennium BC, also seeking the initial roots in the Beaker period, thus offering the wide dispersion of the Celts throughout western Europe, as well as the variability of the different Celtic peoples, and the existence of ancestral traditions and ancient perspective. Using a multidisciplinary approach, Alberto J. Lorrio and Gonzalo Ruiz Zapatero reviewed and built on Almagro Gorbea's work to present a model for the origin of the Celtic archaeological groups in the Iberian Peninsula (Celtiberian, Vetton, Vaccean, the Castro culture of the northwest, Asturian-Cantabrian and Celtic of the southwest) and proposing a rethinking of the meaning of "Celtic" from a European perspective. More recently, John Koch and Barry Cunliffe have suggested that Celtic origins lie with the Atlantic Bronze Age, roughly contemporaneous with the Hallstatt culture but positioned considerably to the West, extending along the Atlantic coast of Europe. Stephen Oppenheimer points out that the only written evidence that locates the Keltoi near the source of the Danube (i.e. in the Hallstatt region) is in the "Histories" of Herodotus. However, Oppenheimer shows that Herodotus seemed to believe the Danube rose near the Pyrenees, which would place the Ancient Celts in a region which is more in agreement with later classical writers and historians (i.e. in Gaul and the Iberian peninsula). The Proto-Celtic language is usually dated to the Late Bronze Age. The earliest records of a Celtic language are the Lepontic inscriptions of Cisalpine Gaul (Northern Italy), the oldest of which predate the La Tène period. Other early inscriptions, appearing from the early La Tène period in the area of Massilia, are in Gaulish, which was written in the Greek alphabet until the Roman conquest. Celtiberian inscriptions, using their own Iberian script, appear later, after about 200 BC. Evidence of Insular Celtic is available only from about 400 AD, in the form of Primitive Irish Ogham inscriptions. Besides epigraphical evidence, an important source of information on early Celtic is toponymy. Historically many scholars postulated that there was genetic evidence of a common origin of the European Atlantic populations i.e.: Orkney Islands, Scottish, Irish, British, Bretons, and Iberians (Basques, Galicians). More recent genetic evidence does not support the notion of a significant genetic link between these populations, beyond the fact that they are all West Eurasians. Sardinian-like Neolithic farmers did populate Britain (and all of Northern Europe) during the Neolithic period, however, recent genetics research has claimed that, between 2400BC and 2000BC, over 90% of British DNA was overturned by a North European population of ultimate Russian Steppe origin as part of an ongoing migration process that brought large amounts of Steppe DNA (including the R1b haplogroup) to North and West Europe. Modern autosomal genetic clustering is testament to this fact, as both modern and Iron Age British and Irish samples cluster genetically very closely with other North European populations, and somewhat limited with Galicians, Basques or those from the south of France. Such findings have largely put to rest the theory that there is a significant ancestral genetic link (beyond being Europeans) between the various 'Celtic' peoples in the Atlantic area, instead, they are related in that male lines are brother R1b L151 subclades with the local native maternal line admixture explaining the genetic distance noted. Before the 19th century, scholars assumed that the original land of the Celts was west of the Rhine, more precisely in Gaul, because it was where Greek and Roman ancient sources, namely Caesar, located the Celts. This view was challenged by the 19th-century historian Marie Henri d'Arbois de Jubainville who placed the land of origin of the Celts east of the Rhine. Jubainville based his arguments on a phrase of Herodotus' that placed the Celts at the source of the Danube, and argued that Herodotus had meant to place the Celtic homeland in southern Germany. The finding of the prehistoric cemetery of Hallstat in 1846 by Johan Ramsauer and the finding of the archaeological site of La Tène by Hansli Kopp in 1857 drew attention to this area. The concept that the Hallstatt and La Tène cultures could be seen not just as chronological periods but as "Culture Groups", entities composed of people of the same ethnicity and language, had started to grow by the end of the 19th century. At the beginning of the 20th century the belief that these "Culture Groups" could be thought of in racial or ethnic terms was strongly held by Gordon Childe whose theory was influenced by the writings of Gustaf Kossinna. As the 20th century progressed, the racial ethnic interpretation of La Tène culture became much more strongly rooted, and any findings of La Tène culture and flat inhumation cemeteries were directly associated with the Celts and the Celtic language. The Iron Age Hallstatt (c. 800–475 BC) and La Tène (c. 500–50 BC) cultures are typically associated with Proto-Celtic and Celtic culture. In various academic disciplines the Celts were considered a Central European Iron Age phenomenon, through the cultures of Hallstatt and La Tène. However, archaeological finds from the Halstatt and La Tène culture were rare in the Iberian Peninsula, in southwestern France, northern and western Britain, southern Ireland and Galatia and did not provide enough evidence for a cultural scenario comparable to that of Central Europe. It is considered equally difficult to maintain that the origin of the Peninsular Celts can be linked to the preceding Urnfield culture. This has resulted in a more recent approach that introduces a 'proto-Celtic' substratum and a process of Celticisation, having its initial roots in the Bronze Age Bell Beaker culture. The La Tène culture developed and flourished during the late Iron Age (from 450 BC to the Roman conquest in the 1st century BC) in eastern France, Switzerland, Austria, southwest Germany, the Czech Republic, Slovakia and Hungary. It developed out of the Hallstatt culture without any definite cultural break, under the impetus of considerable Mediterranean influence from Greek, and later Etruscan civilisations. A shift of settlement centres took place in the 4th century. The western La Tène culture corresponds to historical Celtic Gaul. Whether this means that the whole of La Tène culture can be attributed to a unified Celtic people is difficult to assess; archaeologists have repeatedly concluded that language, material culture, and political affiliation do not necessarily run parallel. Frey notes that in the 5th century, "burial customs in the Celtic world were not uniform; rather, localised groups had their own beliefs, which, in consequence, also gave rise to distinct artistic expressions". Thus, while the La Tène culture is certainly associated with the Gauls, the presence of La Tène artefacts may be due to cultural contact and does not imply the permanent presence of Celtic speakers. Polybius published a history of Rome about 150 BC in which he describes the Gauls of Italy and their conflict with Rome. Pausanias in the 2nd century AD says that the Gauls "originally called Celts", "live on the remotest region of Europe on the coast of an enormous tidal sea". Posidonius described the southern Gauls about 100 BC. Though his original work is lost it was used by later writers such as Strabo. The latter, writing in the early 1st century AD, deals with Britain and Gaul as well as Hispania, Italy and Galatia. Caesar wrote extensively about his Gallic Wars in 58–51 BC. Diodorus Siculus wrote about the Celts of Gaul and Britain in his 1st-century history. The Romans knew the Celts then living in present-day France as Gauls. The territory of these peoples probably included the Low Countries, the Alps and present-day northern Italy. Julius Caesar in his "Gallic Wars" described the 1st-century BC descendants of those Gauls. Eastern Gaul became the centre of the western La Tène culture. In later Iron Age Gaul, the social organisation resembled that of the Romans, with large towns. From the 3rd century BC the Gauls adopted coinage. Texts with Greek characters from southern Gaul have survived from the 2nd century BC. Greek traders founded Massalia about 600 BC, with some objects (mostly drinking ceramics) being traded up the Rhone valley. But trade became disrupted soon after 500 BC and re-oriented over the Alps to the Po valley in the Italian peninsula. The Romans arrived in the Rhone valley in the 2nd century BC and encountered a mostly Celtic-speaking Gaul. Rome wanted land communications with its Iberian provinces and fought a major battle with the Saluvii at Entremont in 124–123 BC. Gradually Roman control extended, and the Roman Province of Gallia Transalpina developed along the Mediterranean coast. The Romans knew the remainder of Gaul as Gallia Comata – "Hairy Gaul". In 58 BC the Helvetii planned to migrate westward but Julius Caesar forced them back. He then became involved in fighting the various tribes in Gaul, and by 55 BC had overrun most of Gaul. In 52 BC Vercingetorix led a revolt against the Roman occupation but was defeated at the Siege of Alesia and surrendered. Following the Gallic Wars of 58–51 BC, Caesar's "Celtica" formed the main part of Roman Gaul, becoming the province of Gallia Lugdunensis. This territory of the Celtic tribes was bounded on the south by the Garonne and on the north by the Seine and the Marne. The Romans attached large swathes of this region to neighboring provinces Belgica and Aquitania, particularly under Augustus. Place- and personal-name analysis and inscriptions suggest that the Gaulish Celtic language was spoken over most of what is now France. Until the end of the 19th century, traditional scholarship dealing with the Celts did acknowledge their presence in the Iberian Peninsula as a material culture relatable to the Hallstatt and La Tène cultures. However, since according to the definition of the Iron Age in the 19th century Celtic populations were supposedly rare in Iberia and did not provide a cultural scenario that could easily be linked to that of Central Europe, the presence of Celtic culture in that region was generally not fully recognised. Modern scholarship, however, has clearly proven that Celtic presence and influences were most substantial in what is today Spain and Portugal (with perhaps the highest settlement saturation in Western Europe), particularly in the central, western and northern regions. In addition to Gauls infiltrating from the north of the Pyrenees, the Roman and Greek sources mention Celtic populations in three parts of the Iberian Peninsula: the eastern part of the "Meseta" (inhabited by the Celtiberians), the southwest (Celtici, in modern-day Alentejo) and the northwest (Gallaecia and Asturias). A modern scholarly review found several archaeological groups of Celts in Spain: The origins of the Celtiberians might provide a key to understanding the Celticisation process in the rest of the Peninsula. The process of Celticisation of the southwestern area of the peninsula by the Keltoi and of the northwestern area is, however, not a simple Celtiberian question. Recent investigations about the Callaici and Bracari in northwestern Portugal are providing new approaches to understanding Celtic culture (language, art and religion) in western Iberia. John T. Koch of Aberystwyth University suggested that Tartessian inscriptions of the 8th century BC might be classified as Celtic. This would mean that Tartessian is the earliest attested trace of Celtic by a margin of more than a century. The Canegrate culture represented the first migratory wave of the proto-Celtic population from the northwest part of the Alps that, through the Alpine passes, had already penetrated and settled in the western Po valley between Lake Maggiore and Lake Como (Scamozzina culture). It has also been proposed that a more ancient proto-Celtic presence can be traced back to the beginning of the Middle Bronze Age, when North Westwern Italy appears closely linked regarding the production of bronze artefacts, including ornaments, to the western groups of the Tumulus culture. La Tène cultural material appeared over a large area of mainland Italy, the southernmost example being the Celtic helmet from Canosa di Puglia. Italy is home to Lepontic, the oldest attested Celtic language (from the 6th century BC). Anciently spoken in Switzerland and in Northern-Central Italy, from the Alps to Umbria. According to the "Recueil des Inscriptions Gauloises", more than 760 Gaulish inscriptions have been found throughout present-day France – with the notable exception of Aquitaine – and in Italy, which testifies the importance of Celtic heritage in the peninsula. In 391 BC, Celts "who had their homes beyond the Alps streamed through the passes in great strength and seized the territory that lay between the Apennine mountains and the Alps" according to Diodorus Siculus. The Po Valley and the rest of northern Italy (known to the Romans as Cisalpine Gaul) was inhabited by Celtic-speakers who founded cities such as Milan. Later the Roman army was routed at the battle of Allia and Rome was sacked in 390 BC by the Senones. At the battle of Telamon in 225 BC, a large Celtic army was trapped between two Roman forces and crushed. The defeat of the combined Samnite, Celtic and Etruscan alliance by the Romans in the Third Samnite War sounded the beginning of the end of the Celtic domination in mainland Europe, but it was not until 192 BC that the Roman armies conquered the last remaining independent Celtic kingdoms in Italy. The Celts also expanded down the Danube river and its tributaries. One of the most influential tribes, the Scordisci, had established their capital at Singidunum in the 3rd century BC, which is present-day Belgrade, Serbia. The concentration of hill-forts and cemeteries shows a density of population in the Tisza valley of modern-day Vojvodina, Serbia, Hungary and into Ukraine. Expansion into Romania was however blocked by the Dacians. The Serdi were a Celtic tribe inhabiting Thrace. They were located around and founded Serdika (, , ), now Sofia in Bulgaria, which reflects their ethnonym. They would have established themselves in this area during the Celtic migrations at the end of the 4th century BC, though there is no evidence for their existence before the 1st century BC. "Serdi" are among traditional tribal names reported into the Roman era. They were gradually Thracianized over the centuries but retained their Celtic character in material culture up to a late date. According to other sources they may have been simply of Thracian origin, according to others they may have become of mixed Thraco-Celtic origin. Further south, Celts settled in Thrace (Bulgaria), which they ruled for over a century, and Anatolia, where they settled as the Galatians "(see also: Gallic Invasion of Greece)". Despite their geographical isolation from the rest of the Celtic world, the Galatians maintained their Celtic language for at least 700 years. St Jerome, who visited Ancyra (modern-day Ankara) in 373 AD, likened their language to that of the Treveri of northern Gaul. For Venceslas Kruta, Galatia in central Turkey was an area of dense Celtic settlement. The Boii tribe gave their name to Bohemia, Bologna and possibly Bavaria, and Celtic artefacts and cemeteries have been discovered further east in what is now Poland and Slovakia. A Celtic coin (Biatec) from Bratislava's mint was displayed on the old Slovak 5-crown coin. As there is no archaeological evidence for large-scale invasions in some of the other areas, one current school of thought holds that Celtic language and culture spread to those areas by contact rather than invasion. However, the Celtic invasions of Italy and the expedition in Greece and western Anatolia, are well documented in Greek and Latin history. There are records of Celtic mercenaries in Egypt serving the Ptolemies. Thousands were employed in 283–246 BC and they were also in service around 186 BC. They attempted to overthrow Ptolemy II. All Celtic languages extant today belong to the Insular Celtic languages, derived from the Celtic languages spoken in Iron Age Britain and Ireland. They were separated into a Goidelic and a Brythonic branch from an early period. Linguists have been arguing for many years whether a Celtic language came to Britain and Ireland and then split or whether there were two separate "invasions". The older view of prehistorians was that the Celtic influence in the British Isles was the result of successive invasions from the European continent by diverse Celtic-speaking peoples over the course of several centuries, accounting for the P-Celtic vs. Q-Celtic isogloss. This view has been challenged by the hypothesis that the Celtic languages of the British Isles form a phylogenetic Insular Celtic dialect group. In the 19th and 20th centuries, scholars commonly dated the "arrival" of Celtic culture in Britain (via an invasion model) to the 6th century BC, corresponding to archaeological evidence of Hallstatt influence and the appearance of chariot burials in what is now England. Some Iron Age migration does seem to have occurred but the nature of the interactions with the indigenous populations of the isles is unknown. According to this model, by about the 6th century (Sub-Roman Britain), most of the inhabitants of the Isles were speaking Celtic languages of either the Goidelic or the Brythonic branch. Since the late 20th century, a new model has emerged (championed by archaeologists such as Barry Cunliffe and Celtic historians such as John T. Koch) which places the emergence of Celtic culture in Britain much earlier, in the Bronze Age, and credits its spread not to invasion, but due to a gradual emergence "in situ" out of Proto-Indo-European culture (perhaps introduced to the region by the Bell Beaker People, and enabled by an extensive network of contacts that existed between the peoples of Britain and Ireland and those of the Atlantic seaboard. Classical writers did not apply the terms ("Keltoi") or "Celtae" to the inhabitants of Britain or Ireland, leading a number of scholars to question the use of the term Celt to describe the Iron Age inhabitants of those islands. The first historical account of the islands of Britain and Ireland was by Pytheas, a Greek from the city of Massalia, who around 310–306 BC, sailed around what he called the "Pretannikai nesoi", which can be translated as the "Pretannic Isles". In general, classical writers referred to the inhabitants of Britain as Pretannoi or Britanni. Strabo, writing in the Roman era, clearly distinguished between the Celts and Britons. Under Caesar the Romans conquered Celtic Gaul, and from Claudius onward the Roman empire absorbed parts of Britain. Roman local government of these regions closely mirrored pre-Roman tribal boundaries, and archaeological finds suggest native involvement in local government. The native peoples under Roman rule became Romanised and keen to adopt Roman ways. Celtic art had already incorporated classical influences, and surviving Gallo-Roman pieces interpret classical subjects or keep faith with old traditions despite a Roman overlay. The Roman occupation of Gaul, and to a lesser extent of Britain, led to Roman-Celtic syncretism. In the case of the continental Celts, this eventually resulted in a language shift to Vulgar Latin, while the Insular Celts retained their language. There was also considerable cultural influence exerted by Gaul on Rome, particularly in military matters and horsemanship, as the Gauls often served in the Roman cavalry. The Romans adopted the Celtic cavalry sword, the spatha, and Epona, the Celtic horse goddess. To the extent that sources are available, they depict a pre-Christian Iron Age Celtic social structure based formally on class and kingship, although this may only have been a particular late phase of organization in Celtic societies. Patron-client relationships similar to those of Roman society are also described by Caesar and others in the Gaul of the 1st century BC. In the main, the evidence is of tribes being led by kings, although some argue that there is also evidence of oligarchical republican forms of government eventually emerging in areas which had close contact with Rome. Most descriptions of Celtic societies portray them as being divided into three groups: a warrior aristocracy; an intellectual class including professions such as druid, poet, and jurist; and everyone else. In historical times, the offices of high and low kings in Ireland and Scotland were filled by election under the system of tanistry, which eventually came into conflict with the feudal principle of primogeniture in which succession goes to the first-born son. Little is known of family structure among the Celts. Patterns of settlement varied from decentralised to urban. The popular stereotype of non-urbanised societies settled in hillforts and duns, drawn from Britain and Ireland (there are about 3,000 hill forts known in Britain) contrasts with the urban settlements present in the core Hallstatt and La Tène areas, with the many significant "oppida" of Gaul late in the first millennium BC, and with the towns of Gallia Cisalpina. Slavery, as practised by the Celts, was very likely similar to the better documented practice in ancient Greece and Rome. Slaves were acquired from war, raids, and penal and debt servitude. Slavery was hereditary, though manumission was possible. The Old Irish and Welsh words for 'slave', "cacht" and "caeth" respectively, are cognate with Latin "captus" 'captive' suggesting that the slave trade was an early means of contact between Latin and Celtic societies. In the Middle Ages, slavery was especially prevalent in the Celtic countries. Manumissions were discouraged by law and the word for "female slave", "cumal", was used as a general unit of value in Ireland. Archaeological evidence suggests that the pre-Roman Celtic societies were linked to the network of overland trade routes that spanned Eurasia. Archaeologists have discovered large prehistoric trackways crossing bogs in Ireland and Germany. Due to their substantial nature, these are believed to have been created for wheeled transport as part of an extensive roadway system that facilitated trade. The territory held by the Celts contained tin, lead, iron, silver and gold. Celtic smiths and metalworkers created weapons and jewellery for international trade, particularly with the Romans. The myth that the Celtic monetary system consisted of wholly barter is a common one, but is in part false. The monetary system was complex and is still not understood (much like the late Roman coinages), and due to the absence of large numbers of coin items, it is assumed that "proto-money" was used. This included bronze items made from the early La Tène period and onwards, which were often in the shape of axeheads, rings, or bells. Due to the large number of these present in some burials, it is thought they had a relatively high monetary value, and could be used for "day to day" purchases. Low-value coinages of potin, a bronze alloy with high tin content, were minted in most Celtic areas of the continent and in South-East Britain prior to the Roman conquest of these lands. Higher-value coinages, suitable for use in trade, were minted in gold, silver, and high-quality bronze. Gold coinage was much more common than silver coinage, despite being worth substantially more, as while there were around 100 mines in Southern Britain and Central France, silver was more rarely mined. This was due partly to the relative sparsity of mines and the amount of effort needed for extraction compared to the profit gained. As the Roman civilisation grew in importance and expanded its trade with the Celtic world, silver and bronze coinage became more common. This coincided with a major increase in gold production in Celtic areas to meet the Roman demand, due to the high value Romans put on the metal. The large number of gold mines in France is thought to be a major reason why Caesar invaded. There are only very limited records from pre-Christian times written in Celtic languages. These are mostly inscriptions in the Roman and sometimes Greek alphabets. The Ogham script, an Early Medieval alphabet, was mostly used in early Christian times in Ireland and Scotland (but also in Wales and England), and was only used for ceremonial purposes such as inscriptions on gravestones. The available evidence is of a strong oral tradition, such as that preserved by bards in Ireland, and eventually recorded by monasteries. Celtic art also produced a great deal of intricate and beautiful metalwork, examples of which have been preserved by their distinctive burial rites. In some regards the Atlantic Celts were conservative: for example, they still used chariots in combat long after they had been reduced to ceremonial roles by the Greeks and Romans. However, despite being outdated, Celtic chariot tactics were able to repel the invasion of Britain attempted by Julius Caesar. According to Diodorus Siculus: During the later Iron Age the Gauls generally wore long-sleeved shirts or tunics and long trousers (called "braccae" by the Romans). Clothes were made of wool or linen, with some silk being used by the rich. Cloaks were worn in the winter. Brooches and armlets were used, but the most famous item of jewellery was the torc, a neck collar of metal, sometimes gold. The horned Waterloo Helmet in the British Museum, which long set the standard for modern images of Celtic warriors, is in fact a unique survival, and may have been a piece for ceremonial rather than military wear. Very few reliable sources exist regarding Celtic views on gender divisions and societal status, though some archaeological evidence does suggest that their views of gender roles may differ from contemporary and less egalitarian classical counterparts of the Roman era. There are some general indications from Iron Age burial sites in the Champagne and Bourgogne regions of Northeastern France suggesting that women may have had roles in combat during the earlier "La Tène" period. However, the evidence is far from conclusive. Examples of individuals buried with both female jewellery and weaponry have been identified, such as the Vix Grave, and there are questions about the gender of some skeletons that were buried with warrior assemblages. However, it has been suggested that "the weapons may indicate rank instead of masculinity". Among the insular Celts, there is a greater amount of historic documentation to suggest warrior roles for women. In addition to commentary by Tacitus about Boudica, there are indications from later period histories that also suggest a more substantial role for "women as warriors", in symbolic if not actual roles. Posidonius and Strabo described an island of women where men could not venture for fear of death, and where the women ripped each other apart. Other writers, such as Ammianus Marcellinus and Tacitus, mentioned Celtic women inciting, participating in, and leading battles. Posidonius' anthropological comments on the Celts had common themes, primarily primitivism, extreme ferocity, cruel sacrificial practices, and the strength and courage of their women. Under Brehon Law, which was written down in early Medieval Ireland after conversion to Christianity, a woman had the right to divorce her husband and gain his property if he was unable to perform his marital duties due to impotence, obesity, homosexual inclination or preference for other women. Classical literature records the views of the Celts' neighbours, though historians are not sure how much relation to reality these had. According to Aristotle, most "belligerent nations" were strongly influenced by their women, but the Celts were unusual because their men openly preferred male lovers ("Politics" II 1269b). H. D. Rankin in "Celts and the Classical World" notes that "Athenaeus echoes this comment (603a) and so does Ammianus (30.9). It seems to be the general opinion of antiquity." In book XIII of his "Deipnosophists", the Roman Greek rhetorician and grammarian Athenaeus, repeating assertions made by Diodorus Siculus in the 1st century BC (Bibliotheca historica 5:32), wrote that Celtic women were beautiful but that the men preferred to sleep together. Diodorus went further, stating that "the young men will offer themselves to strangers and are insulted if the offer is refused". Rankin argues that the ultimate source of these assertions is likely to be Posidonius and speculates that these authors may be recording male "bonding rituals". The sexual freedom of women in Britain was noted by Cassius Dio: There are instances recorded where women participated both in warfare and in kingship, although they were in the minority in these areas. Plutarch reports that Celtic women acted as ambassadors to avoid a war among Celts chiefdoms in the Po valley during the 4th century BC. Celtic art is generally used by art historians to refer to art of the La Tène period across Europe, while the Early Medieval art of Britain and Ireland, that is what "Celtic art" evokes for much of the general public, is called Insular art in art history. Both styles absorbed considerable influences from non-Celtic sources, but retained a preference for geometrical decoration over figurative subjects, which are often extremely stylised when they do appear; narrative scenes only appear under outside influence. Energetic circular forms, triskeles and spirals are characteristic. Much of the surviving material is in precious metal, which no doubt gives a very unrepresentative picture, but apart from Pictish stones and the Insular high crosses, large monumental sculpture, even with decorative carving, is very rare; possibly it was originally common in wood. Celts were also able to create developed musical instruments such as the carnyces, these famous war trumpets used before the battle to frighten the enemy, as the best preserved found in Tintignac (Gaul) in 2004 and which were decorated with a boar head or a snake head. The interlace patterns that are often regarded as typical of "Celtic art" were characteristic of the whole of the British Isles, a style referred to as Insular art, or Hiberno-Saxon art. This artistic style incorporated elements of La Tène, Late Roman, and, most importantly, animal Style II of Germanic Migration Period art. The style was taken up with great skill and enthusiasm by Celtic artists in metalwork and illuminated manuscripts. Equally, the forms used for the finest Insular art were all adopted from the Roman world: Gospel books like the Book of Kells and Book of Lindisfarne, chalices like the Ardagh Chalice and Derrynaflan Chalice, and penannular brooches like the Tara Brooch. These works are from the period of peak achievement of Insular art, which lasted from the 7th to the 9th centuries, before the Viking attacks sharply set back cultural life. In contrast the less well known but often spectacular art of the richest earlier Continental Celts, before they were conquered by the Romans, often adopted elements of Roman, Greek and other "foreign" styles (and possibly used imported craftsmen) to decorate objects that were distinctively Celtic. After the Roman conquests, some Celtic elements remained in popular art, especially Ancient Roman pottery, of which Gaul was actually the largest producer, mostly in Italian styles, but also producing work in local taste, including figurines of deities and wares painted with animals and other subjects in highly formalised styles. Roman Britain also took more interest in enamel than most of the Empire, and its development of champlevé technique was probably important to the later Medieval art of the whole of Europe, of which the energy and freedom of Insular decoration was an important element. Rising nationalism brought Celtic revivals from the 19th century. Tribal warfare appears to have been a regular feature of Celtic societies. While epic literature depicts this as more of a sport focused on raids and hunting rather than organised territorial conquest, the historical record is more of tribes using warfare to exert political control and harass rivals, for economic advantage, and in some instances to conquer territory. The Celts were described by classical writers such as Strabo, Livy, Pausanias, and Florus as fighting like "wild beasts", and as hordes. Dionysius said that their Such descriptions have been challenged by contemporary historians. Polybius (2.33) indicates that the principal Celtic weapon was a long bladed sword which was used for hacking edgewise rather than stabbing. Celtic warriors are described by Polybius and Plutarch as frequently having to cease fighting in order to straighten their sword blades. This claim has been questioned by some archaeologists, who note that Noric steel, steel produced in Celtic Noricum, was famous in the Roman Empire period and was used to equip the Roman military. However, Radomir Pleiner, in "The Celtic Sword" (1993) argues that "the metallographic evidence shows that Polybius was right up to a point", as around one third of surviving swords from the period might well have behaved as he describes. Polybius also asserts that certain of the Celts fought naked, "The appearance of these naked warriors was a terrifying spectacle, for they were all men of splendid physique and in the prime of life." According to Livy, this was also true of the Celts of Asia Minor. Celts had a reputation as head hunters. According to Paul Jacobsthal, "Amongst the Celts the human head was venerated above all else, since the head was to the Celt the soul, centre of the emotions as well as of life itself, a symbol of divinity and of the powers of the other-world." Arguments for a Celtic cult of the severed head include the many sculptured representations of severed heads in La Tène carvings, and the surviving Celtic mythology, which is full of stories of the severed heads of heroes and the saints who carry their own severed heads, right down to "Sir Gawain and the Green Knight", where the Green Knight picks up his own severed head after Gawain has struck it off, just as St. Denis carried his head to the top of Montmartre. Physical evidence exists for the ritual importance of the severed head at the religious centre at Roquepertuse (southern France), destroyed by the Romans in 124 BC, where stone pillars with prominent niches for displaying severed heads were found. A further example of this regeneration after beheading lies in the tales of Connemara's St. Feichin, who after being beheaded by Viking pirates carried his head to the Holy Well on Omey Island and on dipping the head into the well placed it back upon his neck and was restored to full health. Diodorus Siculus, in his 1st-century "History" had this to say about Celtic head-hunting: In "Gods and Fighting Men", Lady Gregory's Celtic Revival translation of Irish mythology, heads of men killed in battle are described in the beginning of the story "The Fight with the Fir Bolgs" as pleasing to Macha, one aspect of the war goddess Morrigu. Like other European Iron Age tribal societies, the Celts practised a polytheistic religion. Many Celtic gods are known from texts and inscriptions from the Roman period. Rites and sacrifices were carried out by priests known as druids. The Celts did not see their gods as having human shapes until late in the Iron Age. Celtic shrines were situated in remote areas such as hilltops, groves, and lakes. Celtic religious patterns were regionally variable; however, some patterns of deity forms, and ways of worshipping these deities, appeared over a wide geographical and temporal range. The Celts worshipped both gods and goddesses. In general, Celtic gods were deities of particular skills, such as the many-skilled Lugh and Dagda, while goddesses were associated with natural features, particularly rivers (such as Boann, goddess of the River Boyne). This was not universal, however, as goddesses such as Brighid and The Morrígan were associated with both natural features (holy wells and the River Unius) and skills such as blacksmithing and healing. Triplicity is a common theme in Celtic cosmology, and a number of deities were seen as threefold. This trait is exhibited by The Three Mothers, a group of goddesses worshipped by many Celtic tribes (with regional variations). The Celts had hundreds of deities, some of which were unknown outside a single family or tribe, while others were popular enough to have a following that crossed lingual and cultural barriers. For instance, the Irish god Lugh, associated with storms, lightning, and culture, is seen in similar forms as Lugos in Gaul and Lleu in Wales. Similar patterns are also seen with the continental Celtic horse goddess Epona and what may well be her Irish and Welsh counterparts, Macha and Rhiannon, respectively. Roman reports of the druids mention ceremonies being held in sacred groves. La Tène Celts built temples of varying size and shape, though they also maintained shrines at sacred trees and votive pools. Druids fulfilled a variety of roles in Celtic religion, serving as priests and religious officiants, but also as judges, sacrificers, teachers, and lore-keepers. Druids organised and ran religious ceremonies, and they memorised and taught the calendar. Other classes of druids performed ceremonial sacrifices of crops and animals for the perceived benefit of the community. The Coligny calendar, which was found in 1897 in Coligny, Ain, was engraved on a bronze tablet, preserved in 73 fragments, that originally was wide and high (Lambert p. 111). Based on the style of lettering and the accompanying objects, it probably dates to the end of the 2nd century. It is written in Latin inscriptional capitals, and is in the Gallic language. The restored tablet contains 16 vertical columns, with 62 months distributed over 5 years. The French archaeologist J. Monard speculated that it was recorded by druids wishing to preserve their tradition of timekeeping in a time when the Julian calendar was imposed throughout the Roman Empire. However, the general form of the calendar suggests the public peg calendars (or "parapegmata") found throughout the Greek and Roman world. The Roman invasion of Gaul brought a great deal of Celtic peoples into the Roman Empire. Roman culture had a profound effect on the Celtic tribes which came under the empire's control. Roman influence led to many changes in Celtic religion, the most noticeable of which was the weakening of the druid class, especially religiously; the druids were to eventually disappear altogether. Romano-Celtic deities also began to appear: these deities often had both Roman and Celtic attributes, combined the names of Roman and Celtic deities, and/or included couples with one Roman and one Celtic deity. Other changes included the adaptation of the Jupiter Column, a sacred column set up in many Celtic regions of the empire, primarily in northern and eastern Gaul. Another major change in religious practice was the use of stone monuments to represent gods and goddesses. The Celts had only created wooden idols (including monuments carved into trees, which were known as sacred poles) previously to Roman conquest. While the regions under Roman rule adopted Christianity along with the rest of the Roman empire, unconquered areas of Ireland and Scotland began to move from Celtic polytheism to Christianity in the 5th century. Ireland was converted by missionaries from Britain, such as Saint Patrick. Later missionaries from Ireland were a major source of missionary work in Scotland, Anglo-Saxon parts of Britain, and central Europe (see Hiberno-Scottish mission). Celtic Christianity, the forms of Christianity that took hold in Britain and Ireland at this time, had for some centuries only limited and intermittent contact with Rome and continental Christianity, as well as some contacts with Coptic Christianity. Some elements of Celtic Christianity developed, or retained, features that made them distinct from the rest of Western Christianity, most famously their conservative method of calculating the date of Easter. In 664, the Synod of Whitby began to resolve these differences, mostly by adopting the current Roman practices, which the Gregorian Mission from Rome had introduced to Anglo-Saxon England. In a March 2019 genetic study published in "Science", three Celtiberians buried at La Hoya, Salamanca between 400 BC and 195 BC were examined. They were find to be harboring elevated levels of north-central European ancestry compared to non-Celtic populations of Iberia. One of the males examined was found to be a carrier of the paternal haplogroup I2a1a1a. Geography Organisations
https://en.wikipedia.org/wiki?curid=6546
Claude Monet Oscar-Claude Monet (, , ; 14 November 1840 – 5 December 1926) was a French painter, a founder of French Impressionist painting and the most consistent and prolific practitioner of the movement's philosophy of expressing one's perceptions before nature, especially as applied to plein air landscape painting. The term "Impressionism" is derived from the title of his painting "Impression, soleil levant" ("Impression, Sunrise"), which was exhibited in 1874 in the first of the independent exhibitions mounted by Monet and his associates as an alternative to the Salon de Paris. Monet's ambition of documenting the French countryside led him to adopt a method of painting the same scene many times in order to capture the changing of light and the passing of the seasons. From 1883, Monet lived in Giverny, where he purchased a house and property and began a vast landscaping project which included lily ponds that would become the subjects of his best-known works. He began painting the water lilies in 1899, first in vertical views with a Japanese bridge as a central feature and later in the series of large-scale paintings that was to occupy him continuously for the next 20 years of his life. Claude Monet was born on 14 November 1840 on the fifth floor of 45 rue Laffitte, in the 9th arrondissement of Paris. He was the second son of Claude Adolphe Monet and Louise Justine Aubrée Monet, both of them second-generation Parisians. On 20 May 1841, he was baptized in the local parish church, Notre-Dame-de-Lorette, as Oscar-Claude, but his parents called him simply Oscar. (He signed his juvenilia "O. Monet".) Despite being baptized Catholic, Monet later became an atheist. In 1845, his family moved to Le Havre in Normandy. His father wanted him to go into the family's ship-chandling and grocery business, but Monet wanted to become an artist. His mother was a singer, and supported Monet's desire for a career in art. On 1 April 1851, Monet entered Le Havre secondary school of the arts. Locals knew him well for his charcoal caricatures, which he would sell for ten to twenty francs. Monet also undertook his first drawing lessons from Jacques-François Ochard, a former student of Jacques-Louis David. On the beaches of Normandy around 1856 he met fellow artist Eugène Boudin, who became his mentor and taught him to use oil paints. Boudin taught Monet "en plein air" (outdoor) techniques for painting. Both were influenced by Johan Barthold Jongkind. On 28 January 1857, his mother died. At the age of sixteen, he left school and went to live with his widowed, childless aunt, Marie-Jeanne Lecadre. When Monet traveled to Paris to visit the Louvre, he witnessed painters copying from the old masters. Having brought his paints and other tools with him, he would instead go and sit by a window and paint what he saw. Monet was in Paris for several years and met other young painters, including Édouard Manet and others who would become friends and fellow Impressionists. After drawing a low ballot number in March 1861, Monet was drafted into the First Regiment of African Light Cavalry ("Chasseurs d'Afrique") in Algeria for a seven-year period of military service. His prosperous father could have purchased Monet's exemption from conscription but declined to do so when his son refused to give up painting. While in Algeria, Monet did only a few sketches of casbah scenes, a single landscape, and several portraits of officers, all of which have been lost. In a "Le Temps" interview of 1900 however he commented that the light and vivid colours of North Africa "contained the germ of my future researches". After about a year of garrison duty in Algiers, Monet contracted typhoid fever and briefly went absent without leave. Following convalescence, Monet's aunt intervened to remove him from the army if he agreed to complete a course at an art school. It is possible that the Dutch painter Johan Barthold Jongkind, whom Monet knew, may have prompted his aunt on this matter. Disillusioned with the traditional art taught at art schools, in 1862 Monet became a student of Charles Gleyre in Paris, where he met Pierre-Auguste Renoir, Frédéric Bazille and Alfred Sisley. Together they shared new approaches to art, painting the effects of light "en plein air" with broken colour and rapid brushstrokes, in what later came to be known as Impressionism. In January 1865 Monet was working on a version of "Le déjeuner sur l'herbe", aiming to present it for hanging at the Salon, which had rejected Manet's "Le déjeuner sur l'herbe" two years earlier. Monet's painting was very large and could not be completed in time. (It was later cut up, with parts now in different galleries.) Monet submitted instead a painting of "Camille" or "The Woman in the Green Dress" ("La femme à la robe verte"), one of many works using his future wife, Camille Doncieux, as his model. Both this painting and a small landscape were hung. The following year Monet used Camille for his model in "Women in the Garden", and "On the Bank of the Seine, Bennecourt" in 1868. Camille became pregnant and gave birth to their first child, Jean, in 1867. Monet and Camille married on 28 June 1870, just before the outbreak of the Franco-Prussian War, and, after their excursion to London and Zaandam, they moved to Argenteuil, in December 1871. During this time Monet painted various works of modern life. He and Camille lived in poverty for most of this period. Following the successful exhibition of some maritime paintings, and the winning of a silver medal at Le Havre, Monet's paintings were seized by creditors, from whom they were bought back by a shipping merchant, Gaudibert, who was also a patron of Boudin. From the late 1860s, Monet and other like-minded artists met with rejection from the conservative Académie des Beaux-Arts, which held its annual exhibition at the Salon de Paris. During the latter part of 1873, Monet, Pierre-Auguste Renoir, Camille Pissarro, and Alfred Sisley organized the (Anonymous Society of Painters, Sculptors, and Engravers) to exhibit their artworks independently. At their first exhibition, held in April 1874, Monet exhibited the work that was to give the group its lasting name. He was inspired by the style and subject matter of previous modern painters Camille Pissarro and Edouard Manet. "Impression, Sunrise" was painted in 1872, depicting a Le Havre port landscape. From the painting's title the art critic Louis Leroy, in his review, "L'Exposition des Impressionnistes," which appeared in "Le Charivari", coined the term "Impressionism". It was intended as disparagement but the Impressionists appropriated the term for themselves. After the outbreak of the Franco-Prussian War (19 July 1870), Monet and his family took refuge in England in September 1870, where he studied the works of John Constable and Joseph Mallord William Turner, both of whose landscapes would serve to inspire Monet's innovations in the study of colour. In the spring of 1871, Monet's works were refused authorisation for inclusion in the Royal Academy exhibition. In May 1871, he left London to live in Zaandam, in the Netherlands, where he made twenty-five paintings (and the police suspected him of revolutionary activities). He also paid a first visit to nearby Amsterdam. In October or November 1871, he returned to France. From December 1871 to 1878 he lived at Argenteuil, a village on the right bank of the Seine river near Paris, and a popular Sunday-outing destination for Parisians, where he painted some of his best-known works. In 1873, Monet purchased a small boat equipped to be used as a floating studio. From the boat studio Monet painted landscapes and also portraits of Édouard Manet and his wife; Manet in turn depicted Monet painting aboard the boat, accompanied by Camille, in 1874. In 1874, he briefly returned to Holland. The first Impressionist exhibition was held in 1874 at 35 boulevard des Capucines, Paris, from 15 April to 15 May. The primary purpose of the participants was not so much to promote a new style, but to free themselves from the constraints of the Salon de Paris. The exhibition, open to anyone prepared to pay 60 francs, gave artists the opportunity to show their work without the interference of a jury. Renoir chaired the hanging committee and did most of the work himself, as others members failed to present themselves. In addition to "" (pictured above), Monet presented four oil paintings and seven pastels. Among the paintings he displayed was "The Luncheon" (1868), which features Camille Doncieux and Jean Monet, and which had been rejected by the Paris Salon of 1870. Also in this exhibition was a painting titled "Boulevard des Capucines", a painting of the boulevard done from the photographer Nadar's apartment at no. 35. Monet painted the subject twice, and it is uncertain which of the two pictures, that now in the Pushkin Museum in Moscow, or that in the Nelson-Atkins Museum of Art in Kansas City, was the painting that appeared in the groundbreaking 1874 exhibition, though more recently the Moscow picture has been favoured. Altogether, 165 works were exhibited in the exhibition, including 4 oils, 2 pastels and 3 watercolours by Morisot; 6 oils and 1 pastel by Renoir; 10 works by Degas; 5 by Pissarro; 3 by Cézanne; and 3 by Guillaumin. Several works were on loan, including Cézanne's "Modern Olympia", Morisot's "Hide and Seek" (owned by Manet) and 2 landscapes by Sisley that had been purchased by Durand-Ruel. The total attendance is estimated at 3500, and some works did sell, though some exhibitors had placed their prices too high. Pissarro was asking 1000 francs for "The Orchard" and Monet the same for "Impression: Sunrise", neither of which sold. Renoir failed to obtain the 500 francs he was asking for "La Loge", but later sold it for 450 francs to Père Martin, dealer and supporter of the group. In 1876, Camille Monet became ill with tuberculosis. Their second son, Michel, was born on 17 March 1878. This second child weakened her already fading health. In the summer of that year, the family moved to the village of Vétheuil where they shared a house with the family of Ernest Hoschedé, a wealthy department store owner and patron of the arts. In 1878, Camille Monet was diagnosed with uterine cancer. She died on 5 September 1879 at the age of thirty-two. Monet made a study in oils of his dead wife. Many years later, Monet confessed to his friend Georges Clemenceau that his need to analyse colours was both the joy and torment of his life. He explained, I one day found myself looking at my beloved wife's dead face and just systematically noting the colours according to an automatic reflex! John Berger describes the work as "a blizzard of white, grey, purplish paint ... a terrible blizzard of loss which will forever efface her features. In fact there can be very few death-bed paintings which have been so intensely felt or subjectively expressive." After several difficult months following the death of Camille, Monet began to create some of his best paintings of the 19th century. During the early 1880s, Monet painted several groups of landscapes and seascapes in what he considered to be campaigns to document the French countryside. These began to evolve into series of pictures in which he documented the same scene many times in order to capture the changing of light and the passing of the seasons. Monet's friend Ernest Hoschedé became bankrupt, and left in 1878 for Belgium. After the death of Camille Monet in September 1879, and while Monet continued to live in the house in Vétheuil, Alice Hoschedé helped Monet to raise his two sons, Jean and Michel. She took them to Paris to live alongside her own six children, Blanche (who married Jean Monet), Germaine, Suzanne, Marthe, Jean-Pierre, and Jacques. In the spring of 1880, Alice Hoschedé and all the children left Paris and rejoined Monet at Vétheuil. In 1881, all of them moved to Poissy, which Monet hated. In April 1883, looking out the window of the little train between Vernon and Gasny, he discovered Giverny in Normandy. Monet, Alice Hoschedé and the children moved to Vernon, then to the house in Giverny, where he planted a large garden and where he painted for much of the rest of his life. Following the death of her estranged husband, Monet married Alice Hoschedé in 1892. Monet rented and eventually purchased a house and gardens in Giverny. At the beginning of May 1883, Monet and his large family rented the home and from a local landowner. The house was situated near the main road between the towns of Vernon and Gasny at Giverny. There was a barn that doubled as a painting studio, orchards and a small garden. The house was close enough to the local schools for the children to attend, and the surrounding landscape offered many suitable motifs for Monet's work. The family worked and built up the gardens, and Monet's fortunes began to change for the better as his dealer, Paul Durand-Ruel, had increasing success in selling his paintings. By November 1890, Monet was prosperous enough to buy the house, the surrounding buildings and the land for his gardens. During the 1890s, Monet built a greenhouse and a second studio, a spacious building well lit with skylights. Monet wrote daily instructions to his gardener, precise designs and layouts for plantings, and invoices for his floral purchases and his collection of botany books. As Monet's wealth grew, his garden evolved. He remained its architect, even after he hired seven gardeners. Monet purchased additional land with a water meadow. In 1893 he began a vast landscaping project which included lily ponds that would become the subjects of his best-known works. White water lilies local to France were planted along with imported cultivars from South America and Egypt, resulting in a range of colours including yellow, blue and white lilies that turned pink with age. In 1899 he began painting the water lilies, first in vertical views with a Japanese bridge as a central feature, and later in the series of large-scale paintings that was to occupy him continuously for the next 20 years of his life. This scenery, with its alternating light and mirror-like reflections, became an integral part of his work. By the mid-1910s Monet had achieved: Monet's second wife, Alice, died in 1911, and his oldest son Jean, who had married Alice's daughter Blanche, Monet's particular favourite, died in 1914. After Alice died, Blanche looked after and cared for Monet. It was during this time that Monet began to develop the first signs of cataracts. During World War I, in which his younger son Michel served and his friend and admirer Georges Clemenceau led the French nation, Monet painted a series of weeping willow trees as homage to the French fallen soldiers. In 1923, he underwent two operations to remove his cataracts. The paintings done while the cataracts affected his vision have a general reddish tone, which is characteristic of the vision of cataract victims. It may also be that after surgery he was able to see certain ultraviolet wavelengths of light that are normally excluded by the lens of the eye; this may have had an effect on the colours he perceived. After his operations he even repainted some of these paintings, with bluer water lilies than before. Monet died of lung cancer on 5 December 1926 at the age of 86 and is buried in the Giverny church cemetery. Monet had insisted that the occasion be simple; thus only about fifty people attended the ceremony. At his funeral, his long-time friend Georges Clemenceau removed the black cloth draped over the coffin, stating, "No black for Monet!" and replaced it with a flower-patterned cloth. Monet did not leave a will and so his son Michel inherited his entire estate. Monet's home, garden, and waterlily pond were bequeathed by Michel to the French Academy of Fine Arts (part of the Institut de France) in 1966. Through the "Fondation Claude Monet", the house and gardens were opened for visits in 1980, following restoration. In addition to souvenirs of Monet and other objects of his life, the house contains his collection of Japanese woodcut prints. The house and garden, along with the Museum of Impressionism, are major attractions in Giverny, which hosts tourists from all over the world. Monet has been described as "the driving force behind Impressionism". Crucial to the art of the Impressionist painters was the understanding of the effects of light on the local colour of objects, and the effects of the juxtaposition of colours with each other. Monet's long career as a painter was spent in the pursuit of this aim. In 1856, his chance meeting with Eugene Boudin, a painter of small beach scenes, opened his eyes to the possibility of plein-air painting. From that time, with a short interruption for military service, he dedicated himself to searching for new and improved methods of painterly expression. To this end, as a young man, he visited the Paris Salon and familiarised himself with the works of older painters, and made friends with other young artists. The five years that he spent at Argenteuil, spending much time on the River Seine in a little floating studio, were formative in his study of the effects of light and reflections. He began to think in terms of colours and shapes rather than scenes and objects. He used bright colours in dabs and dashes and squiggles of paint. Having rejected the academic teachings of Gleyre's studio, he freed himself from theory, saying "I like to paint as a bird sings." In 1877 a series of paintings at St-Lazare Station had Monet looking at smoke and steam and the way that they affected colour and visibility, being sometimes opaque and sometimes translucent. He was to further use this study in the painting of the effects of mist and rain on the landscape. The study of the effects of atmosphere was to evolve into a number of series of paintings in which Monet repeatedly painted the same subject (such as his water lilies series) in different lights, at different hours of the day, and through the changes of weather and season. This process began in the 1880s and continued until the end of his life in 1926. His first series exhibited as such was of Haystacks, painted from different points of view and at different times of the day. Fifteen of the paintings were exhibited at the Galerie Durand-Ruel in 1891. In 1892 he produced what is probably his best-known series, twenty-six views of "Rouen Cathedral". In these paintings Monet broke with painterly traditions by cropping the subject so that only a portion of the façade is seen on the canvas. The paintings do not focus on the grand Medieval building, but on the play of light and shade across its surface, transforming the solid masonry. Other series include "Poplars", "Mornings on the Seine", and the "Water Lilies" that were painted on his property at Giverny. Between 1883 and 1908, Monet traveled to the Mediterranean, where he painted landmarks, landscapes, and seascapes, including a series of paintings in Venice. In London he painted four series: "the Houses of Parliament, London", "Charing Cross Bridge", "Waterloo Bridge", and "Views of Westminster Bridge". Helen Gardner writes: In 2004, "London, the Parliament, Effects of Sun in the Fog" ("Londres, le Parlement, trouée de soleil dans le brouillard"; 1904), sold for US$20.1 million. In 2006, the journal "Proceedings of the Royal Society" published a paper providing evidence that these were painted in situ at St Thomas' Hospital over the river Thames. "Falaises près de Dieppe" ("Cliffs Near Dieppe") has been stolen on two separate occasions: once in 1998 (in which the museum's curator was convicted of the theft and jailed for five years and two months along with two accomplices) and most recently in August 2007. It was recovered in June 2008. Monet's "Le Pont du chemin de fer à Argenteuil", an 1873 painting of a railway bridge spanning the Seine near Paris, was bought by an anonymous telephone bidder for a record $41.4 million at Christie's auction in New York on 6 May 2008. The previous record for his painting stood at $36.5 million. A few weeks later, "Le bassin aux nymphéas" (from the water lilies series) sold at Christie's 24 June 2008 auction in London for £40,921,250 ($80,451,178), nearly doubling the record for the artist. This purchase represented one of the top 20 highest prices paid for a painting at the time. In October 2013, Monet's paintings, "L'Eglise de Vetheuil" and "Le Bassin aux Nympheas", became subjects of a legal case in New York against NY-based Vilma Bautista, one-time aide to Imelda Marcos, wife of dictator Ferdinand Marcos, after she sold "Le Bassin aux Nympheas" for $32 million to a Swiss buyer. The said Monet paintings, along with two others, were acquired by Imelda during her husband's presidency and allegedly bought using the nation's funds. Bautista's lawyer claimed that the aide sold the painting for Imelda but did not have a chance to give her the money. The Philippine government seeks the return of the painting. "Le Bassin aux Nympheas", also known as "Japanese Footbridge over the Water-Lily Pond at Giverny", is part of Monet's famed Water Lilies series.
https://en.wikipedia.org/wiki?curid=6548
Carthage Carthage was the center or capital city of the ancient Carthaginian civilization, on the eastern side of the Lake of Tunis in what is now the Tunis Governorate in Tunisia. Carthage was widely considered the most important trading hub of the Ancient Mediterranean and was arguably one of the most affluent cities of the classical world. The city developed from a Phoenician colony into the capital of a Punic empire which dominated large parts of the Southwest Mediterranean during the first millennium BC. The legendary Queen Alyssa or Dido is regarded as the founder of the city, though her historicity has been questioned. According to accounts by Timaeus of Tauromenium, she purchased from a local tribe the amount of land that could be covered by an oxhide. Cutting the skin into strips, she laid out her claim and founded an empire that would become, through the Punic Wars, the only existential threat to Rome until the coming of the Vandals several centuries later. The ancient city was destroyed by the Roman Republic in the Third Punic War in 146 BC and then re-developed as Roman Carthage, which became the major city of the Roman Empire in the province of Africa. The city was sacked and destroyed by Umayyad forces after the Battle of Carthage in 698 to prevent it from being reconquered by the Byzantine Empire. It remained occupied during the Muslim period and was used as a fort by the Muslims until the Hafsid period when it was taken by the Crusaders with its inhabitants massacred during the Eighth Crusade. The Hafsids decided to destroy its defenses so it could not be used as a base by a hostile power again. It also continued to function as an episcopal see. The regional power had shifted to Kairouan and the Medina of Tunis in the medieval period, until the early 20th century, when it began to develop into a coastal suburb of Tunis, incorporated as Carthage municipality in 1919. The archaeological site was first surveyed in 1830, by Danish consul Christian Tuxen Falbe. Excavations were performed in the second half of the 19th century by Charles Ernest Beulé and by Alfred Louis Delattre. The Carthage National Museum was founded in 1875 by Cardinal Charles Lavigerie. Excavations performed by French archaeologists in the 1920s first attracted an extraordinary amount of attention because of the evidence they produced for child sacrifice. There has been considerable disagreement among scholars concerning whether child sacrifice was practiced by ancient Carthage. The open-air Carthage Paleo-Christian Museum has exhibits excavated under the auspices of UNESCO from 1975 to 1984. The name "Carthage" is the Early Modern anglicisation of Middle French "Carthage" , from Latin ' and ' (cf. Greek "Karkhēdōn" () and Etruscan "*Carθaza") from the Punic ' "new city", implying it was a "new Tyre". The Latin adjective "pūnicus", meaning "Phoenician", is reflected in English in some borrowings from Latin—notably the Punic Wars and the Punic language. The Modern Standard Arabic form ("") is an adoption of French "Carthage", replacing an older local toponym reported as "Cartagenna" that directly continued the Latin name. Carthage was built on a promontory with sea inlets to the north and the south. The city's location made it master of the Mediterranean's maritime trade. All ships crossing the sea had to pass between Sicily and the coast of Tunisia, where Carthage was built, affording it great power and influence. Two large, artificial harbors were built within the city, one for harboring the city's massive navy of 220 warships and the other for mercantile trade. A walled tower overlooked both harbors. The city had massive walls, in length, longer than the walls of comparable cities. Most of the walls were located on the shore, thus could be less impressive, as Carthaginian control of the sea made attack from that direction difficult. The of wall on the isthmus to the west were truly massive and were never penetrated. Carthage was one of the largest cities of the Hellenistic period and was among the largest cities in preindustrial history. Whereas by AD 14, Rome had at least 750,000 inhabitants and in the following century may have reached 1 million, the cities of Alexandria and Antioch numbered only a few hundred thousand or less. According to the not-always-reliable history of Herodian, Carthage rivaled Alexandria for second place in the Roman empire. The Punic Carthage was divided into four equally sized residential areas with the same layout, had religious areas, market places, council house, towers, a theater, and a huge necropolis; roughly in the middle of the city stood a high citadel called the Byrsa. Surrounding Carthage were walls "of great strength" said in places to rise above 13 m, being nearly 10 m thick, according to ancient authors. To the west, three parallel walls were built. The walls altogether ran for about to encircle the city. The heights of the Byrsa were additionally fortified; this area being the last to succumb to the Romans in 146 BC. Originally the Romans had landed their army on the strip of land extending southward from the city. Outside the city walls of Carthage is the "Chora" or farm lands of Carthage. "Chora" encompassed a limited area: the north coastal "tell", the lower Bagradas river valley (inland from Utica), Cape Bon, and the adjacent "sahel" on the east coast. Punic culture here achieved the introduction of agricultural sciences first developed for lands of the eastern Mediterranean, and their adaptation to local African conditions. The "urban landscape" of Carthage is known in part from ancient authors, augmented by modern digs and surveys conducted by archeologists. The "first urban nucleus" dating to the seventh century, in area about , was apparently located on low-lying lands along the coast (north of the later harbors). As confirmed by archaeological excavations, Carthage was a "creation "ex nihilo"", built on 'virgin' land, and situated at what was then the end of a peninsula. Here among "mud brick walls and beaten clay floors" (recently uncovered) were also found extensive cemeteries, which yielded evocative grave goods like clay masks. "Thanks to this burial archaeology we know more about archaic Carthage than about any other contemporary city in the western Mediterranean." Already in the eighth century, fabric dyeing operations had been established, evident from crushed shells of murex (from which the 'Phoenician purple' was derived). Nonetheless, only a "meager picture" of the cultural life of the earliest pioneers in the city can be conjectured, and not much about housing, monuments or defenses. The Roman poet Virgil (70–19 BC) imagined early Carthage, when his legendary character Aeneas had arrived there: "Aeneas found, where lately huts had been, marvelous buildings, gateways, cobbled ways, and din of wagons. There the Tyrians were hard at work: laying courses for walls, rolling up stones to build the citadel, while others picked out building sites and plowed a boundary furrow. Laws were being enacted, magistrates and a sacred senate chosen. Here men were dredging harbors, there they laid the deep foundations of a theatre, and quarried massive pillars... ." The two inner harbours, named "cothon" in Punic, were located in the southeast; one being commercial, and the other for war. Their definite functions are not entirely known, probably for the construction, outfitting, or repair of ships, perhaps also loading and unloading cargo. Larger anchorages existed to the north and south of the city. North and west of the "cothon" were located several industrial areas, e.g., metalworking and pottery (e.g., for amphora), which could serve both inner harbours, and ships anchored to the south of the city. About the Byrsa, the citadel area to the north, considering its importance our knowledge of it is patchy. Its prominent heights were the scene of fierce combat during the fiery destruction of the city in 146 BC. The Byrsa was the reported site of the Temple of Eshmun (the healing god), at the top of a stairway of sixty steps. A temple of Tanit (the city's queen goddess) was likely situated on the slope of the 'lesser Byrsa' immediately to the east, which runs down toward the sea. Also situated on the Byrsa were luxury homes. South of the citadel, near the "cothon" was the "tophet", a special and very old cemetery, which when begun lay outside the city's boundaries. Here the "Salammbô" was located, the "Sanctuary of Tanit", not a temple but an enclosure for placing stone stelae. These were mostly short and upright, carved for funeral purposes. The presence of infant skeletons from here may indicate the occurrence of child sacrifice, as claimed in the Bible, although there has been considerable doubt among archeologists as to this interpretation and many consider it simply a cemetery devoted to infants. Probably the "tophet" burial fields were "dedicated at an early date, perhaps by the first settlers." Recent studies, on the other hand, indicate that child sacrifice was practiced by the Carthaginians. Between the sea-filled "cothon" for shipping and the Byrsa heights lay the "agora" [Greek: "market"], the city-state's central marketplace for business and commerce. The "agora" was also an area of public squares and plazas, where the people might formally assemble, or gather for festivals. It was the site of religious shrines, and the location of whatever were the major municipal buildings of Carthage. Here beat the heart of civic life. In this district of the Carthage, more probably, the ruling suffets presided, the council of elders convened, the tribunal of the 104 met, and justice was dispensed at trials in the open air. Early residential districts wrapped around the Byrsa from the south to the north east. Houses usually were whitewashed and blank to the street, but within were courtyards open to the sky. In these neighborhoods multistory construction later became common, some up to six stories tall according to an ancient Greek author. Several architectural floorplans of homes have been revealed by recent excavations, as well as the general layout of several city blocks. Stone stairs were set in the streets, and drainage was planned, e.g., in the form of soakways leaching into the sandy soil. Along the Byrsa's southern slope were located not only fine old homes, but also many of the earliest grave-sites, juxtaposed in small areas, interspersed with daily life. Artisan workshops were located in the city at sites north and west of the harbours. The location of three metal workshops (implied from iron slag and other vestiges of such activity) were found adjacent to the naval and commercial harbours, and another two were further up the hill toward the Byrsa citadel. Sites of pottery kilns have been identified, between the "agora" and the harbours, and further north. Earthenware often used Greek models. A fuller's shop for preparing woolen cloth (shrink and thicken) was evidently situated further to the west and south, then by the edge of the city. Carthage also produced objects of rare refinement. During the 4th and 3rd centuries, the sculptures of the sarcophagi became works of art. "Bronze engraving and stone-carving reached their zenith." The elevation of the land at the promontory on the seashore to the north-east (now called Sidi Bou Saïd), was twice as high above sea level as that at the Byrsa (100 m and 50 m). In between runs a ridge, several times reaching 50 m; it continues northwestward along the seashore, and forms the edge of a plateau-like area between the Byrsa and the sea. Newer urban developments lay here in these northern districts. Due to the Roman's leveling of the city, the original Punic urban landscape of Carthage was largely lost. Since 1982, French archaeologist Serge Lancel excavated a residential area of the Punic Carthage on top of Byrsa hill near the Forum of the Roman Carthage. The neighborhood can be dating back to early second century BC, and with its houses, shops, and private spaces, is significant for what it reveals about daily life of the Punic Carthage. The remains have been preserved under embankments, the substructures of the later Roman forum, whose foundation piles dot the district. The housing blocks are separated by a grid of straight streets about wide, with a roadway consisting of clay; "in situ" stairs compensate for the slope of the hill. Construction of this type presupposes organization and political will, and has inspired the name of the neighborhood, "Hannibal district", referring to the legendary Punic general or sufet (consul) at the beginning of the second century BC. The habitat is typical, even stereotypical. The street was often used as a storefront/shopfront; cisterns were installed in basements to collect water for domestic use, and a long corridor on the right side of each residence led to a courtyard containing a sump, around which various other elements may be found. In some places, the ground is covered with mosaics called punica pavement, sometimes using a characteristic red mortar. Punic culture and agricultural sciences, when arrived at Carthage from eastern Mediterranean, gradually adapted to the local African conditions. The merchant harbor at Carthage was developed after settlement of the nearby Punic town of Utica, and eventually the surrounding African countryside was brought into the orbit of the Punic urban centers, first commercially, then politically. Direct management over cultivation of neighbouring lands by Punic owners followed. A 28-volume work on agriculture written in Punic by Mago, a retired army general (c. 300), was translated into Latin and later into Greek. The original and both translations have been lost; however, some of Mago's text has survived in other Latin works. Olive trees (e.g., grafting), fruit trees (pomegranate, almond, fig, date palm), viniculture, bees, cattle, sheep, poultry, implements, and farm management were among the ancient topics which Mago discussed. As well, Mago addresses the wine-maker's art (here a type of sherry). In Punic farming society, according to Mago, the small estate owners were the chief producers. They were, two modern historians write, not absent landlords. Rather, the likely reader of Mago was "the master of a relatively modest estate, from which, by great personal exertion, he extracted the maximum yield." Mago counselled the rural landowner, for the sake of their own 'utilitarian' interests, to treat carefully and well their managers and farm workers, or their overseers and slaves. Yet elsewhere these writers suggest that rural land ownership provided also a new power base among the city's nobility, for those resident in their country villas. By many, farming was viewed as an alternative endeavour to an urban business. Another modern historian opines that more often it was the urban merchant of Carthage who owned rural farming land to some profit, and also to retire there during the heat of summer. It may seem that Mago anticipated such an opinion, and instead issued this contrary advice (as quoted by the Roman writer Columella): The man who acquires an estate must sell his house, lest he prefer to live in the town rather than in the country. Anyone who prefers to live in a town has no need of an estate in the country." "One who has bought land should sell his town house, so that he will have no desire to worship the household gods of the city rather than those of the country; the man who takes greater delight in his city residence will have no need of a country estate. The issues involved in rural land management also reveal underlying features of Punic society, its structure and stratification. The hired workers might be considered 'rural proletariat', drawn from the local Berbers. Whether there remained Berber landowners next to Punic-run farms is unclear. Some Berbers became sharecroppers. Slaves acquired for farm work were often prisoners of war. In lands outside Punic political control, independent Berbers cultivated grain and raised horses on their lands. Yet within the Punic domain that surrounded the city-state of Carthage, there were ethnic divisions in addition to the usual quasi feudal distinctions between lord and peasant, or master and serf. This inherent instability in the countryside drew the unwanted attention of potential invaders. Yet for long periods Carthage was able to manage these social difficulties. The many amphorae with Punic markings subsequently found about ancient Mediterranean coastal settlements testify to Carthaginian trade in locally made olive oil and wine. Carthage's agricultural production was held in high regard by the ancients, and rivaled that of Rome—they were once competitors, e.g., over their olive harvests. Under Roman rule, however, grain production ([wheat] and barley) for export increased dramatically in 'Africa'; yet these later fell with the rise in Roman Egypt's grain exports. Thereafter olive groves and vineyards were re-established around Carthage. Visitors to the several growing regions that surrounded the city wrote admiringly of the lush green gardens, orchards, fields, irrigation channels, hedgerows (as boundaries), as well as the many prosperous farming towns located across the rural landscape. Accordingly, the Greek author and compiler Diodorus Siculus (fl. 1st century BC), who enjoyed access to ancient writings later lost, and on which he based most of his writings, described agricultural land near the city of Carthage circa 310 BC: It was divided into market gardens and orchards of all sorts of fruit trees, with many streams of water flowing in channels irrigating every part. There were country homes everywhere, lavishly built and covered with stucco. ... Part of the land was planted with vines, part with olives and other productive trees. Beyond these, cattle and sheep were pastured on the plains, and there were meadows with grazing horses. Greek cities contested with Carthage for the Western Mediterranean culminating in the Sicilian Wars and the Pyrrhic War over Sicily, while the Romans fought three wars against Carthage, known as the Punic Wars, "Punic" meaning "Phoenician" in Latin, as Carthage was a Phoenician colony grown into a kingdom. The Carthaginian republic was one of the longest-lived and largest states in the ancient Mediterranean. Reports relay several wars with Syracuse and finally, Rome, which eventually resulted in the defeat and destruction of Carthage in the Third Punic War. The Carthaginians were Phoenician settlers originating in the Mediterranean coast of the Near East. They spoke Canaanite, a Semitic language, and followed a local variety of the ancient Canaanite religion. The fall of Carthage came at the end of the Third Punic War in 146 BC at the Battle of Carthage. Despite initial devastating Roman naval losses and Rome's recovery from the brink of defeat after the terror of a 15-year occupation of much of Italy by Hannibal, the end of the series of wars resulted in the end of Carthaginian power and the complete destruction of the city by Scipio Aemilianus. The Romans pulled the Phoenician warships out into the harbor and burned them before the city, and went from house to house, capturing and enslaving the people. About 50,000 Carthaginians were sold into slavery. The city was set ablaze and razed to the ground, leaving only ruins and rubble. After the fall of Carthage, Rome annexed the majority of the Carthaginian colonies, including other North African locations such as Volubilis, Lixus, Chellah. The legend that the city was sown with salt remains widely accepted despite a lack of evidence among ancient historical accounts; According to R.T. Ridley, the earliest such claim is attributable to B.L. Hallward's chapter in "Cambridge Ancient History", published in 1930. Ridley contended that Hallward's claim may have gained traction due to historical evidence of other salted-earth instances such as Abimelech's salting of Shechem in Judges 9:45. B.H. Warmington admitted he had repeated Hallward's error, but posited that the legend precedes 1930 and inspired repetitions of the practice. He also suggested that it is useful to understand how subsequent historical narratives have been framed and that the symbolic value of the legend is so great and enduring that it mitigates a deficiency of concrete evidence. For many years but especially beginning in the 19th century, various texts claim that after defeating the city of Carthage in the Third Punic War (146 BC), the Roman general Scipio Aemilianus Africanus ordered the city be sacked, forced its surviving inhabitants into slavery, plowed it over and sowed it with salt. However, no ancient sources exist documenting the salting itself. The element of salting is therefore probably a later invention modeled on the Biblical story of Shechem. The ritual of symbolically drawing a plow over the site of a city is mentioned in ancient sources, but not in reference to Carthage specifically. When Pope Boniface VIII destroyed Palestrina in 1299, he issued a papal bull that it be plowed "following the old example of Carthage in Africa" and also salted. "I have run the plough over it, like the ancient Carthage of Africa, and I have had salt sown upon it..." When Carthage fell, its nearby rival Utica, a Roman ally, was made capital of the region and replaced Carthage as the leading center of Punic trade and leadership. It had the advantageous position of being situated on the outlet of the Medjerda River, Tunisia's only river that flowed all year long. However, grain cultivation in the Tunisian mountains caused large amounts of silt to erode into the river. This silt accumulated in the harbor until it became useless, and Rome was forced to rebuild Carthage. By 122 BC, Gaius Gracchus founded a short-lived colony, called "Colonia Iunonia", after the Latin name for the Punic goddess Tanit, "Iuno Caelestis". The purpose was to obtain arable lands for impoverished farmers. The Senate abolished the colony some time later, to undermine Gracchus' power. After this ill-fated attempt, a new city of Carthage was built on the same land by Julius Caesar in the period from 49 to 44 BC, and by the first century, it had grown to be the second-largest city in the western half of the Roman Empire, with a peak population of 500,000. It was the center of the province of Africa, which was a major breadbasket of the Empire. Among its major monuments was an amphitheater. Carthage also became a center of early Christianity (see Carthage (episcopal see)). In the first of a string of rather poorly reported councils at Carthage a few years later, no fewer than 70 bishops attended. Tertullian later broke with the mainstream that was increasingly represented in the West by the primacy of the Bishop of Rome, but a more serious rift among Christians was the Donatist controversy, which Augustine of Hippo spent much time and parchment arguing against. At the Council of Carthage (397), the biblical canon for the western Church was confirmed. The Christians at Carthage conducted persecutions against the pagans, during which the pagan temples, notably the famous Temple of Juno Caelesti, were destroyed. The political fallout from the deep disaffection of African Christians is supposedly a crucial factor in the ease with which Carthage and the other centers were captured in the fifth century by Gaiseric, king of the Vandals, who defeated the Roman general Bonifacius and made the city the capital of the Vandal Kingdom. Gaiseric was considered a heretic, too, an Arian, and though Arians commonly despised Catholic Christians, a mere promise of toleration might have caused the city's population to accept him. The Vandals during their conquest are said to have destroyed parts of Carthage by Victor Vitensis in "Historia Persecutionis Africanae Provincia" including various buildings and churches. After a failed attempt to recapture the city in the fifth century, the Eastern Roman Empire finally subdued the Vandals in the Vandalic War in 533–534. Thereafter, the city became the seat of the praetorian prefecture of Africa, which was made into an exarchate during the emperor Maurice's reign, as was Ravenna on the Italian Peninsula. These two exarchates were the western bulwarks of the Byzantine Empire, all that remained of its power in the West. In the early seventh century Heraclius the Elder, the exarch of Carthage, overthrew the Byzantine emperor Phocas, whereupon his son Heraclius succeeded to the imperial throne. The Roman Exarchate of Africa was not able to withstand the seventh-century Muslim conquest of the Maghreb. The Umayyad Caliphate under Abd al-Malik ibn Marwan in 686 sent a force led by Zuhayr ibn Qays, who won a battle over the Romans and Berbers led by King Kusaila of the Kingdom of Altava on the plain of Kairouan, but he could not follow that up. In 695, Hassan ibn al-Nu'man captured Carthage and advanced into the Atlas Mountains. An imperial fleet arrived and retook Carthage, but in 698, Hasan ibn al-Nu'man returned and defeated Emperor Tiberios III at the 698 Battle of Carthage. Roman imperial forces withdrew from all of Africa except Ceuta. Fearing that the Byzantine Empire might reconquer it, they decided to destroy Roman Carthage in a scorched earth policy and establish their headquarters somewhere else. Its walls were torn down, the water supply from its aqueducts cut off, the agricultural land was ravaged and its harbors made unusable. The destruction of the Exarchate of Africa marked a permanent end to the Byzantine Empire's influence in the region. It is visible from archaeological evidence, that the town of Carthage continued to be occupied. The neighborhood of Bjordi Djedid continued to be occupied. The Baths of Antoninus continued to function in the Arab period and the historian Al-Bakri stated that they were still in good condition. They also had production centers nearby. It is difficult to determine whether the continued habitation of some other buildings belonged to Late Byzantine or Early Arab period. The Bir Ftouha church might have continued to remain in use though it is not clear when it became uninhabited. Constantine the African was born in Carthage. The Medina of Tunis, originally a Berber settlement, was established as the new regional center under the Umayyad Caliphate in the early 8th century. Under the Aghlabids, the people of Tunis revolted numerous times, but the city profited from economic improvements and quickly became the second most important in the kingdom. It was briefly the national capital, from the end of the reign of Ibrahim II in 902, until 909, when the Shi'ite Berbers took over Ifriqiya and founded the Fatimid Caliphate. Carthage remained a residential see until the high medieval period, mentioned in two letters of Pope Leo IX dated 1053, written in reply to consultations regarding a conflict between the bishops of Carthage and Gummi. In each of the two letters, Pope Leo declares that, after the Bishop of Rome, the first archbishop and chief metropolitan of the whole of Africa is the bishop of Carthage. Later, an archbishop of Carthage named Cyriacus was imprisoned by the Arab rulers because of an accusation by some Christians. Pope Gregory VII wrote him a letter of consolation, repeating the hopeful assurances of the primacy of the Church of Carthage, "whether the Church of Carthage should still lie desolate or rise again in glory". By 1076, Cyriacus was set free, but there was only one other bishop in the province. These are the last of whom there is mention in that period of the history of the see. The fortress of Carthage was used by the Muslims until Hafsid era and was captured by the Crusaders during the Eighth Crusade. The inhabitants of Carthage were slaughtered by the Crusaders after they took it, and it was used as a base of operations against the Hafsids. After repelling them, Muhammad I al-Mustansir decided to destroy Cathage's defenses completely to prevent a repeat. Carthage is some east-northeast of Tunis; the settlements nearest to Carthage were the town of Sidi Bou Said to the north and the village of Le Kram to the south. Sidi Bou Said was a village which had grown around the tomb of the eponymous sufi saint (d. 1231), which had been developed into a town under Ottoman rule in the 18th century. Le Kram was developed in the late 19th century under French administration as a settlement close to the port of La Goulette. In 1881, Tunisia became a French protectorate, and in the same year Charles Lavigerie, who was archbishop of Algiers, became apostolic administrator of the vicariate of Tunis. In the following year, Lavigerie became a cardinal. He "saw himself as the reviver of the ancient Christian Church of Africa, the Church of Cyprian of Carthage", and, on 10 November 1884, was successful in his great ambition of having the metropolitan see of Carthage restored, with himself as its first archbishop. In line with the declaration of Pope Leo IX in 1053, Pope Leo XIII acknowledged the revived Archdiocese of Carthage as the primatial see of Africa and Lavigerie as primate. The Acropolium of Carthage (Saint Louis Cathedral of Carthage) was erected on Byrsa hill in 1884. The Danish consul Christian Tuxen Falbe conducted a first survey of the topography of the archaeological site (published in 1833). Antiquarian interest was intensified following the publication of Flaubert's "Salammbô" in 1858. Charles Ernest Beulé performed some preliminary excavations of Roman remains on Byrsa hill in 1860. A more systematic survey of both Punic and Roman-era remains is due to Alfred Louis Delattre, who was sent to Tunis by cardinal Charles Lavigerie in 1875 on both an apostolic and an archaeological mission. Audollent (1901, p. 203) cites Delattre and Lavigerie to the effect that in the 1880s, locals still knew the area of the ancient city under the name of "Cartagenna" (i.e. reflecting the Latin "n"-stem "Carthāgine"). Auguste Audollent divides the area of Roman Carthage into four quarters, "Cartagenna", "Dermèche", "Byrsa" and "La Malga". Cartagenna and Dermèche correspond with the lower city, including the site of Punic Carthage; Byrsa is associated with the upper city, which in Punic times was a walled citadel above the harbour; and "La Malga" is linked with the more remote parts of the upper city in Roman times. French-led excavations at Carthage began in 1921, and from 1923 reported finds of a large quantity of urns containing a mixture of animal and children's bones. René Dussaud identified a 4th-century BC stela found in Carthage as depicting a child sacrifice. A temple at Amman (1400–1250 BC) excavated and reported upon by J.B. Hennessy in 1966, shows the possibility of bestial and human sacrifice by fire. While evidence of child sacrifice in Canaan was the object of academic disagreement, with some scholars arguing that merely children's cemeteries had been unearthed in Carthage, the mixture of children's with animal bones as well as associated epigraphic evidence involving mention of "mlk" led some to believe that, at least in Carthage, child sacrifice was indeed common practice. However, though the animals were surely sacrificed, this does not entirely indicate that the infants were, and in fact the bones indicate the opposite. Rather, the animal sacrifice was likely done to, in some way, honour the deceased. In 2016, an ancient Carthaginian individual, who was excavated from a Punic tomb in Byrsa Hill, was found to belong to the rare U5b2c1 maternal haplogroup. The Young Man of Byrsa specimen dates from the late 6th century BCE, and his lineage is believed to represent early gene flow from Iberia to the Maghreb. In 1920, the first seaplane base was built on the Lake of Tunis for the seaplanes of Compagnie Aéronavale. The Tunis Airfield opened in 1938, serving around 5,800 passengers annually on the Paris-Tunis route. During World War II, the airport was used by the United States Army Air Force Twelfth Air Force as a headquarters and command control base for the Italian Campaign of 1943. Construction on the Tunis-Carthage Airport, which was fully funded by France, began in 1944, and in 1948 the airport become the main hub for Tunisair. In the 1950s the Lycée Français de Carthage was established to serve French families in Carthage. In 1961 it was given to the Tunisian government as part of the Independence of Tunisia, so the nearby Collège Maurice Cailloux in La Marsa, previously an annex of the Lycée Français de Carthage, was renamed to the Lycée Français de La Marsa and began serving the "lycée" level. It is currently the Lycée Gustave Flaubert. After Tunisian independence in 1956, the Tunis conurbation gradually extended around the airport, and Carthage (قرطاج " Qarṭāj") is now a suburb of Tunis, covering the area between Sidi Bou Said and Le Kram. Its population as of January 2013 was estimated at 21,276, mostly attracting the more wealthy residents. If Carthage is not the capital, it tends to be the political pole, a « place of emblematic power » according to Sophie Bessis, leaving to Tunis the economic and administrative roles. The Carthage Palace (the Tunisian presidential palace) is located in the coast. The suburb has six train stations of the TGM line between Le Kram and Sidi Bou Said: Carthage Salammbo (named for Salambo, the fictional daughter of Hamilcar), Carthage Byrsa (named for Byrsa hill), Carthage Dermech ("Dermèche"), Carthage Hannibal (named for Hannibal), Carthage Présidence (named for the Presidential Palace) and Carthage Amilcar (named for Hamilcar). The scant remains of what was once a great city are reflected upon in Letitia Elizabeth Landon's poem, "Carthage", published in 1836 with quotes from Sir Grenville Temple's Journal. The merchants of Carthage were in part heirs of the Mediterranean trade developed by Phoenicia, and so also heirs of the rivalry with Greek merchants. Business activity was accordingly both stimulated and challenged. Cyprus had been an early site of such commercial contests. The Phoenicians then had ventured into the western Mediterranean, founding trading posts, including Utica and Carthage. The Greeks followed, entering the western seas where the commercial rivalry continued. Eventually it would lead, especially in Sicily, to several centuries of intermittent war. Although Greek-made merchandise was generally considered superior in design, Carthage also produced trade goods in abundance. That Carthage came to function as a manufacturing colossus was shown during the Third Punic War with Rome. Carthage, which had previously disarmed, then was made to face the fatal Roman siege. The city "suddenly organised the manufacture of arms" with great skill and effectiveness. According to Strabo (63 BC – AD 21) in his "Geographica": [Carthage] each day produced one hundred and forty finished shields, three hundred swords, five hundred spears, and one thousand missiles for the catapults... . Furthermore, [Carthage although surrounded by the Romans] built one hundred and twenty decked ships in two months... for old timber had been stored away in readiness, and a large number of skilled workmen, maintained at public expense. The textiles industry in Carthage probably started in private homes, but the existence of professional weavers indicates that a sort of factory system later developed. Products included embroidery, carpets, and use of the purple murex dye (for which the Carthaginian isle of Djerba was famous). Metalworkers developed specialized skills, i.e., making various weapons for the armed forces, as well as domestic articles, such as knives, forks, scissors, mirrors, and razors (all articles found in tombs). Artwork in metals included vases and lamps in bronze, also bowls, and plates. Other products came from such crafts as the potters, the glassmakers, and the goldsmiths. Inscriptions on votive stele indicate that many were not slaves but 'free citizens'. Phoenician and Punic merchant ventures were often run as a family enterprise, putting to work its members and its subordinate clients. Such family-run businesses might perform a variety of tasks: own and maintain the ships, providing the captain and crew; do the negotiations overseas, either by barter or buying and selling, of their own manufactured commodities and trade goods, and native products (metals, foodstuffs, etc.) to carry and trade elsewhere; and send their agents to stay at distant outposts in order to make lasting local contacts, and later to establish a warehouse of shipped goods for exchange, and eventually perhaps a settlement. Over generations, such activity might result in the creation of a wide-ranging network of trading operations. Ancillary would be the growth of reciprocity between different family firms, foreign and domestic. State protection was extended to its sea traders by the Phoenician city of Tyre and later likewise by the daughter city-state of Carthage. , the well-regarded French historian of ancient North Africa, summarized the major principles guiding the civic rulers of Carthage with regard to its policies for trade and commerce: Both the Phoenicians and the Cathaginians were well known in antiquity for their secrecy in general, and especially pertaining to commercial contacts and trade routes. Both cultures excelled in commercial dealings. Strabo (63BC-AD21) the Greek geographer wrote that before its fall (in 146 BC) Carthage enjoyed a population of 700,000, and directed an alliance of 300 cities. The Greek historian Polybius (c.203–120) referred to Carthage as "the wealthiest city in the world". A "suffet" (possibly two) was elected by the citizens, and held office with no military power for a one-year term. Carthaginian generals marshalled mercenary armies and were separately elected. From about 550 to 450 the Magonid family monopolized the top military position; later the Barcid family acted similarly. Eventually it came to be that, after a war, the commanding general had to testify justifying his actions before a court of 104 judges. Aristotle (384–322) discusses Carthage in his work, "Politica"; he begins: "The Carthaginians are also considered to have an excellent form of government." He briefly describes the city as a "mixed constitution", a political arrangement with cohabiting elements of monarchy, aristocracy, and democracy, i.e., a king (Gk: basileus), a council of elders (Gk: gerusia), and the people (Gk: demos). Later Polybius of Megalopolis (c.204–122, Greek) in his "Histories" would describe the Roman Republic in more detail as a mixed constitution in which the Consuls were the monarchy, the Senate the aristocracy, and the Assemblies the democracy. Evidently Carthage also had an institution of elders who advised the Suffets, similar to a Greek "gerusia" or the Roman Senate. We do not have a Punic name for this body. At times its members would travel with an army general on campaign. Members also formed permanent committees. The institution had several hundred members drawn from the wealthiest class who held office for life. Vacancies were probably filled by recruitment from among the elite, i.e., by co-option. From among its members were selected the 104 Judges mentioned above. Later the 104 would come to evaluate not only army generals but other office holders as well. Aristotle regarded the 104 as most important; he compared it to the ephorate of Sparta with regard to control over security. In Hannibal's time, such a Judge held office for life. At some stage there also came to be independent self-perpetuating boards of five who filled vacancies and supervised (non-military) government administration. Popular assemblies also existed at Carthage. When deadlocked the Suffets and the quasi-senatorial institution of elders might request the assembly to vote; also, assembly votes were requested in very crucial matters in order to achieve political consensus and popular coherence. The assembly members had no "legal" wealth or birth qualification. How its members were selected is unknown, e.g., whether by festival group or urban ward or another method. The Greeks were favourably impressed by the constitution of Carthage; Aristotle had a separate study of it made which unfortunately is lost. In his "Politica" he states: "The government of Carthage is oligarchical, but they successfully escape the evils of oligarchy by enriching one portion of the people after another by sending them to their colonies." "[T]heir policy is to send some [poorer citizens] to their dependent towns, where they grow rich." Yet Aristotle continues, "[I]f any misfortune occurred, and the bulk of the subjects revolted, there would be no way of restoring peace by legal means." Aristotle remarked also: Many of the Carthaginian institutions are excellent. The superiority of their constitution is proved by the fact that the common people remain loyal to the constitution; the Carthaginians have never had any rebellion worth speaking of, and have never been under the rule of a tyrant. Here one may remember that the city-state of Carthage, who citizens were mainly "Libyphoenicians" (of Phoenician ancestry born in Africa), dominated and exploited an agricultural countryside composed mainly of native Berber sharecroppers and farmworkers, whose affiliations to Carthage were open to divergent possibilities. Beyond these more settled Berbers and the Punic farming towns and rural manors, lived the independent Berber tribes, who were mostly pastoralists. In the brief, uneven review of government at Carthage found in his "Politica" Aristotle mentions several faults. Thus, "that the same person should hold many offices, which is a favorite practice among the Carthaginians." Aristotle disapproves, mentioning the flute-player and the shoemaker. Also, that "magistrates should be chosen not only for their merit but for their wealth." Aristotle's opinion is that focus on pursuit of wealth will lead to oligarchy and its evils. [S]urely it is a bad thing that the greatest offices... should be bought. The law which allows this abuse makes wealth of more account than virtue, and the whole state becomes avaricious. For, whenever the chiefs of the state deem anything honorable, the other citizens are sure to follow their example; and, where virtue has not the first place, their aristocracy cannot be firmly established. In Carthage the people seemed politically satisfied and submissive, according to the historian Warmington. They in their assemblies only rarely exercised the few opportunities given them to assent to state decisions. Popular influence over government appears not to have been an issue at Carthage. Being a commercial republic fielding a mercenary army, the people were not conscripted for military service, an experience which can foster the feel for popular political action. But perhaps this misunderstands the society; perhaps the people, whose values were based on small-group loyalty, felt themselves sufficiently connected to their city's leadership by the very integrity of the person-to-person linkage within their social fabric. Carthage was very stable; there were few openings for tyrants. Only after defeat by Rome devastated Punic imperial ambitions did the people of Carthage seem to question their governance and to show interest in political reform. In 196, following the Second Punic War (218–201), Hannibal Barca, still greatly admired as a Barcid military leader, was elected suffet. When his reforms were blocked by a financial official about to become a judge for life, Hannibal rallied the populace against the 104 judges. He proposed a one-year term for the 104, as part of a major civic overhaul. Additionally, the reform included a restructuring of the city's revenues, and the fostering of trade and agriculture. The changes rather quickly resulted in a noticeable increase in prosperity. Yet his incorrigible political opponents cravenly went to Rome, to charge Hannibal with conspiracy, namely, plotting war against Rome in league with Antiochus the Hellenic ruler of Syria. Although the Roman Scipio Africanus resisted such manoeuvre, eventually intervention by Rome forced Hannibal to leave Carthage. Thus, corrupt city officials efficiently blocked Hannibal Barca in his efforts to reform the government of Carthage. Mago (6th century) was King of Carthage; the head of state, war leader, and religious figurehead. His family was considered to possess a sacred quality. Mago's office was somewhat similar to that of a pharaoh, but although kept in a family it was not hereditary, it was limited by legal consent. Picard, accordingly, believes that the council of elders and the popular assembly are late institutions. Carthage was founded by the king of Tyre who had a royal monopoly on this trading venture. Thus it was the royal authority stemming from this traditional source of power that the King of Carthage possessed. Later, as other Phoenician ship companies entered the trading region, and so associated with the city-state, the King of Carthage had to keep order among a rich variety of powerful merchants in their negotiations among themselves and over risky commerce across the Mediterranean. Under these circumstance, the office of king began to be transformed. Yet it was not until the aristocrats of Carthage became wealthy owners of agricultural lands in Africa that a council of elders was institutionalized at Carthage. Most ancient literature concerning Carthage comes from Greek and Roman sources as Carthage's own documents were destroyed by the Romans. Apart from inscriptions, hardly any Punic literature has survived, and none in its own language and script. A brief catalogue would include: "[F]rom the Greek author Plutarch [(c. 46 – c. 120)] we learn of the 'sacred books' in Punic safeguarded by the city's temples. Few Punic texts survive, however." Once "the City Archives, the Annals, and the scribal lists of "suffets"" existed, but evidently these were destroyed in the horrific fires during the Roman capture of the city in 146 BC. Yet some Punic books (Latin: "libri punici") from the libraries of Carthage reportedly did survive the fires. These works were apparently given by Roman authorities to the newly augmented Berber rulers. Over a century after the fall of Carthage, the Roman politician-turned-author Gaius Sallustius Crispus or Sallust (86–34) reported his having seen volumes written in Punic, which books were said to be once possessed by the Berber king, Hiempsal II (r. 88–81). By way of Berber informants and Punic translators, Sallust had used these surviving books to write his brief sketch of Berber affairs. Probably some of Hiempsal II's "libri punici", that had escaped the fires that consumed Carthage in 146 BC, wound up later in the large royal library of his grandson Juba II (r.25 BC-AD 24). Juba II not only was a Berber king, and husband of Cleopatra's daughter, but also a scholar and author in Greek of no less than nine works. He wrote for the Mediterranean-wide audience then enjoying classical literature. The "libri punici" inherited from his grandfather surely became useful to him when composing his "Libyka", a work on North Africa written in Greek. Unfortunately, only fragments of "Libyka" survive, mostly from quotations made by other ancient authors. It may have been Juba II who 'discovered' the five-centuries-old 'log book' of Hanno the Navigator, called the "Periplus", among library documents saved from fallen Carthage. In the end, however, most Punic writings that survived the destruction of Carthage "did not escape the immense wreckage in which so many of Antiquity's literary works perished." Accordingly, the long and continuous interactions between Punic citizens of Carthage and the Berber communities that surrounded the city have no local historian. Their political arrangements and periodic crises, their economic and work life, the cultural ties and social relations established and nourished (infrequently as kin), are not known to us directly from ancient Punic authors in written accounts. Neither side has left us their stories about life in Punic-era Carthage. Regarding "Phoenician" writings, few remain and these seldom refer to Carthage. The more ancient and most informative are cuneiform tablets, ca. 1600–1185, from ancient Ugarit, located to the north of Phoenicia on the Syrian coast; it was a Canaanite city politically affiliated with the Hittites. The clay tablets tell of myths, epics, rituals, medical and administrative matters, and also correspondence. The highly valued works of Sanchuniathon, an ancient priest of Beirut, who reportedly wrote on Phoenician religion and the origins of civilization, are themselves completely lost, but some little content endures twice removed. Sanchuniathon was said to have lived in the 11th century, which is considered doubtful. Much later a "Phoenician History" by Philo of Byblos (64–141) reportedly existed, written in Greek, but only fragments of this work survive. An explanation proffered for why so few Phoenician works endured: early on (11th century) archives and records began to be kept on papyrus, which does not long survive in a moist coastal climate. Also, both Phoenicians and Carthaginians were well known for their secrecy. Thus, of their ancient writings we have little of major interest left to us by Carthage, or by Phoenicia the country of origin of the city founders. "Of the various Phoenician and Punic compositions alluded to by the ancient classical authors, not a single work or even fragment has survived in its original idiom." "Indeed, not a single Phoenician manuscript has survived in the original [language] or in translation." We cannot therefore access directly the line of thought or the contour of their worldview as expressed in their own words, in their own voice. Ironically, it was the Phoenicians who "invented or at least perfected and transmitted a form of writing [the alphabet] that has influenced dozens of cultures including our own." As noted, the celebrated ancient books on agriculture written by Mago of Carthage survives only via quotations in Latin from several later Roman works.
https://en.wikipedia.org/wiki?curid=6555
Coprime integers In number theory, two integers and are said to be relatively prime, mutually prime, or coprime (also written co-prime) if the only positive integer (factor) that divides both of them is 1. Consequently, any prime number that divides one does not divide the other. This is equivalent to their greatest common divisor (gcd) being 1. The numerator and denominator of a reduced fraction are coprime. As specific examples, 14 and 15 are coprime, being commonly divisible only by 1, while 14 and 21 are not coprime, because they are both divisible by 7. Standard notations for relatively prime integers and are: and . Graham, Knuth and Patashnik have proposed that the notation formula_1 be used to indicate that and are relatively prime and that the term "prime" be used instead of coprime (as in is "prime" to ). A fast way to determine whether two numbers are coprime is given by the Euclidean algorithm and its faster variants such as binary GCD algorithm or Lehmer's GCD algorithm. The number of integers coprime to a positive integer , between 1 and , is given by Euler's totient function (or Euler's phi function) . A set of integers can also be called coprime if its elements share no common positive factor except 1. A stronger condition on a set of integers is pairwise coprime, which means that and are coprime for every pair of different integers in the set. The set } is coprime, but it is not pairwise coprime since 2 and 4 are not relatively prime. The numbers 1 and −1 are the only integers coprime to every integer, and they are the only integers that are coprime with 0. A number of conditions are equivalent to and being coprime: As a consequence of the third point, if "a" and "b" are coprime and "br" ≡ "bs" (mod "a"), then "r" ≡ "s" (mod "a"). That is, we may "divide by "b"" when working modulo "a". Furthermore, if "b"1 and "b"2 are both coprime with "a", then so is their product "b"1"b"2 (i.e., modulo "a" it is a product of invertible elements, and therefore invertible); this also follows from the first point by Euclid's lemma, which states that if a prime number "p" divides a product "bc", then "p" divides at least one of the factors "b", "c". As a consequence of the first point, if "a" and "b" are coprime, then so are any powers "a""k" and "b""m". If "a" and "b" are coprime and "a" divides the product "bc", then "a" divides "c". This can be viewed as a generalization of Euclid's lemma. The two integers "a" and "b" are coprime if and only if the point with coordinates ("a", "b") in a Cartesian coordinate system is "visible" from the origin (0,0), in the sense that there is no point with integer coordinates on the line segment between the origin and ("a", "b"). (See figure 1.) In a sense that can be made precise, the probability that two randomly chosen integers are coprime is 6/π2 (see pi), which is about 61%. See below. Two natural numbers "a" and "b" are coprime if and only if the numbers 2"a" − 1 and 2"b" − 1 are coprime. As a generalization of this, following easily from the Euclidean algorithm in base "n" > 1: A set of integers "S" = {"a"1, "a"2, ... "a""n"} can also be called "coprime" or "setwise coprime" if the greatest common divisor of all the elements of the set is 1. For example, the integers 6, 10, 15 are coprime because 1 is the only positive integer that divides all of them. If every pair in a set of integers is coprime, then the set is said to be "pairwise coprime" (or "pairwise relatively prime", "mutually coprime" or "mutually relatively prime"). Pairwise coprimality is a stronger condition than setwise coprimality; every pairwise coprime finite set is also setwise coprime, but the reverse is not true. For example, the integers 4, 5, 6 are (setwise) coprime (because the only positive integer dividing "all" of them is 1), but they are not "pairwise" coprime (because gcd(4, 6) = 2). The concept of pairwise coprimality is important as a hypothesis in many results in number theory, such as the Chinese remainder theorem. It is possible for an infinite set of integers to be pairwise coprime. Notable examples include the set of all prime numbers, the set of elements in Sylvester's sequence, and the set of all Fermat numbers. Two ideals "A" and "B" in the commutative ring "R" are called coprime (or comaximal) if "A" + "B" = "R". This generalizes Bézout's identity: with this definition, two principal ideals ("a") and ("b") in the ring of integers Z are coprime if and only if "a" and "b" are coprime. If the ideals "A" and "B" of "R" are coprime, then "AB" = "A"∩"B"; furthermore, if "C" is a third ideal such that "A" contains "BC", then "A" contains "C". The Chinese remainder theorem can be generalized to any commutative ring, using coprime ideals. Given two randomly chosen integers "a" and "b", it is reasonable to ask how likely it is that "a" and "b" are coprime. In this determination, it is convenient to use the characterization that "a" and "b" are coprime if and only if no prime number divides both of them (see Fundamental theorem of arithmetic). Informally, the probability that any number is divisible by a prime (or in fact any integer) formula_3 is formula_4; for example, every 7th integer is divisible by 7. Hence the probability that two numbers are both divisible by "p" is formula_5, and the probability that at least one of them is not is formula_6. Any finite collection of divisibility events associated to distinct primes is mutually independent. For example, in the case of two events, a number is divisible by primes "p" and "q" if and only if it is divisible by "pq"; the latter event has probability 1/"pq". If one makes the heuristic assumption that such reasoning can be extended to infinitely many divisibility events, one is led to guess that the probability that two numbers are coprime is given by a product over all primes, Here "ζ" refers to the Riemann zeta function, the identity relating the product over primes to "ζ"(2) is an example of an Euler product, and the evaluation of "ζ"(2) as "π"2/6 is the Basel problem, solved by Leonhard Euler in 1735. There is no way to choose a positive integer at random so that each positive integer occurs with equal probability, but statements about "randomly chosen integers" such as the ones above can be formalized by using the notion of "natural density". For each positive integer "N", let "P""N" be the probability that two randomly chosen numbers in formula_8 are coprime. Although "P""N" will never equal formula_9 exactly, with work one can show that in the limit as formula_10, the probability formula_11 approaches formula_9. More generally, the probability of "k" randomly chosen integers being coprime is formula_13. All pairs of positive coprime numbers formula_14 (with formula_15) can be arranged in two disjoint complete ternary trees, one tree starting from formula_16 (for even-odd and odd-even pairs), and the other tree starting from formula_17 (for odd-odd pairs). The children of each vertex formula_18 are generated as follows: This scheme is exhaustive and non-redundant with no invalid members.
https://en.wikipedia.org/wiki?curid=6556
Control unit The control unit (CU) is a component of a computer's central processing unit (CPU) that directs the operation of the processor. It tells the computer's memory, arithmetic and logic unit and input and output devices how to respond to the instructions that have been sent to the processor. It directs the operation of the other units by providing timing and control signals. Most computer resources are managed by the CU. It directs the flow of data between the CPU and the other devices. John von Neumann included the control unit as part of the von Neumann architecture. In modern computer designs, the control unit is typically an internal part of the CPU with its overall role and operation unchanged since its introduction. The simplest computers use a multicycle microarchitecture. These were the earliest designs. They are still popular in the very smallest computers, such as the embedded systems that operate machinery. In a multicycle computer, the control unit often steps through the Von Neumann Cycle: Fetch the instruction, Fetch the operands, do the instruction, write the results. When the next instruction is placed in the control unit, it changes the behavior of the control unit to finish the instruction correctly. So, the bits of the instruction directly control the control unit, which in turn controls the computer. The control unit may include a binary counter to tell the control unit's logic what step it should do. Multicycle control units typically use both the rising and falling edges of their square-wave timing clock. They operate a step of their operation on each edge of the timing clock, so that a four-step operation completes in two clock cycles. Many computers have two different types of unexpected events. An interrupt occurs because some type of input or output needs software attention in order to operate correctly. An exception is caused by the computer's operation. One crucial difference is that the timing of an interrupt cannot be predicted. Another is that some exceptions (e.g. a memory-not-available exception) can be caused by an instruction that needs to be restarted. Control units can be designed to handle interrupts in one of two typical ways. If a quick response is most important, a control unit is designed to abandon work to handle the interrupt. In this case, the work in process will be restarted after the last completed instruction. If the computer is to be very inexpensive, very simple, very reliable, or to get more work done, the control unit will finish the work in process before handling the interrupt. Finishing the work is inexpensive, because it needs no register to record the last finished instruction. It is simple and reliable because it has the least number of states. It also wastes the least amount of work. Exceptions can be made to operate like interrupts in very simple computers. If virtual memory is required, then a memory-not-available exception must retry the failing instruction. It is common for multicycle computers to use more cycles. Sometimes it takes longer to take a conditional jump, because the program counter has to be reloaded. Sometimes they do multiplication or division instructions by a process something like binary long multiplication and division. Very small computers might do arithmetic one or a few bits at a time. Some computers have very complex instructions that take many steps. Many medium-complexity computers pipeline instructions. This design is popular because of its economy and speed. In a pipelined computer, instructions flow through the computer. This design has several stages. For example, it might have one stage for each step of the Von Neumann cycle. A pipelined computer usually has "pipeline registers" after each stage. These store the bits calculated by a stage so that the logic gates of the next stage can use the bits to do the next step. It is common for even numbered stages to operate on one edge of the square-wave clock, while odd-numbered stages operate on the other edge. In a pipelined computer, the control unit arranges for the flow to start, continue, and stop as a program commands. The instruction data is usually passed in pipeline registers from one stage to the next, with a somewhat separated piece of control logic for each stage. The control unit also assures that the instruction in each stage does not harm the operation of instructions in other stages. For example, if two stages must use the same piece of data, the control logic assures that the uses are done in the correct sequence. When operating efficiently, a pipelined computer will have an instruction in each stage. It is then working on all of those instructions at the same time. It can finish about one instruction for each cycle of its clock. When a program makes a decision, and switches to a different sequence of instructions, the pipeline sometimes must discard the data in process and restart. This is called a "stall." When two instructions could interfere, sometimes the control unit must stop processing a later instruction until an earlier instruction completes. This is called a "pipeline bubble" because a part of the pipeline is not processing instructions. Pipeline bubbles can occur when two instructions operate on the same register. Interrupts and unexpected exceptions also stall the pipeline. If a pipelined computer abandons work for an interrupt, more work is lost than in a multicycle computer. Predictable exceptions do not need to stall. For example, if an exception instruction is used to enter the operating system, it does not cause a stall. Speed? For the same speed of electronic logic, it can do more instructions per second than a multicycle computer. Also, even though the electronic logic has a fixed maximum speed, a pipelined computer can be made faster or slower by varying the number of stages in the pipeline. With more stages, each stage does less work, and so the stage has fewer delays from the logic gates. Economy? A pipelined model of a computer often has the least logic gates per instruction per second, less than either a multicycle or out-of-order computer. Why? The average stage is less complex than a multicycle computer. An out of order computer usually has large amounts of idle logic at any given instant. Similar calculations usually show that a pipelined computer uses less energy per instruction. However, a pipelined computer is usually more complex and more costly than a comparable multicycle computer. It typically has more logic gates, registers and a more complex control unit. In a like way, it might use more total energy, while using less energy per instruction. Out of order CPUs can usually do more instructions per second because they can do several instructions at once. Control units use many methods to keep a pipeline full and avoid stalls. For example, even simple control units can assume that a backwards branch, to a lower-numbered, earlier instruction, is a loop, and will be repeated. So, a control unit with this design will always fill the pipeline with the backwards branch path. If a compiler can detect the most frequently-taken direction of a branch, the compiler can just produce instructions so that the most frequently taken branch is the preferred direction of branch. In a like way, a control unit might get hints from the compiler: Some computers have instructions that can encode hints from the compiler about the direction of branch. Some control units do branch prediction: A control unit keeps an electronic list of the recent branches, encoded by the address of the branch instruction. This list has a few bits for each branch to remember the direction that was taken most recently. Some control units can do speculative execution, in which a computer might have two or more pipelines, calculate both directions of a branch,then discard the calculations of the unused direction. Results from memory can become available at unpredictable times because very fast computers cache memory. That is, they copy limited amounts of memory data into very fast memory. The CPU must be designed to process at the very fast speed of the cache memory. Therefore, the CPU might stall when it must access main memory directly. In modern PCs, main memory is as much as three hundred times slower than cache. To help this, out-of-order CPUs and control units were developed to process data as it becomes available. (See next section) But what if all the calculations are complete, but the CPU is still stalled, waiting for main memory? Then, a control unit can switch to an alternative thread of execution whose data has been fetched while the thread was idle. A thread has its own program counter, a stream of instructions and a separate set of registers. Designers vary the number of threads depending on current memory technologies and the type of computer. Typical computers such as PCs and smart phones usually have control units with a few threads, just enough to keep busy with affordable memory systems. Database computers often have about twice as many threads, to keep their much larger memories busy. Graphic processing units (GPUs) usually have hundreds or thousands of threads, because they have hundreds or thousands of execution units doing repetitive graphic calculations. When a control unit permits threads, the software also has to be designed to handle them. In general-purpose CPUs like PCs and smartphones, the threads are usually made to look very like normal time-sliced processes. At most, the operating system might need some awareness of them. In GPUs, the thread scheduling usually cannot be hidden from the application software, and is often controlled with a specialized subroutine library. A control unit can be designed to finish what it can. If several instructions can be completed at the same time, the control unit will arrange it. So, the fastest computers can process instructions in a sequence that can vary somewhat, depending on when the operands or instruction destinations become available. Most supercomputers and many PC CPUs use this method. The exact organization of this type of control unit depends on the slowest part of the computer. When the execution of calculations is the slowest, instructions flow from memory into pieces of electronics called "issue units." An issue unit holds an instruction until both its operands and an execution unit are available. Then, the instruction and its operands are "issued" to an execution unit. The execution unit does the instruction. Then the resulting data is moved into a queue of data to be written back to memory or registers. If the computer has multiple execution units, it can usually do several instructions per clock cycle. It is common to have specialized execution units. For example, a modestly priced computer might have only one floating-point execution unit, because floating point units are expensive. The same computer might have several integer units, because these are relatively inexpensive, and can do the bulk of instructions. One kind of control unit for issuing uses an array of electronic logic, a "scoreboard"" that detects when an instruction can be issued. The "height" of the array is the number of execution units, and the "length" and "width" are each the number of sources of operands. When all the items come together, the signals from the operands and execution unit will cross. The logic at this intersection detects that the instruction can work, so the instruction is "issued" to the free execution unit. An alternative style of issuing control unit implements the Tomasulo algorithm, which reorders a hardware queue of instructions. In some sense, both styles utilize a queue. The scoreboard is an alternative way to encode and reorder a queue of instructions, and some designers call it a queue table. With some additional logic, a scoreboard can compactly combine execution reordering, register renaming and precise exceptions and interrupts. Further it can do this without the power-hungry, complex content-addressable memory used by the Tomasulo algorithm. If the execution is slower than writing the results, the memory write-back queue always has free entries. But what if the memory writes slowly? Or what if the destination register will be used by an "earlier" instruction that has not yet issued? Then the write-back step of the instruction might need to be scheduled. This is sometimes called "retiring" an instruction. In this case, there must be scheduling logic on the back end of execution units. It schedules access to the registers or memory that will get the results. Retiring logic can also be designed into an issuing scoreboard or a Tomasulo queue, by including memory or register access in the issuing logic. Out of order controllers require special design features to handle interrupts. When there are several instructions in progress, it is not clear where in the instruction stream an interrupt occurs. For input and output interrupts, almost any solution works. However when a computer has virtual memory, an interrupt occurs to indicate that a memory access failed. This memory access must be associated with an exact instruction and an exact processor state, so that the processor's state can be saved and restored by the interrupt. A usual solution preserves copies of registers until a memory access completes. Also, out of order CPUs have even more problems with stalls from branching, because they can complete several instructions per clock cycle, and usually have many instructions in various stages of progress. So, these control units might use all of the solutions used by pipelined processors. Some computers translate each single instruction into a sequence of simpler instructions. The advantage is that an out of order computer can be simpler in the bulk of its logic, while handling complex multi-step instructions. x86 Intel CPUs since the Pentium Pro translate complex CISC x86 instructions to more RISC-like internal micro-operations. In these, the "front" of the control unit manages the translation of instructions. Operands are not translated. The "back" of the CU is an out-of-order CPU that issues the micro-operations and operands to the execution units and data paths. Many modern computers have controls that minimize power usage. In battery-powered computers, such as those in cell-phones, the advantage is longer battery life. In computers with utility power, the justification is to reduce the cost of power, cooling or noise. Most modern computers use CMOS logic. CMOS wastes power in two common ways: By changing state, i.e. "active power," and by unintended leakage. The active power of a computer can be reduced by turning off control signals. Leakage current can be reduced by reducing the electrical pressure, the voltage, making the transistors with larger depletion regions or turning off the logic completely. Active power is easier to reduce because data stored in the logic is not affected. The usual method reduces the CPU's clock rate. Most computer systems use this method. It's common for a CPU to idle during the transition to avoid side-effects from the changing clock. Most computers also have a "halt" instruction. This was invented to stop non-interrupt code so that interrupt code has reliable timing. However, designers soon noticed that a halt instruction was also a good time to turn off a CPU's clock completely, reducing the CPU's active power to zero. The interrupt controller might continue to need a clock, but that usually uses much less power than the CPU. These methods are relatively easy to design, and became so common that others were invented for commercial advantage. Many modern low-power CMOS CPUs stop and start specialized execution units and bus interfaces depending on the needed instruction. Some computers even arrange the CPU's microarchitecture to use transfer-triggered multiplexers so that each instruction only utilises the exact pieces of logic needed. Theoretically, computers at lower clock speeds could also reduce leakage by reducing the voltage of the power supply. This affects the reliability of the computer in many ways, so the engineering is expensive, and it is uncommon except in relatively expensive computers such as PCs or cellphones. Some designs can use very low leakage transistors, but these usually add cost. The depletion barriers of the transistors can be made larger to have less leakage, but this makes the transistor larger and thus both slower and more expensive. Some vendors use this technique in selected portions of an IC by constructing low leakage logic from large transistors that some processes provide for analog circuits. Some processes place the transistors above the surface of the silicon, in "fin fets", but these processes have more steps, so are more expensive. Special transistor doping materials (e.g. hafnium) can also reduce leakage, but this adds steps to the processing, making it more expensive. Some semiconductors have a larger band-gap than silicon. However, these materials and processes are currently (2020) more expensive than silicon. Managing leakage is more difficult, because before the logic can be turned-off, the data in it must be moved to some type of low-leakage storage. One common method is to spread the load to many CPUs, and turn off unused CPUs as the load reduces. The operating system's task switching logic saves the CPUs' data to memory. In some cases, one of the CPUs can be simpler and smaller, literally with fewer logic gates. So, it has low leakage, and it is the last to be turned off, and the first to be turned on. Also it then is the only CPU that requires special low-power features. A similar method is used in most PCs, which usually have an auxiliary embedded CPU that manages the power system. However in PCs, the software is usually in the BIOS, not the operating system. Some CPUs make use of a special type of flip-flop (to store a bit) that couples a fast, high-leakage storage cell to a slow, large (expensive) low-leakage cell. These two cells have separated power supplies. When the CPU enters a power saving mode (e.g. because of a halt that waits for an interrupt), data is transferred to the low-leakage cells, and the others are turned off. When the CPU leaves a low-leakage mode (e.g. because of an interrupt), the process is reversed. Older designs would copy the CPU state to memory, or even disk, sometimes with specialized software. Very simple embedded systems sometimes just restart. All modern CPUs have control logic to attach the CPU to the rest of the computer. In modern computers, this is usually a bus controller. When an instruction reads or writes memory, the control unit either controls the bus directly, or controls a bus controller. Many modern computers use the same bus interface for memory, input and output. This is called "memory-mapped I/O". To a programmer, the registers of the I/O devices appear as numbers at specific memory addresses. x86 PCs use an older method, a separate I/O bus accessed by I/O instructions. A modern CPU also tends to include an interrupt controller. It handles interrupt signals from the system bus. The control unit is the part of the computer that responds to the interrupts. There is often a cache controller to cache memory. The cache controller and the associated cache memory is often the largest physical part of a modern, higher-performance CPU. When the memory, bus or cache is shared with other CPUs, the control logic must communicate with them to assure that no computer ever gets out-of-date old data. Many historic computers built some type of input and output directly into the control unit. For example, many historic computers had a front panel with switches and lights directly controlled by the control unit. These let a programmer directly enter a program and debug it. In later production computers, the most common use of a front panel was to a enter a small bootstrap program to read the operating system from disk. This was annoying. So, front panels were replaced by bootstrap programs in read-only memory. Most PDP-8 models had a data bus designed to let I/O devices borrow the control unit's memory read and write logic. This reduced the complexity and expense of high speed I/O controllers, e.g. for disk. The Xerox Alto had a multitasking microprogammable control unit that performed almost all I/O. This design provided most of the features of a modern PC with only a tiny fraction of the electronic logic. The dual-thread computer was run by the two lowest-priority microthreads. These performed calculations whenever I/O was not required. High priority microthreads provided (in decreasing priority) video, network, disk, a periodic timer, mouse, and keyboard. The microprogram did the complex logic of the I/O device, as well as the logic to integrate the device with the computer. For the actual hardware I/O, the microprogram read and wrote shift registers for most I/O, sometimes with resistor networks and transistors to shift output voltage levels (e.g. for video). To handle outside events, the microcontroller had microinterrupts to switch threads at the end of a thread's cycle, e.g. at the end of an instruction, or after a shift-register was accessed. The microprogram could be rewritten and reinstalled, which was very useful for a research computer. Thus a program of instructions in memory will cause the CU to configure a CPU's data flows to manipulate the data correctly between instructions. This results in a computer that could run a complete program and require no human intervention to make hardware changes between instructions (as had to be done when using only punch cards for computations before stored programmed computers with CUs were invented). Hardwired control units are implemented through use of combinational logic units, featuring a finite number of gates that can generate specific results based on the instructions that were used to invoke those responses. Hardwired control units are generally faster than the microprogrammed designs. This design uses a fixed architecture—it requires changes in the wiring if the instruction set is modified or changed. It can be convenient for simple, fast computers. A controller that uses this approach can operate at high speed; however, it has little flexibility. A complex instruction set can overwhelm a designer who uses ad-hoc logic design. The hardwired approach has become less popular as computers have evolved. Previously, control units for CPUs used ad-hoc logic, and they were difficult to design. The idea of microprogramming was introduced by Maurice Wilkes in 1951 as an intermediate level to execute computer program instructions. Microprograms were organized as a sequence of "microinstructions" and stored in special control memory. The algorithm for the microprogram control unit, unlike the hardwired control unit, is usually specified by flowchart description. The main advantage of the microprogram control unit is the simplicity of its structure. Outputs of the controller are organized in microinstructions and they can be easily replaced. A popular variation on microcode is to debug the microcode using a software simulator. Then, the microcode is a table of bits. This is a logical truth table, that translates a microcode address into the control unit outputs. This truth table can be fed to a computer program that produces optimized electronic logic. The resulting control unit is almost as easy to design as microprogramming, but it has the fast speed and low number of logic elements of a hard wired control unit. The practical result resembles a Mealy machine or Richards controller.
https://en.wikipedia.org/wiki?curid=6557
Cello The cello ( ; plural celli or cellos) or violoncello ( ; ) is a bowed (and occasionally plucked) string instrument of the violin family. Its four strings are usually tuned in perfect fifths: from low to high, C2, G2, D3 and A3, an octave lower than the viola. Music for the cello is generally written in the bass clef, with tenor clef and treble clef used for higher-range passages. Played by a "cellist" or "violoncellist", it enjoys a large solo repertoire with and without accompaniment, as well as numerous concerti. The cello often plays the bass part, both in chamber music such as string quartets and the orchestra's string section, where the cello may be reinforced an octave lower by the double basses. Figured bass music of the Baroque-era typically assumes a cello, viola da gamba or bassoon as part of the basso continuo group alongside chordal instruments such as organ, harpsichord, lute or theorbo. Cellos are found in many other ensembles, from modern Chinese orchestras to cello rock bands. The name "cello" is derived from the ending of the Italian "violoncello", which means "little violone". Violone ("big viola") was a large-sized member of viol (viola da gamba) family or the violin (viola da braccio) family. The term "violone" today usually refers to the lowest-pitched instrument of the viols, a family of stringed instruments that went out of fashion around the end of the 17th century in most countries except England and, especially, France, where they survived another half-century before the louder violin family came into greater favour in that country as well. In modern symphony orchestras, it is the second largest stringed instrument (the double bass is the largest). Thus, the name "violoncello" contained both the augmentative ""-one"" ("big") and the diminutive ""-cello"" ("little"). By the turn of the 20th century, it had become common to shorten the name to 'cello, with the apostrophe indicating the missing stem. It is now customary to use "cello" without apostrophe as the full designation. "Viol" is derived from the root "viola", which was derived from Medieval Latin "vitula", meaning stringed instrument. Cellos are tuned in fifths, starting with C2 (two octaves below middle C), followed by G2, D3, and then A3. It is tuned in the same intervals as the viola, but an octave lower. Unlike the violin or viola but similar to the double bass, the cello has an endpin that rests on the floor to support the instrument's weight. The cello is most closely associated with European classical music. The instrument is a part of the standard orchestra, as part of the string section, and is the bass voice of the string quartet (although many composers give it a melodic role as well), as well as being part of many other chamber groups. Among the most well-known Baroque works for the cello are Johann Sebastian Bach's six unaccompanied Suites. Other significant include Sonatas and Concertos by Vivaldi, and earlier works by Gabrieli, Geminiani, and Bononcini. As a basso continuo instrument basso continuo the cello may have been used in works by Francesca Caccini (1587–1641), Barbara Strozzi (1619–1677) with pieces such as "Il primo libro di madrigali, per 2–5 voci e basso continuo, op. 1" and Elisabeth Jacquet de La Guerre (1665–1729) who wrote six sonatas for violin and basso continuo. From the Classical era, the two concertos by Joseph Haydn in C major and D major stand out, as do the five sonatas for cello and pianoforte of Ludwig van Beethoven, which span the important three periods of his compositional evolution. Other outstanding examples include the three Concerti by C.P.E. Bach, Capricci by dall'Abaco, and Sonatas by Flackton, Boismortier, and Boccherini. A "Divertimento for Piano, Clarinet, Viola and Cello" is among the surviving works by Duchess Anna Amalia of Brunswick-Wolfenbüttel (1739–1807). Well-known works of the era include the Robert Schumann Concerto, the Antonín Dvořák Concerto as well as the two sonatas and the Double Concerto by Johannes Brahms. A review of compositions for cello in the Romantic era must include the German composer Fanny Mendelssohn (1805–1847) who wrote the Fantasy in G minor for cello and piano and a Capriccio in A-flat for cello. Compositions from the late-19th and early 20th century include three cello sonatas (including the Cello Sonata in C Minor written in 1880) by Dame Ethel Smyth (1858–1944), Edward Elgar's Cello Concerto in E minor, Claude Debussy's Sonata for Cello and Piano, and unaccompanied cello sonatas by Zoltán Kodály and Paul Hindemith. Pieces including cello were written by American Music Cente founder Marion Bauer (1882–1955) (two trio sonatas for flute, cello and piano) and Ruth Crawford Seeger (1901–1953) (Diaphonic suite No. 2 for bassoon and cello). The cello's versatility made it popular with many composers in this era, such as Sergei Prokofiev, Dmitri Shostakovich, Benjamin Britten, György Ligeti, Witold Lutoslawski and Henri Dutilleux. Polish composer Grażyna Bacewicz (1909–1969) was writing for cello in the mid 20th century with Concerto No. 1 for Cello and Orchestra (1951), Concerto No. 2 for Cello and Orchestra (1963) and in 1964 composed her Quartet for four cellos. Well-known cellists from the 20th century include Jacqueline du Pre, Pablo Casals, Yo-Yo Ma, Emanuel Feuermann, Guilhermina Suggia, Mstislav Rostropovich and Beatrice Harrison. Others include Raya Garbousova, Anner Bylsma, Zara Nelsova, Alfred Wallenstein, Han-Na Chang, Mischa Maisky, Hildur Gudnadottir, and Gregor Piatigorsky. See the comprehensive list of cellists here. In the 2010s, the instrument is found in popular music, but was more commonly used in 1970s pop and disco music. Today it is sometimes featured in pop and rock recordings, examples of which are noted later in this article. The cello has also appeared in major hip-hop and R & B performances, such as singers Rihanna and Ne-Yo's 2007 performance at the American Music Awards. The instrument has also been modified for Indian classical music by Nancy Lesh and Saskia Rao-de Haas.[5] The violin family, including cello-sized instruments, emerged c. 1500 as family of instruments distinct from the viola da gamba family. The earliest depictions of the violin family, from northern Italy c. 1530, show three sizes of instruments, roughly corresponding to what we now call violins, violas, and cellos. Contrary to a popular misconception, the cello did not evolve from the viola da gamba, but existed alongside it for about two and a half centuries. The violin family is also known as the viola da braccio (meaning viola of the arm) family, a reference to the primary way the members of the family are held. This is to distinguish it from the viola da gamba (meaning viola of the leg) family, in which all the members are all held with the legs. The likely predecessors of the violin family include the lira da braccio and the rebec. The earliest surviving cellos are made by Andrea Amati, the first known member of the celebrated Amati family of luthiers. The direct ancestor to the violoncello was the bass violin. Monteverdi referred to the instrument as "basso de viola da braccio" in "Orfeo" (1607). Although the first bass violin, possibly invented as early as 1538, was most likely inspired by the viol, it was created to be used in consort with the violin. The bass violin was actually often referred to as a ""violone"", or "large viola", as were the viols of the same period. Instruments that share features with both the bass violin and the "viola da gamba" appear in Italian art of the early 16th century. The invention of wire-wound strings (fine wire around a thin gut core), around 1660 in Bologna, allowed for a finer bass sound than was possible with purely gut strings on such a short body. Bolognese makers exploited this new technology to create the cello, a somewhat smaller instrument suitable for solo repertoire due to both the timbre of the instrument and the fact that the smaller size made it easier to play virtuosic passages. This instrument had disadvantages as well, however. The cello's light sound was not as suitable for church and ensemble playing, so it had to be doubled by organ, theorbo or violone. Around 1700, Italian players popularized the cello in northern Europe, although the bass violin (basse de violon) continued to be used for another two decades in France. Many existing bass violins were literally cut down in size to convert them into cellos according to the smaller pattern developed by Stradivarius, who also made a number of old pattern large cellos (the 'Servais'). The sizes, names, and tunings of the cello varied widely by geography and time. The size was not standardized until around 1750. Despite similarities to the viola da gamba, the cello is actually part of the viola da braccio family, meaning "viol of the arm", which includes, among others, the violin and viola. Though paintings like Bruegel's "The Rustic Wedding", and Jambe de Fer in his "Epitome Musical" suggest that the bass violin had alternate playing positions, these were short-lived and the more practical and ergonomic "a gamba" position eventually replaced them entirely. Baroque-era cellos differed from the modern instrument in several ways. The neck has a different form and angle, which matches the baroque bass-bar and stringing. Modern cellos have an endpin at the bottom to support the instrument (and transmit some of the sound through the floor), while Baroque cellos are held only by the calves of the player. Modern bows curve in and are held at the frog; Baroque bows curve out and are held closer to the bow's point of balance. Modern strings normally have a metal core, although some use a synthetic core; Baroque strings are made of gut, with the G and C strings wire-wound. Modern cellos often have fine-tuners connecting the strings to the tailpiece, which make it much easier to tune the instrument, but such pins are rendered ineffective by the flexibility of the gut strings used on Baroque cellos. Overall, the modern instrument has much higher string tension than the Baroque cello, resulting in a louder, more projecting tone, with fewer overtones. Few educational works specifically devoted to the cello existed before the 18th century, and those that do exist contain little value to the performer beyond simple accounts of instrumental technique. One of the earliest cello manuals is Michel Corrette's "Méthode, thèorique et pratique pour apprendre en peu de temps le violoncelle dans sa perfection" (Paris, 1741). Cellos are part of the standard symphony orchestra, which usually includes eight to twelve cello players. The cello section, in standard orchestral seating, is located on stage left (the audience's right) in the front, opposite the first violin section. However, some orchestras and conductors prefer switching the positioning of the viola and cello sections. The "principal" cellist is the section leader, determining bowings for the section in conjunction with other string principals, playing solos and leading entrances (when the section begins to play its part). Principal players always sit closest to the audience. The cellos are a critical part of orchestral music; all symphonic works involve the cello section, and many pieces require cello soli or solos. Much of the time, cellos provide part of the low-register harmony for the orchestra. Often, the cello section plays the melody for a brief period, before returning to the harmony role. There are also cello concertos, which are orchestral pieces that feature a solo cellist accompanied by an entire orchestra. There are numerous cello concertos – where a solo cello is accompanied by an orchestra – notably 25 by Vivaldi, 12 by Boccherini, at least three by Haydn, three by C. P. E. Bach, two by Saint-Saëns, two by Dvořák, and one each by Robert Schumann, Lalo, and Elgar. There were also some composers who, while not otherwise cellists, did write cello-specific repertoire, such as Nikolaus Kraft who wrote six cello concertos. Beethoven's Triple Concerto for Cello, Violin and Piano and Brahms' Double Concerto for Cello and Violin are also part of the concertante repertoire although in both cases the cello shares solo duties with at least one other instrument. Moreover, several composers wrote large-scale pieces for cello and orchestra, which are concertos in all but name. Some familiar "concertos" are Richard Strauss' tone poem "Don Quixote", Tchaikovsky's "Variations on a Rococo Theme", Bloch's "Schelomo" and Bruch's "Kol Nidrei". In the 20th century, the cello repertoire grew immensely. This was partly due to the influence of virtuoso cellist Mstislav Rostropovich, who inspired, commissioned and premiered dozens of new works. Among these, Prokofiev's "Symphony-Concerto", Britten's "Cello Symphony", the concertos of Shostakovich and Lutosławski as well as Dutilleux's "Tout un monde lointain..." have already become part of the standard repertoire. Other major composers who wrote concertante works for him include Messiaen, Jolivet, Berio and Penderecki. In addition, Arnold, Barber, Glass, Hindemith, Honegger, Ligeti, Myaskovsky, Penderecki, Rodrigo, Villa-Lobos and Walton also wrote major concertos for other cellists, notably for Gaspar Cassadó, Aldo Parisot, Gregor Piatigorsky, Siegfried Palm and Julian Lloyd Webber. There are also many sonatas for cello and piano. Those written by Beethoven, Mendelssohn, Chopin, Brahms, Grieg, Rachmaninoff, Debussy, Fauré, Shostakovich, Prokofiev, Poulenc, Carter, and Britten are particularly well known. Other important pieces for cello and piano include Schumann's five "Stücke im Volkston" and transcriptions like Schubert's Arpeggione Sonata (originally for arpeggione and piano), César Franck's Cello Sonata (originally a violin sonata, transcribed by Jules Delsart with the composer's approval), Stravinsky's "Suite italienne" (transcribed by the composer – with Gregor Piatigorsky – from his ballet "Pulcinella") and Bartók's first rhapsody (also transcribed by the composer, originally for violin and piano). There are pieces for cello solo, J. S. Bach's six Suites for Cello (which are among the best-known solo cello pieces), Kodály's Sonata for Solo Cello and Britten's three Cello Suites. Other notable examples include Hindemith's and Ysaÿe's Sonatas for Solo Cello, Dutilleux's "Trois Strophes sur le Nom de Sacher", Berio's "Les Mots Sont Allés", Cassadó's Suite for Solo Cello, Ligeti's Solo Sonata, Carter's two "Figment"s and Xenakis' "Nomos Alpha" and "Kottos". The cello is a member of the traditional string quartet as well as string quintets, sextet or trios and other mixed ensembles. There are also pieces written for two, three, four or more cellos; this type of ensemble is also called a "cello choir" and its sound is familiar from the introduction to Rossini's William Tell Overture as well as Zaccharia's prayer scene in Verdi's Nabucco. Tchaikovsky's 1812 Overture also starts with a cello ensemble, with four cellos playing the top lines and two violas playing the bass lines. As a self-sufficient ensemble, its most famous repertoire is Villa-Lobos' first of his Bachianas Brasileiras for cello ensemble (the fifth is for soprano and 8 cellos). Other examples are Offenbach's cello duets, quartet, and sextet, Pärt's Fratres for eight cellos and Boulez' "Messagesquisse" for seven cellos, or even Villa-Lobos' rarely played "Fantasia Concertante" (1958) for 32 cellos. The 12 cellists of the Berlin Philharmonic Orchestra (or "the Twelve" as they have since taken to being called) specialize in this repertoire and have commissioned many works, including arrangements of well-known popular songs. The cello is less common in popular music than in classical music. Several bands feature a cello in their standard line-up, including Hoppy Jones of the Ink Spots and Joe Kwon of the Avett Brothers. The more common use in pop and rock is to bring the instrument in for a particular song. In the 1960s, artists such as the Beatles and Cher used the cello in popular music, in songs such as The Beatles' "Yesterday", "Eleanor Rigby" and "Strawberry Fields Forever", and Cher's "Bang Bang (My Baby Shot Me Down)". "Good Vibrations" by the Beach Boys includes the cello in its instrumental ensemble, which includes a number of instruments unusual for this sort of music. Bass guitarist Jack Bruce, who had originally studied music on a performance scholarship for cello, played a prominent cello part in "As You Said" on Cream's "Wheels of Fire" studio album (1968). In the 1970s, the Electric Light Orchestra enjoyed great commercial success taking inspiration from so-called "Beatlesque" arrangements, adding the cello (and violin) to the standard rock combo line-up and in 1978 the UK based rock band, Colosseum II, collaborated with cellist Julian Lloyd Webber on the recording "Variations". Most notably, Pink Floyd included a cello solo in their 1970 epic instrumental "Atom Heart Mother". Bass guitarist Mike Rutherford of Genesis was originally a cellist and included some cello parts in their "Foxtrot" album. Established non-traditional cello groups include Apocalyptica, a group of Finnish cellists best known for their versions of Metallica songs, Rasputina, a group of cellists committed to an intricate cello style intermingled with Gothic music, the Massive Violins, an ensemble of seven singing cellists known for their arrangements of rock, pop and classical hits, Von Cello, a cello fronted rock power trio, Break of Reality who mix elements of classical music with the more modern rock and metal genre, Cello Fury, a cello rock band that performs original rock/classical crossover music, and Jelloslave, a Minneapolis-based Cello duo with two percussionists. These groups are examples of a style that has become known as cello rock. The crossover string quartet bond also includes a cellist. Silenzium and Cellissimo Quartet are Russian (Novosibirsk) groups playing rock and metal and having more and more popularity in Siberia. Cold Fairyland from Shanghai, China is using a cello along a Pipa as the main solo instrument to create East meets West progressive (folk) rock. More recent bands using the cello are Clean Bandit, Aerosmith, The Auteurs, Nirvana, Oasis, Smashing Pumpkins, James, Talk Talk, Phillip Phillips, OneRepublic, and the baroque rock band Arcade Fire. An Atlanta-based trio, King Richard's Sunday Best, also uses a cellist in their lineup. So-called "chamber pop" artists like Kronos Quartet, The Vitamin String Quartet and Margot and the Nuclear So and So's have also recently made cello common in modern alternative rock. Heavy metal band System of a Down has also made use of the cello's rich sound. The indie rock band The Stiletto Formal are known for using a cello as a major staple of their sound, similarly, the indie rock band Canada employs two cello players in their lineup. The orch-rock group, The Polyphonic Spree, which has pioneered the use of stringed and symphonic instruments, employs the cello in very creative ways for many of their "psychedelic-esque" melodies. The first wave screamo band "I Would Set Myself On Fire For You" featured a cello as well as a viola to create a more folk-oriented sound. The band, Panic! at the Disco uses a cello in their song, "Build God, Then We'll Talk". The lead vocalist of the band, Brendon Urie, also did the recording of the cello solo. The Lumineers added cellist Nela Pekarek to the band in 2010. She plays cello, sings harmony and duets. In jazz, bassists Oscar Pettiford and Harry Babasin were among the first to use the cello as a solo instrument; both tuned their instrument in fourths, an octave above the double bass. Fred Katz (who was not a bassist) was one of the first notable jazz cellists to use the instrument's standard tuning and arco technique. Contemporary jazz cellists include Abdul Wadud, Diedre Murray, Ron Carter, Dave Holland, David Darling, Lucio Amanti, Akua Dixon, Ernst Reijseger, Fred Lonberg-Holm, Tom Cora and Erik Friedlander. Modern musical theatre pieces like Jason Robert Brown's The Last Five Years, Duncan Sheik's Spring Awakening, Adam Guettel's Floyd Collins, and Ricky Ian Gordon's My Life with Albertine use small string ensembles (including solo cellos) to a prominent extent. In Indian Classical music Saskia Rao-de Haas is a well established soloist as well as playing duets with her sitarist husband Pt. Shubhendra Rao. Other cellists performing Indian classical music are: Nancy Lesh (Dhrupad) and Anup Biswas. Both Rao and Lesh play the cello sitting cross-legged on the floor. The cello can also be used in bluegrass and folk music, with notable players including Ben Sollee of the Sparrow Quartet and the "Cajun cellist" Sean Grissom, as well as Vyvienne Long who, in addition to her own projects, has played for those of Damien Rice. Cellists such as Natalie Haas, Abby Newton and Liz Davis Maxfield have contributed significantly to the use of cello playing in Celtic folk music, often with the cello featured as a primary melodic instrument and employing the skills and techniques of traditional fiddle playing. Lindsay Mac is becoming well known for playing the cello like a guitar, with her cover of The Beatles' "Blackbird". The cello is typically made from carved wood, although other materials such as carbon fiber or aluminum may be used. A traditional cello has a spruce top, with maple for the back, sides, and neck. Other woods, such as poplar or willow, are sometimes used for the back and sides. Less expensive cellos frequently have tops and backs made of laminated wood. Laminated cellos are widely used in elementary and secondary school orchestras and youth orchestras, because they are much more durable than carved wood cellos (i.e., they are less likely to crack if bumped or dropped) and they are much less expensive. The top and back are traditionally hand-carved, though less expensive cellos are often machine-produced. The sides, or ribs, are made by heating the wood and bending it around forms. The cello body has a wide top bout, narrow middle formed by two C-bouts, and wide bottom bout, with the bridge and F holes just below the middle. The top and back of the cello has decorative border inlay known as purfling. While purfling is attractive, it is also functional: if the instrument is struck, the purfling can prevent cracking of the wood. A crack may form at the rim of the instrument, but spreads no further. Without purfling, cracks can spread up or down the top or back. Playing, traveling and the weather all affect the cello and can increase a crack if purfling is not in place. Less expensive instruments typically have painted purfling. In the late 1920s and early 1930s, the Aluminum Company of America (Alcoa) as well as German luthier G.A. Pfretzschner produced an unknown number of aluminum cellos (in addition to aluminum double basses and violins). Cello manufacturer Luis & Clark constructs cellos from carbon fibre. Carbon fibre instruments are particularly suitable for outdoor playing because of the strength of the material and its resistance to humidity and temperature fluctuations. Luis & Clark has produced over 1000 cellos, some of which are owned by cellists such as Yo-Yo Ma and Josephine van Lier. Above the main body is the carved neck. The neck has a curved cross-section on its underside, which is where the player's thumb runs along the neck during playing. The neck which leads to a pegbox and the scroll. The neck, pegbox, and scroll are normally carved out of a single piece of wood, usually maple. The fingerboard is glued to the neck and extends over the body of the instrument. The fingerboard is given a curved shape, matching the curve on the bridge. Both the fingerboard and bridge need to be curved so that the performer can bow individual strings. If the cello were to have a flat fingerboard and bridge, as with a typical guitar, the performer would only be able to bow the "outer" two strings or bow all the strings. The performer would not be able to play the "inner" two strings alone. The nut is a raised piece of wood, fitted where the fingerboard meets the pegbox, in which the strings rest in shallow slots or grooves to keep them the correct distance apart. The pegbox houses four tapered tuning pegs, one for each string. The pegs are used to tune the cello by either tightening or loosening the string. The pegs are called "friction pegs", because they maintain their position by friction. The scroll is a traditional ornamental part of the cello and a feature of all other members of the violin family. Ebony is usually used for the tuning pegs, fingerboard, and nut, but other hardwoods, such as boxwood or rosewood, can be used. Black fittings on low-cost instruments are often made from inexpensive wood that has been blackened or "ebonized" to look like ebony, which is much harder and more expensive. Ebonised parts such as tuning pegs may crack or split, and the black surface of the fingerboard will eventually wear down to reveal the lighter wood underneath. Historically, cello strings had cores made out of catgut, which, despite its name is made from sheep or goat intestines which are dried out. Most modern strings used in the 2010s are wound with metallic materials like aluminum, titanium and chromium. Cellists may mix different types of strings on their instruments. The pitches of the open strings are C, G, D, and A (black note heads in the playing range figure above), unless alternative tuning (scordatura) is specified by the composer. Some composers (e.g. Ottorino Respighi in the final movement of ‘’The Pines of Rome’’) ask that the low C be tuned down to a B-flat so that the performer can play a different low note on the lowest open string. The tailpiece and endpin are found in the lower part of the cello. The tailpiece is the part of the cello to which the "ball ends" of the strings are attached by passing them through holes. The tailpiece is attached to the bottom of the cello. The tailpiece is traditionally made of ebony or another hard wood, but can also be made of plastic or steel on lower-cost instruments. It attaches the strings to the lower end of the cello, and can have one or more fine tuners. The fine tuners are used to make smaller adjustments to the pitch of the string. The fine tuners can increase the tension of each string (raising the pitch) or decrease the tension of the string (lowering the pitch). When the performer is putting on a new string, the fine tuner for that string is normally reset to a middle position, and then the peg is turned to bring the string up to pitch. The fine turners are used for subtle, minor adjustments to pitch, such as tuning a cello to the oboe's 440 Hz A note or to tune the cello to a piano. The endpin or spike is made of wood, metal or rigid carbon fibre and supports the cello in playing position. The endpin can be retracted into the hollow body of the instrument when the cello is being transported in its case. This makes the cello easier to move about. When the performer wishes to play the cello, the endpin is pulled out to lengthen it. The endpin is locked into the player's preferred length with a screw mechanism. The adjustable nature of endpins enables performers of different ages and body sizes to adjust the endpin length to suit them. In the Baroque period the cello was held between the calves, as there was no endpin at that time. The endpin was "introduced by Adrien Servais 1845 to give the instrument greater stability". Modern endpins are retractable and adjustable; older ones were removed when not in use. (The word "endpin" sometimes also refers to the button of wood located at this place in all instruments in the violin family, but this is usually called "tailpin".) The sharp tip of the cello's endpin is sometimes capped with a rubber tip that protects the tip from dulling and prevents the cello from slipping on the floor. Many cellists use a rubber pad with a metal cup to keep the tip from slipping on the floor. A number of accessories to keep the endpin from slipping; these include ropes which attach to the chair leg and other devices. The bridge holds the strings above the cello and transfers their vibrations to the top of the instrument and the soundpost inside (see below). The bridge is not glued, but rather held in place by the tension of the strings. The bridge is usually positioned by the cross point of the "f-hole" (i.e., where the horizontal line occurs in the "f"). The f-holes, named for their shape, are located on either side of the bridge, and allow air to move in and out of the instrument as part of the sound-production process. They probably actually stand for an old style medial S, for words related to Sound. The f-holes also act as access points to the interior of the cello for repairs or maintenance. Sometimes a small length of rubber hose containing a water-soaked sponge, called a Dampit, is inserted through the f-holes, and serves as a humidifier. This keeps the wood components of the cello from drying out. Internally, the cello has two important features: a bass bar, which is glued to the underside of the top of the instrument, and a round wooden sound post, a solid wooden cylinder which is wedged between the top and bottom plates. The bass bar, found under the bass foot of the bridge, serves to support the cello's top and distribute the vibrations from the strings to the body of the instrument. The sound post, found under the treble side of the bridge, connects the back and front of the cello. Like the bridge, the sound post is not glued, but is kept in place by the tensions of the bridge and strings. Together, the bass bar and sound post transfer the strings' vibrations to the top (front) of the instrument (and to a lesser extent the back), acting as a diaphragm to produce the instrument's sound. Cellos are constructed and repaired using hide glue, which is strong but reversible, allowing for disassembly when needed. Tops may be glued on with diluted glue, since some repairs call for the removal of the top. Theoretically, hide glue is weaker than the body's wood, so as the top or back shrinks side-to-side, the glue holding it lets go, so the plate does not crack. Cellists repairing cracks in their cello do not use regular wood glue, because it cannot be steamed open when a repair has to be made by a luthier. Traditionally, bows are made from pernambuco or brazilwood. Both come from the same species of tree ("Caesalpinia echinata"), but pernambuco, used for higher-quality bows, is the heartwood of the tree and is darker in color than brazilwood (which is sometimes stained to compensate). Pernambuco is a heavy, resinous wood with great elasticity, which makes it an ideal wood for instrument bows. Horsehair is stretched out between the two ends of the bow. The taut horsehair is drawn over the strings to produce the cello's characteristic tone. A small knob is twisted to increase or decrease the tension of the horsehair. The tension on the bow is released when the instrument is not being used. The amount of tension a cellist puts on the bow hair depends on the preferences of the player, the style of music being played, and for students, the preferences of their teacher. Bows are also made from other materials, such as carbon-fibre—stronger than wood—and fiberglass (often used to make inexpensive, lower-quality student bows). An average cello bow is long (shorter than a violin or viola bow) high (from the frog to the stick) and wide. The frog of a cello bow typically has a rounded corner like that of a viola bow, but is wider. A cello bow is roughly heavier than a viola bow, which in turn is roughly heavier than a violin bow. Bow hair is traditionally horsehair, though synthetic hair, in varying colors, is also used. Prior to playing, the musician tightens the bow by turning a screw to pull the frog (the part of the bow under the hand) back, and increase the tension of the hair. Rosin is applied by the player to make the hair sticky. Bows need to be re-haired periodically. Baroque style (1600–1750) cello bows were much thicker and were formed with a larger outward arch when compared to modern cello bows. The inward arch of a modern cello bow produces greater tension, which in turn gives off a louder sound. The cello bow has also been used to play electric guitars. Jimmy Page pioneered its application on tracks such as "Dazed and Confused". The post-rock Icelandic band Sigur Rós's lead singer often plays a guitar using a cello bow. In 1989, the German cellist Michael Bach began developing a curved bow, encouraged by John Cage, Dieter Schnebel, Mstislav Rostropovich and Luigi Colani: and since then many pieces have been composed especially for it. This curved bow ("BACH.Bow") is a convex curved bow which, unlike the ordinary bow, renders possible polyphonic playing on the various strings of the instrument. The solo repertoire for violin and cello by J. S. Bach the BACH.Bow is particularly suited to it: and it was developed with this in mind, polyphonic playing being required, as well as monophonic. When a string is bowed or plucked, it vibrates and moves the air around it, producing sound waves. Because the string is quite thin, not much air is moved by the string itself, and consequently if the string was not mounted on a hollow body, the sound would be weak. In acoustic stringed instruments such as the cello, this lack of volume is solved by mounting the vibrating string on a larger hollow wooden body. The vibrations are transmitted to the larger body, which can move more air and produce a louder sound. Different designs of the instrument produces variations in the instrument’s vibrational patterns and thus changes the character of the sound produced. A string’s fundamental pitch can be adjusted by changing its stiffness, which depends on tension and length. Tightening a string stiffens it by increasing both the outward forces along its length and the net forces it experiences during a distortion. A cello can be tuned by adjusting the tension of its strings, by turning the tuning pegs mounted on its pegbox, and tension adjusters (fine tuners) on the tail piece. A string's length also affects its fundamental pitch. Shortening a string stiffens it by increasing its curvature during a distortion and subjecting it to larger net forces. Shortening the string also reduces its mass, but does not alter the mass per unit length, and it is the latter ratio rather than the total mass which governs the frequency. The string vibrates in a standing wave whose speed of propagation is given by , where "T" is the tension and "m" is the mass per unit length; there is a node at either end of the vibrating length, and thus the vibrating length "l" is half a wavelength. Since the frequency of any wave is equal to the speed divided by the wavelength, we have frequency = . (Note that some writers, including Muncaster (cited below) use the Greek letter "μ" in place of "m".) Thus shortening a string increases the frequency, and thus the pitch. Because of this effect, you can raise and change the pitch of a string by pressing it against the fingerboard in the cello’s neck and effectively shortening it. Likewise strings with less mass per unit length, if under the same tension, will have a higher frequency and thus higher pitch than more massive strings. This is a prime reason why the different strings on all string instruments have different fundamental pitches, with the lightest strings having the highest pitches. A played note of E or F has a frequency which is often very close to the natural resonating frequency of the body of the instrument, and if the problem is not addressed this can set the body into near resonance. This may cause an unpleasant sudden amplification of this pitch, and additionally a loud beating sound results from the interference produced between these nearby frequencies; this is known as the “wolf tone” because it is an unpleasant growling sound. The wood resonance appears to be split into two frequencies by the driving force of the sounding string. These two periodic resonances beat with each other. This wolf tone must be eliminated or significantly reduced for the cello to play the nearby notes with a pleasant tone. This can be accomplished by modifying the cello front plate, attaching a wolf eliminator (a metal cylinder or a rubber cylinder encased in metal), or moving the sound post. When a string is bowed or plucked to produce a note, the fundamental note is accompanied by higher frequency overtones. Each sound has a particular recipe of frequencies that combine to make the total sound. Playing the cello is done while seated with the instrument supported on the floor by the endpin. The left hand fingertips stop the strings on the fingerboard, determining the pitch of the fingered note. The right hand plucks or bows the strings to sound the notes. The left hand fingertips stop the strings along their length, determining the pitch of each fingered note. Stopping the string closer to the bridge results in higher-pitched sound, because the vibrating string length has been shortened. In the "neck" positions (which use just less than half of the fingerboard, nearest the top of the instrument), the thumb rests on the back of the neck; in "thumb position" (a general name for notes on the remainder of the fingerboard) the thumb usually rests alongside the fingers on the string and the side of the thumb is used to play notes. The fingers are normally held curved with each knuckle bent, with the fingertips in contact with the string. If a finger is required on two (or more) strings at once to play perfect fifths (in double stops or chords) it is used flat. In slower, or more expressive playing, the contact point can move slightly away from the nail to the pad of the finger, allowing a fuller vibrato. Vibrato is a small oscillation in the pitch of a note, usually considered expressive. Harmonics played on the cello fall into two classes; natural and artificial. Natural harmonics are produced by lightly touching (but not depressing) the string with the finger at certain places, and then bowing (or, rarely, plucking) the string. For example, the halfway point of the string will produce a harmonic that is one octave above the unfingered (open) string. Natural harmonics only produce notes that are part of the harmonic series on a particular string. Artificial harmonics (also called false harmonics or stopped harmonics), in which the player depresses the string fully with one finger while touching the same string lightly with another finger, can produce any note above middle C. Glissando (Italian for "sliding") is an effect played by sliding the finger up or down the fingerboard without releasing the string. This causes the pitch to rise and fall smoothly, without separate, discernible steps. In cello playing, the bow is much like the breath of a wind instrument player. Arguably, it is the major determinant in the expressiveness of the playing. The right hand holds the bow and controls the duration and character of the notes. The bow is drawn across the strings roughly halfway between the end of the fingerboard and the bridge, in a direction perpendicular to the strings. The bow is held with all five fingers of the right hand, the thumb opposite the fingers and closer to the cellist's body. Tone production and volume of sound depend on a combination of several factors. The three most important ones are: bow speed, weight applied to the string, and point of contact of the bow hair with the string. Double stops involve the playing of two notes at the same time. Two strings are fingered simultaneously, and the bow is drawn so as to sound them both at once. In pizzicato playing, the string is plucked directly with the fingers or thumb. Pizzicato is often abbreviated as "Pizz.". Position of the hand is slightly over the finger board and away from the bridge. A player using the col legno technique strikes or rubs the strings with the wood of the bow rather than the hair. In spiccato playing, the strings are not "drawn" by the bow hair but struck by it, while still retaining some horizontal motion, to generate a more percussive, crisp sound. In staccato, the player moves the bow a small distance and stops it on the string, making a short sound, the rest of the written duration being taken up by silence. Legato is a technique where the notes are smoothly connected without accents or breaks. It is noted by a slur (curved line) above or below – depending on their position on the staff – the notes of the passage that is to be played legato. "Sul ponticello" ("on the bridge") refers to bowing closer to the bridge, while "sul tasto" ("on the fingerboard") calls for bowing nearer the end of the fingerboard. Sul tasto produces a more flute-like sound, with more emphasis on the fundamental frequency of the note, and softer overtones. Standard-sized cellos are referred to as "full-size" or "" but are also made in smaller (fractional) sizes (e.g. , , , , , , ). The fractions refer to volume rather than length, so a 1/2 size cello is much longer than half the length of a full size. The smaller cellos are identical to standard cellos in construction, range, and usage, but are simply scaled-down for the benefit of children and shorter adults. Cellos in sizes larger than do exist, and cellists with unusually large hands may require such a non-standard instrument. Cellos made before approximately 1700 tended to be considerably larger than those made and commonly played today. Around 1680, changes in string-making technology made it possible to play lower-pitched notes on shorter strings. The cellos of Stradivari, for example, can be clearly divided into two models: the style made before 1702, characterized by larger instruments (of which only three exist in their original size and configuration), and the style made during and after 1707, when Stradivari began making smaller cellos. This later model is the design most commonly used by modern luthiers. The scale length of a cello is about . The new size offered fuller tonal projection and greater range of expression. The instrument in this form was able to contribute to more pieces musically and offered the possibility of greater physical dexterity for the player to develop technique. There are many accessories for the cello. Cellos are made by luthiers, specialists in building and repairing stringed instruments, ranging from guitars to violins. The following luthiers are notable for the cellos they have produced: A person who plays the cello is called a "cellist". For a list of notable cellists, see the list of cellists and . Specific instruments are famous (or become famous) for a variety of reasons. An instrument's notability may arise from its age, the fame of its maker, its physical appearance, its acoustic properties, and its use by notable performers. The most famous instruments are generally known for all of these things. The most highly prized instruments are now collector's items, and are priced beyond the reach of most musicians. These instruments are typically owned by some kind of organization or investment group, which may loan the instrument to a notable performer. (For example, the Davidov Stradivarius, which is currently in the possession of one of the most widely known living cellists, Yo-Yo Ma, is actually owned by the Vuitton Foundation.) Some notable cellos: "La Mariée", a 1950 painting by Marc Chagall which is prominently featured in the 1999 film "Notting Hill", portrays a goat playing a cello.
https://en.wikipedia.org/wiki?curid=6558
Control store A control store is the part of a CPU's control unit that stores the CPU's microprogram. It is usually accessed by a microsequencer. A control store implementation whose contents are unalterable is known as a Read Only Memory (ROM) or Read Only Storage (ROS); one whose contents are alterable is known as a Writable Control Store (WCS). Early control stores were implemented as a diode-array accessed via address decoders, a form of read-only memory. This tradition dates back to the "program timing matrix" on the MIT Whirlwind, first described in 1947. Modern VLSI processors instead use matrices of field-effect transistors to build the ROM and/or PLA structures used to control the processor as well as its internal sequencer in a microcoded implementation. IBM System/360 used a variety of techniques: CCROS (Card Capacitor Read-Only Storage) on the Model 30, TROS (Transformer Read-Only Storage) on the Model 40, and BCROS (Balanced Capacitor Read-Only Storage) on the Model 50. Some computers were built using "writable microcode" — rather than storing the microcode in ROM or hard-wired logic, the microcode was stored in a RAM called a "writable control store" or "WCS". Such a computer is sometimes called a "Writable Instruction Set Computer" or "WISC". Many of these machines were experimental laboratory prototypes, such as the WISC CPU/16 and the RTX 32P. The original System/360 models had read-only control store, but later System/360, System/370 and successor models loaded part or all of their microprograms from floppy disks or other DASD into a writable control store consisting of ultra-high speed random-access read-write memory. The System/370 architecture included a facility called Initial-Microprogram Load (IML or IMPL) that could be invoked from the console, as part of Power On Reset (POR) or from another processor in a tightly coupled multiprocessor complex. This permitted IBM to easily repair microprogramming defects in the field. Even when the majority of the control store is stored in ROM, computer vendors would often sell writable control store as an option, allowing the customers to customize the machine's microprogram. Other vendors, e.g., IBM, use the WCS to run microcode for emulator features and hardware diagnostics. Other commercial machines that used writable microcode include the Burroughs Small Systems (1970s and 1980s), the Xerox processors in their Lisp machines and Xerox Star workstations, the DEC VAX 8800 ("Nautilus") family, and the Symbolics L- and G-machines (1980s). Some DEC PDP-10 machines stored their microcode in SRAM chips (about 80 bits wide x 2 Kwords), which was typically loaded on power-on through some other front-end CPU. Many more machines offered user-programmable writable control stores as an option (including the HP 2100, DEC PDP-11/60 and Varian Data Machines V-70 series minicomputers). The Mentec M11 and Mentec M1 stored its microcode in SRAM chips, loaded on power-on through another CPU. The Data General Eclipse MV/8000 ("Eagle") had a SRAM writable control store, loaded on power-on through another CPU. WCS offered several advantages including the ease of patching the microprogram and, for certain hardware generations, faster access than ROMs could provide. User-programmable WCS allowed the user to optimize the machine for specific purposes. Some CPU designs compile the instruction set to a writable RAM or FLASH inside the CPU (such as the Rekursiv processor and the Imsys Cjip), or an FPGA (reconfigurable computing). Several Intel CPUs in the x86 architecture family have writable microcode, starting with the Pentium Pro in 1995. This has allowed bugs in the Intel Core 2 microcode and Intel Xeon microcode to be fixed in software, rather than requiring the entire chip to be replaced. Such fixes can be installed by Linux, FreeBSD, Microsoft Windows, or the motherboard BIOS. The control store usually has a register on its outputs. The outputs that go back into the sequencer to determine the next address have to go through some sort of register to prevent the creation of a race condition. In most designs all of the other bits also go through a register. This is because the machine will work faster if the execution of the next microinstruction is delayed by one cycle. This register is known as a pipeline register. Very often the execution of the next microinstruction is dependent on the result of the current microinstruction, which will not be stable until the end of the current microcycle. It can be seen that either way, all of the outputs of the control store go into one big register. Historically it used to be possible to buy EPROMs with these register bits on the same chip. The clock signal determining the clock rate, which is the cycle time of the system, primarily clocks this register.
https://en.wikipedia.org/wiki?curid=6559
Columba Columba (, 'church dove'; Scots Gaelic: "Calum Cille", ; 7 December 521 – 9 June 597) was an Irish abbot and missionary evangelist credited with spreading Christianity in what is today Scotland at the start of the Hiberno-Scottish mission. He founded the important abbey on Iona, which became a dominant religious and political institution in the region for centuries. He is the patron saint of Derry. He was highly regarded by both the Gaels of Dál Riata and the Picts, and is remembered today as a Catholic saint and one of the Twelve Apostles of Ireland. In Ireland, he is commonly known as Colmcille. Columba studied under some of Ireland's most prominent church figures and founded several monasteries in the country. Around 563 he and his twelve companions crossed to Dunaverty near Southend, Argyll, in Kintyre before settling in Iona in Scotland, then part of the Ulster kingdom of Dál Riata, where they founded a new abbey as a base for spreading Celtic Christianity among the northern Pictish kingdoms who were pagan. He remained active in Irish politics, though he spent most of the remainder of his life in Scotland. Three surviving early medieval Latin hymns may be attributed to him. Columba was born to Fedlimid and Eithne of the Cenel Conaill in Gartan, a district beside Lough Gartan, in Tír Chonaill (mainly modern County Donegal) in the north of Ireland. On his father's side, he is claimed as being great-great-grandson of Niall of the Nine Hostages, a pseudo-historical Irish high king of the 5th century. He was baptised in Temple-Douglas, in the County Donegal parish of Conwal (midway between Gartan and Letterkenny), by his teacher and foster-uncle Cruithnechán. It is not known for sure if his name at birth was Colmcille or if he adopted this name later in life; Adomnán (Eunan) of Iona thought it was his birth name but other Irish sources have claimed his name at birth was Crimthann (meaning 'fox'). In the Irish language his name means 'dove', which is the same name as the Prophet Jonah (Jonah in Hebrew is also 'dove'), which Adomnán of Iona as well as other early Irish writers were aware of, although it is not clear if he was deliberately named after Jonah or not. When sufficiently advanced in letters he entered the monastic school of Movilla, at Newtownards, under Finnian of Movilla who had studied at Ninian's "Magnum Monasterium" on the shores of Galloway. He was about twenty, and a deacon when, having completed his training at Movilla, he travelled southwards into Leinster, where he became a pupil of an aged bard named Gemman. On leaving him, Columba entered the monastery of Clonard, governed at that time by Finnian, noted for sanctity and learning. Here he imbibed the traditions of the Welsh Church, for Finnian had been trained in the schools of Saint David. In early Christian Ireland the druidic tradition collapsed due to the spread of the new Christian faith. The study of Latin learning and Christian theology in monasteries flourished. Columba became a pupil at the monastic school at Clonard Abbey, situated on the River Boyne in modern County Meath. During the sixth century, some of the most significant names in the history of Celtic Christianity studied at the Clonard monastery. The average number of scholars under instruction at Clonard was said to be 300. Columba was one of twelve students of St Finnian who became known as the Twelve Apostles of Ireland. He became a monk and eventually was ordained a priest. Another preceptor of Columba was St Mobhi, whose monastery at Glasnevin was frequented by such famous men as St Canice, St Comgall, and St Ciarán. A pestilence which devastated Ireland in 544 caused the dispersion of Mobhi's disciples, and Columba returned to Ulster, the land of his kindred. He was a striking figure of great stature and powerful build, with a loud, melodious voice which could be heard from one hilltop to another. The following years were marked by the foundation of several important monasteries: Derry, at the southern edge of Inishowen; Durrow, County Offaly; Kells, County Meath; and Swords. While at Derry it is said that he planned a pilgrimage to Rome and Jerusalem, but did not proceed farther than Tours. Thence he brought a copy of those gospels that had lain on the bosom of St. Martin for the space of 100 years. This relic was deposited in Derry. Some traditions assert that sometime around 560 Columba became involved in a quarrel with Finnian of Moville of Movilla Abbey over a psalter. Columba copied the manuscript at the scriptorium under Finnian, intending to keep the copy. Finnian disputed his right to keep it. There is a suggestion that this conflict resulted in the Battle of Cúl Dreimhne in Cairbre Drom Cliabh (now in County Sligo) in 561, during which many men were killed. Richard Sharpe, translator of Adomnán's Life of St. Columba (referenced in the bibliography below) makes a stern caution at this point against accepting the many references that link the battle and Columba's leaving of Ireland, even though there is evidence in the annals that Columba supported his own king against the high king. Political conflicts that had existed for some time resulted in the clan Neill's battle against King Diarmait at Cooldrevny in 561. An issue, for example, was the king's violation of the right of sanctuary belonging to Columba's person as a monk on the occasion of the murder of Prince Curnan, the Columba's kinsman. Prince Curnan of Connacht, who had fatally injured a rival in a hurling match and had taken refuge with Columba, was dragged from his protector's arms and slain by Diarmaid's men, in defiance of the rights of sanctuary. A synod of clerics and scholars threatened to excommunicate him for these deaths, but Brendan of Birr spoke on his behalf. Eventually the process was deemed a miscarriage of justice. Columba's own conscience was uneasy, and on the advice of an aged hermit, Molaise, he resolved to expiate his sense of offence by departing Ireland. The term "exile" is used in some references. This, too, can be disputed, for the term "pilgrimage" is used more frequently in the literature about him. A marker at Stroove Beach on the Inishowen Peninsula commemorates the place where St. Columba set sail for Scotland. He left Ireland, but through the following years he would return several times in relationships with the communities he had founded there.Columba's copy of the psalter has been traditionally associated with the Cathach of St. Columba. In 574/5 during his return for the Synod of Drum Ceat he founded the monastery of Drumcliff in Cairbre, now County Sligo, near the battlefield. In 563, he travelled to Scotland with twelve companions (said to include Odran of Iona) in a wicker currach covered with leather. According to legend he first landed on the Kintyre Peninsula, near Southend. However, being still in sight of his native land, he moved farther north up the west coast of Scotland. The island of Iona was made over to him by his kinsman Conall mac Comgaill King of Dál Riata, who perhaps had invited him to come to Scotland in the first place. However, there is a sense in which he was not leaving his native people, as the Ulster Gaels had been colonising the west coast of Scotland for the previous couple of centuries. Aside from the services he provided guiding the only centre of literacy in the region, his reputation as a holy man led to his role as a diplomat among the tribes. There are also many stories of miracles which he performed during his work to convert the Picts, the most famous being his encounter with an unidentified animal that some have equated with the Loch Ness Monster in 565. It is said that he banished a ferocious "water beast" to the depths of the River Ness after it had killed a Pict and then tried to attack Columba's disciple named Lugne (see Vita Columbae Book 2 below). He visited the pagan King Bridei, King of Fortriu, at his base in Inverness, winning Bridei's respect, although not his conversion. He subsequently played a major role in the politics of the country. He was also very energetic in his work as a missionary, and, in addition to founding several churches in the Hebrides, he worked to turn his monastery at Iona into a school for missionaries. He was a renowned man of letters, having written several hymns and being credited with having transcribed 300 books. One of the few, if not the only, times he left Scotland was towards the end of his life, when he returned to Ireland to found the monastery at Durrow. Columba died on Iona and was buried in 597 by his monks in the abbey he created. In 794 the Vikings descended on Iona. Columba's relics were finally removed in 849 and divided between Scotland and Ireland. The parts of the relics which went to Ireland are reputed to be buried in Downpatrick, County Down, with Saint Patrick and Brigid of Kildare or at Saul Church neighbouring Downpatrick. (Names of Iona), Inchcolm and Eilean Chaluim Chille. Colmcille is one of the three patron saints of Ireland, after Patrick and Brigid of Kildare. Colmcille is the patron-saint of the city of Derry, where he founded a monastic settlement in c. 540. The name of the city in Irish is "Doire Colmcille" and is derived from the native oak trees in the area and the city's association with Colmcille. The Catholic Church of Saint Colmcille's Long Tower, and the Church of Ireland St Augustine's Church both claim to stand at the spot of this original settlement. The Church of Ireland Cathedral, St Columb's Cathedral, and the largest park in the city, St. Columb's Park, are named in his honour. St. Columba's Primary School in Drumcondra is a girl's school named after the saint. St. Colmcille's Primary School and St. Colmcille's Community School are two schools in Knocklyon, Dublin, named after him, with the former having an annual day dedicated to the saint on 9 June. The town of Swords, Dublin was reputedly founded by Colmcille in 560 AD. St Colmcille's Boys’ National School and St. Colmcille's Girls’ National School, both located in the town of Swords, are also named after the Saint as is one of the local gaelic teams, Naomh Colmcille. The Columba Press, a religious and spiritual book company based in Dublin, is named after Colmcille. Aer Lingus, Ireland's national flag carrier has named one of its Airbus A330 aircraft in commemoration of the saint (reg: EI-DUO). Columba is credited as being a leading figure in the revitalisation of monasticism. The Clan Malcolm/Clan McCallum claims its name from Columba and was reputedly founded by the descendants of his original followers. It is also said that Clan Robertson Clan Donnachaidh / Duncan are heirs of Columba. Clan MacKinnon may also have some claim to being spiritual descendants of St Columba as after he founded his monastery on Isle Iona, the MacKinnons were the abbots of the Church for centuries. This would also account for the fact that Clan MacKinnon is amongst the ancient clans of Scotland. The cathedral of the Catholic Diocese of Argyll and the Isles is placed under the patronage of Saint Columba, as are numerous Catholic schools and parishes throughout the nation. The Scottish Episcopal Church, the Church of Scotland, and the Evangelical Lutheran Church of England also have parishes dedicated to him. The village of Kilmacolm in Renfrewshire is also derived from Colmcille's name. St Columba's Hospice, a prominent hospice in Edinburgh, is named after the saint. Columba currently has two poems attributed to him: "Adiutor Laborantium" and "Altus Prosator". Both poems are examples of Abecedarian hymns in Latin written while Columba was at the Iona Abbey. The shorter of the two poems, "Adiutor Laborantium" consists of twenty-seven lines of eight syllables each, with each line following the format of an Abecedarian hymn using the Classical Latin alphabet save for lines 10-11 and 25–27. The content of the poem addresses God as a helper, ruler, guard, defender and lifter for those who are good and an enemy of sinners whom he will punish. "Altus Prosator" consists of twenty-three stanzas sixteen syllables long, with the first containing seven lines and six lines in each subsequent stanza. It uses the same format and alphabet as "Adiutor Laborantium" except with each stanza starting with a different letter rather than each line. The poem tells a story over three parts split into the beginning of time, history of Creation, and the Apocalypse or end of time. As of 2011, Canadians who are of Scottish ancestry are the third largest ethnic group in the country and thus Columba's name is to be found attached to Catholic, Anglican and Presbyterian parishes. This is particularly the case in eastern Canada, apart from French-speaking Quebec. Throughout the US there are numerous parishes within the Catholic and Episcopalian denominations dedicated to Columba. Within the Protestant tradition the Presbyterian Church (which has its roots in Scottish Presbyterianism) also has parishes named in honour of Columba. There is even an Orthodox Church monastery dedicated to the saint in the Massachusetts town of Southbridge. Columba is the patron saint of the Roman Catholic Diocese of Youngstown, Ohio. The Cathedral there is named for him. Iona College, a small Catholic liberal arts college in New Rochelle, New York, is named after the island on which Columba established his first monastery in Scotland, as is Iona College in Windsor, Ontario, Iona Presentation College, Perth, and Iona College Geelong in Charlemont, Victoria. There are at least four pipe bands named for him; one each from Tullamore, Ireland, from Derry, Northern Ireland, from Kearny, New Jersey, and from Cape Cod, Massachusetts. St. Columba's School one of the most prominent English-Medium schools in India run by the Irish Christian Brothers is also named after the saint. The Munich GAA is named München Colmcilles. Saint Columba's Feast Day, 9 June, has been designated as International Celtic Art Day. The Book of Kells and the Book of Durrow, great medieval masterpieces of Celtic art, are associated with Columba. Benjamin Britten composed "A Hymn of St Columba" for choir and organ in 1962, setting a poem by the saint, on the occasion of the 1,400th anniversary of his voyage to Iona. The main source of information about Columba's life is the Vita Columbae, a hagiography written in the style of "saint's lives" narratives that had become widespread throughout medieval Europe. Compiled and drafted by scribes and clergymen, these accounts were written in Latin and served as written collections of the deeds and miracles attributed to the saint, both during his or her life or after death. The canonization of a saint, especially one who had lived on the fringes of the medieval Christian world like Columba, required a well-written hagiography to be submitted to Rome, but popular belief and local cults of sainthood often led to the veneration of these men and women without official approval from the Church. Writing a century after the death of Columba, the author Adomnán (also known as Eunan), served as the ninth Abbot of Iona until his death in 704. James Earle Fraser asserts that Adomnán drew extensively from an existing body of accounts regarding the life of Columba, including a Latin collection entitled "De uirtutibus sancti Columbae", composed c. 640 A.D. This earlier work is attributed to Cummene Find, who became the abbot of Iona and served as the leader of the monastic island community from 656 until his death in 668 A.D. or 669 A.D. While the Vita Columbae often conflicts with contemporaneous accounts of various battles, figures, and dates, it remains the most important surviving work from early medieval Scotland and provides a wealth of knowledge regarding the Picts and other ethnic and political groups from this time period. The Vita also offers a valuable insight into the monastic practices of Iona and the daily life of the early medieval Gaelic monks. The surviving manuscripts include: Instead of relying on chronological order, Adomnán categorises the events recorded in the "Vita Columbae" into three different books: Columba's Prophecies, Columba's Miracles, and Columba's Apparitions. In the first book, the author Adomnán lists Columba's prophetic revelations, which come as a result of the his ability to view the present and the future simultaneously. Most of the short chapters begin with Columba informing his fellow monks that a person will soon arrive on the island or an event will imminently occur. In one notable instance, Columba appears in a dream to King Oswald of Northumbria, and announces the king's incoming victory against the King Catlon (Cadwallon of Wales) in the Battle of Heavenfield. The people of Britain promise to convert to Christianity and receive baptism after the conclusion of the war. This victory signals the re-Christianizing of pagan England, and establishes King Oswald as ruler of the entirety of Britain. Columba's other prophecies include when he sends a man named Batain off to perform his penance, but then Columba turns to his friends and says Batain will instead return to Scotia and be killed by his enemies. Several of Saint Columba's prophecies reflect the scribal culture in which he was immersed, such his miraculous knowledge of the missing letter "I” from Baithene's psalter or when he prophecies that an eager man will knock over his inkhorn and spill its contents. In the second book, Columba performs various miracles such as healing people with diseases, expelling malignant spirits, subduing wild beasts, calming storms, and even returning the dead to life. He also performs agricultural miracles that would hold a special significance to the common people of Ireland and the Britain such as when he casts a demon out of a pail and restores the spilt milk to its container. The Vita contains a story that has been interpreted as the first reference to the Loch Ness Monster. According to Adomnán, Columba came across a group of Picts burying a man who had been killed by the monster. Columba saves a swimmer from the monster with the sign of the Cross and the imprecation, "Thou shalt go no further, nor touch the man; go back with all speed." The beast flees, terrified, to the amazement of the assembled Picts who glorified Columba's God. Whether or not this incident is true, Adomnan's text specifically states that the monster was swimming in the River Ness – the river flowing from the loch – rather than in Loch Ness itself. In book three, Adomnán describes different apparitions of the Saint, both that Columba receives and those that are seen by others regarding him. He mentions that, "For indeed after the lapse of many years, ... St. Columba was excommunicated by a certain synod for some pardonable and very trifling reasons, and indeed unjustly" (P.79- 80). In one of the accounts, Columba, in this period of excommunication, goes to a meeting held against him in Teilte. Brendán, despite of all the negative reactions among the seniors toward Columba, kisses him reverently and assures that Columba is the man of God and that he sees Holy Angels accompanying Columba on his journey through the plain. In the last Chapter, Columba foresees his own death when speaking to his attendant: This day in the Holy Scriptures is called the Sabbath, which means rest. And this day is indeed a Sabbath to me, for it is the last day of my present laborious life, and on it I rest after the fatigues of my labours; and this night at midnight, which commenceth the solemn Lord's Day, I shall, according to the sayings of Scripture, go the way of our fathers. For already my Lord Jesus Christ deigneth to invite me; and to Him, I say, in the middle of this night shall I depart, at His invitation. For so it hath been revealed to me by the Lord himself. And when the bell strikes midnight, Columba goes to the church and kneels beside the altar. His attendant witnesses heavenly light in the direction of Columba, and angels join him in his passage to the Lord: And having given them his holy benediction in this way, he immediately breathed his last. After his soul had left the tabernacle of the body, his face still continued ruddy, and brightened in a wonderful way by his vision of the angels, and that to such a degree that he had the appearance, not so much of one dead, as of one alive and sleeping. Both the "Vita Columbae" and the Bede (672/673-735) record Columba's visit to Bridei. Whereas Adomnán just tells us that Columba visited Bridei, Bede relates a later, perhaps Pictish tradition, whereby Columba actually converts the Pictish king. Another early source is a poem in praise of Columba, most probably commissioned by Columba's kinsman, the King of the Uí Néill clan. It was almost certainly written within three or four years of Columba's death and is the earliest vernacular poem in European history. It consists of 25 stanzas of four verses of seven syllables each, called the Amra Coluim Chille. Through the reputation of its venerable founder and its position as a major European centre of learning, Columba's Iona became a place of pilgrimage. Columba is historically revered as a warrior saint, and was often invoked for victory in battle. His relics were finally removed in 849 and divided between Alba and Ireland. Relics of Columba were carried before Scottish armies in the reliquary made at Iona in the mid-8th century, called the Brecbennoch. Legend has it that the Brecbennoch was carried to the Battle of Bannockburn (24 June 1314) by the vastly outnumbered Scots army and the intercession of Columba helped them to victory. Since the 19th century the "Brecbennoch of St. Columba" has been identified with the Monymusk Reliquary, although this is now doubted by scholars. In the Antiphoner of Inchcolm Abbey, the "Iona of the East" (situated on an island in the Firth of Forth), a 14th-century prayer begins "O Columba spes Scotorum..." "O Columbus, hope of the Scots".
https://en.wikipedia.org/wiki?curid=6561
Conditional proof A conditional proof is a proof that takes the form of asserting a conditional, and proving that the antecedent of the conditional necessarily leads to the consequent. The assumed antecedent of a conditional proof is called the conditional proof assumption (CPA). Thus, the goal of a conditional proof is to demonstrate that if the CPA were true, then the desired conclusion necessarily follows. The validity of a conditional proof does not require that the CPA be actually true, only that "if it were true" it would lead to the consequent. Conditional proofs are of great importance in mathematics. Conditional proofs exist linking several otherwise unproven conjectures, so that a proof of one conjecture may immediately imply the validity of several others. It can be much easier to show a proposition's truth to follow from another proposition than to prove it independently. A famous network of conditional proofs is the NP-complete class of complexity theory. There is a large number of interesting tasks, and while it is not known if a polynomial-time solution exists for any of them, it is known that if such a solution exists for any of them, one exists for all of them. Similarly, the Riemann hypothesis has many consequences already proven. As an example of a conditional proof in symbolic logic, suppose we want to prove A → C (if A, then C) from the first two premises below:
https://en.wikipedia.org/wiki?curid=6562
Conjunction introduction Conjunction introduction (often abbreviated simply as conjunction and also called and introduction) is a valid rule of inference of propositional logic. The rule makes it possible to introduce a conjunction into a logical proof. It is the inference that if the proposition "p" is true, and proposition "q" is true, then the logical conjunction of the two propositions "p and q" is true. For example, if it is true that "it's raining", and it is true that "I'm inside", then it is true that "it's raining and I'm inside". The rule can be stated: where the rule is that wherever an instance of "formula_2" and "formula_3" appear on lines of a proof, a "formula_4" can be placed on a subsequent line. The "conjunction introduction" rule may be written in sequent notation: where formula_2 and formula_3 are propositions expressed in some formal system, and formula_8 is a metalogical symbol meaning that formula_4 is a syntactic consequence if formula_2 and formula_3 are each on lines of a proof in some logical system;
https://en.wikipedia.org/wiki?curid=6563
English in the Commonwealth of Nations The use of the English language in most current and former member countries of the Commonwealth of Nations was inherited from British colonisation. Mozambique is an exception – although English is widely spoken there, it is a former Portuguese colony which joined the Commonwealth in 1996. English is spoken as a first or second language in most of the Commonwealth. In a few countries, such as Cyprus and Malaysia, it does not have official status, but is widely used as a lingua franca. Many regions, notably Australia, Brunei, Canada, Hong Kong, India, Ireland, Malaysia, New Zealand, Pakistan, Singapore, South Africa, Sri Lanka and the Caribbean, have developed their own native varieties of the language. Written English as used in the current and former Commonwealth generally favours British spelling as opposed to American, with some exceptions in Canada, where there is a strong influence from neighbouring American English (collectively, the US and Canadian dialects form North American English). Few said countries besides Canada and Australia have produced their own dictionaries and style guides from major publishers, and rely on those produced in the United Kingdom, especially for formal writing. The report of the Inter-Governmental Group on Criteria for Commonwealth Membership states that English is a symbol of Commonwealth heritage and unity. Southern Hemisphere native varieties of English began to develop during the 18th century, with the colonisation of Australasia and South Africa. Australian English and New Zealand English are closely related to each other, and share some similarities with South African English (though it has unique influences from indigenous African languages, and Dutch influences it inherited along with the development of Afrikaans from Dutch). The vocabularies of these dialects draw from both British English (in the main) and American English (especially for recent terminology), as well as numerous native peculiarities. Canadian English contains elements of British English and American English, as well as many Canadianisms and some French influences. It is the product of several waves of immigration and settlement, from Britain, Ireland, France, the United States, and around the world, over a period of almost two centuries. Modern Canadian English has taken significant vocabulary and spelling from the shared political and social institutions of Commonwealth countries. Caribbean English is influenced by the English-based Creole varieties spoken, but they are not one and the same. There is a great deal of variation in the way English is spoken, with a "Standard English" at one end of a bipolar linguistic continuum and Creole languages at the other. These dialects have roots in 17th-century British and Irish English, and African languages, plus localised influences from other colonial languages including French, Spanish, and Dutch; unlike most native varieties of English, West Indian dialects often tend to be syllable-timed rather than stress-timed. Second-language varieties of English in Africa and Asia have often undergone "indigenisation"; that is, each English-speaking community has developed (or is in the process of developing) its own standards of usage, often under the influence of local languages. These dialects are sometimes referred to as "New Englishes" (McArthur, p. 36); most of them inherited non-rhoticity from Southern British English. Several dialects of West African English exist, with a lot of regional variation and some influence from indigenous languages. West African English tends to be syllable-timed, and its phoneme inventory is much simpler than that of Received Pronunciation; this sometimes affects mutual intelligibility with native varieties of English. A distinctive East African English, often with significant influences from Bantu languages such as Swahili, is spoken in countries such as Kenya or Tanzania, particularly in Nairobi and other cities where there is an expanding middle class, for whom English is increasingly being used in the home as the first language. Small communities of native English speakers can be found in Zimbabwe, Botswana, and Namibia; the dialects spoken are similar to native South African English. India has the largest English-speaking population in the Commonwealth, although comparatively few speakers of Indian English are first-language speakers. The same is true of English spoken in other parts of South Asia, e.g. Pakistani English, Sri Lankan English, and Bangladeshi English. South Asian English phonology is highly variable; stress, rhythm and intonation are generally different from those of native varieties. There are also several peculiarities at the levels of morphology, syntax and usage, some of which can also be found among educated speakers. Southeast Asian English comprises Singapore English, Malaysian English, and Brunei English; it features some influence from Malay and Chinese languages, as well as Indian English. Hong Kong ceased to be part of the Commonwealth in 1997. Nonetheless, the English language there still enjoys status as an official language. Other languages:
https://en.wikipedia.org/wiki?curid=6566
Charles McCarry Charles McCarry (June 14, 1930 – February 26, 2019) was an American writer, primarily of spy fiction, and a former undercover operative for the Central Intelligence Agency whom "The Wall Street Journal" described in 2013 as "the dean of American spy writers"; "The New Republic" magazine calls him "poet laureate of the CIA."; and Otto Penzler says he has produced some "poetic masterpieces". William Zinsser calls him a "political novelist:" Jonathan Yardley, Pulitzer Prize-winning critic for the "Washington Post", calls him a "'serious' novelist" whose work may include "the best novel ever written about life in high-stakes Washington, DC." P.J. O'Rourke called him "the best modern writer on the subject of intrigue." O'Rourke learned about McCarry from a working covert operative who called McCarry "very realistic." His family came from The Berkshires area of western Massachusetts, McCarry was born in Pittsfield, and he lived in Virginia. McCarry believed that "the best novels are about ordinary things: love, betrayal, death, trust, loneliness, marriage, fatherhood." He also said "if you write a political novel, you're writing what you believe instead of what you know." McCarry said, "the themes of my novels have been ordinary things – love, death, betrayal and the American dream." In a 1988 essay published in the Washington "Post", McCarry wrote, "[I]n 1973 when I turned in the manuscript of my novel "The Tears of Autumn", [the publisher] summoned me to New York and, in his office high above lower Park Avenue, banged the manuscript on his desk. ‘This book is talky, it's slow, and nobody is going to believe a goddamn word of the plot,’ he said. ‘Where's the car chase? Where's the torture scene? Where's the sex? Where's the good Russian? Do you call this a thriller?’ ‘No,’ I said. He didn't hear me." McCarry wrote that: "After I resigned [from the CIA], intending to spend the rest of my life writing fiction and knowing what tricks the mind can play when the gates are thrown wide open, as they are by the act of writing, between the imagination and that part of the brain in which information is stored, I took the precaution of writing a closely remembered narrative of my clandestine experiences. After correcting the manuscript, I burned it. What I kept for my own use was the atmosphere of secret life: How it worked on the five senses and what it did to the heart and mind. All the rest went up in flames, setting me free henceforth to make it all up. In all important matters, such as the creation of characters and the invention of plots, with rare and minor exceptions, that is what I have done. And, as might be expected, when I have been weak enough to use something that really happened as an episode in a novel, it is that piece of scrap, buried in a landfill of the imaginary, readers invariably refuse to believe." Throughout McCarry's fiction are statements and descriptions such as "the average intelligence officer is a sort of latter-day Marcel Proust. He lies abed in a cork-lined room, hoping to profit by secrets that other people slip under the door." Snippets from McCarry's CIA years can be found in his non-fiction writing for newspapers. For example, "In the early days of the cold war, a colleague of mine who had worked his way through college playing the saxophone was recruited by a certain United States intelligence agency. For his first assignment, he was posted to Cambodia and told to find an apartment near the royal palace, open the windows at night, and play “Muskrat Ramble.” Prince Sihanouk, the eccentric ruler of the country, was reputed to be an amateur musician who loved jazz and had his own band. Who knew but what he might hear that plaintive sax in the jungle night and invite the nice young American to come on over and sit in?" McCarry's first published novel came out in 1973, which means he was 43; in contrast, most successful novelists have their first novel published while they are still in their twenties. A critic for "Tin House" magazine approaches McCarry through what the critic calls "the art of the sentence," citing as an example a description in the opening pages of McCarry's "The Secret Lovers": “The sun shone feebly through the overcast, like a lamp covered by a woman’s scarf in a shabby hotel room.” A recurrent statement found in McCarry's non-fiction writing is that events in real life are so strange that if presented as fiction no one would believe them. McCarry's novels were all republished in 2005—which means major critics revisited his work—some of which was more than a third-of-a-century old Rereading "The Tears of Autumn" one critic called it "a perfect spy novel." McCarry said in 1995, "If I had to do it over again, I would have written novels about a pediatrician, anonymously." Why? Because he thinks of himself as a novelist, not a writer of spy fiction. And yet, almost of his novels—including the non-Paul Christopher books—have a strong focus on espionage. McCarry began his writing career in the United States Army as a correspondent for "Stars and Stripes". Afterwards, in the 1950s, serving as a speechwriter in the early Administration of President Dwight D. Eisenhower; a typical McCarry item was the 1953 Labor Day Proclamation, which read, in part, "Free American labor has won for itself the enjoyment of a standard of living unmatched in history. The contemporary world knows no comparison with it. There is only brutal contrast to it. To this, there is no more pitiful and dramatic testimony than the food which this free people has been able to send to feed hundreds of thousands suffering the peculiar torments of the proletarian paradise of Eastern Germany." In the late 1950s, he accepted a post with the CIA, for whom he traveled the globe as a deep cover operative—his son, Nathan McCarry, CEO of Pluribus International Corporation, in 2014 described his father's work for the CIA as "trying for the family." He left the CIA, in 1967, becoming a writer of spy novels McCarry rarely spoke or wrote directly about those years, saying simply, "For a decade at the height of the Cold War, I worked abroad under cover as an intelligence agent." In the mid-to-late 1970s, several books by former CIA operatives helped trigger and fuel what became known as the U.S. Senate Church Committee hearings that resulted in legislation limiting the power and secrecy of the CIA. McCarry had been "outed"—publicly identified as a secret CIA operative—in 1975, but was never called to testify. McCarry was an editor-at-large for "National Geographic" and contributed pieces to "The New York Times", "The Wall Street Journal", "The Washington Post", the "Saturday Evening Post," and other national publications. In as essay published by the "Washington Post", he said that "for a writer in America, going out to dinner is like living as an American in Europe: Total strangers think they can say anything they like to you." Ten of McCarry's novels involve the life story of a fictional character named Paul Christopher, who — in McCarry's telling — grew up in pre-Nazi Germany, and later became an operative for a U.S. government entity that is clearly the Central Intelligence Agency. These books are, in order of publication: Alternately, in chronological order of events depicted: The Paul Christopher novels, together and separately, resemble a Christopher Nolan movie in that time sequences become jumbled; e.g. only as Paul Christopher becomes an old man do readers learn about his parents and childhood. One critic notes that “As far as recurring characters go, one must look to John Updike’s ‘Rabbit’ Angstrom books or Philip Roth's Nathan Zuckerman novels for equivalents in the scope and breadth of what McCarry accomplishes with one character and his movements through the events of the twentieth century.” "It’s tempting to say that Charles McCarry’s "The Tears of Autumn" is the greatest espionage novel ever written by an American, if only because it’s hard to conceive of one that could possibly be better. But since no one can claim to have read every America espionage novel ever written, let’s just say that "The Tears of Autumn" is a perfect spy novel, and that its hero, Paul Christopher, should by all rights be known the world over as the thinking man’s James Bond — and woman’s too."—Brendan Bernard, "The Great American Spy Novel," March 31, 2005', LA Weekly" ""Old Boys" is a large yarn that will make yummy reading between long looks at Nantucket Sound this summer. (And a boffo movie in the right hands.) But it is a tale that travels from the outlandish to the absurd. As long as readers don't expect the taut realism we have come to expect from the man, they'll be fine. If they're looking for vintage McCarry, though, this will produce unhappy campers. The book does not approach his better grownup fiction. It is not in the same league, for example, with "The Miernik Dossier", the small gem that made McCarry's career. Rather, it is something of a "Treasure Island" for lovers of spook fiction, a near-juvenile adventure that entrances adults who know better with fabulous writing. What they do get is a fleeting reprise of McCarry's great creation, Paul Christopher. Christopher, the spy whom many first met in McCarry's bestseller "The Tears of Autumn", is now an opaque older man and an ascetic survivor of a Chinese prison camp."--Sam Allis, "McCarry's thriller 'Old Boys' is a trip past believable," Boston "Globe", July 26, 2004. McCarry's most recent work has been cited for its "postmodern skepticism" and "epistemic aporia"--"literary reconstruction that profess to unmask" state-sponsored secrecy. These judgments are based on the work of Robert Snyder. In 2007, novelist and former presidential speechwriter Patrick Andersson wrote that “a new generation of American spy novelists soon began [in the 1970s] to produce a body of work that has surpassed that of current British writers. The four most important are Charles McCarry, Robert Littell, Daniel Silva, and Alan Furst” --"...the most credible account of President Kennedy’s assassination. You will believe it’s what really happened."--"New York" magazine on "Tears of Autumn"—In "Lucky Bastards", JFK seems to have an illegitimate son. In real life, no such person is known to exist. --In "Tears", Paul Christopher wakes up ten days after the JFK killing and intuitively and instantly knows "who had arranged the death of the President"—the family and followers of recently assassinated South Vietnamese leader Ngo Dinh Diem and his two brothers. In real life, the wife of one of Diem's murdered brothers attracted media attention for predicting the JFK assassination ("Anything that happens in Vietnam will find its equivalent in the United States"), and later telling reporters that JFK had got what he deserved. --McCarry was a top aide to Henry Cabot Lodge, traveling with him as chief speechwriter in 1960, for example, when Lodge was the Republican Party's vice presidential candidate. After losing the election, Lodge served as U.S. Ambassador to South Vietnam under JFK, and in that capacity played an active role in the plot to assassinate Diem—which McCarry used as the center of action in "The Tears of Autumn". Note: In McCarry's fiction, the CIA is called "the Outfit." —Computer algorithms that analyses media content and specify—with accuracy—when a physical war between two countries will break out. "The Better Angels", 1979.—Terrorist suicide bombers appear in "Better Angels" (1979); the "New York Times" first reported this form of terrorism in 1983, when describing the Lebanese civil war. Suicide terrorists use fully loaded passenger planes as weapons in "Better Angels"—which did not occur in real life until September 11, 2001. --Suicide bombers begin to blow themselves up at American iconic sites like the Alamo. The Lincoln Memorial and Rotunda of the U.S. Capitol soon follow.—On June 10. 2004, the "Wall Street Journal" published a review entitled, "He Has Seen The Future: It's in His Work; Charles McCarry's novels keep coming true. And his new book is about the end of the world."—Someone who thinks he is JFK's "love child" becomes President of the U.S..--"Lucky Bastard" --"ARK" (2004) has people equipped with "artificial hornets as their primary defensive weapon;" as of 2017, experts discuss the impending possibility of "drones the size of bumblebees that could be programmed to kill certain people, or certain categories of people, by grabbing their skulls with tiny metal talons and drilling into their head."—Worsening and more frequent earthquakes and severe storms like hurricanes threaten society in "Ark". --Soviet KGB has long-term operative control over a person who becomes President of the U.S.-- "Lucky Bastard"—A thirty-person expedition (15 men and 15 women) lands on Ganymede, one of Jupiter's moons--"The Better Angels", p. 164. After the collapse of the USSR, McCarry turned to China in "The Shanghai Factor", "Old Boys", "Second Sight" and "Last Supper". —Jacob Heilbrunn writes in the "N.Y. Times" (2006): "McCarry never succumbs to a bogus moral equivalence in which Western operatives are as nefarious as their Communist counterparts. He instructs us that the real problem is not so much moral quicksand as incompetent scheming. At a moment when the C.I.A.'s travails are evoking nostalgia for a golden age when it supposedly operated effectively, McCarry offers a useful reminder that such an era never existed." --"The truth, once discovered, is of no use: people deny what they have done, forget what they had believed, and make the same mistakes over and over again."—Paul Christopher is eating dinner with a beautiful young woman in wartime Saigon. They discuss the morality of killing. "So you believe in nothing," she says to him. "I believe in consequences," he responds.--"Tears of Autumn" --"You think the truth will make men free. But it only makes them angry."--"Tears of Autumn". McCarry challenged accepted wisdom about JFK by hypothesizing that Kennedy caused his own death; asserting that efforts to force Richard Nixon from the White House constituted a coup.; and giving Ronald Reagan credit for triggering the collapse of the Soviet Union. Paul Christopher as a central character never is older than his forties—see "Tears of Autumn"—which is the same as McCarry was when he wrote it; after Tears, Christopher spends twenty (unseen) years at hard labor in a remote Chinese prison camp, and then is a minor character in his own rescue. The novelist Alan Furst has written in "The Book of Spies" that "[Graham] Greene, [John] le Carré, [Somerset] Maugham, and McCarry write with a kind of cloaked anger, a belief that the world is a place where political power is maintained by treachery and betrayal..." In its subtitle, Furst's book calls such writing "literary espionage." "Your name is the one key thing that cannot be taken from you," says Jan Scruggs, founder of the Vietnam Veterans Memorial, "It captures who you are and what happens to you throughout your life." In Existentialist literature, lack of a name—or of a full name—symbolizes human aloneness; the hero of Franz Kafka's "The Trial", for example is "Michael K.," and the hero of Ralph Ellison's "Invisible Man" has no name at all. Named characters fill McCarry's last three novels, but each book's hero (in one case, heroine) never has a name; one has a "funny name" that is used for payroll and administrative paperwork, but is fictitious and meaningless. While McCarry's heroes with no name are different people in each story, the Man with No Name in three Clint Eastwood movies from the mid-1960s is the same person. The film "Wrong is Right" (1982), starring Sean Connery, was loosely based on McCarry's novel, "The Better Angels" (1979). McCarry was an admirer of the work of W. Somerset Maugham, especially the stories. He was also an admirer of Richard Condon, author of "The Manchurian Candidate" (1959), "Prizzi's Honor" (1982), and numerous other novels. --Stories include: In March 1981, shortly after taking office, Ronald Reagan was shot; Secretary of State Haig appeared in the White House press room and announced, "I am in charge here!" Otto Penzler, ed.: Note: The fictional Paul Christopher had several books of poetry published before he joined the CIA;
https://en.wikipedia.org/wiki?curid=6569
Cleveland Browns The Cleveland Browns are a professional American football team based in Cleveland. Named after original coach and co-founder Paul Brown, they compete in the National Football League (NFL) as a member club of the American Football Conference (AFC) North division. The Browns play their home games at FirstEnergy Stadium, which opened in 1999, with administrative offices and training facilities in Berea, Ohio. The Browns' official club colors are brown, orange, and white. They are unique among the 32 member franchises of the NFL in that they do not have a logo on their helmets. The franchise was founded in 1945 by Brown and businessman Arthur B. McBride as a charter member of the All-America Football Conference (AAFC). The Browns dominated the AAFC, compiling a 47–4–3 record in the league's four seasons and winning its championship in each. When the AAFC folded after the 1949 season, the Browns joined the NFL along with the San Francisco 49ers and the original Baltimore Colts. The team won a championship in their inaugural NFL season, as well as in the 1954, 1955, and 1964 seasons, and in a feat unequaled in any of the North American major professional sports, played in their league championship game in each of the Browns' first ten years of existence. From 1965 to 1995, they qualified to play in the NFL playoffs 14 times, but did not win another championship or play in the Super Bowl during that period. In 1995, owner Art Modell, who had purchased the Browns in 1961, announced plans to move the team to Baltimore. After threats of legal action from the city of Cleveland and fans, a compromise was reached in early 1996 that allowed Modell to establish the Baltimore Ravens as a new franchise while retaining the contracts of all Browns personnel. The Browns' intellectual property, including team name, logos, training facility, and history, were kept in trust and the franchise was regarded by the NFL as suspended, with a new team to be established by 1999 either by expansion or relocation. The Browns were announced as an expansion team in 1998 and resumed play in 1999. Since resuming operations in 1999, the Browns have struggled to find success. They have had only two winning seasons (in 2002 and 2007), one playoff appearance (2002), and no playoff wins, winning only approximately a third of their games in total. The franchise has also been noted for a lack of stability with head coaches (10 full time and two interim since 1999), and quarterbacks (30 different starters since 1999), along with the longest active playoff drought in the NFL at 17 seasons. The history of the Cleveland Browns American football team began in 1944 when taxi-cab magnate Arthur B. "Mickey" McBride secured a Cleveland franchise in the newly formed All-America Football Conference (AAFC). Paul Brown was the team's namesake and first coach. The Browns began play in 1946 in the AAFC. The Browns won each of the league's four championship games before the league dissolved in 1949. The team then moved to the more established National Football League (NFL), where it continued to dominate. Between 1950 and 1955, Cleveland reached the NFL championship game every year, winning three times. McBride and his partners sold the team to a group of Cleveland businessmen in 1953 for a then-unheard-of $600,000. Eight years later, the team was sold again, this time to a group led by New York advertising executive Art Modell. Modell fired Brown before the 1963 season, but the team continued to win behind running back Jim Brown. The Browns won the championship in 1964 and reached the title game the following season, losing to the Green Bay Packers. When the AFL and NFL merged before the 1970 season, Cleveland became part of the new American Football Conference (AFC). While the Browns made it back to the playoffs in 1971 and 1972, they fell into mediocrity through the mid-1970s. A revival of sorts took place in 1979 and 1980, when quarterback Brian Sipe engineered a series of last-minute wins and the Browns came to be called the "Kardiac Kids". Under Sipe, however, the Browns did not make it past the first round of the playoffs. Quarterback Bernie Kosar, whom the Browns drafted in 1985, led the team to three AFC Championship games in the late 1980s but lost each time to the Denver Broncos. In 1995, Modell announced he was relocating the Browns to Baltimore, sowing a mix of outrage and bitterness among Cleveland's dedicated fan base. Negotiations and legal battles led to an agreement where Modell would be allowed to take his personnel to Baltimore as an expansion franchise, called the Baltimore Ravens, but would leave Cleveland the Browns' colors, logos and heritage for a reactivated Browns franchise that would take the field no later than 1999. After three years of inactivity while Cleveland Stadium was demolished and FirstEnergy Stadium was built on its site, the Browns was reactivated and started play again in 1999 under new owner Al Lerner. The Browns struggled throughout the 2000s and 2010s, posting a record of 101–234–1 () since their 1999 return. The Browns have only posted two winning seasons and one playoff appearance (2002) since returning to the NFL. The team's struggles have been magnified since 2012, when the Lerner family sold the team to businessman Jimmy Haslam. In six seasons under Haslam’s ownership, the Browns went through four head coaches and four general managers, none of whom had found success. In 2016 and 2017 under head coach Hue Jackson, the Browns went 1–31 () (including a winless 0–16 season in 2017), the worst two-year stretch in NFL history, and received the number one overall draft pick in both of those years. The Browns are the only National Football League team without a helmet logo. The logoless helmet serves as the Browns' official logo. The organization has used several promotional logos throughout the years; players' numbers were painted on the helmets from 1957 to 1960; and an unused "CB" logo was created in 1965, But for much of their history, the Browns' helmets have been an unadorned burnt orange color with a top stripe of dark brown (officially called "seal brown") divided by a white stripe. The team has had various promotional logos throughout the years, such as the "Brownie Elf" mascot or a Brown "B" in a white football. While Art Modell did away with the Brownie Elf in the mid-1960s, believing it to be too childish, its use has been revived under the current ownership. The popularity of the Dawg Pound section at First Energy Stadium has led to a brown and orange dog being used for various Browns functions. But overall, the orange, logo-less helmet continues as the primary trademark of the Cleveland Browns. On February 24, 2015, the team unveiled its new logos and word marks, the only differences being minor color changes to the helmet with the helmet design remaining largely as is. The original designs of the jerseys, pants, and socks remained mostly the same, but the helmets went through many significant revisions throughout the years. The Browns uniforms saw their first massive change prior to the 2015 season. Jerseys: Pants: Socks: Helmet: Solid white (1946–1949); solid white for day games and solid orange for night games (1950–1951); orange with a single white stripe (1952–1956); orange with a single white stripe and brown numerals on the sides (1957–1959); orange with a brown-white-brown stripe sequence and brown numerals on the sides (1960); orange with a brown-white-brown stripe sequence (1961–1995 and 1999–present). Over the years, the Browns have had on-and-off periods of wearing white for their home games, particularly in the 1970s and 80s, as well as in the early 2000s after the team returned to the league. Until recently, when more NFL teams have started to wear white at home at least once a season, the Browns were the only non-subtropical team north of the Mason-Dixon line to wear white at home on a regular basis. Secondary numerals (called "TV numbers") first appeared on the jersey sleeves in 1961. Over the years, there have been minor revisions to the sleeve stripes, the first occurring in 1968 (brown jerseys worn in early season) and 1969 (white and brown jerseys) when stripes began to be silk screened onto the sleeves and separated from each other to prevent color bleeding. However, the basic five-stripe sequence has remained intact (with the exception of the 1984 season). A recent revision was the addition of the initials "AL" to honor team owner Al Lerner who died in 2002; this was removed in 2013 upon Jimmy Haslam assuming ownership of the team. Orange pants with a brown-white-brown stripe sequence were worn from 1975 to 1983 and become symbolic of the "Kardiac Kids" era. The orange pants were worn again occasionally in 2003 and 2004. Other than the helmet, the uniform was completely redesigned for the 1984 season. New striping patterns appeared on the white jerseys, brown jerseys and pants. Solid brown socks were worn with brown jerseys and solid orange socks were worn with white jerseys. Brown numerals on the white jerseys were outlined in orange. White numerals on the brown jerseys were double outlined in brown and orange. (Orange numerals double outlined in brown and white appeared briefly on the brown jerseys in one pre-season game.) However, this particular uniform set was not popular with the fans, and in 1985 the uniform was returned to a look similar to the original design. It remained that way until 1995. In 1999, the expansion Browns adopted the traditional design with two exceptions: first, the TV numbers, previously on the sleeves, were moved to the shoulders; and second, the orange-brown-orange pants stripes were significantly widened. Experimentation with the uniform design began in 2002. An alternate orange jersey was introduced that season as the NFL encouraged teams to adopt a third jersey, and a major design change was made when solid brown socks appeared for the first time since 1984 and were used with white, brown and orange jerseys. Other than 1984, striped socks (matching the jersey stripes) had been a signature design element in the team's traditional uniform. The white striped socks appeared occasionally with the white jerseys in 2003–2005 and 2007. Experimentation continued in 2003 and 2004 when the traditional orange-brown-orange stripes on the white pants were replaced by two variations of a brown-orange-brown sequence, one in which the stripes were joined (worn with white jerseys) and the other in which they were separated by white (worn with brown jerseys). The joined sequence was used exclusively with both jerseys in 2005. In 2006, the traditional orange-brown-orange sequence returned. Additionally in 2006, the team reverted to an older uniform style, featuring gray face masks; the original stripe pattern on the brown jersey sleeves (The white jersey has had that sleeve stripe pattern on a consistent basis since the 1985 season.) and the older, darker shade of brown. The Browns wore brown pants for the first time in team history on August 18, 2008, preseason game against the New York Giants. The pants contain no stripes or markings. The team had the brown pants created as an option for their away uniform when they integrated the gray facemask in 2006. They were not worn again until the Browns "family" scrimmage on August 9, 2009 with white-striped socks. The Browns have continued to wear the brown pants throughout the 2009 season. Browns quarterback Brady Quinn supported the team's move to wearing the brown pants full-time, claiming that the striped pattern on the white pants "prohibit[ed] mobility". However, the fans generally did not like the brown pants, and after being used for only one season, the team returned to their white shirt-on-white pants in 2010. Coach Eric Mangini told "The Plain Dealer" the Browns won't use the brown pants anymore. "It wasn't very well-received," Mangini said. "I hope we can get to the point where we can wear fruit on our heads and people wouldn't notice." At the time, the brown pants weren't officially dropped by the team, but simply not used. The Browns chose to wear white at home for the 2011 season, and wound up wearing white for all 16 games as when they were on the road, the home team would wear their darker colored uniform. The Browns brought back the brown pants in their home game against the Buffalo Bills on October 3, 2013 on "Thursday Night Football", pairing them with the brown jerseys. It marked the first time the team wore an all-brown combination in team history. On April 14, 2015, the Cleveland Browns unveiled their new uniform combinations, consisting of the team's colors of orange, brown and white. The Browns brought back the all-brown look for the NFL Color Rush program in 2016, minus the white elements. In 2018 the uniform was worn at home three times. For the 2019 season, the Browns promoted the Color Rush uniform to their primary home uniform and will wear it for six home games as well as any away game in which the home team chooses to wear white (although for the Week 3 game against the Los Angeles Rams, the Browns wore all-orange socks instead of brown. For the two remaining home games, the browns will wear their previous brown home uniforms as alternate uniforms. The club unveiled a new uniform design on April 15, 2020. The new uniform design pays homage to the Browns' classic uniform design from years past. The Cleveland Browns have rivalries with all three of its AFC North opponents. In addition, the team has had historical rivalries with the Indianapolis Colts, Denver Broncos, Buffalo Bills, and Detroit Lions. The team's biggest rival in the AAFC was the San Francisco 49ers, though this has cooled and in some cases turned into a friendly relationship, as the Browns now play in the AFC and the 49ers play in the NFC. Additionally, many 49ers personnel helped the Browns relaunch in 1999 as well as former team President Mike Holmgren having started his NFL career in San Francisco. Also, 49ers owners John York and Denise DeBartolo York reside in Youngstown, southeast of Cleveland. Former long-time veteran placekicker and fan favorite, Phil Dawson, signed with the 49ers in 2014, along with backup quarterback Colt McCoy. Often called the "Turnpike Rivalry", the Browns' biggest rival has long been the Pittsburgh Steelers. Former Browns owner Art Modell scheduled home games against the Steelers on Saturday nights from 1964 to 1970 to help fuel the rivalry. The rivalry has also been fueled by the proximity of the two teams, number of championships both teams have won, players and personnel having played and/or coached for both sides, and personal bitterness. The teams have played twice annually since 1950, making it the oldest rivalry in the AFC and the fifth-oldest rivalry in the NFL. Though the Browns dominated this rivalry early in the series (winning the first eight meetings and posting a 31–9 record in the 1950s and 1960s), the Steelers went 15–5 in the 1970s and 34–6–1 since the Browns returned to the league in 1999. The Steelers have been particularly dominant in Pittsburgh, posting a 42–6 record when hosting the Browns since 1970, including to winning streaks of 16 games (1970–85, 2004–present). The Steelers currently hold a 76–59–1 lead. The Browns and Steelers met in the playoffs in and , with the Steelers winning both meetings. Though the rivalry has cooled in Pittsburgh due to the Modell move as well as the Browns' poor play since 1999, the Steelers still remain the top rival for Cleveland. Originally conceived due to the personal animosity between Paul Brown and Art Modell, the "Battle of Ohio" between the Browns and the Cincinnati Bengals has been fueled by the sociocultural differences between Cincinnati and Cleveland, a shared history between the two teams, and similar team colors, as Brown used the exact shade of orange for the Bengals that he used for the Browns. (Though this has changed since then, as the Bengals now use a brighter shade of orange.) Modell, in fact, moved the Browns to the AFC after the AFL–NFL merger in order to have a rivalry with the Bengals. The rivalry has also produced two of the eight highest-scoring games in NFL history. Cincinnati has the all-time edge 51–42. While the Bengals have a 27–15 edge since the Browns returned to the NFL in 1999, the Browns-Bengals series has been more competitive than the Browns' series with their other division rivals. Created as a result of the Browns' relocation to Baltimore, the rivalry between the Browns and Baltimore Ravens was more directed at Art Modell than the team itself, and is simply considered a divisional game in Baltimore. This matchup is more bitter for Cleveland than the others due to the fact that the draft picks for 1995 to 1998 resulted in the rosters that won the Super Bowl for the Ravens in 2000. Had the Browns stayed in Cleveland, these teams (drafted by general manager Ozzie Newsome) might have given the Browns the title after a 35-year drought. This bitterness was compounded when the Ravens won their second Super Bowl in 2012. The Ravens lead the overall series 31–11. The two teams have not met in the playoffs. The Detroit Lions rivalry began in the 1950s, when the Browns and Lions played each other in four NFL championships. The Lions won three of those championships, while the Browns won one. This was arguably one of the NFL's best rivalries in the 1950s. Since the NFL-AFL merger of 1970, the teams have met much less frequently with the Browns' move to the AFC. From 2002 to 2014, the two teams played an annual preseason game known as the "Great Lakes Classic". The Bills rivalry had its roots back to the days of the AAFC, when there was a team from Buffalo with the same name in that league. The Browns and AAFC Bills played six games, including a league championship game, before the Browns were selected to merge into the NFL and the Bills left out. After the current incarnation of the Bills joined the NFL, the Browns and Bills have played each other from time to time. Though the Browns and Bills are in different AFC divisions, a mellow rivalry defined by mutual respect has since developed between the teams due to the similarities between Buffalo and Cleveland and shared misfortune between the teams. Despite this "rivalry" being known for ugly games, such as an 8–0 Browns win played in a blizzard in 2007 and a 6–3 Browns win in 2009 in which Browns quarterback Derek Anderson only completed 2 of 17 passes, there have been some competitive moments between the Bills and Browns as well, such as a playoff game in 1990 and two games with playoff-implications in 2007 and 2014. The Browns fans' respect for the Bills is partly due to Bills founding owner Ralph Wilson being one of only two NFL owners to vote against the decision to move the original Browns team to Baltimore. The Browns had a brief rivalry with the Denver Broncos that arose from three AFC Championship Games from 1986 to 1989. In the 1986 AFC Championship, quarterback John Elway led The Drive to secure a tie in the waning moments at Cleveland Municipal Stadium; the Broncos went on to win in 23–20 in overtime. One year later, the two teams met again in the 1987 AFC Championship game at Mile High Stadium. Denver took a 21–3 lead, but Browns' quarterback Bernie Kosar threw four touchdown passes to tie the game at 31–31 halfway through the 4th quarter. After a long drive, John Elway threw a 20-yard touchdown pass to running back Sammy Winder to give Denver a 38–31 lead. Cleveland advanced to Denver's 8-yard line with 1:12 left, but Broncos' safety Jeremiah Castille stripped Browns' running back Earnest Byner of the football at the 2-yard line—a play that has been called The Fumble by Browns' fans. The Broncos recovered it, gave Cleveland an intentional safety, and went on to win 38–33. The two teams met yet again in the 1989 AFC Championship at Mile High Stadium, which the Broncos easily won by a score of 37–21. A 2006 study conducted by "Bizjournal" determined that Browns fans are the most loyal fans in the NFL. The study, while not scientific, was largely based on fan loyalty during winning and losing seasons, attendance at games, and challenges confronting fans (such as inclement weather or long-term poor performance of their team). The study noted that Browns fans filled 99.8% of the seats at Cleveland Browns Stadium during the last seven seasons, despite a combined record of 36–76 over that span. Perhaps the most visible Browns fans are those that can be found in the Dawg Pound. Originally the name for the bleacher section located in the open (east) end of old Cleveland Municipal Stadium, the current incarnation is likewise located in the east end of FirstEnergy Stadium and still features hundreds of orange and brown clad fans sporting various canine-related paraphernalia. The fans adopted that name in 1984 after members of the Browns defense used it to describe the team's defense. Retired cornerback Hanford Dixon, who played his entire career for the Browns (1981–1989), is credited with naming the Cleveland Browns defense 'The Dawgs' in the mid-1980s. Dixon and teammates Frank Minnifield and Eddie Johnson would bark at each other and to the fans in the bleachers at the Cleveland Stadium to fire them up. It was from Dixon's naming that the "Dawg Pound" subsequently took its title. The fans adopted that name in the years after. Due to this nickname, since the team's revival the Browns have used a bulldog as an alternate logo. The most prominent organization of Browns fans is the "Browns Backers Worldwide" (BBW). The organization has approximately 305,000 members and Browns Backers clubs can be found in every major city in the United States, and in a number of military bases throughout the world, with the largest club being in Phoenix, Arizona. In addition, the organization has a sizable foreign presence in places as far away as Egypt, Australia, Japan, Sri Lanka, and McMurdo Station in Antarctica. According to The Official Fan Club of the Cleveland Browns, the two largest international fan clubs are in Alon Shvut, West Bank and Niagara, Canada, with Alon Shvut having 129 members and Niagara having 310. Following former Browns owner Randy Lerner's acquisition of English soccer club Aston Villa, official Villa outlets started selling Cleveland Browns goods such as jerseys and NFL footballs. This has raised interest in England and strengthened the link between the two sporting clubs. Aston Villa supporters have set up an organization known as the Aston (Villa) Browns Backers of Birmingham. The Cleveland Browns were the favorite team of Elvis Presley. This was because his friend Gene Hickerson – with whom he had played football in their common youth in Memphis – was contracted by the Browns in 1957 and played there during his entire career until 1973. Also defender Bobby Franklin, who had played from 1960 to 1966 for the Browns, was a friend of Presley. WWE Hall of Fame wrestler and commentator Jerry "The King" Lawler – though he has spent most of his life in Memphis – spent part of his childhood in the Cleveland area and is a fan of the Browns. Fellow WWE wrestlers The Miz and Dolph Ziggler (both Cleveland natives) are also fans. Another fan of the team is baseball legend Hank Aaron. Other famous Browns fans include LeBron James, Arsenio Hall, Drew Carey, Patricia Heaton (her father, Chuck Heaton, was a sportswriter for "The Plain Dealer", which covered the Browns and wrote two books about the team), Terri Garr, Martin Mull, Condoleezza Rice, Two-time UFC Heavyweight Champion Stipe Miocic, Valerie Bertinelli (her husband is from the Northeast Ohio area, and she starred in "Hot in Cleveland"), Machine Gun Kelly, Paul Adelstein, Iron Chef, Tom Holland, Michael Symon, C. J. McCollum, and Brad Paisley. The Cleveland Browns have the fourth largest number of players enshrined in the Pro Football Hall of Fame with a total of 17 enshrined players elected based on their performance with the Browns, and nine more players or coaches elected who spent at least one year with the Browns franchise. No Browns players were inducted in the inaugural induction class of 1963. Otto Graham was the first Browns player to be enshrined as a member of the class of 1965, and the most recent Browns player to be included in the Pro Football Hall of Fame is Mac Speedie, who was a member of the class of 2020. All of the Browns' Pro Football Hall of Fame inductees thus far have been from the pre-1996 incarnation; no members of the Hall of Fame played for the Browns after 1999. The Cleveland Browns legends program honors former Browns who made noteworthy contributions to the history of the franchise. In addition to all the Hall of Famers listed above, the Legends list includes: Beginning in 2010, the Browns established a Ring of Honor, honoring the greats from the past by having their names displayed around the upper deck of FirstEnergy Stadium. The inaugural class in the Browns Ring of Honor was unveiled during the home opener on September 19, 2010, and featured the 16 Hall of Famers listed above who went into the Hall of Fame as Browns. In 2018, Joe Thomas was entered into the Ring of Honor with the number 10,363 – commemorating his NFL record of consecutive snaps played on offense. In 2019, four-time Pro Bowl linebacker Clay Matthews was entered into the Ring of Honor. Numerous Browns players and staff have had statues made in their honor: In and around First Energy Stadium In and around Cleveland WKNR (850 AM), WKRK-FM (92.3 FM), and WNCX (98.5 FM) serve as co-flagship stations for the Cleveland Browns Radio Network. Play-by-play announcer Jim Donovan calls games on-site alongside color analyst Doug Dieken, a former Browns left tackle, and sideline reporter Nathan Zegura – who made news when he had to serve an eight-game suspension due to arguing with officials during a game in 2018. WKRK-FM personality Ken Carman and WKNR personalities Tony Rizzo & Je'Rod Cherry host the network pregame show, while WKRK-FM personalities Jeff Phelps and Dustin Fox host the network postgame show. WEWS-TV (TV channel 5) serves as the broadcast TV home of the Browns, airing year-round team programming as well as non-network preseason games. Jay Crawford serves as play-by-play announcer, former Browns quarterback Tim Couch serves as color analyst, and former Browns quarterback Bernie Kosar and Dustin Fox serve as sideline analysts. SportsTime Ohio (STO) is the cable outlet for the team, airing various Browns related programming during the season, STO had previously served as the team's cable outlet from its founding in 2006 until 2014. The Browns in-house production team won a pair of Lower Great Lakes Emmy Awards in 2005. One was for a primetime special honoring the 1964 NFL Championship team ("The 1964 Championship Show") and one was for a commercial spot ("The Paperboy"). The Browns have (either directly or indirectly) been featured in various movies and TV shows over the years. Notable examples include:
https://en.wikipedia.org/wiki?curid=6576
Carbine A carbine ( or ) is a long gun firearm but with a shorter barrel than a standard rifle or musket. Many carbines are shortened versions of full-length rifles, shooting the same ammunition, while others fire lower-powered ammunition, including types designed for pistols. The smaller size and lighter weight of carbines make them easier to handle. They are typically issued to high-mobility troops such as special-operations soldiers and paratroopers, as well as to mounted, artillery, logistics, or other non-infantry personnel whose roles do not require full-sized rifles, although there is a growing tendency for carbines to be issued to front-line soldiers to offset the increasing weight of other issued equipment. An example of this is the US Army's M4 carbine, which is standard issue. The name comes from its first users — cavalry troopers called "carabiniers", from the French "carabine", from Old French "carabin" (soldier armed with a musket), whose origin is unclear. One theory connects it to an "ancient engine of war" called a "calabre"; another connects it to Medieval Latin "Calabrinus" 'Calabrian'; yet another, "less likely", to "escarrabin", gravedigger, from the scarab beetle. The carbine was originally developed for cavalry. The start of early modern warfare about the 16th century had infantry armed with firearms, prompting cavalry to do the same, even though reloading muzzle loading firearms while moving mounted was highly impractical. Some cavalry, such as the German Reiters, added one or more pistols, while other cavalry, such as harquebusiers, tried various shorter, lightened versions of the infantry arquebus weapons – the first carbines. But these weapons were still difficult to reload while mounted, and the saber often remained main weapon of such cavalry. Dragoons and other mounted infantry that dismounted for battles usually adopted standard infantry firearms, though some favored versions that were less encumbering when riding – something that could be arranged to hang clear of the rider's elbows and horse's legs. While more portable, carbines had the general disadvantages of less accuracy and power than the longer guns of the infantry. During Napoleonic warfare, pistol and carbine-armed cavalry generally transitioned into traditional melee cavalry or dragoons. Carbines found increased use outside of standard cavalry and infantry, such as support and artillery troops, who might need to defend themselves from attack but would be hindered by keeping full-sized weapons with them continuously; a common title for many short rifles in the late 19th century was "artillery carbine". As the rifled musket replaced the smoothbore firearms for infantry in the mid 19th century, carbine versions were also developed; this was often developed separately from the infantry rifles and, in many cases, did not even use the same ammunition, which made for supply difficulties. A notable weapon developed towards the end of the American Civil War by the Union was the Spencer carbine, one of the very first breechloading, repeating weapons. It had a spring-powered, removable tube magazine in the buttstock which held seven rounds and could be reloaded by inserting spare tubes. It was intended to give the cavalry a replacement weapon which could be fired from horseback without the need for awkward reloading after each shot – although it saw service mostly with dismounted troopers, as was typical of cavalry weapons during that war. In the late 19th century, it became common for a number of nations to make bolt-action rifles in both full-length and carbine versions. One of the most popular and recognizable carbines were the lever-action Winchester carbines, with several versions available firing revolver cartridges. This made it an ideal choice for cowboys and explorers, as well as other inhabitants of the American West, who could carry a revolver and a carbine, both using the same ammunition. The Lee Enfield Cavalry Carbine (LEC) a shortened version of the standard British Army infantry rifle was introduced in 1896, although it did not become the standard British cavalry weapon until 1903. In the decades following World War I, the standard battle rifle used by armies around the world had been growing shorter, either by redesign or by the general issue of carbine versions instead of full-length rifles. This move was initiated by the US Model 1903 Springfield, which was originally produced in 1907 with a short 24-inch barrel, providing a short rifle that was longer than a carbine but shorter than a typical rifle, so it could be issued to all troops without need for separate versions. Other nations followed suit after World War I, when they learned that their traditional long-barreled rifles provided little benefit in the trenches and merely proved a hindrance to the soldiers. Examples include the Russian Model 1891 rifle, originally with an barrel, later shortened to in 1930, and to in 1938, the German Mauser Gewehr 98 rifles went from in 1898 to in 1935 as the "Karabiner 98k" (K98k or Kar98k), or "short carbine". The barrel lengths in rifles used by the United States did not change between the bolt-action M1903 rifle of World War I and the World War II M1 Garand rifle, because the barrel on the M1903 was still shorter than even the shortened versions of the Model 1891 and Gewehr 98. The US M1 carbine was more of a traditional carbine in that it was significantly shorter and lighter, with a barrel, than the M1 Garand rifle, and that it was intended for rear-area troops who couldn't be hindered with full-sized rifles but needed something more powerful and accurate than a Model 1911 pistol (although this didn't stop soldiers from using them on the front line). Contrary to popular belief, and even what some books claim, in spite of both being designated "M1", the M1 Carbine was "not" a shorter version of the .30-06 M1 Garand, as is typical for most rifles and carbines, but a wholly different design firing a smaller, less-powerful cartridge. The "M1" designates each as the first model in the new US designation system, which no longer used the year of introduction, but a sequential series of numbers starting at "1": the M1 "Carbine" and M1 "Rifle". The United Kingdom also developed a "Jungle Carbine" version of their Lee–Enfield service rifle, featuring a shorter barrel, flash suppressor, and manufacturing modifications designed to decrease the rifle's weight. Officially titled "Rifle, No. 5 Mk I", it was introduced in the closing months of World War II, but it did not see widespread service until the Korean War, the Mau Mau Uprising, and the Malayan Emergency as well the Vietnam War. A shorter weapon was more convenient when riding in a truck, armored personnel carrier, helicopter, or aircraft, and also when engaged in close-range combat. Based on the combat experience of World War II, the criteria used for selecting infantry weapons began to change. Unlike previous wars, which were often fought mainly from fixed lines and trenches, World War II was a highly mobile war, often fought in cities, forests, or other areas where mobility and visibility were restricted. In addition, improvements in artillery made moving infantry in open areas even less practical than it had been. The majority of enemy contacts were at ranges of less than , and the enemy was exposed to fire for only short periods of time as they moved from cover to cover. Most rounds fired were not aimed at an enemy combatant, but instead fired in the enemy's direction to keep them from moving and firing back (see suppressive fire). These situations did not require a heavy rifle, firing full-power rifle bullets with long-range accuracy. A less-powerful weapon would still produce casualties at the shorter ranges encountered in actual combat, and the reduced recoil would allow more shots to be fired in the short amount of time an enemy was visible. The lower-powered round would also weigh less, allowing a soldier to carry more ammunition. With no need of a long barrel to fire full-power ammunition, a shorter barrel could be used. A shorter barrel made the weapon weigh less, was easier to handle in tight spaces, and was easier to shoulder quickly to fire a shot at an unexpected target. Full-automatic fire was also considered a desirable feature, allowing the soldier to fire short bursts of three to five rounds, increasing the probability of a hit on a moving target. The Germans had experimented with selective-fire carbines firing rifle cartridges during the early years of World War II. These were determined to be less than ideal, as the recoil of full-power rifle cartridges caused the weapon to be uncontrollable in full-automatic fire. They then developed an intermediate-power cartridge round, which was accomplished by reducing the power and the length of the standard 7.92×57mm Mauser rifle cartridge to create the 7.92×33mm "Kurz" (Short) cartridge. A selective-fire weapon was developed to fire this shorter cartridge, eventually resulting in the Sturmgewehr 44, later translated as "assault rifle" (also frequently called "machine carbines" by Allied intelligence, a quite accurate assessment, in fact). Very shortly after World War II, the USSR would adopt a similar weapon, the ubiquitous AK-47, the first model in the famed Kalashnikov-series, which became the standard Soviet infantry weapon, and which has been produced and exported in extremely large numbers up through the present day. Although the United States had developed the M2 Carbine, a selective-fire version of the M1 Carbine during WW2, the .30 Carbine cartridge was closer to a pistol round in power, making it more of a submachine gun than an assault rifle. It was also adopted only in very small numbers and issued to few troops (the semi-automatic M1 carbine was produced in a 10-to-1 ratio to the M2), while the AK47 was produced by the millions and was standard-issue to all Soviet troops, as well as those of many other nations. The US was slow to follow suit, insisting on retaining a full-power, 7.62×51mm NATO rifle, the M14 (although this "was" selective fire), until too-hastily adopting the 5.56mm M16 rifle in the mid-1960s, with initially poor results due to the rapidity of its introduction (but later to become a highly successful line of rifles and carbines). In the 1950s, the British developed the .280 British, an intermediate cartridge, and a select-fire bullpup assault rifle to fire it, the EM-2. They pressed for the US to adopt it so it could become a NATO-standard round, but the US insisted on retaining a full-power, .30 caliber round. This forced NATO to adopt the 7.62×51mm NATO round (which in reality is only slightly different ballistically to the .308 Winchester), to maintain commonality. The British eventually adopted the 7.62mm FN FAL, and the US adopted the 7.62mm M14. These rifles are both what is known as "battle rifles" and were a few inches shorter than the standard-issue rifles they replaced (22" barrel as opposed to 24" for the M1 Garand), although they were still full-powered rifles, with selective fire capability. These can be compared to the even shorter, less-powerful assault rifle, which might be considered the "carbine branch of weapons development", although indeed, there are now carbine variants of many of the assault rifles which had themselves seemed quite small and light when adopted. By the 1960s, after becoming involved in War in Vietnam, the US did an abrupt about-face and decided to standardize on the intermediate 5.56×45mm round (based on the .223 Remington varmint cartridge) fired from the new, lightweight M16 rifle, leaving NATO to hurry and catch up. Many of the NATO countries couldn't afford to re-equip so soon after the recent 7.62mm standardization, leaving them armed with full-power 7.62mm battle rifles for some decades afterwards, although by this point, the 5.56mm has been adopted by almost all NATO countries and many non-NATO nations as well. This 5.56mm NATO round was even lighter and smaller than the Soviet 7.62×39mm AK-47 cartridge, but possessed higher velocity. In U.S. service, the M16 assault rifle replaced the M14 as the standard infantry weapon, although the M14 continued to be used by designated marksmen. Although at 20", the barrel of the M16 was shorter than that of the M14, it was still designated a "rifle" rather than a "carbine", and it was still longer than the AK, which used a 16" barrel. (The SKS – an interim, semi-automatic, weapon adopted a few years before the AK-47 was put into service – was designated a carbine, even though it's 20" barrel was significantly longer than the AK series' 16.3". This is because of the Kalashnikov's revolutionary nature, which altered the old paradigm. Compared to previous rifles, particularly the Soviets' initial attempts at semi-automatic rifles, such as the 24" SVT-40, the SKS was significantly shorter. The Kalashnikov altered traditional notions and ushered in a change in what was considered a "rifle" in military circles.) In 1974, shortly after the introduction of the 5.56mm NATO, the USSR began to issue a new Kalashnikov variant, the AK-74, chambered in the small-bore 5.45×39mm cartridge, which was a standard 7.62×39mm necked down to take a smaller, lighter, faster bullet. It soon became standard issue in Soviet nations, although many of the nations with export Kalashnikovs retained the larger 7.62×39mm round. In 1995, the People's Republic of China adopted a new 5.8×42mm cartridge to match the modern trend in military ammunition, replacing the previous 7.62×39mm and 5.45×39mm round as standard. Later, even lighter carbines variants of many of these short-barreled assault rifles came to be adopted as the standard infantry weapon. In much modern tactical thinking, only a certain number of soldiers now need to retain longer-range weapons, these serving as designated marksmen. The rest can carry lighter, shorter-ranged weapons for close-quarters combat and suppressive fire. This is basically a more extreme extension of the idea that brought the original assault rifle. Another factor is that with the increasing weight of technology, sighting systems, ballistic armor, etc., the only way to reduce the burden on the modern soldier was to equip them with a smaller, lighter weapon. Also, modern soldiers rely a great deal on vehicles and helicopters to transport them around the battle area, and a longer weapon can be a serious hindrance to entering and exiting these vehicles. Development of lighter assault rifles continued, matched by developments in even lighter carbines. In spite of the short barrels of the new assault rifles, carbines variants like the 5.45×39mm AKS-74U and Colt Commando were being developed for use when mobility was essential and a submachine gun wasn't sufficiently powerful. The AKS-74U featured an extremely short 8.1" barrel which necessitated redesigning and shortening the gas-piston and integrating front sights onto the gas tube; the Colt Commando was a bit longer, at 11.5". Neither was adopted as standard issue, although the US did later adopt the somewhat-longer M4 carbine, with a 14.5" barrel. By the 1990s, the US had adopted the M4 carbine, a derivative of the M16 family which fired the same 5.56mm cartridge but was lighter and shorter (in overall length and barrel length), resulting in marginally reduced range and power, although offering better mobility and lighter weight to offset the weight of equipment and armor that a modern soldier has to carry. However, in spite of the benefits of the modern carbine, many armies are experiencing a certain backlash against the universal equipping of soldiers with carbines and lighter rifles in general, and are equipping selected soldiers, usually called Designated Marksmen, or DM, with higher-power rifles. Another problem comes from the loss of muzzle velocity caused by the shorter barrel, which when coupled with the typical small, lightweight bullets, causes effectiveness to be diminished; a 5.56mm gets its lethality from its high velocity, and when fired from the 14.5" M4 carbine, its power, penetration, and range are diminished. Thus, there has been a move towards adopting a slightly more powerful round tailored for high performance from both long and short barrels. The US has done experiments regarding adopting a new, slightly larger and heavier caliber such as the 6.5mm Grendel or 6.8mm Remington SPC, which are heavier and thus retain more effectiveness at lower muzzle velocities, but has for the time decided to retain the 5.56mm NATO round as standard issue. While the US Army adopted the M4 carbine in the 1990s, the US Marine Corps retained their 20" barrel M16A4 rifles long afterwards, citing the increased range and effectiveness over the carbine version; officers were required to carry an M4 carbine rather than an M9 pistol, as Army officers do. Due to the Marine Corps emphasis on being riflemen, the lighter carbine was considered a suitable compromise between a rifle and a pistol. Marines with restricted mobility such as vehicle operators, or a greater need for mobility such as squad leaders, were also issued M4 carbines. In July 2015, the Marine Corps approved the M4 carbine for standard issue to front-line Marines, replacing the M16A4 rifle. The rifles will be issued to support troops while the carbines go to the front-line Marines, in a reversal of the traditional roles of "rifles for the front line, carbines for the rear". Special forces need to perform fast, decisive operations, frequently airborne or boat-mounted. A pistol, though light and quick to operate, is viewed as not having enough power, firepower, or range. A submachine gun has selective fire, but firing a pistol cartridge and having a short barrel and sight radius, it is not accurate or powerful enough at longer ranges. Submachine guns also tend to have poorer armor and cover penetration than rifles and carbines firing rifle ammunition. Consequently, carbines have gained wide acceptance among SOCOM, UKSF, and other communities, having relatively light weight, large magazine capacity, selective fire, and much better range and penetration than a submachine gun. The smaller size and relative lighter weight of carbines makes them easier to handle in close-quarter situations such as urban engagements, when deploying from military vehicles, or in any situation where space is confined. The disadvantages of carbines relative to rifles include inferior long-range accuracy and a shorter effective range. These comparisons refer to carbines (short-barreled rifles) of the same power and class as the regular full-sized rifles. Compared to submachine guns, carbines have a greater effective range and are capable of penetrating helmets and body armor when used with armor piercing ammunition. However, submachine guns are still used by military special forces and police SWAT teams for close quarters battle (CQB) because they are "a pistol caliber weapon that's easy to control, and less likely to over-penetrate the target." Also, carbines are harder to maneuver in tight encounters where superior range and stopping power at distance are not great considerations. Firing the same ammunition as standard-issue rifles or pistols gives carbines the advantage of standardization over those personal defense weapons (PDWs) that require proprietary cartridges. The modern usage of the term carbine covers much the same scope as it always had, namely lighter weapons (generally rifles) with barrels up to 20 inches in length. These weapons can be considered carbines, while rifles with barrels longer than 20 inches are generally not considered carbines unless specifically named so. Conversely, many rifles have barrels "shorter" than 20", yet aren't considered carbines. The AK series rifles has an almost universal barrel length of 16.3", well within carbine territory, yet has always been considered a rifle, perhaps because it was designed as such and not shortened from a longer weapon. Modern carbines use ammunition ranging from that used in light pistols up to powerful rifle cartridges, with the usual exception of high-velocity magnum cartridges. In the more powerful cartridges, the short barrel of a carbine has significant disadvantages in velocity, and the high residual pressure, and frequently still-burning powder and gases, when the bullet exits the barrel results in substantially greater muzzle blast. Flash suppressors are a common, partial solution to this problem, although even the best flash suppressors are hard put to deal with the excess flash from the still-burning powder leaving the short barrel (and they also add several inches to the length of the barrel, diminishing the purpose of having a short barrel in the first place). The shorter the barrel, the more difficult it is to hide the flash; the AKS-74U has a complex, effective muzzle-booster/flash suppressor, yet it still suffers from extreme muzzle flash. The typical carbine is the pistol-caliber carbine. These first appeared soon after metallic cartridges became common. These were developed as "companions" to the popular revolvers of the day, firing the same cartridge but allowing more velocity and accuracy than the revolver. These were carried by cowboys, lawmen, and others in the Old West. The classic combination would be a Winchester lever-action carbine and a Colt Single Action Army revolver in .44-40 or .38-40. During the 20th century, this trend continued with more modern and powerful smokeless revolver cartridges, in the form of Winchester and Marlin lever action carbines chambered in .38 Special/.357 Magnum and .44 Special/.44 Magnum. Modern equivalents include the Ruger Police Carbine, which uses the same magazine as the Ruger pistols of the same caliber, and the (discontinued) Marlin Camp Carbine, which, in .45 ACP, used M1911 magazines. The Ruger Model 44 and Ruger Deerfield Carbine were both carbines chambered in .44 Magnum. The Beretta Cx4 Storm shares magazines with many Beretta pistols and is designed to be complementary to the Beretta Px4 Storm pistol. The Hi-Point 995TS are popular, economical and reliable alternatives to other pistol caliber carbines in the United States, and their magazines can be used in the Hi-Point C-9 pistol. Another example is the Kel-Tec SUB-2000 series chambered in either 9mm Luger or .40 S&W, which can be configured to accept Glock, Beretta, S&W, or SIG pistol magazines. The SUB-2000 also has the somewhat unusual (although not unique) ability to fold in half. The primary advantage of a carbine over a pistol using the same ammunition is controllability. The combination of firing from the shoulder, longer sight-radius, 3 points of contact (firing hand, support hand & shoulder), and precision offer a significantly more user-friendly platform. Carbines like the Kel-Tec SUB-2000, Hi Point 995TS and Beretta Cx4 Storm have the ability to mount user friendly optics, lights and lasers thanks to them having accessory rails, which make target acquisition and engagement much easier. The longer barrel can offer increased velocity and, with it, greater energy and effective range due to the propellant having more time to burn. However, loss in bullet velocity can happen where the propellant is utilised before the bullet reaches the muzzle, combined with the friction from the barrel on the bullet. As long guns, pistol-caliber carbines may be less legally restricted than handguns in some jurisdictions. Compared to carbines chambered in intermediate or rifle calibers, such as .223 Remington and 7.62×54mmR, pistol-caliber carbines generally experience less of an increase in external ballistic properties as a result of the propellant. The drawback is that one loses the primary benefits of a handgun, i.e. portability and concealability, resulting in a weapon almost the size of, but less accurate than, a long-gun, but not much more powerful than a pistol. Also widely produced are semi-automatic and typically longer-barreled derivatives of select-fire submachine guns, such as the FN PS90, HK USC, KRISS Vector, Thompson carbine, CZ Scorpion S1 Carbine and the Uzi carbine. In order to be sold legally in many countries, the barrel must meet a minimum length (16" in the USA). So the original submachine gun in given a legal-length barrel and made into a semi-automatic, transforming it into a carbine. Though less common, pistol-caliber conversions of centerfire rifles like the AR-15 are commercially available. Some handguns used to come from the factory with mounting lugs for a shoulder stock, notably including the "Broomhandle" Mauser C96, Luger P.08, and Browning Hi-Power. In the case of the first two, the pistol could come with a hollow wooden stock that doubled as a holster. Carbine conversion kits are commercially available for many other pistols, including M1911 and most Glocks. These can either be simple shoulder stocks fitted to a pistol or full carbine conversion kits, which are at least long and replace the pistol's barrel with one at least long for compliance with the US law. In the US, fitting a shoulder stock to a handgun with a barrel less than 16" long legally turns it into a short-barreled rifle, which is in violation of the National Firearms Act. Under the National Firearms Act of 1934, firearms with shoulder stocks or originally manufactured as a rifle and barrels less than in length are classified as short-barreled rifles. Short-barreled rifles are restricted similarly to short-barreled shotguns, requiring a $200 tax paid prior to manufacture or transfer – a process which can take several months. Because of this, firearms with barrels of less than and a shoulder stock are uncommon. A list of firearms not covered by the NFA due to their antique status may be found here or due to their "Curio and Relic" status may be found here; these lists includes a number of carbines with barrels less than the minimum legal length and firearms that are "primarily collector's items and are not likely to be used as weapons and, therefore, are excluded from the provisions of the National Firearms Act." Machine guns, as their own class of firearm, are not subject to requirements of other class firearms. Distinct from simple shoulder stock kits, full carbine conversion kits are not classified as short-barreled rifles. By replacing the pistol barrel with one at least in length and having an overall length of at least , a carbine converted pistol may be treated as a standard rifle under Title I of the Gun Control Act of 1968 (GCA). However, certain "Broomhandle" Mauser C96, Luger, and Browning Hi-Power Curio & Relic pistols with their originally issued stock attached only may retain their pistol classification. Carbines without a stock and not originally manufactured as a rifle are not classified as rifles or short barreled rifles. A carbine manufactured under in length without a forward vertical grip will be a pistol and, state law notwithstanding, can be carried concealed without creating an unregistered Any Other Weapon. A nearly identical carbine with an overall length of or greater is simply an unclassified firearm under Title I of the Gun Control Act of 1968, as the Any Other Weapon catch-all only applies to firearms under or that have been concealed. However, a modification intending to fire from the shoulder and bypass the regulation of short-barreled rifles is considered the unlawful possession and manufacture of an unregistered short-barreled rifle. In some historical cases, the term "machine carbine" was the official title for submachine guns, such as the British Sten and Australian Owen guns. The semiautomatic-only version of the Sterling submachine gun was also officially called a "carbine". The original Sterling semi-auto would be classed a "short barrel rifle" under the U.S. National Firearms Act, but fully legal long-barrel versions of the Sterling have been made for the U.S. collector market.
https://en.wikipedia.org/wiki?curid=6579
Chinese cuisine Chinese cuisine is an important part of Chinese culture, which includes cuisine originating from the diverse regions of China, as well as from Overseas Chinese who have settled in other parts of the world. Because of the Chinese diaspora and historical power of the country, Chinese cuisine has influenced many other cuisines in Asia, with modifications made to cater to local palates. Chinese food staples such as rice, soy sauce, noodles, tea, and tofu, and utensils such as chopsticks and the wok, can now be found worldwide. The preference for seasoning and cooking techniques of Chinese provinces depend on differences in historical background and ethnic groups. Geographic features including mountains, rivers, forests and deserts also have a strong effect on the local available ingredients, considering that the climate of China varies from tropical in the south to subarctic in the northeast. Imperial, royal and noble preference also plays a role in the change of Chinese cuisines. Because of imperial expansion and trading, ingredients and cooking techniques from other cultures are integrated into Chinese cuisines over time. The most praised "Four Major Cuisines" are Chuan, Lu, Yue and Huaiyang, representing West, North, South and East China cuisine correspondingly. The modern "Eight Cuisines" of China are Anhui (), Cantonese (), Fujian (), Hunan (), Jiangsu (), Shandong (), Sichuan (), and Zhejiang () cuisines. Color, smell and taste are the three traditional aspects used to describe Chinese food, as well as the meaning, appearance and nutrition of the food. Cooking should be appraised with respect to the ingredients used, knifework, cooking time and seasoning. Chinese society greatly valued gastronomy, and developed an extensive study of the subject based on its traditional medical beliefs. Chinese culture initially centered around the North China Plain. The first domesticated crops seem to have been the foxtail and broomcorn varieties of millet, while rice was cultivated in the south. By 2000 BC, wheat had arrived from western Asia. These grains were typically served as warm noodle soups instead of baked into bread as in Europe. Nobles hunted various wild game and consumed mutton, pork and dog as these animals were domesticated. Grain was stored against famine and flood and meat was preserved with salt, vinegar, curing, and fermenting. The flavor of the meat was enhanced by cooking it in animal fats though this practice was mostly restricted to the wealthy. By the time of Confucius in the late Zhou, gastronomy had become a high art. Confucius discussed the principles of dining: "The rice would never be too white, the meat would never be too finely cut... When it was not cooked right, man would not eat. When it was cooked bad, man would not eat. When the meat was not cut properly, man would not eat. When the food was not prepared with the right sauce, man would not eat. Although there are plenty of meats, they should not be cooked more than staple food. There is no limit for alcohol, before a man gets drunk." During Shi Huangdi's Qin dynasty, the empire expanded into the south. By the time of the Han dynasty, the different regions and cuisines of China's people were linked by major canals and leading to a greater complexity in the different regional cuisines. Not only is food seen as giving "qi", energy, but food is also about maintaining yin and yang. The philosophy behind it was rooted in the "I Ching" and Chinese traditional medicine: food was judged for color, aroma, taste, and texture and a good meal was expected to balance the Four Natures ('hot', warm, cool, and 'cold') and the Five Tastes (pungent, sweet, sour, bitter, and salty). Salt was used as a preservative from early times, but in cooking was added in the form of soy sauce, and not at the table. By the Later Han period (2nd century), writers frequently complained of lazy aristocrats who did nothing but sit around all day eating smoked meats and roasts. During the Han dynasty, the Chinese developed methods of food preservation for military rations during campaigns such as drying meat into jerky and cooking, roasting, and drying grain. Chinese legends claim that the roasted, flat bread shaobing was brought back from the "Xiyu" (the Western Regions, a name for Central Asia) by the Han dynasty General Ban Chao, and that it was originally known as hubing (, lit. "barbarian bread"). The shaobing is believed to be descended from the hubing. Shaobing is believed to be related to the Persian "nan" and Central Asian "nan", as well as the Middle Eastern pita. Foreign westerners made and sold sesame cakes in China during the Tang dynasty. During the Southern and Northern Dynasties non-Han people like the Xianbei of Northern Wei introduced their cuisine to northern China, and these influences continued up to the Tang dynasty, popularizing meat like mutton and dairy products like goat milk, yogurts, and Kumis among even Han people. It was during the Song dynasty that Han Chinese developed an aversion to dairy products and abandoned the dairy foods introduced earlier. The Han Chinese rebel Wang Su who received asylum in the Xianbei Northern Wei after fleeing from Southern Qi, at first could not stand eating dairy products like goat's milk and meat like mutton and had to consume tea and fish instead, but after a few years he was able to eat yogurt and lamb, and the Xianbei Emperor asked him which of the foods of China (Zhongguo) he preferred, fish vs mutton and tea vs yogurt. The great migration of Chinese people south during the invasions preceding and during the Song dynasty increased the relative importance of southern Chinese staples such as rice and congee. Su Dongpo has improved the red braised pork as Dongpo pork. The Yuan and Qing dynasties introduced Mongolian and Manchu cuisine, warm northern dishes that popularized hot pot cooking. During the Yuan dynasty many Muslim communities emerged in China, who practiced a porkless cuisine now preserved by Hui restaurants throughout the country. Yunnan cuisine is unique in China for its cheeses like Rubing and Rushan cheese made by the Bai people, and its yogurt, the yogurt may have been due to a combination of Mongolian influence during the Yuan dynasty, the Central Asian settlement in Yunnan, and the proximity and influence of India and Tibet on Yunnan. As part of the last leg of the Columbian Exchange, Spanish and Portuguese traders began introducing foods from the New World to China through the port cities of Canton and Macau. Mexican chili peppers became essential ingredients in Sichuan cuisine and calorically-dense potatoes and corn became staple foods across the northern plains. During the Qing Dynasty, Chinese gastronomes such as Yuan Mei focused upon a primary goal of extracting the maximum flavor of each ingredient. As noted in his culinary work the "Suiyuan shidan", however, the fashions of cuisine at the time were quite varied and in some cases were flamboyantly ostentatious, especially when the display served also a formal ceremonial purpose, as in the case of the Manchu Han Imperial Feast. As the pace of life increases in modern China, fast food like fried noodles, fried rice and "gaifan" (dish over rice) become more and more popular. There are a variety of styles of cooking in China, but Chinese chefs have classified eight regional cuisines according to their distinct tastes and local characteristics. A number of different styles contribute to Chinese cuisine but perhaps the best known and most influential are Cantonese cuisine, Shandong cuisine, Jiangsu cuisine (specifically Huaiyang cuisine) and Sichuan cuisine. These styles are distinctive from one another due to factors such as availability of resources, climate, geography, history, cooking techniques and lifestyle. One style may favour the use of garlic and shallots over chili and spices, while another may favour preparing seafood over other meats and fowl. Jiangsu cuisine favours cooking techniques such as braising and stewing, while Sichuan cuisine employs baking. Zhejiang cuisine focuses more on serving fresh food and is more like Japanese food. Fujian cuisine is famous for its delicious seafood and soups and the precise use of scintillating spices. Hunan cuisine is famous for its hot and sour taste. Anhui cuisine incorporates wild food for an unusual taste and is wilder than Fujian cuisine. Based on the raw materials and ingredients used, the method of preparation and cultural differences, a variety of foods with different flavors and textures are prepared in different regions of the country. Many traditional regional cuisines rely on basic methods of preservation such as drying, salting, pickling and fermentation. In addition, the "rice theory" attempts to describe cultural differences between north and south China; in the north, noodles are more consumed due to wheat being widely grown whereas in the south, rice is more preferred as it has historically been more cultivated there. Chinese ancestors successfully planted millet, rice, and other grains about 9,000 and 8,000 years ago. As for wheat, another staple, it took another three or four thousand years. For the first time, grains provided people with a steady supply of food. Because of the lack of food, Chinese people have to adapt to the new eating habits. The meat was scarce at that time, so people cooked with small amounts of meat and rice or noodles. Rice is a major staple food for people from rice farming areas in southern China. Steamed rice, usually white rice, is the most commonly eaten form. People in southern China also like to use rice to make congee as breakfast. Rice is also used to produce beer, baijiu and vinegars. Glutinous rice ("sticky rice") is a variety of rice used in specialty dishes such as lotus leaf rice and glutinous rice balls. In wheat-farming areas in Northern China, people largely rely on flour-based food, such as noodles, "bing" (bread), "jiaozi" (a kind of Chinese dumplings), and "mantou" (a type of steamed buns). Chinese noodles come dry or fresh in a variety of sizes, shapes and textures and are often served in soups or fried as toppings. Some varieties, such as Shou Mian (寿面, literally noodles of longevity), is an avatar of long life and good health according to Chinese traditions. Noodles can be served hot or cold with different toppings, with broth, and occasionally dry (as is the case with mi-fen). Noodles are commonly made with rice flour or wheat flour, but other flours such as soybean are also used in minor groups. Tofu is made of soybeans and is another popular food product that supplies protein. The production process of tofu varies from region to region, resulting in different kinds of tofu with a wide range of texture and taste. Other products such as soy milk, soy paste, soy oil, and fermented soy sauce are also important in Chinese cooking. There are many kinds of soybean products, including tofu skin, smoked tofu, dried tofu, and fried tofu. Stinky tofu is fermented tofu. Like blue cheese or durian, it has a very distinct, potent and strong smell, and is an acquired taste. Hard stinky tofu is often deep-fried and paired with soy sauce or salty spice. Soft stinky tofu are usually used as a spread on steamed buns. Doufuru is another type of fermented tofu that has a salty taste. Doufuru can be pickled together with soy beans, red yeast rice or chili to create different color and flavor. This is more of a pickled type of tofu and is not as strongly scented as stinky tofu. Doufuru has the consistency of slightly soft blue cheese, and a taste similar to Japanese miso paste, but less salty. Doufuru can be used as a spread on steamed buns, or paired with rice congee. Apart from vegetables that can be commonly seen, some unique vegetables used in Chinese cuisine include baby corn, bok choy, snow peas, Chinese eggplant, Chinese broccoli, and straw mushrooms. Other vegetables including bean sprouts, pea vine tips, watercress, lotus roots, water chestnuts, and bamboo shoots are also used in different cuisines of China. Because of different climate and soil conditions, cultivars of green beans, peas, and mushrooms can be found in rich variety. A variety of dried or pickled vegetables are also processed, especially in drier or colder regions where fresh vegetables were hard to get out of season. Seasonings such as fresh ginger root, garlic, scallion, cilantro and sesame are widely used in many regional cuisines. Sichuan peppercorns, star anise, cinnamon, fennel, cloves and white peppers are also used in different regions. To add extra flavors to dishes, many Chinese cuisines also contain dried Chinese mushrooms, dried baby shrimp, dried tangerine peel, and dried Sichuan chillies. When it comes to sauces, China is home to soy sauce, which is made from fermented soybeans and wheat. A number of sauces are also based on fermented soybeans, including hoisin sauce, ground bean sauce and yellow bean sauce. There are also different sauces preferred by regional cuisines, oyster sauce, fish sauce and furu (fermented tofu) are also widely used. Vinegar also has a variety with different flavors: clear rice vinegar, Chinkiang black rice vinegar, Shanxi vinegar, Henghe vinegar etc. Generally, seasonal fruits serve as the most common form of dessert consumed after dinner. Dim sum (點心), originally means small portion of food, can refer to dessert, pastries. Later to avoid the disambiguation, tian dian (甜點) and gao dian (糕點) are used to describe desserts and pastries. Chinese desserts are sweet foods and dishes that are served with tea, usually during the meal, or at the end of meals in Chinese cuisine. Besides served as a dim sum along with tea, pastries are used for celebration of traditional festivals. The most famous one is moon cake, used to celebrate the Mid-Autumn Festival. A wide variety of Chinese desserts are available, mainly including steamed and boiled sweet snacks. Bing is an umbrella term for all breads in Chinese, also including pastries and sweets. These are baked wheat flour based confections, with different stuffings including red bean paste, jujube and various of others. Su (酥) is another kind of pastry made with more amount of oil, making the confection more friable. Chinese candies and sweets, called "táng" (糖) are usually made with cane sugar, malt sugar, honey, nuts and fruit. Gao or Guo are rice based snacks that are typically steamed and may be made from glutinous or normal rice. Another cold dessert is called "baobing", which is shaved ice with sweet syrup. Chinese jellies are known collectively in the language as "ices". Many jelly desserts are traditionally set with agar and are flavored with fruits, though gelatin based jellies are also common in contemporary desserts. Chinese dessert soups are typically sweet and served hot. There are also western pastries in China, like mille-feuille, crème brûlée and cheesecake, but they are generally not as popular because the Chinese preference of dessert is mildly sweet and less oily. Many types of street foods, which vary from region to region, can be eaten as snacks or light dinner. Prawn crackers are an often-consumed snack in Southeast China. Chinese in earlier dynasties evidently drank milk and ate dairy products, although not necessarily from cows, but perhaps "koumiss" (fermented mare's milk) or goat's milk. Many Chinese have until recently avoided milk, partly because pasturage for milk producers in a monsoon rice ecology is not economic, and partly because of the high rate of lactose intolerance among the Chinese population. As such the use of dairy products in Chinese cuisine has historically been rare, with regional exceptions such as the "double skin milk" dessert in Guangdong Province or the Rubing (milk cake) cheese in Yunnan. Today ice cream is commonly available and popular throughout China. Cold dishes are usually served before the main meal. Besides salad and pickles as appetizers, they can range from jelly, beancurd, noodle salad, cooked meat and sausages, to jellyfish or cold soups. Chinese sausages vary from region to region. The most common sausage is made of pork and pork fat. Flavor is generally salty-sweet in Southern China. In other parts of China, sausages are salted to be preserved. Chinese sausage is prepared in many different ways, including oven-roasting, stir-fry, and steaming. In some part of South China, soups are served between the cold dishes and the main dishes. In other parts of China, soups are served between the main dish and staple foods, before desserts or fruit salad. There are many traditional Chinese soups, such as wonton soup, herbal chicken soup, hot and sour soup, winter melon soup and so on. Tea plays an important role in Chinese dining culture. Baijiu and huangjiu as strong alcoholic beverages are preferred by many people as well. Wine is not so popular as other drinks in China that are consumed whilst dining, although they are usually available in the menu. As well as with dim sum, many Chinese drink their tea with snacks such as nuts, plums, dried fruit (in particular jujube), small sweets, melon seeds, and waxberry. China was the earliest country to cultivate and drink tea, which is enjoyed by people from all social classes. Tea processing began after the Qin and Han Dynasties. The different types of Chinese tea include black, white, green, yellow, oolong, and dark tea. Chinese tea is often classified into several different categories according to the species of plant from which it is sourced, the region in which it is grown, and the method of production used. Some of these types are green tea, oolong tea, black tea, scented tea, white tea, and compressed tea. There are four major tea plantation regions: Jiangbei, Jiangnan, Huanan and the southwestern region. Well known types of green tea include Longjing, Huangshan, Mao Feng, Bilochun, Putuofeng Cha, and Liu'an Guapian. China is the world's largest exporter of green tea. One of the most ubiquitous accessories in modern China, after a wallet or purse and an umbrella, is a double-walled insulated glass thermos with tea leaves in the top behind a strainer. The importance of "baijiu" (lit. "white liquor") in China (99.5% of its alcoholic market) makes it the most-consumed alcoholic spirit in the world. It dates back to the introduction of distilling during the Song dynasty; can be made from wheat, corn, or rice; and is usually around 120 proof (60% ABV). The most ubiquitous brand is the cheap Er guo tou, but Mao Tai is the premium "baijiu". Other popular brands Kang, Lu Zhou Te Qu, and Wu Liang Ye. "Huangjiu" (lit. "yellow liquor") is not distilled and is a strong rice wine (10–15% ABV). Popular brands include Shaoxing Lao Jiu, Shaoxing Hua Diao, and Te Jia Fan. Chinese herb tea, also known as "medicinal herbal tea", is a kind of tea made from Chinese medicinal herbs. Soy milk, almond milk, walnut milk and coconut milk are also drunk during the meal in different regions. In some parts of China, hawthorn and jujube juice are preferred. A small shot of fruit vinegar is served as an appetizer in Shanxi. Where there are historical immigrant Chinese populations, the style of food has evolved and been adapted to local tastes and ingredients, and modified by the local cuisine, to greater or lesser extents. This has resulted in a deep Chinese influence on other national cuisines such as Cambodian cuisine, Filipino cuisine, Thai cuisine and Vietnamese cuisine. There are also a large number of forms of fusion cuisine, often popular in the country in question. Some, such as ramen (Japanese Chinese cuisine) have become popular internationally. Deep fried meat combined with sweet and sour sauce as a cooking style receives an enormous preference outside of China. Therefore, many similar international Chinese cuisines are invented based on sweet and sour sauce, including Sweet and sour chicken (Europe and North America), Manchurian chicken (India) or "tangsuyuk" (South Korea). The Hawaiian pizza was inspired by Chinese sweet and sour flavors. Apart from the host country, the dishes developed in overseas Chinese cuisines are heavily dependent on the cuisines derived from the origin of the Chinese immigrants. In Korean Chinese cuisine, the dishes derive primarily from Shandong cuisine while Filipino Chinese cuisine is strongly influenced by Fujian cuisine. The large population having Chinese ancestors in the United States operates many restaurants, has developed distinctive dishes (such as chop suey) based originally on Cantonese cuisine, while those are not popular among Chinese-American people. According to the report released by China's largest on-demand service platform in 2018, there are over 600,000 Chinese restaurants overseas. The report also pointed out that hotpot is the most popular food in the foreign market. Sichuan cuisine and some Chinese snacks and fast food followed in second and third place, respectively. Guests at a meal should greet the host and other guests after arriving at the banquet. The Chinese dining etiquette has that youths should not sit at the table before the elders. If the guest of honor or most senior member is not seated, other people are not allowed to be seated. Youths should not start eating before the elders start eating. When eating with a bowl, one should not hold it with its bottom part, because it resembles the act of begging. Chopsticks are the main eating utensils for Chinese food, which can be used to cut and pick up food. When someone taking a break from eating at the table, they should not put the chopstick into the rice vertically, because it resembles the Chinese traditional funeral tribute, which involves putting chopstick inside a bowl of rice vertically. It is considered inappropriate to use knives on the dining table. Chopsticks should not be waved around in the air or played with. Food should first be taken from the plate in front. It is considered impolite to stare at a plate. Watching TV, using mobile phones or doing other activities while eating is considered a bad habit. If an older person puts food in a younger person's bowl, the younger person should thank them. In Chinese philosophy, food is frequently used as the message that the author is trying to convey. A Chinese philosophy I Ching says, “Gentlemen use eating as a way to attain happiness. They should be aware of what they say, and refrain from eating too much." In Chinese folk religion, ancestor veneration is conducted by offering food to ancestors and Chinese festivals involve the consumption and preparation of specific foods which have symbolic meanings attached to them. Specific religions in China have their own cuisines such as the Taoist diet, Buddhist cuisine and Chinese Islamic Cuisine. The Kaifeng Jews in Henan province once had their own Chinese Jewish cuisine but the community has largely died out in the modern era and not much is known about the specifics of their cuisine but they did influence foods eaten in their region and some of their dishes remain.
https://en.wikipedia.org/wiki?curid=6583
Constantin Brâncuși Constantin Brâncuși (; February 19, 1876 – March 16, 1957) was a Romanian sculptor, painter and photographer who made his career in France. Considered a pioneer of modernism, one of the most influential sculptors of the 20th-century, Brâncuși is called the patriarch of modern sculpture. As a child he displayed an aptitude for carving wooden farm tools. Formal studies took him first to Bucharest, then to Munich, then to the École des Beaux-Arts in Paris from 1905 to 1907. His art emphasizes clean geometrical lines that balance forms inherent in his materials with the symbolic allusions of representational art. Brâncuși sought inspiration in non-European cultures as a source of primitive exoticism, as did Paul Gauguin, Pablo Picasso, André Derain and others. However, other influences emerge from Romanian folk art traceable through Byzantine and Dionysian traditions. Brâncuși grew up in the village of Hobiţa, Gorj, near Târgu Jiu, close to Romania's Carpathian Mountains, an area known for its rich tradition of folk crafts, particularly woodcarving. Geometric patterns of the region are seen in his later works. His parents Nicolae and Maria Brâncuși were poor peasants who earned a meager living through back-breaking labor; from the age of seven, Constantin herded the family's flock of sheep. He showed talent for carving objects out of wood, and often ran away from home to escape the bullying of his father and older brothers. At the age of nine, Brâncuși left the village to work in the nearest large town. At 11 he went into the service of a grocer in Slatina; and then he became a domestic in a public house in Craiova where he remained for several years. When he was 18, Brâncuși created a violin by hand with materials he found around his workplace. Impressed by Brâncuși's talent for carving, an industrialist enrolled him in the Craiova School of Arts and Crafts ("școala de arte și meserii"), where he pursued his love for woodworking, graduating with honors in 1898. He then enrolled in the Bucharest School of Fine Arts, where he received academic training in sculpture. He worked hard, and quickly distinguished himself as talented. One of his earliest surviving works, under the guidance of his anatomy teacher, Dimitrie Gerota, is a masterfully rendered écorché (statue of a man with skin removed to reveal the muscles underneath) which was exhibited at the Romanian Athenaeum in 1903. Though just an anatomical study, it foreshadowed the sculptor's later efforts to reveal essence rather than merely copy outward appearance. In 1903, Brâncuși traveled to Munich, and from there to Paris. In Paris, he was welcomed by the community of artists and intellectuals brimming with new ideas. He worked for two years in the workshop of Antonin Mercié of the École des Beaux-Arts, and was invited to enter the workshop of Auguste Rodin. Even though he admired the eminent Rodin he left the Rodin studio after only two months, saying, "Nothing can grow under big trees." After leaving Rodin's workshop, Brâncuși began developing the revolutionary style for which he is known. His first commissioned work, "The Prayer", was part of a gravestone memorial. It depicts a young woman crossing herself as she kneels, and marks the first step toward abstracted, non-literal representation, and shows his drive to depict "not the outer form but the idea, the essence of things." He also began doing more carving, rather than the method popular with his contemporaries, that of modeling in clay or plaster which would be cast in metal, and by 1908 he worked almost exclusively by carving. In the following few years he made many versions of "Sleeping Muse" and "The Kiss", further simplifying forms to geometrical and sparse objects. His works became popular in France, Romania and the United States. Collectors, notably John Quinn, bought his pieces, and reviewers praised his works. In 1913 Brâncuși's work was displayed at both the Salon des Indépendants and the first exhibition in the U.S. of modern art, the Armory Show. In 1920, he developed a notorious reputation with the entry of "Princess X" in the Salon. The phallic appearance of this large, gleaming bronze piece scandalized the Salon and, despite Brâncuși's explanation that it was simply meant to represent the essence of womanhood, removed it from the exhibition. "Princess X" was revealed to be Princess Marie Bonaparte, direct descendant of the younger brother of Napoleon Bonaparte. The sculpture has been interpreted by some as symbolizing her obsession with the penis and her lifelong quest to achieve vaginal orgasm, with the help of Sigmund Freud. Around this time Brâncuși began crafting the bases for his sculptures with much care and originality because he considered them important to the works themselves. One of his major groups of sculptures involved the "Bird in Space" — simple abstract shapes representing a bird in flight. The works are based on his earlier "Măiastra" series. In Romanian folklore the Măiastra is a beautiful golden bird who foretells the future and cures the blind. Over the following 20 years, Brâncuși made multiple versions of "Bird in Space" out of marble or bronze. Athena Tacha Spear's book, "Brâncuși's Birds," (CAA monographs XXI, NYU Press, New York, 1969), first sorted out the 36 versions and their development, from the early "Măiastra", to the "Golden Bird" of the late teens, to the "Bird in Space", which emerged in the early 1920s and which Brâncuși developed throughout his life. One of these versions caused a major controversy in 1926, when photographer Edward Steichen purchased it and shipped it to the United States. Customs officers did not accept the "Bird" as a work of art and assessed customs duty on its import as an industrial item. After protracted court proceedings, this assessment was overturned, thus confirming the Bird's status as a duty-exempt work of art. The ruling also established the important principle that "art" does not have to involve a realistic representation of nature, and that it was legitimate for it to simply represent an abstract concept – in this case "flight". His work became increasingly popular in the U.S, where he visited several times during his life. Worldwide fame in 1933 brought him the commission of building a meditation temple, the Temple of Deliverance, in India for the Maharajah of Indore, Yeshwant Rao Holkar. Holkar had commissioned three "L'Oiseau dans l'Espace"—in bronze, black and white marble—previously, but when Brâncuși went to India in 1937 to complete the plans and begin construction, the Mahrajah was away and, supposedly, lost interest in the project which was to be an homage to his wife, the Maharani Margaret Holkar, who had died when he returned. Of the three birds, the bronze one is in the collection of the Norton Simon Museum in Pasadena, California, and the two marble birds are currently in the permanent collection of the National Gallery of Australia in Canberra, Australia. In 1938, he finished the World War I monument in Târgu-Jiu where he had spent much of his childhood. "Table of Silence", "The Gate of the Kiss", and "Endless Column" commemorate the courage and sacrifice of Romanians who in 1916 defended Târgu Jiu from the forces of the Central Powers. The restoration of this ensemble was spearheaded by the World Monuments Fund and was completed in 2004. The Târgu Jiu ensemble marks the apex of his artistic career. In his remaining 19 years he created less than 15 pieces, mostly reworking earlier themes, and while his fame grew he withdrew. In 1955 "Life" magazine reported, "Wearing white pajamas and a yellow gnome-like cap, Brâncuși today hobbles about his studio tenderly caring for and communing with the silent host of fish, birds, heads, and endless columns which he created." Brâncuși was cared for in his later years by a Romanian refugee couple. He became a French citizen in 1952 in order to make the caregivers his heirs, and to bequeath his studio and its contents to the Musée National d'Art Moderne in Paris. Brâncuși always dressed in the simple ways the Romanian peasants did. His studio was reminiscent of the houses of the peasants from his native region: there was a big slab of rock as a table and a primitive fireplace, similar to those found in traditional houses in his native Oltenia, while the rest of the furniture was made by him out of wood. Brâncuși would cook his own food, traditional Romanian dishes, with which he would treat his guests. Brâncuși held a large spectrum of interests, from science to music. He was a good violinist and he would sing old Romanian folk songs, often expressing by them his feelings of homesickness. After the installment of communism, he never considered moving back permanently to his native Romania, but he did visit it eight times. His circle of friends included artists and intellectuals in Paris such as Amedeo Modigliani, Ezra Pound, Henri Pierre Roché, Guillaume Apollinaire, Louise Bourgeois, Pablo Picasso, Man Ray, Marcel Duchamp, Henri Rousseau, Peggy Guggenheim, Tristan Tzara and Fernand Léger. He was an old friend of Romany Marie, who was also Romanian, and referred Isamu Noguchi to her café in Greenwich Village. Although surrounded by the Parisian avant-garde, Brâncuși never lost contact with Romania and had friends from the community of Romanian artists and intellectuals living in Paris, including Benjamin Fondane, George Enescu, Theodor Pallady, Camil Ressu, Nicolae Dărăscu, Panait Istrati, Traian Vuia, Eugène Ionesco, Emil Cioran, Natalia Dumitresco and Paul Celan. Another Romanian scholar wrote on Brancusi, Mircea Eliade. Brâncuși held a particular interest in mythology, especially Romanian mythology, folk tales, and traditional art (which also had a strong influence on his works), but he became interested in African and Mediterranean art as well. A talented handyman, he built his own phonograph and made most of his furniture, utensils, and doorways. His worldview valued "differentiating the essential from the ephemeral," with Plato, Lao-Tzu, and Milarepa as influences. Reportedly, he had a copy of the first ever translation from the Tibetan into French of Jacques Bacot's Le poete tibetain Milarepa: ses crimes, ses épreuves, son Nirvana that he kept by his bedside. He identified closely with Milarepa's mountain existence since Brancusi himself came from the Carpathian Mountains of Romania and he often thought of he was a reincarnation of Milarepa. He was a saint-like idealist and near ascetic, turning his workshop into a place where visitors noted the deep spiritual atmosphere. However, particularly through the 1910s and 1920s, he was known as a pleasure seeker and merrymaker in his bohemian circle. He enjoyed cigarettes, good wine, and the company of women. He had one child, John Moore, with the New Zealand pianist Vera Moore, whom he never acknowledged. Brâncuși died on March 16, 1957, aged 81. He was buried in the Cimetière du Montparnasse in Paris. This cemetery also displays statues that Brâncuși carved for deceased artists. In 1962, Georg Olden used Brâncuși's "Bird in Space" as the inspiration behind his design of the Clio Award statuette. At his death Brâncuși left 1200 photographs and 215 sculptures. He bequeathed part of his collection to the French state on condition that his workshop be rebuilt as it was on the day he died. This reconstruction of his studio, adjacent to the Pompidou Centre, is open to the public. Brâncuși's studio inspired Swedish architect Klas Anshelm's design of the Malmö Konsthall, which opened in 1975. Brâncuși was elected posthumously to the Romanian Academy in 1990. Google commemorated his 135th birthday with a Doodle in 2011 consisting of seven of his works. Brâncuși's works are housed in the National Museum of Art of Romania (Bucharest), the Museum of Modern Art (New York) and other museums around the world. The Philadelphia Museum of Art holds the largest collection of Brâncuși sculptures in the United States. In 2015 the Romanian Parliament declared February 19 "The Brâncuși Day", a working holiday in Romania. Brâncuși's piece "Madame L.R." sold for €29.185 million ($37.2 million) in 2009, setting a record price for a sculpture sold at auction. In May 2018, "La Jeune Fille Sophistiquée" ("Portrait de Nancy Cunard"), a polished bronze on a carved marble base (1932), sold for US$71 million (with fees) at Christie's New York, setting a world record auction price for the artist. Both "Bird in Space" and "Sleeping Muse I" are sculptures of animate objects; however, unlike ones from Ancient Greece or Rome, or those from the High Renaissance period, these works of art are more abstract in style. "Bird in Space" is a series from the 1920s. One of these, constructed in 1925 using wood, stone, and marble (Richler 178) stands around 72 inches tall and consists of a narrow feather standing erect on a wooden base. Similar models, but made from materials such as bronze, were also produced by Brâncuși and placed in exhibitions. "Sleeping Muse I" has different versions as well; one, from 1909–10, is made of marble and measures 6 ¾ in. in height (Adams 549). This is a model of a head, without a body, with markings to show features such as hair, nose, lips, and closed eyes. In "A History of Western Art", Adams says that the sculpture has "an abstract, curvilinear quality and a smooth contour that create an impression of elegance" (549). The qualities which produce the effect can particularly be seen in the shape of the eyes and in the set of the mouth.
https://en.wikipedia.org/wiki?curid=6585
Cadillac, Michigan Cadillac is a city in the U.S. state of Michigan. The city is the county seat of Wexford County. The population was 10,355 at the 2010 census. Today US 131, M-55 and M-115 have a junction at this city. The geographic center of Michigan is approximately five miles (8.05 km) north-northwest of Cadillac. Cadillac became the county seat after the so-called "Battle of Manton". Local officials engaged in a show of force to enforce the controversial state legislative decision to move the county seat from Manton. European explorers and fur traders visited this area from the 18th century, most of them initially French and French-Canadians who traded with regional Native Americans. More permanent communities were not established until the late 19th century. Initial settlements developed from logging camps and the logging industry. In 1871, the first sawmill began operations at Cadillac. Originally called the Pioneer Mill, it was built by John R. Yale. That same year, George A. Mitchell, a prominent Cadillac banker and railroad entrepreneur, and Adam Gallinger, a local carpenter, formed the Clam Lake Canal Improvement and Construction Company. Two years later, the Clam Lake Canal was constructed between Big and Little Clam lakes, known as present-day lakes Mitchell and Cadillac. Sawmill owners used the canal to transport timber from Big Clam Lake to the mills and railroad sites on Little Clam Lake. The Grand Rapids and Indiana Railroad (|G.R. & I. Railroad) had reached the area in 1872. This settlement was originally named Clam Lake and was incorporated as a village in 1874. George Mitchell was elected as the first mayor. The village was incorporated as a city in 1877 and renamed Cadillac, after Antoine Laumet de La Mothe, sieur de Cadillac, a French colonist who started the first permanent settlement at Detroit in 1701. The Wexford County seat of government, originally located in Sherman, was moved to Manton in 1881, as the result of a compromise between the feuding residents of Cadillac and Sherman. Cadillac partisans, however, won the county seat by a county-wide vote in April 1882. The day following the election, a sheriff's posse left the city for Manton by special train to seize the county records. After they arrived and collected a portion of the materials, however, an angry crowd confronted the Cadillac men and drove them out of town. When the sheriff returned to Cadillac, he encountered a force consisting of several hundred armed men; this group reportedly included a brass band. The Sheriff's force, some of whom may have been intoxicated, traveled back to Manton to seize the remaining records. Although Manton residents confronted the Cadillac men and barricaded the courthouse, the posse successfully seized the documents. They returned to Cadillac in dubious glory. In 1878, Ephraim Shay perfected his Shay locomotive, which was particularly effective in its ability to climb steep grades, maneuver sharp turns, and accommodate imperfections in railroad tracks. Cadillac was home to the Michigan Iron Works Company, which manufactured the Shay locomotive for a short time in the early 1880s. The lumber industry continued to dominate the city, attracting a large immigrant labor force, most of whom were Swedish. (Later Cadillac made sister city arrangements with Mölnlycke, Sweden, and Rovaniemi, Finland). In 1899, the Cadillac Club formed, the forerunner of the Cadillac Area Chamber of Commerce. Gradually, various manufacturing firms found success in Cadillac. By the early 20th century, with the lumber depleted, the timber industry was in decline. Industrial development soon dominated the local economy, and it continues to do so today. Cadillac's range of industries includes the manufacture of pleasure boats, automotive parts, water-well components, vacuum cleaners, and rubber products. In 1936, the U.S. Forest Service and the Civilian Conservation Corps developed the Caberfae Ski Area during the Great Depression as an investment in future economic development. This resulted in promotion of this area as a tourist center. Caberfae remains in operation today, as the oldest ski resort in the midwest. Tourism and outdoor recreation have since become an important sector of Cadillac's economy. In the summer, tourists travel to the city and region for boating, fishing, hiking, mountain biking, and camping. During the fall, hunting and color tours are popular. The winter is possibly the busiest season; the area can be found packed with downhill skiers, cross-country skiers, ice-fishers, snow-shoers and–most of all-snowmobilers. The North American Snowmobile Festival (NASF) is held on frozen Lake Cadillac every winter. Thirsty's, a gas station on M-55 west of Cadillac, was the home of Samantha or "Sam The Bear" from the 1970s through the late 1990s, when Sam died of old age. Sam was the only brown bear in captivity in the US at the time to hibernate naturally. Sam lived in a large cage in front of the gas station and was fed ice cream cones by tourists every summer. In October 1975 the rock group Kiss visited Cadillac and performed at the Cadillac High School gymnasium. They played the concert to honor the Cadillac High School football team. In previous years, the team had compiled a record of sixteen consecutive victories, but the 1974 squad opened the season with two losses. The assistant coach, Jim Neff, an English teacher and rock'n'roll fan, thought to inspire the team by playing Kiss music in the locker room. He also connected the team's game plan, K-I-S-S or "Keep It Simple Stupid", with the band. The team went on to win seven straight games and their conference co-championship. After learning of their association with the team's success, the band decided to visit the school and play for the homecoming game. Cadillac maintains a number of state historic landmarks. Most are marked with a green "Michigan Historical Marker" sign, which includes a description of the landmark. Six sites with the city are marked: Cadillac Carnegie Library, Charles T. Mitchell House, Clam Lake Canal, Cobbs & Mitchell Building, Cobbs & Mitchell No. 1, and the Shay Locomotive (pictured at the right). Two more are in the near Cadillac area: Caberfae Ski Resort and Greenwood Disciples of Christ Church; and another two are in surrounding Wexford County, marking Battle of Manton and the First Wexford County Court House. According to the United States Census Bureau, the city has a total area of , of which is land and is water. The Lake Cadillac is entirely within the city limits. The larger, Lake Mitchell is nearby on the west side of the city, with of shoreline within the city's municipal boundary. The lakes were connected by a stream which was replaced in 1873 by the Clam Lake Canal. The canal was featured on Ripley's Believe It or Not in the 1970s due to the phenomenon that in winter the canal freezes before the lakes and then after the lakes freeze, the canal thaws and remains unfrozen for the rest of the winter. Cadillac is located at the eastern edge of what is now managed as the Manistee National Forest. The surrounding area is heavily wooded, with mixed hardwood and conifer forests. Christmas tree farming has been important to the area agricultural industry. Cadillac was chosen in 1988 to donate the holiday tree installed at the lawn of the U.S. Capitol building in Washington, D.C. The area surrounding Cadillac is primarily rural, and is considered to be part of Northern Michigan. Given the small size of nearby communities, the city is a major commercial and industrial hub of the region. The commercial center of the city is located on the eastern edge of Lake Cadillac. Most downtown buildings range from two to five stories in height. Many face Mitchell Street, the city's tree-lined main street and traditional corridor of travel through town. The downtown contains a movie theater, gift shops, restaurants, a bookstore, specialty food stores, jewelers, clothing retailers, and various other businesses. The Courthouse Hill Historic District, recognized in April 2005, lies adjacent to the city's commercial center. The District contains a number of large Victorian-style residences built by the lumber barons and businessmen who helped develop the city in the 1870s. Population and building density is highest in this area. On the western bank of Lake Cadillac, where M-55 intersects M-115, is what is locally referred to as Cadillac West. This is a small commercial district, bordering Mitchell State Park and the two lakes; it caters mostly to tourists. It contains a number of motels and restaurants. Along the northern and southern stretches of the lake are the main residential areas of the city. They are generally of low to moderate density, characterized primarily by single-family structures. Cadillac experiences a typical northern Michigan climate, undergoing temperate seasonal changes, influenced by the presence of Lake Michigan and the inevitable lake effect. Winters are generally cold with large amounts of snowfall. Summers are warm. The average high temperature in July is 80 °F (27 °C) and the average low is in February, at 9 °F (−13 °C). Summer temperatures can exceed 90 °F (32 °C), and winter temperatures can drop below 0 °F (−18 °C). Average annual rainfall is 30 inches (76 cm), and average annual snowfall is 81 inches (206 cm) . Snowfall typically occurs between the months of November and March. According to the Köppen climate classification system, Cadillac has a humid continental climate, abbreviated "Dfb" on climate maps. Cadillac has two superfund sites, according to the U.S. Environmental Protection Agency. One is located at 1100 Wright Street, the former site of Kysor Industrial Corp, which operations resulted in toxic wastes. The other is located at 1002 6th Street, the former site of Northernaire Plating. Its operations also produced hazardous wastes, which produced contamination. As of the census of 2010, there were 10,355 people, 4,280 households, and 2,625 families residing in the city. The population density was . There were 4,927 housing units at an average density of . The racial makeup of the city was 95.6% White, 0.5% African American, 0.6% Native American, 1.0% Asian, 0.4% from other races, and 1.8% from two or more races. Hispanic or Latino of any race were 1.8% of the population. There were 4,280 households, of which 32.9% had children under the age of 18 living with them, 39.2% were married couples living together, 16.4% had a female householder with no husband present, 5.7% had a male householder with no wife present, and 38.7% were non-families. 32.0% of all households were made up of individuals, and 14% had someone living alone who was 65 years of age or older. The average household size was 2.34 and the average family size was 2.90. The median age in the city was 36.5 years. 24.7% of residents were under the age of 18; 10% were between the ages of 18 and 24; 24.4% were from 25 to 44; 23.8% were from 45 to 64; and 17.1% were 65 years of age or older. The gender makeup of the city was 47.4% male and 52.6% female. As of the census of 2000, there were 10,000 people, 4,118 households, and 2,577 families residing in the city. The population density was 1,466.0 per square mile (566.1/km2). There were 4,466 housing units at an average density of 654.7 per square mile (252.8/km2). The racial makeup of the city was 96.55% White, 0.21% Black or African American, 0.92% Native American, 0.63% Asian, 0.03% Pacific Islander, 0.28% from other races, and 1.38% from two or more races. 1.18% of the population were Hispanic or Latino of any race. There were 4,118 households, out of which 32.2% had children under the age of 18 living with them, 43.9% were married couples living together, 14.2% had a female householder with no husband present, and 37.4% were non-families. 31.8% of all households were made up of individuals, and 14.4% had someone living alone who was 65 years of age or older. The average household size was 2.37 and the average family size was 2.96. In the city, the population was spread out, with 26.2% under the age of 18, 9.6% from 18 to 24, 27.9% from 25 to 44, 19.6% from 45 to 64, and 16.7% who were 65 years of age or older. The median age was 36 years. For every 100 females, there were 91.4 males. For every 100 females age 18 and over, there were 84.4 males. The median income for a household in the city was $29,899, and the median income for a family was $36,825. Males had a median income of $29,773 versus $21,283 for females. The per capita income for the city was $16,801. About 10.9% of families and 13.7% of the population were below the poverty line, including 15.4% of those under age 18 and 13.3% of those age 65 or over. Cadillac was incorporated as a city in 1877. It is a home rule city with a Council-Manager form of government-one. Current council members are Shari Spoelman, Antoinette Schippers, Arthur Stevens, James Dean and Carla Filkins (mayor). The present City Manager is Marcus Peccia. Cadillac is located in Michigan's 2nd congressional district, represented by Republican Bill Huizenga. Manufacturing has been the greatest employer in Cadillac since the logging industry. More than 26% of the city's labor force is employed in manufacturing. Three industrial parks are located within the city limits, comprising 7% of the total land use in Cadillac. Their operations generate 47% of the city's tax base. Much of the city's economic performance is determined by the fortunes of local industry. Major manufacturers include Four Winns, AAR Manufacturing, Avon Rubber, FIAMM Technologies, Michigan Rubber Products and Rexair. The center of the city is generally perceived to have a "small-town-feel". In the summer, the downtown fills with tourists, many from southern Michigan. The city center is one block from Lake Cadillac. For visitors by boat who dock at the public docks, it is nearly as accessible by boat as it is by car. The city's immediate proximity to two lakes, as well as Manistee National Forest, Pere Marquette State Forest, Mitchell State Park and a number of major highways, has established tourism as a significant sector of the local economy. During the winter months, Lake Cadillac and Lake Mitchell freeze over and the city becomes covered with snow. Cadillac is connected to a number of trail systems popular with winter recreation enthusiasts. The city integrates unusually well into the corridors of travel created by snowmobilers. Cadillac is also known as Chestnut Town, USA. The local area has a relatively high number of American chestnut trees, planted by pioneers from New York and Pennsylvania who settled in western Michigan. A blight in the early 20th century killed nearly every American Chestnut tree, but those in western Michigan had developed a mysterious resistance and survived. Cadillac's public education system has a total of 10 schools, with approximately 3,100 students and 166 teachers with a student:teacher ratio of 19.1:1. Cadillac has 4 private primary and secondary schools with approximately 394 students, 20 teachers and a student:teacher ratio of 20:1. The city has two high schools: Cadillac High School and Innovation High School. The area also has a junior high school, covering grades 7 and 8, located adjacent to the high school, and a middle school, Mackinaw Trail Middle School, covering grades 5 and 6. There are four elementary schools, Forest View Elementary, Franklin Elementary, Kenwood Elementary, and Lincoln Elementary. Cadillac also has an alternative high school, located in the building that formerly housed Cooley Elementary School. Adult high school and GED courses are offered there as well. As a whole, the programs at Cooley are part of a curriculum that aids individuals in overcoming the exceptional obstacles to their educational and workforce goals. Vocational career training is available to high school students free of charge in Cadillac and nearby schools at Wexford-Missaukee Independent School District (ISD) Career Tech Center (formerly Wexford-Missaukee Vocational Center or Voc-Tech). Students are bussed for part of the day to the Career Tech Center from their respective schools and receive credits toward high school graduation. Students are also able to earn certification in a chosen trade. Courses include: Cosmetology is offered through the Career Tech Center, but at an off-campus location in downtown Cadillac. Adults can attend the vocational or cosmetology school with tuition or financial aid for certification. Cadillac hosts the Wexford-Missaukee ISD Special Education for residents of the two counties who are in need of special services. This school is on the same campus as the Career Tech Center. The class of 2006 was the largest class to go through Cadillac Public Schools. Cadillac offers several options for private religious education. Cadillac Heritage Christian offers nondenominational Christian education from pre-K through 12th grade. It is a coed school with 98 students and a teacher:student ratio of 1:11. Graduating classes are typically between 3–12 students. Northview SDA Christian School has 5 students in grades 1–9. It is a coed Seventh Day Adventist school. Noah's Ark Day School is a small alternative non-denominational Christian school for students in pre-K through first grade only. It is coed with 42 students and 1 teacher. Cadillac's largest and most well-known private school is St. Ann School, a coed private Roman Catholic school with 236 students in grades pre-K through 7. The teacher:student ratio is 1:26. St. Ann is a member of the National Catholic Education Association. No Catholic high school education is offered at St. Ann School, and students typically attend public school for grades 8–12. Northwoods Aviation, located at Wexford County Airport, offers training programs for piloting and servicing aircraft. Northwoods Aviation also offers primary instruction for those interested in sport pilot, private, and commercial certificates. The Cadillac Institute of Cosmetology (formerly Cadillac Academy of Beauty) is a full service teaching salon in downtown Cadillac that offers training for general cosmetologists and specialized technicians to high school students through a partnership with Wexford-Missaukee Intermediate School District. Training is also available to adult students though private courses on a tuition basis. Upon completion of the program, students are qualified to take the state board exam to become a licensed cosmetologist or specialty technician. The Baker College-Cadillac campus occupies just outside the City of Cadillac. The school has an enrollment of more than 1,300 students and offers Associate's and bachelor's degrees, in addition to professional certifications. Cadillac is situated as the confluence of three highways: US 131, M-55 and M-115. Prior to 2001, the northern end of the freeway portion of US 131 was located at the southern entrance to Cadillac. With the construction of a bypass, the US 131 freeway was extended around the east side of the city. The former route of the highway through downtown Cadillac was redesignated as BUS US 131. In the city, BUS US 131 is named Mitchell Street, after George Mitchell, but may be referred to as main street. The city is serviced by rail via the Great Lakes Central Railroad. This is primarily a freight line, although passenger service is expected in the future. The White Pine Trail's northern terminus is in Cadillac. The trail, which stretches and originates from Comstock Park, follows an abandoned railroad bed into the center of the city. The trail is paved from the village of Leroy 16 miles north to Cadillac. Cadillac is also home to NewsNet, a national digital broadcast news channel.
https://en.wikipedia.org/wiki?curid=6587
COINTELPRO COINTELPRO (syllabic abbreviation derived from COunter INTELligence PROgram) (1956–1979, and beyond) is a series of covert and illegal projects conducted by the United States Federal Bureau of Investigation (FBI) aimed at surveilling, infiltrating, discrediting, and disrupting American political organizations. FBI records show that COINTELPRO resources targeted groups and individuals that the FBI deemed subversive, including feminist organizations, the Communist Party USA, anti–Vietnam War organizers, activists of the civil rights movement or Black Power movement (e.g. Martin Luther King Jr., the Nation of Islam, and the Black Panther Party), environmentalist and animal rights organizations, the American Indian Movement (AIM), independence movements (such as Puerto Rican independence groups like the Young Lords), and a variety of organizations that were part of the broader New Left. The program also targeted the Ku Klux Klan in 1964. In 1971 in San Diego, the FBI financed, armed, and controlled an extreme right-wing group of former members of the Minutemen anti-communist para-military organization, transforming it into a group called the Secret Army Organization that targeted groups, activists, and leaders involved in the Anti-War Movement, using both intimidation and violent acts. The FBI has used covert operations against domestic political groups since its inception; however, covert operations under the official COINTELPRO label took place between 1956 and 1971. COINTELPRO tactics are still used to this day and have been alleged to include discrediting targets through psychological warfare; smearing individuals and groups using forged documents and by planting false reports in the media; harassment; wrongful imprisonment; and illegal violence, including assassination. According to a senate report, the FBI's motivation was "protecting national security, preventing violence, and maintaining the existing social and political order". Beginning in 1969, leaders of the Black Panther Party were targeted by the COINTELPRO and "neutralized" by being assassinated, imprisoned, publicly humiliated or falsely charged with crimes. Some of the Black Panthers affected included Fred Hampton, Mark Clark, Zayd Shakur, Geronimo Pratt, Mumia Abu-Jamal, and Marshall Conway. Common tactics used by COINTELPRO were perjury, witness harassment, witness intimidation, and withholding of evidence. FBI Director J. Edgar Hoover issued directives governing COINTELPRO, ordering FBI agents to "expose, disrupt, misdirect, discredit, or otherwise neutralize" the activities of these movements and especially their leaders. Under Hoover, the agent in charge of COINTELPRO was William C. Sullivan. Attorney General Robert F. Kennedy personally authorized some of the programs. Although Kennedy only gave written approval for limited wiretapping of Martin Luther King's phones "on a trial basis, for a month or so", Hoover extended the clearance so his men were "unshackled" to look for evidence in any areas of King's life they deemed worthy. Centralized operations under COINTELPRO officially began in August 1956 with a program designed to "increase factionalism, cause disruption and win defections" inside the Communist Party USA (CPUSA). Tactics included anonymous phone calls, Internal Revenue Service (IRS) audits, and the creation of documents that would divide the American communist organization internally. An October 1956 memo from Hoover reclassified the FBI's ongoing surveillance of black leaders, including it within COINTELPRO, with the justification that the movement was infiltrated by communists. In 1956, Hoover sent an open letter denouncing Dr. T. R. M. Howard, a civil rights leader, surgeon, and wealthy entrepreneur in Mississippi who had criticized FBI inaction in solving recent murders of George W. Lee, Emmett Till, and other African Americans in the South. When the Southern Christian Leadership Conference (SCLC), an African-American civil rights organization, was founded in 1957, the FBI began to monitor and target the group almost immediately, focusing particularly on Bayard Rustin, Stanley Levison, and eventually Martin Luther King Jr. After the 1963 March on Washington for Jobs and Freedom, Hoover singled out King as a major target for COINTELPRO. Under pressure from Hoover to focus on King, Sullivan wrote: In the light of King's powerful demagogic speech. ... We must mark him now if we have not done so before, as the most dangerous Negro of the future in this nation from the standpoint of communism, the Negro, and national security. Soon after, the FBI was systematically bugging King's home and his hotel rooms, as they were now aware that King was growing in stature daily as the most prominent leader of the civil rights movement. In the mid-1960s, King began to publicly criticize the Bureau for giving insufficient attention to the use of terrorism by white supremacists. Hoover responded by publicly calling King the most "notorious liar" in the United States. In his 1991 memoir "Washington Post", journalist Carl Rowan asserted that the FBI had sent at least one anonymous letter to King encouraging him to commit suicide. Historian Taylor Branch documents an anonymous November 21, 1964 "suicide package" sent by the FBI that contained audio recordings obtained through tapping King's phone and placing bugs throughout various hotel rooms over the past two years, and that was created two days after the announcement of King's impending Nobel Peace Prize. The tape, which was prepared by FBI audio technician John Matter, documented a series of King's sexual indiscretions combined with a letter telling him: "There is only one way out for you. You better take it before your filthy, abnormal, fraudulent self is bared to the nation". King was subsequently informed that the audio would be released to the media if he did not acquiesce and commit suicide prior to accepting his Nobel Peace Prize. When King refused to satisfy their coercion tactics, FBI Associate Director, Cartha D. DeLoach, commenced a media campaign offering the surveillance transcript to various news organizations, including "Newsweek" and "Newsday". And even by 1969, as has been noted elsewhere, "[FBI] efforts to 'expose' Martin Luther King Jr. had not slackened even though King had been dead for a year. [The Bureau] furnished ammunition to opponents that enabled attacks on King's memory, and ... tried to block efforts to honor the slain leader." During the same period the program also targeted Malcolm X. While an FBI spokesman has denied that the FBI was "directly" involved in Malcolm's murder in 1965, it is documented that the Bureau worked to "widen the rift" between Malcolm and Elijah Muhammad through infiltration and the "sparking of acrimonious debates within the organization", rumor-mongering, and other tactics designed to foster internal disputes, which ultimately led to Malcolm's assassination. The FBI heavily infiltrated Malcolm's Organization of Afro-American Unity in the final months of his life. The Pulitzer Prize-winning asserts that most of the men who plotted Malcolm's assassination were never apprehended and that the full extent of the FBI's involvement in his death cannot be known. Amidst the urban unrest of July–August 1967, the FBI began "COINTELPRO–BLACK HATE", which focused on King and the SCLC, as well as the Student Nonviolent Coordinating Committee (SNCC), the Revolutionary Action Movement (RAM), the Deacons for Defense and Justice, Congress of Racial Equality (CORE), and the Nation of Islam. BLACK HATE established the Ghetto Informant Program and instructed 23 FBI offices to "disrupt, misdirect, discredit, or otherwise neutralize the activities of black nationalist hate type organizations". A March 1968 memo stated the program's goal was to "prevent the coalition of militant black nationalist groups"; to "Prevent the RISE OF A 'MESSIAH' who could unify ... the militant black nationalist movement"; "to pinpoint potential troublemakers and neutralize them before they exercise their potential for violence [against authorities]."; to "Prevent militant black nationalist groups and leaders from gaining RESPECTABILITY, by discrediting them to ... both the responsible community and to liberals who have vestiges of sympathy..."; and to "prevent the long-range GROWTH of militant black organizations, especially among youth". Dr. King was said to have potential to be the "messiah" figure, should he abandon nonviolence and integrationism, and Stokely Carmichael was noted to have "the necessary charisma to be a real threat in this way" as he was portrayed as someone who espoused a much more militant vision of "black power". While the FBI was particularly concerned with leaders and organizers, they did not limit their scope of target to the heads of organizations. Individuals such as writers were also listed among the targets of operations. This program coincided with a broader federal effort to prepare military responses for urban riots and began increased collaboration between the FBI, Central Intelligence Agency, National Security Agency, and the Department of Defense. The CIA launched its own domestic espionage project in 1967 called Operation CHAOS. A particular target was the Poor People's Campaign, a national effort organized by King and the SCLC to occupy Washington, DC. The FBI monitored and disrupted the campaign on a national level, while using targeted smear tactics locally to undermine support for the march. The Black Panther Party was another targeted organization, wherein the FBI collaborated to destroy the party from the inside out. Overall, COINTELPRO encompassed disruption and sabotage of the Socialist Workers Party (1961), the Ku Klux Klan (1964), the Nation of Islam, the Black Panther Party (1967), and the entire New Left social/political movement, which included antiwar, community, and religious groups (1968). A later investigation by the Senate's Church Committee (see below) stated that "COINTELPRO began in 1956, in part because of frustration with Supreme Court rulings limiting the Government's power to proceed overtly against dissident groups." Official congressional committees and several court cases have concluded that COINTELPRO operations against communist and socialist groups exceeded statutory limits on FBI activity and violated constitutional guarantees of freedom of speech and association. The program was secret until 1971, when the Citizens' Commission to Investigate the FBI burgled an FBI field office in Media, Pennsylvania, took several dossiers, and exposed the program by passing this material to news agencies. The boxing match known as the Fight of the Century between Muhammad Ali and Joe Frazier in March 1971 provided cover for the activist group to successfully pull off the burglary; Muhammad Ali was himself a COINTELPRO target due to his involvement with the Nation of Islam and the anti-war movement. Many news organizations initially refused to publish the information. Within the year, Director J. Edgar Hoover declared that the centralized COINTELPRO was over, and that all future counterintelligence operations would be handled on a case-by-case basis. Additional documents were revealed in the course of separate lawsuits filed against the FBI by NBC correspondent Carl Stern, the Socialist Workers Party, and a number of other groups. In 1976 the Select Committee to Study Governmental Operations with Respect to Intelligence Activities of the United States Senate, commonly referred to as the "Church Committee" after its chairman, Senator Frank Church of Idaho, launched a major investigation of the FBI and COINTELPRO. Many released documents have been partly or entirely redacted. The Final Report of the Select Committee castigated the conduct of the intelligence community in its domestic operations (including COINTELPRO) in no uncertain terms: The Church Committee documented a history of the FBI exercising political repression as far back as World War I, through the 1920s, when agents were charged with rounding up "anarchists, communists, socialists, reformists and revolutionaries" for deportation. The domestic operations were increased against political and anti-war groups from 1936 through 1976. The intended effect of the FBI's COINTELPRO was to "expose, disrupt, misdirect, or otherwise neutralize" groups that the FBI officials believed were "subversive" by instructing FBI field operatives to: At its inception, the program's main target was the Communist Party. In an interview with the BBC's Andrew Marr in February 1996, Noam Chomsky—a political activist and MIT professor of linguistics—spoke about the purpose and the targets of COINTELPRO, saying: COINTELPRO was a program of subversion carried out not by a couple of petty crooks but by the national political police, the FBI, under four administrations... by the time it got through, I won't run through the whole story, it was aimed at the entire new left, at the women's movement, at the whole black movement, it was extremely broad. Its actions went as far as political assassination. According to the Church Committee: While the declared purposes of these programs were to protect the "national security" or prevent violence, Bureau witnesses admit that many of the targets were nonviolent and most had no connections with a foreign power. Indeed, nonviolent organizations and individuals were targeted because the Bureau believed they represented a "potential" for violence—and nonviolent citizens who were against the war in Vietnam were targeted because they gave "aid and comfort" to violent demonstrators by lending respectability to their cause. The imprecision of the targeting is demonstrated by the inability of the Bureau to define the subjects of the programs. The Black Nationalist program, according to its supervisor, included "a great number of organizations that you might not today characterize as black nationalist but which were in fact primarily black". Thus, the nonviolent Southern Christian Leadership Conference was labeled as a Black Nationalist-"Hate Group". Furthermore, the actual targets were chosen from a far broader group than the titles of the programs would imply. The CPUSA program targeted not only Communist Party members but also sponsors of the National Committee to Abolish the House Un-American Activities Committee and civil rights leaders allegedly under Communist influence or deemed to be not sufficiently "anti-Communist". The Socialist Workers Party program included non-SWP sponsors of anti-war demonstrations which were cosponsored by the SWP or the Young Socialist Alliance, its youth group. The Black Nationalist program targeted a range of organizations from the Panthers to SNCC to the peaceful Southern Christian Leadership Conference, and included every Black Student Union and many other black student groups. New Left targets ranged from the SDS to the InterUniversity Committee for Debate on Foreign Policy, from Antioch College ("vanguard of the New Left") to the New Mexico Free University and other "alternate" schools, and from underground newspapers to students' protesting university censorship of a student publication by carrying signs with four-letter words on them. Examples of surveillance, spanning all presidents from FDR to Nixon, both legal and illegal, contained in the Church Committee report: Groups that were known to be targets of COINTELPRO operations include: The COINTELPRO operators targeted multiple groups at once and encouraged splintering of these groups from within. In letter-writing campaigns (wherein false letters were sent on behalf of members of parties), the FBI ensured that groups would not unite in their causes. For instance, they launched a campaign specifically to alienate the Black Panther Party from the Mau Maus, Young Lords, Young Patriots and SDS. These racially diverse groups had been building alliances, in part due to charismatic leaders such as Fred Hampton and his attempts to create a "Rainbow Coalition". The FBI was concerned with ensuring that groups could not gain traction through unity, specifically across racial lines. One of the main ways of targeting these groups was to arouse suspicion between the different parties and causes. In this way the bureau took on a divide and conquer offensive. The COINTELPRO documents show numerous cases of the FBI's intentions to prevent and disrupt protests against the Vietnam War. Many techniques were used to accomplish this task. "These included promoting splits among antiwar forces, encouraging red-baiting of socialists, and pushing violent confrontations as an alternative to massive, peaceful demonstrations." One 1966 COINTELPRO operation tried to redirect the Socialist Workers Party from their pledge of support for the antiwar movement. The FBI has said that it no longer undertakes COINTELPRO or COINTELPRO-like operations. However, critics have claimed that agency programs in the spirit of COINTELPRO targeted groups such as the Committee in Solidarity with the People of El Salvador, the American Indian Movement, Earth First!, and the anti-globalization movement. According to attorney Brian Glick in his book "War at Home", the FBI used five main methods during COINTELPRO: The FBI specifically developed tactics intended to heighten tension and hostility between various factions in the black power movement, for example between the Black Panthers and the US Organization. For instance, the FBI sent a fake letter to the US Organization exposing a supposed Black Panther plot to murder the head of the US Organization, Ron Karenga. They then intensified this by spreading falsely attributed cartoons in the black communities pitting the Black Panther Party against the US Organization. This resulted in numerous deaths, among which were San Diego Black Panther Party members John Huggins, Bunchy Carter and Sylvester Bell. Another example of the FBI's anonymous letter writing campaign is how they turned the Blackstone Rangers head, Jeff Fort, against former ally Fred Hampton, by stating that Hampton had a hit on Fort. They also were instrumental in developing the rift between Black Panther Party leaders Eldridge Cleaver and Huey Newton, as executed through false letters inciting the two leaders of the Black Panther Party. Dhoruba Bin Wahad, a former Black Panther, reflects on how these tactics made him feel, saying he had a combat mentality and felt like he was at war with the government. When asked about why he thinks the Black Panthers were targeted he said, "In the United States, the equivalent of the military was the local police. During the early sixties, at the height of the civil rights movement, and the human rights movement, the police in the United States became increasingly militaristic. They began to train out of military bases in the United States. The Law Enforcement Assistance Act supplied local police with military technology, everything from assault rifles to army personnel carriers. In his opinion, the Counterintelligence Program went hand-in-hand with the militarization of the police in the Black community, with the militarization of police in America." The FBI also conspired with the police departments of many U.S. cities (San Diego, Los Angeles, San Francisco, Oakland, Philadelphia, Chicago) to encourage repeated raids on Black Panther homes—often with little or no evidence of violations of federal, state, or local laws—which resulted directly in the police killing many members of the Black Panther Party, most notably Chicago Black Panther Party Chairman Fred Hampton on December 4, 1969. Before the death of Hampton, long-term infiltrator, William O'Neal, shared floor plans of his apartment with the COINTELPRO team. He then gave Hampton a dose of secobarbital that rendered Hampton unconscious during the raid on his home. In order to eliminate black militant leaders whom they considered dangerous, the FBI is believed to have worked with local police departments to target specific individuals, accuse them of crimes they did not commit, suppress exculpatory evidence and falsely incarcerate them. Elmer "Geronimo" Pratt, a Black Panther Party leader, was incarcerated for 27 years before a California Superior Court vacated his murder conviction, ultimately freeing him. Appearing before the court, an FBI agent testified that he believed Pratt had been framed, because both the FBI and the Los Angeles Police Department knew he had not been in the area at the time the murder occurred. Some sources claim that the FBI conducted more than 200 "black bag jobs", which were warrantless surreptitious entries, against the targeted groups and their members. In 1969 the FBI special agent in San Francisco wrote Hoover that his investigation of the Black Panther Party had concluded that in his city, at least, the Panthers were primarily engaged in feeding breakfast to children. Hoover fired back a memo implying the agent's career goals would be directly affected by his supplying evidence to support Hoover's view that the Black Panther Party was "a violence-prone organization seeking to overthrow the Government by revolutionary means". Hoover supported using false claims to attack his political enemies. In one memo he wrote: "Purpose of counterintelligence action is to disrupt the Black Panther Party and it is immaterial whether facts exist to substantiate the charge." In one particularly controversial 1965 incident, white civil rights worker Viola Liuzzo was murdered by Ku Klux Klansmen, who gave chase and fired shots into her car after noticing that her passenger was a young black man; one of the Klansmen was Gary Thomas Rowe, an acknowledged FBI informant. The FBI spread rumors that Liuzzo was a member of the Communist Party and had abandoned her children to have sexual relationships with African Americans involved in the civil rights movement. FBI records show that J. Edgar Hoover personally communicated these insinuations to President Johnson. FBI informant Rowe has also been implicated in some of the most violent crimes of the 1960s civil rights era, including attacks on the Freedom Riders and the 1963 Birmingham, Alabama 16th Street Baptist Church bombing. The FBI also financed, armed, and controlled an extreme right-wing group of former Minutemen, transforming it into a group called the Secret Army Organization that targeted groups, activists, and leaders involved in the Anti-War Movement, using both intimidation and violent acts. Hoover ordered preemptive action "to pinpoint potential troublemakers and neutralize them before they exercise their potential for violence." The final report of the Church Committee concluded: Too many people have been spied upon by too many Government agencies and too much information has been illegally collected. The Government has often undertaken the secret surveillance of citizens on the basis of their political beliefs, even when those beliefs posed no threat of violence or illegal acts on behalf of a hostile foreign power. The Government, operating primarily through secret and biased informants, but also using other intrusive techniques such as wiretaps, microphone "bugs", surreptitious mail opening, and break-ins, has swept in vast amounts of information about the personal lives, views, and associations of American citizens. Investigations of groups deemed potentially dangerous—and even of groups suspected of associating with potentially dangerous organizations—have continued for decades, despite the fact that those groups did not engage in unlawful activity. Groups and individuals have been assaulted, repressed, harassed and disrupted because of their political views, social beliefs and their lifestyles. Investigations have been based upon vague standards whose breadth made excessive collection inevitable. Unsavory, harmful and vicious tactics have been employed—including anonymous attempts to break up marriages, disrupt meetings, ostracize persons from their professions, and provoke target groups into rivalries that might result in deaths. Intelligence agencies have served the political and personal objectives of presidents and other high officials. While the agencies often committed excesses in response to pressure from high officials in the Executive branch and Congress, they also occasionally initiated improper activities and then concealed them from officials whom they had a duty to inform. Governmental officials—including those whose principal duty is to enforce the law—have violated or ignored the law over long periods of time and have advocated and defended their right to break the law. The Constitutional system of checks and balances has not adequately controlled intelligence activities. Until recently the Executive branch has neither delineated the scope of permissible activities nor established procedures for supervising intelligence agencies. Congress has failed to exercise sufficient oversight, seldom questioning the use to which its appropriations were being put. Most domestic intelligence issues have not reached the courts, and in those cases when they have reached the courts, the judiciary has been reluctant to grapple with them. While COINTELPRO was officially terminated in April 1971, domestic espionage continued. Between 1972 and 1974, it is documented that the Bureau planted over 500 bugs without a warrant and opened over 2,000 pieces of personal mail. More recent targets of covert action include the American Indian Movement (AIM), Earth First!, and Committees in Solidarity with the People of El Salvador. Documents released under the FOIA show that the FBI tracked the late David Halberstam—a Pulitzer Prize-winning journalist and author—for more than two decades. "Counterterrorism" guidelines implemented during the Reagan administration have been described as allowing a return to COINTELPRO tactics. Some radical groups accuse factional opponents of being FBI informants or assume the FBI is infiltrating the movement. COINTELPRO survivor Filiberto Ojeda Rios was killed by the FBI's hostage rescue team in 2005, his death described by a United Nations special committee as an assassination. Environmentalist Eric McDavid convicted on arson charges was released after documents emerged demonstrating that the FBI informant in his Earth Liberation Front group provided crucial leadership, information, and material without which the crime could not have been committed, repeating the same pattern of behaviour of COINTELPRO. It has been claimed these sorts of practices have become widespread in FBI counter-terrorism cases targeting left wing politics and Muslims in the 2009 Bronx terrorism plot and others. Authors such as Ward Churchill, Rex Weyler, and Peter Matthiessen allege that the federal government intended to acquire uranium deposits on the Lakota tribe's reservation land, and that this motivated a larger government conspiracy against AIM activists on the Pine Ridge reservation. Others believe COINTELPRO continues and similar actions are being taken against activist groups. Caroline Woidat says that, with respect to Native Americans, COINTELPRO should be understood within a historical context in which "Native Americans have been viewed and have viewed the world themselves through the lens of conspiracy theory." Other authors argue that while some conspiracy theories related to COINTELPRO are unfounded, the issue of ongoing government surveillance and repression is real. FBI Agent Richard G. Held is known to have increased FBI support for the Guardians of the Oglala Nation (GOON) squads accused of the assault and murder of hundreds of AIM supporters. The Bureau refused to investigate the 64 cases of homicide directly linked to GOON, but committed its resources overwhelmingly to prosecute AIM. In April 2018, the "Atlanta Black Star" characterized the FBI as still engaging in COINTELPRO behavior by surveilling the Black Lives Matter movement. Internal documents dated as late as 2017 showed that the FBI had surveilled the movement. In 2014, the FBI tracked a Black Lives Matter activist using surveillance tactics which "The Intercept" found "reminiscent of a rich American history of targeting black Americans," including COINTELPRO. This practice, along with the imprisonment of black activists for their views, has been associated with the new FBI designation of "Black Identity Extremists." Defending Rights & Dissent, a civil liberties group, cataloged known instances of First Amendment abuses and political surveillance by the FBI since 2010. The organization found that the feds devoted disproportionate resources to spy on peaceful left-leaning civil society groups, including Occupy Wall Street, economic justice advocates, racial justice movements, environmentalists, Abolish ICE, and various anti-war movements. During the effort to violently target Occupy Wall Street, the FBI and the DHS conducted their operations against activists in coordination with banks, the local police, and the New York Stock Exchange.
https://en.wikipedia.org/wiki?curid=6589
Cruise missile A cruise missile is a guided missile used against terrestrial targets, that remains in the atmosphere and flies the major portion of its flight path at approximately constant speed. Cruise missiles are designed to deliver a large warhead over long distances with high precision. Modern cruise missiles are capable of travelling at supersonic or high subsonic speeds, are self-navigating, and are able to fly on a non-ballistic, extremely low-altitude trajectory. The idea of an "aerial torpedo" was shown in the British 1909 film "The Airship Destroyer", where flying torpedoes controlled wirelessly are used to bring down airships bombing London. In 1916, American aviator Lawrence Sperry built and patented an "aerial torpedo", the Hewitt-Sperry Automatic Airplane, a small biplane carrying a TNT charge, a Sperry autopilot and a barometric altitude control. Inspired by these experiments, the United States Army developed a similar flying bomb called the Kettering Bug. Germany had also flown trials with remote-controlled aerial gliders "(Torpedogleiter)" built by Siemens-Schuckert beginning in 1916. In the period between the World Wars the UK developed the Larynx (Long Range Gun with Lynx Engine), which underwent a few flight tests in the 1920s. In the Soviet Union, Sergei Korolev headed the GIRD-06 cruise missile project from 1932 to 1939, which used a rocket-powered boost-glide bomb design. The 06/III (RP-216) and 06/IV (RP-212) contained gyroscopic guidance systems. The vehicle was designed to boost to 28 km altitude and glide a distance of 280 km, but test flights in 1934 and 1936 only reached an altitude of 500 meters. In 1944, Germany deployed the first operational cruise missiles in World War II. The V-1, often called a flying bomb, contained a gyroscope guidance system and was propelled by a simple pulsejet engine, the sound of which gave it the nickname of "buzz bomb" or "doodlebug". Accuracy was sufficient only for use against very large targets (the general area of a city), while the range of 250 km was significantly lower than that of a bomber carrying the same payload. The main advantages were speed (although not sufficient to outperform contemporary propeller-driven interceptors) and expendability. The production cost of a V-1 was only a small fraction of that of a V-2 supersonic ballistic missile, carrying a similar-sized warhead. Unlike the V-2, the initial deployments of the V-1 required stationary launch ramps which were susceptible to bombardment. Nazi Germany, in 1943, also developed the Mistel composite aircraft program, which can be seen as a rudimentary air-launched cruise missile, where a piloted fighter-type aircraft was mounted atop an unpiloted bomber-sized aircraft that was packed with explosives to be released while approaching the target. Bomber-launched variants of the V-1 saw limited operational service near the end of the war, with the pioneering V-1's design reverse-engineered by the Americans as the Republic-Ford JB-2 cruise missile. Immediately after the war the United States Air Force had 21 different guided missile projects, including would-be cruise missiles. All but four were cancelled by 1948, — the Air Materiel Command BANSHEE, the SM-62 Snark, the SM-64 Navaho, and the MGM-1 Matador. The BANSHEE design was similar to Operation Aphrodite; like Aphrodite, it failed, and was cancelled in April 1949. Concurrently, the US Navy's Operation: BUMBLEBEE, was conducted at Topsail Island, North Carolina, from c. 1 June 1946, to 28 July 1948. Operation: BUMBLEBEE produced proof-of-concept technologies that influenced the US military's other missile projects. During the Cold War period both the United States and the Soviet Union experimented further with the concept, deploying early cruise missiles from land, submarines and aircraft. The main outcome of the United States Navy submarine missile project was the SSM-N-8 Regulus missile, based upon the V-1. The United States Air Force's first operational surface-to-surface missile was the winged, mobile, nuclear-capable MGM-1 Matador, also similar in concept to the V-1. Deployment overseas began in 1954, first to West Germany and later to the Republic of China (Taiwan) and South Korea. On 7 November 1956, U.S. Air Force deployed Matador units in West Germany, whose missiles were capable of striking targets in the Warsaw Pact, from their fixed day-to-day sites to unannounced dispersed launch locations. This alert was in response to the crisis posed by the Soviet attack on Hungary which suppressed the Hungarian Revolution of 1956. Between 1957 and 1961 the United States followed an ambitious and well-funded program to develop a nuclear-powered cruise missile, Supersonic Low Altitude Missile (SLAM). It was designed to fly below the enemy's radar at speeds above Mach 3 and carry hydrogen bombs that it would drop along its path over enemy territory. Although the concept was proven sound and the 500 megawatt engine finished a successful test run in 1961, no airworthy device was ever completed. The project was finally abandoned in favor of ICBM development. While ballistic missiles were the preferred weapons for land targets, heavy nuclear and conventional weapon tipped cruise missiles were seen by the USSR as a primary weapon to destroy United States naval carrier battle groups. Large submarines (for example, Echo and Oscar classes) were developed to carry these weapons and shadow United States battle groups at sea, and large bombers (for example, Backfire, Bear, and Blackjack models) were equipped with the weapons in their air-launched cruise missile (ALCM) configuration. Cruise missiles generally consist of a guidance system, payload, and aircraft propulsion system, housed in an airframe with small wings and empennage for flight control. Payloads usually consist of a conventional warhead or a nuclear warhead. Cruise missiles tend to be propelled by jet engine, with turbofan engines in particular being preferred due to their greater efficiency at low altitude and subsonic speed. Guidance systems also vary greatly. Low-cost systems use a radar altimeter, barometric altimeter and clock to navigate a digital strip map. More advanced systems use inertial guidance, satellite guidance and terrain contour matching (TERCOM). Use of an automatic target recognition (ATR) algorithm/device in the guidance system increases accuracy of the missile. The Standoff Land Attack Missile features an ATR unit from General Electric. Cruise missiles can be categorized by size, speed (subsonic or supersonic), and range, and whether launched from land, air, surface ship, or submarine. Often versions of the same missile are produced for different launch platforms; sometimes air- and submarine-launched versions are a little lighter and smaller than land- and ship-launched versions. Guidance systems can vary across missiles. Some missiles can be fitted with any of a variety of navigation systems (Inertial navigation, TERCOM, or satellite navigation). Larger cruise missiles can carry either a conventional or a nuclear warhead, while smaller ones carry only conventional warheads. A hypersonic speed cruise missile would travel at least five times the speed of sound (Mach 5). These missiles travel faster than the speed of sound, usually using ramjet engines. The range is typically 100–500 km, but can be greater. Guidance systems vary. Examples: The United States, Russia, India, United Kingdom, Iran, South Korea, Israel, China and Pakistan have developed several long-range subsonic cruise missiles. These missiles have a range of over and fly at about . They typically have a launch weight of about and can carry either a conventional or a nuclear warhead. Earlier versions of these missiles used inertial navigation; later versions use much more accurate TERCOM and DSMAC systems. Most recent versions can use satellite navigation. Examples: These missiles are about the same size and weight and fly at similar speeds to the above category. Guidance systems vary. Examples: These are subsonic missiles which weigh around and have a range of up to . Examples: The most common mission for cruise missiles is to attack relatively high-value targets such as ships, command bunkers, bridges and dams. Modern guidance systems permit accurate attacks. The United States Air Force (USAF) deploys an air-launched cruise missile, the AGM-86 ALCM. The Boeing B-52 Stratofortress is the exclusive delivery vehicle for the AGM-86 and AGM-129 ACM. Both missile types are configurable for either conventional or nuclear warheads. The USAF adopted the AGM-86 for its bomber fleet while AGM-109 was adapted to launch from trucks and ships and adopted by the USAF and Navy. The truck-launched versions, and also the Pershing II and SS-20 Intermediate Range Ballistic Missiles, were later destroyed under the bilateral INF (Intermediate Range Nuclear Forces) treaty with the USSR. The British Royal Navy (RN) also operates cruise missiles, specifically the U.S.-made Tomahawk, used by the RN's nuclear submarine fleet. UK conventional warhead versions were first fired in combat by the RN in 1999, during the Kosovo War (the United States fired cruise missiles in 1991). The Royal Air Force uses the Storm Shadow cruise missile on its Tornado GR4 aircraft. It is also used by France, where it is known as SCALP EG, and carried by the Armée de l'Air's Mirage 2000 and Rafale aircraft. India and Russia have jointly developed the supersonic cruise missile BrahMos. There are three versions of the Brahmos: ship/land-launched, air-launched and sub-launched. The ship/land-launched version were operational as of late 2007. The Brahmos has the capability to attack targets on land. Russia also continues to operate other cruise missiles: the SS-N-12 Sandbox, SS-N-19 Shipwreck, SS-N-22 Sunburn and SS-N-25 Switchblade. Germany and Spain operate the Taurus missile while Pakistan has made the Babur missile, a variant of the US Tomahawk missile. Both the People's Republic of China and the Republic of China (Taiwan) have designed several cruise missile variants, such as the well-known C-802, some of which are capable of carrying biological, chemical, nuclear, and conventional warheads. China has CJ-10 land attack cruise missile which is capable of carrying a nuclear warhead. The French Force de Frappe nuclear forces include both land and sea-based bombers with Air-Sol Moyenne Portée high speed medium range nuclear cruise missiles. Two models are in use, ASMP and a newer ASMP-A. Approximately 60 nuclear missiles are in service, 50 land based and 10 sea-based. India in 2017 successfully flight-tested its indigenous Nirbhay ('Fearless') land-attack cruise missile, which can deliver nuclear warheads to a strike range of 1,000-km Nirbhay had been flight tested successfully. The Israel Defense Forces reportedly deploy the medium-range air-launched Popeye Turbo ALCM and the Popeye Turbo SLCM medium-long range cruise missile with nuclear warheads on Dolphin class submarines. Pakistan currently has four cruise missile systems: the air-launched Ra'ad; the ground and underwater launched Babur; ship-launched Harbah missile and surface launched Zarb missile. Both, Ra'ad and Babur, can carry nuclear warheads between 10 and 25 kt, and deliver them to targets at a range of and respectively. Babur has been in service with the Pakistan Army since 2010. Russia has Kh-55SM cruise missiles, with similar to United States' AGM-129 range of 3000 km, but are able to carry a more powerful warhead of 200 kt. They are equipped with a TERCOM system which allows them to cruise at an altitude lower than 110 meters at subsonic speeds while obtaining a CEP accuracy of 15 meters with an Inertial navigation system. They are air-launched from either Tupolev Tu-95s, Tupolev Tu-22Ms, or Tupolev Tu-160s, each able to carry 16 for the Tu-95, 12 for the Tu-160, and 4 for the Tu-22M. A stealth version of the missile, the Kh-101 is in development. It has similar qualities as the Kh-55, except that its range has been extended to 5,000 km, equipped with a 1,000 kg conventional warhead, and has stealth features which reduces its probability of intercept. In the late 1950s and early 1960s, the Soviet Union was attempting to develop cruise missiles. In this short time frame, the Soviet Union was working on nearly ten different types of cruise missiles. However, due to resources, most of the initial types of cruise missiles developed by the Soviet Union were Sea- Launched Cruise Missiles or Submarine-Launched Cruise Missiles (SLCMs). The SS-N-1 cruise missile was developed to have different configurations to be fired from a submarine or a ship. However, as the time progressed, the Soviet Union began to work on air launched cruise missiles as well (ALCM). These ACLM missiles were typically delivered via bombers designated as "Blinders" or "Backfire". The missiles in this configuration were called the AS-1, and AS-2 with eventual new variants with more development time. The main purpose of Soviet-based cruise missiles was to have defense and offensive mechanisms against enemy ships; in other words most of the Soviet cruise missiles were anti-ship missiles. the 1980s the Soviet Union had developed an arsenal of cruise missiles nearing 600 platforms which consisted of land, sea, and air delivery systems. The United States has deployed four nuclear cruise missiles at one time or another. Currently cruise missiles are among the most expensive of single-use weapons, up to several million dollars apiece. One consequence of this is that its users face difficult choices in target allocation, to avoid expending the missiles on targets of low value. For instance, during the 2001 strikes on Afghanistan the United States attacked targets of very low monetary value with cruise missiles, which led many to question the efficiency of the weapon. However, proponents of the cruise missile counter that the same argument applies to other types of UAVs: they are cheaper than human pilots when total training and infrastructure costs are taken into account, not to mention the risk of loss of personnel. As demonstrated in Libya in 2011 and prior conflicts, cruise missiles are much more difficult to detect and intercept than other aerial assets (reduced radar cross-section, infrared and visual signature due to smaller size), suiting them to attacks against static air defense systems.
https://en.wikipedia.org/wiki?curid=6590
Crete Crete (, Modern: , , Ancient: "," ";" Egyptian: 𓎡𓆑𓍘𓅱𓈉, "keftiu;" , , ) is the largest and most populous of the Greek islands, the 88th largest island in the world and the fifth largest island in the Mediterranean Sea, after Sicily, Sardinia, Cyprus, and Corsica. It bounds the southern border of the Aegean sea. Crete lies approximately south of the Greek mainland. It has an area of and a coastline of 1,046 km (650 mi). Crete and a number of surrounding islands and islets constitute the Region of Crete (), the southernmost of the 13 top-level administrative units of Greece; the region is the fifth most populous region of Greece. Its capital and largest city is Heraklion, located on the northern shore of the island. , the region had a population of 623,065. The Dodecanese are located to the northeast of Crete, while the Cyclades are situated to the north, separated by the Sea of Crete. The Peloponnese is to the region's northwest. Humans have inhabited the island since at least 130,000 years ago, during the Paleolithic age. Crete was the centre of Europe's first advanced civilization, the Minoans, from 2700 to 1420 BC; the Minoan civilization was overrun by the Mycenaean civilization from mainland Greece. Later, Crete fell under Roman rule, and afterwards the Byzantine Empire, Andalusian Arabs, the Venetian Republic, and the Ottoman Empire successively ruled Crete. The Cretan people, who maintained a desire to join the Greek state, achieved independence from the Ottomans in 1898 as the Cretan State and became part of Greece in December 1913. The island is mostly mountainous, and its character is defined by a high mountain range crossing from west to east; it includes Crete's highest point, Mount Ida, and the range of the White Mountains (Lefka Ori) with 30 summits above 2000 metres in altitude and the Samaria Gorge, a World Biosphere Reserve. Crete forms a significant part of the economy and cultural heritage of Greece, while retaining its own local cultural traits (such as its own poetry and music). The Nikos Kazantzakis airport at Heraklion and the Daskalogiannis airport at Chania serve international travelers. The palace of Knossos, a Bronze Age settlement and ancient Minoan city, lies in Heraklion in Crete. The earliest references to the island of Crete come from texts from the Syrian city of Mari dating from the 18th century BC, where the island is referred to as "Kaptara". This is repeated later in Neo-Assyrian records and the Bible ("Caphtor"). It was known in ancient Egyptian as "Keftiu" or "kftı͗w", strongly suggesting a similar Minoan name for the island. The current name "Crete" is first attested in the 15th century BC in Mycenaean Greek texts, written in Linear B, through the words "ke-re-te" (*"Krētes"; later Greek: , plural of ) and "ke-re-si-jo" (*"Krēsijos"; later Greek: , "Cretan"). In Ancient Greek, the name Crete () first appears in Homer's "Odyssey". In Latin, the name of the island became . The original Arabic name of Crete was ( < , but after the Emirate of Crete's establishment of its new capital at (modern Iraklion), both the city and the island became known as () or (), which gave Latin, Italian, and Venetian "Candia", from which were derived French "Candie" and English "Candy" or "Candia". Under Ottoman rule, in Ottoman Turkish, Crete was called (). Crete is the largest island in Greece and the fifth largest island in the Mediterranean Sea. It is located in the southern part of the Aegean Sea separating the Aegean from the Libyan Sea. The island has an elongated shape: it spans from east to west, is at its widest point, and narrows to as little as (close to Ierapetra). Crete covers an area of , with a coastline of ; to the north, it broaches the Sea of Crete (); to the south, the Libyan Sea (); in the west, the Myrtoan Sea, and toward the east the Karpathian Sea. It lies approximately south of the Greek mainland. Crete is mountainous, and its character is defined by a high mountain range crossing from west to east, formed by six different groups of mountains: These mountains lavish Crete with valleys, such as Amari valley, fertile plateaus, such as Lasithi plateau, Omalos and Nidha; caves, such as Gourgouthakas, Diktaion, and Idaion (the birthplace of the ancient Greek god Zeus); and a number of gorges. Mountains in Crete are the object of tremendous fascination both for locals and tourists. The mountains have been seen as a key feature of the island's distinctiveness especially since the time of Romantic travellers' writing. Contemporary Cretans distinguish between highlanders and lowlanders; the former often claim to reside in places affording a higher/better climatic but also moral environment. In keeping with the legacy of Romantic authors, the mountains are seen as having determined their residents' 'resistance' to past invaders which relates to the oft-encountered idea that highlanders are 'purer' in terms of less intermarriages with occupiers. For residents of mountainous areas, such as Sfakia in western Crete, the aridness and rockiness of the mountains is emphasised as an element of pride and is often compared to the alleged soft-soiled mountains of others parts of Greece or the world. The island has a number of gorges, such as the Samariá Gorge, Imbros Gorge, Kourtaliotiko Gorge, Ha Gorge, Platania Gorge, the Gorge of the Dead (at Kato Zakros, Sitia) and Richtis Gorge and (Richtis) waterfall at Exo Mouliana in Sitia. The rivers of Crete include the Ieropotamos River, the Koiliaris, the Anapodiaris, the Almiros, the Giofyros, and Megas Potamos. There are only two freshwater lakes in Crete: Lake Kournas and Lake Agia, which are both in Chania regional unit. Lake Voulismeni at the coast, at Aghios Nikolaos, was formerly a freshwater lake but is now connected to the sea, in Lasithi. Lakes that were created by dams also exist in Crete. There are three: the lake of Aposelemis Dam, the lake of Potamos Dam, and the lake of Mpramiana Dam. A large number of islands, islets, and rocks hug the coast of Crete. Many are visited by tourists, some are only visited by archaeologists and biologists. Some are environmentally protected. A small sample of the islands includes: Off the south coast, the island of Gavdos is located south of Hora Sfakion and is the southernmost point of Europe. Crete straddles two climatic zones, the Mediterranean and the North African, mainly falling within the former. As such, the climate in Crete is primarily Mediterranean. The atmosphere can be quite humid, depending on the proximity to the sea, while winter is fairly mild. Snowfall is common on the mountains between November and May, but rare in the low-lying areas. While some mountain tops are snow-capped for most of the year, near the coast snow only stays on the ground for a few minutes or hours. However, a truly exceptional cold snap swept the island in February 2004, during which period the whole island was blanketed with snow. During the Cretan summer, average temperatures reach the high 20s-low 30s Celsius (mid 80s to mid 90s Fahrenheit), with maxima touching the upper 30s-mid 40s. The south coast, including the Mesara Plain and Asterousia Mountains, falls in the North African climatic zone, and thus enjoys significantly more sunny days and high temperatures throughout the year. There, date palms bear fruit, and swallows remain year-round rather than migrate to Africa. The fertile region around Ierapetra, on the southeastern corner of the island, is renowned for its exceptional year-round agricultural production, with all kinds of summer vegetables and fruit produced in greenhouses throughout the winter. Western Crete (Chania province) receives more rain and is more erosive compared to the Eastern part of Crete. Crete is the most populous island in Greece with a population of more than 600,000 people. Approximately 42% live in Crete's main cities and towns whilst 45% live in rural areas. Crete with its nearby islands form the Crete Region (, , ), one of the 13 regions of Greece which were established in the 1987 administrative reform. Under the 2010 Kallikratis plan, the powers and authority of the regions were redefined and extended. The region is based at Heraklion and is divided into four regional units (pre-Kallikratis prefectures). From west to east these are: Chania, Rethymno, Heraklion, and Lasithi. These are further subdivided into 24 municipalities. The region's governor is, since 1 January 2011, Stavros Arnaoutakis, who was elected in the November 2010 local administration elections for the Panhellenic Socialist Movement. Heraklion is the largest city and capital of Crete. Chania was the capital until 1971. The principal cities are: The economy of Crete is predominantly based on services and tourism. However, agriculture also plays an important role and Crete is one of the few Greek islands that can support itself independently without a tourism industry. The economy began to change visibly during the 1970s as tourism gained in importance. Although an emphasis remains on agriculture and stock breeding, because of the climate and terrain of the island, there has been a drop in manufacturing, and an observable expansion in its service industries (mainly tourism-related). All three sectors of the Cretan economy (agriculture/farming, processing-packaging, services), are directly connected and interdependent. The island has a per capita income much higher than the Greek average, whereas unemployment is at approximately 4%, one-sixth of that of the country overall. As in many regions of Greece, viticulture and olive groves are significant; oranges, citrons and avocadoes are also cultivated. Until recently there were restrictions on the import of bananas to Greece, therefore bananas were grown on the island, predominantly in greenhouses. Dairy products are important to the local economy and there are a number of speciality cheeses such as mizithra, anthotyros, and kefalotyri. The Gross domestic product (GDP) of the region was 9.4 billion € in 2018, accounting for 5.1% of Greek economic output. GDP per capita adjusted for purchasing power was 17,800 € or 59% of the EU27 average in the same year. The GDP per employee was 68% of the EU average. Crete is the region in Greece with the fifth highest GDP per capita. The island has three significant airports, Nikos Kazantzakis at Heraklion, the Daskalogiannis airport at Chania and a smaller one in Sitia. The first two serve international routes, acting as the main gateways to the island for travellers. There is a long-standing plan to replace Heraklion airport with a completely new airport at Kastelli, where there is presently an air force base. The island is well served by ferries, mostly from Athens, by ferry companies such as Minoan Lines and ANEK Lines. Although the road network leads almost everywhere, there is a lack of modern highways, although this is gradually changing with the completion of the northern coastal spine highway. In addition, a European study has been devised from European Union to promote a modern highway that will connect the North and the South parts of the island via a tunnel. According to the study the project should be include 15.7 km of section of road between the villages Agia Varvara and Agia Deka in central Crete, benefits both tourists and local people by improving the accessibility to the southern part of the island and lessen the accidents. The new road section forms part of the route between Messara in the south and Crete's capital city Heraklion, which provides the island's airport and principal sea port link with mainland Greece. Traffic speeds on the new road will increase by 19 km/hour (from 29 km/hours to 48 km/hour), which should reduce journey times between Messara and Heraklion by 55 minutes. The scheme is also expected to improve road safety by cutting the number of accidents along the route. Building works include construction of three road tunnels, five bridges and three junctions. This project is expected to create 44 jobs during the implementation phase. The investment falls under Greece's "Improvement of Accessibility" Operational Programme. The programme aims to improve the country's transport infrastructures as well as its international connections. It will therefore have a key role to play in making Greece's remote and landlocked regions more accessible and economically attractive. This Operational Programme works to link Greece's more prosperous and less developed regions, which should help to promote greater territorial cohesion. Total investment for the project "Completion of construction of the section of Ag. Varvara - Ag. Deka (Kastelli) (22+170 km to 37+900 km) of the vertical road axis Irakleio – Messara in the prefecture of Irakleio, Kriti" is EUR 102 273 321, of which the EU's European Regional Development Fund is contributing EUR 86 932 323 from the Operational Programme "Improvement of Accessibility" for the 2007 to 2013 programming period. Work falls under the priority "Road Transport – trans-European and trans-regional route network of the regions on the Convergence objective". Also, during the 1930s there was a narrow-gauge industrial railway in Heraklion, from Giofyros in the west side of the city to the port. There are now no railway lines on Crete. The government is planning the construction of a line from Chania to Heraklion via Rethymno. Newspapers have reported that the Ministry of Mercantile Marine is ready to support the agreement between Greece, South Korea, Dubai Ports World and China for the construction of a large international container port and free trade zone in southern Crete near Tympaki; the plan is to expropriate 850 ha of land. The port would handle 2 million containers per year, but the project has not been universally welcomed because of its environmental, economic and cultural impact. As of January 2013, the project has still not been confirmed, although there is mounting pressure to approve it, arising from Greece's difficult economic situation. There are plans for underwater cables going from mainland Greece to Israel and Egypt passing by Crete and Cyprus: EuroAfrica Interconnector and EuroAsia Interconnector. They would connect Crete electrically with mainland Greece, ending energy isolation of Crete. Now Hellenic Republic covers for Crete electricity costs difference of around €300 million per year. Hominids settled in Crete at least 130,000 years ago. In the later Neolithic and Bronze Age periods, under the Minoans, Crete had a highly developed, literate civilization. It has been ruled by various ancient Greek entities, the Roman Empire, the Byzantine Empire, the Emirate of Crete, the Republic of Venice and the Ottoman Empire. After a brief period of independence (1897–1913) under a provisional Cretan government, it joined the Kingdom of Greece. It was occupied by Nazi Germany during the Second World War. In 2002, the paleontologist Gerard Gierlinski discovered fossil footprints left by ancient human relatives 5,600,000 years ago. The first human settlement in Crete dates before 130,000 years ago, during the Paleolithic age. Settlements dating to the aceramic Neolithic in the 7th millennium BC, used cattle, sheep, goats, pigs and dogs as well as domesticated cereals and legumes; ancient Knossos was the site of one of these major Neolithic (then later Minoan) sites. Other neolithic settlements include those at Kephala, Magasa, and Trapeza. Crete was the centre of Europe's first advanced civilization, the Minoan (). This civilization wrote in the undeciphered script known as Linear A. Early Cretan history is replete with legends such as those of King Minos, Theseus and the Minotaur, passed on orally via poets such as Homer. The volcanic eruption of Thera may have been the cause of the downfall of the Minoan civilization. In 1420 BC, the Minoan civilization was overrun by the Mycenaean civilization from mainland Greece. The oldest samples of writing in the Greek language, as identified by Michael Ventris, is the Linear B archive from Knossos, dated approximately to 1425–1375 BC. After the Bronze Age collapse, Crete was settled by new waves of Greeks from the mainland. A number of city states developed in the Archaic period. There was very limited contact with mainland Greece, and Greek historiography shows little interest in Crete, and as a result, there are very few literary sources. During the 6th to 4th centuries BC, Crete was comparatively free from warfare. The Gortyn code (5th century BC) is evidence for how codified civil law established a balance between aristocratic power and civil rights. In the late 4th century BC, the aristocratic order began to collapse due to endemic infighting among the elite, and Crete's economy was weakened by prolonged wars between city states. During the 3rd century BC, Gortyn, Kydonia (Chania), Lyttos and Polyrrhenia challenged the primacy of ancient Knossos. While the cities continued to prey upon one another, they invited into their feuds mainland powers like Macedon and its rivals Rhodes and Ptolemaic Egypt. In 220 BC the island was tormented by a war between two opposing coalitions of cities. As a result, the Macedonian king Philip V gained hegemony over Crete which lasted to the end of the Cretan War (205–200 BC), when the Rhodians opposed the rise of Macedon and the Romans started to interfere in Cretan affairs. In the 2nd century BC Ierapytna (Ierapetra) gained supremacy on eastern Crete. Crete was involved in the Mithridatic Wars, initially repelling an attack by Roman general Marcus Antonius Creticus in 71 BC. Nevertheless, a ferocious three-year campaign soon followed under Quintus Caecilius Metellus, equipped with three legions and Crete was finally conquered by Rome in 69 BC, earning for Metellus the title ""Creticus"". Gortyn was made capital of the island, and Crete became a Roman province, along with Cyrenaica that was called Creta et Cyrenaica. Archaeological remains suggest that Crete under Roman rule witnessed prosperity and increased connectivity with other parts of the Empire. In the 2nd century AD, at least three cities in Crete (Lyttos, Gortyn, Hierapytna) joined the Panhellenion, a league of Greek cities founded by the emperor Hadrian. When Diocletian redivided the Empire, Crete was placed, along with Cyrene, under the diocese of Moesia, and later by Constantine I to the diocese of Macedonia. Crete was separated from Cyrenaica . It remained a province within the eastern half of the Roman Empire, usually referred to as the Eastern Roman (Byzantine) Empire after the establishment of a second capital in Constantinople by Constantine in 330. Crete was subjected to an attack by Vandals in 467, the great earthquakes of 365 and 415, a raid by Slavs in 623, Arab raids in 654 and the 670s, and again in the 8th century. In , the Emperor Leo III the Isaurian transferred the island from the jurisdiction of the Pope to that of the Patriarchate of Constantinople. In the 820s, after 900 years as a Roman, and then Eastern Roman (Byzantine) island, Crete was captured by Andalusian Muladis led by Abu Hafs, who established the Emirate of Crete. The Byzantines launched a campaign that took most of the island back in 842 and 843 under Theoktistos. Further Byzantine campaigns in 911 and 949 failed. In 960/1, Nikephoros Phokas' campaign completely restored Crete to the Byzantine Empire, after a century and a half of Arab control. In 961, Nikephoros Phokas returned the island to Byzantine rule after expelling the Arabs. Extensive efforts at conversion of the populace were undertaken, led by John Xenos and Nikon "the Metanoeite". The reconquest of Crete was a major achievement for the Byzantines, as it restored Byzantine control over the Aegean littoral and diminished the threat of Saracen pirates, for which Crete had provided a base of operations. In 1204, the Fourth Crusade seized and sacked the imperial capital of Constantinople. Crete was initially granted to leading Crusader Boniface of Montferrat in the partition of spoils that followed. However, Boniface sold his claim to the Republic of Venice, whose forces made up the majority of the Crusade. Venice's rival the Republic of Genoa immediately seized the island and it was not until 1212 that Venice secured Crete as a colony. From 1212, during Venice's rule, which lasted more than four centuries, a Renaissance swept through the island as is evident from the plethora of artistic works dating to that period. Known as The Cretan School or Post-Byzantine Art, it is among the last flowerings of the artistic traditions of the fallen empire. The most notable representatives of this Cretan renaissance were the painter El Greco and the writers Nicholas Kalliakis (1645–1707), Georgios Kalafatis (professor) (–1720), Andreas Musalus (–1721) and Vitsentzos Kornaros. Under the rule of the Catholic Venetians, the city of Candia was reputed to be the best fortified city of the Eastern Mediterranean. The three main forts were located at Gramvousa, Spinalonga, and Fortezza at Rethymnon. Other fortifications include the Kazarma fortress at Sitia. In 1492, Jews expelled from Spain settled on the island. In 1574–77, Crete was under the rule of Giacomo Foscarini as Proveditor General, Sindace and Inquisitor. According to Starr's 1942 article, the rule of Giacomo Foscarini was a Dark Age for Jews and Greeks. Under his rule, non-Catholics had to pay high taxes with no allowances. In 1627, there were 800 Jews in the city of Candia, about seven percent of the city's population. Marco Foscarini was the Doge of Venice during this time period. The Ottomans conquered Crete in 1669, after the siege of Candia. Many Greek Cretans fled to other regions of the Republic of Venice after the Ottoman–Venetian Wars, some even prospering such as the family of Simone Stratigo (c. 1733 – c. 1824) who migrated to Dalmatia from Crete in 1669. Islamic presence on the island, aside from the interlude of the Arab occupation, was cemented by the Ottoman conquest. Most Cretan Muslims were local Greek converts who spoke Cretan Greek, but in the island's 19th-century political context they came to be viewed by the Christian population as Turks. Contemporary estimates vary, but on the eve of the Greek War of Independence (1830), as much as 45% of the population of the island may have been Muslim. A number of Sufi orders were widespread throughout the island, the Bektashi order being the most prevalent, possessing at least five tekkes. Many Cretan Turks fled Crete because of the unrest, settling in Turkey, Rhodes, Syria, Libya and elsewhere. By 1900, 11% of the population was Muslim. Those remaining were relocated in the 1924 population exchange between Greece and Turkey. During Easter of 1770, a notable revolt against Ottoman rule, in Crete, was started by Daskalogiannis, a shipowner from Sfakia who was promised support by Orlov's fleet which never arrived. Daskalogiannis eventually surrendered to the Ottoman authorities. Today, the airport at Chania is named after him. Crete was left out of the modern Greek state by the London Protocol of 1830, and soon it was yielded to Egypt by the Ottoman sultan. Egyptian rule was short-lived and sovereignty was returned to the Ottoman Empire by the Convention of London on 3 July 1840. Heraklion was surrounded by high walls and bastions and extended westward and southward by the 17th century. The most opulent area of the city was the northeastern quadrant where all the elite were gathered together. The city had received another name under the rule of the Ottomans, "the deserted city". The urban policy that the Ottoman applied to Candia was a two-pronged approach. The first was the religious endowments. It made the Ottoman elite contribute to building and rehabilitating the ruined city. The other method was to boost the population and the urban revenue by selling off urban properties. According to Molly Greene (2001) there were numerous records of real-estate transactions during the Ottoman rule. In the deserted city, minorities received equal rights in purchasing property. Christians and Jews were also able to buy and sell in the real-estate market. The Cretan Revolt of 1866–1869 or Great Cretan Revolution () was a three-year uprising against Ottoman rule, the third and largest in a series of revolts between the end of the Greek War of Independence in 1830 and the establishment of the independent Cretan State in 1898. A particular event which caused strong reactions among the liberal circles of western Europe was the "Holocaust of Arkadi". The event occurred in November 1866, as a large Ottoman force besieged the Arkadi Monastery, which served as the headquarters of the rebellion. In addition to its 259 defenders, over 700 women and children had taken refuge in the monastery. After a few days of hard fighting, the Ottomans broke into the monastery. At that point, the abbot of the monastery set fire to the gunpowder stored in the monastery's vaults, causing the death of most of the rebels and the women and children sheltered there. Following the repeated uprisings in 1841, 1858, 1889, 1895 and 1897 by the Cretan people, who wanted to join Greece, the Great Powers decided to restore order and in February 1897 sent in troops. The island was subsequently garrisoned by troops from Great Britain, France, Italy and Russia; Germany and Austro-Hungary withdrawing from the occupation in early 1898. During this period Crete was governed through a committee of admirals from the remaining four Powers. In March 1898 the Powers decreed, with the very reluctant consent of the Sultan, that the island would be granted autonomy under Ottoman suzerainty in the near future. In September 1898 an outbreak of rioting in Candia, modern Heraklion, left over 500 Cretan Christians, and 14 British servicemen, dead. As a result, the Admirals ordered the expulsion of all Ottoman troops and administrators from the island, a move that was ultimately completed by early November. The decision to grant autonomy to the island was enforced and a High Commissioner, Prince George of Greece, appointed, arriving to take up his post in December 1898. The flag of the Cretan State was chosen by the Powers, with the white star representing the Ottoman suzerainty over the island. In 1905, disagreements between Prince George and minister Eleftherios Venizelos over the question of the "enosis" (union with Greece), such as the Prince's autocratic style of government, resulted in the Theriso revolt, one of the leaders being Eleftherios Venizelos. Prince George resigned as High Commissioner and was replaced by Alexandros Zaimis, a former Greek prime minister, in 1906. In 1908, taking advantage of domestic turmoil in Turkey as well as the timing of Zaimis's vacation away from the island, the Cretan deputies unilaterally declared union with Greece. With the break out of the First Balkan War, the Greek government declared that Crete was now Greek territory. This was not recognised internationally until 1 December 1913. During World War II, the island was the scene of the famous Battle of Crete in May 1941. The initial 11-day battle was bloody and left more than 11,000 soldiers and civilians killed or wounded. As a result of the fierce resistance from both Allied forces and civilian Cretan locals, the invasion force suffered heavy casualties, and Adolf Hitler forbade further large-scale paratroop operations for the rest of the war. During the initial and subsequent occupation, German firing squads routinely executed male civilians in reprisal for the death of German soldiers; civilians were rounded up randomly in local villages for the mass killings, such as at the Massacre of Kondomari and the Viannos massacres. Two German generals were later tried and executed for their roles in the killing of 3,000 of the island's inhabitants. Crete was one of the most popular holiday destinations in Greece. 15% of all arrivals in Greece come through the city of Heraklion (port and airport), while charter journeys to Heraklion seven years ago made up 20% of all charter flights in Greece. Overall, more than two million tourists visited Crete some years back, when the increase in tourism was reflected in the number of hotel beds, rising by 53% in the period between 1986 and 1991. Today, the island's tourism infrastructure caters to all tastes, including a very wide range of accommodation; the island's facilities take in large luxury hotels with their complete facilities, swimming pools, sports and recreation, smaller family-owned apartments, camping facilities and others. Visitors reach the island via two international airports in Heraklion and Chania and a smaller airport in Sitia (international charter and domestic flights starting May 2012) or by boat to the main ports of Heraklion, Chania, Rethimno, Agios Nikolaos and Sitia. Popular tourist attractions include the archaeological sites of the Minoan civilisation, the Venetian old city and port of Chania, the Venetian castle at Rethymno, the gorge of Samaria, the islands of Chrysi, Elafonisi, Gramvousa, Spinalonga and the Palm Beach of Vai, which is the largest natural palm forest in Europe. Crete has an extensive bus system with regular services across the north of the island and from north to south. There are two regional bus stations in Heraklion. Bus routes and timetables can be found on KTEL website. Crete's mild climate attracts interest from northern Europeans who want a holiday home or residence on the island. EU citizens have the right to freely buy property and reside with little formality. A growing number of real estate companies cater to mainly British immigrants, followed by German, Dutch, Scandinavian and other European nationalities wishing to own a home in Crete. The British immigrants are concentrated in the western regional units of Chania and Rethymno and to a lesser extent in Heraklion and Lasithi. The area has a large number of archaeological sites, including the Minoan sites of Knossos, Malia (not to be confused with the town of the same name), Petras and Phaistos, the classical site of Gortys, and the diverse archaeology of the island of Koufonisi, which includes Minoan, Roman, and World War II era ruins (nb. due to conservation concerns, access to the latter has been restricted for the last few years, so it is best to check before heading to a port). There are a number of museums throughout Crete. The Heraklion Archaeological Museum displays most of the archaeological finds from the Minoan era and was reopened in 2014. Helen Briassoulis proposed in the Journal of Sustainable Tourism that Crete is a victim of external tourist systems applying pressure to it to develop at an unhealthy rate, and that informal, internal systems within the country are forced to adapt. According to her, these forces have strengthened in 3 stages: from the period from 1960–1970, 1970–1990, and 1990 to the present. During this first period, tourism was a largely positive force, pushing modern developments like running water and electricity onto the largely rural countryside. However, beginning in the second period and especially in the third period leading up to the present day, tourist companies became more pushy with deforestation and pollution of Crete's natural resources. The country is then pulled into an interesting parity, where these companies only upkeep those natural resources that are directly essential to their industry. Crete is isolated from mainland Europe, Asia, and Africa, and this is reflected in the diversity of the fauna and flora. As a result, the fauna and flora of Crete have many clues to the evolution of species. There are no animals that are dangerous to humans on the island of Crete in contrast to other parts of Greece. Indeed, the ancient Greeks attributed the lack of large mammals such as bears, wolves, jackals, and venomous snakes, to the labour of Hercules (who took a live Cretan bull to the Peloponnese). Hercules wanted to honor the birthplace of Zeus by removing all "harmful" and "venomous" animals from Crete. Later, Cretans believed that the island was cleared of dangerous creatures by the Apostle Paul, who lived on the island of Crete for two years, with his exorcisms and blessings. There is a natural history museum, the Natural History Museum of Crete, operating under the direction of the University of Crete and two aquariums – Aquaworld in Hersonissos and Cretaquarium in Gournes, displaying sea creatures common in Cretan waters. Dwarf elephants, dwarf hippopotamus, dwarf mammoths, dwarf deer, and giant flightless owls were native to Pleistocene Crete. Mammals of Crete include the vulnerable kri-kri, "Capra aegagrus cretica" that can be seen in the national park of the Samaria Gorge and on Thodorou, Dia and Agioi Pantes (islets off the north coast), the Cretan wildcat and the Cretan spiny mouse. Other terrestrial mammals include subspecies of the Cretan marten, the Cretan weasel, the Cretan badger, the long-eared hedgehog, and the edible dormouse. The Cretan shrew, a type of white-toothed shrew is considered endemic to the island of Crete because this species of shrew is unknown elsewhere. It is a relic species of the "crocidura" shrews of which fossils have been found that can be dated to the Pleistocene era. In the present day it can only be found in the highlands of Crete. It is considered to be the only surviving remnant of the endemic species of the Pleistocene Mediterranean islands. Bat species include: Blasius's horseshoe bat, the lesser horseshoe bat, the greater horseshoe bat, the lesser mouse-eared bat, Geoffroy's bat, the whiskered bat, Kuhl's pipistrelle, the common pipistrelle, Savi's pipistrelle, the serotine bat, the long-eared bat, Schreibers' bat and the European free-tailed bat. A large variety of birds includes eagles (can be seen in Lasithi), swallows (throughout Crete in the summer and all the year in the south of the island), pelicans (along the coast), and cranes (including Gavdos and Gavdopoula). The Cretan mountains and gorges are refuges for the endangered lammergeier vulture. Bird species include: the golden eagle, Bonelli's eagle, the bearded vulture or lammergeier, the griffon vulture, Eleanora's falcon, peregrine falcon, lanner falcon, European kestrel, tawny owl, little owl, hooded crow, alpine chough, red-billed chough, and the hoopoe. Tortoises can be seen throughout the island. Snakes can be found hiding under rocks. Toads and frogs reveal themselves when it rains. Reptiles include the Aegean wall lizard, Balkan green lizard, common chameleon, ocellated skink, snake-eyed skink, moorish gecko, Turkish gecko, Kotschy's gecko, spur-thighed tortoise, and the Caspian turtle. There are four species of snake on the island and these are not dangerous to humans. The four species include the leopard snake (locally known as Ochendra), the Balkan whip snake (locally called Dendrogallia), the dice snake (called Nerofido in Greek), and the only venomous snake is the nocturnal cat snake which has evolved to deliver a weak venom at the back of its mouth to paralyse geckos and small lizards, and is not dangerous to humans. Sea turtles include the green turtle and the loggerhead turtle which are both threatened species. The loggerhead turtle nests and hatches on north-coast beaches around Rethymno and Chania, and south-coast beaches along the gulf of Mesara. Amphibians include the European green toad, American bullfrog (introduced), European tree frog, and the Cretan marsh frog (endemic). Crete has an unusual variety of insects. Cicadas, known locally as "Tzitzikia", make a distinctive repetitive "tzi tzi" sound that becomes louder and more frequent on hot summer days. Butterfly species include the swallowtail butterfly. Moth species include the hummingbird moth. There are several species of scorpion such as Euscorpius carpathicus whose venom is generally no more potent than a mosquito bite. River crabs include the semi-terrestrial "Potamon potamios" crab. Edible snails are widespread and can cluster in the hundreds waiting for rainfall to reinvigorate them. Apart from terrestrial mammals, the seas around Crete are rich in large marine mammals, a fact unknown to most Greeks at present, although reported since ancient times. Indeed, the Minoan frescoes depicting dolphins in Queen's Megaron at Knossos indicate that Minoans were well aware of and celebrated these creatures. Apart from the famous endangered Mediterranean monk seal, which lives in almost all the coasts of the country, Greece hosts whales, sperm whales, dolphins and porpoises. These are either permanent residents of the Mediterranean or just occasional visitors. The area south of Crete, known as the Greek Abyss, hosts many of them. Squid and octopus can be found along the coast and sea turtles and hammerhead sharks swim in the sea around the coast. The Cretaquarium and the Aquaworld Aquarium, are two of only three aquariums in the whole of Greece. They are located in Gournes and Hersonissos respectively. Examples of the local sealife can be seen there. Some of the fish that can be seen in the waters around Crete include: scorpion fish, dusky grouper, east Atlantic peacock wrasse, five-spotted wrasse, weever fish, common stingray, brown ray, mediterranean black goby, pearly razorfish, star-gazer, painted comber, damselfish, and the flying gurnard. The Minoans contributed to the deforestation of Crete. Further deforestation occurred in the 1600s "so that no more local supplies of firewood were available". Common wildflowers include: camomile, daisy, gladiolus, hyacinth, iris, poppy, cyclamen and tulip, among others. There are more than 200 different species of wild orchid on the island and this includes 14 varieties of "Ophrys cretica". Crete has a rich variety of indigenous herbs including common sage, rosemary, thyme, and oregano. Rare herbs include the endemic Cretan dittany. and ironwort, "Sideritis syriaca", known as Malotira (Μαλοτήρα). Varieties of cactus include the edible prickly pear. Common trees on the island include the chestnut, cypress, oak, olive tree, pine, plane, and tamarisk. Trees tend to be taller to the west of the island where water is more abundant. There are a number of environmentally protected areas. One such area is located at the island of Elafonisi on the coast of southwestern Crete. Also, the palm forest of Vai in eastern Crete and the Dionysades (both in the municipality of Sitia, Lasithi), have diverse animal and plant life. Vai has a palm beach and is the largest natural palm forest in Europe. The island of Chrysi, south of Ierapetra, has the largest naturally-grown "Juniperus macrocarpa" forest in Europe. Samaria Gorge is a World Biosphere Reserve and Richtis Gorge is protected for its landscape diversity. Crete has a rich mythology mostly connected with the ancient Greek Gods but also connected with the Minoan civilisation. According to Greek Mythology, The Diktaean Cave at Mount Dikti was the birthplace of the god Zeus. The Paximadia islands were the birthplace of the goddess Artemis and the god Apollo. Their mother, the goddess Leto, was worshipped at Phaistos. The goddess Athena bathed in Lake Voulismeni. The ancient Greek god Zeus launched a lightning bolt at a giant lizard that was threatening Crete. The lizard immediately turned to stone and became the island of Dia. The island can be seen from Knossos and it has the shape of a giant lizard. The islets of Lefkai were the result of a musical contest between the Sirens and the Muses. The Muses were so anguished to have lost that they plucked the feathers from the wings of their rivals; the Sirens turned white and fell into the sea at Aptera ("featherless") where they formed the islands in the bay that were called Lefkai (the islands of Souda and Leon). Heracles, in one of his labors, took the Cretan bull to the Peloponnese. Europa and Zeus made love at Gortys and conceived the kings of Crete: Rhadamanthys, Sarpedon, and Minos. The labyrinth of the Palace of Knossos was the setting for the myth of Theseus and the Minotaur in which the Minotaur was slain by Theseus. Icarus and Daedalus were captives of King Minos and crafted wings to escape. After his death King Minos became a judge of the dead in Hades, while Rhadamanthys became the ruler of the Elysian fields. Crete has its own distinctive Mantinades poetry. The island is known for its Mantinades-based music (typically performed with the Cretan lyra and the laouto) and has many indigenous dances, the most noted of which is the Pentozali. Since the 1980s and certainly in the 90s onwards there has been a proliferation of Cultural Associations that teach dancing (in Western Crete many focus on rizitiko singing). These Associations often perform in official events but also become stages for people to meet up and engage in traditionalist practices. The topic of tradition and the role of Cultural Associations in reviving it is very often debated throughout Crete. Cretan authors have made important contributions to Greek literature throughout the modern period; major names include Vikentios Kornaros, creator of the 17th-century epic romance "Erotokritos" (Greek Ερωτόκριτος), and, in the 20th century, Nikos Kazantzakis. In the Renaissance, Crete was the home of the Cretan School of icon painting, which influenced El Greco and through him subsequent European painting. Crete is also famous for its traditional cuisine. The nutritional value of the Cretan cuisine was discovered by the American epidemiologist Ancel Keys in the 1960, being later often mentioned by epidemiologists as one of the best examples of the Mediterranean diet. Cretans are fiercely proud of their island and customs, and men often don elements of traditional dress in everyday life: knee-high black riding boots ("stivania"), "vráka" breeches tucked into the boots at the knee, black shirt and black headdress consisting of a fishnet-weave kerchief worn wrapped around the head or draped on the shoulders ("sariki"). Men often grow large mustaches as a mark of masculinity. Cretan society is known in Greece and internationally for family and clan vendettas which persist on the island to date. Cretans also have a tradition of keeping firearms at home, a tradition lasting from the era of resistance against the Ottoman Empire. Nearly every rural household on Crete has at least one unregistered gun. Guns are subject to strict regulation from the Greek government, and in recent years a great deal of effort to control firearms in Crete has been undertaken by the Greek police, but with limited success. Crete has many football clubs playing in the local leagues. During the 2011–12 season, OFI Crete, which plays at Theodoros Vardinogiannis Stadium (Iraklion), and Ergotelis F.C., which plays at the Pankritio Stadium (Iraklion) were both members of the Greek Superleague. During the 2012–13 season, OFI Crete, which plays at Theodoros Vardinogiannis Stadium (Iraklion), and Platanias F.C., which plays at the Perivolia Municipal Stadium, near Chania, are both members of the Greek Superleague. Notable people from Crete include:
https://en.wikipedia.org/wiki?curid=6591
Computer vision Computer vision is an interdisciplinary scientific field that deals with how computers can gain high-level understanding from digital images or videos. From the perspective of engineering, it seeks to understand and automate tasks that the human visual system can do. Computer vision tasks include methods for acquiring, processing, analyzing and understanding digital images, and extraction of high-dimensional data from the real world in order to produce numerical or symbolic information, e.g. in the forms of decisions. Understanding in this context means the transformation of visual images (the input of the retina) into descriptions of the world that make sense to thought processes and can elicit appropriate action. This image understanding can be seen as the disentangling of symbolic information from image data using models constructed with the aid of geometry, physics, statistics, and learning theory. The scientific discipline of computer vision is concerned with the theory behind artificial systems that extract information from images. The image data can take many forms, such as video sequences, views from multiple cameras, multi-dimensional data from a 3D scanner or medical scanning device. The technological discipline of computer vision seeks to apply its theories and models to the construction of computer vision systems. Sub-domains of computer vision include scene reconstruction, event detection, video tracking, object recognition, 3D pose estimation, learning, indexing, motion estimation, visual servoing, 3D scene modeling, and image restoration. Computer vision is an interdisciplinary field that deals with how computers can be made to gain high-level understanding from digital images or videos. From the perspective of engineering, it seeks to automate tasks that the human visual system can do. "Computer vision is concerned with the automatic extraction, analysis and understanding of useful information from a single image or a sequence of images. It involves the development of a theoretical and algorithmic basis to achieve automatic visual understanding." As a scientific discipline, computer vision is concerned with the theory behind artificial systems that extract information from images. The image data can take many forms, such as video sequences, views from multiple cameras, or multi-dimensional data from a medical scanner. As a technological discipline, computer vision seeks to apply its theories and models for the construction of computer vision systems. In the late 1960s, computer vision began at universities which were pioneering artificial intelligence. It was meant to mimic the human visual system, as a stepping stone to endowing robots with intelligent behavior. In 1966, it was believed that this could be achieved through a summer project, by attaching a camera to a computer and having it "describe what it saw". What distinguished computer vision from the prevalent field of digital image processing at that time was a desire to extract three-dimensional structure from images with the goal of achieving full scene understanding. Studies in the 1970s formed the early foundations for many of the computer vision algorithms that exist today, including extraction of edges from images, labeling of lines, non-polyhedral and polyhedral modeling, representation of objects as interconnections of smaller structures, optical flow, and motion estimation. The next decade saw studies based on more rigorous mathematical analysis and quantitative aspects of computer vision. These include the concept of scale-space, the inference of shape from various cues such as shading, texture and focus, and contour models known as snakes. Researchers also realized that many of these mathematical concepts could be treated within the same optimization framework as regularization and Markov random fields. By the 1990s, some of the previous research topics became more active than the others. Research in projective 3-D reconstructions led to better understanding of camera calibration. With the advent of optimization methods for camera calibration, it was realized that a lot of the ideas were already explored in bundle adjustment theory from the field of photogrammetry. This led to methods for sparse 3-D reconstructions of scenes from multiple images. Progress was made on the dense stereo correspondence problem and further multi-view stereo techniques. At the same time, variations of graph cut were used to solve image segmentation. This decade also marked the first time statistical learning techniques were used in practice to recognize faces in images (see Eigenface). Toward the end of the 1990s, a significant change came about with the increased interaction between the fields of computer graphics and computer vision. This included image-based rendering, image morphing, view interpolation, panoramic image stitching and early light-field rendering. Recent work has seen the resurgence of feature-based methods, used in conjunction with machine learning techniques and complex optimization frameworks. The advancement of Deep Learning techniques has brought further life to the field of computer vision. The accuracy of deep learning algorithms on several benchmark computer vision data sets for tasks ranging from classification, segmentation and optical flow has surpassed prior methods. Areas of artificial intelligence deal with autonomous planning or deliberation for robotic systems to navigate through an environment. A detailed understanding of these environments is required to navigate through them. Information about the environment could be provided by a computer vision system, acting as a vision sensor and providing high-level information about the environment and the robot. Artificial intelligence and computer vision share other topics such as pattern recognition and learning techniques. Consequently, computer vision is sometimes seen as a part of the artificial intelligence field or the computer science field in general. Computer vision is often considered to be part of information engineering. Solid-state physics is another field that is closely related to computer vision. Most computer vision systems rely on image sensors, which detect electromagnetic radiation, which is typically in the form of either visible or infra-red light. The sensors are designed using quantum physics. The process by which light interacts with surfaces is explained using physics. Physics explains the behavior of optics which are a core part of most imaging systems. Sophisticated image sensors even require quantum mechanics to provide a complete understanding of the image formation process. Also, various measurement problems in physics can be addressed using computer vision, for example motion in fluids. A third field which plays an important role is neurobiology, specifically the study of the biological vision system. Over the last century, there has been an extensive study of eyes, neurons, and the brain structures devoted to processing of visual stimuli in both humans and various animals. This has led to a coarse, yet complicated, description of how "real" vision systems operate in order to solve certain vision-related tasks. These results have led to a sub-field within computer vision where artificial systems are designed to mimic the processing and behavior of biological systems, at different levels of complexity. Also, some of the learning-based methods developed within computer vision ("e.g." neural net and deep learning based image and feature analysis and classification) have their background in biology. Some strands of computer vision research are closely related to the study of biological vision – indeed, just as many strands of AI research are closely tied with research into human consciousness, and the use of stored knowledge to interpret, integrate and utilize visual information. The field of biological vision studies and models the physiological processes behind visual perception in humans and other animals. Computer vision, on the other hand, studies and describes the processes implemented in software and hardware behind artificial vision systems. Interdisciplinary exchange between biological and computer vision has proven fruitful for both fields. Yet another field related to computer vision is signal processing. Many methods for processing of one-variable signals, typically temporal signals, can be extended in a natural way to processing of two-variable signals or multi-variable signals in computer vision. However, because of the specific nature of images there are many methods developed within computer vision which have no counterpart in processing of one-variable signals. Together with the multi-dimensionality of the signal, this defines a subfield in signal processing as a part of computer vision. Beside the above-mentioned views on computer vision, many of the related research topics can also be studied from a purely mathematical point of view. For example, many methods in computer vision are based on statistics, optimization or geometry. Finally, a significant part of the field is devoted to the implementation aspect of computer vision; how existing methods can be realized in various combinations of software and hardware, or how these methods can be modified in order to gain processing speed without losing too much performance. Computer vision is also used in fashion ecommerce, inventory management, patent search, furniture, and the beauty industry. The fields most closely related to computer vision are image processing, image analysis and machine vision. There is a significant overlap in the range of techniques and applications that these cover. This implies that the basic techniques that are used and developed in these fields are similar, something which can be interpreted as there is only one field with different names. On the other hand, it appears to be necessary for research groups, scientific journals, conferences and companies to present or market themselves as belonging specifically to one of these fields and, hence, various characterizations which distinguish each of the fields from the others have been presented. Computer graphics produces image data from 3D models, computer vision often produces 3D models from image data. There is also a trend towards a combination of the two disciplines, "e.g.", as explored in augmented reality. The following characterizations appear relevant but should not be taken as universally accepted:: Photogrammetry also overlaps with computer vision, e.g., stereophotogrammetry vs. computer stereo vision. Applications range from tasks such as industrial machine vision systems which, say, inspect bottles speeding by on a production line, to research into artificial intelligence and computers or robots that can comprehend the world around them. The computer vision and machine vision fields have significant overlap. Computer vision covers the core technology of automated image analysis which is used in many fields. Machine vision usually refers to a process of combining automated image analysis with other methods and technologies to provide automated inspection and robot guidance in industrial applications. In many computer-vision applications, the computers are pre-programmed to solve a particular task, but methods based on learning are now becoming increasingly common. Examples of applications of computer vision include systems for: One of the most prominent application fields is medical computer vision, or medical image processing, characterized by the extraction of information from image data to diagnose a patient. An example of this is detection of tumours, arteriosclerosis or other malign changes; measurements of organ dimensions, blood flow, etc. are another example. It also supports medical research by providing new information: "e.g.", about the structure of the brain, or about the quality of medical treatments. Applications of computer vision in the medical area also includes enhancement of images interpreted by humans—ultrasonic images or X-ray images for example—to reduce the influence of noise. A second application area in computer vision is in industry, sometimes called machine vision, where information is extracted for the purpose of supporting a manufacturing process. One example is quality control where details or final products are being automatically inspected in order to find defects. Another example is measurement of position and orientation of details to be picked up by a robot arm. Machine vision is also heavily used in agricultural process to remove undesirable food stuff from bulk material, a process called optical sorting. Military applications are probably one of the largest areas for computer vision. The obvious examples are detection of enemy soldiers or vehicles and missile guidance. More advanced systems for missile guidance send the missile to an area rather than a specific target, and target selection is made when the missile reaches the area based on locally acquired image data. Modern military concepts, such as "battlefield awareness", imply that various sensors, including image sensors, provide a rich set of information about a combat scene which can be used to support strategic decisions. In this case, automatic processing of the data is used to reduce complexity and to fuse information from multiple sensors to increase reliability. One of the newer application areas is autonomous vehicles, which include submersibles, land-based vehicles (small robots with wheels, cars or trucks), aerial vehicles, and unmanned aerial vehicles (UAV). The level of autonomy ranges from fully autonomous (unmanned) vehicles to vehicles where computer-vision-based systems support a driver or a pilot in various situations. Fully autonomous vehicles typically use computer vision for navigation, "e.g." for knowing where it is, or for producing a map of its environment (SLAM) and for detecting obstacles. It can also be used for detecting certain task specific events, "e.g.", a UAV looking for forest fires. Examples of supporting systems are obstacle warning systems in cars, and systems for autonomous landing of aircraft. Several car manufacturers have demonstrated systems for autonomous driving of cars, but this technology has still not reached a level where it can be put on the market. There are ample examples of military autonomous vehicles ranging from advanced missiles to UAVs for recon missions or missile guidance. Space exploration is already being made with autonomous vehicles using computer vision, "e.g.", NASA's "Curiosity" and CNSA's "Yutu-2" rover. Materials such as rubber and silicon are being used to create sensors that allow for applications such as detecting micro undulations and calibrating robotic hands. Rubber can be used in order to create a mold that can be placed over a finger, inside of this mold would be multiple strain gauges. The finger mold and sensors could then be placed on top of a small sheet of rubber containing an array of rubber pins. A user can then wear the finger mold and trace a surface. A computer can then read the data from the strain gauges and measure if one or more of the pins is being pushed upward. If a pin is being pushed upward then the computer can recognize this as an imperfection in the surface. This sort of technology is useful in order to receive accurate data of the imperfections on a very large surface. Another variation of this finger mold sensor are sensors that contain a camera suspended in silicon. The silicon forms a dome around the outside of the camera and embedded in the silicon are point markers that are equally spaced. These cameras can then be placed on devices such as robotic hands in order to allow the computer to receive highly accurate tactile data. Other application areas include: Each of the application areas described above employ a range of computer vision tasks; more or less well-defined measurement problems or processing problems, which can be solved using a variety of methods. Some examples of typical computer vision tasks are presented below. Computer vision tasks include methods for acquiring, processing, analyzing and understanding digital images, and extraction of high-dimensional data from the real world in order to produce numerical or symbolic information, "e.g.", in the forms of decisions. Understanding in this context means the transformation of visual images (the input of the retina) into descriptions of the world that can interface with other thought processes and elicit appropriate action. This image understanding can be seen as the disentangling of symbolic information from image data using models constructed with the aid of geometry, physics, statistics, and learning theory. The classical problem in computer vision, image processing, and machine vision is that of determining whether or not the image data contains some specific object, feature, or activity. Different varieties of the recognition problem are described in the literature: Currently, the best algorithms for such tasks are based on convolutional neural networks. An illustration of their capabilities is given by the ImageNet Large Scale Visual Recognition Challenge; this is a benchmark in object classification and detection, with millions of images and hundreds of object classes. Performance of convolutional neural networks, on the ImageNet tests, is now close to that of humans. The best algorithms still struggle with objects that are small or thin, such as a small ant on a stem of a flower or a person holding a quill in their hand. They also have trouble with images that have been distorted with filters (an increasingly common phenomenon with modern digital cameras). By contrast, those kinds of images rarely trouble humans. Humans, however, tend to have trouble with other issues. For example, they are not good at classifying objects into fine-grained classes, such as the particular breed of dog or species of bird, whereas convolutional neural networks handle this with ease. Several specialized tasks based on recognition exist, such as: Several tasks relate to motion estimation where an image sequence is processed to produce an estimate of the velocity either at each points in the image or in the 3D scene, or even of the camera that produces the images . Examples of such tasks are: Given one or (typically) more images of a scene, or a video, scene reconstruction aims at computing a 3D model of the scene. In the simplest case the model can be a set of 3D points. More sophisticated methods produce a complete 3D surface model. The advent of 3D imaging not requiring motion or scanning, and related processing algorithms is enabling rapid advances in this field. Grid-based 3D sensing can be used to acquire 3D images from multiple angles. Algorithms are now available to stitch multiple 3D images together into point clouds and 3D models. The aim of image restoration is the removal of noise (sensor noise, motion blur, etc.) from images. The simplest possible approach for noise removal is various types of filters such as low-pass filters or median filters. More sophisticated methods assume a model of how the local image structures look, to distinguish them from noise. By first analysing the image data in terms of the local image structures, such as lines or edges, and then controlling the filtering based on local information from the analysis step, a better level of noise removal is usually obtained compared to the simpler approaches. An example in this field is inpainting. The organization of a computer vision system is highly application-dependent. Some systems are stand-alone applications that solve a specific measurement or detection problem, while others constitute a sub-system of a larger design which, for example, also contains sub-systems for control of mechanical actuators, planning, information databases, man-machine interfaces, etc. The specific implementation of a computer vision system also depends on whether its functionality is pre-specified or if some part of it can be learned or modified during operation. Many functions are unique to the application. There are, however, typical functions that are found in many computer vision systems. Image-understanding systems (IUS) include three levels of abstraction as follows: low level includes image primitives such as edges, texture elements, or regions; intermediate level includes boundaries, surfaces and volumes; and high level includes objects, scenes, or events. Many of these requirements are entirely topics for further research. The representational requirements in the designing of IUS for these levels are: representation of prototypical concepts, concept organization, spatial knowledge, temporal knowledge, scaling, and description by comparison and differentiation. While inference refers to the process of deriving new, not explicitly represented facts from currently known facts, control refers to the process that selects which of the many inference, search, and matching techniques should be applied at a particular stage of processing. Inference and control requirements for IUS are: search and hypothesis activation, matching and hypothesis testing, generation and use of expectations, change and focus of attention, certainty and strength of belief, inference and goal satisfaction. There are many kinds of computer vision systems; however, all of them contain these basic elements: a power source, at least one image acquisition device (camera, ccd, etc.), a processor, and control and communication cables or some kind of wireless interconnection mechanism. In addition, a practical vision system contains software, as well as a display in order to monitor the system. Vision systems for inner spaces, as most industrial ones, contain an illumination system and may be placed in a controlled environment. Furthermore, a completed system includes many accessories such as camera supports, cables and connectors. Most computer vision systems use visible-light cameras passively viewing a scene at frame rates of at most 60 frames per second (usually far slower). A few computer vision systems use image-acquisition hardware with active illumination or something other than visible light or both, such as structured-light 3D scanners, thermographic cameras, hyperspectral imagers, radar imaging, lidar scanners, magnetic resonance images, side-scan sonar, synthetic aperture sonar, etc. Such hardware captures "images" that are then processed often using the same computer vision algorithms used to process visible-light images. While traditional broadcast and consumer video systems operate at a rate of 30 frames per second, advances in digital signal processing and consumer graphics hardware has made high-speed image acquisition, processing, and display possible for real-time systems on the order of hundreds to thousands of frames per second. For applications in robotics, fast, real-time video systems are critically important and often can simplify the processing needed for certain algorithms. When combined with a high-speed projector, fast image acquisition allows 3D measurement and feature tracking to be realised. Egocentric vision systems are composed of a wearable camera that automatically take pictures from a first-person perspective. As of 2016, vision processing units are emerging as a new class of processor, to complement CPUs and graphics processing units (GPUs) in this role.
https://en.wikipedia.org/wiki?curid=6596
Curry Curry is a variety of dishes originating in the Indian subcontinent that use a complex combination of spices or herbs, usually including ground turmeric, cumin, coriander, ginger, and fresh or dried chilies. Curry is generally prepared in a sauce. Curry dishes prepared in the southern states of India, where the word also originated, may be spiced with leaves from the curry tree. There are many varieties of dishes called 'curries'. For example, in original traditional cuisines, the precise selection of spices for each dish is a matter of national or regional cultural tradition, religious practice, and, to some extent, family preference. Such dishes are called by specific names that refer to their ingredients, spicing, and cooking methods. Spices are used both whole and ground, cooked or raw, and they may be added at different times during the cooking process to produce different results. The main spices found in most curry powders of the Indian subcontinent are coriander, cumin, and turmeric. A wide range of additional spices may be included depending on the geographic region and the foods being included (fish, lentils, red or white meat, rice, and vegetables). Curry powder, a commercially prepared mixture of spices, is largely a Western creation, dating to the 18th century. Such mixtures are commonly thought to have first been prepared by Indian merchants for sale to members of the British Colonial government and army returning to Britain. Outside of the Indian subcontinent, "curry" may also be used to describe the various unrelated native dishes of Island Southeast Asia, Mainland Southeast Asia, and Oceania which use coconut milk or spice pastes and are commonly eaten over rice (like the Filipino "ginataan" and Thai "gaeng" class of dishes). Dishes called 'curry' may contain fish, meat, poultry, or shellfish, either alone or in combination with vegetables. Additionally, many instead are entirely vegetarian, eaten especially among those who hold ethical or religious proscriptions against eating meat or seafood. Curries may be either 'dry' or 'wet'. Dry curries are cooked with very little liquid which is allowed to evaporate, leaving the other ingredients coated with the spice mixture. Wet curries contain significant amounts of sauce or gravy based on broth, coconut cream or coconut milk, dairy cream or yogurt, or legume purée, sautéed crushed onion, or tomato purée. "Curry" is an anglicised form of the Tamil word "" meaning 'sauce' or 'relish for rice' that uses the leaves of the curry tree ("Murraya koenigii"). The word "kari" is also used in other Dravidian languages, namely in Malayalam, Kannada and Kodava with the meaning of "vegetables (or meat) of any kind (raw or boiled), curry". "Kaṟi" is described in a mid-17th century Portuguese cookbook by members of the British East India Company, who were trading with Tamil merchants along the Coromandel Coast of southeast India, becoming known as a "spice blend ... called "kari podi" or curry powder". The first known appearance in its anglicised form (spelled "currey") appears in a 1747 book of recipes published by Hannah Glasse. The word "cury" appears in the 1390s English cookbook, "The Forme of Cury", but is unrelated and comes from the Middle French word "cuire", meaning 'to cook' Archaeological evidence dating to 2600 BCE from Mohenjo-daro suggests the use of mortar and pestle to pound spices including mustard, fennel, cumin, and tamarind pods with which they flavoured food. Black pepper is native to the Indian subcontinent and Southeast Asia and has been known to Indian cooking since at least 2000 BCE. Domesticated Pacific coconuts and coconut milk extraction technology was introduced to southern India and Sri Lanka by the ancient maritime trade network of the Austronesian peoples, which influenced the curries of these regions. The establishment of the Mughal Empire, in the early 15th century, also influenced some curries, especially in the north. Another influence was the establishment of the Portuguese trading centre in Goa in 1510, resulting in the introduction of chili pepper to India from the Americas, as a byproduct of the Columbian Exchange. Curry was introduced to English cuisine starting with Anglo-Indian cooking in the 17th century as spicy sauces were added to plain boiled and cooked meats. The 1758 edition of Hannah Glasse's "The Art of Cookery" contains a recipe "To make a Currey the India Way". Curry was first served in coffee houses in Britain from 1809, and has been increasingly popular in Great Britain, with major jumps in the 1940s and the 1970s. During the 19th century, curry was also carried to the Caribbean by Indian indentured workers in the British sugar industry. Since the mid-20th century, curries of many national styles have become popular far from their origins, and increasingly become part of international fusion cuisine. From the culinary point of view, it is useful to consider the Indian subcontinent to be the entire historical region encompassed prior to independence since August 1947; that is, the modern countries of India, Bangladesh, Pakistan and Sri Lanka. It is usual to distinguish broadly between northern and southern styles of Indian cuisine, recognising that within those categories are innumerable sub-styles and variations. The distinction is commonly made with reference to the staple starch: wheat in the form of unleavened breads in the north; rice in the east; rice and millet in the south. Bengali cuisine, which refers to the cuisine of Bangladesh and the West Bengal state of India, includes curries, including seafood and fresh fish. Mustard seeds and mustard oil are added to many recipes, as are poppy seeds. Emigrants from the Sylhet district of Bangladesh founded the curry house industry in Britain and in Sylhet some restaurants run by expatriates specialise in British-style Indian food. Curries are the most well-known part of Indian cuisine. Most Indian dishes are usually curry based, prepared by adding different types of vegetables, lentils or meats in the curry. The content of the curry and style of preparation varies per the region. Most curries are water based, with occasional use of dairy and coconut milk. Curry dishes are usually thick and spicy and are eaten along with steamed rice and variety of Indian breads. Although wet curries play a smaller role in Gujarat than elsewhere, there are a number of vegetarian examples with gravies based on buttermilk or coconut milk. The main ingredient may variously be brinjal (eggplant/aubergine), potatoes, fresh corn kernels, okra, tomatoes, etc. In addition, there are several common kofta dishes which substitute vegetables for meat. Undhiyu, a Gujarati specialty, is a spicy 'wet' mixed-vegetable 'casserole' cooked in an earthenware pot, often eaten during the winter months. The curries of Maharashtra vary from mildly spicy to very spicy and include vegetarian, mutton, chicken and fish. Coastal Maharashtrian – Konkani – curries use coconut extensively along with spices. In western Maharashtra, curries are very spicy, often with peanut powder. Vidharba's cuisine is usually spicier than that of the coastal and southern regions. The ingredients commonly used are besan (gram flour), or chickpea flour, and groundnut powder. As a result of the Mughal rule in the region, the cuisine of Aurangabad has been highly influenced by the North Indian method of cooking. Khandeshi food is very spicy and the most famous dish is shev bhaji. Others include Eggplant bharta (wangyache bhareet), (urid dal), stuffed eggplant (bharleli wangi), bhaakari with thecha etc. The majority of Maharashtrian people are farmers living in the rural areas and therefore their traditional food is very simple. Most Punjabi dishes are prepared using "tadka", which is made with the frying of a "masala", which is a mix of ginger, garlic, onions and tomatoes with some dried spices. This is followed by the addition of other ingredients, water, and occasionally milk. Normally spicy, spice levels vary greatly depending on the household itself. Ghee and mustard oil are the most commonly used cooking fats. Many popular Punjabi dishes such as butter chicken and rajma are curry-based. These dishes are usually served with steamed rice and chapatis. Rajasthani cuisine was influenced both by the war-like lifestyles of its inhabitants and the availability of ingredients in this arid region. Food that could last for several days and could be eaten without heating was preferred. Scarcity of water and fresh green vegetables have each had their effect on the cooking. Hence the curries in Rajasthan are usually made using dry spices and herbs and other dry items like gram flour. "Kadhi" is a popular gram flour curry, usually served with steamed rice and bread. To decrease the use of water in this desert state they use a lot of milk and milk-products to cook curries. Laal maans is a popular meat curry from Rajasthan. The food in general from Andhra Pradesh and Telangana, both with Telugu-speaking natives, is considered the hottest in India. The state, being the leading producer of red chilli and green chilli, influences the liberal use of spices, making their curries, chutneys, savories and pickles the hottest and spiciest in taste. Curries known as vindaloo have become well known in Great Britain, America, and elsewhere, where the name is usually used simply to indicate a fiery dish of lamb or chicken frequently including potatoes. Such dishes are far from the Goan originals. The name "vindaloo" derives from the Portuguese "vinha d'alhos" or wine ("vinho") and garlic ("alho"), the two definitive flavour ingredients. The dish was originally made with pork, not taboo to the Christian Portuguese. The inclusion of potatoes was a later Indian addition, thought to be the result of confusion with the Hindi word for potato, "aloo". Throughout the years "vindaloo" has been altered to appeal to many people by adding spices and different wines. The curries of Karnataka are typically vegetarian or with meat and fish around mostly coastal areas. They use a wide variety of vegetables, spices, coconut and jaggery. There are dry and sauce-based curries. Some typical sauce-based dishes include saaru, gojju, thovve, huli, majjige huli (which is similar to the "kadi" made in the north), sagu or kootu, which is eaten mixed with hot rice. Malayali curries of Kerala typically contain shredded coconut paste or coconut milk, curry leaves, and various spices. Mustard seeds are used in almost every dish, along with onions, curry leaves, and sliced red chilies fried in hot oil. Most of the non-vegetarian dishes are heavily spiced. Kerala is known for its traditional sadya, a vegetarian meal served with boiled rice and a host of side dishes such as parippu (green gram), papadum, ghee, sambar, rasam, aviyal, kaalan, kichadi, pachadi, injipuli, Koottukari, pickles (mango, lime), thoran, one to four types of payasam, boli, olan, pulissery, moru (buttermilk), upperi, and banana chips. The sadya is customarily served on a banana leaf. Tamil cuisine's distinctive flavour and aroma is achieved by a blend and combination of spices including curry leaves, tamarind, coriander, ginger, garlic, chili, pepper, poppy seeds, mustard seeds, cinnamon, cloves, cardamom, cumin, fennel or anise seeds, fenugreek seeds, nutmeg, coconut, turmeric root or powder, and rosewater. Lentils, vegetables and dairy products are essential accompaniments and are often served with rice. Traditionally vegetarian foods dominate the menu with a range of non-vegetarian dishes including freshwater fish and seafood cooked with spices and seasoning. In the West, the best-known Kashmiri curry is "rogan josh", a wet curry of lamb with a brilliant red gravy whose colour is derived from a combination of Kashmiri chillies and an extract derived from the red flowers of the cockscomb plant ("mawal"). "Goshtaba" (large lamb meatballs cooked in yoghurt gravy) is another curry dish from the Wazwan tradition occasionally found in Western restaurants. The most important curry in the cuisine of the Maldives is cooked with diced fresh tuna and is known as "mas riha". "Kukulhu riha", chicken curry, is cooked with a different mixture of spices. Traditional vegetable curries in the Maldives include those that use "bashi" (eggplant/aubergine), "tora" ("Luffa aegyptiaca"), "barabō" (pumpkin), "chichanda" ("Trichosanthes cucumerina") and "muranga" ("Moringa oleifera"), as well as green unripe bananas and certain leaves as their main ingredients. Pieces of Maldive fish are normally added to give the vegetable curry a certain flavour. The curries of Nepalese cuisine have been influenced by its neighbours, mainly India and Tibet. Well known Indian spices are used less. Goat is a popular meat in the Himalayan region of Nepal. Daal bhaat (rice and lentil soup) is a staple dish of Nepal. Newa cuisine is a type of cuisine developed over centuries by the Newars of Nepal. Pakistani curries, especially in the provinces of Punjab and Sindh, are basically similar to their counterparts in northern India. Mutton and beef are common ingredients. A typical Pakistani lunch or dinner often consists of some form of bread (such as naan or roti) or rice with a meat or vegetable-based curry. Barbecue style or roasted meats are also very popular in the form of kebabs. It is worth noting that the term "curry" is virtually never used inside the country; instead, regional words such as "salan" or "shorba" are used to denote what is known outside the country as a "curry". Several different types of curries exist, depending on the cooking style, such as bhuna, bharta, roghan josh, qorma, qeema, and shorba. A favourite Pakistani curry is karahi, which is either mutton or chicken cooked in a cooking utensil called karahi, which is similar in shape to a wok. Lahori karahi incorporates garlic, ginger, fresh chillies, tomatoes and select spices. Peshawari karahi is another very popular version made with just meat, salt, tomatoes, and coriander. The cuisine from the Khyber Pakhtunkhwa province of Pakistan is somewhat similar to the cuisine of neighbouring Afghanistan. Extreme winters in some areas made the supply of fresh vegetables impossible, so a lot of dried fruits and vegetables are incorporated in the cuisine. The province still produces a large amount of nuts which are used abundantly in traditional cooking, along with cereals like wheat, maize, barley, and rice. Accompanying these staples are dairy products (yoghurt, whey), various nuts, native vegetables, and fresh and dried fruits. Peshawari karahi from the provincial capital of Peshawar is a popular curry all over the country. Cuisine in Pakistani Punjab differs from Indian Punjab on account of contents and religious diet rules. A typical Punjabi meal consists of some form of bread or rice with a "salan" (curry). Most preparations start with the frying of a masala which is a concoction of ginger, garlic, onions, tomatoes, and dried spices. Various other ingredients are then added. Spice level varies greatly depending on the sub-region as well as the household itself. A popular cooking fat is desi ghee with some dishes enriched with liberal amounts of butter and cream. There are certain dishes that are exclusive to Punjab, such as maash di dal and saron da saag (sarson ka saag). In Punjab and Kashmir, the only dish known as kardhi (curry) is a dish made of dahi (yogurt) and flour dumplings. In Pakistan, the provinces of Sindh and Balochistan border the Arabian Sea. Due to this, the Sindhi cuisine often has abundant use of fish in curries. Among Pakistani food, the Sindhi curries generally tend to be the hottest. The daily food in most Sindhi households consists of wheat-based flatbread (phulka) and rice accompanied by two dishes, one gravy and one dry. In Sri Lankan cuisine, rice, which is usually consumed daily, can be found at any special occasion, while spicy curries are favourite dishes for lunch and dinner. "Rice and curry" refers to a range of Sri Lankan dishes. Burmese cuisine is based on a very different understanding of curries. The principal ingredients of almost all Burmese curries are fresh onion (which provides the gravy and main body of the curry), Indian spices and red chilies. Usually, meat and fish are the main ingredients for popular curries. Burmese curries can be generalised into two types – the hot spicy dishes which exhibit north Indian or Pakistani influence, and the milder "sweet" curries. Burmese curries almost overwhelmingly lack coconut milk, setting them apart from most southeast Asian curries. Regular ingredients include fresh onion, garlic and chili paste. Common spices include garam masala, dried chili powder, cumin powder, turmeric and ngapi, a fermented paste made from either fish or prawns. Burmese curries are quite oily, as the extra oil helps the food to last longer. A spaghetti equivalent called Nan gyi thohk exists, in which wheat or rice noodles are eaten with thick chicken curry. In Indonesia curry is called "kari" or "kare". The most common type of "kari" consumed in Indonesia is "kari ayam" (chicken curry) and "kari kambing" (goat meat curry). In Aceh and North Sumatra roti cane is often eaten with "kari kambing". Other dishes such as gulai and opor are dishes based on curry. They are often highly localised and reflect the meat and vegetables available. They can therefore employ a variety of meats (chicken, beef, water buffalo and goat as in the flavoursome "gulai kambing"), seafood (such as prawn, crab, mussel, clam, and squid), fish (tuna, mackerel, carp, pangasius, catfish), or vegetables (young jackfruit, common beans, cassava leaf) dishes in a spiced sauce. They use local ingredients such as chili peppers, kaffir lime leaves, lemongrass, galangal, Indonesian bay leaves (salam leaf), candlenuts, turmeric, turmeric leaves, asam gelugur and asam kandis (sour mangosteens similar to tamarind), shrimp paste (terasi), cumin, coriander seed and coconut milk. In Aceh, curries use "daun salam koja" or "daun kari" ("Murraya koenigii") translated as "curry leaves". One popular dish, rendang from West Sumatran cuisine, is often described as caramelised beef dry curry. In Indonesia, rendang is usually not considered to be curry since it is richer and contains less liquid than is normal for Indonesian curries. Authentic rendang uses water buffalo meat slow-cooked in thick coconut milk for a number of hours to tenderise, caramelise, and flavour the meat. Opor Ayam is another variation of curry, which tastes very similar to gulai. Opor is usually whitish in colour and uses neither cinnamon nor turmeric, while gulai may contain either or both. Opor is also often part of a family meal around Lebaran, while gulai can be commonly found in Padang restaurants. Being at the crossroads of ancient trade routes has left a mark on Malaysian cuisine. While curry may have initially found its way to Malaysian shores via the Indian population, it has since become a staple among the Malays and Chinese. Malaysian curries differ from state to state, even within similar ethnic groupings, as they are influenced by many factors, be they cultural, religious, agricultural or economical. Malaysian curries typically use turmeric-rich curry powders, coconut milk, shallots, ginger, belacan (shrimp paste), chili peppers, and garlic. Tamarind is also often used. Rendang is another form of curry consumed in Malaysia, Singapore, Indonesia and the Philippines; it is drier and contains mostly meat and more coconut milk than a conventional Malaysian curry. Rendang was mentioned in Malay literature Hikayat Amir Hamzah (1550s) and is popular among Indonesians, Singaporeans and Malaysians. All sorts of things are curried in Malaysia, including mutton, chicken, tofu, shrimp, cuttlefish, fish, eggplants, eggs, and vegetables. In the Philippines, two kinds of curry traditions are seen corresponding with the cultural divide between the Hispanicised north and Indianised/Islamised south. In the northern areas, a linear range of new curry recipes could be seen. The most common is a variant of the native "ginataang manok" (chicken is cooked in coconut milk) dish with the addition of curry powder, known as the "Filipino chicken curry". This is the usual curry dish that northern Filipinos are familiar with. Similarly, other northern Filipino dishes that can be considered "curries" are usually "ginataan" (cooked with coconut milk) variants of other native meat or seafood dishes such as "adobo", "kaldereta", and "mechado", that simply add curry powder or non-native Indian spices. In southern areas of the Visayas, Mindanao, the Sulu Archipelago and southern Palawan, various older curry recipes are seen, and owe their origins to the limited influence of the Spanish in these regions that preserved older culinary traditions; as well as closer historical ties to Malay states like the Sultanate of Brunei. These Mindanaoan curries include "kulma", synonymous with the Indian "korma"; "tiyula itum" which is a beef curry blackened with burned coconut-meat powder; and "rendang", also eaten in Indonesia and Malaysia. Meats used in these curries include beef, goat, mutton, lamb, seafood and chicken. Pork is not used, in accordance with Islamic dietary laws. In Thai cuisine, curries are called "kaeng", and usually consist of meat, fish or vegetables in a sauce based on a paste made from chilies, onions or shallots, garlic, and shrimp paste. Additional spices and herbs define the type of curry. Local ingredients, such as chili peppers, kaffir lime leaves, lemongrass, galangal are used and, in central and southern Thai cuisine, coconut milk. Northern and northeastern Thai curries generally do not contain coconut milk. Due to the use of sugar and coconut milk, Thai curries tend to be sweeter than Indian curries. In the West, some of the Thai curries are described by colour; red curries use red chilies while green curries use green chilies. Yellow curry—called "kaeng kari" (by various spellings) in Thai, of which a literal translation could be "curry soup"—is more similar to Indian curries, with the use of turmeric, cumin, and other dried spices. A few stir-fried Thai dishes also use an Indian style curry powder (Thai: "phong kari"). Thai curries: In Vietnam where curry is called "cà ri", curry features include coconut milk, potato, sweet potato, taro roots, chicken garnished with coriander, and green onion. It is more soup-like than Indian curry. The curry is usually eaten with a baguette, rice vermicelli or steamed rice. Some dishes use a curry-based stew, such as snail dishes, phá lấu, stewed frogs or eels. Although not an integral part of Chinese cuisine, curry powder is added to some dishes in southern part of China. The curry powder sold in Chinese grocery stores is similar to Madras curry powder but with addition of star anise and cinnamon. Chinese curries (咖哩, gā lǐ) typically consist of chicken, beef, fish, lamb, or other meats, green peppers, onions, large chunks of potatoes, and a variety of other ingredients and spices in a mildly spicy yellow curry sauce, and topped over steamed rice. White pepper, soy sauce, hot sauce or hot chili oil may be applied to the sauce to enhance the flavour of the curry. The most common Chinese variety of curry sauce is usually sold in powder form. The ethnic Cantonese being dominant in Kuala Lumpur, this yellow Chinese-Malaysian variety was naturally introduced to China by the Cantonese. It features typically in Hong Kong cuisine, where curry is often cooked with brisket or fish balls. Malay satay seems to have been introduced to China with wider success by the ethnic Teochew, who make up the second largest group of Chinese of Singapore and are the dominant group in Thailand. In Hong Kong, curry fish balls are a street snack, and curried brisket is a typical main course in cha chaan teng and fast food restaurants. is usually eaten as "karē raisu" — curry, rice, and often pickled vegetables, served on the same plate and eaten with a spoon, a common lunchtime canteen dish. It is less spicy and seasoned than Indian and Southeast Asian curries, being more of a thick stew than a curry. British people brought curry from the Indian colony back to Britain and introduced it to Japan during the Meiji period (1868 to 1912), after Japan ended its policy of national self-isolation ("sakoku"), and curry in Japan was categorised as a Western dish. Its spread across the country is commonly attributed to its use in the Japanese Army and Navy which adopted it extensively as convenient field and naval canteen cooking, allowing even conscripts from the remotest countryside to experience the dish. The Japan Maritime Self-Defense Force traditionally have curry every Friday for lunch and many ships have their own unique recipes. The standard Japanese curry contains onions, carrots, potatoes, and sometimes celery, and a meat that is cooked in a large pot. Sometimes grated apples or honey are added for additional sweetness and other vegetables are sometimes used instead. For the meat, pork, beef, and chicken are the most popular, in order of decreasing popularity. In northern and eastern Japan including Tokyo, pork is the most popular meat for curry. Beef is more common in western Japan, including Osaka, and in Okinawa, chicken is favoured. Curry seasoning is commonly sold in the form of a condensed brick, similar to a bouillon cube, which dissolves in the mixture of meat and vegetables. Sometimes the curry-rice is topped with breaded pork cutlet (tonkatsu); this is called "katsukarē". Korokke (potato croquettes) are also a common topping. Apart from with rice, curry is also served over noodles, possibly even on top of broth in addition, in dishes such as curry udon and curry ramen. It is also used as the filling in a fried curry bread pastry. Though curry was introduced to Korea in the 1940s, the Indian dish was only popularized decades later, when Ottogi entered the Korean food industry by launching its powder-type curry product in 1969. Korean curry, usually served with rice, is characterized by the golden yellow colour from turmeric. Curry tteokbokki, along with curry rice, is one of the most popular curry dishes in Korea. It is made of tteok (rice cakes), eomuk (fish cakes), eggs, vegetables, and curry. Curry can be added to various Korean dishes such as bokkeumbap (fried rice), sundubujjigae (silken tofu stew), fried chicken, vegetable stir-fries, and salads. Curry is also used in Korean-style western food such as pasta and steak, as well as Korean-style Japanese food such as cream udon. Curry is very popular in the United Kingdom, with a curry house in nearly every town. Such is the popularity of curry in the United Kingdom, it has frequently been called its "adopted national dish". It was estimated that in 2016 there were 12,000 curry houses, employing 100,000 people and with annual combined sales of approximately £4.2 billion. In general the food offered is Indian food cooked to British taste; however, there is increasing demand for authentic Indian food. As of 2015 curry houses accounted for a fifth of the restaurant business in the U.K. but, being historically a low wage sector, they were plagued by a shortage of labour. Established Indian immigrants from South Asia were moving on to other occupations; there were difficulties in training Europeans to cook curry; and immigration restrictions, which require payment of a high wage to skilled immigrants, had crimped the supply of new cooks. Historically, the word "curry" was first used in British cuisine to denote dishes of meat (often leftover lamb) in a Western-style sauce flavoured with curry powder. The first curry recipe in Britain appeared in "The Art of Cookery made Plain and Easy" by Hannah Glasse in 1747. The first edition of her book used only black pepper and coriander seeds for seasoning of "currey". By the fourth edition of the book, other ingredients such as turmeric and ginger were called for. The use of hot spices was not mentioned, which reflected the limited use of chili in India — chili plants had only been introduced into India around the late 16th century and at that time were only popular in southern India. Many curry recipes are contained in 19th century cookbooks such as those of Charles Elmé Francatelli and Mrs Beeton. In "Mrs Beeton's Book of Household Management", a recipe for curry powder is given that contains coriander, turmeric, cinnamon, cayenne, mustard, ginger, allspice and fenugreek; although she notes that it is more economical to purchase the powder at "any respectable shop". According to legend, one 19th century attempt at curry resulted in the invention of Worcestershire sauce. Throughout the 19th and early 20th centuries, curry grew increasingly popular in Britain owing to the large number of British civil servants and military personnel associated with the British Raj. Following World War II, curry became even more popular in Britain owing to the large number of immigrants from South Asia. Curry has become an integral part of British cuisine, so much so that, since the late 1990s, chicken tikka masala has been referred to as "a true British national dish". Other British curry derivatives include "Coronation chicken", a cold dish, often used as a sandwich filling, invented to commemorate the coronation of Queen Elizabeth II in 1953 – and curry sauce (or curry gravy), usually served warm with traditional British fast food dishes such as chips. Curry sauce occasionally includes sultanas or other dried fruits. In 1810, the entrepreneur Sake Dean Mahomed, from the Bengal Presidency, opened the first Indian curry house in England: the Hindoostanee Coffee House in London. (Curry was served prior to this in some London coffee houses.) The first modern "upscale" Indian restaurant in Britain is thought to have been The Shafi in 1915, followed by Veeraswamy in London's Regent Street, founded in 1926; the latter is still standing and is the oldest surviving Indian restaurant in Britain. Bengalis in the UK settled in big cities with industrial employment. In London, they settled in the East End, which for centuries has been the first port of call for many immigrants working in the docks and shipping from east Bengal. Their regular stopover paved the way for food and curry outlets to be opened up catering for an all-male workforce as family migration and settlement took place some decades later. Brick Lane in the East London Borough of Tower Hamlets is famous for its many curry houses. Until the early 1970s, more than three-quarters of Indian restaurants in Britain were identified as being owned and run by people of Bengali origin. Most were run by migrants from East Pakistan, which became Bangladesh in 1971. Bangladeshi restaurateurs overwhelmingly come from the northeastern division of Sylhet. Until 1998, as many as 85% of curry restaurants in the UK were British Bangladeshi restaurants, but in 2003 this figure declined to just over 65%. The dominance of Bangladeshi restaurants is generally declining in some parts of London and the further north one travels. In Glasgow, there are more restaurants of Punjabi origin than any other. In the early 2010s the popularity of the curry house saw a decline. This has been attributed to the sale of this style of food in generic restaurants, increased home cooking of this style of food with easy supermarket availability of ingredients, and immigration restrictions brought in from 2008 making the availability of low-wage chefs and other staff difficult. Regardless of the ethnic origin of a restaurant's ownership, the menu will often be influenced by the wider South Asia (sometimes including Nepalese dishes), and sometimes cuisines from further afield (such as Persian dishes). Some British variations on Indian food are now being exported from the U.K. to India. Curry restaurants of more-or-less the British style are also popular in Canada, Australia and New Zealand. This cuisine is characterised by the use of a common base for all the sauces to which spices are added when individual dishes are prepared. The standard "feedstock" is usually a sautéed mixture of onion, garlic and fresh ginger, to which various spices are added, depending on the recipe, but which may include: cloves, cinnamon, cardamom, chilies, peppercorns, cumin and mustard seeds. Ground coriander seed is widely used as a thickening agent, and turmeric is added for colour and its digestive qualities. Fresh or canned tomatoes and bell peppers are a common addition. Better quality restaurants will normally make up new sauces on a daily basis, using fresh ingredients wherever possible and grinding their own spices. More modest establishments are more likely to resort to frozen or dried ingredients and pre-packaged spice mixtures. Restaurants in Great Britain have adopted a number of Indian terms to identify popular dishes. Although the names may derive from traditional dishes, often the recipes do not. Representative names include: The tandoor was introduced into Britain in the 1960s, and tandoori and tikka chicken became popular dishes. Other dishes may feature with varying strengths, with those of north Indian origin, such as butter chicken, tending to be mild, and recipes from the south of India tending to be hotter. Baltis are a style of curry thought to have been developed in Birmingham, England which have spread to other western countries and are traditionally cooked and served in the same pot, typically made of cast iron, called karahi or "balty". African curries, Cape Malay curries and Natal curries include the traditional Natal curry, the Durban curry, bunny chow, and roti rolls. South African curries appear to have been founded in two distinct regions – one in the east (KwaZulu-Natal) and the other in the west (Western Cape) – with a variety of other curries developing across the country over the late 20th century and early 21st century to include ekasi, coloured, and Afrikaner curries. Durban has the largest single population of Indians outside of India, who have been developing traditional Natal curries since their arrival in the late 19th century. Natal curries are mostly based on South Indian dishes and mostly consist of simple spiced lamb and chicken dishes (with large amounts of ghee and oils), but also include very complex and elaborate seafood, chicken and lamb specialties (chicken and prawn curry is a Natal favourite). Continental and British recipes have also evolved alongside Indian South African curries. Continental and British versions use mainly traditional recipes with the addition of red wine, milk, cream, vanilla or butter instead of ghee. Bunny chow or a "set", a South African standard, has spread in popularity throughout the country and into other southern African countries and countries with large South African immigrant populations. It consists of either lamb, chicken or bean curry poured into a tunnelled-out loaf of bread to be eaten with one's fingers. The roti roll is another classic takeaway curry that could either be a curry in a flat roti bread (similar to a kebab bread) or the classic "chip, cheese and curry" roll which basically consists of fried chips with melted cheese and curry gravy rolled into a roti roll. In the West Indies, curry is a very popular dish. The Indian indentured servants that were brought over from India by different European powers brought this dish to the West Indies. In Jamaica and Trinidad, curried goat is prominently featured. Curry can be found at both inexpensive and upscale Caribbean restaurants, and ingredients can range from chicken or vegetables to shellfish such as shrimp and scallops. Examples of curries in the West Indies include: In Fiji curries are made in most Indian homes and are eaten with rice or roti. Roti (circle or square) is mainly eaten for breakfast with vegetable curries. Lunch is often dal and rice with some side dishes. Most working people take roti and curry for their lunch. Dinner is usually curry, rice with some chutneys. Curries are normally cooked in vegetable oil. Ghee is mainly used to fry dal, to make puris or sweets. To make a curry, spices like cumin, fenugreek, mustard, and curry leaves are added to the hot oil. Onion is chopped or sliced and garlic crushed and added to the pot. Once the onion and garlic have turned slightly golden then turmeric and garam masala are added. For every 1 tsp turmeric normally 2 tsp masala is added. Salt and chillies are added according to taste. Curry is simmered on low heat until well cooked. Water is added so that it can be mixed with rice. If coriander leaves are available then they are added for extra flavour. Sometimes potatoes or vegetables are also added to curries to increase volume and make them more nutritious. Often coconut cream is added to seafood curries, such as prawn, crab or fish curries. Dal is often cooked with only turmeric and then fried in cumin, onion, and garlic. Sometimes carrots and leafy vegetables like "chauraiya" or "saijan" are added for extra flavor and nutrients. Curry powder is a spice mixture of widely varying composition developed by the British during the days of the Raj as a means of approximating the taste of Indian cuisine at home. Masala refers to spices, and this is the name given to the thick and pasty sauce based on a combination of spices with ghee (clarified butter), butter, palm oil or coconut milk. Most commercial curry powders available in Britain, the U.S. and Canada rely heavily on ground turmeric, in turn producing a very yellow sauce. Lesser ingredients in these Western yellow curry powders are often coriander, cumin, fenugreek, mustard, chili, black pepper and salt. By contrast, curry powders and curry pastes produced and consumed in India are extremely diverse; some red, some yellow, some brown; some with five spices and some with as many as 20 or more. Besides the previously mentioned spices, other commonly found spices in different curry powders in India are allspice, white pepper, ground mustard, ground ginger, cinnamon, roasted cumin, cloves, nutmeg, mace, green cardamom seeds or black cardamom pods, bay leaves and coriander seeds. Curry powder is used as an incidental ingredient in other cuisines, including for example a "curry sauce" ("sauce au curry", sometimes even "au cari") variation of the classic French béchamel.
https://en.wikipedia.org/wiki?curid=6597
Camel A camel is an even-toed ungulate in the genus "Camelus" that bears distinctive fatty deposits known as "humps" on its back. Camels have long been domesticated and, as livestock, they provide food (milk and meat) and textiles (fiber and felt from hair). Camels are working animals especially suited to their desert habitat and are a vital means of transport for passengers and cargo. There are three surviving species of camel. The one-humped dromedary makes up 94% of the world's camel population, and the two-humped Bactrian camel makes up 6%. The Wild Bactrian camel is a separate species and is now critically endangered. The word "camel" is also used informally in a wider sense, where the more correct term is "camelid", to include all seven species of the family Camelidae: the true camels (the above three species), along with the "New World" camelids: the llama, the alpaca, the guanaco, and the vicuña. The word itself is derived via and ("kamēlos") from Hebrew or Phoenician: "gāmāl". 3 species are extant: An extinct species of camel in the separate genus "Camelops", known as "C. hesternus", lived in western North America until the end of the Pleistocene, roughly 11,000 years ago. The average life expectancy of a camel is 40 to 50 years. A full-grown adult dromedary camel stands at the shoulder and at the hump. Bactrian camels can be a foot taller. Camels can run at up to in short bursts and sustain speeds of up to . Bactrian camels weigh and dromedaries . The widening toes on a camel's hoof provide supplemental grip for varying soil sediments. The male dromedary camel has an organ called a dulla in its throat, a large, inflatable sac he extrudes from his mouth when in rut to assert dominance and attract females. It resembles a long, swollen, pink tongue hanging out of the side of its mouth. Camels mate by having both male and female sitting on the ground, with the male mounting from behind. The male usually ejaculates three or four times within a single mating session. Camelids are the only ungulates to mate in a sitting position. Camels do not directly store water in their humps; they are reservoirs of fatty tissue. Concentrating body fat in their humps minimizes the insulating effect fat would have if distributed over the rest of their bodies, helping camels survive in hot climates. When this tissue is metabolized, it yields more than one gram of water for every gram of fat processed. This fat metabolization, while releasing energy, causes water to evaporate from the lungs during respiration (as oxygen is required for the metabolic process): overall, there is a net decrease in water. Camels have a series of physiological adaptations that allow them to withstand long periods of time without any external source of water. The dromedary camel can drink as seldom as once every 10 days even under very hot conditions, and can lose up to 30% of its body mass due to dehydration. Unlike other mammals, camels' red blood cells are oval rather than circular in shape. This facilitates the flow of red blood cells during dehydration and makes them better at withstanding high osmotic variation without rupturing when drinking large amounts of water: a camel can drink of water in three minutes. Camels are able to withstand changes in body temperature and water consumption that would kill most other mammals. Their temperature ranges from at dawn and steadily increases to by sunset, before they cool off at night again. In general, to compare between camels and the other livestock, camels lose only 1.3 liters of fluid intake every day while the other livestock lose 20 to 40 liters per day (Breulmann, et al., 2007). Maintaining the brain temperature within certain limits is critical for animals; to assist this, camels have a rete mirabile, a complex of arteries and veins lying very close to each other which utilizes countercurrent blood flow to cool blood flowing to the brain. Camels rarely sweat, even when ambient temperatures reach . Any sweat that does occur evaporates at the skin level rather than at the surface of their coat; the heat of vaporization therefore comes from body heat rather than ambient heat. Camels can withstand losing 25% of their body weight to sweating, whereas most other mammals can withstand only about 12–14% dehydration before cardiac failure results from circulatory disturbance. When the camel exhales, water vapor becomes trapped in their nostrils and is reabsorbed into the body as a means to conserve water. Camels eating green herbage can ingest sufficient moisture in milder conditions to maintain their bodies' hydrated state without the need for drinking. The camel's thick coat insulates it from the intense heat radiated from desert sand; a shorn camel must sweat 50% more to avoid overheating. During the summer the coat becomes lighter in color, reflecting light as well as helping avoid sunburn. The camel's long legs help by keeping its body farther from the ground, which can heat up to . Dromedaries have a pad of thick tissue over the sternum called the "pedestal". When the animal lies down in a sternal recumbent position, the pedestal raises the body from the hot surface and allows cooling air to pass under the body. Camels' mouths have a thick leathery lining, allowing them to chew thorny desert plants. Long eyelashes and ear hairs, together with nostrils that can close, form a barrier against sand. If sand gets lodged in their eyes, they can dislodge it using their transparent third eyelid. The camels' gait and widened feet help them move without sinking into the sand. The kidneys and intestines of a camel are very efficient at reabsorbing water. Camels' kidneys have a 1:4 cortex to medulla ratio. Thus, the medullary part of a camel's kidney occupies twice as much area as a cow's kidney. Secondly, renal corpuscles have a smaller diameter, which reduces surface area for filtration. These two major anatomical characteristics enable camels to conserve water and limit the volume of urine in extreme desert conditions. Camel urine comes out as a thick syrup, and camel faeces are so dry that they do not require drying when the Bedouins use them to fuel fires. The camel immune system differs from those of other mammals. Normally, the Y-shaped antibody molecules consist of two heavy (or long) chains along the length of the Y, and two light (or short) chains at each tip of the Y. Camels, in addition to these, also have antibodies made of only two heavy chains, a trait that makes them smaller and more durable. These "heavy-chain-only" antibodies, discovered in 1993, are thought to have developed 50 million years ago, after camelids split from ruminants and pigs. The karyotypes of different camelid species have been studied earlier by many groups, but no agreement on chromosome nomenclature of camelids has been reached. A 2007 study flow sorted camel chromosomes, building on the fact that camels have 37 pairs of chromosomes (2n=74), and found that the karyotype consisted of one metacentric, three submetacentric, and 32 acrocentric autosomes. The Y is a small metacentric chromosome, while the X is a large metacentric chromosome. The hybrid camel, a hybrid between Bactrian and dromedary camels, has one hump, though it has an indentation deep that divides the front from the back. The hybrid is at the shoulder and tall at the hump. It weighs an average of and can carry around , which is more than either the dromedary or Bactrian can. According to molecular data, the wild Bactrian camel ("C. ferus") separated from the domestic Bactrian camel ("C. bactrianus") about 1 million years ago. New World and Old World camelids diverged about 11 million years ago. In spite of this, these species can hybridize and produce viable offspring. The cama is a camel-llama hybrid bred by scientists to see how closely related the parent species are. Scientists collected semen from a camel via an artificial vagina and inseminated a llama after stimulating ovulation with gonadotrophin injections. The cama is halfway in size between a camel and a llama and lacks a hump. It has ears intermediate between those of camels and llamas, longer legs than the llama, and partially cloven hooves. Like the mule, camas are sterile, despite both parents having the same number of chromosomes. The earliest known camel, called "Protylopus", lived in North America 40 to 50 million years ago (during the Eocene). It was about the size of a rabbit and lived in the open woodlands of what is now South Dakota. By 35 million years ago, the "Poebrotherium" was the size of a goat and had many more traits similar to camels and llamas. The hoofed "Stenomylus", which walked on the tips of its toes, also existed around this time, and the long-necked "Aepycamelus" evolved in the Miocene. An early relative of extant Old World camels, "Paracamelus", existed in the upper Miocene to Middle Pleistocene. Around 3–5 million years ago, the North American Camelidae spread to South America as part of the Great American Interchange via the newly formed Isthmus of Panama, where they gave rise to guanacos and related animals, and to Asia via the Bering land bridge. Surprising finds of fossil "Paracamelus" on Ellesmere Island beginning in 2006 in the high Canadian Arctic suggest that the extant Old World camels may descend from a larger, boreal browser whose hump may have evolved as an adaptation in a cold climate. This creature is estimated to have stood around nine feet (2.7 metres) tall. The Bactrican camel diverged from the dromedary about 1 million years ago, according to the fossil record. The last camel native to North America was "Camelops hesternus", which vanished along with horses, short-faced bears, mammoths and mastodons, ground sloths, sabertooth cats, and many other megafauna, coinciding with the migration of humans from Asia. Like horses before their extinction in their continent of origin, camels spread across the Bering land bridge, moving in the opposite direction from the Asian immigration to America. They survived in the Old World, and eventually humans domesticated them and spread them globally. Along with many other megafauna in North America, the original wild camels were wiped out during the spread of Native Americans from Asia into North America, 12,000 to 10,000 years ago; although fossils have never been associated with definitive evidence of hunting. Most camels surviving today are domesticated. Although feral populations exist in Australia, India and Kazakhstan, wild camels survive only in the wild Bactrian camel population of the Gobi Desert. Humans may have first domesticated dromedaries in Somalia and southern Arabia around 3,000 BC, and Bactrian camels in central Asia around 2,500 BC, as at Shahr-e Sukhteh (also known as the Burnt City), Iran. Martin Heide's 2010 work on the domestication of the camel tentatively concludes that humans had domesticated the Bactrian camel by at least the middle of the third millennium somewhere east of the Zagros Mountains, with the practice then moving into Mesopotamia. Heide suggests that mentions of camels "in the patriarchal narratives may refer, at least in some places, to the Bactrian camel", while noting that the camel is not mentioned in relationship to Canaan. Recent excavations in the Timna Valley by Lidar Sapir-Hen and Erez Ben-Yosef discovered what may be the earliest domestic camel bones yet found in Israel or even outside the Arabian Peninsula, dating to around 930 BC. This garnered considerable media coverage, as it is strong evidence that the stories of Abraham, Jacob, Esau, and Joseph were written after this time. The existence of camels in Mesopotamia—but not in the eastern Mediterranean lands—is not a new idea. The historian Richard Bulliet did not think that the occasional mention of camels in the Bible meant that the domestic camels were common in the Holy Land at that time. The archaeologist William F. Albright, writing even earlier, saw camels in the Bible as an anachronism. The official report by Sapir-Hen and Ben-Joseph notes: The introduction of the dromedary camel (Camelus dromedarius) as a pack animal to the southern Levant ... substantially facilitated trade across the vast deserts of Arabia, promoting both economic and social change (e.g., Kohler 1984; Borowski 1998: 112–116; Jasmin 2005). This ... has generated extensive discussion regarding the date of the earliest domestic camel in the southern Levant (and beyond) (e.g., Albright 1949: 207; Epstein 1971: 558–584; Bulliet 1975; Zarins 1989; Köhler-Rollefson 1993; Uerpmann and Uerpmann 2002; Jasmin 2005; 2006; Heide 2010; Rosen and Saidel 2010; Grigson 2012). Most scholars today agree that the dromedary was exploited as a pack animal sometime in the early Iron Age (not before the 12th century [BC]) and concludes: Current data from copper smelting sites of the Aravah Valley enable us to pinpoint the introduction of domestic camels to the southern Levant more precisely based on stratigraphic contexts associated with an extensive suite of radiocarbon dates. The data indicate that this event occurred not earlier than the last third of the 10th century [BC] and most probably during this time. The coincidence of this event with a major reorganization of the copper industry of the region—attributed to the results of the campaign of Pharaoh Shoshenq I—raises the possibility that the two were connected, and that camels were introduced as part of the efforts to improve efficiency by facilitating trade. Desert tribes and Mongolian nomads use camel hair for tents, yurts, clothing, bedding and accessories. Camels have outer guard hairs and soft inner down, and the fibers are sorted by color and age of the animal. The guard hairs can be felted for use as waterproof coats for the herdsmen, while the softer hair is used for premium goods. The fiber can be spun for use in weaving or made into yarns for hand knitting or crochet. Pure camel hair is recorded as being used for western garments from the 17th century onwards, and from the 19th century a mixture of wool and camel hair was used. By at least 1200 BC the first camel saddles had appeared, and Bactrian camels could be ridden. The first saddle was positioned to the back of the camel, and control of the Bactrian camel was exercised by means of a stick. However, between 500 and 100 BC, Bactrian camels came into military use. New saddles, which were inflexible and bent, were put over the humps and divided the rider's weight over the animal. In the seventh century BC the military Arabian saddle evolved, which again improved the saddle design slightly. Military forces have used camel cavalries in wars throughout Africa, the Middle East, and into the modern-day Border Security Force (BSF) of India (though as of July 2012, the BSF planned the replacement of camels with ATVs). The first documented use of camel cavalries occurred in the Battle of Qarqar in 853 BC. Armies have also used camels as freight animals instead of horses and mules. The East Roman Empire used auxiliary forces known as "dromedarii", whom the Romans recruited in desert provinces. The camels were used mostly in combat because of their ability to scare off horses at close range (horses are afraid of the camels' scent), a quality famously employed by the Achaemenid Persians when fighting Lydia in the Battle of Thymbra (547 BC). The United States Army established the U.S. Camel Corps, stationed in California, in the late 19th century. One may still see stables at the Benicia Arsenal in Benicia, California, where they nowadays serve as the Benicia Historical Museum. Though the experimental use of camels was seen as a success (John B. Floyd, Secretary of War in 1858, recommended that funds be allocated towards obtaining a thousand more camels), the outbreak of the American Civil War in 1861 saw the end of the Camel Corps: Texas became part of the Confederacy, and most of the camels were left to wander away into the desert. France created a "méhariste" camel corps in 1912 as part of the Armée d'Afrique in the Sahara in order to exercise greater control over the camel-riding Tuareg and Arab insurgents, as previous efforts to defeat them on foot had failed. The Free French Camel Corps fought during World War II, and camel-mounted units remained in service until the end of French rule over Algeria in 1962. In 1916, the British created the Imperial Camel Corps. It was originally used to fight the Senussi, but was later used in the Sinai and Palestine Campaign in World War I. The Imperial Camel Corps comprised infantrymen mounted on camels for movement across desert, though they dismounted at battle sites and fought on foot. After July 1918, the Corps began to become run down, receiving no new reinforcements, and was formally disbanded in 1919. In World War I, the British Army also created the Egyptian Camel Transport Corps, which consisted of a group of Egyptian camel drivers and their camels. The Corps supported British war operations in Sinai, Palestine, and Syria by transporting supplies to the troops. The Somaliland Camel Corps was created by colonial authorities in British Somaliland in 1912; it was disbanded in 1944. Bactrian camels were used by Romanian forces during World War II in the Caucasian region. At the same period the Soviet units operating around Astrakhan in 1942 adopted local camels as draft animals due to shortage of trucks and horses, and kept them even after moving out of the area. Despite severe losses, some of these camels came as far West as to Berlin itself. The Bikaner Camel Corps of British India fought alongside the British Indian Army in World Wars I and II. The "Tropas Nómadas" (Nomad Troops) were an auxiliary regiment of Sahrawi tribesmen serving in the colonial army in Spanish Sahara (today Western Sahara). Operational from the 1930s until the end of the Spanish presence in the territory in 1975, the "Tropas Nómadas" were equipped with small arms and led by Spanish officers. The unit guarded outposts and sometimes conducted patrols on camelback. Camel milk is a staple food of desert nomad tribes and is sometimes considered a meal itself; a nomad can live on only camel milk for almost a month. Camel milk can readily be made into yogurt, but can only be made into butter if it is soured first, churned, and a clarifying agent is then added. Until recently, camel milk could not be made into camel cheese because rennet was unable to coagulate the milk proteins to allow the collection of curds. Developing less wasteful uses of the milk, the FAO commissioned Professor J.P. Ramet of the École Nationale Supérieure d'Agronomie et des Industries Alimentaires, who was able to produce curdling by the addition of calcium phosphate and vegetable rennet in the 1990s. The cheese produced from this process has low levels of cholesterol and is easy to digest, even for the lactose intolerant. Camel milk can also be made into ice cream. They provide food in the form of meat and milk (Tariq "et al.",2010). Approximately 3.3 million camels and camelids are slaughtered each year for meat worldwide. A camel carcass can provide a substantial amount of meat. The male dromedary carcass can weigh , while the carcass of a male Bactrian can weigh up to . The carcass of a female dromedary weighs less than the male, ranging between . The brisket, ribs and loin are among the preferred parts, and the hump is considered a delicacy. The hump contains "white and sickly fat", which can be used to make the "khli" (preserved meat) of mutton, beef, or camel. On the other hand, camel milk and meat are rich in protein, vitamins, glycogen, and other nutrients making them essential in the diet of many people. From chemical composition to meat quality, the dromedary camel is the preferred breed for meat production. It does well even in arid areas due to its unusual physiological behaviors and characteristics, which include tolerance to extreme temperatures, radiation from the sun, water paucity, rugged landscape and low vegetation. Camel meat is reported to taste like coarse beef, but older camels can prove to be very tough, although camel meat becomes tenderer the more it is cooked. The Abu Dhabi Officers' Club serves a camel burger mixed with beef or lamb fat in order to improve the texture and taste. In Karachi, Pakistan, some restaurants prepare nihari from camel meat. Specialist camel butchers provide expert cuts, with the hump considered the most popular. Camel meat has been eaten for centuries. It has been recorded by ancient Greek writers as an available dish at banquets in ancient Persia, usually roasted whole. The Roman emperor Heliogabalus enjoyed camel's heel. Camel meat is mainly eaten in certain regions, including Eritrea, Somalia, Djibouti, Saudi Arabia, Egypt, Syria, Libya, Sudan, Ethiopia, Kazakhstan, and other arid regions where alternative forms of protein may be limited or where camel meat has had a long cultural history. Camel blood is also consumable, as is the case among pastoralists in northern Kenya, where camel blood is drunk with milk and acts as a key source of iron, vitamin D, salts and minerals. A 2005 report issued jointly by the Saudi Ministry of Health and the United States Centers for Disease Control and Prevention details four cases of human bubonic plague resulting from the ingestion of raw camel liver. Camel meat is also occasionally found in Australian cuisine: for example, a camel lasagna is available in Alice Springs. Australia has exported camel meat, primarily to the Middle East but also to Europe and the US, for many years. The meat is very popular among North African Australians, such as Somalis, and other Australians have also been buying it. The feral nature of the animals means they produce a different type of meat to farmed camels in other parts of the world, and it is sought after because it is disease-free, and a unique genetic group. Demand is outstripping supply, and governments are being urged not to cull the camels, but redirect the cost of the cull into developing the market. Australia has seven camel dairies, which produce milk, cheese and skincare products in addition to meat. Camel meat is "halal" (, 'allowed') for Muslims. However, according to some Islamic schools of thought, a state of impurity is brought on by the consumption of it. Consequently, these schools hold that Muslims must perform "wudhu" (ablution) before the next time they pray after eating camel meat. Also, some Islamic schools of thought consider it "haram" (, 'forbidden') for a Muslim to perform "Salat" in places where camels lie, as it is said to be a dwelling place of the "Shaytan" (, 'Devil'). According to Abu Yusuf, the urine of camel may be used for medical treatment if necessary, but according to Abū Ḥanīfah, the drinking of camel urine is discouraged. The Islamic texts contain several stories featuring camels. In the story of the people of Thamud, the Prophet Salih miraculously brings forth a "naqat" (, 'she-camel') out of a rock. After the Prophet Muhammad migrated from Mecca to Medina, he allowed his she-camel to roam there; the location where the camel stopped to rest determined the location where he would build his house in Medina. According to Jewish tradition, camel meat and milk are not kosher. Camels possess only one of the two kosher criteria; although they chew their cud, they do not possess cloven hooves: "But these you shall not eat among those that bring up the cud and those that have a cloven hoof: the camel, because it brings up its cud, but does not have a [completely] cloven hoof; it is unclean for you." There are around 14 million camels alive , with 90% being dromedaries. Dromedaries alive today are domesticated animals (mostly living in the Horn of Africa, the Sahel, Maghreb, Middle East and South Asia). The Horn region alone has the largest concentration of camels in the world, where the dromedaries constitute an important part of local nomadic life. They provide nomadic people in Somalia and Ethiopia with milk, food, and transportation. Around 700,000 dromedary camels are now feral in Australia, descended from those introduced as a method of transport in the 19th and early 20th centuries. This population is growing about 8% per year. Representatives of the Australian government have culled more than 100,000 of the animals in part because the camels use too much of the limited resources needed by sheep farmers. A small population of introduced camels, dromedaries and Bactrians, wandered through Southwestern United States after having been imported in the 19th century as part of the U.S. Camel Corps experiment. When the project ended, they were used as draft animals in mines and escaped or were released. Twenty-five U.S. camels were bought and exported to Canada during the Cariboo Gold Rush. The Bactrian camel is, , reduced to an estimated 1.4 million animals, most of which are domesticated. The Wild Bactrian camel is a separate species and is the only truly wild (as opposed to feral) camel in the world. The wild camels are critically endangered and number approximately 1400, inhabiting the Gobi and Taklamakan Deserts in China and Mongolia.
https://en.wikipedia.org/wiki?curid=6598
Chaldea Chaldea (, also spelled Chaldaea) was a country that existed between the late 10th or early 9th and mid-6th centuries BC, after which the country and its people were absorbed and assimilated into Babylonia. Semitic-speaking, it was located in the marshy land of the far southeastern corner of Mesopotamia and briefly came to rule Babylon. The Hebrew Bible uses the term ("Kaśdim") and this is translated as "Chaldaeans" in the Greek Old Testament, although there is some dispute as to whether "Kasdim" in fact means "Chaldean" or refers to the south Mesopotamian "Kaldu". During a period of weakness in the East Semitic-speaking kingdom of Babylonia, new tribes of West Semitic-speaking migrants arrived in the region from the Levant between the 11th and 9th centuries BCE. The earliest waves consisted of Suteans and Arameans, followed a century or so later by the Kaldu, a group who became known later as the Chaldeans or the Chaldees. These migrations did not affect the powerful kingdom of Assyria in the northern half of Mesopotamia, which repelled these incursions. The short-lived 11th dynasty of the Kings of Babylon (6th century BCE) is conventionally known to historians as the Chaldean Dynasty, although the last rulers, Nabonidus and his son Belshazzar, were from Assyria. (This is not found in Georges Roux - Ancient Iraq. Roux states that Nabonidus "was the son of a certain Nabu-Balatsu-Iqbi, who belonged to the Babylonian nobility but was not of royal blood and of a votaress... of the city of Harran.") These nomadic Chaldeans settled in the far southeastern portion of Babylonia, chiefly on the left bank of the Euphrates. Though for a short time the name commonly referred to the whole of southern Mesopotamia in Hebraic literature, this was a geographical and historical misnomer as Chaldea proper was in fact only the plain in the far southeast formed by the deposits of the Euphrates and the Tigris, extending about four hundred miles along the course of these rivers and averaging about a hundred miles in width. The name "Chaldaea" is a latinization of the Greek ' (), a hellenization of Akkadian ' or '. The name appears in Hebrew in the Bible as ' () and in Aramaic as "" (). The Hebrew word possibly appears in the Bible (Genesis 22:22) in the name "Kesed"(כשד), the singular form of "Kasdim"(כַּשְׂדִּים), meaning Chaldeans. Kesed is identified as son of Abraham's brother Nahor (and brother of Kemuel the father of Aram), residing in Aram Naharaim. Jewish historian Flavius Josephus also links Arphaxad and Chaldaea, in his Antiquities of the Jews, stating, “Arphaxad named the Arphaxadites, who are now called Chaldeans.” In the early period, between the early 9th century and late 7th century BCE, "mat Kaldi" was the name of a small sporadically independent migrant-founded territory under the domination of the Neo-Assyrian Empire (911-605 BCE) in southeastern Babylonia, extending to the western shores of the Persian Gulf. The expression " mat Bit Yâkin" is also used, apparently synonymously. "Bit Yâkin" was the name of the largest and most powerful of the five tribes of the Chaldeans, or equivalently, their territory. The original extension of "Bit Yâkin" is not known precisely, but it extended from the lower Tigris into Arabia. Sargon II mentions it as extending as far as Dilmun or "sea-land" (littoral Eastern Arabia). "Chaldea" or "mat Kaldi" generally referred to the low, marshy, alluvial land around the estuaries of the Tigris and Euphrates, which at the time discharged their waters through separate mouths into the sea. The tribal capital "Dur Yâkin" was the original seat of Marduk-Baladan. The king of Chaldea was also called the king of Bit Yakin, just as the kings of Babylonia and Assyria were regularly styled simply king of Babylon or Assur, the capital city in each case. In the same way, what is now known as the Persian Gulf was sometimes called "the Sea of Bit Yakin", and sometimes "the Sea of the Land of Chaldea". "Chaldea" came to be used in a wider sense, of Mesopotamia in general, following the ascendancy of the Chaldeans during 608–557 BCE. This is especially the case in the Hebrew Bible, which was substantially composed during this period (roughly corresponding to the period of Babylonian captivity). The Book of Jeremiah makes frequent reference to the Chaldeans (KJV "Chaldees" following LXX ; in Biblical Hebrew as "Kasdîy(mâh)" "Kassites"). Habbakuk 1:6 calls them "that bitter and hasty nation" (). Unlike the East Semitic Akkadian-speaking Akkadians, Assyrians and Babylonians, whose ancestors had been established in Mesopotamia since at least the 30th century BCE, the Chaldeans were not a native Mesopotamian people, but were late 10th or early 9th century BCE West Semitic Levantine migrants to the southeastern corner of the region, who had played no part in the previous 3,000 years or so of Sumero-Akkadian and Assyro-Babylonian Mesopotamian civilization and history. The ancient Chaldeans seem to have migrated into Mesopotamia sometime between c. 940–860 BCE, a century or so after other new Semitic arrivals, the Arameans and the Suteans, appeared in Babylonia, c. 1100 BCE. They first appear in written record in the annals of the Assyrian king Shalmaneser III during the 850s BCE. This was a period of weakness in Babylonia, and its ineffectual native kings were unable to prevent new waves of semi-nomadic foreign peoples from invading and settling in the land. Though belonging to the same West Semitic speaking ethnic group and migrating from the same Levantine regions like the earlier arriving Aramaeans, they are to be differentiated; the Assyrian king Sennacherib, for example, carefully distinguishes them in his inscriptions. The Chaldeans were able to keep their identity despite the dominant Assyro-Babylonian culture although some were not able to, as was the case for the earlier Amorites, Kassites and Suteans before them by the time Babylon fell in 539 BCE. The Chaldeans originally spoke a West Semitic language similar to but distinct from Aramaic. During the Assyrian Empire, the Assyrian king Tiglath-Pileser III introduced an Eastern Aramaic dialect as the lingua franca of his empire in the mid-8th century BCE. As a result of this innovation, in late periods both the Babylonian and Assyrian dialects of Akkadian became marginalised, and Mesopotamian Aramaic took its place across Mesopotamia, including among the Chaldeans. This language in the form of Eastern Aramaic neo-Aramaic dialects still remains the mother-tongue of the now Christian Assyrian people of northern Iraq, north-east Syria, south-eastern Turkey and north-western Iran to this day. One form of this once widespread language is used in Daniel and Ezra, but the use of the name "Chaldee" to describe it, first introduced by Jerome, is linguistically correct and accurate in the sense that the Chaldeans were using this language. In the Hebrew Bible, the prophet Abraham is stated to have originally come from "Ur of the Chaldees" (Ur Kaśdim). The region that the Chaldeans eventually made their homeland was in relatively poor southeastern Mesopotamia, at the head of the Persian Gulf. They appear to have migrated into southern Babylonia from the Levant at some unknown point between the end of the reign of Ninurta-kudurri-usur II (a contemporary of Tiglath-Pileser II) circa 940 BCE, and the start of the reign of Marduk-zakir-shumi I in 855 BCE, although there is no historical proof of their existence prior to the late 850s BCE. For perhaps a century or so after settling in the area, these semi-nomadic migrant Chaldean tribes had no impact on the pages of history, seemingly remaining subjugated by the native Akkadian speaking kings of Babylon or by perhaps regionally influential Aramean tribes. The main players in southern Mesopotamia during this period were Babylonia and Assyria, together with Elam to the east and the Aramaeans, who had already settled in the region a century or so prior to the arrival of the Chaldeans. The very first "written historical attestation" of the existence of Chaldeans occurs in 852 BCE, in the annals of the Assyrian king Shalmaneser III, who mentions invading the southeastern extremes of Babylonia and subjugating one Mushallim-Marduk, the chief of the Amukani tribe and overall leader of the Kaldu tribes, together with capturing the town of Baqani, extracting tribute from Adini, chief of the Bet-Dakkuri, another Chaldean tribe. Shalmaneser III had invaded Babylonia at the request of its own king, Marduk-zakir-shumi I, the Babylonian king being threatened by his own rebellious relations, together with powerful Aramean tribes pleaded with the more powerful Assyrian king for help. The subjugation of the Chaldean tribes by the Assyrian king appears to have been an aside, as they were not at that time a powerful force or a threat to the native Babylonian king. Important Kaldu regions in southeastern Babylonia were Bit-Yâkin (the original area the Chaldeans settled in on the Persian Gulf), Bet-Dakuri, Bet-Adini, Bet-Amukkani, and Bet-Shilani. Chaldean leaders had by this time already adopted Assyro-Babylonian names, religion, language, and customs, indicating that they had become Akkadianized to a great degree. The Chaldeans remained quietly ruled by the native Babylonians (who were in turn subjugated by their Assyrian relations) for the next seventy-two years, only coming to historical prominence for the first time in Babylonia in 780 BCE, when a previously unknown Chaldean named Marduk-apla-usur usurped the throne from the native Babylonian king Marduk-bel-zeri (790–780 BCE). The latter was a vassal of the Assyrian king Shalmaneser IV (783–773 BCE), who was otherwise occupied quelling a civil war in Assyria at the time. This was to set a precedent for all future Chaldean aspirations on Babylon during the Neo Assyrian Empire; always too weak to confront a strong Assyria alone and directly, the Chaldeans awaited periods when Assyrian kings were distracted elsewhere in their vast empire, or engaged in internal conflicts, then, in alliance with other powers stronger than themselves (usually Elam), they made a bid for control over Babylonia. Shalmaneser IV attacked and defeated Marduk-apla-user, retaking northern Babylonia and forcing on him a border treaty in Assyria's favour. The Assyrians allowed him to remain on the throne, although subject to Assyria. Eriba-Marduk, another Chaldean, succeeded him in 769 BCE and his son, Nabu-shuma-ishkun in 761 BCE, with both being dominated by the new Assyrian king Ashur-Dan III (772–755 BCE). Babylonia appears to have been in a state of chaos during this time, with the north occupied by Assyria, its throne occupied by foreign Chaldeans, and continual civil unrest throughout the land. The Chaldean rule proved short-lived. A native Babylonian king named Nabonassar (748–734 BCE) defeated and overthrew the Chaldean usurpers in 748 BCE, restored indigenous rule, and successfully stabilised Babylonia. The Chaldeans once more faded into obscurity for the next three decades. During this time both the Babylonians and the Chaldean and Aramean migrant groups who had settled in the land once more fell completely under the yoke of the powerful Assyrian king Tiglath-Pileser III (745–727 BCE), a ruler who introduced Imperial Aramaic as the lingua franca of his empire. The Assyrian king at first made Nabonassar and his successor native Babylonian kings Nabu-nadin-zeri, Nabu-suma-ukin II and Nabu-mukin-zeri his subjects, but decided to rule Babylonia directly from 729 BCE. He was followed by Shalmaneser V (727–722 BCE), who also ruled Babylon in person. When Sargon II (722–705 BCE) ascended the throne of the Assyrian Empire in 722 BCE after the death of Shalmaneser V, he was forced to launch a major campaign in his subject states of Persia, Mannea and Media in Ancient Iran to defend his territories there. He defeated and drove out the Scythians and Cimmerians who had attacked Assyria's Persian and Medina vassal colonies in the region. At the same time, Egypt began encouraging and supporting the rebellion against Assyria in Israel and Canaan, forcing the Assyrians to send troops to deal with the Egyptians. These events allowed the Chaldeans to once more attempt to assert themselves. While the Assyrian king was otherwise occupied defending his Iranian colonies from the Scythians and Cimmerians and driving the Egyptians from Canaan, Marduk-apla-iddina II (the Biblical Merodach-Baladan) of Bit-Yâkin, allied himself with the powerful Elamite kingdom and the native Babylonians, briefly seizing control of Babylon between 721 and 710 BCE. With the Scythians and Cimmerians vanquished, the Medes and Persians pledging loyalty, and the Egyptians defeated and ejected from southern Canaan, Sargon II was free at last to deal with the Chaldeans, Babylonians and Elamites. He attacked and deposed Marduk-apla-adding II in 710 BCE, also defeating his Elamite allies in the process. After defeat by the Assyrians, Merodach-Baladan fled to his protectors in Elam. In 703, Merodach-Baladan very briefly regained the throne from a native Akkadian-Babylonian ruler Marduk-zakir-shumi II, who was a puppet of the new Assyrian king, Sennacherib (705–681 BCE). He was once more soundly defeated at Kish, and once again fled to Elam where he died in exile after one final failed attempt to raise a revolt against Assyria in 700 BCE, this time not in Babylon, but in the Chaldean tribal land of Bit-Yâkin. A native Babylonian king named Bel-ibni (703–701 BCE) was placed on the throne as a puppet of Assyria. The next challenge to Assyrian domination came from the Elamites in 694 BCE, with Nergal-ushezib deposing and murdering Ashur-nadin-shumi (700–694 BCE), the Assyrian prince who was king of Babylon and son of Sennacherib. The Chaldeans and Babylonians again allied with their more powerful Elamite neighbors in this endeavour. This prompted the enraged Assyrian king Sennacherib to invade and subjugate Elam and Chaldea and to sack Babylon, laying waste to and largely destroying the city. Babylon was regarded as a sacred city by all Mesopotamians, including the Assyrians, and this act eventually resulted in Sennacherib's being murdered by his own sons while he was praying to the god Nisroch in Nineveh. Esarhaddon (681–669 BCE) succeeded Sennacherib as ruler of the Assyrian Empire. He completely rebuilt Babylon and brought peace to the region. He conquered Egypt, Nubia and Libya and entrenched his mastery over the Persians, Medes, Parthians, Scythians, Cimmerians, Arameans, Israelites, Phoenicians, Canaanites, Urartians, Pontic Greeks, Cilicians, Phrygians, Lydians, Manneans and Arabs. For the next 60 or so years, Babylon and Chaldea remained peacefully under direct Assyrian control. The Chaldeans remained subjugated and quiet during this period, and the next major revolt in Babylon against the Assyrian empire was fermented not by a Chaldean, Babylonian or Elamite, but by Shamash-shum-ukin, who was an Assyrian king of Babylon, and elder brother of Ashurbanipal (668-627 BCE), the new ruler of the Neo-Assyrian Empire. Shamash-shum-ukin (668–648 BCE) had become infused with Babylonian nationalism after sixteen years peacefully subject to his brother, and despite being Assyrian himself, declared that the city of Babylon and not Nineveh or Ashur should be the seat of the empire. In 652 BCE, he raised a powerful coalition of peoples resentful of their subjugation to Assyria against his own brother Ashurbanipal. The alliance included the Babylonians, Persians, Chaldeans, Medes, Elamites, Sultans, Arameans, Israelites, Arabs and Canaanites, together with some disaffected elements among the Assyrians themselves. After a bitter struggle lasting five years, the Assyrian king triumphed over his rebellious brother in 648 BCE, Elam was utterly destroyed, and the Babylonians, Persians, Medes, Chaldeans, Arabs, and others were savagely punished. An Assyrian governor named Kandalanu was then placed on the throne of Babylon to rule on behalf of Ashurbanipal. The next 22 years were peaceful, and neither the Babylonians nor Chaldeans posed a threat to the dominance of Ashurbanipal. However, after the death of the mighty Ashurbanipal (and Kandalanu) in 627 BCE, the Neo Assyrian Empire descended into a series of bitter internal dynastic civil wars that were to be the cause of its downfall. Ashur-etil-ilani (626–623 BCE) ascended to the throne of the empire in 626 BCE but was immediately engulfed in a torrent of fierce rebellions instigated by rival claimants. He was deposed in 623 BCE by an Assyrian general ("turtanu") named Sin-shumu-lishir (623–622 BCE), who was also declared king of Babylon. Sin-shar-ishkun (622–612 BCE), the brother of Ashur-etil-ilani, took back the throne of empire from Sin-shumu-lishir in 622 BCE, but was then himself faced with unremitting rebellion against his rule by his own people. Continual conflict among the Assyrians led to a myriad of subject peoples, from Cyprus to Persia and The Caucasus to Egypt, quietly reasserting their independence and ceasing to pay tribute to Assyria. Nabopolassar, a previously obscure and unknown Chaldean chieftain, followed the opportunistic tactics laid down by previous Chaldean leaders to take advantage of the chaos and anarchy gripping Assyria and Babylonia and seized the city of Babylon in 620 BCE with the help of its native Babylonian inhabitants. Sin-shar-ishkun amassed a powerful army and marched into Babylon to regain control of the region. Nabopolassar was saved from likely destruction because yet another massive Assyrian rebellion broke out in Assyria proper, including the capital Nineveh, which forced the Assyrian king to turn back in order to quell the revolt. Nabopolassar took advantage of this situation, seizing the ancient city of Nippur in 619 BCE, a mainstay of pro-Assyrianism in Babylonia, and thus Babylonia as a whole. However, his position was still far from secure, and bitter fighting continued in the Babylonian heartlands from 620 to 615 BCE, with Assyrian forces encamped in Babylonia in an attempt to eject Nabopolassar. Nabopolassar attempted a counterattack, marched his army into Assyria proper in 616 BCE, and tried to besiege Assur and Arrapha (modern Kirkuk), but was defeated by Sin-shar-ishkun and chased back into Babylonia after being driven from Idiqlat (modern Tikrit) at the southernmost end of Assyria. A stalemate seemed to have ensued, with Nabopolassar unable to make any inroads into Assyria despite its greatly weakened state, and Sin-shar-ishkun unable to eject Nabopolassar from Babylonia due to constant rebellions and civil war among his own people. Nabopolassar's position, and the fate of the Assyrian empire, was sealed when he entered into an alliance with another of Assyria's former vassals, the Medes, the now dominant people of what was to become Persia. The Median Cyaxares had also recently taken advantage of the anarchy in the Assyrian Empire, while officially still a vassal of Assyria, he took the opportunity to meld the Iranian peoples; the Medes, Persians, Sagartians and Parthians, into a large and powerful Median-dominated force. The Medes, Persians, Parthians, Chaldeans and Babylonians formed an alliance that also included the Scythians and Cimmerians to the north. While Sin-shar-ishkun was fighting both the rebels in Assyria and the Chaldeans and Babylonians in southern Mesopotamia, Cyaxares (hitherto a vassal of Assyria), in alliance with the Scythians and Cimmerians launched a surprise attack on civil-war-bleaguered Assyria in 615 BCE, sacking Kalhu (the Biblical Calah/Nimrud) and taking Arrapkha (modern Kirkuk). Nabopolassar, still pinned down in southern Mesopotamia, was not involved in this major breakthrough against Assyria. From this point however, the alliance of Medes, Persians, Chaldeans, Babylonians, Sagartians, Scythians and Cimmerians fought in unison against Assyria. Despite the sorely depleted state of Assyria, bitter fighting ensued. Throughout 614 BCE the alliance of powers continued to make inroads into Assyria itself, although in 613 BCE the Assyrians somehow rallied to score a number of counterattacking victories over the Medes-Persians, Babylonians-Chaldeans and Scythians-Cimmerians. This led to a coalition of forces ranged against it to unite and launch a massive combined attack in 612 BCE, finally besieging and sacking Nineveh in late 612 BCE, killing Sin-shar-ishkun in the process. A new Assyrian king, Ashur-uballit II (612–605 BCE), took the crown amidst the house-to-house fighting in Nineveh, and refused a request to bow in vassalage to the rulers of the alliance. He managed to fight his way out of Nineveh and reach the northern Assyrian city of Harran, where he founded a new capital. Assyria resisted for another seven years until 605 BCE, when the remnants of the Assyrian army and the army of the Egyptians (whose dynasty had also been installed as puppets of the Assyrians) were defeated at Karchemish. Nabopolassar and his Median, Scythian and Cimmerian allies were now in possession of much of the huge Neo Assyrian Empire. The Egyptians had belatedly come to the aid of Assyria, fearing that, without Assyrian protection, they would be the next to succumb to the new powers, having already been raided by the Scythians. The Chaldean king of Babylon now ruled all of southern Mesopotamia (Assyria in the north was ruled by the Medes), and the former Assyrian possessions of Aram (Syria), Phoenicia, Israel, Cyprus, Edom, Philistia, and parts of Arabia, while the Medes took control of the former Assyrian colonies in Ancient Iran, Asia Minor and the Caucasus. Nabopolassar was not able to enjoy his success for long, dying in 604 BCE, only one year after the victory at Karchemish. He was succeeded by his son, who took the name Nebuchadnezzar II, after the unrelated 12th century BCE native Akkadian-Babylonian king Nebuchadnezzar I, indicating the extent to which the migrant Chaldeans had become infused with native Mesopotamian culture. Nebuchadnezzar II and his allies may well have been forced to deal with remnants of Assyrian resistance based in and around Dur-Katlimmu, as Assyrian imperial records continue to be dated in this region between 604 and 599 BCE. In addition, the Egyptians remained in the region, possibly in an attempt to aid their former Assyrian masters and to carve out an empire of their own. Nebuchadnezzar II was to prove himself to be the greatest of the Chaldean rulers, rivaling another non-native ruler, the 18th century BCE Amorite king Hammurabi, as the greatest king of Babylon. He was a patron of the cities and a spectacular builder, rebuilding all of Babylonia's major cities on a lavish scale. His building activity at Babylon, expanding on the earlier major and impressive rebuilding of the Assyrian king Esarhaddon, helped to turn it into the immense and beautiful city of legend. Babylon covered more than three square miles, surrounded by moats and ringed by a double circuit of walls. The Euphrates flowed through the center of the city, spanned by a beautiful stone bridge. At the center of the city rose the giant ziggurat called Etemenanki, "House of the Frontier Between Heaven and Earth," which lay next to the Temple of Marduk. He is also believed by many historians to have built The Hanging Gardens of Babylon (although others believe these gardens were built much earlier by an Assyrian king in Nineveh) for his wife, a Median princess from the green mountains, so that she would feel at home. A capable leader, Nebuchadnezzar II conducted successful military campaigns; cities like Tyre, Sidon and Damascus were subjugated. He also conducted numerous campaigns in Asia Minor against the Scythians, Cimmerians, and Lydians. Like their Assyrian relations, the Babylonians had to campaign yearly in order to control their colonies. In 601 BCE, Nebuchadnezzar II was involved in a major but inconclusive battle against the Egyptians. In 599 BCE, he invaded Arabia and routed the Arabs at Qedar. In 597 BCE, he invaded Judah, captured Jerusalem, and deposed its king Jehoiachin. Egyptian and Babylonian armies fought each other for control of the Near East throughout much of Nebuchadnezzar's reign, and this encouraged king Zedekiah of Judah to revolt. After an eighteen-month siege, Jerusalem was captured in 587 BCE, thousands of Jews were deported to Babylon, and Solomon's Temple was razed to the ground. Nebuchadnezzar successfully fought the Pharaohs Psammetichus II and Apries throughout his reign, and during the reign of Pharaoh Amasis in 568 BCE it is rumoured that he may have briefly invaded Egypt itself. By 572, Nebuchadnezzar was in full control of Babylonia, Chaldea, Aramea (Syria), Phonecia, Israel, Judah, Philistia, Samarra, Jordan, northern Arabia, and parts of Asia Minor. Nebuchadnezzar died of illness in 562 BCE after a one-year co-reign with his son, Amel-Marduk, who was deposed in 560 BCE after a reign of only two years. Neriglissar succeeded Amel-Marduk. It is unclear as to whether he was in fact an ethnic Chaldean or a native Babylonian nobleman, as he was not related by blood to Nabopolassar's descendants, having married into the ruling family. He conducted successful military campaigns against the Hellenic inhabitants of Cilicia, which had threatened Babylonian interests. Neriglissar reigned for only four years and was succeeded by the youthful Labashi-Marduk in 556 BCE. Again, it is unclear whether he was a Chaldean or a native Babylonian. Labashi-Marduk reigned only for a matter of months, being deposed by Nabonidus in late 556 BCE. Nabonidus was certainly not a Chaldean, but an Assyrian from Harran, the last capital of Assyria, and proved to be the final native Mesopotamian king of Babylon. He and his son, the regent Belshazzar, were deposed by the Persians under Cyrus II in 539 BCE. When the Babylonian Empire was absorbed into the Persian Achaemenid Empire, the name "Chaldean" lost its meaning in reference to a particular ethnicity or land, but lingered for a while as a term solely and explicitly used to describe a societal class of astrologers and astronomers in southern Mesopotamia. The original Chaldean tribe had long ago became Akkadianized, adopting Akkadian culture, religion, language and customs, blending into the majority native population, and eventually wholly disappearing as a distinct race of people, as had been the case with other preceding migrant peoples, such as the Amorites, Kassites, Suteans and Arameans of Babylonia. The Persians considered this "Chaldean societal class" to be masters of reading and writing, and especially versed in all forms of incantation, sorcery, witchcraft, and the magical arts. They spoke of astrologists and astronomers as "Chaldeans", and it is used with this specific meaning in the Book of Daniel (Dan. i. 4, ii. 2 et seq.) and by classical writers, such as Strabo. The disappearance of the Chaldeans as an ethnicity and Chaldea as a land is evidenced by the fact that the Persian rulers of the Achaemenid Empire (539–330 BCE) did not retain a province called "Chaldea", nor did they refer to "Chaldeans" as a race of people in their written annals. This is in contrast to Assyria, and for a time Babylonia also, where the Persians retained the names Assyria and Babylonia as designations for distinct geo-political entities within the Achaemenid Empire. In the case of the Assyrians in particular, Achaemenid records show Assyrians holding important positions within the empire, particularly with regards to military and civil administration. This absence of Chaldeans from the historical record continues throughout the Macedonian Empire, Seleucid Empire, Parthian Empire, Roman Empire, Sassanid Empire, Byzantine Empire and after the Arab Islamic conquest and Mongol Empire. By the time of Cicero in the 2nd century BCE, "Chaldean" appears to have completely disappeared even as a societal term for Babylonian astronomers and astrologers; Cicero refers to "Babylonian astrologers" rather than Chaldean astrologers. Horace does the same, referring to "Babylonian horoscopes" rather than Chaldean in his famous Carpe Diem ode. Cicero views the Babylonian astrologers as holding obscure knowledge, while Horace thinks that they are wasting their time and would be happier "going with the flow". The terms Chaldee and Chaldean were henceforth only found in Hebraic and Biblical sources dating from the 6th and 5th centuries BCE, and referring specifically to the period of the Chaldean Dynasty of Babylon. After an absence from history of 2,236 years, the name was revived in 1683 AD by the Roman Catholic Church in the form of the Chaldean Catholic Church as the new name for the "Church of Assyria and Mosul" (so named in 1553 AD). This was a church founded and populated not by the long extinct Chaldean tribe of southeastern Mesopotamia, who had disappeared from the pages of history over twenty two centuries previously, but founded in northern Mesopotamia by a breakaway group of ethnic Assyrians long indigenous to Upper Mesopotamia (Assyria) who had hitherto been members of the Assyrian Church of the East before entering communion with Rome.
https://en.wikipedia.org/wiki?curid=6599
Currying In mathematics and computer science, currying is the technique of converting a function that takes multiple arguments into a sequence of functions that each take a single argument. For example, currying a function formula_1 that takes three arguments creates three functions: formula_2 Or more abstractly, a function that takes two arguments, one from formula_3 and one from formula_4, and produces outputs in formula_5 by currying is translated into a function that takes a single argument from formula_3 and produces as outputs "functions" from formula_4 to formula_8 Currying is related to, but not the same as, partial application. Currying is useful in both practical and theoretical settings. In functional programming languages, and many others, it provides a way of automatically managing how arguments are passed to functions and exceptions. In theoretical computer science, it provides a way to study functions with multiple arguments in simpler theoretical models which provide only one argument. The most general setting for the strict notion of currying and uncurrying is in the closed monoidal categories, which underpins a vast generalization of the Curry–Howard correspondence of proofs and programs to a correspondence with many other structures, including quantum mechanics, cobordisms and string theory. It was introduced by Gottlob Frege, developed by Moses Schönfinkel, and further developed by Haskell Curry. Uncurrying is the dual transformation to currying, and can be seen as a form of defunctionalization. It takes a function formula_1 whose return value is another function formula_10, and yields a new function formula_11 that takes as parameters the arguments for both formula_1 and formula_10, and returns, as a result, the application of formula_1 and subsequently, formula_10, to those arguments. The process can be iterated. Currying provides a way for working with functions that take multiple arguments, and using them in frameworks where functions might take only one argument. For example, some analytical techniques can only be applied to functions with a single argument. Practical functions frequently take more arguments than this. Frege showed that it was sufficient to provide solutions for the single argument case, as it was possible to transform a function with multiple arguments into a chain of single-argument functions instead. This transformation is the process now known as currying. All "ordinary" functions that might typically be encountered in mathematical analysis or in computer programming can be curried. However, there are categories in which currying is not possible; the most general categories which allow currying are the closed monoidal categories. Some programming languages almost always use curried functions to achieve multiple arguments; notable examples are ML and Haskell, where in both cases all functions have exactly one argument. This property is inherited from lambda calculus, where multi-argument functions are usually represented in curried form. Currying is related to, but not the same as partial application. In practice, the programming technique of closures can be used to perform partial application and a kind of currying, by hiding arguments in an environment that travels with the curried function. Suppose we have a function formula_16 which takes two real arguments and outputs real numbers, and it is defined by formula_17. Currying translates this into a function formula_18 which takes a single real argument and outputs functions from formula_19 to formula_19. In symbols, formula_21, where formula_22denotes the set of all functions that take a single real argument and produce real outputs. For every real number formula_23, define the function formula_24 by formula_25, and then define the function formula_21 by formula_27. So for instance, formula_28 is the function that sends its real argument formula_29 to the output formula_30, or formula_31. We see that in general so that the original function formula_1 and its currying formula_18 convey exactly the same information. In this situation, we also write This also works for functions with more than two arguments. If formula_1 were a function of three arguments formula_37, its currying formula_18 would have the property The name "currying", coined by Christopher Strachey in 1967, is a reference to logician Haskell Curry. The alternative name "Schönfinkelisation" has been proposed as a reference to Moses Schönfinkel. In the mathematical context, the principle can be traced back to work in 1893 by Frege. Currying is most easily understood by starting with an informal definition, which can then be molded to fit many different domains. First, there is some notation to be established. The notation formula_40 denotes all functions from formula_3 to formula_4. If formula_1 is such a function, we write formula_44. Let formula_45 denote the ordered pairs of the elements of formula_3 and formula_4 respectively, that is, the Cartesian product of formula_3 and formula_4. Here, formula_3 and formula_4 may be sets, or they may be types, or they may be other kinds of objects, as explored below. Given a function currying constructs a new function That is, formula_54 takes an argument from formula_3 and returns a function that maps formula_4 to formula_57. It is defined by for formula_23 from formula_3 and formula_29 from formula_4. We then also write Uncurrying is the reverse transformation, and is most easily understood in terms of its right adjoint, the function formula_64 In set theory, the notation formula_65 is used to denote the set of functions from the set formula_3 to the set formula_4. Currying is the natural bijection between the set formula_68 of functions from formula_69 to formula_70, and the set formula_71 of functions from formula_72 to the set of functions from formula_73 to formula_70. In symbols: Indeed, it is this natural bijection that justifies the exponential notation for the set of functions. As is the case in all instances of currying, the formula above describes an adjoint pair of functors: for every fixed set formula_73, the functor formula_77 is left adjoint to the functor formula_78. In the category of sets, the object formula_65 is called the exponential object. In the theory of function spaces, such as in functional analysis or homotopy theory, one is commonly interested in continuous functions between topological spaces. One writes formula_80 (the Hom functor) for the set of "all" functions from formula_3 to formula_4, and uses the notation formula_65 to denote the subset of continuous functions. Here, formula_84 is the bijection while uncurrying is the inverse map. If the set formula_65 of continuous functions from formula_3 to formula_4 is given the compact-open topology, and if the space formula_4 is locally compact Hausdorff, then is a homeomorphism. This is also the case when formula_3, formula_4 and formula_65 are compactly generated, although there are more cases. One useful corollary is that a function is continuous if and only if its curried form is continuous. Another important result is that the application map, usually called "evaluation" in this context, is continuous (note that eval is a strictly different concept in computer science.) That is, formula_94 is continuous when formula_65 is compact-open and formula_4 locally compact Hausdorff. These two results are central for establishing the continuity of homotopy, i.e. when formula_3 is the unit interval formula_98, so that formula_99 can the thought of as either a homotopy of two functions from formula_4 to formula_57, or, equivalently, a single (continuous) path in formula_102. In algebraic topology, currying serves as an example of Eckmann–Hilton duality, and, as such, plays an important role in a variety of different settings. For example, loop space is adjoint to reduced suspensions; this is commonly written as where formula_104 is the set of homotopy classes of maps formula_105, and formula_106 is the suspension of "A", and formula_107 is the loop space of "A". In essence, the suspension formula_108 can be seen as the cartesian product of formula_3 with the unit interval, modulo an equivalence relation to turn the interval into a loop. The curried form then maps the space formula_3 to the space of functions from loops into formula_57, that is, from formula_3 into formula_113. Then formula_84 is the adjoint functor that maps suspensions to loop spaces, and uncurrying is the dual. The duality between the mapping cone and the mapping fiber (cofibration and fibration) can be understood as a form of currying, which in turn leads to the duality of the long exact and coexact Puppe sequences. In homological algebra, the relationship between currying and uncurrying is known as tensor-hom adjunction. Here, an interesting twist arises: the Hom functor and the tensor product functor might not lift to an exact sequence; this leads to the definition of the Ext functor and the Tor functor. In order theory, that is, the theory of lattices of partially ordered sets, formula_84 is a continuous function when the lattice is given the Scott topology. Scott-continuous functions were first investigated in the attempt to provide a semantics for lambda calculus (as ordinary set theory is inadequate to do this). More generally, Scott-continuous functions are now studied in domain theory, which encompasses the study of denotational semantics of computer algorithms. Note that the Scott topology is quite different than many common topologies one might encounter in the category of topological spaces; the Scott topology is typically finer, and is not sober. The notion of continuity makes its appearance in homotopy type theory, where, roughly speaking, two computer programs can be considered to be homotopic, i.e. compute the same results, if they can be "continuously" refactored from one to the other. In theoretical computer science, currying provides a way to study functions with multiple arguments in very simple theoretical models, such as the lambda calculus, in which functions only take a single argument. Consider a function formula_116 taking two arguments, and having the type formula_117, which should be understood to mean that "x" must have the type formula_3, "y" must have the type formula_4, and the function itself returns the type formula_57. The curried form of "f" is defined as where formula_122 is the abstractor of lambda calculus. Since curry takes, as input, functions with the type formula_123, one concludes that the type of curry itself is The → operator is often considered right-associative, so the curried function type formula_125 is often written as formula_126. Conversely, function application is considered to be left-associative, so that formula_127 is equivalent to That is, the parenthesis are not required to disambiguate the order of the application. Curried functions may be used in any programming language that supports closures; however, uncurried functions are generally preferred for efficiency reasons, since the overhead of partial application and closure creation can then be avoided for most function calls. In type theory, the general idea of a type system in computer science is formalized into a specific algebra of types. For example, when writing formula_44, the intent is that formula_3 and formula_4 are types, while the arrow formula_132 is a type constructor, specifically, the function type or arrow type. Similarly, the Cartesian product formula_45 of types is constructed by the product type constructor formula_134. The type-theoretical approach is expressed in programming languages such as ML and the languages derived from and inspired by it: CaML, Haskell and F#. The type-theoretical approach provides a natural complement to the language of category theory, as discussed below. This is because categories, and specifically, monoidal categories, have an internal language, with simply-typed lambda calculus being the most prominent example of such a language. It is important in this context, because it can be built from a single type constructor, the arrow type. Currying then endows the language with a natural product type. The correspondence between objects in categories and types then allows programming languages to be re-interpreted as logics (via Curry–Howard correspondence), and as other types of mathematical systems, as explored further, below. Under the Curry–Howard correspondence, the existence of currying and uncurrying is equivalent to the logical theorem formula_135, as tuples (product type) corresponds to conjunction in logic, and function type corresponds to implication. The exponential object formula_136 in the category of Heyting algebras is normally written as material implication formula_137. Distributive Heyting algebras are Boolean algebras, and the exponential object has the explicit form formula_138, thus making it clear that the exponential object really is material implication. The above notions of currying and uncurrying find their most general, abstract statement in category theory. Currying is a universal property of an exponential object, and gives rise to an adjunction in cartesian closed categories. That is, there is a natural isomorphism between the morphisms from a binary product formula_52 and the morphisms to an exponential object formula_140. This generalizes to a broader result in closed monoidal categories: Currying is the statement that the tensor product and the internal Hom are adjoint functors; that is, for every object formula_72 there is a natural isomorphism: Here, "Hom" denotes the (external) Hom-functor of all morphisms in the category, while formula_143 denotes the internal hom functor in the closed monoidal category. For the category of sets, the two are the same. When the product is the cartesian product, then the internal hom formula_143 becomes the exponential object formula_145. Currying can break down in one of two ways. One is if a category is not closed, and thus lacks an internal hom functor (possibly because there is more than one choice for such a functor). Another ways is if it is not monoidal, and thus lacks a product (that is, lacks a way of writing down pairs of objects). Categories that do have both products and internal homs are exactly the closed monoidal categories. The setting of cartesian closed categories is sufficient for the discussion of classical logic; the more general setting of closed monoidal categories is suitable for quantum computation. The difference between these two is that the product for cartesian categories (such as the category of sets, complete partial orders or Heyting algebras) is just the Cartesian product; it is interpreted as an ordered pair of items (or a list). Simply typed lambda calculus is the internal language of cartesian closed categories; and it is for this reason that pairs and lists are the primary types in the type theory of LISP, scheme and many functional programming languages. By contrast, the product for monoidal categories (such as Hilbert space and the vector spaces of functional analysis) is the tensor product. The internal language of such categories is linear logic, a form of quantum logic; the corresponding type system is the linear type system. Such categories are suitable for describing entangled quantum states, and, more generally, allow a vast generalization of the Curry–Howard correspondence to quantum mechanics, to cobordisms in algebraic topology, and to string theory. The linear type system, and linear logic are useful for describing synchronization primitives, such as mutual exclusion locks, and the operation of vending machines. Currying and partial function application are often conflated. One of the significant differences between the two is that a call to a partially applied function returns the result right away, not another function down the currying chain; this distinction can be illustrated clearly for functions whose arity is greater than two. Given a function of type formula_146, currying produces formula_147. That is, while an evaluation of the first function might be represented as formula_148, evaluation of the curried function would be represented as formula_149, applying each argument in turn to a single-argument function returned by the previous invocation. Note that after calling formula_150, we are left with a function that takes a single argument and returns another function, not a function that takes two arguments. In contrast, partial function application refers to the process of fixing a number of arguments to a function, producing another function of smaller arity. Given the definition of formula_1 above, we might fix (or 'bind') the first argument, producing a function of type formula_152. Evaluation of this function might be represented as formula_153. Note that the result of partial function application in this case is a function that takes two arguments. Intuitively, partial function application says "if you fix the first argument of the function, you get a function of the remaining arguments". For example, if function "div" stands for the division operation "x"/"y", then "div" with the parameter "x" fixed at 1 (i.e., "div" 1) is another function: the same as the function "inv" that returns the multiplicative inverse of its argument, defined by "inv"("y") = 1/"y". The practical motivation for partial application is that very often the functions obtained by supplying some but not all of the arguments to a function are useful; for example, many languages have a function or operator similar to codice_1. Partial application makes it easy to define these functions, for example by creating a function that represents the addition operator with 1 bound as its first argument. Partial application can be seen as evaluating a curried function at a fixed point, e.g. given formula_146 and formula_155 then formula_156 or simply formula_157 where formula_158 curries f's first parameter. Thus, partial application is reduced to a curried function at a fixed point. Further, a curried function at a fixed point is (trivially), a partial application. For further evidence, note that, given any function formula_116, a function formula_160 may be defined such that formula_161. Thus, any partial application may be reduced to a single curry operation. As such, curry is more suitably defined as an operation which, in many theoretical cases, is often applied recursively, but which is theoretically indistinguishable (when considered as an operation) from a partial application. So, a partial application can be defined as the objective result of a single application of the curry operator on some ordering of the inputs of some function.
https://en.wikipedia.org/wiki?curid=6600
Cyrus Cyrus (Persian: کوروش) is a male given name. It is the given name of a number of Persian kings. Most notably it refers to Cyrus the Great. Cyrus is also the name of Cyrus I of Anshan (ca. 650 BC), King of Persia the grandfather of Cyrus the Great; and Cyrus the Younger (died 401 BC), brother to the Persian King Artaxerxes II of Persia. Cyrus, as a word in English, is the Latinized form of the Greek Κῦρος, "Kȳros", from Old Persian "Kūruš". According to the inscriptions the name is reflected in Elamite "Kuraš", Babylonian "Ku(r)-raš/-ra-áš" and Imperial Aramaic "kwrš". The modern Persian form of the name is Kourosh. The etymology of Cyrus has been and continues to be a topic of discussion amongst historians, linguists, and scholars of Iranology. The Old Persian name "kuruš" has been interpreted in various forms from "the Sun", "like Sun", "young", "hero" to "humiliator of the enemy in verbal contest" and the Elamite "kuraš" has been translated as one "who bestows care". The name has appeared on many monuments and inscriptions in Old Persian. There is also the record of a small inscription in Morghab (southwestern Iran) on which there is the sentence ("adam kūruš xšāyaƟiya haxāmanišiya") in Old Persian meaning ("I am Cyrus the Achaemenian King"). After a questionable proposal by the German linguist F. H. Weissbach that Darius the Great was the first to inscribe in Persian, it had previously been concluded by some scholars that the inscription in Morghab refers to Cyrus the Younger. This proposal was the result of a false interpretation of a passage in paragraph 70 of Behistun inscription by Darius the Great. Based on many arguments, the accepted theory among modern scholars is that the inscription does belong to Cyrus the Great. There are interpretations of name of Cyrus by classical authors identifying with or referring to the Persian word for “Sun”. The Historian Plutarch (46 - 120) states that "the sun, which, in the Persian language, is called Cyrus". Also the Physician Ctesias who served in the court of the Persian king Artaxerxes II of Persia writes in his book "Persica" as summarized by Photios that the name Cyrus is from Persian word "Khur" (the sun). These are, however, not accepted by modern scholars. Regarding the etymology of Old Persian "kuruš", linguists have proposed various etymologies based on Iranian languages as well as non-Indo-European ones. According to Tavernier, the name "kuraš", attested in Elamite texts, is likely "the original form" as there is no Elamite or Babylonian spelling "ku-ru-uš" in the transcriptions of Old Persian "ku-u-r(u)-u-š". That is, according to Tavernier, "kuraš" is an Elamite name and means "to bestow care". Others, such as Schmitt, Hoffmann maintain that the Persian "Kuruš", which according to Skalmowsky, may be connected to (or a borrowing from) the IE "Kúru-" from Old Indic can give an etymology of the Elamite "kuraš". In this regard the Old Persian "kuruš" is considered with the following etymologies: One proposal is discussed by the linguist Janos Harmatta that refers to the common Iranian root "kur-" (be born) of many words in Old, middle, and new Iranian languages (e.g. Kurdish). Accordingly, the name Kūruš means "young, youth...". Other Iranian etymologies have been proposed. The Indian proposal of Skalmowsky goes down to "to do, accomplish". Another theory is the suggestion of Karl Hoffmann that "kuruš" goes down to a "-ru" derivation from the IE root "*(s)kau" meaning "to humiliate" and accordingly "kuruš" (hence "Cyrus") means ""humiliator" (of the enemy in verbal contest)". People and fictional characters named Cyrus include:
https://en.wikipedia.org/wiki?curid=6601