text
stringlengths
10
951k
source
stringlengths
39
44
Feminism Feminism is a range of social movements, political movements, and ideologies that aim to define, establish, and achieve the political, economic, personal, and social equality of the sexes. Feminism incorporates the position that societies prioritize the male point of view, and that women are treated unfairly within those societies. Efforts to change that include fighting gender stereotypes and seeking to establish educational and professional opportunities for women that are equal to those for men. Feminist movements have campaigned and continue to campaign for women's rights, including the right to vote, to hold public office, to work, to earn fair wages, equal pay and eliminate the gender pay gap, to own property, to receive education, to enter contracts, to have equal rights within marriage, and to have maternity leave. Feminists have also worked to ensure access to legal abortions and social integration and to protect women and girls from rape, sexual harassment, and domestic violence. Changes in dress and acceptable physical activity have often been part of feminist movements. Some scholars consider feminist campaigns to be a main force behind major historical societal changes for women's rights, particularly in the West, where they are near-universally credited with achieving women's suffrage, gender-neutral language, reproductive rights for women (including access to contraceptives and abortion), and the right to enter into contracts and own property. Although feminist advocacy is, and has been, mainly focused on women's rights, some feminists, including bell hooks, argue for the inclusion of men's liberation within its aims, because they believe that men are also harmed by traditional gender roles. Feminist theory, which emerged from feminist movements, aims to understand the nature of gender inequality by examining women's social roles and lived experience; it has developed theories in a variety of disciplines in order to respond to issues concerning gender. Numerous feminist movements and ideologies have developed over the years and represent different viewpoints and aims. Some forms of feminism have been criticized for taking into account only white, middle class, and college-educated perspectives. This criticism led to the creation of ethnically specific or multicultural forms of feminism, including black feminism and intersectional feminism. Charles Fourier, a utopian socialist and French philosopher, is credited with having coined the word "féminisme" in 1837. The words "féminisme" ("feminism") and "féministe" ("feminist") first appeared in France and the Netherlands in 1872, Great Britain in the 1890s, and the United States in 1910. The "Oxford English Dictionary" lists 1852 as the year of the first appearance of "feminist" and 1895 for "feminism". Depending on the historical moment, culture and country, feminists around the world have had different causes and goals. Most western feminist historians contend that all movements working to obtain women's rights should be considered feminist movements, even when they did not (or do not) apply the term to themselves. Other historians assert that the term should be limited to the modern feminist movement and its descendants. Those historians use the label "protofeminist" to describe earlier movements. The history of the modern western feminist movement is divided into four "waves". The first comprised women's suffrage movements of the 19th and early-20th centuries, promoting women's right to vote. The second wave, the women's liberation movement, began in the 1960s and campaigned for legal and social equality for women. In or around 1992, a third wave was identified, characterized by a focus on individuality and diversity. The fourth wave, from around 2012, used social media to combat sexual harassment, violence against women and rape culture; it is best known for the Me Too movement. First-wave feminism was a period of activity during the 19th and early-20th centuries. In the UK and US, it focused on the promotion of equal contract, marriage, parenting, and property rights for women. New legislation included the Custody of Infants Act 1839 in the UK, which introduced the tender years doctrine for child custody and gave women the right of custody of their children for the first time. Other legislation, such as the Married Women's Property Act 1870 in the UK and extended in the 1882 Act, became models for similar legislation in other British territories. Victoria passed legislation in 1884 and New South Wales in 1889; the remaining Australian colonies passed similar legislation between 1890 and 1897. With the turn of the 19th century, activism focused primarily on gaining political power, particularly the right of women's suffrage, though some feminists were active in campaigning for women's sexual, reproductive, and economic rights too. Women's suffrage (the right to vote and stand for parliamentary office) began in Britain's Australasian colonies at the close of the 19th century, with the self-governing colonies of New Zealand granting women the right to vote in 1893; South Australia followed suit in 1895. This was followed by Australia granting female suffrage in 1902. In Britain, the suffragettes and suffragists campaigned for the women's vote, and in 1918 the Representation of the People Act was passed granting the vote to women over the age of 30 who owned property. In 1928 this was extended to all women over 21. Emmeline Pankhurst was the most notable activist in England. "Time" named her one of the , stating: "she shaped an idea of women for our time; she shook society into a new pattern from which there could be no going back." In the US, notable leaders of this movement included Lucretia Mott, Elizabeth Cady Stanton, and Susan B. Anthony, who each campaigned for the abolition of slavery before championing women's right to vote. These women were influenced by the Quaker theology of spiritual equality, which asserts that men and women are equal under God. In the US, first-wave feminism is considered to have ended with the passage of the Nineteenth Amendment to the United States Constitution (1919), granting women the right to vote in all states. The term "first wave" was coined retroactively when the term "second-wave feminism" came into use. During the late Qing period and reform movements such as the Hundred Days' Reform, Chinese feminists called for women's liberation from traditional roles and Neo-Confucian gender segregation. Later, the Chinese Communist Party created projects aimed at integrating women into the workforce, and claimed that the revolution had successfully achieved women's liberation. According to Nawar al-Hassan Golley, Arab feminism was closely connected with Arab nationalism. In 1899, Qasim Amin, considered the "father" of Arab feminism, wrote "The Liberation of Women", which argued for legal and social reforms for women. He drew links between women's position in Egyptian society and nationalism, leading to the development of Cairo University and the National Movement. In 1923 Hoda Shaarawi founded the Egyptian Feminist Union, became its president and a symbol of the Arab women's rights movement. The Iranian Constitutional Revolution in 1905 triggered the Iranian women's movement, which aimed to achieve women's equality in education, marriage, careers, and legal rights. However, during the Iranian revolution of 1979, many of the rights that women had gained from the women's movement were systematically abolished, such as the Family Protection Law. In France, women obtained the right to vote only with the Provisional Government of the French Republic of 21 April 1944. The Consultative Assembly of Algiers of 1944 proposed on 24 March 1944 to grant eligibility to women but following an amendment by Fernand Grenier, they were given full citizenship, including the right to vote. Grenier's proposition was adopted 51 to 16. In May 1947, following the November 1946 elections, the sociologist Robert Verdier minimized the "gender gap", stating in "Le Populaire" that women had not voted in a consistent way, dividing themselves, as men, according to social classes. During the baby boom period, feminism waned in importance. Wars (both World War I and World War II) had seen the provisional emancipation of some women, but post-war periods signalled the return to conservative roles. By the mid-20th century, women still lacked significant rights. In Switzerland, women gained the right to vote in federal elections in 1971; but in the canton of Appenzell Innerrhoden women obtained the right to vote on local issues only in 1991, when the canton was forced to do so by the Federal Supreme Court of Switzerland. In Liechtenstein, women were given the right to vote by the women's suffrage referendum of 1984. Three prior referendums held in 1968, 1971 and 1973 had failed to secure women's right to vote. Feminists continued to campaign for the reform of family laws which gave husbands control over their wives. Although by the 20th century coverture had been abolished in the UK and US, in many continental European countries married women still had very few rights. For instance, in France, married women did not receive the right to work without their husband's permission until 1965. Feminists have also worked to abolish the "marital exemption" in rape laws which precluded the prosecution of husbands for the rape of their wives. Earlier efforts by first-wave feminists such as Voltairine de Cleyre, Victoria Woodhull and Elizabeth Clarke Wolstenholme Elmy to criminalize marital rape in the late 19th century had failed; this was only achieved a century later in most Western countries, but is still not achieved in many other parts of the world. French philosopher Simone de Beauvoir provided a Marxist solution and an existentialist view on many of the questions of feminism with the publication of "Le Deuxième Sexe" ("The Second Sex") in 1949. The book expressed feminists' sense of injustice. Second-wave feminism is a feminist movement beginning in the early 1960s and continuing to the present; as such, it coexists with third-wave feminism. Second-wave feminism is largely concerned with issues of equality beyond suffrage, such as ending gender discrimination. Second-wave feminists see women's cultural and political inequalities as inextricably linked and encourage women to understand aspects of their personal lives as deeply politicized and as reflecting sexist power structures. The feminist activist and author Carol Hanisch coined the slogan "The Personal is Political", which became synonymous with the second wave. Second- and third-wave feminism in China has been characterized by a reexamination of women's roles during the communist revolution and other reform movements, and new discussions about whether women's equality has actually been fully achieved. In 1956, President Gamal Abdel Nasser of Egypt initiated "state feminism", which outlawed discrimination based on gender and granted women's suffrage, but also blocked political activism by feminist leaders. During Sadat's presidency, his wife, Jehan Sadat, publicly advocated further women's rights, though Egyptian policy and society began to move away from women's equality with the new Islamist movement and growing conservatism. However, some activists proposed a new feminist movement, Islamic feminism, which argues for women's equality within an Islamic framework. In Latin America, revolutions brought changes in women's status in countries such as Nicaragua, where feminist ideology during the Sandinista Revolution aided women's quality of life but fell short of achieving a social and ideological change. In 1963, Betty Friedan's book "The Feminine Mystique" helped voice the discontent that American women felt. The book is widely credited with sparking the beginning of second-wave feminism in the United States. Within ten years, women made up over half the First World workforce. Third-wave feminism is traced to the emergence of the Riot grrrl feminist punk subculture in Olympia, Washington, in the early 1990s, and to Anita Hill's televised testimony in 1991—to an all-male, all-white Senate Judiciary Committee—that Clarence Thomas, nominated for the Supreme Court of the United States, had sexually harassed her. The term "third wave" is credited to Rebecca Walker, who responded to Thomas's appointment to the Supreme Court with an article in "Ms." magazine, "Becoming the Third Wave" (1992). She wrote: Third-wave feminism also sought to challenge or avoid what it deemed the second wave's essentialist definitions of femininity, which, third-wave feminists argued, over-emphasized the experiences of upper middle-class white women. Third-wave feminists often focused on "micro-politics" and challenged the second wave's paradigm as to what was, or was not, good for women, and tended to use a post-structuralist interpretation of gender and sexuality. Feminist leaders rooted in the second wave, such as Gloria Anzaldúa, bell hooks, Chela Sandoval, Cherríe Moraga, Audre Lorde, Maxine Hong Kingston, and many other non-white feminists, sought to negotiate a space within feminist thought for consideration of race-related subjectivities. Third-wave feminism also contained internal debates between difference feminists, who believe that there are important psychological differences between the sexes, and those who believe that there are no inherent psychological differences between the sexes and contend that gender roles are due to social conditioning. Standpoint theory is a feminist theoretical point of view stating that a person's social position influences their knowledge. This perspective argues that research and theory treat women and the feminist movement as insignificant and refuses to see traditional science as unbiased. Since the 1980s, standpoint feminists have argued that the feminist movement should address global issues (such as rape, incest, and prostitution) and culturally specific issues (such as female genital mutilation in some parts of Africa and Arab societies, as well as glass ceiling practices that impede women's advancement in developed economies) in order to understand how gender inequality interacts with racism, homophobia, classism and colonization in a "matrix of domination". Fourth-wave feminism refers to a resurgence of interest in feminism that began around 2012 and is associated with the use of social media. According to feminist scholar Prudence Chamberlain, the focus of the fourth wave is justice for women and opposition to sexual harassment and violence against women. Its essence, she writes, is "incredulity that certain attitudes can still exist". Fourth-wave feminism is "defined by technology", according to Kira Cochrane, and is characterized particularly by the use of Facebook, Twitter, Instagram, YouTube, Tumblr, and blogs such as Feministing to challenge misogyny and further gender equality. Issues that fourth-wave feminists focus on include street and workplace harassment, campus sexual assault and rape culture. Scandals involving the harassment, abuse, and murder of women and girls have galvanized the movement. These have included the 2012 Delhi gang rape, 2012 Jimmy Savile allegations, the Bill Cosby allegations, 2014 Isla Vista killings, 2016 trial of Jian Ghomeshi, 2017 Harvey Weinstein allegations and subsequent Weinstein effect, and the 2017 Westminster sexual scandals. Examples of fourth-wave feminist campaigns include the Everyday Sexism Project, No More Page 3, Stop Bild Sexism, "Mattress Performance", "10 Hours of Walking in NYC as a Woman", #YesAllWomen, Free the Nipple, One Billion Rising, the 2017 Women's March, the 2018 Women's March, and the #MeToo movement. In December 2017, "Time" magazine chose several prominent female activists involved in the #MeToo movement, dubbed "the silence breakers", as Person of the Year. The term postfeminism is used to describe a range of viewpoints reacting to feminism since the 1980s. While not being "anti-feminist", postfeminists believe that women have achieved second wave goals while being critical of third- and fourth-wave feminist goals. The term was first used to describe a backlash against second-wave feminism, but it is now a label for a wide range of theories that take critical approaches to previous feminist discourses and includes challenges to the second wave's ideas. Other postfeminists say that feminism is no longer relevant to today's society. Amelia Jones has written that the postfeminist texts which emerged in the 1980s and 1990s portrayed second-wave feminism as a monolithic entity. Dorothy Chunn notes a "blaming narrative" under the postfeminist moniker, where feminists are undermined for continuing to make demands for gender equality in a "post-feminist" society, where "gender equality has (already) been achieved." According to Chunn, "many feminists have voiced disquiet about the ways in which rights and equality discourses are now used against them." Feminist theory is the extension of feminism into theoretical or philosophical fields. It encompasses work in a variety of disciplines, including anthropology, sociology, economics, women's studies, literary criticism, art history, psychoanalysis and philosophy. Feminist theory aims to understand gender inequality and focuses on gender politics, power relations, and sexuality. While providing a critique of these social and political relations, much of feminist theory also focuses on the promotion of women's rights and interests. Themes explored in feminist theory include discrimination, stereotyping, objectification (especially sexual objectification), oppression, and patriarchy. In the field of literary criticism, Elaine Showalter describes the development of feminist theory as having three phases. The first she calls "feminist critique", in which the feminist reader examines the ideologies behind literary phenomena. The second Showalter calls "gynocriticism", in which the "woman is producer of textual meaning". The last phase she calls "gender theory", in which the "ideological inscription and the literary effects of the sex/gender system are explored". This was paralleled in the 1970s by French feminists, who developed the concept of "écriture féminine" (which translates as 'female or feminine writing'). Helene Cixous argues that writing and philosophy are "" and along with other French feminists such as Luce Irigaray emphasize "writing from the body" as a subversive exercise. The work of Julia Kristeva, a feminist psychoanalyst and philosopher, and Bracha Ettinger, artist and psychoanalyst, has influenced feminist theory in general and feminist literary criticism in particular. However, as the scholar Elizabeth Wright points out, "none of these French feminists align themselves with the feminist movement as it appeared in the Anglophone world". More recent feminist theory, such as that of Lisa Lucile Owens, has concentrated on characterizing feminism as a universal emancipatory movement. Many overlapping feminist movements and ideologies have developed over the years. Some branches of feminism closely track the political leanings of the larger society, such as liberalism and conservatism, or focus on the environment. Liberal feminism seeks individualistic equality of men and women through political and legal reform without altering the structure of society. Catherine Rottenberg has argued that the neoliberal shirt in Liberal feminism has led to that form of feminism being individualized rather than collectivized and becoming detached from social inequality. Due to this she argues that Liberal Feminism cannot offer any sustained analysis of the structures of male dominance, power, or privilege. Radical feminism considers the male-controlled capitalist hierarchy as the defining feature of women's oppression and the total uprooting and reconstruction of society as necessary. Conservative feminism is conservative relative to the society in which it resides. Libertarian feminism conceives of people as self-owners and therefore as entitled to freedom from coercive interference. Separatist feminism does not support heterosexual relationships. Lesbian feminism is thus closely related. Other feminists criticize separatist feminism as sexist. Ecofeminists see men's control of land as responsible for the oppression of women and destruction of the natural environment; ecofeminism has been criticized for focusing too much on a mystical connection between women and nature. Rosemary Hennessy and Chrys Ingraham say that materialist forms of feminism grew out of Western Marxist thought and have inspired a number of different (but overlapping) movements, all of which are involved in a critique of capitalism and are focused on ideology's relationship to women. Marxist feminism argues that capitalism is the root cause of women's oppression, and that discrimination against women in domestic life and employment is an effect of capitalist ideologies. Socialist feminism distinguishes itself from Marxist feminism by arguing that women's liberation can only be achieved by working to end both the economic and cultural sources of women's oppression. Anarcha-feminists believe that class struggle and anarchy against the state require struggling against patriarchy, which comes from involuntary hierarchy. Sara Ahmed argues that Black and Postcolonial feminisms pose a challenge "to some of the organizing premises of Western feminist thought." During much of its history, feminist movements and theoretical developments were led predominantly by middle-class white women from Western Europe and North America. However, women of other races have proposed alternative feminisms. This trend accelerated in the 1960s with the civil rights movement in the United States and the collapse of European colonialism in Africa, the Caribbean, parts of Latin America, and Southeast Asia. Since that time, women in developing nations and former colonies and who are of colour or various ethnicities or living in poverty have proposed additional feminisms. Womanism emerged after early feminist movements were largely white and middle-class. Postcolonial feminists argue that colonial oppression and Western feminism marginalized postcolonial women but did not turn them passive or voiceless. Third-world feminism and Indigenous feminism are closely related to postcolonial feminism. These ideas also correspond with ideas in African feminism, motherism, Stiwanism, negofeminism, femalism, transnational feminism, and Africana womanism. In the late twentieth century various feminists began to argue that gender roles are socially constructed, and that it is impossible to generalize women's experiences across cultures and histories.
https://en.wikipedia.org/wiki?curid=11185
Federal Aviation Administration The Federal Aviation Administration (FAA) is a governmental body of the United States with powers to regulate all aspects of civil aviation in that nation as well as over its surrounding international waters. Its powers include the construction and operation of airports, air traffic management, the certification of personnel and aircraft, and the protection of U.S. assets during the launch or re-entry of commercial space vehicles. Powers over neighboring international waters were delegated to the FAA by authority of the International Civil Aviation Organization. Created in August 1958, the FAA replaced the former Civil Aeronautics Administration (CAA) and later became an agency within the U.S. Department of Transportation. The FAA's roles include: The FAA is divided into four "lines of business" (LOB). Each LOB has a specific role within the FAA. The FAA is headquartered in Washington, D.C. as well as the William J. Hughes Technical Center in Atlantic City, New Jersey, the Mike Monroney Aeronautical Center in Oklahoma City, Oklahoma, and its nine regional offices: The Air Commerce Act of May 20, 1926, is the cornerstone of the federal government's regulation of civil aviation. This landmark legislation was passed at the urging of the aviation industry, whose leaders believed the airplane could not reach its full commercial potential without federal action to improve and maintain safety standards. The Act charged the Secretary of Commerce with fostering air commerce, issuing and enforcing air traffic rules, licensing pilots, certifying aircraft, establishing airways, and operating and maintaining aids to air navigation. The newly created Aeronautics Branch, operating under the Department of Commerce assumed primary responsibility for aviation oversight. In fulfilling its civil aviation responsibilities, the U.S. Department of Commerce initially concentrated on such functions as safety regulations and the certification of pilots and aircraft. It took over the building and operation of the nation's system of lighted airways, a task initiated by the Post Office Department. The Department of Commerce improved aeronautical radio communications—before the founding of the Federal Communications Commission in 1934, which handles most such matters today—and introduced radio beacons as an effective aid to air navigation. The Aeronautics Branch was renamed the Bureau of Air Commerce in 1934 to reflect its enhanced status within the Department. As commercial flying increased, the Bureau encouraged a group of airlines to establish the first three centers for providing air traffic control (ATC) along the airways. In 1936, the Bureau itself took over the centers and began to expand the ATC system. The pioneer air traffic controllers used maps, blackboards, and mental calculations to ensure the safe separation of aircraft traveling along designated routes between cities. In 1938, the Civil Aeronautics Act transferred the federal civil aviation responsibilities from the Commerce Department to a new independent agency, the Civil Aeronautics Authority. The legislation also expanded the government's role by giving the CAA the authority and the power to regulate airline fares and to determine the routes that air carriers would serve. President Franklin D. Roosevelt split the authority into two agencies in 1940: the Civil Aeronautics Administration (CAA) and the Civil Aeronautics Board (CAB). CAA was responsible for ATC, airman and aircraft certification, safety enforcement, and airway development. CAB was entrusted with safety regulation, accident investigation, and economic regulation of the airlines. The CAA was part of the Department of Commerce. The CAB was an independent federal agency. On the eve of America's entry into World War II, CAA began to extend its ATC responsibilities to takeoff and landing operations at airports. This expanded role eventually became permanent after the war. The application of radar to ATC helped controllers in their drive to keep abreast of the postwar boom in commercial air transportation. In 1946, meanwhile, Congress gave CAA the added task of administering the federal-aid airport program, the first peacetime program of financial assistance aimed exclusively at development of the nation's civil airports. The approaching era of jet travel, and a series of midair collisions (most notable was the 1956 Grand Canyon mid-air collision) prompted passage of the Federal Aviation Act of 1958. This legislation gave the CAA's functions to a new independent body, the Federal Aviation Agency. The act transferred air safety regulation from the CAB to the new FAA, and also gave the FAA sole responsibility for a common civil-military system of air navigation and air traffic control. The FAA's first administrator, Elwood R. Quesada, was a former Air Force general and adviser to President Eisenhower. The same year witnessed the birth of the National Aeronautics and Space Administration (NASA), created in the wake of the Soviets launching the first artificial satellite and assuming NACA's role of aeronautical research. In 1967, a new U.S. Department of Transportation (DOT) combined major federal responsibilities for air and surface transport. The Federal Aviation Agency's name changed to the Federal Aviation Administration as it became one of several agencies (e.g., Federal Highway Administration, Federal Railroad Administration, the Coast Guard, and the Saint Lawrence Seaway Commission) within DOT (albeit the largest). The FAA administrator would no longer report directly to the president but would instead report to the Secretary of Transportation. New programs and budget requests would have to be approved by DOT, which would then include these requests in the overall budget and submit it to the president. At the same time, a new National Transportation Safety Board took over the Civil Aeronautics Board's (CAB) role of investigating and determining the causes of transportation accidents and making recommendations to the secretary of transportation. CAB was merged into DOT with its responsibilities limited to the regulation of commercial airline routes and fares. The FAA gradually assumed additional functions. The hijacking epidemic of the 1960s had already brought the agency into the field of civil aviation security. In response to the hijackings on September 11, 2001, this responsibility is now primarily taken by the Department of Homeland Security. The FAA became more involved with the environmental aspects of aviation in 1968 when it received the power to set aircraft noise standards. Legislation in 1970 gave the agency management of a new airport aid program and certain added responsibilities for airport safety. During the 1960s and 1970s, the FAA also started to regulate high altitude (over 500 feet) kite and balloon flying. By the mid-1970s, the agency had achieved a semi-automated air traffic control system using both radar and computer technology. This system required enhancement to keep pace with air traffic growth, however, especially after the Airline Deregulation Act of 1978 phased out the CAB's economic regulation of the airlines. A nationwide strike by the air traffic controllers union in 1981 forced temporary flight restrictions but failed to shut down the airspace system. During the following year, the agency unveiled a new plan for further automating its air traffic control facilities, but progress proved disappointing. In 1994, the FAA shifted to a more step-by-step approach that has provided controllers with advanced equipment. In 1979, Congress authorized the FAA to work with major commercial airports to define noise pollution contours and investigate the feasibility of noise mitigation by residential retrofit programs. Throughout the 1980s, these charters were implemented. In the 1990s, satellite technology received increased emphasis in the FAA's development programs as a means to improvements in communications, navigation, and airspace management. In 1995, the agency assumed responsibility for safety oversight of commercial space transportation, a function begun eleven years before by an office within DOT headquarters. The agency was responsible for the decision to ground flights after the September 11 attacks. In December 2000, an organization within the FAA called the Air Traffic Organization, (ATO) was set up by presidential executive order. This became the air navigation service provider for the airspace of the United States and for the New York (Atlantic) and Oakland (Pacific) oceanic areas. It is a full member of the Civil Air Navigation Services Organisation. The FAA issues a number of awards to holders of its certificates. Among these are demonstrated proficiencies as an aviation mechanic (the AMT Awards), a flight instructor (Gold Seal certification), a 50-year aviator (Wright Brothers Master Pilot Award), a 50-year mechanic (Charles Taylor Master Mechanic Award) or as a proficient pilot. The latter, the FAA "WINGS Program", provides a lifetime series of grouped proficiency activities at three levels (Basic, Advanced, and Master) for pilots who have undergone several hours of ground and flight training since their last WINGS award, or "Phase". For more information on all these programs, visit www.faasafety.gov or inquire at an FAA Flight Standards District Office (FSDO). For more information on the WINGS program, please see FAA Advisory Circular AC 61-91J. The FAA encourages volunteerism in the promotion of aviation safety. The FAA Safety Team, or FAASTeam, works with Volunteers at several levels and promotes safety education and outreach nationwide. For more information, inquire at an FAA Flight Standards District Office (FSDO). On March 18, 2008, the FAA ordered its inspectors to reconfirm that airlines are complying with federal rules after revelations that Southwest Airlines flew dozens of aircraft without certain mandatory inspections. The FAA exercises surprise Red Team drills on national airports annually. On October 31, 2013, after outcry from media outlets, including heavy criticism from Nick Bilton of "The New York Times", the FAA announced it will allow airlines to expand the passengers use of portable electronic devices during all phases of flight, but mobile phone calls would still be prohibited (and use of cellular networks during any point when aircraft doors are closed remains prohibited to-date). Implementation initially varied among airlines. The FAA expected many carriers to show that their planes allow passengers to safely use their devices in airplane mode, gate-to-gate, by the end of 2013. Devices must be held or put in the seat-back pocket during the actual takeoff and landing. Mobile phones must be in airplane mode or with mobile service disabled, with no signal bars displayed, and cannot be used for voice communications due to Federal Communications Commission regulations that prohibit any airborne calls using mobile phones. From a technological standpoint, cellular service would not work in-flight because of the rapid speed of the airborne aircraft: mobile phones cannot switch fast enough between cellular towers at an aircraft's high speed. However, the ban is due to potential radio interference with aircraft avionics. If an air carrier provides Wi-Fi service during flight, passengers may use it. Short-range Bluetooth accessories, like wireless keyboards, can also be used. In July 2014, in the wake of the downing of Malaysia Airlines Flight 17, the FAA suspended flights by U.S. airlines to Ben Gurion Airport during the 2014 Israel–Gaza conflict for 24 hours. The ban was extended for a further 24 hours but was lifted about six hours later. The FAA Reauthorization Act of 2018 gives the FAA one year to establish minimum pitch, width and length for airplane seats, to ensure they are safe for passengers. On March 19, 2019, President Donald Trump announced he would nominate Stephen Dickson, a former executive and pilot at Delta Air Lines, to be the next FAA Administrator. On July 24, 2019, the Senate confirmed Dickson by a vote of 52-40. He was sworn in as Administrator by Transportation Secretary Elaine Chao on August 12, 2019. The FAA has been cited as an example of regulatory capture, "in which the airline industry openly dictates to its regulators its governing rules, arranging for not only beneficial regulation, but placing key people to head these regulators." Retired NASA Office of Inspector General Senior Special Agent Joseph Gutheinz, who used to be a Special Agent with the Office of Inspector General for the Department of Transportation and with FAA Security, is one of the most outspoken critics of FAA. Rather than commend the agency for proposing a $10.2 million fine against Southwest Airlines for its failure to conduct mandatory inspections in 2008, he was quoted as saying the following in an Associated Press story: "Penalties against airlines that violate FAA directives should be stiffer. At $25,000 per violation, Gutheinz said, airlines can justify rolling the dice and taking the chance on getting caught. He also said the FAA is often too quick to bend to pressure from airlines and pilots." Other experts have been critical of the constraints and expectations under which the FAA is expected to operate. The dual role of encouraging aerospace travel and regulating aerospace travel are contradictory. For example, to levy a heavy penalty upon an airline for violating an FAA regulation which would impact their ability to continue operating would not be considered encouraging aerospace travel. On July 22, 2008, in the aftermath of the Southwest Airlines inspection scandal, a bill was unanimously approved in the House to tighten regulations concerning airplane maintenance procedures, including the establishment of a whistleblower office and a two-year "cooling off" period that FAA inspectors or supervisors of inspectors must wait before they can work for those they regulated. The bill also required rotation of principal maintenance inspectors and stipulated that the word "customer" properly applies to the flying public, not those entities regulated by the FAA. The bill died in a Senate committee that year. In September 2009, the FAA administrator issued a directive mandating that the agency use the term "customers" to refer to only the flying public. In 2007, two FAA whistleblowers, inspectors Charalambe "Bobby" Boutris and Douglas E. Peters, alleged that Boutris said he attempted to ground Southwest after finding cracks in the fuselage of an aircraft, but was prevented by supervisors he said were friendly with the airline. This was validated by a report by the Department of Transportation which found FAA managers had allowed Southwest Airlines to fly 46 airplanes in 2006 and 2007 that were overdue for safety inspections, ignoring concerns raised by inspectors. Audits of other airlines resulted in two airlines grounding hundreds of planes, causing thousands of flight cancellations. The House Transportation and Infrastructure Committee held hearings in April 2008. Jim Oberstar, former chairman of the committee, said its investigation uncovered a pattern of regulatory abuse and widespread regulatory lapses, allowing 117 aircraft to be operated commercially although not in compliance with FAA safety rules. Oberstar said there was a "culture of coziness" between senior FAA officials and the airlines and "a systematic breakdown" in the FAA's culture that resulted in "malfeasance, bordering on corruption". In 2008 the FAA proposed to fine Southwest $10.2 million for failing to inspect older planes for cracks, and in 2009 Southwest and the FAA agreed that Southwest would pay a $7.5 million penalty and would adopt new safety procedures, with the fine doubling if Southwest failed to follow through. In 2014, the FAA modified its approach to air traffic control hiring. It launched more "off the street bids", allowing anyone with either a four-year degree or five years of full-time work experience to apply, rather than the closed college program or VRA bids, something that had last been done in 2008. Thousands have been picked up, including veterans, CTI grads, and people who are true "off the street" hires. The move was made to open the job up to more people who might make good controllers but did not go to a college that offered a CTI program. Before the change, candidates who had completed coursework at participating colleges and universities could be "fast-tracked" for consideration. However, the CTI program had no guarantee of a job offer, nor was the goal of the program to teach people to work actual traffic. The goal of the program was to prepare people for the FAA Academy in Oklahoma City, OK. Having a CTI certificate allowed a prospective controller to skip the Air Traffic Basics part of the academy, about a 30- to 45-day course, and go right into Initial Qualification Training (IQT). All prospective controllers, CTI or not, have had to pass the FAA Academy in order to be hired as a controller. Failure at the academy means FAA employment is terminated. In January 2015 they launched another pipeline, a "prior experience" bid, where anyone with an FAA Control Tower Operator certificate (CTO) and 52 weeks of experience could apply. This was a revolving bid, every month the applicants on this bid were sorted out, and eligible applicants were hired and sent directly to facilities, bypassing the FAA academy entirely. In the process of promoting diversity, the FAA revised its hiring process. The FAA later issued a report that the "bio-data" was not a reliable test for future performance. However, the "Bio-Q" was not the determinating factor for hiring, it was merely a screening tool to determine who would take a revised Air Traffic Standardized Aptitude Test (ATSAT). Due to cost and time, it was not practical to give all 30,000 some applicants the revised ATSAT, which has since been validated. In 2015 Fox News levied unsubstantiated criticism that the FAA discriminated against qualified candidates. In December 2015, a reverse discrimination lawsuit was filed against the FAA seeking class-action status for the thousands of men and women who spent up to $40,000 getting trained under FAA rules before they were abruptly changed. The prospects of the lawsuit are unknown, as the FAA is a self-governing entity and therefore can alter and experiment with its hiring practices, and there was never any guarantee of a job in the CTI program. A May 2017 letter from staff of the U.S. House of Representatives Committee on Transportation and Infrastructure to members of the same committee sent before a meeting to discuss air traffic control privatization noted a 35-year legacy of failed air traffic control modernization management, including NextGen. The letter said the FAA initially described NextGen as fundamentally transforming how air traffic would be managed. In 2015, however, the National Research Council noted that NextGen, as currently executed, was not broadly transformational and that it is a set of programs to implement a suite of incremental changes to the National Airspace System (NAS). More precise PBN can reduce fuel burn, emissions, and noise exposure for a majority of communities, but the concentration of flight tracks also can increase noise exposure for people who live directly under those flight paths. A feature of the NextGen program is GPS-based waypoints, which result in consolidated flight paths for planes. The result of this change is that many localities experience huge increases in air traffic over previously quiet areas. Complaints have risen with the added traffic and multiple municipalities have filed suit. As a result of the March 10, 2019 Ethiopian Airlines Flight 302 crash and the Lion Air Flight 610 crash five months earlier, most airlines and countries began grounding the Boeing 737 MAX 8 (and in many cases all MAX variants) due to safety concerns, but the FAA declined to temporarily ground Boeing 737 Max 8 aircraft operating in the U.S. On March 12, the FAA said that its ongoing review showed "no systemic performance issues and provides no basis to order grounding the aircraft." Some U.S. Senators called for the FAA to ground the aircraft until an investigation into the cause of the Ethiopian Airlines crash was complete. U.S. Transportation Secretary Elaine Chao said that "If the FAA identifies an issue that affects safety, the department will take immediate and appropriate action." On March 13, President Donald Trump announced that all Boeing 737 MAX airplanes within U.S. territory would be grounded, and the FAA issued the official order, citing new evidence of similarity between the two accidents. Three major U.S. airlines--Southwest, United, and American Airlines—were affected by this decision. A Designated Engineering Representative (DER) is an engineer who is appointed under 14 CFR section 183.29 to act on behalf of a company or as an independent consultant (IC). A DAR is an individual appointed in accordance with 14 CFR 183.33 who may perform examination, inspection, and testing services necessary to the issuance of certificates. There are two types of DARs: manufacturing, and maintenance. Specialized Experience – Amateur-Built and Light-Sport Aircraft DARs Both Manufacturing DARs and Maintenance DARs may be authorized to perform airworthiness certification of light-sport aircraft. DAR qualification criteria and selection procedures for amateur-built and light-sport aircraft airworthiness functions are provided in Order 8100.8. A Continued Airworthiness Notification to the International Community (commonly abbreviated as CANIC) is a notification from the United States Federal Aviation Administration (FAA) to civil airworthiness authorities of foreign countries of pending significant safety actions. The FAA Airworthiness Directives Manual, states the following: 8. Continued Airworthiness Notification to the International Community (CANIC). The FAA issued a CANIC to state the continued airworthiness of the Boeing 737 MAX, following the crash of Ethiopian Airlines Flight 302. U.S. law requires that the FAA's budget and mandate be reauthorized on a regular basis. On July 18, 2016, President Obama signed a second short-term extension of the FAA authorization, replacing a previous extension that was due to expire that day. The 2016 extension (set to expire itself in September 2017) left out a provision pushed by Republican House leadership, including House Transportation and Infrastructure (T&I) Committee Chairman Bill Shuster (R-PA). The provision would have moved authority over air traffic control from the FAA to a non-profit corporation, as many other nations, such as Canada, Germany and the United Kingdom, have done. Shuster's bill, the Aviation Innovation, Reform, and Reauthorization (AIRR) Act, expired in the House at the end of the 114th Congress. The House T&I Committee began the new reauthorization process for the FAA in February 2017. It is expected that the committee will again urge Congress to consider and adopt air traffic control reform as part of the reauthorization package. Shuster has additional support from President Trump, who, in a meeting with aviation industry executives in early 2017 said the U.S. air control system is "...totally out of whack."
https://en.wikipedia.org/wiki?curid=11186
French Revolution The French Revolution ( ) was a period of far-reaching social and political upheaval in France and its colonies beginning in 1789 and ending in 1799. The Revolution overthrew the monarchy; established a republic; catalyzed violent periods of political turmoil; and finally culminated in a dictatorship under Napoleon, who brought many of the revolution's principles to areas he conquered in Western Europe and beyond. Inspired by liberal and radical ideas such as equality before the law, the Revolution made a profound impression on the course of modern history, influencing the decline of absolute monarchies while replacing them with republics and liberal democracies. Historians widely regard the Revolution as one of the most important events in human history. The causes of the French Revolution are complex and are still debated among historians. The American Revolution helped set the stage for the events of the French Revolution, having shown France that a rebellion based on Enlightenment principles, including natural rights and equality for all citizens, against an authoritarian regime could succeed. German diplomat Friedrich von Gentz (1764-1832) observed in a treatise published in France in 1800: The National Assembly of France even used the American Declaration of Independence as a template when drafting the Declaration of the Rights of Man and of the Citizen in 1789. The Americans' victory over the British may have been the "single greatest impact" on the start of the French Revolution. Following the Seven Years' War and the American Revolutionary War, the French government was deeply in debt. It attempted to restore its financial status through unpopular taxation schemes, which were heavily regressive. Leading up to the Revolution, years of bad harvests worsened by deregulation of the grain industry and fifty consecutive days of below-freezing temperatures in the winter of 1788/1789 inflamed popular resentment of the privileges enjoyed by the aristocracy and the Catholic clergy of the established church. Demands for change were formulated in terms of Enlightenment ideals on democracy, and contributed to the convocation of the Estates General in May 1789. During the first year of the Revolution, members of the Third Estate (commoners) took control; the Bastille was attacked in July; the Declaration of the Rights of Man and of the Citizen was passed in August; and the Women's March on Versailles forced the royal court back to Paris in October. A central event of the first stage, in August 1789, was the abolition of feudalism and the old rules and privileges left over from the "Ancien Régime". The next few years featured political struggles between various liberal assemblies and supporters of the monarchy intent on thwarting major reforms, promoted by the Jacobins, led to the Insurrection of 10 August 1792 and the arrest of Louis XVI and the royal family. The Republic was proclaimed in 22 September after the first French elections and the victory at Valmy. Its goal was to unify France and to introduce the same taxes and democratic elections for more citizens. It opposed prerogatives. In a momentous event that led to international condemnation and an internal struggle in the Convention between the Girondins and Montagnards, Louis XVI was executed in January 1793. External threats closely shaped the course of the Revolution. The French Revolutionary Wars unleashed a wave of global conflicts that extended from the Caribbean to the Middle East. Internally, popular agitation by the Sans-culottes radicalised the Revolution significantly, followed by the Insurrection at the end of May, and the rise of Maximilien Robespierre. A "levée en masse", an army of volunteers to beat the external and internal enemy, culminated in a federalist revolt in the South and the West. The dictatorship imposed by the Committee of Public Safety established price controls on food and soap, introduced a secular Republican calendar, de-established the Catholic church (dechristianised society). During what was called the Reign of Terror, counter-revolutionaries were expelled, arrested or executed; and the borders of the new republic were secured from its enemies. After the Fall of Robespierre and Thermidorian Reaction, an executive council known as the Directory assumed control of the French state in 1795. They suspended elections, repudiated debts (creating financial instability in the process), persecuted the Catholic clergy, and made significant military conquests on the Italian Peninsula. Dogged by charges of corruption, the Directory collapsed in a coup led by Napoleon Bonaparte in 1799. Napoleon, who ended and became the hero of the Revolution, established the Consulate and later the First French Empire. The modern era has unfolded in the shadow of the French Revolution. Almost all future revolutionary movements looked back to the Revolution as their predecessor. Its central phrases and cultural symbols, such as "La Marseillaise" and "Liberté, fraternité, égalité, ou la mort", became the clarion call for other major upheavals in modern history, including the Russian Revolution over a century later. Some of its central documents, such as the Declaration of the Rights of Man and of the Citizen, continued to inspire movements for abolitionism and universal suffrage in the next two centuries. The Revolution resulted in the suppression of the feudal system, emancipation of the individual, a greater division of landed property, abolition of the privileges of noble birth, and nominal establishment of equality among men. The Revolution also witnessed the birth of total war by organising the resources of France and the lives of its citizens towards the objective of national defence. Globally, the Revolution became the focal point for the development of most modern political ideologies, leading to the spread of liberalism, radicalism, nationalism, and secularism, among many others, accelerating the rise of republics and democracies. The values and institutions of the Revolution dominate French politics to this day. Historians have pointed to many events and factors within the "Ancien Régime" that led to the Revolution. Rising social and economic inequality; new political ideas emerging from the Enlightenment; economic mismanagement; environmental factors leading to agricultural failure, unmanageable national debt; and political mismanagement on the part of King Louis XVI, have all been cited as laying the groundwork for the Revolution. Over the course of the 18th century, there emerged what the philosopher Jürgen Habermas called the idea of the "public sphere" in France and elsewhere in Europe. Habermas argued that the dominant cultural model in 17th-century France was a "representational" culture, which was based on a one-sided need to "represent" power with one side active and the other passive. A perfect example would be the Palace of Versailles, which was meant to overwhelm the senses of the visitor and convince one of the greatness of the French state and Louis XIV. Starting in the early 18th century the "public sphere" emerged which was "critical" in that both sides were active. Examples of the public sphere included newspapers, journals, masonic lodges, coffee houses and reading clubs where people either in person or virtually via the printed word debated and discussed issues. In France, the emergence of the public sphere outside of the control of the state led to the shift from Versailles to Paris as the cultural capital of France. Likewise, while in the 17th century the court had decided what was culturally good and what was not, in the 18th century the opinion of the court mattered less and consumers became the arbiters of cultural taste. In the 1750s, during the ""Querelle des Bouffons"" over the question of the quality of Italian vs. French music, the partisans of both sides appealed to the French public "because it alone has the right to decide whether a work will be preserved for posterity or will be used by grocers as wrapping-paper." In 1782, Louis-Sébastien Mercier wrote: "The word court no longer inspires awe amongst us as in the time of Louis XIV. Reigning opinions are no longer received from the court; it no longer decides on reputations of any sort ... The court's judgments are countermanded; one says openly that it understands nothing; it has no ideas on the subject and could have none." Inevitably, the belief that public opinion had the right to decide cultural questions instead of deferring to the court transformed itself into the demand that the public also have a say on political questions as well. The economy in the Ancien Régime during the years preceding the Revolution suffered from instability. The sequence of events leading to the Revolution included the national government's fiscal troubles caused by an unjust, inefficient and deeply hated tax system—the "ferme générale"—and by expenditure on numerous large wars. The attempt to challenge British naval and commercial power in the Seven Years' War was a costly disaster, with the loss of France's colonial possessions in continental North America and the destruction of the French Navy. French forces were rebuilt, and feeling bitter about having lost many of France's overseas colonies to the British Empire during the Seven Years' War, Louis XVI was eager to give the American rebels financial and military support. After the British surrender at the Battle of Saratoga, the French sent 10,000 troops and millions of dollars to the rebels. Despite succeeding in gaining independence for the Thirteen Colonies, France was severely indebted by the American Revolutionary War. France's inefficient and antiquated financial system could not finance this debt. Faced with a financial crisis, the king called an Estates General, recommended by the Assembly of Notables in 1787 for the first time in over a century. France was experiencing such a severe economic depression that there wasn't enough food to go around. Poor harvests lasting several years and an inadequate transportation system both contributed to making food more expensive. As with most monarchies, the upper class was always ensured a stable living, so while the rich remained very wealthy, the majority of the French population was starving. Many were so destitute that they couldn't even feed their families and resorted to theft or prostitution to stay alive. Meanwhile, the royal court at Versailles was isolated from and indifferent to the escalating crisis. While in theory King Louis XVI was an absolute monarch, in practice he was often indecisive and known to back down when faced with strong opposition. While he did reduce government expenditures, opponents in the "parlements" successfully thwarted his attempts at enacting much needed reforms. The Enlightenment had produced many writers, pamphleteers and publishers who could inform or inflame public opinion. The opposition used this resource to mobilise public opinion against the monarchy, which in turn tried to repress the underground literature. Many other factors involved resentments and aspirations given focus by the rise of Enlightenment ideals. These included resentment of royal absolutism; resentment by peasants, labourers and the bourgeoisie towards the traditional seigneurial privileges possessed by the nobility; resentment of the Catholic Church's influence over public policy and institutions; aspirations for freedom of religion; resentment of aristocratic bishops by the poorer rural clergy; aspirations for social, political and economic equality, and (especially as the Revolution progressed) republicanism; hatred of Queen Marie-Antoinette, who was falsely accused of being a spendthrift and an Austrian spy; and anger towards the King for dismissing ministers, including finance minister Jacques Necker, who were popularly seen as representatives of the people. In 1774 Louis XVI ascended to the throne in the middle of a financial crisis in which the state was faced with a budget deficit and was nearing bankruptcy. This was due in part to France's costly involvements in the Seven Years' War and later the American Revolutionary War. In May 1776, finance minister Turgot was dismissed, after failing to enact reforms. The next year, Jacques Necker, a foreigner, was appointed Comptroller-General of Finance. He could not be made an official minister because he was a Protestant. Necker realised that the country's extremely regressive tax system subjected the lower classes to a heavy burden, while numerous exemptions existed for the nobility and clergy. He argued that the country could not be taxed higher; that tax exemptions for the nobility and clergy must be reduced; and proposed that borrowing more money would solve the country's fiscal shortages. Necker published a report to support this claim that underestimated the deficit by roughly 36 million livres, and proposed restricting the power of the "parlements". This was not received well by the King's ministers, and Necker, hoping to bolster his position, argued to be made a minister. The King refused, Necker was dismissed, and Charles Alexandre de Calonne was appointed to the Comptrollership. Calonne initially spent liberally, but he quickly realised the critical financial situation and proposed a new tax code. The proposal included a consistent land tax, which would include taxation of the nobility and clergy. Faced with opposition from the parlements, Calonne organised the summoning of the Assembly of Notables. But the Assembly failed to endorse Calonne's proposals and instead weakened his position through its criticism. The Notables concluded that this kind of change must be approved by the people in the form of the Estates-General. His failures resulted in Calonne's dismissal, and in May 1787 he was replaced with Brienne. Brienne presented a package to the Notables that was not dissimilar to Calonne's, but without the sale of church land. This was again rejected and the Notables dissolved. Brienne attempted to appeal this judgement by sending his reforms to the Parlement in early June; the Parlement endorsed the principal of the reforms, but like the Notables, concluded that the changes could only be made by the Estates-General. They announced their decision in July, following the Day of the Tiles, a formal resistance by the people to Brienne and the King's efforts to push reforms. Louis attempted to compel the Parlement to approve the reforms via a "lit de justice", which the Parlement declared invalid. A Royal Session was called for 19 November 1787, between the Parlement and the King. The Parlement resisted royal mandate even in King's presence. In response, the King had no choice but to announce the calling of the Estates-General for May 1789, the first time the body had been summoned since 1614. This was a signal that the Bourbon monarchy was in a weakened state and subject to the demands of its people. The Estates-General was organised into three estates: the clergy, the nobility, and the rest of France. It had last met in 1614. Elections were held in the spring of 1789; suffrage requirements for the Third Estate were for French-born or naturalised males, aged 25 years or more, who resided where the vote was to take place and who paid taxes. Strong turnout produced 1,201 delegates, including 303 clergy, 291 nobles and 610 members of the Third Estate. The First Estate represented 100,000 Catholic clergy; the Church owned about 10% of the land and collected its own taxes (the tithe) on peasants. The lands were controlled by bishops and abbots of monasteries, but two-thirds of the 303 delegates from the First Estate were ordinary parish priests; only 51 were bishops. The Second Estate represented the nobility, about 400,000 men and women who owned about 25% of the land and collected seigneurial dues and rents from their peasant tenants. About a third of these deputies were nobles, mostly with minor holdings. The Third Estate representation was doubled to 610 men, representing 95% of the population. Half were well educated lawyers or local officials. Nearly a third were in trades or industry; 51 were wealthy land owners. To assist delegates, "Books of grievances" ("cahiers de doléances") were compiled to list problems. The books articulated ideas which would have seemed radical only months before; however, most supported the monarchical system in general. Many assumed the Estates-General would approve future taxes, and Enlightenment ideals were relatively rare. Pamphlets by liberal nobles and clergy became widespread after the lifting of press censorship. The Abbé Sieyès, a theorist and Catholic clergyman, argued the paramount importance of the Third Estate in the pamphlet "Qu'est-ce que le tiers état?" ("What is the Third Estate?") published in January 1789. He asserted: "What is the Third Estate? Everything. What has it been until now in the political order? Nothing. What does it want to be? Something." The Estates-General convened in the Grands Salles des Menus-Plaisirs in Versailles on 5 May 1789 and opened with a three-hour speech by Necker. The Third Estate demanded that the credentials of deputies should be verified by all deputies, rather than each estate verifying the credentials of its own members, but negotiations with the other estates failed to achieve this. The commoners appealed to the clergy, who asked for more time. Necker then stated that each estate should verify its own members' credentials and that the king should act as arbitrator. On 10 June 1789 Abbé Sieyès moved that the Third Estate, now meeting as the "Communes" (English: "Commons") proceed with verifying its own powers and invite the other two estates to take part, but not to wait for them. They proceeded to do so two days later, completing the process on 17 June. Then they voted a measure far more radical, declaring themselves the National Assembly, an assembly not of the Estates but of "the People". They invited the other orders to join them, but made it clear they intended to conduct the nation's affairs with or without them. In an attempt to keep control of the process and prevent the Assembly from convening, Louis XVI ordered the closure of the Salle des États where the Assembly met, making an excuse that the carpenters needed to prepare the hall for a royal speech in two days. Weather did not allow an outdoor meeting, and fearing an attack ordered by Louis XVI, they met in a tennis court just outside Versailles, where they proceeded to swear the Tennis Court Oath (20 June 1789) under which they agreed not to separate until they had given France a constitution. A majority of the representatives of the clergy soon joined them, as did 47 members of the nobility. By 27 June, the royal party had overtly given in, although the military began to arrive in large numbers around Paris and Versailles. Messages of support for the Assembly poured in from Paris and other French cities. By this time, Necker had earned the enmity of many members of the French court for his overt manipulation of public opinion. Marie Antoinette, the King's younger brother the Comte d'Artois, and other conservative members of the King's privy council urged him to dismiss Necker as financial advisor. On 11 July 1789, after Necker published an inaccurate account of the government's debts and made it available to the public, the King fired him, and completely restructured the finance ministry at the same time. Many Parisians presumed Louis' actions to be aimed against the Assembly and began open rebellion when they heard the news the next day. They were also afraid that arriving soldiers – mostly foreign mercenaries – had been summoned to shut down the National Constituent Assembly. The Assembly, meeting at Versailles, went into nonstop session to prevent another eviction from their meeting place. Paris was soon consumed by riots, chaos, and widespread looting. The mobs soon had the support of some of the French Guard, who were armed and trained soldiers. On 14 July, the insurgents set their eyes on the large weapons and ammunition cache inside the Bastille fortress, which was also perceived to be a symbol of royal power. After several hours of combat, the prison fell that afternoon. Despite ordering a ceasefire, which prevented a mutual massacre, Governor Marquis Bernard-René de Launay was beaten, stabbed and decapitated; his head was placed on a pike and paraded about the city. Although the fortress had held only seven prisoners (four forgers, two noblemen kept for immoral behaviour, and a murder suspect) the Bastille served as a potent symbol of everything hated under the "Ancien Régime". Returning to the Hôtel de Ville (city hall), the mob accused the "prévôt des marchands" (roughly, mayor) Jacques de Flesselles of treachery and butchered him. The King, alarmed by the violence, backed down, at least for the time being. The Marquis de Lafayette took up command of the National Guard at Paris. Jean-Sylvain Bailly, president of the Assembly at the time of the Tennis Court Oath, became the city's mayor under a new governmental structure known as the "commune". The King visited Paris, where, on 17 July he accepted a tricolore cockade, to cries of "Vive la Nation" ("Long live the Nation") and "Vive le Roi" ("Long live the King"). Necker was recalled to power, but his triumph was short-lived. An astute financier but a less astute politician, Necker overplayed his hand by demanding and obtaining a general amnesty, losing much of the people's favour. As civil authority rapidly deteriorated, with random acts of violence and theft breaking out across the country, members of the nobility, fearing for their safety, fled to neighbouring countries; many of these "émigrés", as they were called, funded counter-revolutionary causes within France and urged foreign monarchs to offer military support to a counter-revolution. By late July, the spirit of popular sovereignty had spread throughout France. In rural areas, many commoners began to form militias and arm themselves against a foreign invasion: some attacked the châteaux of the nobility as part of a general agrarian insurrection known as ""la Grande Peur"" ("the Great Fear"). In addition, wild rumours and paranoia caused widespread unrest and civil disturbances that contributed to the collapse of law and order. On 4 and 11 August 1789 the National Constituent Assembly abolished privileges and feudalism (numerous peasant revolts had almost brought feudalism to an end) in the August Decrees, sweeping away personal serfdom, exclusive hunting rights and other seigneurial rights of the Second Estate (nobility). Also the tithe (a 10% tax for the Church, gathered by the First Estate (clergy)), which had been the main source of income for many clergymen, was abolished. During the course of a few hours nobles, clergy, towns, provinces, companies and cities lost their special privileges. Historian Georges Lefebvre summarises the night's work: Originally the peasants were supposed to pay for the release of seigneurial dues; these dues affected more than a fourth of the farmland in France and provided most of the income of the large landowners. The majority refused to pay and in 1793 the obligation was cancelled. Thus the peasants got their land free, and also no longer paid the tithe to the church. Furet emphasises that the decisions of August 1789 survived and became an integral part of The old judicial system, based on the 13 regional parlements, was suspended in November 1789, and officially abolished in September 1790. The main institutional pillars of the old regime had vanished. On 26 August 1789 the Assembly published the Declaration of the Rights of Man and of the Citizen, which comprised a statement of principles rather than a constitution with legal effect. The National Constituent Assembly functioned not only as a legislature, but also as a body to draft a new constitution. Fuelled by rumours of a reception for the King's bodyguards on 1 October 1789, at which the national cockade had been trampled upon, on 5 October 1789, crowds of women began to assemble at Parisian markets. The women first marched to the Hôtel de Ville, demanding that city officials address their concerns. The women were responding to the harsh economic situations they faced, especially bread shortages. They also demanded an end to royal efforts to block the National Assembly, and for the King and his administration to move to Paris as a sign of good faith in addressing the widespread poverty. Getting unsatisfactory responses from city officials, as many as 7,000 women joined the march to Versailles, bringing with them cannons and a variety of smaller weapons. Twenty thousand National Guardsmen under the command of Lafayette responded to keep order, and members of the mob stormed the palace, killing several guards. Lafayette ultimately persuaded the king to accede to the demand of the crowd that the monarchy relocate to Paris. On 6 October 1789, the King and the royal family moved from Versailles to Paris under the "protection" of the National Guards, thus legitimising the National Assembly. The Revolution caused a massive shift of power from the Roman Catholic Church to the state. Under the "Ancien Régime", the Church had been the largest single landowner in the country, owning about 10% of the land in the kingdom. The Church was exempt from paying taxes to the government, while it levied a tithe – a 10% tax on income, often collected in the form of crops – on the general population, only a fraction of which it then redistributed to the poor. Resentment towards the Church weakened its power during the opening of the Estates General in May 1789. The Church composed the First Estate with 130,000 members of the clergy. When the National Assembly was later created in June 1789 by the Third Estate, the clergy voted to join them, which perpetuated the destruction of the Estates General as a governing body. The National Assembly began to enact social and economic reform. Legislation sanctioned on 4 August 1789 abolished the Church's authority to impose the tithe. In an attempt to address the financial crisis, the Assembly declared, on 2 November 1789, that the property of the Church was "at the disposal of the nation". They used this property to back a new currency, the assignats. Thus, the nation had now also taken on the responsibility of the Church, which included paying the clergy and caring for the poor, the sick and the orphaned. In December, the Assembly began to sell the lands to the highest bidder to raise revenue, effectively decreasing the value of the assignats by 25% in two years. In autumn 1789, legislation abolished monastic vows and on 13 February 1790 all religious orders were dissolved. Monks and nuns were encouraged to return to private life and a small percentage did eventually marry. The Civil Constitution of the Clergy, passed on 12 July 1790, turned the remaining clergy into employees of the state. This established an election system for parish priests and bishops and set a pay rate for the clergy. Many Catholics objected to the election system because it effectively denied the authority of the Pope in Rome over the French Church. In October a group of 30 bishops wrote a declaration saying they could not accept that law, and this protest also fueled civilian opposition against that law. Eventually, in November 1790, the National Assembly began to require an oath of loyalty to the Civil Constitution from all the members of the clergy. This led to a schism between those clergy who swore the required oath and accepted the new arrangement and those who remained loyal to the Pope. Priests swearing the oath were indicated as 'constitutional', those not taking the oath as 'non-juring' or 'refractory' clergy. Overall, 24% of the clergy nationwide took the oath. This decree stiffened the resistance against the state's interference with the church, especially in the west of France like in Normandy, Brittany and the Vendée, where only a few priests took the oath and the civilian population turned against the revolution. Widespread refusal led to legislation against the clergy, "forcing them into exile, deporting them forcibly, or executing them as traitors". Pope Pius VI never accepted the Civil Constitution of the Clergy, further isolating the Church in France. Necker, Mounier, Lally-Tollendal and others argued unsuccessfully for a senate, with members appointed by the crown on the nomination of the people. The bulk of the nobles argued for an aristocratic upper house elected by the nobles. The popular party carried the day: France would have a single, unicameral assembly. The King retained only a "suspensive veto"; he could delay the implementation of a law, but not block it absolutely. The Assembly eventually replaced the historic provinces with 83 "départements," uniformly administered and roughly equal in area and population. Amid the Assembly's preoccupation with constitutional affairs, the financial crisis had continued largely unaddressed, and the deficit had only increased. Honoré Mirabeau now led the move to address this matter, and the Assembly gave Necker complete financial dictatorship. A new Republican Calendar was established in October 1793, with 10-day weeks that made it very difficult for Catholics to remember Sundays and saints' days. Workers complained it reduced the number of first-day-of-the-week holidays from 52 to 37. During the Reign of Terror, extreme efforts of dechristianisation ensued, including the imprisonment and massacre of priests and destruction of churches and religious images throughout France. An effort was made to replace the Catholic Church altogether, with civic festivals replacing religious ones. The establishment of the Cult of Reason was the final step of radical dechristianisation. These events led to a widespread disillusionment with the Revolution and to counter-rebellions across France. Locals often resisted de-Christianisation by attacking revolutionary agents and hiding members of the clergy who were being hunted. Eventually, Robespierre and the Committee of Public Safety were forced to denounce the campaign, replacing the Cult of Reason with the deist but still non-Christian Cult of the Supreme Being. The Concordat of 1801 between Napoleon and the Church ended the de-Christianisation period and established the rules for a relationship between the Catholic Church and the French State that lasted until it was abrogated by the Third Republic via the separation of church and state on 11 December 1905. The persecution of the Church led to a counter-revolution known as the Revolt in the Vendée. Historians Lynn Hunt and Jack Censer argue that some French Protestants, the Huguenots, wanted an anti-Catholic regime, and that Enlightenment thinkers such as Voltaire helped fuel this resentment. The Enlightenment philosopher Jean-Jacques Rousseau, for example, had told French citizens that it was "manifestly contrary to the law of nature... that a handful of people should gorge themselves with superfluities while the hungry multitude goes in want of necessities." Historian John McManners writes, ""In eighteenth-century France throne and altar were commonly spoken of as in close alliance; their simultaneous collapse ... would one day provide the final proof of their interdependence."" Factions within the Assembly began to clarify. The aristocrat Jacques Antoine Marie de Cazalès and the abbé Jean-Sifrein Maury led what would become known as the right wing, the opposition to revolution (this party sat on the right-hand side of the Assembly). The "Royalist democrats" or "monarchiens", allied with Necker, inclined towards organising France along lines similar to the British constitutional model; they included Jean Joseph Mounier, the Comte de Lally-Tollendal, the comte de Clermont-Tonnerre, and Pierre Victor Malouet, comte de Virieu. The "National Party", representing the centre or centre-left of the assembly, included Honoré Mirabeau, Lafayette, and Bailly; while Adrien Duport, Barnave and Alexandre Lameth represented somewhat more extreme views. Almost alone in his radicalism on the left was the Arras lawyer Maximilien Robespierre, supported by Pétion de Villeneuve and Buzot. Abbé Sieyès led in proposing legislation in this period and successfully forged consensus for some time between the political centre and the left. In Paris, various committees, the mayor, the assembly of representatives, and the individual districts each claimed authority independent of the others. The increasingly middle-class National Guard under Lafayette also slowly emerged as a power in its own right, as did other self-generated assemblies. The Assembly abolished the symbolic paraphernalia of the "Ancien Régime" – armorial bearings, liveries, etc. – which further alienated the more conservative nobles, and added to the ranks of the "émigrés". On 14 July 1790, and for several days following, crowds in the Champ de Mars celebrated the anniversary of the fall of the Bastille with the "Fête de la Fédération"; Talleyrand performed a mass; participants swore an oath of "fidelity to the nation, the law, and the king"; the King and the royal family actively participated. The electors had originally chosen the members of the Estates-General to serve for a single year. However, by the terms of the Tennis Court Oath, the "communes" had bound themselves to meet continuously until France had a constitution. Right-wing elements now argued for a new election, but Mirabeau prevailed, asserting that the status of the assembly had fundamentally changed, and that no new election should take place before completing the constitution. In late 1790 the French army was in considerable disarray. The military officer corps was largely composed of noblemen, who found it increasingly difficult to maintain order within the ranks. In some cases, soldiers (drawn from the lower classes) had turned against their aristocratic commanders and attacked them. At Nancy, General Bouillé successfully put down one such rebellion, only to be accused of being anti-revolutionary for doing so. This and other such incidents spurred a mass desertion as more and more officers defected to other countries, leaving a dearth of experienced leadership within the army. This period also saw the rise of the political "clubs" in French politics. Foremost among these was the Jacobin Club; 152 members had affiliated with the Jacobins by 10 August 1790. The Jacobin Society began as a broad, general organisation for political debate, but as it grew in members, various factions developed with widely differing views. Several of these factions broke off to form their own clubs, such as the Club of '89. Meanwhile, the Assembly continued to work on developing a constitution. A new judicial organisation made all magistracies temporary and independent of the throne. The legislators abolished hereditary offices, except for the monarchy itself. Jury trials started for criminal cases. The King would have the unique power to propose war, with the legislature then deciding whether to declare war. The Assembly abolished all internal trade barriers and suppressed guilds, masterships, and workers' organisations: any individual gained the right to practise a trade through the purchase of a license; strikes became illegal. Louis XVI was increasingly dismayed by the direction of the revolution. His brother, the Comte d'Artois and his queen, Marie Antoinette, urged a stronger stance against the revolution and support for the émigrés, while he was resistant to any course that would see him openly side with foreign powers against the Assembly. Eventually, fearing for his own safety and that of his family, he decided to flee Paris to the Austrian border, having been assured of the loyalty of the border garrisons. Louis cast his lot with General Bouillé, who condemned both the emigration and the Assembly, and promised him refuge and support in his camp at Montmédy. On the night of 20 June 1791 the royal family fled the Tuileries Palace dressed as servants, while their servants dressed as nobles. However, late the next day, the King was recognised and arrested at Varennes and returned to Paris. The Assembly provisionally suspended the King. He and Queen Marie Antoinette remained held under guard. The King's flight had a profound impact on public opinion, turning popular sentiment further against the clergy and nobility, and built momentum for the institution of a constitutional monarchy. As most of the Assembly still favoured a constitutional monarchy rather than a republic, the various groups reached a compromise which left Louis XVI as little more than a figurehead: he was forced to swear an oath to the constitution, and a decree declared that retracting the oath, heading an army for the purpose of making war upon the nation, or permitting anyone to do so in his name would amount to abdication. However, Jacques Pierre Brissot drafted a petition, insisting that in the eyes of the nation Louis XVI was deposed since his flight. An immense crowd gathered in the Champ de Mars to sign the petition. Georges Danton and Camille Desmoulins gave fiery speeches. The Assembly called for the municipal authorities to "preserve public order". The National Guard under Lafayette's command confronted the crowd. The soldiers responded to a barrage of stones by firing into the crowd, killing between 13 and 50 people. The incident cost Lafayette and his National Guard much public support. In the wake of the massacre the authorities closed many of the patriotic clubs, as well as radical newspapers such as Jean-Paul Marat's "L'Ami du Peuple". Danton fled to England; Desmoulins and Marat went into hiding. Meanwhile, in August 1791, a new threat arose from abroad: the King's brother-in-law Holy Roman Emperor Leopold II, King Frederick William II of Prussia, and the King's brother Charles-Philippe, comte d'Artois, issued the Declaration of Pillnitz, declaring their intention to bring the French king in the position "to consolidate the basis of a monarchical government" and that they were preparing their own troops for action, hinting at an invasion of France on the King's behalf. Although Leopold himself sought to avoid war and made the declaration to satisfy the Comte d'Artois and the other émigrés, the reaction within France was ferocious. The French people expressed no respect for the dictates of foreign monarchs, and the threat of force merely hastened their militarisation. Even before the Flight to Varennes, the Assembly members had determined to debar themselves from the legislature that would succeed them, the Legislative Assembly. They now gathered the various constitutional laws they had passed into a single constitution, and submitted it to the recently restored Louis XVI, who accepted it, writing "I engage to maintain it at home, to defend it from all attacks from abroad, and to cause its execution by all the means it places at my disposal". The King addressed the Assembly and received enthusiastic applause from members and spectators. With this capstone, the National Constituent Assembly adjourned in a final session on 30 September 1791. The Legislative Assembly first met on 1 October 1791, elected by those 4 million men – out of a population of 6 million men over the age of 25 – who paid a certain minimum amount of taxes. Under the Constitution of 1791, France would function as a constitutional monarchy. The King had to share power with the elected Legislative Assembly, but he retained his royal veto and the ability to select ministers. Early on, the King vetoed legislation that threatened the "émigrés" with death and that decreed that every non-juring clergyman must take within eight days the civic oath mandated by the Civil Constitution of the Clergy. Over the course of a year, such disagreements would lead to a constitutional crisis. Late in 1791, a group of Assembly members who demanded war against Austria and Prussia was, after a remark by the politician Maximilien Robespierre, henceforth known as the 'Girondins', although not all of them really came from the southern province of Gironde. A group around Robespierre – later known as 'Montagnards' or 'Jacobins' – argued against this war; this dispute between the two groups would harden into a bitter enmity over the next year and a half. In response to the threat of war of August 1791 from Austria and Prussia, leaders of the Assembly saw such a war as a means to strengthen support for their revolutionary government, and the French people as well as the Assembly thought that they would win a war against Austria and Prussia. On 20 April 1792, France declared war on Austria. Late April 1792, France invaded and conquered the Austrian Netherlands (roughly present-day Belgium and Luxembourg). The Legislative Assembly degenerated into chaos before October 1792. Francis Charles Montague concluded in 1911, "In the attempt to govern, the Assembly failed altogether. It left behind an empty treasury, an undisciplined army and navy, and a people debauched by safe and successful riot." Lyons argues that the Constituent Assembly had liberal, rational, and individualistic goals that seem to have been largely achieved by 1791. However, it failed to consolidate the gains of the Revolution, which continued with increasing momentum and escalating radicalism until 1794. Lyons identifies six reasons for this escalation. First, the king did not accept the limitations on his powers, and mobilised support from foreign monarchs to reverse it. Second, the effort to overthrow the Roman Catholic Church, sell off its lands, close its monasteries and its charitable operations, and replace it with an unpopular makeshift system caused deep consternation among the pious and the peasants. Third, the economy was badly hurt by the issuance of ever increasing amounts of paper money ("assignats"), which caused more and more inflation; the rising prices hurt the urban poor who spent most of their income on food. Fourth, the rural peasants demanded liberation from the heavy system of taxes and dues owed to local landowners. Fifth, the working class of Paris and the other cities – the sans-culottes – resented the fact that the property owners and professionals had taken all the spoils of the Revolution. Finally, foreign powers threatened to overthrow the Revolution, which responded with extremism and systematic violence in its own defence. In the summer of 1792, a large number of Parisians were against the monarchy, and hoped that the Assembly would depose the king, but the Assembly hesitated. At dawn of 10 August 1792, a large, angry crowd of insurgents and popular militias, supported by the revolutionary Paris Commune, marched on the Tuileries Palace where the king resided, assailed the Palace and killed the Swiss Guards who were assigned for the protection of the king. Around 8:00 am the king decided to leave his palace and seek safety with his wife and children in the Assembly that was gathered in permanent session in Salle du Manège opposite to the Tuileries. The royal family became prisoners. After 11:00 am, a rump session of the Legislative Assembly 'temporarily relieved the king from his task' and thus suspended the monarchy; little more than a third of the deputies were present, almost all of them Jacobins. In reaction, on 19 August the Prussian general Duke of Brunswick invaded France and besieged Longwy. On 26 August, the Assembly decreed the deportation of refractory priests in the west of France, as "causes of danger to the fatherland", to destinations like French Guiana. In reaction, peasants in the Vendée took over a town, in another step toward civil war. What remained of a national government depended on the support of the insurrectionary Commune. With enemy troops advancing, the Commune looked for potential traitors in Paris. On 2, 3 and 4 September 1792, hundreds of Parisians, supporters of the revolution, infuriated by Verdun being captured by the Prussian enemy, the uprisings in the west of France, and rumours that the incarcerated prisoners in Paris were conspiring with the foreign enemy, raided the Parisian prisons and murdered between 1,000 and 1,500 prisoners, many of them Catholic priests, aristocrats but also common criminals. Jean-Paul Marat, a political ally of Robespierre, in an open letter on 3 September incited the rest of France to follow the Parisian example; Robespierre kept a low profile in regard to the murder orgy. The Assembly and the city council of Paris "(la Commune)" seemed inept and hardly motivated to call a halt to the unleashed bloodshed. The Commune sent gangs of National Guardsmen and "fédérés" into the prisons, and they killed 1,400 or more victims, mostly nonjuring priests. The Commune then sent a circular letter to the other cities of France inviting them to follow this example, and many cities launched their own massacres of prisoners and priests in the "September massacres". The Assembly could offer only feeble resistance. In October, however, there was a counterattack accusing the instigators, especially Marat, of being terrorists. This led to a political contest between the more moderate Girondists and the more radical Montagnards inside the convention, with rumour used as a weapon by both sides. The Girondists lost ground when they seemed too conciliatory. But the pendulum swung again and after Thermidor, the men who had endorsed the massacres were denounced as terrorists. Chaos persisted until the Convention, elected by universal male suffrage and charged with writing a new constitution, met on 20 September 1792 and became the new "de facto" government of France. The next day it abolished the monarchy and declared a republic. The following day – 22 September 1792, the first morning of the new Republic – was later retroactively adopted as the beginning of Year One of the French Republican Calendar. From May 1792 to June 1815 France was engaged almost continuously (with two short breaks) in wars with Britain and a changing coalition of other major powers. The many French successes led to the spread of the French revolutionary ideals into neighbouring countries, and indeed across much of Europe. However, the final defeat of Napoleon in 1814 (and 1815) brought a reaction that reversed some – but not all – of the revolutionary achievements in France and Europe. The Bourbons were restored to the throne, with the brother of King Louis XVI becoming King Louis XVIII. The politics of the period inevitably drove France towards war with Austria and its allies. The King, many of the Feuillants, and the Girondins specifically wanted to wage war. The King (and many Feuillants with him) expected war would increase his personal popularity; he also foresaw an opportunity to exploit any defeat: either result would make him stronger. The Girondins wanted to export the Revolution throughout Europe and, by extension, to defend the Revolution within France. The forces opposing war were much weaker. Barnave and his supporters among the Feuillants feared a war they thought France had little chance to win and which they feared might lead to greater radicalisation of the revolution. On the other end of the political spectrum Robespierre opposed a war on two grounds, fearing that it would strengthen the monarchy and military at the expense of the revolution, and that it would incur the anger of ordinary people in Austria and elsewhere. The Austrian emperor Leopold II, brother of Marie Antoinette, may have wished to avoid war, but he died on 1 March 1792. France preemptively declared war on Austria (20 April 1792) and Prussia joined on the Austrian side a few weeks later. The invading Prussian army faced little resistance until it was checked at the Battle of Valmy (20 September 1792) and forced to withdraw. The new-born Republic followed up on this success with a series of victories in Belgium and the Rhineland in the fall of 1792. The French armies defeated the Austrians at the Battle of Jemappes on 6 November, and had soon taken over most of the Austrian Netherlands. This brought them into conflict with Britain and the Dutch Republic, which wished to preserve the independence of the southern Netherlands from France. After the French king's execution in January 1793, these powers, along with Spain and most other European states, joined the war against France. Almost immediately, French forces suffered defeats on many fronts, and were driven out of their newly conquered territories in the spring of 1793. At the same time, the republican regime was forced to deal with rebellions against its authority in much of western and southern France. But the allies failed to take advantage of French disunity, and by the autumn of 1793 the republican regime had defeated most of the internal rebellions and halted the allied advance into France itself. This stalemate ended in the summer of 1794 with dramatic French victories. The French defeated the allied army at the Battle of Fleurus, leading to a full Allied withdrawal from the Austrian Netherlands. They pushed the allies to the east bank of the Rhine, allowing France, by the beginning of 1795, to conquer the Dutch Republic itself. The House of Orange was expelled and replaced by the Batavian Republic, a French satellite state. These victories led to the collapse of the anti-French coalition. Prussia, having effectively abandoned the coalition in the fall of 1794, made peace with revolutionary France at Basel in April 1795, and soon thereafter Spain also made peace with France. Britain and Austria were the only major powers to remain at war with France. Although the French Revolution had a dramatic impact in numerous areas of Europe, the French colonies felt a particular influence. As the Martinican author Aimé Césaire put it, "there was in each French colony a specific revolution, that occurred on the occasion of the French Revolution, in tune with it." The Haitian Revolution (Saint Domingue) became a central example of slave uprisings in French colonies. In the 1780s, Saint-Domingue had been France's wealthiest colony, producing more sugar than all the British West Indies colonies put together. During the Revolution, the National Convention voted to abolish slavery in February 1794, months after the rebelling slaves had already announced an abolition of slavery in Saint-Domingue. However, the 1794 decree was only implemented in Saint-Domingue, Guadeloupe and Guyane, and was a dead letter in Senegal, Mauritius, Réunion and Martinique, the last of which had been conquered by the British, who maintained the institution of slavery on that Caribbean island. Late in August/early September 1792, elections were held, now under universal male suffrage, for the new National Convention, which replaced the Legislative Assembly on 20 September 1792. From the start the Convention suffered from the bitter division between a group (political faction) around Robespierre, Danton and Marat, referred to as 'Montagnards' or the 'extreme left', and a group around Brissot, referred to as 'Girondins' as moderate republicans. But the majority of the representatives, referred to as 'la Plaine', were member of neither of those two antagonistic groups and managed to preserve some speed in the convention's debates; The Plain was formed by independents (as Barère, Cambon and Carnot) but dominated by the radical Mountain. The Plaine occupied the middle only in the sense that its votes determined which group of revolutionary leaders should enjoy the sanctions of relative legality. Immediately on 21 September the Convention abolished the monarchy, making France the French First Republic. A new French Republican Calendar was introduced on 24 October 1793 to replace the Christian Gregorian calendar, renaming the year 1792 as year 1 of the Republic. With wars against Prussia and Austria having started earlier in 1792, France also declared war on the Kingdom of Great Britain and the Dutch Republic in February 1793. In the course of 1793, the Holy Roman Empire, the kings of Spain, Portugal and Naples and the Grand-Duke of Tuscany declared war against France. In the Brunswick Manifesto, the Imperial and Prussian armies threatened retaliation on the French population if it were to resist their advance or the reinstatement of the monarchy. This among other things made Louis appear to be conspiring with the enemies of France. On 17 January 1793 Louis was condemned to death for "conspiracy against the public liberty and the general safety" by a close majority in Convention: 361 voted to execute the king, 288 voted against, and another 72 voted to execute him subject to a variety of delaying conditions. The former Louis XVI, now simply named "Citoyen Louis Capet" (Citizen Louis Capet) was executed by guillotine on 21 January 1793 on the "Place de la Révolution", former "Place Louis XV", now called the Place de la Concorde. Conservatives across Europe were horrified and monarchies called for war against revolutionary France. When prices rose in February the "sans-culottes" – poor labourers and craftsmen – rioted; the Girondins were held responsible. On 24 February the Convention decreed the first, but unsuccessful Levée en Masse as the attempt to draft new troops set off an uprising in rural France. The Montagnards lost influence in Marseille, Toulon and Lyon. This encouraged the Jacobins to seize power through a parliamentary "coup", backed up by force effected by mobilising public support against the Girondin faction, and by utilising the mob power of the Parisian "sans-culottes". An alliance of Jacobin and "sans-culottes" elements thus became the effective center of the new government. Policy became considerably more radical, as "The Law of the Maximum" set food prices and led to executions of offenders. The price control policy was coeval with the rise to power of the Committee of Public Safety and the Reign of Terror. The Committee first attempted to set the price for only a limited number of grain products, but by September 1793 it expanded the "maximum" to cover all foodstuffs and a long list of other goods. Widespread shortages and famine ensued. The Committee reacted by sending a small revolutionary army into the countryside to arrest farmers and seize crops. This temporarily solved the problem in Paris, but the rest of the country suffered. By the spring of 1794, forced collection of food was not sufficient to feed even Paris, and the days of the committee were numbered. When Robespierre was sent to the guillotine in July 1794, the crowd jeered, "There goes the dirty maximum!" The Reign of Terror, or more commonly The Terror (), refers to a period of the French Revolution when numerous public executions took place in response to revolutionary fervour, anti-clerical, anti-federalist and anti-aristocratic sentiment, and spurious accusations of treason by Maximilien Robespierre and the Committee of Public Safety. There is disagreement among historians over when exactly “the Terror” started, either in June 1793 or September 1793. In July the Committee of Public Safety came under the control of Maximilien Robespierre, and the Jacobins unleashed the Reign of Terror (1793–94). According to archival records, at least 16,594 people died under the guillotine or otherwise after accusations of counter-revolutionary activities. As many as 40,000 accused prisoners may have been summarily executed without trial or died awaiting trial. On 2 June 1793, Paris sections – encouraged by the "enragés" ("enraged ones") Jacques Roux and Jacques Hébert – took over the Convention, calling for administrative and political purges, a low fixed price for bread, and a limitation of the electoral franchise to "sans-culottes" alone. With the backing of the National Guard, they managed to persuade the convention to arrest ten members of the Commission of Twelve and 21 Girondin leaders, including Jacques Pierre Brissot. Following these arrests, the Jacobins gained control of the Committee of Public Safety on 10 June, installing the "revolutionary dictatorship". On 24 June, the Convention adopted the first republican constitution of France, variously referred to as the French Constitution of 1793 or Constitution of the Year I. It was progressive and radical in several respects, in particular by establishing universal male suffrage. It was ratified by public referendum, but normal legal processes were suspended before it could take effect. On 13 July, the assassination of Jean-Paul Marat – a Jacobin leader and journalist known for his bloodthirsty rhetoric – by Charlotte Corday, a Girondin, resulted in further increase of Jacobin political influence. Georges Danton was removed from the Committee and Robespierre, "the Incorruptible", became its most influential member as it moved to take radical measures against the Revolution's domestic and foreign enemies. The Reign of Terror ultimately weakened the revolutionary government, while temporarily ending internal opposition. The Jacobins expanded the size of the army, and Carnot replaced many aristocratic officers with soldiers who had demonstrated their patriotism, if not their ability. The Republican army repulsed the Austrians, Prussians, British, and Spanish. At the end of 1793, the army began to prevail and revolts were defeated with ease. The Ventôse Decrees (February–March 1794) proposed the confiscation of the goods of exiles and opponents of the Revolution, and their redistribution to the needy. However, this policy was never fully implemented. Three approaches attempt to explain the Reign of Terror imposed by the Jacobins in 1793–94. The older Marxist interpretation argued the Terror was a necessary response to outside threats (in terms of other countries going to war with France) and internal threats (of traitors inside France threatening to frustrate the Revolution). In this interpretation, as expressed by the Marxist historian Albert Soboul, Robespierre and the sans-culottes were heroes for defending the revolution from its enemies. François Furet has argued that foreign threats had little to do with the terror. Instead, the extreme violence was an inherent part of the intense ideological commitment of the revolutionaries – their Utopian goals required exterminating opposition. Soboul's Marxist interpretation has been largely abandoned by most historians since the 1990s. Hanson (2009) takes a middle position, recognising the importance of the foreign enemies, and sees the terror as a contingency that was caused by the interaction of a series of complex events and the foreign threat. Hanson says the terror was not inherent in the ideology of the Revolution, but that circumstances made it necessary. Introduction of a nationwide conscription for the army in February 1793 was the spark that in March made the Vendée, already rebellious since 1790 because of the changes imposed on the Roman Catholic Church by the Civil Constitution of the Clergy (1790), ignite into civil (guerrilla) war against the French Revolutionary government in Paris. North of the Loire, similar revolts were started by the so-called Chouans (royalist rebels). In March 1793, France also declared war on Spain, the Vendée rebels won some victories against Paris, and the French army was defeated in Belgium by Austria with the French general Dumouriez defecting to the Austrians: the French Republic's survival was now in real danger. Facing local revolts and foreign invasions in both the East and West of the country, the most urgent government business was the war. On 6 April 1793, to prevent the Convention from losing itself in abstract debate and to streamline government decisions, the "Comité de salut public" (Committee of Public Prosperity) was created, as executive government which was accountable to the convention. In April 1793, the Girondins indicted Jean-Paul Marat before the Revolutionary Tribunal for 'attempting to destroy the sovereignty of the people' and 'preaching plunder and massacre', referring to his behaviour during the September massacres. Marat was quickly acquitted but the incident further exacerbated the 'Girondins' versus 'Montagnards' party strife in the convention. Jacques Hébert, Convention member leaning to the 'Cordeliers' group, on 24 May 1793 called on the "sans-culottes" to rise in revolt against the "henchmen of Capet [the ex-king] and Dumouriez [the defected general]". Hébert was arrested by a Convention committee. While that committee consisted only of members from The Plain and the Girondins, the anger of the sans-culottes was directed towards the Girondins. 25 May, a delegation of "la Commune" (the Paris city council) protested against Hébert's arrest. The convention's President Isnard, a Girondin, answered them: "Members of "la Commune" [...] If by your incessant rebellions something befalls to the representatives of the nation, I declare, in the name of France, that Paris will be totally obliterated". On 2 June 1793, the convention's session in Tuileries Palace degenerated into chaos and pandemonium. Crowds of people swarmed in and around the palace. Incessant screaming from the public galleries suggested that all of Paris was against the Girondins. Petitions circulated, indicting and condemning 22 Girondins. Barère, member of the "Comité de salut public", suggested: to end this division which is harming the Republic, the Girondin leaders should lay down their offices voluntarily. Late that night after much more tumultuous debate, dozens of Girondins had resigned and left the convention. By the summer of 1793, most French departments in one way or another opposed the central Paris government. Girondins who fled from Paris after 2 June led those revolts. In Brittany's countryside, the people rejecting the Civil Constitution of the Clergy of 1790 had taken to a guerrilla warfare known as "Chouannerie". But generally, the French opposition against 'Paris' had now evolved into a plain struggle for power over the country against the 'Montagnards' around Robespierre and Marat now dominating Paris. In June–July 1793, Bordeaux, Lyon, Toulon, Marseilles, Caen, Brittany and the rest of Normandy gathered armies to march on Paris and against 'the revolution'. In July, the deposed 'Montagnard' head of the Lyon city council was guillotined. On 1 August Barère incited the convention to tougher measures against the Vendée, at war with Paris since March: "We'll have peace only when no Vendée remains [...] we'll have to exterminate that rebellious people". In August, Convention troops besieged Lyon. On 17 August 1793, the Convention voted for general conscription, a second "levée en masse", which mobilised all citizens to serve as soldiers or suppliers in the war effort. By August, political disagreement seemed enough to be summoned before the Tribunal; appeal against a Tribunal verdict was impossible. Late August 1793, Adam Philippe Custine an army general had been guillotined on the accusation of choosing too timid strategies on the battlefield. Early September 1793, militants urged the convention to do more to quell the counter-revolution. A delegation of the "Commune" (Paris city council) suggested to form revolutionary armies to arrest hoarders and conspirators. Barère, on 5 September reacted favorably, saying: let's "make terror the order of the day!" Criteria for bringing someone before the Revolutionary Tribunal had always been vast and vague. On 9 September The National Convention voted to establish "sans-culottes" paramilitary forces, revolutionary armies, and to force farmers to surrender grain demanded by the government. On 17 September, the Law of Suspects was passed, which ordered the arrest of suspected counter-revolutionaries and people who had revealed themselves as "enemies of freedom". This decree was one of the causes for some 17,000 legal death sentences until the end of July 1794, an average of 370 per week – reason for historians to label those 10 months 'the (Reign of) Terror'. On 19 September the Vendée rebels again defeated a Republican Convention army. On 29 September, the Convention extended price limits from grain and bread to other household goods and established the Law of the Maximum, intended to prevent price gouging and supply food to the cities. On 1 October Barère repeated his plea to subdue the Vendée: "refuge of fanaticism, where priests have raised their altars". On 9 October the Convention troops captured Lyon and reinstated a Montagnard government there. On 10 October the Convention decreed to recognize the Committee of Public Safety as the supreme "Revolutionary Government", (which was consolidated on 4 December). The provisional government would be revolutionary until peace according to Saint-Just. Though the French Constitution of 1793 was overwhelmingly popular and its drafting and ratification buoyed popular support for the Montagnards, the Convention set it aside indefinitely until a future peace. Mid-October, the widowed former queen Marie Antoinette was on trial for a long list of charges such as "teaching [her husband] Louis Capet the art of dissimulation" and incest with her son, she too was guillotined. At the end of October 1793, 21 former 'Girondins' Convention members who hadn't left Paris after June were convicted to death and executed, on the charge of federalism, verbally supporting the preparation of an insurrection in Caen by fellow-Girondins. 17 October 1793, the 'blue' Republican army near Cholet defeated the 'white' Vendéan insubordinate army and all surviving Vendée residents, counting in tens of thousands, fled over the river Loire north into Brittany. A Convention's representative on mission in Nantes commissioned in October to pacify the region did so by simply drowning prisoners in the river Loire: until February 1794 he drowned at least 4,000. Meanwhile, the instalment of the Republican Calendar on 24 October 1793 caused an anti-clerical uprising. Hébert's and Chaumette's atheist movement campaigned to dechristianise society. The climax was reached with the celebration of the flame of Reason in Notre Dame Cathedral on 10 November. By November 1793, the revolts in Normandy, Bordeaux and Lyon were overcome, in December also that in Toulon. Two representatives on mission sent to punish Lyon between November 1793 and April 1794 executed 2,000 people to death by guillotine or firing-squad. The Vendéan army since October roaming through Brittany on 12 December 1793 again ran up against Republican troops and saw 10,000 of its rebels perish, meaning the end of this once threatening army. Some historians claim that after that Vendéan defeat Convention Republic armies in 1794 massacred 117,000 Vendéan civilians to obliterate the Vendéan people, but others contest that claim. Some historians consider the total civil war to have lasted until 1796 with a toll of 170,000 or 450,000 lives. Because of the extremely brutal forms that the Republican repression took in many places, historians such as Reynald Secher have called the event a "genocide". Historian François Furet concluded that the repression in the Vendee "not only revealed massacre and destruction on an unprecedented scale but also a zeal so violent that it has bestowed as its legacy much of the region's identity." The guillotine became the tool for a string of executions. Louis XVI had already been guillotined before the start of the terror; Queen Marie Antoinette, Barnave, Bailly, Brissot and other leading Girondins, Philippe Égalité (despite his vote for the death of the King), Madame Roland and many others were executed by guillotine. The Revolutionary Tribunal summarily condemned thousands of people to death by the guillotine, while mobs beat other victims to death. At the peak of the terror, the slightest hint of counter-revolutionary thoughts or activities (or, as in the case of Jacques Hébert, revolutionary zeal exceeding that of those in power) could place one under suspicion, and trials did not always proceed according to contemporary standards of due process. Sometimes people died for their political opinions or actions, but many for little reason beyond mere suspicion, or because some others had a stake in getting rid of them. Most of the victims received an unceremonious trip to the guillotine in an open wooden cart (the tumbrel). In the rebellious provinces, the government representatives had unlimited authority and some engaged in extreme repressions and abuses. For example, Jean-Baptiste Carrier became notorious for the "Noyades" ("drownings") he organised in Nantes; his conduct was judged unacceptable even by the Jacobin government and he was recalled. Maximilien Robespierre, since July 1793 member of the Committee of Public Prosperity, on 5 February 1794 in a speech in the Convention identified Jacques Hébert and his faction as "internal enemies" working toward the triumph of tyranny. After a dubious trial Hébert and some allies, charged with counter-revolutionary activities, were guillotined in March. On 5 April, again at the instigation of Robespierre, Danton, a moderate Montagnard, and 13 associated politicians, charged with counter-revolutionary activities, were executed. A week later again 19 politicians. This hushed the Convention deputies: if henceforth they disagreed with Robespierre they hardly dared to speak out. On 7 June 1794, Robespierre advocated a new state religion and recommended the Convention acknowledge the existence of the "Supreme Being". A law enacted on 10 June 1794 (22 Prairial II) further streamlined criminal procedures: if the Revolutionary Tribunal saw sufficient proof of someone being an "enemy of the people" a counsel for defence would not be allowed. The frequency of guillotine executions in Paris now rose from on average three a day to an average of 29 a day. Meanwhile, France's external wars were going well, with victories over Austrian and British troops in May and June 1794 opening up Belgium for French conquest. But cooperation within the Committee of Public Prosperity, since April 1793 the "de facto" executive government, started to break down. On 29 June 1794, three colleagues of Robespierre at 'the Committee' called him a dictator in his face – Robespierre baffled left the meeting. This encouraged other Convention members to also defy Robespierre. On 26 July, a long and vague speech of Robespierre wasn't met with thunderous applause as usual but with hostility; some deputies yelled that Robespierre should have the courage to say which deputies he deemed necessary to be killed next, what Robespierre refused to do. In the Convention session of 27 July 1794 (9 Thermidor of Year II), Robespierre and his allies hardly managed to say a word as they were constantly interrupted by a row of critics such as Tallien, Billaud-Varenne, Vadier, Barère and acting president Thuriot. Finally, even Robespierre's own voice failed on him: it faltered at his last attempt to beg permission to speak. A decree was adopted to arrest Robespierre, Saint-Just and Couthon. 28 July, they and 19 other leading Jacobins were beheaded. 29 July, again 70 members of the Parisian Commune were guillotined. Subsequently, the Law of 22 Prairial (10 June 1794) was repealed, and the 'Girondins' expelled from the Convention in June 1793, if not dead yet, were reinstated as Convention deputies. In November the Jacobin Club was closed and banned. After July 1794, the French government was dominated by 'Girondins', who indulged in revenge and violence and death sentences against people associated with the previous 'Jacobin'/'Montagnard' governments around Robespierre and Marat, in what was known as the White Terror.
https://en.wikipedia.org/wiki?curid=11188
Francis Fukuyama Yoshihiro Francis Fukuyama (, ; born October 27, 1952) is an American political scientist, political economist, and writer. Fukuyama is known for his book "The End of History and the Last Man" (1992), which argued that the worldwide spread of liberal democracies and free-market capitalism of the West and its lifestyle may signal the end point of humanity's sociocultural evolution and become the final form of human government. However, his subsequent book "Trust: Social Virtues and Creation of Prosperity" (1995) modified his earlier position to acknowledge that culture cannot be cleanly separated from economics. Fukuyama is also associated with the rise of the neoconservative movement, from which he has since distanced himself. Fukuyama has been a senior fellow at the Freeman Spogli Institute for International Studies since July 2010 and a Mosbacher Director of the Center on Democracy, Development and the Rule of Law at Stanford University. In August 2019, he was named director of the Ford Dorsey Master's in International Policy at Stanford. Before that, he served as a professor and director of the International Development program at the School of Advanced International Studies of Johns Hopkins University. Previously, he was Omer L. and Nancy Hirst Professor of Public Policy at the School of Public Policy at George Mason University. He is a council member of the International Forum for Democratic Studies founded by the National Endowment for Democracy and was a member of the Political Science Department of the RAND Corporation. He is also one of the 25 leading figures on the Information and Democracy Commission launched by Reporters Without Borders. Francis Fukuyama was born in the Hyde Park neighborhood of Chicago, Illinois, United States. His paternal grandfather fled the Russo-Japanese War in 1905 and started a shop on the west coast before being interned in the Second World War. His father, Yoshio Fukuyama (), a second-generation Japanese American, was trained as a minister in the Congregational Church, received a doctorate in sociology from the University of Chicago, and taught religious studies. His mother, Toshiko Kawata Fukuyama (), was born in Kyoto, Japan, and was the daughter of Shiro Kawata (), founder of the Economics Department of Kyoto University and first president of Osaka City University. Francis grew up in Manhattan as an only child, had little contact with Japanese culture, and did not learn Japanese. His family moved to State College, Pennsylvania, in 1967. Fukuyama received his Bachelor of Arts degree in classics from Cornell University, where he studied political philosophy under Allan Bloom. He initially pursued graduate studies in comparative literature at Yale University, going to Paris for six months to study under Roland Barthes and Jacques Derrida but became disillusioned and switched to political science at Harvard University. There, he studied with Samuel P. Huntington and Harvey Mansfield, among others. He earned his Ph.D. in political science at Harvard for his thesis on Soviet threats to intervene in the Middle East. In 1979, he joined the global policy think tank RAND Corporation. Fukuyama lived at the Telluride House and has been affiliated with the Telluride Association since his undergraduate years at Cornell, an education enterprise that was home to other significant leaders and intellectuals, including Steven Weinberg, Paul Wolfowitz and Kathleen Sullivan. Fukuyama was the Omer L. and Nancy Hirst Professor of Public Policy in the School of Public Policy at George Mason University from 1996 to 2000. Until July 10, 2010, he was the Bernard L. Schwartz Professor of International Political Economy and Director of the International Development Program at the Paul H. Nitze School of Advanced International Studies of Johns Hopkins University in Washington, D.C. He is now Olivier Nomellini Senior Fellow and resident in the Center on Democracy, Development, and the Rule of Law at the Freeman Spogli Institute for International Studies at Stanford University, and director of the Ford Dorsey Master's in International Policy at Stanford. Fukuyama is best known as the author of "The End of History and the Last Man", in which he argued that the progression of human history as a struggle between ideologies is largely at an end, with the world settling on liberal democracy after the end of the Cold War and the fall of the Berlin Wall in 1989. Fukuyama predicted the eventual global triumph of political and economic liberalism: Authors like Ralf Dahrendorf argued in 1990 that the essay gave Fukuyama his 15 minutes of fame, which will be followed by a slide into obscurity. He continued to remain a relevant and cited public intellectual leading American communitarian Amitai Etzioni to declare him "one of the few enduring public intellectuals. They are often media stars who are eaten up and spat out after their 15 minutes. But he has lasted." According to Fukuyama, one of the main reasons for the massive criticism against "The End of History" was the aggressive stance that it took towards postmodernism. Postmodern philosophy had, in Fukuyama's opinion, undermined the ideology behind liberal democracy, leaving the western world in a potentially weaker position. The fact that Marxism and fascism had been proven untenable for practical use while liberal democracy still thrived was reason enough to embrace the hopeful attitude of the Progressive era, as this hope for the future was what made a society worth struggling to maintain. Postmodernism, which, by this time, had become embedded in the cultural consciousness, offered no hope and nothing to sustain a necessary sense of community, instead relying only on lofty intellectual premises. In the 2011 book, Fukuyama describes what makes a state stable. It uses comparative political history to develop a theory of the stability of a political system. According to Fukuyama, an ideal political order needs a modern and effective state, the rule of law governing the state and be accountable. The 2014 book is the second book on political order, following the 2011 book "The Origins of Political Order." In this book, Fukuyama covers events taking place since the French Revolution and shed light on political institutions and their development in different regions. After tracing how a modern and effective government was developed in the U.S., Fukuyama asserts that it is experiencing political decay. Fukuyama believes that political decay can be seen in the deterioration of bureaucracies, special interest groups capturing the legislature, and inevitable but cumbersome judicial processes challenging all types of government action. Fukuyama has written a number of other books, among them "Trust: The Social Virtues and the Creation of Prosperity" and "Our Posthuman Future: Consequences of the Biotechnology Revolution". In the latter, he qualified his original "end of history" thesis, arguing that since biotechnology increasingly allows humans to control their own evolution, it may allow humans to alter human nature, thereby putting liberal democracy at risk. One possible outcome could be that an altered human nature could end in radical inequality. He is a fierce enemy of transhumanism, an intellectual movement asserting that posthumanity is a desirable goal. In another work, "The Great Disruption: Human Nature and the Reconstruction of Social Order", Fukuyama explores the origins of social norms, and analyses the current disruptions in the fabric of our moral traditions, which he considers as arising from a shift from the manufacturing to the information age. This shift is, he thinks, normal and will prove self-correcting, given the intrinsic human need for social norms and rules. In 2006, in "America at the Crossroads", Fukuyama discusses the history of neoconservatism, with particular focus on its major tenets and political implications. He outlines his rationale for supporting the Bush administration, as well as where he believes it has gone wrong. In 2008, Fukuyama published the book "Falling Behind: Explaining the Development Gap Between Latin America and the United States", which resulted from research and a conference funded by Grupo Mayan to gain understanding on why Latin America, once far wealthier than North America, fell behind in terms of development in only a matter of centuries. Discussing this book at a 2009 conference, Fukuyama outlined his belief that inequality within Latin American nations is a key impediment to growth. An unequal distribution of wealth, he stated, leads to social upheaval, which then results in stunted growth. His most recent book, published in 2018, is "Identity: The Demand for Dignity and the Politics of Resentment". As a key Reagan Administration contributor to the formulation of the Reagan Doctrine, Fukuyama is an important figure in the rise of neoconservatism, although his works came out years after Irving Kristol's 1972 book crystallized neoconservatism. Fukuyama was active in the Project for the New American Century think tank starting in 1997, and as a member co-signed the organization's 1998 letter recommending that President Bill Clinton support Iraqi insurgencies in the overthrow of then-President of Iraq Saddam Hussein. He was also among forty co-signers of William Kristol's September 20, 2001 letter to President George W. Bush after the September 11, 2001 attacks that suggested the U.S. not only "capture or kill Osama bin Laden", but also embark upon "a determined effort to remove Saddam Hussein from power in Iraq". In a "New York Times" article from February 2006, Fukuyama, in considering the ongoing Iraq War, stated: "What American foreign policy needs is not a return to a narrow and cynical realism, but rather the formulation of a 'realistic Wilsonianism' that better matches means to ends." In regard to neoconservatism, he went on to say: "What is needed now are new ideas, neither neoconservative nor realist, for how America is to relate to the rest of the world – ideas that retain the neoconservative belief in the universality of human rights, but without its illusions about the efficacy of American power and hegemony to bring these ends about." Fukuyama began to distance himself from the neoconservative agenda of the Bush administration, citing its excessive militarism and embrace of unilateral armed intervention, particularly in the Middle East. By late 2003, Fukuyama had voiced his growing opposition to the Iraq War and called for Donald Rumsfeld's resignation as Secretary of Defense. At an annual dinner of the American Enterprise Institute in February 2004, Dick Cheney and Charles Krauthammer declared the beginning of a unipolar era under American hegemony. "All of these people around me were cheering wildly," Fukuyama remembers. He believes that the Iraq War was being blundered. "All of my friends had taken leave of reality." He has not spoken to Paul Wolfowitz (previously a good friend) since. Fukuyama declared he would not be voting for Bush, and that the Bush administration had made three mistakes: Fukuyama believes the US has a right to promote its own values in the world, but more along the lines of what he calls "realistic Wilsonianism", with military intervention only as a last resort and only in addition to other measures. A latent military force is more likely to have an effect than actual deployment. The US spends 43% of global military spending, but Iraq shows there are limits to its effectiveness. The US should instead stimulate political and economic development and gain a better understanding of what happens in other countries. The best instruments are setting a good example and providing education and, in many cases, money. The secret of development, be it political or economic, is that it never comes from outsiders, but always from people in the country itself. One thing the US proved to have excelled in during the aftermath of World War II was the formation of international institutions. A return to support for these structures would combine American power with international legitimacy, but such measures require a lot of patience. This is the central thesis of his 2006 work "America at the Crossroads". In a 2006 essay in "The New York Times Magazine" strongly critical of the invasion, he identified neoconservatism with Leninism. He wrote that neoconservatives: Fukuyama announced the end of the neoconservative moment and argued for the demilitarization of the War on Terrorism: Fukuyama endorsed Barack Obama in the 2008 US presidential election. He states: In a 2018 interview with "New Statesman", when asked about his views on the resurgence of socialist politics in the United States and the United Kingdom, he responded: In a review for "The Washington Post", Fukuyama discussed Ezra Klein's 2020 book "Why We're Polarized" regarding US politics, and outlined Klein's central conclusion about the importance of race and white identity to Donald Trump voters and Republicans. Fukuyama is a part-time photographer. He also has a keen interest in early-American furniture, which he reproduces by hand. He is keenly interested in sound recording and reproduction, saying, "These days I seem to spend as much time thinking about gear as I do analyzing politics for my day job." Fukuyama is married to Laura Holmgren, whom he met when she was a UCLA graduate student after he started working for the RAND Corporation. He dedicated his book "Trust: The Social Virtues and the Creation of Prosperity" to her. They live in California, with their three children, Julia, David, and John. He is the first cousin to crime novelist Joe Ide. Francis helped him get his first book published.
https://en.wikipedia.org/wiki?curid=11194
Fingerspelling Fingerspelling (or dactylology) is the representation of the letters of a writing system, and sometimes numeral systems, using only the hands. These manual alphabets (also known as finger alphabets or hand alphabets), have often been used in deaf education, and have subsequently been adopted as a distinct part of a number of sign languages; there are about forty manual alphabets around the world. Historically, manual alphabets have had a number of additional applications—including use as ciphers, as mnemonics, and in silent religious settings. As with other forms of manual communication, fingerspelling can be comprehended visually or tactually. The simplest visual form of fingerspelling is tracing the shape of letters in the air, or tactually, tracing letters on the hand. Fingerspelling can be one-handed such as in American Sign Language, French Sign Language and Irish Sign Language, or it can be two-handed such as in British Sign Language. There are two families of manual alphabets used for representing the Latin alphabet in the modern world. The more common of the two is mostly produced on one hand, and can be traced back to alphabetic signs used in Europe from at least the early 15th century. The alphabet, first described completely by Spanish monks, was adopted by the Abbé de l'Épée's deaf school in Paris in the 18th century, and was then spread to deaf communities around the world in the 19th and 20th centuries via educators who had learned it in Paris. Over time, variations have emerged, brought about by natural phonetic changes that occur over time, adaptions for local written forms with special characters or diacritics (which are sometimes represented with the other hand), and avoidance of handshapes that are considered obscene in some cultures. The most widely used modern descendant is the American manual alphabet. Two-handed manual alphabets are used by a number of deaf communities; one such alphabet is shared by users of British Sign Language, Auslan, and New Zealand Sign Language (collectively known as the BANZSL language family), while another is used in Turkish Sign Language. Some of the letters are represented by iconic shapes, and in the BANZSL languages the vowels are represented by pointing to the fingertips. Letters are formed by a dominant hand, which is on top of or alongside the other hand at the point of contact, and a subordinate hand, which uses either the same or a simpler handshape as the dominant hand. Either the left or right hand can be dominant. In a modified tactile form used by deafblind people, the signer's hand acts as the dominant hand, and the receiver's hand becomes the subordinate hand. Some signs, such as the sign commonly used for the letter "C", may be one-handed. Manual alphabets based on the Arabic alphabet, the Ethiopian Ge'ez script and the Korean Hangul script use handshapes that are more or less iconic representations of the characters in the writing system. Some manual representations of non-Roman scripts such as Chinese, Japanese, Devanagari (e.g. the Nepali manual alphabet), Hebrew, Greek, Thai and Russian alphabets are based to some extent on the one-handed Latin alphabet described above. In some cases however, the "basis" is more theory than practice. Thus, for example, in the Japanese manual syllabary only the five vowels (ア /a/, イ /i/, ウ /u/, エ /e/, オ /o/) and the Ca (consonant plus "a' vowel) letters (カ /ka/, サ /sa/, ナ /na/, ハ /ha/, マ /ma/, ヤ /ya/, ラ /ra/, ワ /wa/, but notably "not" タ /ta/ which would resemble a somewhat rude gesture) derive from the American manual alphabet. In the Nepali Sign Language it is only four "letters" which derive from the American manual alphabet: अ /a/, ब /b/, म /m/, and र /r/). The Yugoslav manual alphabet represents characters from the Serbian Cyrillic alphabet as well as Gaj's Latin alphabet. Fingerspelling has been introduced into certain sign languages by educators, and as such has some structural properties that are unlike the visually motivated and multi-layered signs that are typical in deaf sign languages. In many ways fingerspelling serves as a bridge between the sign language and the oral language that surrounds it. Fingerspelling is used in different sign languages and registers for different purposes. It may be used to represent words from an oral language which have no sign equivalent, or for emphasis, clarification, or when teaching or learning a sign language. In American Sign Language (ASL), more lexical items are fingerspelled in casual conversation than in formal or narrative signing. Different sign language speech communities use fingerspelling to a greater or lesser degree. At the high end of the scale, fingerspelling makes up about 8.7% of casual signing in ASL, and 10% of casual signing in Auslan. The proportion is higher in older signers, suggesting that the use of fingerspelling has diminished over time. Across the Tasman Sea, only 2.5% of the corpus of New Zealand Sign Language was found to be fingerspelling. Fingerspelling has only become a part of NZSL since the 1980s; prior to that, words could be spelled or initialised by tracing letters in the air. Fingerspelling does not seem to be used much in the sign languages of Eastern Europe, except in schools, and Italian Sign Language is also said to use very little fingerspelling, and mainly for foreign words. Sign languages that make no use of fingerspelling at all include Kata Kolok and Ban Khor Sign Language. The speed and clarity of fingerspelling also varies between different signing communities. In Italian Sign Language, fingerspelled words are relatively slow and clearly produced, whereas fingerspelling in standard British Sign Language (BSL) is often rapid so that the individual letters become difficult to distinguish, and the word is grasped from the overall hand movement. Most of the letters of the BSL alphabet are produced with two hands, but when one hand is occupied, the dominant hand may fingerspell onto an "imaginary" subordinate hand, and the word can be recognised by the movement. As with written words, the first and last letters and the length of the word are the most significant factors for recognition. When people fluent in sign language read fingerspelling, they do not usually look at the signer's hand(s), but maintain eye contact and look at the face of the signer as is normal for sign language. People who are learning fingerspelling often find it impossible to understand it using just their peripheral vision and must look directly at the hand of someone who is fingerspelling. Often, they must also ask the signer to fingerspell slowly. It frequently takes years of expressive and receptive practice to become skilled with fingerspelling. Some writers have suggested that the body and hands were used to represent alphabets in Greek, Roman, Egyptian and Assyrian antiquity. Certainly, "finger calculus" systems were widespread, and capable of representing numbers up to 10,000; they are still in use today in parts of the Middle East. The practice of substituting letters for numbers and vice versa, known as gematria, was also common, and it is possible that the two practices were combined to produce a finger calculus alphabet. The earliest known manual alphabet, described by the Benedictine monk Bede in 8th century Northumbria, did just that. While the usual purpose of the Latin and Greek finger alphabets described by Bede is unknown, they were unlikely to have been used by deaf people for communication — even though Bede lost his own hearing later in life. Historian Lois Bragg concludes that these alphabets were "only a bookish game." Beginning with R. A. S. Macalister in 1938, several writers have speculated that the 5th century Irish Ogham script, with its quinary alphabet system, was derived from a finger alphabet that predates even Bede. European monks from at least the time of Bede have made use of forms of manual communication, including alphabetic gestures, for a number of reasons: communication among the monastery while observing vows of silence, administering to the ill, and as mnemonic devices. They also may have been used as ciphers for discreet or secret communication. Clear antecedents of many of the manual alphabets in use today can be seen from the 16th century in books published by friars in Spain and Italy. From the same time, monks such as the Benedictine Fray Pedro Ponce de León began tutoring deaf children of wealthy patrons — in some places, literacy was a requirement for legal recognition as an heir — and the manual alphabets found a new purpose. They were originally part of the earliest known Mouth Hand Systems. The first book on deaf education, published in 1620 by Juan Pablo Bonet in Madrid, included a detailed account of the use of a manual alphabet to teach deaf students to read and speak. Meanwhile, in Britain, manual alphabets were also in use for a number of purposes, such as secret communication, public speaking, or used for communication by deaf people. In 1648, John Bulwer described "Master Babington", a deaf man proficient in the use of a manual alphabet, "contryved on the joynts of his fingers", whose wife could converse with him easily, even in the dark through the use of tactile signing. In 1680, George Dalgarno published "Didascalocophus, or, The deaf and dumb mans tutor", in which he presented his own method of deaf education, including an "arthrological" alphabet, where letters are indicated by pointing to different joints of the fingers and palm of the left hand. Arthrological systems had been in use by hearing people for some time; some have speculated that they can be traced to early Ogham manual alphabets. The vowels of this alphabet have survived in the contemporary alphabets used in British Sign Language, Auslan and New Zealand Sign Language. The earliest known printed pictures of consonants of the modern two-handed alphabet appeared in 1698 with "Digiti Lingua", a pamphlet by an anonymous author who was himself unable to speak. He suggested that the manual alphabet could also be used by mutes, for silence and secrecy, or purely for entertainment. Nine of its letters can be traced to earlier alphabets, and 17 letters of the modern two-handed alphabet can be found among the two sets of 26 handshapes depicted. Charles de La Fin published a book in 1692 describing an alphabetic system where pointing to a body part represented the first letter of the part (e.g. Brow=B), and vowels were located on the fingertips as with the other British systems. He described codes for both English and Latin. By 1720, the British manual alphabet had found more or less its present form. Descendants of this alphabet have been used by deaf communities (or at least in classrooms) in former British colonies India, Australia, New Zealand, Uganda and South Africa, as well as the republics and provinces of the former Yugoslavia, Grand Cayman Island in the Caribbean, Indonesia, Norway, Germany and the USA. Yoel (2009) demonstrated that American Sign Language is influencing the lexicon and grammar of Maritime Sign Language in various ways, including the fact that the original BANZSL two-handed manual alphabet is no longer used in the Maritimes and has been replaced by the one-handed American manual alphabet, which has been influencing lexicalisation. Although all participants in her survey had learnt and could still produce the BANSZL fingerspelling, they had difficulty doing so, and all participants indicated that it had been a long time since they last used it. Power et al. (2020) conducted a large-scale data study into the evolution and contemporary character of 76 current and defunct manual alphabets (MAs) of sign languages, postulating the existence of eight groups: a Afghan–Jordanian Group, an Austrian-origin Group (with a Danish Subgroup), a British-origin Group, a French-origin Group, a Polish Group, a Russian Group, a Spanish Group, and a Swedish Group. Notably, several defunct versions of German, Austrian, Hungarian and Danish manual alphabets were part of the Austrian-origin group, while the current MAs of these sign languages are closely related to the French, American, International Sign and other MAs in the French-origin Group. Latvian Sign Language's MA dangled somewhere between the Polish and Russian Groups, Finnish Sign Language (which belongs to the Swedish Sign Language family) had a French-origin MA, while Indo-Pakistani Sign Language (whose lexicon and grammar have independent origins) currently used a two-handed manual alphabet of British origin. Engravings of (Bonet, 1620): Other historical manual alphabets
https://en.wikipedia.org/wiki?curid=11195
Four Weddings and a Funeral Four Weddings and a Funeral is a 1994 British romantic comedy film directed by Mike Newell. It was the first of several films by screenwriter Richard Curtis to feature Hugh Grant, and follows the adventures of Charles (Grant) and his circle of friends through a number of social occasions as they each encounter romance. Andie MacDowell stars as Charles's love interest Carrie, with Kristin Scott Thomas, James Fleet, Simon Callow, John Hannah, Charlotte Coleman, David Bower, Corin Redgrave, and Rowan Atkinson in supporting roles. The film was made in six weeks, cost under £3 million, and became an unexpected success and the highest-grossing British film in history at the time, with worldwide box office in excess of $245.7 million, and receiving Academy Award nominations for Best Picture and Best Original Screenplay. Additionally, Grant won the Golden Globe Award for Best Actor - Motion Picture Musical or Comedy and the BAFTA Award for Best Actor in a Leading Role, and the film won the BAFTA Awards Best Film, Best Direction, and Best Actress in a Supporting Role for Scott Thomas. The film's success propelled Hugh Grant to international stardom, particularly in the United States. In 1999, "Four Weddings and a Funeral" placed 23rd, on the British Film Institute's 100 greatest British films of the 20th century. In 2016, "Empire" magazine ranked it 21st in their list of the 100 best British films. A 2017 poll of 150 actors, directors, writers, producers, and critics for "Time Out" magazine ranked it the 74th best British film ever. Curtis reunited director Newell and the surviving cast for a 25th anniversary reunion Comic Relief short entitled "One Red Nose Day and a Wedding", which aired in the UK during Red Nose Day on 15 March 2019. At the wedding of Angus and Laura in Somerset, unmarried best man Charles and his friends – his flatmate Scarlett, Fiona and her brother Tom, partners Gareth and Matthew, and Charles's brother David – endure the festivities. At the reception, Charles becomes smitten with Carrie, an American, and the two spend the night together. In the morning, Carrie jokingly demands that Charles propose to her, observing that they may have "missed a great opportunity", and she leaves for America. Three months later, at the wedding of Bernard and Lydia – who became romantically involved at the previous wedding – Charles meets Carrie again, now accompanied by her new fiancé Hamish. Charles faces further humiliation from several of his ex-girlfriends, including the distraught Henrietta, and retreats to an empty hotel suite; he watches Carrie and Hamish depart, and becomes trapped in a cupboard when the newlyweds stumble in to have sex. Charles makes an awkward exit and is confronted by Henrietta about his habit of “serial monogamy” and his fear of letting anyone get too close. Carrie reappears, and she and Charles spend another night together. A month later, Charles receives an invitation to Carrie's wedding. While shopping for a gift, he runs into Carrie and helps her select a wedding dress. She recounts her 33 sexual partners; Charles makes an awkward confession of his love to her, but to no avail. Another month later, Charles and his friends attend Carrie's wedding in Scotland. The gregarious Gareth instructs the group to seek potential mates; Scarlett hits it off with an American named Chester. As Charles watches Carrie and Hamish dance, Fiona deduces his heartbreak. When Charles asks why Fiona is single, she confesses that she has loved Charles since they first met; though empathetic, Charles does not requite her feelings. During the groom's toast, Gareth dies of a heart attack. At Gareth's funeral, Matthew recites the poem "Funeral Blues" by W. H. Auden, commemorating their relationship. Afterward, Carrie and Charles share a moment, and Charles and Tom ponder the fact that despite their clique's pride in being single, Gareth and Matthew were a "married" couple all the while, and whether the search for "one true love" is a futile effort. Ten months later, Charles's own wedding day arrives; the bride is revealed to be Henrietta. Shortly before the ceremony, Carrie arrives and tells Charles that she and Hamish have separated. Charles has a crisis of confidence and is counseled by David and Matthew, but proceeds with the wedding. When the vicar asks for any reason why the couple should not marry, David (who is deaf) asks Charles to translate for him: he declares in sign language that he suspects the groom loves someone else, which Charles confirms. Henrietta punches him, and the wedding is halted. Carrie tries to apologize to Charles, who confesses that at the altar, he realised the one person he truly loved was her. Charles makes a proposal of lifelong commitment without marriage to Carrie, and she accepts. Henrietta marries an officer in the Grenadier Guards; David marries his girlfriend Serena; Scarlett marries Chester; Tom marries Deirdre (a distant cousin whom he met at Charles's wedding); Matthew finds a new partner; Fiona is shown in a picture with a Prince Charles cut-out; and Charles and Carrie have a baby. Screenwriter Richard Curtis's own experiences as a wedding attendee inspired the premise for "Four Weddings and a Funeral". According to Curtis he began writing the script at age 34, after realizing he had attended 65 different weddings in an 11-year period. At one wedding he was propositioned by a fellow guest, but he turned her down and forever regretted it; accordingly he based the origin of Charles and Carrie's romance on that situation. It took Curtis 17 drafts to reach the final version. He has commented on director Mike Newell's influence; ”"I come from a school where making it funny is what matters. Mike was obsessed with keeping it real. Every character, no matter how small, has a story, not just three funny lines. It's a romantic film about love and friendship that swims in a sea of jokes." Curtis chose to omit any mention of the characters' careers, because he didn't think a group of friends would realistically discuss their jobs while together at a wedding. Curtis, Newell and the producers began the casting process for "Four Weddings" in early 1992, and continued scouting possible actors even after funding fell through in mid-1992. Roughly 70 actors auditioned for the role of Charles before Hugh Grant did. Hugh Grant was ready to give up acting as a career when he received the script for "Four Weddings and a Funeral"; he stated in 2016 that: "I wasn't really getting any work at all, and then to my great surprise this script came through the letterbox from my agent, and it was really good. And I rang on and said there must be a mistake, you've sent me a good script." Initially, writer Richard Curtis, who had modelled the character of Charles after himself, was opposed to casting Grant in the role because he thought he was too handsome. He was eventually persuaded by Mike Newell and the film's producers to approve of Grant. The original choice for the role of Carrie was Jeanne Tripplehorn, but she had to drop out because of a death in her family. Marisa Tomei and, reportedly, Sarah Jessica Parker, were also considered before Andie MacDowell was cast. Grant's participation hit another stumbling block when his agent requested a £5,000 rise over the £35,000 salary Grant was offered. The producers initially refused because of the extremely tight budget, but eventually agreed. The supporting cast-members were paid £17,500 apiece. Pre-production for the movie was a long process because funding was erratic, falling through in mid-1992 and leading to much uncertainty. Finally in early 1993, Working Title Films stepped in to close the gap. Nonetheless, another $1.2 million was cut just before production began in the summer of 1993, forcing the film to be made in just 36 days with a final budget of £2.7 million (appr. $4.4 million in 1994). The budget was so tight that extras had to wear their own wedding clothes, while Rowan Atkinson was retained as the Vicar for two of the weddings so production wouldn't have to pay another actor. Future Home Secretary and MP Amber Rudd was given the credit of "Aristocracy Coordinator" after she arranged for several aristocrats to make uncredited appearances as wedding extras, including the then – Marquess of Hartington, Peregrine Cavendish and the Earl of Woolton, who conveniently wore their own morning suits. To make Grant look more nerdy, the producers styled him with shaggy hair, glasses, and deliberately unflattering, ill-fitting clothes. Grant was encouraged by director Mike Newell to mess up and trip over his lines, written in "convoluted syntax" as Grant describes them, in order to give Charles a stammering, nervous quality. Grant, who struggled with hay fever throughout filming, was unsure of Newell's direction and his own performance, which he thought was "atrocious"; on Newell he commented that: "He seemed to be giving direction against what I thought were the natural beats of the comedy. He was making a film with texture, grounding it, playing the truths rather than the gags". The film was shot mainly in London and the Home Counties, including Hampstead, Islington where the final moments take place on Highbury Terrace, Greenwich Hospital, Betchworth in Surrey, Amersham in Buckinghamshire, St Bartholomew-the-Great (wedding number four) and West Thurrock in Essex. Exterior shots of guests arriving for the funeral were filmed in Thurrock, Essex overlooking the River Thames with the backdrop of the Dartford River Crossing and at stately homes in Bedfordshire (Luton Hoo for wedding two's reception) and Hampshire. According to Hugh Grant, the initial screening of a rough-cut of "Four Weddings" went very badly. Throughout production, Gramercy Pictures, the U.S. distributor for the film, sent frequent transatlantic faxes objecting to the explicit language and sexual content, fearing the final product would not be suitable for American distribution or television airings. They particularly objected to the opening scene of the movie, in which Charles and Scarlett say the word "Fuck" over and over, after an initial screening of the movie in Salt Lake City led the conservative Mormon members of the city council to walk out. Accordingly, Mike Newell and the actors agreed to reshoot the scene with the British swear word "Bugger" to be used in the American version. The executives also objected to the title, believing "Four Weddings and a Funeral" would turn off male viewers from the film. In its place they suggested such titles as "True Love and Near Misses", "Loitering in Sacred Places", "Skulking Around", and "Rolling in the Aisles", none of which were accepted. The original score was composed by British composer Richard Rodney Bennett. The movie also featured a soundtrack of popular songs, including a cover version of The Troggs' "Love Is All Around" performed by Wet Wet Wet that remained at number 1 on the UK Singles Chart for fifteen weeks and was then the ninth (now twelfth) biggest selling single of all time in Britain. This song would later be adapted into "Christmas Is All Around" and sung by the character of Billy Mack in Richard Curtis' 2003 film "Love Actually", in which Grant also stars. The movie made its world premiere in February 1994 at the Sundance Film Festival in Salt Lake City, Utah. The box office receipts from the first five days of the film's general release in North America averaged $27,697 per screen, which so impressed the movie's distributor that it decided to spend lavishly on promotion, buying full-page newspaper ads and TV-spots totaling some $11 million, against a budget that had only amounted to $4.4 million. The movie also benefited from much free publicity because of Grant's reception in the United States, where he became an instant sex symbol and undertook a successful media tour promoting the film. Producer Duncan Kenworthy stated that "It was the most amazing luck that when Hugh went on the publicity trail he turned out to be incredibly funny, and very like the character of Charles. That doesn't ever happen." Hugh Grant's then-girlfriend Elizabeth Hurley also garnered much publicity for the film when she wore a black Versace safety-pin dress to the premiere in Leicester Square which became a sensation in the press. The film earned £27.8 million in the United Kingdom. Upon its North American limited release on 11 March 1994, "Four Weddings and a Funeral" opened with $138,486 in five theatres. But upon its wide release on 15 April 1994, the film topped the box office with $4,162,489. The film would continue to gross $52.7 million in North America with an additional $193 million internationally, earning $245.7 million worldwide. The film was very well received with critics, currently holding a 96% "Certified Fresh" approval on review aggregator Rotten Tomatoes based on 67 reviews, with an average rating of 7.64/10. The site's consensus states, "While frothy to a fault, "Four Weddings and a Funeral" features irresistibly breezy humor, and winsome performances from Hugh Grant and Andie MacDowell." Film critic Roger Ebert gave the film three-and-a-half stars out of four, calling it "delightful and sly", and directed with "light-hearted enchantment" by Newell. He praised Grant's performance, describing it as a kind of "endearing awkwardness". Metacritic gave the film a score of 81 based on 19 reviews, indicating "universal acclaim". Producer Duncan Kenworthy attributes much of the stimulus for "Four Weddings"' box office success to its first glowing review by Todd McCarthy in "Variety", who called it a "truly beguiling romantic comedy" which was "frequently hilarious without being sappily sentimental or tiresomely retrograde." The film had its detractors. Writing for the "Chicago Reader", Jonathan Rosenbaum called the film "generic" and "standard issue", stating that the audience shouldn't "expect to remember it ten minutes later". "Time" magazine writer Richard Corliss was less scathing, but agreed that it was forgettable, saying that people would "forget all about [the movie] by the time they leave the multiplex," even joking at the end of his review that he had forgotten the film's name. The character of Carrie was voted one of the most annoying film characters of all time in a British online poll. The film was voted the 27th greatest comedy film of all time by readers of "Total Film" in 2000. In 2004, the same magazine named it the 34th greatest British film of all time. It is number 96 on Bravo's "100 Funniest Movies". The Guardian, in a 20th anniversary retrospective of "Four Weddings", stated that "Its influence on the British film industry, on romantic-comedy writing, on the pop charts, on funeral readings, on "haircuts", was enormous." Hugh Grant commented of the experience of the film's phenomenal success and its effect on his career in 2016: "I was making "An Awfully Big Adventure" at the time that "Four Weddings" came out, with Mike Newell again, same director, even tinier budget, in Dublin. And we'd get back from brutal days on the set, very long and no money, and the fax machines...were coming out saying that now your film "Four Weddings" is #5 in America, now it's #3, now it's #1 and here's an offer Hugh, for "Captain Blood" and they'll pay you $1 million. It was completely surreal." It was reported in November 2017 that the streaming service Hulu was developing an eponymous anthology television series based upon the film, to be written and executive produced by Mindy Kaling and Matt Warburton, with Richard Curtis also serving as an executive producer. In October 2018, it was announced Jessica Williams, Nikesh Patel, Rebecca Rittenhouse, and John Reynolds had joined the cast. On 5 December 2018, it was announced that Richard Curtis had written "One Red Nose Day and a Wedding", a 25th anniversary reunion short film. The original film's director, Mike Newell, returned, along with the film's surviving cast, including Hugh Grant, Andie MacDowell, Kristin Scott Thomas, John Hannah, Rowan Atkinson, James Fleet, David Haig, Sophie Thompson, David Bower, Robin McCaffrey, Anna Chancellor, Rupert Vansittart, Simon Kunz, Sara Crowe and Timothy Walker. It was filmed from 13–14 December 2018 at St James' Church, Islington, London. The 14-minute film premiered in the UK during Red Nose Day on Friday 15 March 2019. It centered on the reunion of all the characters from the original film at the wedding of Charles and Carrie's daughter to Fiona's daughter. The involvement of additional cast members Lily James and Alicia Vikander was not announced until the day the film aired in the UK, because they played the young women getting married. The film aired in the US on their Red Nose Day on Thursday 23 May 2019.
https://en.wikipedia.org/wiki?curid=11223
Ferrari Ferrari (; ) is an Italian-Dutch luxury sports car manufacturer based in Maranello, Italy. Founded by Enzo Ferrari in 1939 out of Alfa Romeo's race division as "Auto Avio Costruzioni", the company built its first car in 1940. However, the company's inception as an auto manufacturer is usually recognized in 1947, when the first Ferrari-badged car was completed. In 2014 Ferrari was rated the world's most powerful brand by Brand Finance. In June 2018, the 1963 250 GTO became the most expensive car in history, setting an all-time record selling price of $70 million. Fiat S.p.A. acquired 50% of Ferrari in 1969 and expanded its stake to 90% in 1988. In October 2014 Fiat Chrysler Automobiles N.V. (FCA) announced its intentions to separate Ferrari S.p.A. from FCA; as of the announcement FCA owned 90% of Ferrari. The separation began in October 2015 with a restructuring that established Ferrari N.V. (a company incorporated in the Netherlands) as the new holding company of the Ferrari group and the subsequent sale by FCA of 10% of the shares in an IPO and concurrent listing of common shares on the New York Stock Exchange. Through the remaining steps of the separation, FCA's interest in Ferrari's business was distributed to shareholders of FCA, with 10% continuing to be owned by Piero Ferrari. The spin-off was completed on 3 January 2016. Throughout its history, the company has been noted for its continued participation in racing, especially in Formula One, where it is the oldest and most successful racing team, holding the most constructors championships (16) and having produced the highest number of drivers' championship wins (15). Ferrari road cars are generally seen as a symbol of speed, luxury and wealth. Ferrari cars are built at the 165,000 square-meter (16.5-hectare) Maranello factory. Enzo Ferrari was not initially interested in the idea of producing road cars when he formed Scuderia Ferrari in 1929, with headquarters in Modena. Scuderia Ferrari (pronounced ) literally means "Ferrari Stable" and is usually used to mean "Team Ferrari." Ferrari bought, prepared, and fielded Alfa Romeo racing cars for gentleman drivers, functioning as the racing division of Alfa Romeo. In 1933, Alfa Romeo withdrew its in-house racing team and Scuderia Ferrari took over as its works team: the "Scuderia" received Alfa's Grand Prix cars of the latest specifications and fielded many famous drivers such as Tazio Nuvolari and Achille Varzi. In 1938, Alfa Romeo brought its racing operation again in-house, forming "Alfa Corse" in Milan and hired Enzo Ferrari as manager of the new racing department; therefore the Scuderia Ferrari was disbanded. In September 1939, Ferrari left Alfa Romeo under the provision he would not use the Ferrari name in association with races or racing cars for at least four years. A few days later he founded "Auto Avio Costruzioni", headquartered in the facilities of the old "Scuderia Ferrari". The new company ostensibly produced machine tools and aircraft accessories. In 1940, Ferrari produced a race car – the Tipo 815, based on a Fiat platform. It was the first Ferrari car and debuted at the 1940 Mille Miglia, but due to World War II it saw little competition. In 1943, the Ferrari factory moved to Maranello, where it has remained ever since. The factory was bombed by the Allies and subsequently rebuilt including works for road car production. The first Ferrari-badged car was the 1947 125 S, powered by a 1.5 L V12 engine; Enzo Ferrari reluctantly built and sold his automobiles to fund Scuderia Ferrari. The "Scuderia Ferrari" name was resurrected to denote the factory racing cars and distinguish them from those fielded by customer teams. In 1960 the company was restructured as a public corporation under the name SEFAC S.p.A. (Società Esercizio Fabbriche Automobili e Corse). Early in 1969, Fiat took a 50% stake in Ferrari. An immediate result was an increase in available investment funds, and work started at once on a factory extension intended to transfer production from Fiat's Turin plant of the Ferrari engined Fiat Dino. New model investment further up in the Ferrari range also received a boost. In 1988, Enzo Ferrari oversaw the launch of the Ferrari F40, the last new Ferrari launched before his death later that year. In 1989, the company was renamed Ferrari S.p.A. From 2002 to 2004, Ferrari produced the Enzo, their fastest model at the time, which was introduced and named in honor of the company's founder, Enzo Ferrari. It was to be called the F60, continuing on from the F40 and F50, but Ferrari was so pleased with it, they called it the Enzo instead. It was initially offered to loyal and recurring customers, each of the 399 made (minus the 400th which was donated to the Vatican for charity) had a price tag of $650,000 apiece (equivalent to £400,900). On 15 September 2012, 964 Ferrari cars worth over $162 million (£99.95 million) attended the Ferrari Driving Days event at Silverstone Circuit and paraded round the Silverstone Circuit setting a world record. Ferrari's former CEO and Chairman, Luca di Montezemolo, resigned from the company after 23 years, who was succeeded by Amedeo Felisa and finally on 3 May 2016 Amedeo resigned and was succeeded by Sergio Marchionne, CEO and Chairman of Fiat Chrysler Automobiles, Ferrari's parent company. In July 2018, Marchionne was replaced by board member Louis Camilleri as CEO and by John Elkann as chairman. On 29 October 2014, the FCA group, resulting from the merger between manufacturers Fiat and Chrysler, announced the split of its luxury brand, Ferrari. The aim is to turn Ferrari into an independent brand which 10% of stake will be sold in an IPO in 2015. Ferrari officially priced its initial public offering at $52 a share after the market close on 20 October 2015. Since the company's beginnings, Ferrari has been involved in motorsport, competing in a range of categories including Formula One and sports car racing through its Scuderia Ferrari sporting division as well as supplying cars and engines to other teams and for one-make race series. 1940 AAC 815 was the first racing car to be designed by Enzo Ferrari, although it was not badged as a Ferrari model. "Scuderia" Ferrari has participated in several classes of motorsport, though it is currently only officially involved in Formula One. It is the only team to have competed in the Formula One World Championship continuously since its inception in 1950. José Froilán González gave the team its first F1 victory at the 1951 British Grand Prix. Alberto Ascari gave Ferrari its first Drivers Championship a year later. Ferrari is the oldest team in the championship, and the most successful: the team holds nearly every Formula One record. , the team's records include 15 World Drivers Championship titles, 16 World Constructors Championship titles, 221 Grand Prix victories, 6736.27 points, 679 podium finishes, 207 pole positions, and 230 fastest laps in 890 Grands Prix contested. Of the 19 tracks used in 2014, 8 have lap records set by the F2004, with a further 3 set by the F2003-GA, F2008 and F10. Ferrari drivers include: Tazio Nuvolari, José Froilán González, Juan Manuel Fangio, Alberto Ascari, Luigi Chinetti, Maurice Trintignant, Wolfgang von Trips, Phil Hill, Olivier Gendebien, Mike Hawthorn, Peter Collins, Giancarlo Baghetti, Ricardo Rodríguez, Chris Amon, John Surtees, Lorenzo Bandini, Ludovico Scarfiotti, Jacky Ickx, Mario Andretti, Clay Regazzoni, Niki Lauda, Carlos Reutemann, Jody Scheckter, Gilles Villeneuve, Didier Pironi, Patrick Tambay, René Arnoux, Michele Alboreto, Gerhard Berger, Nigel Mansell, Alain Prost, Jean Alesi, Michael Schumacher, Eddie Irvine, Rubens Barrichello, Felipe Massa, Kimi Räikkönen, Fernando Alonso, Sebastian Vettel, and Charles Leclerc. At the end of the 2006 season, the team courted controversy by continuing to allow Marlboro to sponsor them after they, along with the other F1 teams, made a promise to end sponsorship deals with tobacco manufacturers. A five-year deal was agreed and although this was not due to end until 2011, in April 2008 Marlboro dropped their on-car branding on Ferrari. In addition to Formula One, Ferrari also entered cars in sportscar racing, the two programs existing in parallel for many years. In 1949, Luigi Chinetti drove a 166 M to Ferrari's first win in motorsports, the 24 Hours of Le Mans. Ferrari went on to dominate the early years of the World Sportscar Championship which was created in 1953, winning the title seven out of its first nine years. When the championship format changed in 1962, Ferrari earned titles in at least one class each year through to 1965 and then again in 1967. Ferrari would win one final title, the 1972 World Championship of Makes before Enzo decided to leave sports car racing after 1973 and allow Scuderia Ferrari to concentrate solely on Formula One. During Ferrari's seasons of the World Sportscars Championship, they also gained more wins at the 24 Hours of Le Mans, with the factory team earning their first in . Another win would come in , followed by five consecutive wins from to . Luigi Chinetti's North American Racing Team (NART) would take Ferrari's final victory at Le Mans in . Although Scuderia Ferrari no longer participated in sports cars after 1973, they have occasionally built various successful sports cars for privateers. These include the BB 512 LM in the 1970s, the 333 SP which won the IMSA GT Championship in the 1990s, and currently the 458 GT2 and GT3 which are currently winning championships in their respective classes. Throughout its history, Ferrari has supplied racing cars to other entrants, aside from its own works Scuderia Ferrari team. In the 1950s and '60s, Ferrari supplied Formula One cars to a number of private entrants and other teams. One famous example was Tony Vandervell's team, which raced the Thinwall Special modified Ferraris before building their own Vanwall cars. The North American Racing Team's entries in the final three rounds of the 1969 season were the last occasions on which a team other than Scuderia Ferrari entered a World Championship Grand Prix with a Ferrari car. Ferrari supplied cars complete with V8 engines for the A1 Grand Prix series, from the 2008–09 season. The car was designed by Rory Byrne and is styled to resemble the 2004 Ferrari Formula one car. Ferrari currently runs a customer GT program for a racing version of its 458 and has done so for the 458's predecessors, dating back to the 355 in the late 1990s. Such private teams as the American Risi Competizione and Italian AF Corse teams have been very successful with Ferrari GT racers over the years. This car, made for endurance sportscar racing to compete against such racing versions of the Audi R8, McLaren MP4-12C, and BMW Z4 (E89) has proven to be successful, but not as successful as its predecessor, the F430. The Ferrari Challenge is a one-make racing series for the Ferrari 458. The FXX is not road legal and is therefore only used for track events. The first vehicle made with the Ferrari name was the 125 S. This was primarily a sports/racing model. In 1949, the 166 Inter was introduced marking the company's significant move into the grand touring road car market. The first 166 Inter was a four seat (2+2) berlinetta coupe with body work designed by Carrozzeria Touring Superleggera. Road cars quickly became the bulk of Ferrari sales. The early Ferrari cars typically featured bodywork designed and customised by independent coachbuilders such as Pininfarina, Scaglietti, Zagato, Vignale and Bertone. Starting in the early 2010s with the LaFerrari, the focus was shifted to what is now the standard, Ferrari relying on in-house design from the Centro Stile Ferrari. The original road cars were typically two-seat front-engined V12s. This platform served Ferrari very well through the 1950s and 1960s. In 1968 the Dino was introduced as the first two-seat rear mid-engined Ferrari. The Dino was produced primarily with a V6 engine, however, a V8 model was also developed. This rear mid-engine layout would go on to be used in many Ferraris of the 1980s, 1990s and to the present day. Current road cars typically use V8 or V12 engines, with V8 models making up well over half of the marque's total production. Historically, Ferrari has also produced flat 12 engines. For a time, Ferrari built 2+2 versions of its mid-engined V8 cars. Although they looked quite different from their 2-seat counterparts, both the GT4 and Mondial were closely related to the 308 GTB. The company has also produced several front-engined 2+2 cars, culminating in the current V12 model Lusso and V8 models Roma, Portofino and Lusso T. The California is credited with initiating the popular current model line of V8 front-engined 2+2 grand touring performance sports cars. Ferrari entered the mid-engined 12-cylinder fray with the Berlinetta Boxer in 1973. The later Testarossa (also mid-engined 12 cylinders) remains one of the most popular and famous Ferrari road cars of all time. In the 1950s and 1960s, clients often personalized their vehicles as they came straight from the factory. This philosophy added to the mystique of the brand. Every Ferrari that comes out of Maranello is built to an individual customer's specification. In this sense, each vehicle is a unique result of a specific client's desire. Ferrari formalized this concept with its earlier Carrozzeria Scaglietti programme. The options offered here were more typical such as racing seats, rearview cameras, and other special trim. In late 2011, Ferrari announced a significant update of this philosophy. The Tailor Made programme allows clients to work with designers in Maranello to make decisions at every step of the process. Through this program almost any trim, any exterior color or any interior material is possible. The program carries on the original tradition and emphasizes the idea of each car being unique. The 1984 288 GTO may be considered the first in the line of Ferrari supercars. This pedigree extends through the Enzo Ferrari to the LaFerrari. In February 2019, at the 89th Geneva International Motor Show, Ferrari revealed its latest mid-engine V8 supercar, the F8 Tributo. Ferrari SF90 Stradale is the first-ever Ferrari to feature PHEV (Plug-in Hybrid Electric Vehicle) architecture which sees the internal combustion engine integrated with three electric motors, two of which are independent and located on the front axle, with the third at the rear between the engine and the gearbox. Ferrari has produced a number of concept cars, such as the Mythos. While some of these were quite radical (such as the Modulo) and never intended for production, others such as the Mythos have shown styling elements which were later incorporated into production models. The most recent concept car to be produced by Ferrari themselves was the 2010 Millechili. A number of one-off special versions of Ferrari road cars have also been produced, commissioned to coachbuilders by wealthy owners. Recent examples include the P4/5 and the 412 Kappa. The Special Projects programme was launched in the late 2000s as Ferrari's ultimate in-house personalization service, enabling customers to own bespoke bodied one-offs based on modern Ferrari road cars. Engineering and design is done by Ferrari, sometimes in cooperation with external design houses like Pininfarina or Fioravanti, and the vehicles receive full homologation to be road legal. The first car to be completed under this programme was the 2008 SP1, commissioned by a Japanese business executive, the second was the P540 Superfast Aperta, commissioned by an American collector. The following is a list of Special Projects cars that have been made public: An F430 Spider that runs on ethanol was displayed at the 2008 Detroit Auto Show. At the 2010 Geneva Motor Show, Ferrari unveiled a hybrid version of their flagship 599. Called the "HY-KERS Concept", Ferrari's hybrid system adds more than 100 horsepower on top of the 599 Fiorano's 612 HP. Also in mid-2014, the flagship LaFerrari was put into production. From the beginning, the Ferrari naming convention consisted of a three-digit unitary displacement of an engine cylinder with an additional suffix representing the purpose of a vehicle. Therefore, Ferrari 125 S had 1.5 L (1496.77 cc) V12 engine with a unitary displacement of 124.73 cc; whilst S-suffix represented Sport. Other race cars also received names invoking particular races like Ferrari 166 MM for Mille Miglia. With the introduction of road-going models, suffix Inter was added, inspired by Scuderia Inter racing team of Igor Troubetzkoy. Popular at that time 166-series had 2.0 L (1995.02 cc) engines with 166.25 cc of unitary displacement and a very diverse 250-series had 3.0 L (2953.21 cc) of total displacement and 246.10 cc of unitary. Later series of road cars were renamed Europa and top of the line series: America and Superamerica. Until the early 1990s, Ferrari followed a three-number naming scheme based on engine displacement and a number of cylinders: Most Ferraris were also given designations referring to their body style. In general, the following conventions were used: This naming system can be confusing, as some entirely different vehicles used the same engine type and body style. Many Ferraris also had other names affixed (like Daytona) to identify them further. Many such names are actually not official factory names. The Daytona name commemorates Ferrari's triple success in the February 1967 24 Hours of Daytona with the 330 P4. Only in the 1973 Daytona 24 Hours, a 365 GTB/4 run by NART (who raced Ferraris in America) ran second, behind a Porsche 911. The various Dino models were named for Enzo's son, Dino Ferrari, and were marketed as Dinos by Ferrari and sold at Ferrari dealers—for all intents and purposes they are Ferraris. In the mid-1990s, Ferrari added the letter "F" to the beginning of all models (a practice abandoned after the F512 M and F355, but adopted again with the F430, but not with its successor, the Ferrari 458 ). The famous symbol of the Ferrari race team is the "Cavallino Rampante" ("prancing horse") black prancing stallion on a yellow shield, usually with the letters "S F" (for "Scuderia Ferrari"), with three stripes of green, white and red (the Italian national colors) at the top. The road cars have a rectangular badge on the hood (see picture at top of page), and, optionally, the shield-shaped race logo on the sides of both front wings, close to the door. On 17 June 1923, Enzo Ferrari won a race at the Savio track in Ravenna where he met Countess Paolina, mother of Count Francesco Baracca, an ace of the Italian air force and national hero of World War I, who used to paint a horse on the side of his planes. The Countess asked Enzo to use this horse on his cars, suggesting that it would bring him good luck. The original "prancing horse" on Baracca's airplane was painted in red on a white cloud-like shape, but Ferrari chose to have the horse in black (as it had been painted as a sign of grief on Baracca's squadron planes after the pilot was killed in action) and he added a canary yellow background as this is the color of the city of Modena, his birthplace. The Ferrari horse was, from the very beginning, markedly different from the Baracca horse in most details, the most noticeable being the tail that in the original Baracca version was pointing downward. Ferrari has used the "cavallino rampante" on official company stationery since 1929. Since the Spa 24 Hours of 9 July 1932, the "cavallino rampante" has been used on Alfa Romeos raced by Scuderia Ferrari. The motif of a prancing horse is old, it can be found on ancient coins. A similar black horse on a yellow shield is the Coat of Arms of the German city of Stuttgart, home of Mercedes-Benz and the design bureau of Porsche, both being main competitors of Alfa and Ferrari in the 1930s. The city's name derives from "Stutengarten", an ancient form of the German word "Gestüt", which translates into English as "stud farm" and into Italian as "scuderia". Porsche also includes the Stuttgart sign in its corporate logo, centred in the emblem of the state of Württemberg. Stuttgart's "Rössle" has both rear legs firmly planted on the soil, like Baracca's horse, but unlike Ferrari's "cavallino". Fabio Taglioni used the "cavallino rampante" on his Ducati motorbikes, as Taglioni was born at Lugo di Romagna like Baracca, and his father too was a military pilot during WWI (although not part of Baracca's squadron, as is sometimes mistakenly reported). As Ferrari's fame grew, Ducati abandoned the horse- perhaps the result of a private agreement between the two companies. The "cavallino rampante" is the visual symbol of Ferrari. Cavallino Magazine uses the name, but not the logo. Other companies use similar logos: Avanti, an Austrian company operating over 100 filling stations, uses a prancing horse logo which is nearly identical to Ferrari's, as does Iron Horse Bicycles and Norfolk Southern Railway. Since the 1920s, Italian race cars of Alfa Romeo, Maserati and later Ferrari and Abarth were (and often still are) painted in "race red" ("Rosso Corsa"). This was the customary national racing color of Italy, as recommended between the World Wars by the organizations that later would become the FIA. It refers to the nationality of the competing team, not that of the car manufacturer or driver. In that scheme, French-entered cars such as Bugatti were blue, German such as Auto Union and Mercedes white (since 1934 also bare sheet metal silver), and British green such as the mid-1960s Lotus and BRM, for instance. Ferrari won the 1964 World championship with John Surtees by competing for the last two races in North America with cars painted in the US-American race colors white and blue, as these were not entered by the Italian factory themselves, but by the U.S.-based North American Racing Team (NART) team. This was done as a protest concerning arguments between Ferrari and the Italian Racing Authorities regarding the homologation of a new mid-engined Ferrari race car. In 1963, Enzo Ferrari was approached by the Ford Motor Company about a possible buy out. Ford audited Ferrari's assets but legal negotiations and talks were unilaterally cut off by Ferrari when he realized that the deal offered by Ford would not enable him to stay at the helm of the company racing program. Henry Ford II consequently directed his racing division to negotiate with Lotus, Lola, and Cooper to build a car capable of beating Ferrari on the world endurance circuit, eventually resulting in the production of the Ford GT40 in 1964. As the Ford deal fell through, FIAT approached Ferrari with a more flexible proposal and purchased controlling interests in the company in 1969. Enzo Ferrari retained a 10% share, which is currently owned by his son Piero Lardi Ferrari. Ferrari has an internally managed merchandising line that licenses many products bearing the Ferrari brand, including eyewear, pens, pencils, electronic goods, perfume, cologne, clothing, high-tech bicycles, watches, cell phones, and laptop computers. Ferrari also runs a museum, the Museo Ferrari in Maranello, which displays road and race cars and other items from the company's history. In 1997 Ferrari launched a long term master planned effort to improve overall corporate efficiency, production and employee happiness. The program was called Formula Uomo and became a case study in social sustainability. It took over ten years to fully implement and included over €200 million (2008) in investment. Ferrari has had a long-standing relationship with Shell Oil. It is a technical partnership with Ferrari and Ducati to test as well as supply fuel and oils to the Formula One, MotoGP and World Superbike racing teams. For example, the Shell V-Power premium gasoline fuel has been developed with the many years of technical expertise between Shell and Ferrari. Ferrari has had agreements to supply Formula One engines to a number of other teams over the years, and currently supply Sauber F1 Team, and Haas F1 Team. As of the end of 2019, the total of Ferrari built and sold cars in their whole company history is 219,062. Approximately thirty Ferrari boutiques exist worldwide, with two owned by Ferrari and the rest operating as franchises. The stores sell branded clothes, accessories and racing memorabilia. Clothing includes upscale and lower-priced collections for men, women, and children. Some stores include race car simulation games for entertainment.
https://en.wikipedia.org/wiki?curid=11225
Freemasonry Freemasonry or Masonry consists of fraternal organisations that trace their origins to the local fraternities of stonemasons that from the end of the fourteenth century regulated the qualifications of stonemasons and their interaction with authorities and clients. Modern Freemasonry broadly consists of two main recognition groups. Freemasonry has been the subject of numerous conspiracy theories throughout the years. The basic, local organisational unit of Freemasonry is the Lodge. These private Lodges are usually supervised at the regional level (usually coterminous with either a state, province, or national border) by a Grand Lodge or Grand Orient. There is no international, worldwide Grand Lodge that supervises all of Freemasonry; each Grand Lodge is independent, and they do not necessarily recognise each other as being legitimate. The degrees of Freemasonry retain the three grades of medieval craft guilds, those of Apprentice, Journeyman or fellow (now called Fellowcraft), and Master Mason. The candidate of these three degrees is progressively taught the meanings of the symbols of Freemasonry, and entrusted with grips, signs and words to signify to other members that he has been so initiated. The degrees are part allegorical morality play and part lecture. Three degrees are offered by Craft (or Blue Lodge) Freemasonry, and members of any of these degrees are known as Freemasons or Masons. There are additional degrees, which vary with locality and jurisdiction, and are usually administered by their own bodies (separate from those who administer the Craft degrees). The Masonic lodge is the basic organisational unit of Freemasonry. The Lodge meets regularly to conduct the usual formal business of any small organisation (pay bills, organise social and charitable events, elect new members, etc.). In addition to business, the meeting may perform a ceremony to confer a Masonic degree or receive a lecture, which is usually on some aspect of Masonic history or ritual. At the conclusion of the meeting, the Lodge might adjourn for a formal dinner, or "festive board", sometimes involving toasting and song. The bulk of Masonic ritual consists of degree ceremonies. Candidates for Freemasonry are progressively "initiated" into Freemasonry, first in the degree of Entered Apprentice. Some time later, in a separate ceremony, they will be "passed" to the degree of Fellowcraft, and finally they will be "raised" to the degree of Master Mason. In all of these ceremonies, the candidate is first obligated, then entrusted with passwords, signs and grips (secret handshakes) peculiar to his new rank. Another ceremony is the annual installation of the Master and officers of the Lodge. In some jurisdictions "Installed Master" is valued as a separate rank, with its own secrets to distinguish its members. In other jurisdictions, the grade is not recognised, and no inner ceremony conveys new secrets during the installation of a new Master of the Lodge. Most Lodges have some sort of social calendar, allowing Masons and their partners to meet in a less ritualised environment. Often coupled with these events is the obligation placed on every Mason to contribute to charity. This occurs at both Lodge and Grand Lodge level. Masonic charities contribute to many fields, such as education or disaster relief. These private local Lodges form the backbone of Freemasonry, and a Freemason will necessarily have been initiated into one of these. There also exist specialist Lodges where Masons meet to celebrate events, such as sport or Masonic research. The rank of Master Mason also entitles a Freemason to explore Masonry further through other degrees, administered separately from the Craft, or "Blue Lodge" degrees described here, but having a similar format to their meetings. There is very little consistency in Freemasonry. Because each Masonic jurisdiction is independent, each sets its own procedures. The wording of the ritual, the number of officers present, the layout of the meeting room, etc. varies from jurisdiction to jurisdiction. The officers of the Lodge are elected or appointed annually. Every Masonic Lodge has a Master, two Wardens, a secretary and a treasurer. There is also a Tyler, or outer guard, who is always present outside the door of a working Lodge, to secure its privacy. Other offices vary between jurisdictions. Each Masonic Lodge exists and operates according to a set of ancient principles known as the "Landmarks of Freemasonry". These principles have thus far eluded any universally accepted definition. Candidates for Freemasonry will have met most active members of the Lodge they are joining before they are initiated. The process varies between jurisdictions, but the candidate will typically have been introduced by a friend at a Lodge social function, or at some form of open evening in the Lodge. In modern times, interested people often track down a local Lodge through the Internet. The onus is on candidates to ask to join; while candidates may be encouraged to ask, they are never invited. Once the initial inquiry is made, an interview usually follows to determine the candidate's suitability. If the candidate decides to proceed from here, the Lodge ballots on the application before he (or she, depending on the Masonic Jurisdiction) can be accepted. The absolute minimum requirement of any body of Freemasons is that the candidate must be free, and considered to be of good character. There is usually an age requirement, varying greatly between Grand Lodges, and (in some jurisdictions) capable of being overridden by a dispensation from the Grand Lodge. The underlying assumption is that the candidate should be a mature adult. Additionally, most Grand Lodges require the candidate to declare a belief in a Supreme Being. In a few cases, the candidate may be required to be of a specific religion. The form of Freemasonry most common in Scandinavia (known as the Swedish Rite), for example, accepts only Christians. At the other end of the spectrum, "Liberal" or Continental Freemasonry, exemplified by the Grand Orient de France, does not require a declaration of belief in any deity, and accepts atheists (a cause of discord with the rest of Freemasonry). During the ceremony of initiation, the candidate is expected to undertake or swear (usually on a volume of sacred text appropriate to his personal religious faith) to fulfil certain obligations as a Mason. In the course of three degrees, new masons will promise to keep the secrets of their degree from lower degrees and outsiders, and to support a fellow Mason in distress (as far as practicality and the law permit). There is instruction as to the duties of a Freemason, but on the whole, Freemasons are left to explore the craft in the manner they find most satisfying. Some will simply enjoy the dramatics or the management and administration of the lodge, others will further explore the ritual and symbolism of the craft, others will focus their involvement on their Lodge's social side, perhaps in association with other lodges, while still others will concentrate on the charitable functions of the lodge. Grand Lodges and Grand Orients are independent and sovereign bodies that govern Masonry in a given country, state, or geographical area (termed a "jurisdiction"). There is no single overarching governing body that presides over worldwide Freemasonry; connections between different jurisdictions depend solely on mutual recognition. Freemasonry, as it exists in various forms all over the world, has a membership estimated by the United Grand Lodge of England at around 6 million worldwide. The fraternity is administratively organised into independent Grand Lodges (or sometimes Grand Orients), each of which governs its own Masonic jurisdiction, which consists of subordinate (or "constituent") Lodges. The largest single jurisdiction, in terms of membership, is the United Grand Lodge of England (with a membership estimated at around a quarter million). The Grand Lodge of Scotland and Grand Lodge of Ireland (taken together) have approximately 150,000 members. In the United States, total membership is just under 2 million. Relations between Grand Lodges are determined by the concept of "Recognition". Each Grand Lodge maintains a list of other Grand Lodges that it recognises. When two Grand Lodges recognise and are in Masonic communication with each other, they are said to be "in amity", and the brethren of each may visit each other's Lodges and interact Masonically. When two Grand Lodges are not in amity, inter-visitation is not allowed. There are many reasons one Grand Lodge will withhold or withdraw recognition from another, but the two most common are "Exclusive Jurisdiction" and "Regularity". Exclusive Jurisdiction is a concept whereby normally only one Grand Lodge will be recognised in any geographical area. If two Grand Lodges claim jurisdiction over the same area, the other Grand Lodges will have to choose between them, and they may not all decide to recognise the same one. (In 1849, for example, the Grand Lodge of New York split into two rival factions, each claiming to be the legitimate Grand Lodge. Other Grand Lodges had to choose between them until the schism was healed.) Exclusive Jurisdiction can be waived when the two overlapping Grand Lodges are themselves in Amity and agree to share jurisdiction (for example, since the Grand Lodge of Connecticut is in Amity with the Prince Hall Grand Lodge of Connecticut, the principle of Exclusive Jurisdiction does not apply, and other Grand Lodges may recognise both, likewise the five distinct kinds of lodges in Germany have nominally united under one Grand Lodge, in order to obtain international recognition. Regularity is a concept based on adherence to Masonic Landmarks, the basic membership requirements, tenets and rituals of the craft. Each Grand Lodge sets its own definition of what these landmarks are, and thus what is Regular and what is Irregular (and the definitions do not necessarily agree between Grand Lodges). Essentially, every Grand Lodge will hold that "its" landmarks (its requirements, tenets and rituals) are Regular, and judge other Grand Lodges based on those. If the differences are significant, one Grand Lodge may declare the other "Irregular" and withdraw or withhold recognition. The most commonly shared rules for Recognition (based on Regularity) are those given by the United Grand Lodge of England in 1929: Blue Lodge (a term not used in the United Kingdom, which simply refers to the Craft) basic Freemasonry offers only three traditional degrees and, in most jurisdictions, the rank of past or installed master. Master Masons are also able to extend their Masonic experience by taking further degrees, in appendant or other bodies whether or not approved by their own Grand Lodge. The Ancient and Accepted Scottish Rite is a system of 33 degrees (including the three Blue Lodge degrees) administered by a local or national Supreme Council. This system is popular in North America, South America and in Continental Europe. In America, the York Rite, with a similar range, administers three orders of Masonry, namely the Royal Arch, Cryptic Masonry, and Knights Templar. In Britain, separate bodies administer each order. Freemasons are encouraged to join the Holy Royal Arch, which is linked to Mark Masonry in Scotland and Ireland, but completely separate in England. In England, the Royal Arch is closely associated with the Craft, automatically having many Grand Officers in common, including H.R.H the Duke of Kent as both Grand Master of the Craft and First Grand Principal of the Royal Arch. The English Knights Templar and Cryptic Masonry share the Mark Grand Lodge offices and staff at Mark Masons Hall. In the Nordic countries, the Swedish Rite is dominant; a variation of it is also used in parts of Germany. Freemasonry describes itself as a "'beautiful system of morality, veiled in allegory and illustrated by symbols". The symbolism is mainly, but not exclusively, drawn from the tools of stonemasons – the square and compasses, the level and plumb rule, the trowel, the rough and smooth ashlars, among others. Moral lessons are attributed to each of these tools, although the assignment is by no means consistent. The meaning of the symbolism is taught and explored through ritual, and in lectures and articles by individual masons who offer their personal insights and opinions. All Freemasons begin their journey in the "craft" by being progressively "initiated", "passed" and "raised" into the three degrees of Craft, or Blue Lodge Masonry. During these three rituals, the candidate is progressively taught the Masonic symbols, and entrusted with grips or tokens, signs and words to signify to other Masons which degrees he has taken. The dramatic allegorical ceremonies include explanatory lectures, and revolve around the construction of the Temple of Solomon, and the artistry and death of the chief architect, Hiram Abiff. The degrees are those of "Entered apprentice", "Fellowcraft" and "Master Mason". While many different versions of these rituals exist, with various lodge layouts and versions of the Hiramic legend, each version is recognisable to any Freemason from any jurisdiction. In some jurisdictions, the main themes of each degree are illustrated by tracing boards. These painted depictions of Masonic themes are exhibited in the lodge according to which degree is being worked, and are explained to the candidate to illustrate the legend and symbolism of each degree. The idea of Masonic brotherhood probably descends from a 16th-century legal definition of a "brother" as one who has taken an oath of mutual support to another. Accordingly, Masons swear at each degree to keep the contents of that degree secret, and to support and protect their brethren unless they have broken the law. In most Lodges the oath or obligation is taken on a "Volume of Sacred Law", whichever book of divine revelation is appropriate to the religious beliefs of the individual brother (usually the Bible in the Anglo-American tradition). In "Progressive" continental Freemasonry, books other than scripture are permissible, a cause of rupture between Grand Lodges. Since the middle of the 19th century, Masonic historians have sought the origins of the movement in a series of similar documents known as the Old Charges, dating from the Regius Poem in about 1425 to the beginning of the 18th century. Alluding to the membership of a lodge of operative masons, they relate it to a mythologised history of the craft, the duties of its grades, and the manner in which oaths of fidelity are to be taken on joining. The 15th century also sees the first evidence of ceremonial regalia. There is no clear mechanism by which these local trade organisations became today's Masonic Lodges. The earliest rituals and passwords known, from operative lodges around the turn of the 17th–18th centuries, show continuity with the rituals developed in the later 18th century by accepted or speculative Masons, as those members who did not practice the physical craft gradually came to be known. The minutes of the Lodge of Edinburgh (Mary's Chapel) No. 1 in Scotland show a continuity from an operative lodge in 1598 to a modern speculative Lodge. It is reputed to be the oldest Masonic Lodge in the world. Alternatively, Thomas De Quincey in his work titled "Rosicrucians and Freemasonry" put forward the theory which suggested that Freemasonry was possibly an outgrowth of Rosicrucianism. The theory had also been postulated in 1803 by German professor; J. G. Buhle. The first Grand Lodge, the Grand Lodge of London and Westminster (later called the Grand Lodge of England (GLE)), was founded on St John's Day, 24 June 1717, when four existing London Lodges met for a joint dinner. Many English Lodges joined the new regulatory body, which itself entered a period of self-publicity and expansion. However, many Lodges could not endorse changes which some Lodges of the GLE made to the ritual (they came to be known as the Moderns), and a few of these formed a rival Grand Lodge on 17 July 1751, which they called the "Antient Grand Lodge of England." These two Grand Lodges vied for supremacy until the Moderns promised to return to the ancient ritual. They united on 27 December 1813 to form the United Grand Lodge of England (UGLE). The Grand Lodge of Ireland and the Grand Lodge of Scotland were formed in 1725 and 1736 respectively, although neither persuaded all of the existing lodges in their countries to join for many years. The earliest known American lodges were in Pennsylvania. The Collector for the port of Pennsylvania, John Moore, wrote of attending lodges there in 1715, two years before the putative formation of the first Grand Lodge in London. The Premier Grand Lodge of England appointed a Provincial Grand Master for North America in 1731, based in Pennsylvania, leading to the creation of the Grand Lodge of Pennsylvania. In present-day Canada, Erasmus James Philipps became a Freemason while working on a commission to resolve boundaries in New England and, in 1739, became provincial grand master for Nova Scotia; Philipps founded the first Masonic lodge in Canada at Annapolis Royal, Nova Scotia. Other lodges in the colony of Pennsylvania obtained authorisations from the later Antient Grand Lodge of England, the Grand Lodge of Scotland, and the Grand Lodge of Ireland, which was particularly well represented in the travelling lodges of the British Army. Many lodges came into existence with no warrant from any Grand Lodge, applying and paying for their authorisation only after they were confident of their own survival. After the American Revolution, independent U.S. Grand Lodges developed within each state. Some thought was briefly given to organising an overarching "Grand Lodge of the United States," with George Washington (who was a member of a Virginian lodge) as the first Grand Master, but the idea was short-lived. The various state Grand Lodges did not wish to diminish their own authority by agreeing to such a body. Freemasonry was imported to Jamaica by British immigrants who colonized the island for over 300 years. In 1908, there were eleven recorded Masonic Lodges which included three Grand Lodges, two Craft Lodges, and two Rose Croix Chapters. During slavery, the Lodges were open to all "freeborn" men. According to the Jamaican 1834 census, that potentially included 5,000 free black men and 40,000 free coloureds (mixed-race). After the full abolition of slavery in 1838, the Lodges were open to all Jamaican men of any race. Jamaica also kept close relationships with Masons from other countries. Jamaican Freemasonry historian Jackie Ranston, noted that: On 25 May 2017, Masons around the world celebrated the 300th anniversary of the fraternity. Jamaica hosted one of the regional gatherings for this celebration. Prince Hall Freemasonry exists because of the refusal of early American lodges to admit African Americans. In 1775, an African American named Prince Hall, along with 14 other African-American men, was initiated into a British military lodge with a warrant from the Grand Lodge of Ireland, having failed to obtain admission from the other lodges in Boston. When the British military Lodge left North America after the end of the Revolution, those 15 men were given the authority to meet as a Lodge, but not to initiate Masons. In 1784, these individuals obtained a Warrant from the Premier Grand Lodge of England (GLE) and formed African Lodge, Number 459. When the UGLE was formed in 1813, all U.S.-based Lodges were stricken from their rolls – due largely to the War of 1812. Thus, separated from both UGLE and any concordantly recognised U.S. Grand Lodge, African Lodge retitled itself as the African Lodge, Number 1 – and became a "de facto" Grand Lodge. (This lodge is not to be confused with the various Grand Lodges in Africa.) As with the rest of U.S. Freemasonry, Prince Hall Freemasonry soon grew and organised on a Grand Lodge system for each state. Widespread racial segregation in 19th- and early 20th-century North America made it difficult for African Americans to join Lodges outside of Prince Hall jurisdictions – and impossible for inter-jurisdiction recognition between the parallel U.S. Masonic authorities. By the 1980s, such discrimination was a thing of the past. Today most U.S. Grand Lodges recognise their Prince Hall counterparts, and the authorities of both traditions are working towards full recognition. The United Grand Lodge of England has no problem with recognising Prince Hall Grand Lodges. While celebrating their heritage as lodges of black Americans, Prince Hall is open to all men regardless of race or religion. English Freemasonry spread to France in the 1720s, first as lodges of expatriates and exiled Jacobites, and then as distinctively French lodges which still follow the ritual of the Moderns. From France and England, Freemasonry spread to most of Continental Europe during the course of the 18th century. The Grande Loge de France formed under the Grand Mastership of the Duke of Clermont, who exercised only nominal authority. His successor, the Duke of Orléans, reconstituted the central body as the Grand Orient de France in 1773. Briefly eclipsed during the French Revolution, French Freemasonry continued to grow in the next century, at first under the leadership of Alexandre Francois Auguste de Grasse, Comte de Grassy-Tilly. A career Army officer, he had lived with his family in Charleston, South Carolina from 1793 to the early 1800s, after leaving Saint-Domingue (now Haiti) during the years of the Haitian Revolution. The ritual form on which the Grand Orient of France was based was abolished in England in the events leading to the formation of the United Grand Lodge of England in 1813. However the two jurisdictions continued in amity (mutual recognition) until events of the 1860s and 1870s drove a seemingly permanent wedge between them. In 1868 the "Supreme Council of the Ancient and Accepted Scottish Rite of the State of Louisiana" appeared in the jurisdiction of the Grand Lodge of Louisiana, recognised by the Grand Orient de France, but regarded by the older body as an invasion of their jurisdiction. The new Scottish Rite body admitted blacks. The resolution of the Grand Orient the following year that neither colour, race, nor religion could disqualify a man from Masonry prompted the Grand Lodge to withdraw recognition, and it persuaded other American Grand Lodges to do the same. A dispute during the Lausanne Congress of Supreme Councils of 1875 prompted the Grand Orient de France to commission a report by a Protestant pastor which concluded that, as Freemasonry was not a religion, it should not require a religious belief. The new constitutions read, "Its principles are absolute liberty of conscience and human solidarity", the existence of God and the immortality of the soul being struck out. It is possible that the immediate objections of the United Grand Lodge of England were at least partly motivated by the political tension between France and Britain at the time. The result was the withdrawal of recognition of the Grand Orient of France by the United Grand Lodge of England, a situation that continues today. Not all French lodges agreed with the new wording. In 1894, lodges favouring the compulsory recognition of the Great Architect of the Universe formed the Grande Loge de France. In 1913, the United Grand Lodge of England recognised a new Grand Lodge of Regular Freemasons, a Grand Lodge that follows a similar rite to Anglo-American Freemasonry with a mandatory belief in a deity. There are now three strands of Freemasonry in France, which extend into the rest of Continental Europe:- The term Continental Freemasonry was used in Mackey's 1873 "Encyclopedia of Freemasonry" to "designate the Lodges on the Continent of Europe which retain many usages which have either been abandoned by, or never were observed in, the Lodges of England, Ireland, and Scotland, as well as the United States of America". Today, it is frequently used to refer to only the Liberal jurisdictions typified by the Grand Orient de France. The majority of Freemasonry considers the Liberal (Continental) strand to be Irregular, and thus withhold recognition. For the Continental lodges, however, having a different approach to Freemasonry was not a reason for severing masonic ties. In 1961, an umbrella organisation, Centre de Liaison et d'Information des Puissances maçonniques Signataires de l'Appel de Strasbourg (CLIPSAS) was set up, which today provides a forum for most of these Grand Lodges and Grand Orients worldwide. Included in the list of over 70 Grand Lodges and Grand Orients are representatives of all three of the above categories, including mixed and women's organisations. The United Grand Lodge of England does not communicate with any of these jurisdictions, and expects its allies to follow suit. This creates the distinction between Anglo-American and Continental Freemasonry. In the early 20th century Freemasonry was an influential semi-secret force in Italian politics with a strong presence among professionals and the middle class across Italy, as well as among the leadership of the parliament, public administration, and the army. The two main organisations were the Grand Orient and the Grand Lodge of Italy. They had 25,000 members in 500 or more lodges. Freemasons took on the challenge of mobilizing the press, public opinion and the leading political parties in support of Italy's joining the Allies of the First World War in 1914–1915. Traditionally, they promoted Italian nationalism focused on unification, and undermining the power of the Catholic Church. In 1914-15 they dropped the traditional pacifistic rhetoric and used instead the powerful language of Italian nationalism. Freemasonry had always promoted cosmopolitan universal values, and by 1917 onwards they demanded a League of Nations to promote a new post-war universal order based upon the peaceful coexistence of independent and democratic nations. The status of women in the old guilds and corporations of medieval masons remains uncertain. The principle of "femme sole" allowed a widow to continue the trade of her husband, but its application had wide local variations, such as full membership of a trade body or limited trade by deputation or approved members of that body. In masonry, the small available evidence points to the less empowered end of the scale. At the dawn of the Grand Lodge era, during the 1720s, James Anderson composed the first printed constitutions for Freemasons, the basis for most subsequent constitutions, which specifically excluded women from Freemasonry. As Freemasonry spread, continental masons began to include their ladies in Lodges of Adoption, which worked three degrees with the same names as the men's but different content. The French officially abandoned the experiment in the early 19th century. Later organisations with a similar aim emerged in the United States, but distinguished the names of the degrees from those of male masonry. Maria Deraismes was initiated into Freemasonry in 1882, then resigned to allow her lodge to rejoin their Grand Lodge. Having failed to achieve acceptance from any masonic governing body, she and Georges Martin started a mixed masonic lodge that worked masonic ritual. Annie Besant spread the phenomenon to the English-speaking world. Disagreements over ritual led to the formation of exclusively female bodies of Freemasons in England, which spread to other countries. Meanwhile, the French had re-invented Adoption as an all-female lodge in 1901, only to cast it aside again in 1935. The lodges, however, continued to meet, which gave rise, in 1959, to a body of women practising continental Freemasonry. In general, Continental Freemasonry is sympathetic to Freemasonry amongst women, dating from the 1890s when French lodges assisted the emergent co-masonic movement by promoting enough of their members to the 33rd degree of the Ancient and Accepted Scottish Rite to allow them, in 1899, to form their own grand council, recognised by the other Continental Grand Councils of that Rite. The United Grand Lodge of England issued a statement in 1999 recognising the two women's grand lodges there to be regular in all but the participants. While they were not, therefore, recognised as regular, they were part of Freemasonry "in general". The attitude of most regular Anglo-American grand lodges remains that women Freemasons are not legitimate Masons. In 2018 guidance was released by the United Grand Lodge of England stating that, in regard to transgender women, "A Freemason who after initiation ceases to be a man does not cease to be a Freemason". The guidance also states that transgender men are allowed to apply to become Freemasons. "Anti-Masonry" (alternatively called "Anti-Freemasonry") has been defined as "opposition to Freemasonry", but there is no homogeneous anti-Masonic movement. Anti-Masonry consists of widely differing criticisms from diverse (and often incompatible) groups who are hostile to Freemasonry in some form. Critics have included religious groups, political groups, and conspiracy theorists, in particular, those espousing Masonic conspiracy theories or the Judeo-Masonic conspiracy theory. Certain prominent Anti-Masons, such as Nesta Helen Webster (1876 – 1960), have exclusively criticized "Continental Masonry" while considering "Regular Masonry" an honorable association. There have been many disclosures and exposés dating as far back as the 18th century. These often lack context, may be outdated for various reasons, or could be outright hoaxes on the part of the author, as in the case of the Taxil hoax. These hoaxes and exposés have often become the basis for criticism of Masonry, often religious or political in nature or are based on suspicion of corrupt conspiracy of some form. The political opposition that arose after the American "Morgan Affair" in 1826 gave rise to the term "Anti-Masonry", which is still in use in America today, both by Masons in referring to their critics and as a self-descriptor by the critics themselves. Freemasonry has attracted criticism from theocratic states and organised religions for supposed competition with religion, or supposed heterodoxy within the fraternity itself and has long been the target of conspiracy theories, which assert Freemasonry to be an occult and evil power. Although members of various faiths cite objections, certain Christian denominations have had high-profile negative attitudes to Masonry, banning or discouraging their members from being Freemasons. The denomination with the longest history of objection to Freemasonry is the Catholic Church. The objections raised by the Catholic Church are based on the allegation that Masonry teaches a naturalistic deistic religion which is in conflict with Church doctrine. A number of Papal pronouncements have been issued against Freemasonry. The first was Pope Clement XII's "In eminenti apostolatus," 28 April 1738; the most recent was Pope Leo XIII's "Ab apostolici," 15 October 1890. The "1917 Code of Canon Law" explicitly declared that joining Freemasonry entailed automatic excommunication, and banned books favouring Freemasonry. In 1983, the Church issued a new code of canon law. Unlike its predecessor, the "1983 Code of Canon Law" did not explicitly name Masonic orders among the secret societies it condemns. It states: "A person who joins an association which plots against the Church is to be punished with a just penalty; one who promotes or takes office in such an association is to be punished with an interdict." This named omission of Masonic orders caused both Catholics and Freemasons to believe that the ban on Catholics becoming Freemasons may have been lifted, especially after the perceived liberalisation of Vatican II. However, the matter was clarified when Cardinal Joseph Ratzinger (later Pope Benedict XVI), as the Prefect of the Congregation for the Doctrine of the Faith, issued a Declaration on Masonic Associations, which states: "... the Church's negative judgment in regard to Masonic association remains unchanged since their principles have always been considered irreconcilable with the doctrine of the Church and therefore membership in them remains forbidden. The faithful who enrol in Masonic associations are in a state of grave sin and may not receive Holy Communion." For its part, Freemasonry has never objected to Catholics joining their fraternity. Those Grand Lodges in amity with UGLE deny the Church's claims. The UGLE now states that "Freemasonry does not seek to replace a Mason's religion or provide a substitute for it." In contrast to Catholic allegations of rationalism and naturalism, Protestant objections are more likely to be based on allegations of mysticism, occultism, and even Satanism. Masonic scholar Albert Pike is often quoted (in some cases misquoted) by Protestant anti-Masons as an authority for the position of Masonry on these issues. However, Pike, although undoubtedly learned, was not a spokesman for Freemasonry and was also controversial among Freemasons in general. His writings represented his personal opinion only, and furthermore an opinion grounded in the attitudes and understandings of late 19th century Southern Freemasonry of the US. Notably, his book carries in the preface a form of disclaimer from his own Grand Lodge. No one voice has ever spoken for the whole of Freemasonry. Free Methodist Church founder B.T. Roberts was a vocal opponent of Freemasonry in the mid 19th century. Roberts opposed the society on moral grounds and stated, "The god of the lodge is not the God of the Bible." Roberts believed Freemasonry was a "mystery" or "alternate" religion and encouraged his church not to support ministers who were Freemasons. Freedom from secret societies is one of the "frees" upon which the Free Methodist Church was founded. Since the founding of Freemasonry, many Bishops of the Church of England have been Freemasons, such as Archbishop Geoffrey Fisher. In the past, few members of the Church of England would have seen any incongruity in concurrently adhering to Anglican Christianity and practising Freemasonry. In recent decades, however, reservations about Freemasonry have increased within Anglicanism, perhaps due to the increasing prominence of the evangelical wing of the church. The former Archbishop of Canterbury, Dr Rowan Williams, appeared to harbour some reservations about Masonic ritual, whilst being anxious to avoid causing offence to Freemasons inside and outside the Church of England. In 2003 he felt it necessary to apologise to British Freemasons after he said that their beliefs were incompatible with Christianity and that he had barred the appointment of Freemasons to senior posts in his diocese when he was Bishop of Monmouth. In 1933, the Orthodox Church of Greece officially declared that being a Freemason constitutes an act of apostasy and thus, until he repents, the person involved with Freemasonry cannot partake of the Eucharist. This has been generally affirmed throughout the whole Eastern Orthodox Church. The Orthodox critique of Freemasonry agrees with both the Catholic and Protestant versions: "Freemasonry cannot be at all compatible with Christianity as far as it is a secret organisation, acting and teaching in mystery and secret and deifying rationalism." Regular Freemasonry has traditionally not responded to these claims, beyond the often repeated statement that those Grand Lodges in amity with UGLE explicitly adhere to the principle that "Freemasonry is not a religion, nor a substitute for religion. There is no separate 'Masonic deity,' and there is no separate proper name for a deity in Freemasonry." Christian men, who were discouraged from joining the Freemasons by their Churches or who wanted a more religiocentric society, joined similar fraternal organisations, such as the Knights of Columbus for Catholic Christians, and the Loyal Orange Institution for Protestant Christians, although these fraternal organisations have been "organized in part on the style of and use many symbols of Freemasonry". There are some elements of Freemasonry within the temple rituals of Mormonism. Many Islamic anti-Masonic arguments are closely tied to both antisemitism and Anti-Zionism, though other criticisms are made such as linking Freemasonry to Al-Masih ad-Dajjal (the false Messiah in Islamic Scripture). Some Muslim anti-Masons argue that Freemasonry promotes the interests of the Jews around the world and that one of its aims is to destroy the Al-Aqsa Mosque in order to rebuild the Temple of Solomon in Jerusalem. In article 28 of its Covenant, Hamas states that Freemasonry, Rotary, and other similar groups "work in the interest of Zionism and according to its instructions ..." Many countries with a majority Muslim population do not allow Masonic establishments within their borders. However, countries such as Turkey and Morocco have established Grand Lodges, while in countries such as Malaysia and Lebanon there are District Grand Lodges operating under a warrant from an established Grand Lodge. In Pakistan in 1972, Zulfiqar Ali Bhutto, then Prime Minister of Pakistan, placed a ban on Freemasonry. Lodge buildings were confiscated by the government. Masonic lodges existed in Iraq as early as 1917, when the first lodge under the United Grand Lodge of England (UGLE) was opened. Nine lodges under UGLE existed by the 1950s, and a Scottish lodge was formed in 1923. However, the position changed following the revolution, and all lodges were forced to close in 1965. This position was later reinforced under Saddam Hussein; the death penalty was "prescribed" for those who "promote or acclaim Zionist principles, including freemasonry, or who associate [themselves] with Zionist organisations." In 1799, English Freemasonry almost came to a halt due to Parliamentary proclamation. In the wake of the French Revolution, the Unlawful Societies Act banned any meetings of groups that required their members to take an oath or obligation. The Grand Masters of both the Moderns and the Antients Grand Lodges called on Prime Minister William Pitt (who was not a Freemason) and explained to him that Freemasonry was a supporter of the law and lawfully constituted authority and was much involved in charitable work. As a result, Freemasonry was specifically exempted from the terms of the Act, provided that each private lodge's Secretary placed with the local "Clerk of the Peace" a list of the members of his lodge once a year. This continued until 1967, when the obligation of the provision was rescinded by Parliament. Freemasonry in the United States faced political pressure following the 1826 kidnapping of William Morgan by Freemasons and his subsequent disappearance. Reports of the "Morgan Affair", together with opposition to Jacksonian democracy (Andrew Jackson was a prominent Mason), helped fuel an Anti-Masonic movement. The short-lived Anti-Masonic Party was formed, which fielded candidates for the presidential elections of 1828 and 1832. In Italy, Freemasonry has become linked to a scandal concerning the Propaganda Due lodge (a.k.a. P2). This lodge was chartered by the Grande Oriente d'Italia in 1877, as a lodge for visiting Masons unable to attend their own lodges. Under Licio Gelli's leadership, in the late 1970s, P2 became involved in the financial scandals that nearly bankrupted the Vatican Bank. However, by this time the lodge was operating independently and irregularly, as the Grand Orient had revoked its charter and expelled Gelli in 1976. Conspiracy theorists have long associated Freemasonry with the New World Order and the Illuminati, and state that Freemasonry as an organisation is either bent on world domination or already secretly in control of world politics. Historically Freemasonry has attracted criticism—and suppression—from both the politically far right (e.g., Nazi Germany) and the far left (e.g. the former Communist states in Eastern Europe). Freemasonry is viewed with distrust even in some modern democracies. In the UK, Masons working in the justice system, such as judges and police officers, were from 1999 to 2009 required to disclose their membership. While a parliamentary inquiry found that there has been no evidence of wrongdoing, the government believed that Masons' potential loyalties to support fellow Masons, should be transparent to the public. The policy of requiring a declaration of masonic membership by applicants for judicial office (judges and magistrates) was ended in 2009 by Justice Secretary Jack Straw (who had initiated the requirement in the 1990s). Straw stated that the rule was considered disproportionate, since no impropriety or malpractice had been shown as a result of judges being Freemasons. Freemasonry is both successful and controversial in France. As of the early 21st century, membership is rising, but reporting in the popular media is often negative. In some countries anti-Masonry is often related to antisemitism and anti-Zionism. For example, in 1980 the Iraqi legal and penal code was changed by Saddam Hussein's ruling Ba'ath Party, making it a felony to "promote or acclaim Zionist principles, including Freemasonry, or who associate [themselves] with Zionist organisations". Professor Andrew Prescott of the University of Sheffield writes: "Since at least the time of the Protocols of the Elders of Zion, antisemitism has gone hand in hand with anti-masonry, so it is not surprising that allegations that 11 September was a Zionist plot have been accompanied by suggestions that the attacks were inspired by a masonic world order". The preserved records of the "Reichssicherheitshauptamt" (the Reich Security Main Office) show the persecution of Freemasons during the Holocaust. RSHA Amt VII (Written Records) was overseen by Professor Franz Six and was responsible for "ideological" tasks, by which was meant the creation of antisemitic and anti-Masonic propaganda. While the number of victims is not accurately known, historians estimate that between 80,000 and 200,000 Freemasons were killed under the Nazi regime. Masonic concentration camp inmates were classified as political prisoners and wore an inverted red triangle. Hitler believed Freemasons had succumbed to Jews conspiring against Germany. The small blue forget-me-not flower was first used by the Grand Lodge "Zur Sonne", in 1926, as a Masonic emblem at the annual convention in Bremen, Germany. In 1938 a forget-me-not badge—made by the same factory as the Masonic badge—was chosen for the annual Nazi Party "Winterhilfswerk", the annual charity drive of the National Socialist People's Welfare, the welfare branch of the Nazi party. This coincidence enabled Freemasons to wear the forget-me-not badge as a secret sign of membership. After World War II, the forget-me-not flower was used again as a Masonic emblem in 1948 at the first Annual Convention of the United Grand Lodges of Germany in 1948. The badge is now worn in the coat lapel by Freemasons around the world to remember all who suffered in the name of Freemasonry, especially those during the Nazi era.
https://en.wikipedia.org/wiki?curid=11227
Fulham F.C. Fulham Football Club is a professional association football club based in Fulham, London, England. They compete in the EFL Championship, the second tier of English football. Founded in 1879, they are the oldest football club from London to play in the Football League. The club has spent 26 seasons in English football's top division, the majority of these in two spells during the 1960s and 2000s. The latter spell was associated with former chairman Mohamed Al-Fayed, after the club had climbed up from the fourth tier in the 1990s. Fulham have reached two major finals: in 1975, as a Second Division team, they contested the FA Cup Final for the only time in their history, losing 2–0 to West Ham United, and in 2010 they reached the final of the UEFA Europa League, which they contested with Atlético Madrid in Hamburg, losing 2–1 after extra time. Fulham's main rivalries are with fellow West London clubs Chelsea, Queens Park Rangers and Brentford. Fulham were formed in 1879 as Fulham St Andrew's Church Sunday School F.C., founded by worshipers (mostly adept at cricket) at the Church of England on Star Road, West Kensington (St Andrew's, Fulham Fields). Fulham's mother church still stands today with a plaque commemorating the team's foundation. They won the West London Amateur Cup in 1887 and, having shortened the name from Fulham St Andrews to its present form in December 1888, they then won the West London League in 1893 at the first attempt. One of the club's first ever kits was half red, half white shirts with white shorts worn in the 1886–87 season. Fulham started playing at their current ground at Craven Cottage in 1896, their first game against now defunct rivals Minerva. Fulham are one of the oldest established clubs in southern England currently playing professional football, though there are many non-league sides like Kent side Cray Wanderers who are several decades older. The club gained professional status on 12 December 1898, the same year that they were admitted into the Southern League's Second Division. They were the second club from London to turn professional, following Arsenal, then named Royal Arsenal 1891. They adopted a red and white kit during the 1896-97 season. In 1902–03, the club won promotion from this division, entering the Southern League First Division. The club's first recorded all-white club kit came in 1903, and ever since then the club has been playing in all-white shirts and black shorts, with socks going through various evolutions of black and/or white, but are now normally white-only. The club won the Southern League twice, in 1905–06 and 1906–07. Fulham joined The Football League after the second of their Southern League triumphs. The club's first league game, playing in the Second Division's 1907–08 season, saw them lose 1–0 at home to Hull City in September 1907. The first win came a few days later at Derby County's Baseball Ground by a score line of 1–0. Fulham finished the season three points short of promotion in fourth place. The club progressed all the way to the semi-final of that season's FA Cup, a run that included an 8–3 away win at Luton Town. In the semi-final, however, they were heavily beaten, 6–0, by Newcastle United. This is still a record loss for an FA Cup semi-final game. Two years later, the club won the London Challenge Cup in the 1909–10 season. Fulham's first season in Division Two turned out to be the highest that the club would finish for 21 years, until in 1927–28 when the club were relegated to the 3rd Division South, created in 1920. Hussein Hegazi, an Egyptian forward, was one of the first non-British players to appear in The Football League, though he only played one game for Fulham in 1911, marked with a goal, afterwards playing for non-league Dulwich Hamlet. During this period, businessman and politician Henry Norris was the club chairman and curiously he had an indirect role in the foundation of Fulham's local rivals Chelsea. When he rejected an offer from businessman Gus Mears to move Fulham to land where the present-day Chelsea stadium Stamford Bridge is situated, Mears decided to create his own team to occupy the ground. In 1910, Norris started to combine his role at Fulham with the chairmanship of Arsenal. Fulham became the first British team to sell hot dogs at their ground in 1926. Fulham had several high-profile international players during the 1920s, including Len Oliver and Albert Barrett. After finishing fifth, seventh and ninth (out of 22 teams) in their first three seasons in the Third Division South, Fulham won the division in the 1931–32 season. In doing so they beat Torquay United 10–2, won 24 out of 42 games and scored 111 goals, thus being promoted back to the Second Division. The next season they missed out on a second consecutive promotion, finishing third behind Tottenham Hotspur and Stoke City. A mixed bag of league performances followed, although the club also reached another FA Cup semi-final during the 1935–36 season. Fulham were also to draw with Austria in 1936 before Anschluss. On 8 October 1938, Craven Cottage saw its all-time highest attendance at a match against Millwall, with a crowd of 49,335 watching the game. League and cup football were severely disrupted by the outbreak of World War II in 1939, with the Football League split into regional divisions temporarily, with a national Football League War Cup and a London War Cup up for grabs. Craven Cottage was used like many grounds for fitness and training of the army youth reserves. Post-war, a full league programme was only restored for 1946–47. In the third season of what is now considered the modern era of football, Fulham finished top of the Second Division, with a win-loss-draw record of 24–9–9 (identical to that which won them the Third Division South 17 years previously). John Fox Watson made a pioneering transfer to Real Madrid in 1948, becoming one of the first players from the United Kingdom to sign for a high-profile side abroad. Promotion to the top tier of English football saw the club perform poorly, finishing 17th in their first year and 18th in their second. In only their third season of First Division football, Fulham finished rock bottom of the 22-team league in the 1951–52 season, winning only eight of 42 games. On 20 May 1951, Fulham played one of their first ever games in North America in an exhibition match against Celtic at Delorimier Stadium in Montreal in front of 29,000 spectators. Possibly the single most influential character in Fulham's history is Johnny Haynes. "Mr. Fulham" or "The Maestro," as Haynes later came to be known, signed for The Cottagers as a schoolboy in 1950, making his first team debut on Boxing Day against Southampton at Craven Cottage in the 1951/52 relegation season. Haynes played for another 18 years, notching 657 appearances (along with many other club records too), his last appearance for Fulham coming on 17 January 1970. He is often considered as the greatest player in Fulham history, and never played for another team in Britain. He gained 56 caps for England (22 as captain), with many being earned while playing for Fulham in the Second Division. Haynes was injured in a car accident in Blackpool in 1962, but by his own admissions never regained the fitness or form to play for England again, missing out on England's victory in the FIFA World Cup 1966 for which he would have stood a chance of being selected. The Stevenage Road Stand was renamed in his honour after his death in a car crash in 2005. Fulham reached the 1957-58 FA Cup semi-finals, the best cup run of Haynes' career and nearest he came to a major trophy win playing in England. They were eliminated in a replay by the remnants of Manchester United's Busby Babes team that had been decimated in the Munich air disaster the month before. United were the first top division team Fulham played in that cup run. Fulham won promotion back to the First Division in the following season by finishing second to Sheffield Wednesday. Also joining Fulham in 1958 was Graham Leggat, who went on to score 134 goals in 277 appearances, (making him the club's fifth all-time top scorer). In the 1959–60 season, they achieved tenth position in the First Division, which until finishing ninth in the 2003–04 season was their highest-ever league position. This accompanied another appearance in the last four of the FA Cup in 1962. By this time, the club were regularly playing in front of 30,000 plus crowds at Craven Cottage, despite struggling in the League. The club experienced several close escapes from relegation, none more so than in 1965–66. On the morning of 26 February 1966, Fulham were bottom with just 15 points from 29 matches. The last 13 games saw Fulham win nine and draw two to reach safety. Eventually, however, the club suffered relegation in the 1967–68 season, having won just ten out of their 42 games. Even that, however, was not as catastrophic as the calamity of next season. Winning only seven in 42, the club were again relegated to the Third Division. (Note that this is not the same as the Third Division South, as the regional Third Divisions had been removed with the 1959 creation of the Fourth Division). The aforementioned Third Division hiatus lasted only two seasons before the club was promoted back to the Second Division as runners-up in 1970–71. This spell also saw Fulham invited to the Anglo-Italian Cup, which saw the club draw four out of four games in 1972–73 season. This preceded a period of high-profile signings for the club under Alec Stock in the mid-1970s, including Alan Mullery and Bobby Moore. Fulham reached their only FA Cup final in 1975, having won their first semi-final in five attempts. The club then lost to West Ham United in the final. This gained the club qualification to another European tournament, the Anglo-Scottish Cup, where they made the final, losing to Middlesbrough. George Best played 47 times for the club in the 1976–77 season. Rodney Marsh, who having grown up with Fulham in the 1960s went on to play First Division football and play for England, rejoined the club in the same season, playing only 16 games. This capped one of the most successful eras in Fulham history. The club were relegated again after winning only 11 in 42 matches in the 1979–80 season, which eventually resulted in Bobby Campbell's sacking in October 1980, to be replaced by Malcolm Macdonald. With a strong squad during his 1980–1984 period in charge (with players such as Ray Houghton, Tony Gale, Paul Parker, Gerry Peyton and Ray Lewington), they won promotion again in 1981–82 back to the Second Division, although the promotion was overshadowed by the suicide of former defender Dave Clement a few weeks before promotion was sealed. In 1980, Fulham founded the rugby league club that is now London Broncos designed to be an extra stream of income for the football club, but which made financial losses every year while linked to Fulham F.C. Then called "Fulham Rugby League," they played at Craven Cottage until moving away from the parent club in 1984. In 1978, Fulham had signed Gordon "Ivor" Davies who, during two spells at Fulham, became the club's leading goalscorer of all time with a total of 178 goals in all competitions; the record still stands. Fulham narrowly missed out on back-to-back promotions to the First Division, losing 1–0 to Derby County away on the last day of the 1982–83 season – although the match was abandoned after 88 minutes due to a pitch invasion and inexplicably never replayed or finished. The side which had shown so much promise was quickly sold off as the club were in debt, so it was little surprise when the club were relegated again to the Third Division in 1986. The club nearly went out of business in 1987 via an ill-advised merger attempt with Queens Park Rangers. It was only the intervention of ex-player Jimmy Hill that allowed the club to stay in business as a re-structured Fulham FC 1987 Ltd. In 1987, the club took part in what was then the longest penalty deciders ever recorded – it needed 28 spot kicks to sort out a winner between them and Aldershot following a Freight Rover Trophy match. In 1992, the foundation of the Premier League, and the resignation of 22 clubs from The Football League, restored Fulham to that league's Second Division. However, the club were relegated to the new Third Division after a poor 1993–94 season, following which Ian Branfoot was appointed as team manager. After an eighth-place finish in Branfoot's first season in charge, the club hit its lowest-ever final league position in the 1995–96 season, finishing 17th out of 24. Branfoot was sacked as manager, but remained at the club in other capacities for a short while. In February 1996, Micky Adams became player-manager. Adams oversaw an upturn in form that lifted the side out of relegation danger. The next season, he engineered a second-place league finish, missing out on first place because several years previously the league had dropped the old "goal difference" system in favour of a "goals scored" tally, meaning Fulham finished behind Wigan Athletic. The club's chairman Jimmy Hill had argued that goals scored should decide places of teams tied on points, and the Football League clubs had voted the system in. Egyptian businessman Mohamed Al-Fayed bought the club for £6.25 million in summer 1997. The club was purchased via Bill Muddyman's Muddyman Group. Micky Adams was replaced by Al-Fayed in the aftermath of a mid-table start to the season. He installed a two-tier management "dream team" of Ray Wilkins as First Team Manager and Kevin Keegan as chief operating officer, pledging that the club would reach the Premier League within five years. After an argument over team selection, Wilkins left the club in May 1998 to hand over the full managerial duties to Keegan, who helped steer the club to promotion the next season, winning 101 points out of a possible 138, after spending £1.1 million to sign Paul Peschisolido from West Bromwich Albion, who was top scorer and captained by Chris Coleman – then the most expensive footballer outside the top two divisions of the English league. In 1999, Keegan left Fulham to become manager of England, and Paul Bracewell was put in charge. Bracewell was sacked in March 2000, as Fulham's promising early season form dwindled away to a mid-table finish. Frenchman Jean Tigana was put in charge and, having signed a number of young stars (including French striker Louis Saha), he guided Fulham to their third promotion in five seasons in the 2000–01 season, giving Fulham top-flight status for the first time since 1968. Fulham once again amassed 101 points out of a possible 138 in their scintillating title run, which was crowned with an open-top bus parade down Fulham Palace Road. They are the only team to have twice reached 100 points in a season. During the season, Chris Coleman was involved in a car crash that put him out of action for well over a year and eventually ended his playing career after he failed to make a sufficient recovery. Fulham's run through the divisions saw a large turnover of players, with the only player to play for the club in all four leagues being Sean Davis. Fulham returned to the top division of English football, and competed in the Premier League for the first time. The club finished the 2001–02 season in 13th place. Fulham were the only team to host top-flight football with some standing areas in the 21st century, but due to restrictions on standing, this was not allowed to continue; clubs promoted from the second division had only three years to make their ground all-seater. Fulham were forced to groundshare with QPR at Loftus Road during the 2002–03 and 2003–04 seasons while Craven Cottage was rebuilt as an all-seated stadium. There were fears that Fulham would not return to the Cottage, after it was revealed that Al-Fayed had sold the first right to build on the ground to a property development firm. In 2002–03, Fulham spent most of the season in the lower half of the table. Chairman Al-Fayed told manager Jean Tigana that his contract would not be renewed at the end of the season. However, with five games left to play and relegation still possible, Tigana was sacked, and Chris Coleman was temporarily put in charge. Fulham won 10 points from a possible 15 and managed to avoid relegation. Coleman was appointed manager on a permanent basis in the summer of 2003; despite predictions that the inexperience of Coleman would result in Fulham's relegation, he kept the club well clear of relegation, guiding them to a club record ninth-place finish in his debut season. This might have been greater had the club not come under significant financial pressure to sell Louis Saha to Manchester United, for whom they received a club record £13 million. Fulham lost a legal case against former manager Tigana in 2004 after Al-Fayed wrongly alleged that Tigana had overpaid more than £7 million for new players and had negotiated transfers in secret. Coleman notched up another satisfactory performance in the 2004–05 season and guided Fulham to a secure 13th-place finish. The following season Fulham improved by one place, finishing 12th – the high point of the season was a 1–0 win over local rivals and reigning champions Chelsea in the West London derby – Chelsea had only lost two games in two and a half years. The 2006–07 season proved to be Coleman's last, as on 10 April 2007, Fulham terminated his contract with immediate effect. His replacement was Northern Ireland manager Lawrie Sanchez. Fulham only gained four points from five games with Sanchez as caretaker manager. They ensured top-flight survival that season by defeating a weakened Liverpool side 1–0 in the penultimate match of the season, and Sanchez was appointed manager. Sanchez received strong financial backing from the board and made a number of signings during the summer break, but, after just two league wins in the first five months of the season and with Fulham in the relegation zone, he was dismissed on 21 December 2007 after a defeat to Newcastle United. Roy Hodgson was named as the new manager of Fulham on 28 December 2007 and took up his contractual duties on 30 December, just two days before the January transfer window opened. Hodgson's tenure did not start well and it took a month to secure his first win, against Aston Villa, courtesy of a Jimmy Bullard free-kick. Fulham continued to struggle and a 3–1 home defeat in April at the hands of fellow strugglers Sunderland left Hodgson on the verge of tears in the post-match press conference and many pundits writing off Fulham's survival chances. Despite the negative press, Hodgson continued to believe survival was attainable. The turning point of the season came in the third-to-last match, against Manchester City. Fulham trailed 2–0 at half-time and had the Premier League scores at that time become results, they would have been relegated. However, the introduction of Diomansy Kamara heralded the start of a fantastic comeback—Kamara struck twice as Fulham registered an amazing 3–2 victory. Fulham then won a crucial match against fellow strugglers Birmingham City at Craven Cottage, leaving survival in the club's own hands. Barring a goal-rush from fellow strugglers Reading, a win against a Portsmouth side looking ahead to their fourth FA Cup final would guarantee survival. With 15 minutes to play at Portsmouth, Fulham were drawing, and with Birmingham City and Reading leading comfortably against Blackburn Rovers and Derby County respectively, they looked likely to be relegated. However, Fulham earned a free-kick with 76 minutes played; Jimmy Bullard's delivery found Danny Murphy, who headed home the decisive goal, sparking manic celebrations from the travelling fans. Hodgson had ensured survival against all odds, breaking several club records in the process and cementing his place in Fulham folklore. Fulham narrowly missed out on a UEFA Cup place via Fairplay by a dubious 0.8 of a point behind Manchester City, who lost 8–1 at Middlesbrough. In the 2008–09 season, Fulham finished seventh, their highest-ever league placing, earning qualification for the inaugural UEFA Europa League, the second time that the club had entered a UEFA competition. 2009–10 was arguably the most successful season in the club's history. They were eliminated from the FA Cup in the quarter-finals for the second year running, and finished 12th in the Premier League, despite fielding weakened teams in the last few matches. In the inaugural Europa League season, however, Fulham reached the final, meeting Spanish club Atlético Madrid, who had dropped down from the Champions League, at the Volksparkstadion in Hamburg. In their first European cup final, the Cottagers were beaten 2–1 after extra time, having drawn 1–1 after full-time. The achievement of taking Fulham so unexpectedly far, beating famous teams like Hamburger SV, Juventus, holders Shakhtar Donetsk and Basel in the competition, led to Roy Hodgson being voted the LMA Manager of the Year by the widest margin in the history of the award. The home match in the round of 16 was arguably Fulham's greatest result in the history of the club. Despite losing 3–1 in the first leg at Italian giants Juventus and falling behind minutes into the second leg at Craven Cottage, Fulham scored four goals with no reply from Juventus. At the end of the season, Hodgson left Fulham to manage Liverpool. On 29 July 2010, Mark Hughes was named the successor to Hodgson, signing a two-year contract with the club. Hughes had previously managed Manchester City, the Welsh national team and Blackburn. Hughes' first match in charge was against Bolton Wanderers at the Reebok Stadium. The highlight of the season was a 4–0 win in the FA Cup over London rivals Tottenham Hotspur, all goals coming in the first half. Hughes resigned as manager of Fulham on 2 June 2011, having spent fewer than 11 months at the club. The Whites had an encouraging finish in eighth position and qualified for the Europa League via Fairplay. On 7 June 2011, Martin Jol signed a two-year contract with Fulham, becoming successor to Hughes. Jol's first match was a 3–0 Europa League win against NSÍ Runavík of the Faroe Islands on 30 June. Fulham then navigated their way with some ease to the group stage in the Europa League through late summer. However, the Cottagers were knocked out with the last seconds of the group stage matches, Odense BK equalising to make a draw, leaving Fulham in third place, with Polish side Wisła Kraków instead progressing to the next round. Fulham's Premier League form in the 2011–12 season was mixed, with the continuing away-record hangover of previous seasons dragging on. In October 2011, Fulham had an emphatic 6–0 home win over neighbours QPR, with Andrew Johnson scoring a hat-trick for Fulham in the match. The January 2012 transfer window saw Bobby Zamora move over the Hammersmith flyover to Loftus Road, with Russian striker Pavel Pogrebnyak coming in place from VfB Stuttgart. The New Year saw two further hat-tricks scored by Clint Dempsey. On 11 February 2012, Progrebnyak scored on his debut in the 2–1 win over Stoke City. In March 2012, a 5–0 win against Wolverhampton Wanderers saw a hat-trick from Pogrebnyak. The Cottagers broke their historic drought on Merseyside with a 1–0 win over Liverpool at Anfield on May Day and another win against Sunderland in the last home game meant Fulham were only one point short of equalling their largest points haul in the Premier League, with just one game remaining. However, they failed to achieve this after losing their last game away at Tottenham. In the 2012–13 season, Fulham ended a seven-match winless run by beating Swansea City 3–0 away at the Liberty Stadium on the final game of the season on 19 May 2013. Fulham finished the season in 12th place. Shahid Khan took over as chairman in July 2013, but after a poor start to the 2013–14 season, having only amassed 10 points from 13 games, Martin Jol was sacked as manager on 1 December 2013, with René Meulensteen taking charge as head coach. Meulensteen was replaced by Felix Magath after just 17 games in charge following no upturn in form, but fortunes did not improve, and Fulham were eventually relegated to the Championship after a 4–1 defeat away to Stoke on 3 May. Post-season, the media criticised chairman Shahid Khan's decision to sack Meulensteen and appoint the third manager of the season in Magath Fulham broke the Championship transfer record that summer in a restructuring of the squad by Magath, but after an alarming start to the new season, amassing just one point in seven games, Magath was sacked in September 2014, with Kit Symons appointed as caretaker manager. Fulham eventually finished the season in an uneventful 17th place. The following season, the team suffered an inconsistent start to the season and after a 5–2 loss at home to Birmingham City, and lying in 12th place, Kit Symons was sacked as manager in November 2015. It paved the way for Serbian Slaviša Jokanović to be appointed on 27 December 2015. Fulham's fortunes did not improve greatly following Jokanović's appointment, but the team finished the 2015–16 Championship season in 20th place, avoiding relegation by 11 points. The 2016–17 season saw huge improvements in both results and performances. Despite an inconsistent start, the team saw a significant improvement from October onwards which propelled them up the table. Fulham secured a 6th-place finish after an impressive finish to the season and entered the play-offs, but they were undone by a controversial Yann Kermorgant penalty, which saw Reading triumph 2–1 on aggregate. Despite a poor start to the following season, the club went on a club-record 23 game unbeaten run in the league which led to a 3rd-place finish, narrowly missing out automatic promotion, and the team went on to win the EFL Championship play-off Final against Aston Villa to return to the Premier League on 26 May 2018. Following a poor start to life back in the Premier League, Jokanović was sacked on 14 November 2018 and replaced with former Leicester manager Claudio Ranieri. Results ultimately did not improve under Ranieri and he too was sacked in February 2019 after just 4 months in charge. He was replaced by coach Scott Parker as caretaker manager who couldn't save the club from relegation on 3 April 2019. Parker was appointed as manager on a permanent basis on 10 May 2019. Fulham Football Club is owned by Shahid Khan. Khan completed his purchase of the club from Mohamed Al-Fayed on 12 July 2013 for a reported £150–200 million. During his ownership of Fulham, Al-Fayed had provided the club with £187 million in interest-free loans. In March 2011, Fulham posted annual losses of £16.9 million, with Al-Fayed stating that he would "continue to make funds available to achieve our goals both on and off the pitch" and that "the continued success of Fulham and its eventual financial self-sustainability is my priority." As of January 2013, Fulham were effectively debt-free as Al-Fayed converted the loans into equity in the club. Fulham's sponsorship by Betfair in 2002–03 was the first gambling sponsorship in English football, and came before the Gambling Act 2005 permitted the industry to advertise on television and radio; within fifteen years half of Premier League teams were sponsored by such companies. Fulham are a member of the European Club Association, having qualified four times for European Competition, firstly the UEFA Intertoto Cup after their inaugural season in the Premier League, then the UEFA Cup as a result of winning that, and then the UEFA Europa League twice. Fulham are unbeaten at home in European competition, in 23 games, with a record of 17 wins and six draws. In 2010, Fulham reached the UEFA Europa League final, which they lost 2–1 to Atletico Madrid. Fulham fans consider their main rivals to be Chelsea. Despite this fixture not being played that often in the years preceding Fulham's ascent to the top division, this is a clear local derby as Chelsea's ground, Stamford Bridge, is actually within Fulham. Fulham consider their secondary rivals to be Queens Park Rangers. Fulham last played QPR in the 2000–01 season before meeting them again twice in the 2011–12 Premier League season in which Fulham were the victors with a 6–0 victory at Craven Cottage, and beating them 1–0 away from home at Loftus Road. Fulham also have rivalries with other London clubs, including Brentford and Crystal Palace. Outside of London, Gillingham are still considered rivals to Fulham supporters despite the two clubs having played in different divisions for the past 11 years. Fulham and Gillingham have been involved in several ill-tempered matches in the lower leagues, including the death of a Fulham supporter. Fulham's fan base has fluctuated over the years, with high crowds coinciding with the club's success in the Premier League so that the club now averages in the top 20 home attendances in the country. Fulham supporters have played a vital role in the club's long term stay at Craven Cottage. When the club moved temporarily to Loftus Road, a committee known as Back to the Cottage was formed, committed to ensuring the club continued to play at their traditional home. Fulham fans have traditionally come from the Fulham and Hammersmith areas, and also from other areas in South-West London, such as Putney, Richmond, Sutton and Worcester Park. In July 2012, the club website asked supporters using Facebook and Twitter to pick their best FFC Premier League XI from 2001 to the present. The supporters picked their favourite goalkeeper, full-backs, centre-backs, wingers, centre midfielders and forwards in a classic 4–4–2 formation. Fulham have had 37 managers in 114 years. Prior to the appointment of the first manager at the club (Bradshaw in 1904), the duties normally assigned to a modern-day manager would have been shared between club secretary, captain, and other officials. Managerial records: Temporary managers at the club have included: Between the years 1879 and when Fulham had a ground to call their own in 1896, they played at a number of stadiums, only some of which were recorded and this should not be regarded as a full or complete list. Only rivals and former landlords Queens Park Rangers have played at more home stadiums. Some of the early grounds listed below are likely to have been parks and parkland, which have now been developed. Even when the club purchased Craven Cottage and the surrounding land in 1894, they had to wait two years before they could play a game there. The Fulham mascot is Billy the Badger, who was the winning design sent in by Kyle Jackson after an online competition by the club. Billy the Badger wears the number 79 Fulham shirt, in reference to the club's year of founding, 1879. Controversy first surrounded Billy when he tried to cheer up Chelsea manager Avram Grant during a home match in front of the television cameras. Secondly, Billy was seen on television being sent off during the home game against Aston Villa on 3 February 2008 for break-dancing in the corner of the pitch after the referee had commenced the game. Billy blamed his badger hearing and eyesight for the incident, and apologised to referee Chris Foy. On 11 March 2009, Billy walked across the goal during a match although it was not spotted by the referee. The former mascot for Fulham was Sir Craven of Cottage, the Knight. The cheerleaders were known as the Cravenettes.
https://en.wikipedia.org/wiki?curid=11228
Frankie Goes to Hollywood Frankie Goes to Hollywood were a British band formed in Liverpool, England, in the 1980s. The group was fronted by Holly Johnson (vocals), with Paul Rutherford (vocals), Peter Gill (drums, percussion), Mark O'Toole (bass guitar) and Brian Nash (guitar). The group's 1983 debut single "Relax" was banned by the BBC in 1984 while at number six in the charts and subsequently topped the UK Singles Chart for five consecutive weeks, going on to enjoy prolonged chart success throughout that year and ultimately becoming the seventh-best-selling UK single of all time. It also won the 1985 Brit Award for Best British Single. Their debut album, "Welcome to the Pleasuredome", reached number one in the UK in 1984 with advanced sales of over one million. After the follow-up success of "Two Tribes" and "The Power of Love", the group became only the second act in the history of the UK charts to reach number one with their first three singles; the first being fellow Liverpudlians Gerry and the Pacemakers in the 1960s. This record remained unbeaten until the Spice Girls achieved a six-single streak in 1996–1997. In 1985 the band won the Brit Award for British Breakthrough Act. Associated with the Second British Invasion of the US, they also received Grammy Award and MTV Video Music Award nominations for Best New Artist. Songwriters Johnson, Gill and O'Toole received the 1984 Ivor Novello Award from the British Academy of Songwriters, Composers, and Authors for Best Song Musically and Lyrically for "Two Tribes". In 2015, the song was voted by the British public as the nation's 14th-favourite 1980s number one in a poll for ITV. On the B-side to the group's first single, Johnson explained that the group's name derived from a page from "The New Yorker" magazine, featuring the headline "Frankie Goes Hollywood" and a picture of Frank Sinatra, although the magazine page Johnson referred to was actually a pop art poster by Guy Peellaert. The original group named "Frankie Goes to Hollywood" dates from 1980. The nucleus of the group emerged from the late 1970s Liverpool punk scene. Lead singer Johnson had played bass with Big in Japan and had also released two solo singles. Local musicians Peter Gill (drums), Jed O'Toole (bass), and O'Toole's cousin Brian Nash (guitar) initially joined Johnson, calling themselves the Sons of Egypt. This line-up secured a number of small local gigs before disbanding. The group was reprised when Johnson joined Mark O'Toole (bass) and Peter "Ped" Gill to form FGTH. During a particularly fluid period of personnel changes, Jed O'Toole joined FGTH on guitar, and a female vocalist, Sonia Mazumder, was also a band member for the first Frankie gig at the Leeds nightclub "The Warehouse", supporting Hambi & The Dance. Paul Rutherford—a member of the headline act who had also sung in seminal Liverpool punk band The Spitfire Boys—apparently got so caught up in Frankie's performance that he effectively replaced Mazumder that very night. The new all-male musical line-up subsequently toured locally with a leather-clad female duo known as "The Leatherpets" and managed to fund promotional videos and demos, despite being eventually turned down by both Arista Records and Phonogram Inc. In October 1982, the group recorded a John Peel Session for BBC Radio 1, comprising the originals "Krisco Kisses", "Two Tribes", "Disneyland", and "The World Is My Oyster". Around this time Jed O'Toole left the group, to be replaced by the returning Brian Nash. In February 1983, the group was invited to record a video for "Relax" by the Channel 4 show "The Tube" at the Liverpool State Ballroom. After the broadcast, the Peel session was repeated on radio, and a new session recorded for the BBC, comprising "Welcome to the Pleasuredome", "The Only Star in Heaven" and "Relax". These performances, along with a repeat of the "Tube" video, convinced Trevor Horn to sign the group for his new label, ZTT Records, in May 1983. "Relax" was released by ZTT in October 1983, with production and remix directed by Trevor Horn, received a modicum of airplay, allowing it steady progress into the UK Top 40. Following a debut on the BBC's "Top of the Pops" on 5 January 1984 while at number 35, the single rose to number six the following week. On 11 January 1984, BBC Radio 1 disc jockey Mike Read was playing the record on his show when he noticed the front cover design (by Yvonne Gilbert). Read apparently became outraged by the "overtly sexual" nature of both the record sleeve and the printed lyrics, which prompted him to remove the disc from the turntable live on air, branding it "obscene". Two days later—almost three months after the single's initial release, and just eight days after the group's "Top of the Pops" appearance—the BBC banned the record from all its TV and radio outlets. "Relax" immediately shot to number one in the UK charts and stayed there for five weeks, during which time the BBC could not feature the nation's best-selling single on "Top of the Pops". The original video was directed by Bernard Rose and depicted a gay S&M parlour where the band members were admired by muscular leathermen, a bleached blonde drag queen, and a large-bodied man dressed as a Roman emperor. The video featured a scene where one of the band members wrestled a live tiger, to the admiration of the clubgoers, and ended where the "emperor" was so excited he shimmied out of his toga. Filmed in the unused East London theatre Wilton's Music Hall, it was promptly banned by both the BBC and MTV, resulting in the production of a substitute video directed by filmmaker Brian De Palma to coincide with the release of his film "Body Double". There have been 4 official music videos for "Relax." The BBC lifted its ban on "Relax" at the end of 1984 to allow the band to perform it on the Christmas edition of "Top of the Pops" (it had been, aside from Band Aid of which Holly Johnson was a participant, the biggest-selling single of the year). "Relax" remained in the charts when the follow-up, "Two Tribes", was released in May 1984. The anti-conflict song was given an aggressively topical nuclear war slant. Featuring sirens, the unmistakable voice of Patrick Allen (who had voiced the British Government's actual nuclear warning ads, "Protect and Survive", two years earlier) and another innovative electronic backing, it went straight into the UK charts at Number One and stayed there for nine weeks (the first single to do so since John Travolta and Olivia Newton-John's "You're the One That I Want" in 1978), with total sales exceeding 1.5 million copies and becoming one of the top 30 best-selling records in the UK ever. Directed by Kevin Godley and Lol Creme, the video featured lookalikes of Cold War leaders Ronald Reagan and Konstantin Chernenko wrestling in a marquee while band members and others laid bets on the outcome. Ultimately, the audience—consisting of other world leaders—were brought into the fight, and eventually Earth was seen to explode. "Two Tribes" was a successful single in its own right, but its reign at the top of the charts was made even more notable by the continuing success of its predecessor. "Relax" had made a natural decline down the charts by May 1984, but on the release of "Two Tribes" its sales began to increase again, to the extent that FGTH held the top two spots in the UK charts during July 1984, the first active group to do so since the early 1960s. The release of "Two Tribes" also coincided with an extensive and iconic T-shirt marketing campaign during the British summer of 1984, featuring such slogans as "Frankie Say War! Hide Yourself" and "Frankie Say Relax Don't Do It!" FGTH released a third single, "The Power of Love", at the end of 1984. Unlike the earlier singles, this song was a slower-paced ballad, but it also went to Number One in December and making the band the first act for two decades (since Gerry and the Pacemakers, a fellow Liverpool band, in 1963) to achieve chart-toppers with its first three releases. The video was not banned on this occasion but still caused trouble for the group—because it depicted a nativity scene (and on its first showing did not feature any members of the band, who were subsequently added as picture framing). The song's release was preceded by an advertising campaign that, cheekily, declared it to be the band's third number one single, as if this was a fait accompli. The Band Aid project, for which Johnson recorded a message for the B-side, meant that FGTH managed only one week at the top this time before it was replaced by "Do They Know It's Christmas?" The title track from FGTH's album, "Welcome to the Pleasuredome", was released as a fourth single in March 1985. Early promotional posters for the single proclaimed it as "their fourth number one", even prior to the single's release. However, the single peaked at Number 2. The twelve-inch singles featured Greek mythology/Samuel Taylor Coleridge spoken introduction. Geoffrey Palmer narrates on the second 12", known as the "fruitness" mix. Only one new track appeared in the next eighteen months; "Disneyland", was released on the ZTT Records "Zang Tuum Tumb Sampled" album in late 1985. In 1986, FGTH appeared at the Montreux Rock Festival which was broadcast on UK television. This performance saw the first airings of two future singles, namely "Rage Hard" and "Warriors of the Wasteland". Both versions were different from the versions eventually released. In August 1986, the long-awaited new Frankie Goes to Hollywood single, "Rage Hard", was released, reaching number 4 in the UK. Initially showcased promotionally with songs like "Warriors of the Wasteland", the group's sound had developed a significantly harder edge with a less flamboyant, more nitty-gritty lyrical side. The album, "Liverpool", was released in October 1986 and reached UK No. 5. It was generally panned by the music press and chart returns declined rapidly with the follow-up singles "Warriors of the Wasteland" (No. 19) and "Watching the Wildlife" (No. 28). The group meanwhile threatened to implode of its own accord, in the course of a tour promoting the new album. Johnson kept himself markedly separate from the rest of the band when offstage during this period, tensions becoming exacerbated during a backstage altercation between Johnson and O'Toole at Wembley Arena in January 1987, reflecting the generally collapsing relationship between lead singer and the rest of the band. Things were so bad that fellow Liverpudlian singer Pete Wylie was approached to replace Johnson but declined the offer. FGTH completed the tour, but Johnson ultimately left the group thereafter, citing musical estrangement. In the aftermath of the group split, Johnson was offered a solo recording agreement with MCA Records. However, ZTT, which maintained they had invested heavily in "Liverpool" (to the extent that the digital recording system used to record the album was very nearly treated as a sixth member of the band on the sleeve of the "Warriors of the Wasteland" single), had other ideas, and promptly sued Johnson in an attempt to hold him to his original contract with the label. Among other things, ZTT believed that as a departing member of FGTH, Johnson was required to release all solo material through the label until the band's original multiple-album agreement was fulfilled. The suit was bitterly fought, exposing the inner workings of the ZTT/Frankie machine to a giddy UK music press. After two years, the High Court found in Johnson's favour, holding that the highly restrictive terms of the contract constituted an unreasonable restraint of trade. The court case also effectively freed the remaining members of FGTH from their ZTT contract. Johnson's solo career at MCA commenced in 1989, with a succession of high-placed singles and the number one album "Blast". The remix collection "Hollelujah" followed, trailed by a second studio album, "Dreams That Money Can't Buy". However, Johnson's relations with MCA cooled with this release, and he would ultimately become a reclusive but successful painter, after announcing in 1993 that he was HIV-positive. The following year, Johnson recounted his version of Frankie's history in his autobiography "A Bone in My Flute". His self-issued 1999 album "Soulstream" included a re-recording of "The Power of Love", which was also released as a single. Paul Rutherford, the other openly gay member of the band, released the partially ABC produced album "Oh World" and a handful of singles before retiring with his New Zealander partner to Waiheke Island. The "other three", as "Smash Hits" labelled them, continued to work together in what turned out to be a vain attempt to resurrect "Frankie" with various singers including Dee Harris from Fashion and Grant Boult (Jeckyl Ice) from The Premise, who had opened the shows on the band's UK and European tours. Under the name Boss Dog, with Boult on vocals, the band were offered a major deal with Virgin Records but on the condition they work as Frankie Goes to Hollywood. Johnson challenged the use of the name and the deal soured. Boult and Brian Nash continued with the material written by The Shuffle Brothers and under the name Low they released "Tearing My Soul Apart" in 1992 on Swanyard Records. As "Nasher", Nash released a 2002 solo album entitled "Ripe". Ped worked behind the scenes and scored a top ten hit with the group "Lovestation". Mark O'Toole moved to Florida and played with punk outfit "Trapped by Mormons". The band's name lived on to the extent that re-issues of "Relax" and "The Power of Love" both returned to the UK Top 10 in 1993. Remixes of "The Power of Love" (which became a dance anthem from its original ballad format) and "Two Tribes" were Top 20 hits again in 2000, while "Welcome to the Pleasuredome" also got commercially successful remix treatment, to the extent of a Top 20 placing four years earlier. The group's first two singles appeared sixth and 22nd, respectively, in the official all-time UK best-selling singles list issued in 2002. In 1998, a band calling itself alternately "Frankie Goes to Hollywood" and "The New Frankie Goes to Hollywood Featuring Davey Johnson" began to tour the United States. The band consisted of none of the original members of the band and formed without their knowledge or consent. The impostor band was led by an American using the stage name Davey Johnson, who alternately claimed he was Holly Johnson's brother and had performed as an uncredited session musician on "Welcome to the Pleasuredome". The members of the actual band and their producer Trevor Horn refuted both claims. Mark O'Toole, who had been living in Florida, became aware of the band and warned concert promoters not to hire them. Likewise, A Flock of Seagulls frontman Mike Score, who had been a Liverpool acquaintance of the members of Frankie Goes to Hollywood, kicked the impostor band off his tour after discovering they were a fraud. After Holly Johnson contacted the trade magazine "Pollstar" to confirm that the American-based act was unauthorized, the impostor band was dropped by a booking agent, but continued to be booked by small clubs throughout the southern United States. The fake group continued to perform until at least September 2000, when a feature on the controversy was published in that month's issue of "Spin". In 2003, the VH-1 program "Bands Reunited" brought Johnson, Rutherford, Gill, Nash, and Mark O'Toole together, in the hope of their agreeing to perform impromptu on the show. However, a reunion performance did not transpire. Both Johnson and Nash had reservations about performing at short notice in the contrived manner dictated by the TV show format. In 2004, a celebration of the 25th anniversary of Trevor Horn's involvement in the music industry a special concert took place at Wembley Arena in November, featuring three of the original FGTH line-up, Mark O'Toole, Peter (Ped) Gill and Paul Rutherford and another former member, Jed O'Toole. Original vocalist Johnson, who announced via the Internet that he would not be appearing, and guitarist Nash, who declined to appear for his own reasons, did not take part. Jed O'Toole took over guitar duties for the event, whilst an open audition was held for a new singer for the concert. Ryan Molloy was recruited as a result. The same lineup reunited for a tour in 2005 playing festivals in Europe. They headlined the Big Gay Out festival at Hyde Park in London. After some confusion with ensuing tour dates, the band posted a warning on their website that many of the tour dates listed by ticket promoters were inaccurate. The band became increasingly focused on the release of a new album during 2007. However, in early April 2007 came the news that Gill, Rutherford, J. O'Toole and Molloy had formed Forbidden Hollywood to play their new songs alongside old FGTH material. This was to avoid legal issues with Holly Johnson over use of the Frankie Goes to Hollywood name. Live dates were announced but in June 2007, these were cancelled. ! Year !! Awards !! Work !! Category !! Result In 1985, a computer game entitled "Frankie Goes to Hollywood" was developed by Denton Designs and published by Ocean Software. Based on the band's music, imagery and slogans, the objective of the game was to explore the town of Mundanesville to find the Pleasuredome. It also came boxed with an extra cassette with a live version of "Relax" on it.
https://en.wikipedia.org/wiki?curid=11229
Factors of production In economics, factors of production, resources, or inputs are what is used in the production process to produce output—that is, finished goods and services. The utilized amounts of the various inputs determine the quantity of output according to the relationship called the production function. There are three "basic" resources or factors of production: land, labour and capital. The factors are also frequently labeled "producer goods or services" to distinguish them from the goods or services purchased by consumers, which are frequently labeled "consumer goods". There are two types of factors: "primary" and "secondary". The previously mentioned primary factors are land, labour and capital. Materials and energy are considered secondary factors in classical economics because they are obtained from land, labour, and capital. The primary factors facilitate production but neither becomes part of the product (as with raw materials) nor becomes significantly transformed by the production process (as with fuel used to power machinery). Land includes not only the site of production but also natural resources above or below the soil. Recent usage has distinguished human capital (the stock of knowledge in the labor force) from labor. Entrepreneurship is also sometimes considered a factor of production. Sometimes the overall state of technology is described as a factor of production. The number and definition of factors vary, depending on theoretical purpose, empirical emphasis, or school of economics. In the interpretation of the currently dominant view of classical economic theory developed by neoclassical economists, the term "factors" did not exist until after the classical period and is not to be found in any of the literature of that time. Differences are most stark when it comes to deciding which factor is the most important. Physiocracy (from the Greek for "government of nature") is an economic theory developed by a group of 18th century Enlightenment French economists who believed that the wealth of nations was derived solely from the value of "land agriculture" or "land development" and that agricultural products should be highly priced. The classical economics of Adam Smith, David Ricardo, and their followers focus on physical resources in defining its factors of production and discuss the distribution of cost and value among these factors. Adam Smith and David Ricardo referred to the "component parts of price" as the costs of using: The classical economists also employed the word "capital" in reference to money. Money, however, was not considered to be a factor of production in the sense of capital stock since it is not used to directly produce any good. The return to loaned money or to loaned stock was styled as interest while the return to the actual proprietor of capital stock (tools, etc.) was styled as profit. See also returns. Marx considered the "elementary factors of the labor-process" or "productive forces" to be: The "subject of labor" refers to natural resources and raw materials, including land. The "instruments of labor" are tools, in the broadest sense. They include factory buildings, infrastructure, and other human-made objects that facilitate labor's production of goods and services. This view seems similar to the classical perspective described above. But unlike the classical school and many economists today, Marx made a clear distinction between labor actually done and an individual's "labor power" or ability to work. Labor done is often referred to nowadays as "effort" or "labor services." Labor-power might be seen as a stock which can produce a flow of labor. Labor, not labor power, is the key factor of production for Marx and the basis for Marx's labor theory of value. The hiring of labor power only results in the production of goods or services ("use-values") when organized and regulated (often by the "management"). How much labor is actually done depends on the importance of conflict or tensions within the labor process. Neoclassical economics, one of the branches of mainstream economics, started with the classical factors of production of land, labor, and capital. However, it developed an alternative theory of value and distribution. Many of its practitioners have added various further factors of production (see below). Further distinctions from classical and neoclassical microeconomics include the following: Ecological economics is an alternative to neoclassical economics). It integrates, among other things, the first and second laws of thermodynamics (see: Laws of thermodynamics) to formulate more realistic economic systems that adhere to fundamental physical limitations. In addition to the neoclassical focus on efficient allocation, ecological economics emphasizes sustainability of scale and just distribution. Ecological economics also differ from neoclassical theories in its definitions of factors of production, replacing them with the following: Integral to ecological economics is the following notion: at the maximum rates of sustainable matter and energy uptake, the only way to increase productivity would be through an increase in design intelligence. This provides the basis for a core tenet of ecological economics, namely that infinite growth is impossible. In the first half of the 20th century, some authors added the work of organization or entrepreneurship as a fourth factor of production. This became standard in the post-war Neoclassical synthesis. For example, J. B. Clark saw the co-ordinating function in production and distribution as being served by entrepreneurs; Frank Knight introduced managers who co-ordinate using their own money (financial capital) and the financial capital of others. In contrast, many economists today consider "human capital" (skills and education) as the fourth factor of production, with entrepreneurship as a form of human capital. Yet others refer to intellectual capital. More recently, many have begun to see "social capital" as a factor, as contributing to production of goods and services. In markets, entrepreneurs combine the other factors of production, land, labor, and capital, to make a profit. Often these entrepreneurs are seen as innovators, developing new ways to produce and new products. In a planned economy, central planners decide how land, labor, and capital should be used to provide for maximum benefit for all citizens. Just as with market entrepreneurs, the benefits may mostly accrue to the entrepreneurs themselves. The sociologist C. Wright Mills refers to "new entrepreneurs" who work within and between corporate and government bureaucracies in new and different ways. Others (such as those practicing public choice theory) refer to "political entrepreneurs", i.e., politicians and other actors. Much controversy rages about the benefits produced by entrepreneurship. But the real issue is about how well institutions they operate in (markets, planning, bureaucracies, government) serve the public. This concerns such issues as the relative importance of market failure and government failure. In the book "Accounting of Ideas", "intequity", a neologism, is abstracted from equity to add a newly researched production factor of the capitalist system. Equity, which is regarded as part of capital, was divided into equity and intequity. Entrepreneurship was divided into network-related matters and creating-related matters. Network-related matters function in the sphere of equity, and creating-related matters in spheres of intequities. Ayres and Warr (2010) are among the economists who criticize orthodox economics for overlooking the role of natural resources and the effects of declining resource capital. See also: "Natural resource economics" Exercise can be seen as individual factor of production, with an elastication larger than labor. A cointegration analysis support results derived from linear exponential (LINEX) production functions. C. H. Douglas disagreed with classical economists who recognized only three factors of production. While Douglas did not deny the role of these factors in production, he considered the “Cultural heritage” as the primary factor. He defined cultural inheritance as the knowledge, techniques, and processes that have accrued to us incrementally from the origins of civilization (i.e., progress). Consequently, mankind does not have to keep "reinventing the wheel". "We are merely the administrators of that cultural inheritance, and to that extent, the cultural inheritance is the property of all of us, without exception. Adam Smith, David Ricardo, and Karl Marx claimed that labor creates all value. While Douglas did not deny that all costs ultimately relate to labour charges of some sort (past or present), he denied that the present labour of the world creates all wealth. Douglas carefully distinguished between value, costs and prices. He claimed that one of the factors resulting in a misdirection of thought in terms of the nature and function of money was economists' near-obsession about values and their relation to prices and incomes. While Douglas recognized "value in use" as a legitimate theory of values, he also considered values as subjective and not capable of being measured in an objective manner. Peter Kropotkin argued for the common ownership of all intellectual and useful property due to the collective work that went into creating it. Kropotkin does not argue that the product of a worker's labor should belong to the worker. Instead, Kropotkin asserts that every individual product is essentially the work of everyone since every individual relies on the intellectual and physical labor of those who came before them as well as those who built the world around them. Because of this, Kropotkin proclaims that every human deserves an essential right to well-being because every human contributes to the collective social product: Kropotkin goes on to say that the central obstacle preventing humanity from claiming this right is the state's violent protection of private property. Kropotkin compares this relationship to feudalism, saying that even if the forms have changed, the essential relationship between the propertied and the landless is the same as the relationship between a feudal lord and their serfs.
https://en.wikipedia.org/wiki?curid=11231
Fort Wayne, Indiana Fort Wayne is a city in the U.S. state of Indiana and the seat of Allen County, United States. Located in northeastern Indiana, the city is west of the Ohio border and south of the Michigan border. With a population of 253,691 in the 2010 census, it is the second-most populous city in Indiana after Indianapolis, and the 75th-most populous city in the United States. It is the principal city of the Fort Wayne metropolitan area, consisting of Allen, Wells, and Whitley counties, a combined population of 419,453 as of 2011. Fort Wayne is the cultural and economic center of northeastern Indiana. The city is within a 300-mile (482.803 km) radius of major population centers, including Chicago, Cincinnati, Cleveland, Columbus, Detroit, Indianapolis, Louisville, Lexington, and Milwaukee. In addition to the three core counties, the combined statistical area (CSA) includes Adams, DeKalb, Huntington, Noble, and Steuben counties, with an estimated population of 615,077. Fort Wayne was built in 1794 by the United States Army under the direction of American Revolutionary War general Anthony Wayne, the last in a series of forts built near the Miami village of Kekionga. Named in Wayne's honor, the European-American settlement developed at the confluence of the St. Joseph, St. Marys, and Maumee rivers as a trading post for pioneers. The village was platted in 1823 and underwent tremendous growth after completion of the Wabash and Erie Canal and advent of the railroad. Once a booming manufacturing town located in what became known as the Rust Belt, Fort Wayne's economy in the 21st century is based upon distribution, transportation and logistics, healthcare, professional and business services, leisure and hospitality, and financial services. The city is a center for the defense industry which employs thousands. There are also many jobs through local healthcare providers Parkview Health and Lutheran Health Network. Fort Wayne was an All-America City Award recipient in 1982, 1998, and 2009. The city also received an Outstanding Achievement City Livability Award by the U.S. Conference of Mayors in 1999. This area at the confluence of rivers was long occupied by successive cultures of indigenous peoples. The Miami tribe established its settlement of Kekionga at the confluence of the Maumee, St. Joseph, and St. Marys rivers. It was the capital of the Miami nation and related Algonquian tribes. In 1696, Comte de Frontenac appointed Jean Baptiste Bissot, Sieur de Vincennes as commander of the outpost. The French built Fort Miami in 1697 as part of a group of forts and trading posts built between Quebec and St. Louis. In 1721, a few years after Bissot's death, Fort Miami was replaced by Fort St. Philippe des Miamis. The first census in 1744 recorded a population of approximately 40 Frenchmen and 1,000 Miami. Increasing tension between France and Great Britain developed over control of the territory. In 1760, France ceded the area to Britain after its forces in North America surrendered during the Seven Years' War, known on the North American front as the French and Indian War. In 1763, various Native American nations rebelled against British rule and retook the fort as part of Pontiac's Rebellion. The Miami regained control of Kekionga, ruling it for more than 30 years. In 1790, after the United States achieved independence, President George Washington ordered the United States Army to secure Indiana Territory. Three battles were fought at Kekionga against Little Turtle and the Miami Confederacy. Miami warriors defeated U.S. forces in the first two battles. General Anthony Wayne led a third expedition resulting in the destruction of Kekionga and the start of peace negotiations between Little Turtle and the U.S. After General Wayne refused to negotiate, tribal forces advanced to Fallen Timbers, where they were defeated on August 20, 1794. On October 22, 1794, U.S. forces captured the Wabash–Erie portage from the Miami Confederacy and built Fort Wayne, named in honor of the general. The first settlement started in 1815. In 1819, the military garrison abandoned the fort and moved to Detroit. In 1822, a federal land office opened to sell land ceded by local Native Americans by the Treaty of St. Mary's in 1818. Platted in 1823 at the Ewing Tavern, the village became an important frontier outpost, and was incorporated as the Town of Fort Wayne in 1829, with a population of 300. The Wabash and Erie Canal's opening improved travel conditions to the Great Lakes and Mississippi River, exposing Fort Wayne to expanded economic opportunities. The population topped 2,000 when the town was incorporated as the City of Fort Wayne on February 22, 1840. Pioneer newspaperman George W. Wood was elected the city's first mayor. Fort Wayne's "Summit City" nickname dates from this period, referring to the city's position at the highest elevation along the canal's route. As influential as the canal was to the city's earliest development, it quickly became obsolete after briefly competing with the city's first railroad, the Pittsburgh, Fort Wayne and Chicago Railway, completed in 1854. At the turn of the 20th century, the city's population reached nearly 50,000, attributed to a large influx of German and Irish immigrants. Fort Wayne's "urban working class" thrived in industrial and railroad-related jobs. The city's economy was substantially based on manufacturing, ushering in an era of innovation with several notable inventions and developments coming out of the city over the years, such as gasoline pumps (1885), the refrigerator (1913), and in 1972, the first home video game console. A 1913 flood caused seven deaths, left 15,000 homeless, and damaged over 5,500 buildings in the worst natural disaster in the city's history. As the automobile's prevalence grew, Fort Wayne became a fixture on the Lincoln Highway. Aviation arrived in 1919 with the opening of the city's first airport, Smith Field. The airport served as Fort Wayne's primary commercial airfield until Baer Field (now Fort Wayne International Airport) was transferred to the city in 1947 after serving as a military base during World War II. Fort Wayne was hit by the Great Depression beginning in 1929, with most factories cutting their workforce. The stock market crash did not discourage plans to build the city's first skyscraper and Indiana's tallest building at the time, the Lincoln Bank Tower. By 1935, the New Deal's WPA put over 7,000 residents back to work through local infrastructure improvements, including the construction of new parks, bridges, viaducts, and a $5.2 million sewage treatment facility. The helped the city prosper once again. Between 1950 and 1955, more than 5,000 homes were built, many in large subdivisions in rural Allen County. In 1950, Fort Wayne's first bypass, Coliseum Boulevard, opened on the north side of the city, followed by the city's first arena, War Memorial Coliseum, bringing new opportunities for suburban expansion. The Coliseum was home to the NBA's from 1952 to 1957. The opening of enclosed shopping malls and the construction of Interstate 69 through rural areas north and west of the city proper further drove the exodus of retail from downtown through the 1960s. According to the Fort Wayne Home Builders Association estimates, more than 80 percent of new home construction occurred outside the city proper in the 1970s. Like many cities in the Rust Belt, deindustrialization in the 1980s brought urban blight, increased crime, and a decrease in blue-collar manufacturing jobs. Downtown and surrounding neighborhoods continued declining as residents and businesses sprawled further into rural Allen County. A 1982 flood forced an evacuation of 9,000 residents, damaging 2,000 buildings, and costing $56.1 million (1982 USD, $137 million 2015 USD). The 1990s marked a turnaround for the city, as local leaders focused on crime reduction, economic diversification, and downtown redevelopment. By 1999, Fort Wayne's crime rate decreased to levels not seen since 1974, and the city's economy recovered, with the unemployment rate hovering at 2.4 percent in 1998. Clearing blighted buildings downtown resulted in new public greenspaces, including Headwaters Park, which has become the premier community gathering space and centerpiece in the city's $50 million flood control project. Fort Wayne celebrated its bicentennial in 1994. The city continued to concentrate on downtown redevelopment and investment in the 2000s. The decade saw the beginnings of its transformation, with renovations and expansions of the Allen County Public Library, Grand Wayne Convention Center, and Fort Wayne Museum of Art. In 2007, the $130 million Harrison Square development was launched, creating Parkview Field. Suburban growth continued, with the opening of Fort Wayne's first lifestyle center, Jefferson Pointe, and the half-billion dollar Parkview Regional Medical Center in 2012. Fort Wayne is in the East North Central region of the Midwestern United States, in northeastern Indiana, west of Ohio and south of Michigan. According to the 2010 census, Fort Wayne has a total area of , of which (or 99.81%) is land and (or 0.19%) is water. For a regional summit, the city is situated on flat land characterized by little topographical relief, a result of the Wisconsin glaciation episode. Receding glaciers eroded the land, depositing an evenly distributed layer of sediment during the last glacial period. The most distinguishable topographical feature is Cedar Creek Canyon, just north of the city proper near Huntertown. The Fort Wayne Moraine follows two of the city's three rivers: the St. Marys and St. Joseph. The two rivers converge to form the Maumee, which eventually empties into Lake Erie. Land east of the moraine includes the former Great Black Swamp, a lacustrine plain formed by Glacial Lake Maumee. The Little River flows southwest of Fort Wayne, a tributary of the Wabash River, and remnant of the Maumee Torrent. The city lies along the St. Lawrence Continental Divide which separates the Great Lakes Basin from the Gulf of Mexico watershed. The most important geographical feature of the area is the short distance overland between the Three Rivers system, which eventually flows to the Atlantic, and the Wabash system, which eventually flows to the Gulf of Mexico. This came to be the "portage" or carrying place, over which travelers could transport their cargoes from one system to the next. This natural crossroads attracted the Native Americans for thousands of years. It later attracted the European explorers and traders and the American pioneer settlers who continued to develop the area as a transportation and communications center. Chief Little Turtle of the Miami Nation expressed its importance eloquently at the treaty of Greenville in 1795 when he called it "that glorious gate...through which all the words of our chiefs had to pass through from north to south and from east to west." Fort Wayne's urban tree canopy is 29 percent, double the state average of 14.5 percent and above the national average of 27.1 percent. The canopy is decreasing, notably from development and the emerald ash borer infestation. Fort Wayne has been designated a Tree City USA since 1990. Historically, Fort Wayne has been divided into four unofficial quadrants: northeast, northwest, southeast, and southwest. Calhoun Street divides the southwest and southeast, while the St. Joseph River divides the northwest and northeast quadrants. The Maumee River separates the northeast and southeast, while portions of the St. Marys River and Chicago, Fort Wayne and Eastern Railroad separate the northwest and southwest quadrants. Fort Wayne's early-20th century development was influenced by the City Beautiful movement and centered on a "park and boulevard plan" conceived by urban planner Charles Mulford Robinson in 1909 and finalized by landscape architect George Kessler in 1912. The master plan proposed a network of parkways and boulevards connecting the city's three rivers and Spy Run Creek to dozens of neighborhoods and parks. Several parks were designed by noted landscape architect Arthur Asahel Shurcliff. Much of the original plan was implemented by 1955. In 2010, the Fort Wayne Park and Boulevard System was listed on the National Register of Historic Places, consisting of 11 public parks, four parkways, and ten boulevards, covering . During the 19th century, Fort Wayne was dominated by Greek Revival, Gothic Revival, and Italianate architecture. Examples of Greek Revival architecture remain in the city, with one being the Richardville House (1827), a National Historic Landmark. Gothic and Gothic Revival architecture can be found in some of the city's most prominent churches, including Trinity English Lutheran Church (1846), Cathedral of the Immaculate Conception (1860), Trinity Episcopal Church (1865), and Saint Paul's Evangelical Lutheran Church (1889). Popular early-20th century architectural styles found in the city include Queen Anne, Richardsonian Romanesque, Neoclassical, Colonial Revival, Dutch Colonial Revival, Tudor Revival, Prairie, American Craftsman, American Foursquare, and Art Deco. Richardsonian Romanesque buildings include Fort Wayne City Hall (1893) and John H. Bass Mansion (1902), each designed by Wing & Mahurin. Notable examples of Neoclassical architecture include the Masonic Temple (1926) and North Side High School (1927). Beaux-Arts, an architectural style closely related to Neoclassical, gained popularity during the City Beautiful movement of the 1890s and early 1900s, which is reflected in the Allen County Courthouse (1902). The Allen County Courthouse is one of two National Historic Landmarks in the city. The Pennsylvania Railroad Station, also known as Baker Street Station (1914), was designed in American Craftsman style. At , the Art Deco-style Lincoln Bank Tower was Fort Wayne's first high-rise and Indiana's tallest building from 1930 to 1962. The E. Ross Adair Federal Building and United States Courthouse (1932) is another example of Art Deco architecture. Williams–Woodland Park Historic District includes examples of Queen Anne and Colonial Revival residential homes, while the Forest Park Boulevard Historic District includes Tudor Revival homes. Modern and Postmodern architecture can be found in buildings constructed during the second half of the 20th century in Fort Wayne. The John D. Haynes House (1952) was designed by Frank Lloyd Wright, while the campus of Concordia Theological Seminary (1953) was designed by Eero Saarinen. Postmodern architect Michael Graves' first commissions were built in the city, including Hanselmann House (1967) and Snyderman House (1972, now demolished). Louis Kahn's design for the Arts United Center (1973) was inspired by a violin and its case. Other notable buildings include Indiana Michigan Power Center (1982), the tallest building in the city and tallest building in Indiana outside of Indianapolis, at . Fort Wayne lies in the humid continental climate zone (Köppen: "Dfa"), experiencing four distinct seasons. The city is located in USDA hardiness zones 5b and 6a. Summer is warm, humid, and wet. Winter is cold with moderate snowfall. The average annual precipitation is , recorded at Fort Wayne International Airport. During the winter season, snowfall accumulation averages per year. Lake-effect snow is not uncommon to the region, but usually appears in the form of light snow flurries. The National Weather Service reports the highest recorded temperature in the city at , most recently on June 28, 2012, and the lowest recorded temperature at on January 12, 1918. The wettest month on record was June 2015, with of precipitation. The greatest 24-hour rainfall was on August 1, 1926. The snowiest month on record was January 2014, with of snowfall. The greatest calendar-day snowfall was on February 28, 1900. Severe weather is not uncommon, particularly in the spring and summer months; the city experiences an average of 39 thunderstorm days and about 10 severe weather days annually. An EF2 tornado struck northern Fort Wayne on May 26, 2001, injuring three and causing damage along the Coliseum Boulevard corridor and a subdivision. Fort Wayne experienced wind gusts in the June 2012 North American derecho, knocking out power to 78,000, uprooting approximately 500 trees, and costing $2.5 million. According to the 2010 Census, there were 420,690 people and 113,541 households. The racial makeup of the city is 73.62% White, 15.41% Black or African American, 0.37% Native American or Alaska Native, 3.3% Asian (1.4% Burmese, 0.4% Indian, 0.3% Vietnamese, 0.2% Chinese, 0.2% Filipino, 0.1% Korean, 0.1% Laotian, 0.1% Thai), 0.06% Pacific Islander, 3.72% from other races, and 3.52% from two or more races. 7.96% of the population are Hispanic or Latino of any race. Among the Hispanic population, 6.1% are Mexican, 0.4% Puerto Rican, and 0.3% Guatemalan. Non-Hispanic Whites were 70.3% of the population in 2010, down from 87.7% in 1970. There were 101,585 households of which 30.1% had children under the age of 18 living with them, 42.3% were married couples living together, 14.8% had a female householder with no husband present, 4.9% had a male householder with no wife present, and 38.0% were non-families. 31.2% of all households were made up of individuals and 9.7% had someone living alone who was 65 years of age or older. The average household size was 2.44 and the average family size was 3.09. The median age in the city was 34.5 years. 26.4% of residents were under the age of 18; 10.2% were between the ages of 18 and 24; 26.5% were from 25 to 44; 24.9% were from 45 to 64; and 12% were 65 years of age or older. The gender makeup of the city was 48.4% male and 51.6% female. Fort Wayne has the largest Burmese American population in the U.S., estimated at 6,000. Burmese refugee settlement and "secondary migrants" doubled the city's Asian population between 2000 and 2010. Fort Wayne is sometimes referred to as the "City of Churches," an unofficial moniker dating to the late-19th century when the city was the regional hub of Catholic, Lutheran, and Episcopal faiths. Today, there are 360 churches in the city. 54 percent of Fort Wayne residents identify as religious, where 16 percent are Catholic, 9 percent are Lutheran, 6.5 percent are Baptist, 5 percent are Methodist, and 0.14 percent are Jewish, with 16.5 percent adhering to other Christian faiths. An increasing religious minority is found among the city's immigrant communities, including Buddhism, Hinduism, and Islam. Major churches include the Cathedral of the Immaculate Conception, Saint Paul's Evangelical Lutheran Church and Trinity Episcopal Church. Fort Wayne's Reform Judaism population is served by Congregation Achduth Vesholom, the oldest Jewish congregation in Indiana, founded in 1848. In 2013, construction began on the first Burmese Muslim mosque to be built worldwide since the mid-1970s. As of December 2012, four national Christian denominations were headquartered in the city: the American Association of Lutheran Churches, the Fundamental Baptist Fellowship Association, the Missionary Church and the Fellowship of Evangelical Churches. Fort Wayne is the seat of the Roman Catholic Diocese of Fort Wayne–South Bend, covering 14 counties in Northern Indiana, and the Lutheran Church–Missouri Synod Indiana District, encompassing all of Indiana and north central Kentucky. In 2017, the Fort Wayne metropolitan area had a gross domestic product (GDP) of $25.7 billion. The top four industries were: manufacturing ($8.1B), health care ($2.54B), retail trade ($1.4B), and finance and insurance ($1.3B) Government, if it had been a private industry, would have tied for third, generating $1.4 billion. Manufacturing is deeply rooted in Fort Wayne's economic history, dating to the earliest days of the city's growth as an important trade stop along the Wabash and Erie Canal. Railroads, introduced shortly after the canal's arrival, eased travel from Fort Wayne to other booming industrial centers along the Great Lakes, such as Chicago, Detroit, Toledo, and Cleveland. Throughout the early and mid-20th century, manufacturing dominated the city's economic landscape. From 1900 to 1930, Fort Wayne's industrial output expanded by 747 percent, with total production valued at $95 million in 1929, up from $11 million in 1899. The total workforce also increased from 18,000 in 1900 to nearly 50,000 in 1930. Companies that had a significant presence in the city include Dana Holding Corporation, Falstaff Brewing Corporation, Fruehauf Corporation, General Electric, International Harvester, Magnavox, Old Crown Brewing Corporation, and Tokheim, among several others, producing goods such as refrigerators, washing machines, automatic phonographs, meat packing products, televisions, garbage disposals, automotive parts and motors, trailers, gasoline pumps, trucks, beer, tents and awnings. Magnet wire production became an especially vital component to the city's economy. In 1960, Fort Wayne was at the center of the United States magnet wire industry, home to New Haven Wire and Cable Company, Phelps Dodge, Rea Magnet Wire, Superior Essex, and an operation at General Electric, producing nearly 90 percent of North America's magnet wire. The 1970s and 1980s were times of economic depression in Fort Wayne, when much of the city's manufacturing foundation eroded and the blue-collar workforce shrank. Fort Wayne joined several other cities reeling economically within the Rust Belt. At the same time, General Electric also downsized much of its more than 10,000-person workforce. Amid other area plant closures and downsizing, coupled with the early 1980s recession, the city lost 30,000 jobs and reached a 12.1 percent unemployment rate. The arrival of General Motors in 1987 helped fill the void from shuttered manufacturers and aided in the area's recovery, employing 3,000 at its Fort Wayne Assembly. In 2017, General Motors was the largest manufacturer in the city, employing 4,100 assembling Chevrolet Silverado regular and double cab light- and heavy-duty pickup trucks. Through the 1990s and into the 2000s, the city diversified its economy; manufacturing now employs 16.9 percent of Allen County's workforce. Other sectors include distribution, transportation, and logistics (23.1 percent), health care (17.9 percent), professional and business services (12.1 percent), leisure and hospitality (11.1 percent), and financial services (6.3 percent). The leisure and hospitality sector has especially grown, with 5.8 million visitors spending $545 million in 2013, a 4.3 percent increase over the previous year. The city is a center for the defense industry, employing thousands at such companies as BAE Systems (1,150), Harris Corporation (888), Raytheon Systems (950), and the Fort Wayne Air National Guard Station (423). Despite economic diversification, the city was significantly impacted by the Great Recession. According to a report from Pew Research Center, the city lost nearly a quarter of its manufacturing jobs and 11% of its economic status between 2000 and 2014. Economic Innovation Group's 2016 Distressed Communities Index Report ranked Fort Wayne among the most unequal large cities in the U.S. in terms of linking economic opportunities to its distressed zip codes. , Allen County's labor force was 180,637 with an unemployment rate of 2.5 percent. Companies based in Fort Wayne include Brotherhood Mutual, Do it Best, Franklin Electric, Frontier Communications – Central Region, Genteq, Indiana Michigan Power, K&K Insurance, MedPro Group, North American Van Lines, Rea Magnet Wire, Steel Dynamics, Sweetwater Sound, and Vera Bradley. Steel Dynamics is the only Fortune 500 company headquartered in the city, ranking 354th. Founded in 1905, Lincoln Financial Group was based in Fort Wayne until its move to suburban Philadelphia in 1999. The company maintains a large presence in the city, employing nearly 2,000. The Embassy Theatre is a 2,471-seat performing arts theater which hosts over 200,000 patrons annually. Since its founding in 1944, the Fort Wayne Philharmonic Orchestra has often been hosted at the Embassy. The University of Saint Francis Robert Goldstine Performing Arts Center, located on its Downtown Campus, contains a 2,086-seat auditorium. Since its establishment in 2010, the Cultural District has been home to several of the city's cultural institutions, including the Fort Wayne Museum of Art, Auer Center for Arts and Culture, Arts United Center, and Hall Community Arts Center. Arts United Center houses the Fort Wayne Civic Theater, Fort Wayne Dance Collective, and Fort Wayne Youtheatre. Auer Center for Arts and Culture houses Fort Wayne Ballet. Hall Community Arts Center houses Cinema Center, an independent film venue. Though used mainly for exhibitions and conventions, the Grand Wayne Convention Center hosts dance and choir productions, such as the annual Foundation for Art and Music in Education (FAME) Northeast Festival. Foellinger Theatre, a 2,500-seat amphitheater in Franke Park, hosts seasonal acts and outdoor concerts during warmer months. Located west of downtown, Arena Dinner Theatre is a nonprofit community arts corporation with a focus on live theater production, annually hosting seven full-length theatrical productions. The Fort Wayne Children's Zoo has been lauded as one of the nation's foremost zoos. Covering and containing 1,000 animals of 200 different species, the zoo is the largest regional attraction, regularly drawing over 500,000 visitors annually. The Foellinger-Freimann Botanical Conservatory gardens cover , displaying over 1,200 plants of 502 different species and 72 types of cacti. Science Central, an interactive science center, contains permanent displays and temporary exhibits, drawing 130,000 visitors annually. Established in 1921, the Fort Wayne Museum of Art (FWMoA) is accredited by the American Alliance of Museums, specializing in the collection and exhibition of American art. The FWMoA annually receives 100,000 visitors. The History Center, located in Fort Wayne's Old City Hall, manages a collection of more than 23,000 artifacts recalling the region's history. The center is overseen by the Allen County–Fort Wayne Historical Society, which maintains the Richardville House, one of two National Historic Landmarks in the city. Historic Fort Wayne, a replica of the 1815 fortification, hosts scheduled tours and historical reenactments throughout the year. Other cultural museums include the African/African–American Historical Museum, Fort Wayne Firefighters Museum, Greater Fort Wayne Aviation Museum, and Baer Field Heritage Air Park. The Allen County Public Library's Fred J. Reynolds Historical Genealogy Department is the second-largest genealogy collection in North America. The collection contains 350,000 printed volumes and 513,000 items of microfilm and microfiche. The city hosts a variety of cultural festivals and events annually. Festivals commemorating ethnic food, dance, music, and art include Germanfest, Greek Festival, and Japanese Cherry Blossom Festival. Initiated in 1997, Fort Wayne Pride celebrates northeast Indiana's LGBTQ community. BBQ RibFest showcases barbecue rib cooks and live entertainment, attracting 40,000 visitors annually. Fort4Fitness is a certified half marathon, run/walk, and health fair. Over 9,000 participated in the 2011 half marathon. In 2012, Fort4Fitness debuted a spring cycle, Bike-the-Fort, which included three bicycling tours with over 1,000 participants. HolidayFest begins with the Night of Lights on Thanksgiving eve, with the lighting of the PNC Santa and Reindeer, Wells Fargo Holiday Display, and Indiana Michigan Power Christmas Wreath, ending with a fireworks finale at Parkview Field. The largest annual events in the city are the Johnny Appleseed Festival, Taste of the Arts, Middlewaves and the Three Rivers Festival. The Johnny Appleseed Festival draws 300,000 visitors. The festival is held at Johnny Appleseed Park, where American folklore legend John Chapman is believed to be buried. Apple-themed cuisine, crafts, and historical demonstrations recalling 19th century American pioneering are among some of the festival's events. Three Rivers Festival, a celebration of Fort Wayne, spans nine days each July, attracting 400,000 visitors. Three Rivers features over 200 events, including a parade, midway, hot dog eating contest, bed race, raft race, arts fair, and fireworks spectacular. Other annual events include the Allen County Fair, BAALS Music Festival, National Soccer Festival, and the Vera Bradley Outlet Sale. Fort Wayne is home to three minor league sports franchises: the ECHL's Fort Wayne Komets, the Midwest League's Fort Wayne TinCaps, and the NBA G League's Fort Wayne Mad Ants who are owned and operated by their parent club, the Indiana Pacers. Fort Wayne also hosts the Fort Wayne Derby Girls of the Women's Flat Track Derby Association Division 2. These teams compete at the Allen County War Memorial Coliseum. Parkview Field is home to the TinCaps. The city has been home to other professional sports franchises, including the National Basketball Association's Fort Wayne Pistons (which moved to Detroit in 1957), the Fort Wayne Daisies of the All-American Girls Professional Baseball League, and the Fort Wayne Kekiongas of the National Association of Professional Base Ball Players (precursor to Major League Baseball). Intercollegiate sports in the city include the Purdue Fort Wayne Mastodons, representing Purdue University Fort Wayne (PFW) in the NCAA's Division I Horizon League, and NAIA schools Indiana Tech (Wolverine–Hoosier Athletic Conference) and University of Saint Francis (Crossroads League and Mid-States Football Association). The Mastodons had represented Indiana University – Purdue University Fort Wayne (IPFW) prior to its 2018 split into two separate institutions (see below), and from 2016 to 2018 were branded as the Fort Wayne Mastodons, but the athletic brand was changed to "Purdue Fort Wayne" shortly before the split took effect. Some notable events in sports history occurred in Fort Wayne. On June 2, 1883, Fort Wayne hosted the Quincy Professionals for one of the first lighted evening baseball games ever recorded. Fort Wayne is also credited as the birthplace of the NBA, as Pistons' coach Carl Bennett brokered the merger of the BAA and the NBL in 1948 from his Alexander Street home. On March 10, 1961, Wilt Chamberlain became the first player in the NBA to reach 3,000 points in a single season while competing at the War Memorial Coliseum. Fort Wayne Parks and Recreation maintains 86 public parks totaling . Three public and 20 private golf courses are located in Allen County. Franke Park is the most extensive city park, covering . Franke is home to the Foellinger Theatre, Shoaff Lake, and the Fort Wayne Children's Zoo. Other notable parks include Johnny Appleseed Park (home to a campground and John Chapman's grave) and McCulloch Park (home to Samuel Bigger's grave). Foellinger-Freimann Botanical Conservatory, Headwaters Park, Lawton Skatepark, and Historic Fort Wayne are located downtown. Hurshtown Reservoir, near Grabill, is the largest body of water in Allen County and is popular with watersports enthusiasts for sailing and fishing. Some 300 lakes are located within of the city. Located downtown along the St. Marys River, Fort Wayne Outfitters offers canoe, kayak, stand-up paddle board, and pontoon boat rentals for recreation along the three rivers. Starting in the 1970s, the city developed a system of recreational trails along the riverbanks, known as the Rivergreenway, with the aim of beautifying the riverfronts and promoting active lifestyles for residents. The Rivergreenway was designated a National Recreation Trail in 2009. As of 2018, the Rivergreenway had expanded with additional trails to encompass nearly throughout the city and county, with about 550,000 annual users. With the expansion of trails in recent years, cycling has become an emerging mode of transportation for residents. In 2009, the city's first bicycle lanes were established with the installation of 250 bike parking places. In 2016, Fort Wayne was designated a Bronze Level bicycle friendly community by the League of American Bicyclists. According to the Trust for Public Land's 2017 ParkScore Index, some 56% of Fort Wayne residents are underserved. Fort Wayne has a mayor–council government. The mayor, city clerk, and city council members serve four-year terms. Fort Wayne's mayor is Tom Henry, a Democrat, who was elected in 2007. Henry succeeded Democrat Graham Richard who chose not to run for re-election after two terms as mayor. Henry was re-elected to a third term in 2015. Karl Bandemer was appointed deputy mayor in 2013. Lana Keesling was elected city clerk in 2015. Fort Wayne City Council has nine elected members, one representative from each of the city's six council districts and three at-large members, serving four-year terms. The city is represented in the Indiana General Assembly by three Senate Districts and seven House Districts. Fort Wayne's state senators include Dennis Kruse (14th District), Liz Brown (15th), and David Long (16th). Representatives include Dan Leonard (50th District), Ben Smaltz (52nd), Phil GiaQuinta (80th), Martin Carbaugh (81st), Christopher Judy (83rd), Bob Morris (84th), and Dave Heine (85th). Federally, Fort Wayne is part of Indiana's 3rd congressional district, represented by Republican Jim Banks, who was first elected in 2016. Under the Unigov provision of Indiana Law, Fort Wayne would have automatically consolidated with Allen County when its population exceeded 250,000, previously the minimum population for a first class city in Indiana. Fort Wayne nearly met the state requirements for first class city designation on January 1, 2006 when of neighboring Aboite Township (and a small section of Wayne Township) including 25,094 people were annexed. However, a 2004 legislative change raised the population threshold for first-class status from 250,000 to 600,000, which ensured Indianapolis' status as the only first class city in Indiana. Fort Wayne's E. Ross Adair Federal Building and U.S. Courthouse houses the United States District Court for the Northern District of Indiana, which was authorized by Congress in 1928. Municipal and state laws are enforced by the Fort Wayne Police Department, an organization of 460 officers. In 2006, Fort Wayne's crime rate was 5104.1 per 100,000 people, slightly above the national average of 4479.3. There were 18 murders, 404 robberies, and 2,128 burglaries in 2006. Steve Reed was appointed to the position of police chief in 2016. In 2014, former police chief Rusty York was appointed to the position of director of public safety. York previously served as police chief from 2000 to 2014. As of 2010, the Fort Wayne Fire Department included 375 uniformed firefighters and 18 fire stations. Eric Lahey was appointed fire chief in 2014. Fort Wayne Community Schools (FWCS) is the largest public school district in Indiana, enrolling 30,981 students as of the 2013–2014 academic year. FWCS operate 51 facilities, including 31 elementary schools, ten middle schools, and five high schools. The student body is diverse, with 75 spoken languages in the district. East Allen County Schools (EACS) operate 20 schools, with a total enrollment of 9,114. Northwest Allen County Schools (NACS) operate seven elementary schools, two middle schools, and one high school, with a total enrollment of 6,853. Southwest Allen County Schools (SACS) operate six elementary schools, two middle schools, and one high school, with a total enrollment of 6,995. Private primary and secondary education is offered largely through Lutheran Schools of Indiana and the Roman Catholic Diocese of Fort Wayne–South Bend. Amish Parochial Schools of Indiana has schools through eighth grade in rural eastern Allen County. Fort Wayne hosts institutions affiliated with both of Indiana's major state university systems. Indiana University Fort Wayne (IU Fort Wayne) and Purdue University Fort Wayne (PFW) were established in July 2018 after the dissolution of Indiana University – Purdue University Fort Wayne (IPFW), which had enrolled over 13,000 students prior to its closure and was the state's fifth-largest public university. IPFW's degree programs in health sciences are now operated by IU Fort Wayne; as such, that institution is now home to the Fort Wayne Center for Medical Education, a branch of the Indiana University School of Medicine. All remaining IPFW degree programs were taken over by PFW. Ivy Tech Community College of Indiana also contains two campuses in the city. Three private universities are located in the city, including Concordia Theological Seminary, Indiana Institute of Technology, and the University of Saint Francis. Private universities with regional branches in Fort Wayne include Crossroads Bible College, Grace College and Theological Seminary, Huntington University, Indiana Wesleyan University, Manchester University College of Pharmacy, and Trine University. For-profit institutions include Harrison College and International Business College. Composed of 14 branches, the Allen County Public Library is among the 20 largest public libraries in the U.S., and ranks 89th factoring in academic libraries, with 3.4 million volumes. The library's foundation is also among the nation's largest, with $14 million in assets. The entire library system underwent an $84.1 million overhaul from 2002 to 2007. In 2009, over 7.4 million materials were borrowed by patrons, with over 3 million visits made throughout the library system. Major broadcasting network affiliates include WANE-TV (CBS), WPTA-TV (ABC/NBC), WISE-TV (CW), WFFT-TV (Fox), and WFWA-TV (PBS), Northeast Indiana's PBS member station. Religious broadcasters include WINM and W07CL. Access Fort Wayne maintains Fort Wayne and Allen County's Public Access capabilities serving from the Allen County Public Library. Two National Public Radio stations are based in the city, WBNI and WBOI with the new WELT Community Radio Station transmitting from the Allen County Public Library. Fort Wayne is served by two primary newspapers, the "Journal Gazette" and Pulitzer Prize-winning "News-Sentinel". The two dailies have separate editorial departments, but under a joint operating agreement, printing, advertising, and circulation are handled by Fort Wayne Newspapers, Inc. The "News-Sentinel" announced that it would cease printing operations in favor of digital publishing in August 2017." Fort Wayne includes two municipal airports, both managed by the Fort Wayne–Allen County Airport Authority. Fort Wayne International Airport (FWA) is the city's primary commercial airport, with five airlines offering direct service to 13 domestic connections. The airport is Indiana's second busiest, with over 350,000 passenger enplanements in 2015. Fort Wayne International is also home to the 122d Fighter Wing's Fort Wayne Air National Guard Station. Smith Field, in northern Fort Wayne, is used primarily for general aviation. Fort Wayne is served by a single Interstate, (Interstate 69), along with an auxiliary beltway (Interstate 469). Once the State Road 37 expressway between Bloomington and Martinsville is completed in 2018, filling a gap in I-69 that exists south of Indianapolis, the road will run south to Evansville; it currently runs north to the Canada–United States border at Port Huron, Michigan. In the coming years, I-69 will extend to the US–Mexico border in Texas, with branches ending in Laredo, Pharr, and Brownsville. Four U.S. Routes bisect the city, including US 24, US 27, US 30, and US 33. Five Indiana State Roads also meet in the city, including State Road 1, State Road 3, State Road 14, State Road 37, and State Road 930. Airport Expressway, a four-lane divided highway, links Fort Wayne International Airport directly to I-69. About 85 percent of residents commute alone by personal vehicle, while another eight percent carpool. Unlike most cities comparable to its size, Fort Wayne does not have an urban freeway system. In 1946, planners proposed a $27 million federally funded freeway, crossing east–west and north–south through downtown. Opponents successfully campaigned against the proposal, objecting to the demolition of nearly 1,500 homes at the time of the post-World War II housing shortage, while playing on fears that the project would force displaced minorities into white neighborhoods. In 1947, Fort Wayne residents voted down the referendum that would have allowed for its construction, dubbed the 'Anthony Wayne Parkway.' Beginning in 1962, construction commenced for I-69 in suburban Fort Wayne. The I-469 beltway around the southern and eastern fringes of Fort Wayne and New Haven was constructed between 1988 and 1995 as the largest public works project in Allen County history, at $207 million. Amtrak's "Capitol Limited" (Chicago - Toledo - Cleveland - Pittsburgh - Washington, D.C.) and Amtrak's "Lake Shore Limited" (Chicago - Toledo - Cleveland - Buffalo - Albany - split to Boston and to New York City) are the closest passenger rail services to Fort Wayne, located north at Waterloo Station. Service by Amtrak ended in 1990 when the "Broadway Limited" was rerouted away from Fort Wayne's Pennsylvania Station. Until 1961 the Pennsylvania Railroad operated the north–south "Northern Arrow" through the station. Other stations in Fort Wayne served the passenger trains of the Chicago, Indianapolis, and Louisville Railway ('Monon Railroad') and the Wabash Railroad. There has been a movement to bring direct passenger rail service back in the form of Amtrak or high-speed rail service. In 2013, a feasibility study was published outlining the impacts of a proposed Columbus—Fort Wayne—Chicago high-speed rail corridor. At , the route would cost $1.29 billion and generate some $7.1 billion in economic benefits to the region. Freight service is provided by a class I railroad (Norfolk Southern) and two class III railroads. Fort Wayne is headquarters and main operations hub of Norfolk Southern's Triple Crown Services subsidiary, the largest truckload shipper in the U.S. Fort Wayne's mass transit system is managed by the Fort Wayne Public Transportation Corporation (Citilink), providing 12 bus routes through the cities of Fort Wayne and New Haven via downtown's Central Station. CampusLink debuted in 2009 as a free shuttle service for students, faculty, and general public traveling between Ivy Tech's Coliseum and North campuses, IPFW and its student housing on the Waterfield Campus, and shopping and residential areas. MedLink debuted in 2013 connecting Parkview Regional Medical Center with Parkview Health's Randallia campus. Despite annual ridership of 2.2 million, less than one percent of residents commute by public transportation. Fort Wayne is served by two intercity bus providers: Greyhound Lines (Indianapolis—Toledo—Detroit) and Lakefront Lines (Chicago—Columbus—Akron). In 2016, the city introduced its first bike-sharing program, including five stations and 25 bicycles. Fort Wayne is served by ten medical centers belonging to one of two regional healthcare providers in the city: Parkview Health System and Lutheran Health Network. Notable hospitals include Dupont Hospital, Lutheran Hospital of Indiana, Parkview Regional Medical Center, Parkview Hospital Randallia, and St. Joseph Hospital. Over 1,600 patient beds are available throughout the city's healthcare system. , both healthcare systems were the city's first and second largest employers, respectively, and contribute to a total healthcare workforce in Allen County of 34,000. VA Northern Indiana Health Care System's Fort Wayne Campus provides medical services through the Department of Veterans Affairs. City Utilities is the largest municipally owned water utility in Indiana, supplying residents with of water per day from the St. Joseph River via the Three Rivers Water Filtration Plant. Sanitary sewer treatment is also managed by City Utilities. The City of Fort Wayne offers full curbside recycling and solid waste collection services for residents, presently contracted through Red River Waste Solutions. Electricity is provided by Indiana Michigan Power, a subsidiary of American Electric Power, while natural gas is supplied by Northern Indiana Public Service Company (NIPSCO), a subsidiary of NiSource. All tier 1 networks and several additional telecommunication service providers cover the Fort Wayne rate area. Fort Wayne has four sister cities as designated by Sister Cities International: Friendship city
https://en.wikipedia.org/wiki?curid=11232
Fart (word) Fart is a word in the English language most commonly used in reference to flatulence that can be used as a noun or a verb. The immediate roots are in the Middle English words "ferten", "feortan" and "farten", kin of the Old High German word "ferzan". Cognates are found in Old Norse, Slavic and also Greek and Sanskrit. The word "fart" has been incorporated into the colloquial and technical speech of a number of occupations, including computing. It is often considered unsuitable in formal situations as it may be considered vulgar or offensive. The English word "fart" is one of the oldest words in the English lexicon. Its Indo-European origins are confirmed by the many cognate words in some other Indo-European languages: It is cognate with Greek verb πέρδομαι (perdomai), as well as the Latin "pēdĕre", Sanskrit "pardate", Avestan "pərəδaiti", Italian "fare un peto", French "péter", Russian пердеть (perdet') and Polish "pierd" « PIE *perd [break wind loudly] or *pezd [the same, softly], all of which mean the same thing. Like most Indo-European roots in the Germanic languages, it was altered under Grimm's law, so that Indo-European "/p/" > "/f/", and "/d/" > "/t/", as the German cognate "furzen" also manifests. In certain circles the word is considered merely a common profanity with an often humorous connotation. For example, a person may be referred to as a 'fart', or an 'old fart', not necessarily depending on the person's age. This may convey the sense that a person is boring or unduly fussy and be intended as an insult, mainly when used in the second or third person. For example, '"he's a boring old fart!" However the word may be used as a colloquial term of endearment or in an attempt at humorous self-deprecation (e.g., in such phrases as "I know I'm just an old fart" or "you do like to fart about!"). 'Fart' is often only used as a term of endearment when the subject is personally well known to the user. In both cases though, it tends to refer to personal habits or traits that the user considers to be a negative feature of the subject, even when it is a self-reference. For example, when concerned that a person is being overly methodical they might say 'I know I'm being an old fart', potentially to forestall negative thoughts and opinions in others. When used in an attempt to be offensive, the word is still considered vulgar, but it remains a mild example of such an insult. This usage dates back to the Medieval period, where the phrase 'not worth a fart' would be applied to an item held to be worthless. Whilst being accused of farting may be offensive, other phrasal descriptions of flatulent discharge, in phrases such as, "Have you dropped your guts?", "Have you blown one off" or "Who shit?" are certainly no less so. The word "fart" in Middle English occurs in "Sumer Is Icumen In", where one sign of summer is "bucke uerteþ" (the buck farts). It appears in several of Geoffrey Chaucer's "Canterbury Tales". In "The Miller's Tale", Absolon has already been tricked into kissing Alison's buttocks when he is expecting to kiss her face. Her boyfriend Nicholas hangs his buttocks out of a window, hoping to trick Absolon into kissing his buttocks in turn and then farts in the face of his rival. In "The Summoner's Tale", the friars in the story are to receive the smell of a fart through a twelve-spoked wheel. In the early modern period, the word "fart" was not considered especially vulgar; it even surfaced in literary works. For example, Samuel Johnson's "A Dictionary of the English Language", published in 1755, included the word. Johnson defined it with two poems, one by Jonathan Swift, the other by Sir John Suckling. Benjamin Franklin prepared an essay on the topic for the Royal Academy of Brussels in 1781 urging scientific study. In 1607, a group of Members of Parliament had written a ribald poem entitled "The Parliament Fart", as a symbolic protest against the conservatism of the House of Lords and the king, James I. While not one of George Carlin's original seven dirty words, he noted in a later routine that the word "fart" ought to be added to "the list" of words that were not acceptable (for broadcast) in any context (which have non-offensive meanings), and described television as (then) a "fart-free zone". Thomas Wolfe had the phrase "a fizzing and sulphuric fart" cut out of his 1929 work "Look Homeward, Angel" by his publisher. Ernest Hemingway, who had the same publisher, accepted the principle that "fart" could be cut, on the grounds that words should not be used purely to shock. The hippie movement in the 1970s saw a new definition develop, with the use of "fart" as a personal noun, to describe a "detestable person, or someone of small stature or limited mental capacity", gaining wider and more open usage as a result. Rhyming slang developed the alternative form "raspberry tart", later shortened to "raspberry", and occasionally abbreviated further to "razz". This was associated with the phrase "blowing a raspberry". The word has become more prevalent, and now features in children's literature, such as the "Walter the Farting Dog" series of children's books, Robert Munsch's "Good Families Don't" and "The Gas We Pass" by Shinta Cho.
https://en.wikipedia.org/wiki?curid=11236
FA Cup The FA Cup, known officially as The Football Association Challenge Cup, is an annual knockout football competition in men's domestic English football. First played during the 1871–72 season, it is the oldest national football competition in the world. It is organised by and named after The Football Association (The FA). Since 2015, it has been known as The Emirates FA Cup after its headline sponsor. A concurrent women's tournament is also held, the FA Women's Cup. The competition is open to any eligible club down to Level 10 of the English football league system – all 92 professional clubs in the Premier League (Level 1) and the English Football League (Levels 2 to 4), and several hundred non-League teams in Steps 1 to 6 of the National League System (Levels 5 to 10). A record 763 clubs competed in 2011–12. The tournament consists of 12 randomly drawn rounds followed by the semi-finals and the final. Entrants are not seeded, although a system of byes based on league level ensures higher ranked teams enter in later rounds – the minimum number of games needed to win, depending on which round a team enters the competition, ranges from six to fourteen. The first six rounds are the Qualifying Competition, from which 32 teams progress to the first round of the Competition Proper, meeting the first of the 48 professional teams from Leagues One and Two. The last entrants are the Premier League and Championship clubs, into the draw for the Third Round Proper. In the modern era, only one non-League team has ever reached the quarter-finals, and teams below Level 2 have never reached the final. As a result, significant focus is given to the smaller teams who progress furthest, especially if they achieve an unlikely "giant-killing" victory. Winners receive the FA Cup trophy, of which there have been two designs and five actual cups; the latest is a 2014 replica of the second design, introduced in 1911. Winners also qualify for the Europa League and a place in the FA Community Shield match. Manchester City are the current holders, having beaten Watford 6–0 in the 2019 final. Arsenal are the most successful club with 13 titles. Arsène Wenger is the most successful manager in the history of the competition, having won seven finals as manager of Arsenal. In 1863, the newly founded Football Association (the FA) published the Laws of the Game of Association Football, unifying the various different rules in use before then. On 20 July 1871, in the offices of "The Sportsman" newspaper, the FA Secretary C. W. Alcock proposed to the FA committee that "it is desirable that a Challenge Cup should be established in connection with the Association for which all clubs belonging to the Association should be invited to compete". The inaugural FA Cup tournament kicked off in November 1871. After thirteen games in all, Wanderers were crowned the winners in the final, on 16 March 1872. Wanderers retained the trophy the following year. The modern cup was beginning to be established by the 1888–89 season, when qualifying rounds were introduced. Following the 1914–15 edition, the competition was suspended due to the First World War, and did not resume until 1919–20. The 1923 FA Cup Final, commonly known as the "White Horse Final", was the first final to be played in the newly opened Wembley Stadium (known at the time as the Empire Stadium). The 1927 final saw "Abide with Me" being sung for the first time at the Cup final, which has become a pre-match tradition. Due to the outbreak of World War II, the competition was not played between the 1938–39 and 1945–46 editions. Due to the wartime breaks, the competition did not celebrate its centenary year until 1980–81; fittingly the final featured a goal by Ricky Villa which was later voted the greatest goal ever scored in an FA Cup Final, but has since been replaced by Steven Gerrard. Having previously featured replays, the modern day practice of ensuring the semi-final and final matches finish on the day was introduced in 2000. Redevelopment of Wembley saw the final played outside of England for the first time, the 2001–2006 finals being played at the Millennium Stadium in Cardiff. The final returned to Wembley in 2007, followed by the semi-finals from 2008. The competition is open to any club down to Level 10 of the English football league system which meets the eligibility criteria. All clubs in the top four levels (the Premier League and the three divisions of the Football League) are automatically eligible. Clubs in the next six levels (non-league football) are also eligible provided they have played in either the FA Cup, FA Trophy or FA Vase competitions in the previous season. Newly formed clubs, such as F.C. United of Manchester in 2005–06 and also 2006–07, may not therefore play in the FA Cup in their first season. All clubs entering the competition must also have a suitable stadium. It is very rare for top clubs to miss the competition, although it can happen in exceptional circumstances. Manchester United did not defend their title in 1999–2000, as they were already in the inaugural Club World Championship. The club stated that entering both tournaments would overload their fixture schedule and make it more difficult to defend their Champions League and Premier League titles. The club claimed that they did not want to devalue the FA Cup by fielding a weaker side. The move benefited United as they received a two-week break and won the 1999–2000 league title by an 18-point margin, although they did not progress past the group stage of the Club World Championship. The withdrawal from the FA Cup, however, drew considerable criticism as this weakened the tournament's prestige and Sir Alex Ferguson later admitted his regret regarding their handling of the situation. Welsh sides that play in English leagues are eligible, although since the creation of the League of Wales there are only five clubs remaining: Cardiff City (the only non-English team to win the tournament, in 1927), Swansea City, Newport County, Wrexham, and Merthyr Town. In the early years other teams from Wales, Ireland and Scotland also took part in the competition, with Glasgow side Queen's Park losing the final to Blackburn Rovers in 1884 and 1885 before being barred from entering by the Scottish Football Association. In the 2013–14 season the first Channel Island club entered the competition when Guernsey F.C. competed. The number of entrants has increased greatly in recent years. In the 2004–05 season, 660 clubs entered the competition, beating the long-standing record of 656 from the 1921–22 season. In 2005–06 this increased to 674 entrants, in 2006–07 to 687, in 2007–08 to 731 clubs, and for the 2008–09 and 2009–10 competitions it reached 762. The number has varied slightly but remained roughly stable since then, with 759 clubs participating in 2010–11, a record 763 in 2011–12, 758 for 2012–13, 737 for 2013–14 and 736 for 2014–15. By comparison, the other major English domestic cup, the League Cup, involves only the 92 members of the Premier League and Football League. Beginning in August, the competition proceeds as a knockout tournament throughout, consisting of twelve rounds, a semi-final and then a final, in May. A system of byes ensures clubs above Level 9 and 10 enter the competition at later stages. There is no seeding, the fixtures in each round being determined by a random draw. Prior to the fifth round, fixtures ending in a tie are replayed once only. The first six rounds are qualifiers, with the draws organised on a regional basis. The next six rounds are the "proper" rounds where all clubs are in one draw. Entrants from the bottom two levels (9 and 10) begin the competition in the Extra Preliminary Round. Clubs from higher levels are then added in later rounds, as per the table below. The months in which rounds are played are traditional, with exact dates subject to each year's calendar. The qualifying rounds are regionalised to reduce the travel costs for smaller non-league sides. The first and second proper rounds were also previously split into Northern and Southern sections, but this practice was ended after the 1997–98 competition. The final is normally held the Saturday after the Premier League season finishes in May. The only seasons in recent times when this pattern was not followed were 1999–2000, when most rounds were played a few weeks earlier than normal as an experiment, and 2010–11 and 2012–13 when the FA Cup Final was played before the Premier League season had finished, to allow Wembley Stadium to be ready for the UEFA Champions League final, as well as in 2011–12 to allow England time to prepare for that summer's European Championships. The draws for the Extra-Preliminary, Preliminary, and First Qualifying Rounds used to all occur at the same time. Thereafter, the draw for each subsequent round is not made until after the scheduled dates for the previous round, meaning that in the case of replays, clubs will often know their future opponents in advance. This season 2016/17 the draw for the 1st qualifying round was drawn at a later date as per previous season's later rounds. The draw for each of the proper rounds is broadcast live on television, usually taking place at the conclusion of live coverage of one of the games of the previous round. Public interest is particularly high during the draw for the third round, which is where the top-ranked teams are added to the draw. In rounds up to and including the Fourth Round Proper, fixtures resulting in a draw (after normal time) go to a replay, played at the venue of the away team, at a later date; if that replay is still tied, the winner is settled by a period of extra time, and if still necessary, a penalty shootout. Since 2016–17, ties have been settled on the day from the quarter-finals onwards, using extra time and penalties. From 2018–19, Fifth Round ties are also settled by extra time and penalties. Until 1990–91, further replays would be played until one team was victorious. Some ties took as many as six matches to settle; in their 1975 campaign, Fulham played a total of 12 games over six rounds, which remains the most games played by a team to reach a final. Replays were traditionally played three or four days after the original game, but from 1991–92 they were staged at least 10 days later on police advice for the rounds proper. This led to penalty shoot-outs being introduced, the first of which came on 26 November 1991 when Rotherham United eliminated Scunthorpe United. From 1980–81 to 1998–99, the semi-finals went to extra time on the day if the score after 90 minutes was a draw. If the score was still level after extra time, the match would go to a replay. Replays for the semi-finals were scrapped for 1999–2000, the last semi-final to go into a replay was in 1998–99 when Manchester United beat Arsenal 2–1 after extra time. The first game had ended in a 0–0 draw. The first FA Cup Final to go to extra time and a replay was the 1875 final, between the Royal Engineers and the Old Etonians. The initial tie finished 1–1 but the Royal Engineers won the replay 2–0 in normal time. The last replayed final was the 1993 FA Cup Final, when Arsenal and Sheffield Wednesday fought a 1–1 draw. The replay saw Arsenal win the FA Cup, 2–1 after extra time. The last quarter-final to go to a replay was Manchester United vs West Ham United in the 2015–16 FA Cup. The original game at Old Trafford ended in a 1–1 draw, while Manchester United won the replay at the Boleyn Ground, 2–1. It was also the last FA Cup game ever played at the Boleyn Ground. The last fifth round replay saw Tottenham Hotspur defeat Rochdale 6–1 at Wembley in the 2017–18 FA Cup after the first match at Spotland Stadium ended in a 2–2 draw. The FA Cup winners qualify for the following season's UEFA Europa League (formerly named the UEFA Cup; from its launch in 1960 until 1998, they entered the now-defunct UEFA Cup Winners' Cup instead). This European place applies even if the team is relegated or is not in the English top flight. In the past, if the FA Cup winning team also qualified for the following season's Champions League or Europa League through their league or European performance, then the losing FA Cup finalist were given the European berth of the League Cup winners and the League Cup winners will be given the league berth instead (in the Cup Winners' Cup era, teams qualifying for the UEFA Cup via other competitions will be promoted to the Cup Winners' Cup instead). FA Cup winners enter the Europa League at the group stage. Losing finalists, if they haven't qualified for Europe via the league, began earlier, at the play-off or third qualifying round stage. From the 2015–16 UEFA Europa League season, however, UEFA does not allow the runners-up to qualify for the Europa League through the competition. If the winner — and until 2015, the runner-up — has already qualified for Europe through their league position (with the exception of the UEFA Cup until 1998), the FA Cup berth is then given to the highest-place team in the league who has not yet qualified. The FA Cup winners also qualify for the following season's single-match FA Community Shield, the traditional season opener played against the previous season's Premier League champions (or the Premier League runners-up if the FA Cup winners also won the league – the double). Fixtures in the 12 rounds of the competition are usually played at the home ground of one of the two teams. The semi-finals and final are played at a neutral venue – the rebuilt Wembley Stadium (since 2008 and 2007 respectively). In the matches for the twelve competition rounds, the team who plays at home is decided when the fixtures are drawn – simply the first team drawn out for each fixture. Occasionally games may have to be moved to other grounds due to other events taking place, security reasons or a ground not being suitable to host popular teams. In the event of a draw, the replay is played at the ground of the team who originally played away from home. In the days when multiple replays were possible, the second replay (and any further replays) were played at neutral grounds. The clubs involved could alternatively agree to toss for home advantage in the second replay. The semi-finals have been played exclusively at the rebuilt Wembley Stadium since 2008, one year after it opened and after it had already hosted a final (in 2007). For the first decade of the competition, the Kennington Oval was used as the semi-final venue. In the period between this first decade and the reopening of Wembley, semi-finals were played at high-capacity neutral venues around England; usually the home grounds of teams not involved in that semi-final, chosen to be roughly equidistant between the two teams for fairness of travel. The top three most used venues in this period were Villa Park in Birmingham (55 times), Hillsborough in Sheffield (34 times) and Old Trafford in Manchester (23 times). The original Wembley Stadium was also used seven times for semi-final, between 1991 and 2000 (the last held there), but not always for fixtures featuring London teams. In 2005, both were held at the Millennium Stadium. In 2003 the FA took the decision to permanently use the new Wembley for semi-finals to recoup debts in financing the new stadium. This was controversial, with the move seen as both unfair to fans of teams located far from London, as well as taking some of the prestige away from a Wembley final. In defending the move, the FA has also cited the extra capacity Wembley offers, although the 2013 fixture between Millwall and Wigan led to the unprecedented step of placing 6,000 tickets on sale to neutral fans after the game failed to sell out. A fan poll by "The Guardian" in 2013 found 86% opposition to Wembley semi-finals. The final has been played at the rebuilt Wembley Stadium since it opened, in 2007. The rebuilding process meant that between 2001 and 2006 they were hosted at the Millennium Stadium in Cardiff in Wales. Prior to rebuilding, the final was hosted by the original Wembley Stadium since it opened in 1923 (being originally named the Empire Stadium). One exception to this 78 year series of Empire Stadium finals (including five replays) was the 1970 replay between Leeds and Chelsea, held at Old Trafford in Manchester. In the 51 years prior to the Empire Stadium opening, the final (including 8 replays) was held in a variety of locations, predominantly in London, and mainly at the Kennington Oval and then Crystal Palace. It was played 22 times at The Oval (the inaugural competition in 1872, and then all but two times until 1892). After The Oval, Crystal Palace hosted 21 finals from 1895 to 1914, broken up by four replays elsewhere. The other London venues were Stamford Bridge from 1920 to 1922 (the last three finals before the move to Empire Stadium); and the University of Oxford's Lillie Bridge in Fulham for the second ever final, in 1873. The other venues used sparingly in this period were all outside of London, as follows: The FA permitted artificial turf (3G) pitches in all rounds of the competition from the 2014–15 edition and beyond. Under the 2015–16 rules, the pitch must be of FIFA One Star quality, or Two Star for ties if they involve one of the 92 professional clubs. This followed approval two years previously for their use in the qualifying rounds only – if a team with a 3G pitch progressed to the competition proper, they had to switch their tie to the ground of another eligible entrant with a natural grass pitch. Having been strong proponents of the surface, the first match in the proper rounds to be played on a 3G surface was a televised first round replay at Maidstone United's Gallagher Stadium on 20 November 2014. The eventual winners of the competition receive the FA Cup; it is only loaned to the club by the FA, under the current (2015–16) rules it must be returned by 1 March, or earlier if given seven days’ notice. Traditionally, the holders had the Cup until the following year's presentation, although more recently the trophy has been taken on publicity tours by the FA in between finals. The trophy comes in three parts – the cup itself, plus a lid and a base. There have been two designs of trophy in use, but five physical trophies have been presented. The original trophy, known as the "little tin idol", was 18 inches high and made by Martin, Hall & Co. It was stolen in 1895 and never recovered, and so was replaced by an exact replica, used until 1910. The FA decided to change the design after the 1909 winners, Manchester United, made their own replica, leading the FA to realise they did not own the copyright. This new, larger design was by Fattorini and Sons, and was used from 1911. In order to preserve this original, from 1992 it was replaced by an exact replica, although this had to be replaced after just over two decades, after showing wear and tear from being handled more than in previous eras. This third replica, first used in 2014, was built heavier to withstand the increased handling. Of the four surviving trophies, only the 1895 replica has entered private ownership. The name of the winning team is engraved on the silver band around the base as soon as the final has finished, in order to be ready in time for the presentation ceremony. This means the engraver has just five minutes to perform a task which would take 20 under normal conditions, although time is saved by engraving the year on during the match, and sketching the presumed winner. During the final, the trophy is decorated with ribbons in the colours of both finalists, with the loser's ribbons being removed at the end of the game. The tradition of tying ribbons started after Tottenham Hotspur won the 1901 FA Cup Final and the wife of a Spurs director decided to tie blue and white ribbons to the handles of the cup. Traditionally, at Wembley finals, the presentation is made at the Royal Box, with players, led by the captain, mounting a staircase to a gangway in front of the box and returning by a second staircase on the other side of the box. At Cardiff the presentation was made on a podium on the pitch. The tradition of presenting the trophy immediately after the game did not start until the 1882 final; after the first final in 1872 the trophy was not presented to the winners, Wanderers, until a reception held four weeks later in the Pall Mall Restaurant in London. Under the original rules, the trophy was to be permanently presented to any club which won the competition three times, although when inaugural winners Wanderers achieved this feat by the 1876 final, the rules were changed by FA Secretary CW Alcock (who was also captain of Wanderers in their first victory). Portsmouth F.C. have the distinction of being the football club which has held the FA Cup trophy for the longest uninterrupted period - seven years. Portsmouth had defeated Wolverhampton Wanderers 4–1 in the 1939 FA Cup Final and were awarded the trophy as 1938-1939 FA Cup Champions. But with the outbreak of World War II in September 1939, the regular Football League and FA Cup competitions for the 1939–40 season were cancelled for the duration of the war. Portsmouth's manager Jack Tinn was rumoured to have kept the FA Cup trophy 'safe under his bed' throughout the duration of the war, but this is an urban myth. Because the naval city of Portsmouth was a primary strategic military target for German Luftwaffe bombing, the FA Cup trophy was actually taken ten miles to the north of Portsmouth, to the nearby Hampshire village of Lovedean, and there it resided in a quaint thatched roof country pub called "The Bird in Hand" for the seven years of the war. After the conclusion of World War II, the FA Cup trophy was presented back to the Football Association by Portsmouth F.C. in time for the 1946 FA Cup Final. The first trophy, the 'little tin idol', was made by Martin, Hall & Co at a cost of £20. It was stolen from a Birmingham shoe shop window belonging to William Shillcock while held by Aston Villa on 11 September 1895 and was never seen again. Despite a £10 reward for information, the crime was never solved. As it happened while it was in their care, the FA fined Villa £25 to pay for a replacement. Just over 60 years later, 80 year old career criminal Henry (Harry) James Burge claimed to have committed the theft, confessing to a newspaper, with the story being published in the "Sunday Pictorial" newspaper on 23 February 1958. He claimed to have carried out the robbery with two other men, although when discrepancies with a contemporaneous report in the "Birmingham Post" newspaper (the crime pre-dated written police reports) in his account of the means of entry and other items stolen, detectives decided there was no realistic possibility of a conviction and the case was closed. Burge claimed the cup had been melted down to make counterfeit half-crown coins, which matched known intelligence of the time, in which stolen silver was being used to forge coins which were then laundered through betting shops at a local racecourse, although Burge had no past history of forgery in a record of 42 previous convictions for which he had spent 42 years in prison. He had been further imprisoned in 1957 for seven years for theft from cars. Released in 1961, he died in 1964. After being rendered obsolete by the redesign, the 1895 replica was presented in 1911 to the FA's long-serving president Lord Kinnaird. Kinnaird died in 1923, and his family kept it in their possession, out of view, until putting it up for auction in 2005. It was sold at Christie's auction house on 19 May 2005 for £420,000 (£478,400 including auction fees and taxes). The sale price set a new world record for a piece of football memorabilia, surpassing the £254,000 paid for the Jules Rimet World Cup Trophy in 1997. The successful bidder was David Gold, the then joint chairman of Birmingham City; claiming the FA and government were doing nothing proactive to ensure the trophy remained in the country, Gold stated his purchase was motivated by wanting to save it for the nation. Accordingly, Gold presented the trophy to the National Football Museum in Preston on 20 April 2006, where it went on immediate public display. It later moved with the museum to its new location in Manchester. In November 2012, it was ceremonially presented to Royal Engineers, after they beat Wanderers 7–1 in a charity replay of the first FA Cup final. The redesigned trophy first used in 1911 was larger at 61.5 cm (24.2 inches) high, and was designed and manufactured by Fattorini's of Bradford, coincidentally being won by Bradford City in its first outing. On the 27 March 2016 episode of the BBC television programme Antiques Roadshow, this trophy was valued at £1 million by expert Alastair Dickenson, although he suggested that, due to the design featuring depictions of grapes and vines, it may not have been specifically produced for the FA, but was instead an off-the-shelf design originally meant to be a wine or champagne cooler. A smaller, but otherwise identical, replica was also made by Fattorini, the North Wales Coast FA Cup trophy, and is contested annually by members of that regional Association. The 1992 replica was made by Toye, Kenning and Spencer. A copy of this trophy was also produced, in case anything happened to the primary trophy. The 2014 replica was made by Thomas Lyte, handcrafted in sterling 925 silver over 250 hours. A weight increase for greater durability has taken it to . Each club in the final receives 40 winners or runners-up medals to be distributed among players, staff, and officials. Since the start of the 1994–95 season, the FA Cup has been sponsored. However, to protect the identity of the competition, the sponsored name has always included 'The FA Cup' in addition to the sponsor's name, unlike sponsorship deals for the League Cup where the word 'cup' is preceded by only the sponsor's name. Sponsorship deals run for four years, though – as in the case of E.ON – one-year extensions may be agreed. Emirates Airline is the sponsor from 2015 to 2018, renaming the competition as 'The Emirates FA Cup', unlike previous editions, which included 'The FA Cup in association with E.ON' and 'The FA Cup with Budweiser'. From 2006 to 2013, Umbro supplied match balls for all FA Cup matches. They were replaced at the start of the 2013–14 season by Nike, who produced the competition's official match ball for five seasons. Mitre took over for the 2018–19 season, beginning a three-year partnership with the FA. The possibility of unlikely victories in the earlier rounds of the competition, where lower ranked teams beat higher placed opposition in what is known as a "giant killing" or "cupset", is much anticipated by the public. Such upsets are considered an integral part of the tradition and prestige of the competition, and the attention gained by giant-killing teams can be as great as that for winners of the cup. Almost every club in the League Pyramid has a fondly remembered giant-killing act in its history. It is considered particularly newsworthy when a top Premier League team suffers an upset defeat, or where the giant-killer is a non-league club, i.e. from outside the professional levels of The Football League. One analysis of four years of FA Cup results showed that it was 99.85 per cent likely that at least one team would beat one from its next higher division in a given year. The probability drops to 48.8 per cent for a two-division gap, and 39.28 per cent for a three-division gap. The Football League was founded in 1888, 16 years after the first FA Cup competition. Since the creation of the Football League, Tottenham Hotspur is the only non-League side to win the Cup, taking the 1901 FA Cup with a victory over reigning League runners-up Sheffield United. At that time, the Football League consisted of two divisions with a combined total of 36 clubs, mostly teams from Northern England and the Midlands. Spurs competed in the Southern Football League, which ran parallel to the Football League, and were champions. Only two other non-League clubs have even reached the final since the founding of the League: Sheffield Wednesday in 1890 (champions of the Football Alliance, a rival league which was already effectively the tier below the League, which it formally became in 1892 upon formation of the Second Division — Wednesday being let straight into the First Division — and the Southern League's Southampton in 1900 and 1902. Upon the Football League's expansion and creation of the Third Division for 1920–21, all the clubs in the Southern League First Division clubs transferred over and it has been since placed below the League in the English football league system, with the National League sandwiching the two since 1979–80. The most recent examples of a non-league team (Levels 5 to 10) beating a Level 1 opponent are National League side Lincoln City's away victory over Premier League side Burnley in the 2016–17 FA Cup and Conference Premier side Luton Town's away victory over Level 1 Premier League's Norwich City in the 2012–13 Fourth Round Proper. Prior to that game, the last time a non-league side defeated a Level 1 club was in 1989 when Sutton United claimed a 2–1 victory at home over Coventry City, who had won the FA Cup less than two years prior. In the 1971–72 FA Cup, a non-league side achieved a Level 1 giant killing that was voted "best FA Cup tie ever" in a 2007 poll by "The Observer" newspaper. Non-league Hereford United was trailing First Division Newcastle United 0–1 with less than seven minutes left in the Third Round Proper replay, when Hereford's Ronnie Radford scored the equalizer – a goal still shown regularly when FA Cup fixtures are broadcast. Hereford finished the shocking comeback by defeating Newcastle 2–1 in the match. Some small clubs gain a reputation for being "cup specialists" after two or more giant killing feats within a few years. Yeovil Town holds the record for the most victories over league opposition as a non-league team, having recorded 20 wins through the years before it achieved promotion into The Football League. The record for a club which has never entered The Football League is held by Altrincham, with 17 wins against league teams. For non-League teams, reaching the Third Round Proper – where all Level 1 sides now enter – is considered a major achievement. In the 2008–09 FA Cup, a record eight non-League teams achieved this feat. As of the 2016–17 season, only nine non-League teams have reached the Fifth Round Proper (final 16) since 1945, and only Lincoln City have progressed to the Sixth Round (final 8), during the 2016–17 edition of the tournament. Chasetown, while playing at Level 8 of English football during the 2007–08 competition, are the lowest-ranked team to ever play in the Third Round Proper (final 64, of 731 teams entered that season). Chasetown was then a member of the Southern League Division One Midlands (a lower level within the Southern Football League), when they lost to Football League Championship (Level 2) team Cardiff City, the eventual FA Cup runners-up that year. Their success earned the lowly organisation over £60,000 in prize money. In games between League sides, one of the most notable results was the 1992 victory by Wrexham, bottom of the previous season's League (avoiding relegation due to expansion of The Football League), over reigning champions Arsenal. Another similar shock was when Shrewsbury Town beat Everton 2–1 in 2003. Everton finished 7th in the Premier League and Shrewsbury Town were relegated to the Football Conference that same season. Since its establishment, the FA Cup has been won by 43 different teams. Teams shown in "italics" are no longer in existence. Additionally, Queen's Park ceased to be eligible to enter the FA Cup after a Scottish Football Association ruling in 1887. Four clubs have won consecutive FA Cups on more than one occasion: Wanderers (1872, 1873 and 1876, 1877, 1878), Blackburn Rovers (1884, 1885, 1886 and 1890, 1891), Tottenham Hotspur (1961, 1962 and 1981, 1982) and Arsenal (2002, 2003 and 2014, 2015). The record for most winner's medals for a manager is held by Arsène Wenger, who has won seven titles with Arsenal (1998, 2002, 2003, 2005, 2014, 2015, 2017). Manchester City (2019) are the only club to have achieved a domestic League and Cup treble, having beaten Chelsea 4–3 on penalties in the EFL Cup Final, finished at the top of the English Premier League, and beaten Watford 6–0 in the FA Cup Final. Seven clubs have won the FA Cup as part of a League and Cup double, namely Preston North End (1889), Aston Villa (1897), Tottenham Hotspur (1961), Arsenal (1971, 1998, 2002), Liverpool (1986), Manchester United (1994, 1996, 1999) and Chelsea (2010). In 1993, Arsenal became the first side to win both the FA Cup and the League Cup in the same season when they beat Sheffield Wednesday 2–1 in both finals. Liverpool (in 2001), Chelsea (in 2007) and Manchester City (in 2019) have since repeated this feat. In 2012, Chelsea accomplished a different cup double consisting of the FA Cup and the 2012 Champions League. In 1998–99, Manchester United added the 1999 Champions League title to their league and cup double to complete a unique Treble. Two years later, in 2000–01, Liverpool won the FA Cup, League Cup and UEFA Cup to complete a cup treble. The FA Cup has only been won by a non-English team once. Cardiff City achieved this in 1927 when they beat Arsenal in the final at Wembley. They had previously made it to the final only to lose to Sheffield United in 1925 and lost another final to Portsmouth in 2008. Cardiff City is also the only team to win the national cups of two different countries in the same season, having also won the Welsh Cup in 1927. The Scottish team Queen's Park reached and lost the final in both 1884 and 1885. Since the creation of the Football League in 1888, the final has never been contested by two teams from outside the top division, and there have only been eight winners who were not in the top flight: Notts County (1894); Tottenham Hotspur (1901); Wolverhampton Wanderers (1908); Barnsley (1912); West Bromwich Albion (1931); Sunderland (1973), Southampton (1976) and West Ham United (1980). With the exception of Tottenham, these clubs were all playing in the second tier (the old Second Division) – Tottenham were playing in the Southern League and were only elected to the Football League in 1908, meaning they are the only non-League winners of the FA Cup since the League's creation. Other than Tottenham's victory, only 24 finalists have come from outside English football's top tier, with a record of 7 wins and 17 runners-up: and none at all from the third tier or lower, Southampton (1902, then in the Southern League) being the last finalist from outside the top two tiers. Sunderland's win in 1973 was considered a major upset, having beaten Leeds United who finished third in the top flight that season., as was West Ham's victory over Arsenal in 1980 as the Gunners were in their third successive FA Cup Final and were also the cup holders from the previous year as well as just having finished 4th in the First Division, whereas West Ham had ended the season 7th in Division 2. This also marked the last time (as of 2018-19) a team from outside the top division won the FA Cup. Uniquely, in 2008 three of the four semi-finalists (Barnsley, Cardiff City and West Bromwich) were from outside the top division, although the eventual winner was the last remaining top-flight team, Portsmouth. West Bromwich (1931) are the only team to have won the FA Cup and earned promotion to the top flight in the same season; whereas Wigan Athletic (2013) are the only team to have won the Cup and been relegated from the top flight in the same season. The FA Cup Final is one of 10 events reserved for live broadcast on UK terrestrial television under the Ofcom Code on Sports and Other Listed and Designated Events. In the early years of coverage the BBC had exclusive radio coverage with a picture of the pitch marked in the Radio Times with numbered squares to help the listener follow the match on the radio. The first FA Cup Final on Radio was in 1926 between Bolton Wanderers and Manchester City but this was only broadcast in Manchester, the first national final on BBC Radio was between Arsenal and Cardiff in 1927. The first final on BBC Television was in 1937 in a match which featured Sunderland and Preston North End but this was not televised in full. The following season's final between Preston and Huddersfield was covered in full by the BBC. When ITV was formed in 1955 they shared final coverage with the BBC in one of the only club matches shown live on television, during the 1970s and 1980s coverage became more elaborate with BBC and ITV trying to steal viewers from the others by starting coverage earlier and earlier some starting as early as 9 a.m. which was six hours before kick off. The sharing of rights between BBC and ITV continued from 1955 to 1988, when ITV lost coverage to the new Sports Channel which later became Sky Sports. From 1988 to 1997, the BBC and Sky Sports had coverage of the FA Cup, the BBC had highlights on Match of the Day and usually one match per round while Sky had the same deal. From 1997 to 2001, ITV and Sky shared live coverage with both having two matches per round and BBC continuing with highlights on Match of the Day. From 2002 to 2008, BBC and Sky again shared coverage with BBC having two or three matches per round and Sky having one or two. From 2008–09 to 2013–14, FA Cup matches are shown live by ITV across England and Wales, with UTV broadcasting to Northern Ireland but STV refusing to show them. ITV shows 16 FA Cup games per season, including the first pick of live matches from each of the first to sixth rounds of the competition, plus one semi-final exclusively live. The final is also shown live on ITV. Under the same 2008 contract, Setanta Sports showed three games and one replay in each round from round three to five, two quarter-finals, one semi-final and the final. The channel also broadcast ITV's matches exclusively to Scotland, after the ITV franchise holder in Scotland, STV, decided not to broadcast FA Cup games. Setanta entered administration in June 2009 and as a result the FA terminated Setanta's deal to broadcast FA-sanctioned competitions and England internationals. As a result of Setanta going out of business ITV showed the competition exclusively in the 2009–10 season with between three and four matches per round, all quarter finals, semi-finals and final live as the FA could not find a pay TV broadcaster in time. ESPN bought the competition for the 2010–11 to 2012–13 season and during this time Rebecca Lowe became the first woman to host the FA Cup Final in the UK. In October 2009, The FA announced that ITV would show an additional match in the First and Second Rounds on ITV, with one replay match shown on ITV4. One match and one replay match from the first two rounds will broadcast on The FA website for free, in a similar situation to the 2010 World Cup Qualifier between Ukraine and England. The 2009–10 First Round match between Oldham Athletic and Leeds United was the first FA Cup match to be streamed online live. Many expected BSkyB to make a bid to show some of the remaining FA Cup games for the remainder of the 2009–10 season which would include a semi-final and shared rights to the final. ESPN took over the package Setanta held for the FA Cup from the 2010–11 season. The 2011 final was also shown live on Sky 3D in addition to ESPN (who provided the 3D coverage for Sky 3D) and ITV. Following the sale of ESPN's UK and Ireland channels to BT, ESPN's rights package transferred to BT Sport from the 2013–14 season. BBC Radio 5 Live and Talksport provides radio coverage including several full live commentaries per round, with additional commentaries broadcast on BBC Local Radio. Until the 2008–09 season, the BBC and Sky Sports shared television coverage, with the BBC showing three matches in the earlier rounds. Some analysts argued the decision to move away from the Sky and, in particular, the BBC undermined the FA Cup in the eyes of the public. The early rounds of the 2008–09 competition were covered for the first time by ITV's online service, ITV Local. The first match of the competition, between Wantage Town and Brading Town, was broadcast live online. Highlights of eight games of each round were broadcast as catch up on ITV Local. Since ITV Local closed, this coverage did not continue. ITV lost the rights to the FA Cup beginning with the 2014–15 FA Cup, terrestrial rights will return to BBC Sport, with the final being shown on BBC One while BT Sport hold the pay TV rights. Under this deal, the BBC will show around the same number of games as ITV and still having the first pick for each round. Matches involving Welsh clubs are sometimes exclusively broadcast on Welsh language channel S4C, which is also available to view across the rest of the United Kingdom on satellite and cable television, and through the channel's website. A similar arrangement is shared with BBC Cymru Wales from 2014–15, potentially giving the BBC an extra match per round. On the 23rd May 2019, it was announced that ITV would replace BT Sport in broadcasting the FA Cup from the 2021-22 season, this new deal will see BBC and ITV become joint broadcasters of the tournament, this will mean for the very first time that all FA Cup matches would be all be exclusively broadcast on free-to-air television. The FA sells overseas rights separately from the domestic contract.
https://en.wikipedia.org/wiki?curid=11237
Fenway Park Fenway Park is a baseball park located in Boston, Massachusetts, near Kenmore Square. Since 1912, it has been the home for the Boston Red Sox, the city's American League baseball team, and since 1953, its only Major League Baseball (MLB) franchise. While the stadium was built in 1912, it was rebuilt in 1934. It is the oldest active ballpark in MLB. Because of its age and constrained location in Boston's dense Fenway–Kenmore neighborhood, the park has been renovated or expanded many times, resulting in quirky features including "The Triangle," Pesky's Pole, and the Green Monster in left field. It is the fifth-smallest among MLB ballparks by seating capacity, second-smallest by total capacity, and one of eight that cannot accommodate at least 40,000 spectators. Fenway has hosted the World Series 11 times, with the Red Sox winning six of them and the Boston Braves winning one. Besides baseball games it has been the site of many other sporting and cultural events including professional football games for the Boston Redskins, Boston Yanks, and the New England Patriots; concerts; soccer and hockey games (such as the 2010 NHL Winter Classic); and political and religious campaigns. April 20, 2012, marked Fenway Park's centennial. On March 7 of that year, the park was added to the National Register of Historic Places. Former pitcher Bill Lee has called Fenway Park "a shrine". It is a pending Boston Landmark which will regulate any further changes to the park. Today, the park is considered to be one of the most well-known sports venues in the world. The Red Sox moved to Fenway Park from the old Huntington Avenue Grounds. In 1911, owner John I. Taylor purchased the land bordered by Brookline Avenue, Jersey Street, Van Ness Street and Lansdowne Street and developed it into a larger baseball stadium. Taylor claimed the name Fenway Park came from its location in the Fenway neighborhood of Boston, which was partially created late in the nineteenth century by filling in marshland or "fens", to create the Back Bay Fens urban park. However, given that Taylor's family also owned the Fenway Realty Company, the promotional value of the naming at the time has been cited as well. Like many classic ballparks, Fenway Park was constructed on an asymmetrical block, with consequent asymmetry in its field dimensions. The park was designed by architect James E. McLaughlin, and the General Contractor was the Charles Logue Building Company. The first game was played April 20, 1912, with mayor John F. Fitzgerald throwing out the first pitch and Boston defeating the New York Highlanders, 7–6 in 11 innings. Newspaper coverage of the opening was overshadowed by continuing coverage of the "Titanic" sinking a few days earlier. In June 1919 a rally supporting Irish Independence turned out nearly 50,000 supporters to see the President of the Irish Republic, Éamon de Valera, and it was allegedly the largest crowd ever in the ballpark. Fenway Park had historically drawn low attendance, its lowest occurring late in the 1965 season with two games having paid attendance under 500 spectators. Its attendance has risen since the Red Sox' 1967 "Impossible Dream" season, and on September 8, 2008, with a game versus the Tampa Bay Rays, Fenway Park broke the all-time Major League record for consecutive sellouts with 456, surpassing the record previously held by Jacobs Field in Cleveland. On Wednesday, June 17, 2009, the park celebrated its 500th consecutive Red Sox sellout. According to WBZ-TV, the team joined three NBA teams which achieved 500 consecutive home sellouts. The sellout streak ended on April 11, 2013; in all the Red Sox sold out 794 regular season games and an additional 26 postseason games during this streak. The park's address was originally 24 Jersey Street. In 1977, the section of Jersey Street nearest the park was renamed Yawkey Way in honor of longtime Red Sox owner Tom Yawkey, and the park's address was 4 Yawkey Way until 2018, when the street's name was reverted to Jersey Street. The address is now 4 Jersey Street. Some of the changes include: On May 15, 1999, then Red Sox CEO John Harrington announced plans for a new Fenway Park to be built near the existing structure. It was to have seated 44,130 and would have been a modernized replica of the current Fenway Park, with the same field dimensions except for a shorter right field and reduced foul territory. Some sections of the existing ballpark were to be preserved (mainly the original Green Monster and the third base side of the park) as part of the overall new layout. Most of the current stadium was to be demolished to make room for new development, with one section remaining to house a baseball museum and public park. The proposal was highly controversial; it projected that the park had less than 15 years of usable life, would require hundreds of millions of dollars of public investment, and was later revealed to be part of a scheme by current ownership to increase the marketable value of the team as they were ready to sell. Several groups (such as "Save Fenway Park") formed in an attempt to block the move. Discussion took place for several years regarding the new stadium proposal. One plan involved building a "Sports Megaplex" in South Boston, where a new Fenway would be located next to a new stadium for the New England Patriots. The Patriots ultimately built Gillette Stadium in Foxborough, their home throughout most of their history, which ended the Megaplex proposal. The Red Sox and the city of Boston failed to reach an agreement on building the new stadium, and in 2005, the Red Sox ownership group announced that the team would stay at Fenway Park indefinitely. The stadium has since been renovated, and will remain usable until as late as 2061. The park is located along Lansdowne Street and Jersey Street in the Kenmore Square area of Boston. The area includes many buildings of similar height and architecture and thus it blends in with its surroundings. When pitcher Roger Clemens arrived in Boston for the first time in 1984, he took a taxi from Logan Airport and was sure the driver had misunderstood his directions when he announced their arrival at the park. Clemens recalled telling the driver "No, Fenway Park, it's a baseball stadium ... this is a warehouse." Only when the driver told Clemens to look up and he saw the light towers did he realize he was in the right place. Fenway Park is one of the two remaining jewel box ballparks still in use in Major League Baseball (the other being Wrigley Field), and both have a significant number of obstructed view seats, due to pillars supporting the upper deck. These are sold as such, and are a reminder of the architectural limitations of older ballparks. George Will asserts in his book "" that Fenway Park is a "hitters' ballpark", with its short right-field fence (302 feet), narrow foul ground (the smallest of any current major league park), and generally closer-than-normal outfield fences. By Rule 1.04, Note(a), all parks built after 1958 have been required to have foul lines at least long and a center-field fence at least from home plate. Regarding the narrow foul territory, Will writes: Will states that some observers might feel that these unique aspects of Fenway give the Red Sox an advantage over their opponents, given that the Red Sox hitters play 81 games at the home stadium while each opponent plays no more than nine games as visiting teams but Will does not share this view. Fenway Park's bullpen wall is much lower than most other outfield walls; outfielders are known to end up flying over this wall when chasing balls hit that direction, such as with Torii Hunter when chasing a David Ortiz game-tying grand slam that direction in Game 2 of the 2013 ALCS. The Green Monster is the nickname of the left field wall in the park. It is located from home plate; this short distance often benefits right-handed hitters. Part of the original ballpark construction of 1912, the wall is made of wood, but was covered in tin and concrete in 1934 when the scoreboard was added. The wall was covered in hard plastic in 1976. The scoreboard is manually updated throughout the game. Despite the name, the Green Monster was not painted green until 1947; before that it was covered with advertisements. The "Monster" designation is relatively new; for most of its history it was simply called "the wall." In 2003, terrace-style seating was added on top of the wall. "The Triangle" is a region of center field where the walls form a triangle whose far corner is from home plate. That deep right-center point is conventionally given as the center field distance. The true center is unmarked, from home plate, to the left of "The Triangle" when viewed from home plate. There was once a smaller "triangle" at the left end of the bleachers in center field, posted as . The end of the bleachers form a right angle with the Green Monster and the flagpole stands within that little triangle. That is not the true power alley, but deep left-center. The true power alley distance is not posted. The foul line intersects with the Green Monster at nearly a right angle, so the power alley could be estimated at , assuming the power alley is 22.5° away from the foul line as measured from home plate. "Williamsburg" was the name, invented by sportswriters, for the bullpen area built in front of the right-center field bleachers in 1940. It was built there primarily for the benefit of Ted Williams, to enable him and other left-handed batters to hit more home runs, since it was closer than the bleacher wall. The lone red seat in the right field bleachers (Section 42, Row 37, Seat 21) signifies the longest home run ever hit at Fenway. The home run, hit by Ted Williams on June 9, 1946, was officially measured at – well beyond "Williamsburg". According to Hit Tracker Online, the ball, if unobstructed, would have flown . The ball landed on Joseph A. Boucher, penetrating his large straw hat and hitting him in the head. A confounded Boucher was later quoted as saying, There have been other home runs hit at Fenway that have contended for the distance title. In the 2007 book "The Year Babe Ruth Hit 104 Home Runs", researcher Bill Jenkinson found evidence that on May 25, 1926, Babe Ruth hit one in the pre-1934 bleacher configuration which landed five rows from the top in right field. This would have placed it at an estimated from home plate. On June 23, 2001, Manny Ramirez hit one that struck a light tower above the Green Monster, which would have cleared the park had it missed. The park's official estimate placed the home run one foot short of Williams' record at . An April 2019 home run by Rowdy Tellez of the Toronto Blue Jays was initially reported as , but later found to be significantly shorter, approximately . Pesky's Pole is the name for the pole on the right field foul line, which stands from home plate, the shortest outfield distance (left or right field) in Major League Baseball. Like the measurement of the left-field line at Fenway Park, this has been disputed. Aerial shots show it to be noticeably shorter than the (actual) 302 foot line in right field, and Pesky has been quoted as estimating it to be "around 295 feet". Despite the short wall, home runs in this area are relatively rare, as the fence curves away from the foul pole sharply. The pole was named after Johnny Pesky, a light-hitting shortstop and long-time coach for the Red Sox, who hit some of his six home runs at Fenway Park around the pole but never off the pole. Pesky (playing 1942 to 1952, except for 1943 to 1945) was a contact hitter who hit just 17 home runs in his career (6 at Fenway Park). It's not known how many of these six actually landed near the pole. The Red Sox gives credit to pitcher and Sox broadcaster Mel Parnell for coining the name. The most notable for Pesky is a two-run homer in the eighth inning of the 1946 Opening Day game to win the game. According to Pesky, Mel Parnell named the pole after Pesky won a game for Parnell in with a home run down the short right field line, just around the pole. However, Pesky hit just one home run in a game pitched by Parnell, a two-run shot in the first inning of a game against Detroit played on June 11, 1950. The game was eventually won by the visiting Tigers in the 14th inning on a three-run shot by Tigers right fielder Vic Wertz and Parnell earned a no-decision that day. The term, though it had been in use since the 1950s, became far more common when Parnell became a Red Sox broadcaster in 1965. Mark Bellhorn hit what proved to be the game-winning home run off of Julián Tavárez, in Game 1 of the 2004 World Series off that pole's screen. On September 27, 2006, on Pesky's 87th birthday, the Red Sox organization officially dedicated the right field foul pole as Pesky's Pole with a commemorative plaque placed at its base. The seat directly on the foul side of Pesky's Pole in the front row is Section 94, Row E, Seat 5 and is usually sold as a lone ticket. In a ceremony before the Red Sox's 2005 game against the Cincinnati Reds, the pole on the left field foul line atop the Green Monster was named the Fisk Foul Pole, or "Pudge's Pole", in honor of Carlton Fisk. Fisk provided one of baseball's most enduring moments in Game 6 of the 1975 World Series against the Reds. Facing Reds right-hander Pat Darcy in the 12th inning with the score tied at 6, Fisk hit a long fly ball down the left field line. It appeared to be heading foul, but Fisk, after initially appearing unsure of whether or not to continue running to first base, famously jumped and waved his arms to the right as if to somehow direct the ball fair. It ricocheted off the foul pole, winning the game for the Red Sox and sending the series to a seventh and deciding game the next night, which Cincinnati won. Like Johnny Pesky's No. 6, Carlton had his No. 27 player number retired by the team. From 1912 to 1933, there was a high incline in front of the then -high left field wall at Fenway Park, extending from the left-field foul pole to the center field flag pole. As a result, a left fielder had to play part of the territory running uphill (and back down). Boston's first star left fielder, Duffy Lewis, mastered the skill so well that the area became known as "Duffy's Cliff". The incline served two purposes: it was a support for a high wall and it was built to compensate for the difference in grades between the field and Lansdowne Street on the other side of that wall. The wall also served as a spectator-friendly seating area during the dead ball era when overflow crowds, in "front" of the later Green Monster, would sit on the incline behind ropes. As part of the 1934 remodeling of the ballpark, the bleachers, and the wall itself, Red Sox owner Tom Yawkey arranged to flatten the ground along the base of the wall, so that Duffy's Cliff no longer existed. The base of the left field wall is several feet below the grade level of Lansdowne Street, accounting for the occasional rat that might spook the scoreboard operators. There has been debate as to the true left field distance, which was once posted as . A reporter from "The Boston Globe" was able to sneak into Fenway Park and measure the distance. When the paper's evidence was presented to the club in 1995, the distance was remeasured by the Red Sox and restated at . The companion sign remained unchanged until 1998, when it was corrected to . In 1983, private suites were added to the roof behind home plate. In 1988, 610 stadium club seats enclosed in glass and named the "600 Club", were added above the home plate grandstand replacing the existing press box. The press box was then added to the top of the 600 Club. The 1988 addition has been thought to have changed the air currents in the park to the detriment of hitters. In 2002, the organization renamed the club seats the ".406 Club" (in honor of Ted Williams' batting average in 1941). Between the 2005 and 2006 seasons the existing .406 club was rebuilt as part of the continuing ballpark expansion efforts. The second deck now features two open-air levels: the bottom level is the new "Dell EMC Club" featuring 406 seats and concierge services and the upper level, the State Street Pavilion, has 374 seats and a dedicated standing room area. The added seats are wider than the previous seats. In 1990, Mike Rutstein started handing out the first issue of "Boston Baseball Magazine" (originally called "Baseball Underground") outside of the park. He was frustrated with the quality of the program being sold inside the park, which also came out once every two months. The program was sold for $1, half the cost of the programs inside the park. To sell the program, Rutstein's employees would stand outside the park wearing bright red shirts and greet fans by holding a program up and shouting "Program, Scorecard, One Dollar!". By 1992, the Red Sox organization filed complaints with the city code enforcement arguing that the scorecard inside the magazine was not covered under the First Amendment protecting magazines and that Rutstein's employees were operating on the streets without a permit. Despite a lot of attention in the news, Rutstein said the charges were not pursued and no further legal action was taken. In 2012, one of Rutstein's long time employees Sly Egidio quit Boston Baseball to start "The Yawkey Way Report" named after Yawkey Way. By that time, Boston Baseball was selling for $3 per program, $2 cheaper than the in-park programs selling for $5. The Yawkey Way Report cost $1 and Egidio stationed his hawkers close to Boston Baseball's hawkers, starting a "hawker war." The Yawkey Way Report also came with baseball cards, ponchos and tote bags, which caused Rutstein to file his own complaints with Boston city code enforcement. Despite the rivalry, both programs continue to be hawked outside of Fenway Park and are often the first thing fans see when they approach the stadium on game-day. The Red Sox's one-time cross-town rivals, the Boston Braves used Fenway Park for the 1914 World Series and the 1915 season until Braves Field was completed; ironically, the Red Sox would then use Braves Field – which had a much higher seating capacity – for their own World Series games in 1915 and 1916. Neil Diamond's "Sweet Caroline" has been played at Fenway Park since at least 1997, and in the middle of the eighth inning at every game since 2002. On opening night of the 2010 season at Fenway Park, the song was performed by Diamond himself. Since 1990 (except in 2005 when, because of field work, it was held in a minor league ballpark, and 2020, as the tournament was cancelled due to the COVID-19 pandemic), Fenway Park has also hosted the final round of a Boston-area intercollegiate baseball tournament called the Baseball Beanpot, an equivalent to the more well-known hockey Beanpot tourney. The teams play the first rounds in minor league stadiums before moving on to Fenway for the final and a consolation game. Boston College, Harvard University, Northeastern University, and the University of Massachusetts Amherst compete in the four-team tournament. Beginning in 2006, the Red Sox have hosted the "Futures at Fenway" event, where two of their minor-league affiliates play a regular-season doubleheader as the "home" teams. Before the Futures day started, the most recent minor-league game held at Fenway had been the Eastern League All-Star Game in 1977. The 2009 Atlantic Coast Conference Baseball Tournament was scheduled to be held at Fenway Park, but a scheduling conflict caused the 2010 tournament to be scheduled at Fenway Park instead. Due to economic reasons, the ACC elected to move the 2010 tournament from Fenway Park to NewBridge Bank Park in Greensboro, North Carolina, but is still looking to host a tournament at Fenway Park in the future. From 1970 to 1987, the Cape Cod Baseball League (CCBL) played its annual all-star game at various major league stadiums. The games were interleague contests between the CCBL and the Atlantic Collegiate Baseball League (ACBL). The 1975, 1977, 1979, 1981, 1983, 1985 and 1987 games were played at Fenway. The MVP of the 1977 contest was future major league slugger Steve Balboni, who clobbered two home runs over the Green Monster that day. The CCBL returned to Fenway in 2009, 2010 and 2011 for its intraleague all-star game matching the league's East and West divisions. The 2009 game starred East division MVP and future Boston Red Sox Chris Sale of Florida Gulf Coast University. The CCBL also holds an annual workout day at Fenway where CCBL players are evaluated by major league scouts. On October 9, 1920, Fenway Park was the site of the first open-air boxing show in Boston. The card featured four bouts. Although Eddie Shevlin and Paul Doyle fought in the feature bout, Daniel J. Saunders of the "Boston Daily Globe" described heavyweights Battling McCreery and John Lester Johnson as "the only boxers who caused any excitement". McCreery, who according to Saunders, "was to take a flop in five rounds", won by judge's decision in ten rounds. After the fight, Johnson punched McCreery while McCreery was trying to shake his hand. McCreery then knocked Johnson out of the ring and hit him over the head with his chair. The card drew 5,000 spectators (half of what was expected) and brought in $6,100 (several thousand less than what was promised to the fighters). In 1928, New England Welterweight Champion Al Mello headlined three cards at Fenway. He defeated Billy Murphy in front of a crowd of 12,000 on June 26, Charlie Donovan on August 31, and Murphy again on September 13. On July 2, 1930, future World Heavyweight Champion James J. Braddock made his debut in that weight class. He defeated Joe Monte in ten rounds. On September 2, 1930, Babe Hunt defeated Ernie Schaaf in what "The Boston Daily Globe" described as a "dull bout" and a "big disappointment". The undercard included future light heavyweight champion George Nichols, who defeated Harry Allen of Brockton, Massachusetts in ten rounds. In 1932, Eddie Mack promoted ten cards at Fenway Park. The August 2 card featured World Light Heavyweight Champion Maxie Rosenbloom defeating Joe Barlow of Roxbury and Taunton' Henry Emond defeating The Cocoa Kid. On August 23, Dave Shade defeated Norman Conrad of Wilton, New Hampshire in front of 3,500 attendees. The September 6 card was headlined by World junior lightweight champion Kid Chocolate, who defeated Steve Smith. On June 25, 1936, former world heavyweight champion Jack Sharkey defeated Phil Brubaker in what would be his final career victory. In 1937, Rip Valenti and the Goodwin Athletic Club promoted five cards at Fenway. Three of these were headlined by New England Heavyweight Champion Al McCoy. On June 16 McCoy defeated Natie Brown in front of a crowd of 4,516. On July 29 he knocked out Jack McCarthy in the third round. On August 24 he and Tony Shucco fought to a draw. Future WBA featherweight champion Sal Bartolo fought one of his first professional fights on the May 24 undercard. On June 25, 1945, Tami Mauriello knocked out Lou Nova in 2:47. An estimated crowd of 8,000 was in attendance. On July 12, 1954, Tony DeMarco knocked out George Araujo 58 seconds into the fifth round in front of 12,000 spectators. The most recent boxing event at Fenway took place on June 16, 1956. The undercard consisted of Eddie Andrews vs. George Chimenti, Bobby Courchesne vs. George Monroe for the New England Lightweight Championship, and Barry Allison vs. Don Williams for the New England Middleweight Championship. In the main event, Tony DeMarco defeated Vince Martinez by decision. An estimated 15,000 were in attendance - far below promoter Sam Silverman's expectations. On October 17, 1925, the Boston Soccer Club and the Fall River Marksmen of the American Soccer League played a scoreless tie before 4,000 fans. Boston also hosted the Providence Clamdiggers and Indiana Flooring at Fenway later that season. On June 18, 1928, Boston played Rangers F.C. to a 2–2 tie in front of a crowd of 10,000. In 1929, Boston hosted two more matches at Fenway Park; a 3–2 victory over the New Bedford Whalers on August 10 and a 3–2 loss to Fall River on August 17. On May 30, 1931; 8,000 fans were on hand to see the American Soccer League champion New York Yankees defeat Celtic 4–3. The Yankees goalkeeper, Johnny Reder, would later return to play for the Boston Red Sox. During 1968, the park was home to the Boston Beacons of the now-defunct NASL. On July 21, 2010 Fenway hosted an exhibition game between European soccer clubs Celtic F.C. and Sporting C.P. in an event called "Football at Fenway". A crowd of 32,162 watched the two teams play to a 1–1 draw. Celtic won 6–5 on penalty shoot out, winning the first "Fenway football challenge Trophy". Recent matches have taken place between Liverpool, an English Premier League club owned by Fenway Sports Group, and A.S. Roma, an Italian Serie A club owned by FSG partner Thomas R. DiBenedetto. The July 25, 2012 match ended in a 2–1 win for AS Roma before a crowd of 37,169. AS Roma also won the rematch on July 23, 2014 by a score of 1–0. On July 21, 2019 Liverpool returned to Fenway for a preseason match against Sevilla, the Spanish team won 2–1 at the end of full-time. Football has been played at Fenway since at least 1916; In 1926, the first American Football League's Boston Bulldogs played at both Fenway and Braves Field; the Boston Shamrocks of the second AFL did the same in 1936 and 1937. The National Football League's Boston Redskins played at Fenway for four seasons (1933–1936) after playing their inaugural season in 1932 at Braves Field as the Boston Braves. The Boston Yanks played there in the 1940s; and the American Football League's Boston Patriots called Fenway Park home from 1963 to 1968 after moving there from Nickerson Field. At various times in the past, Dartmouth College, Boston College, Brown University, and Boston University teams have also played football games at Fenway Park. Boston College and Notre Dame played a game at Fenway in 2015 as part of Notre Dame's Shamrock Series. The annual Harvard Yale game in November 2018 was played at Fenway. Beginning in 2020, a new bowl game at Fenway called the Fenway Bowl will pit a team from the Atlantic Coast Conference against a team from the American Athletic Conference. The third annual NHL Winter Classic was held at Fenway on New Year's Day in 2010. The Boston Bruins beat the Philadelphia Flyers 2–1 in sudden-death overtime, securing the first home-team victory in the relatively short history of the NHL Winter Classic. The Winter Classic paved the way for the Frozen Fenway series of ice skating and hockey events at the ballpark. Frozen Fenway is an annual series of collegiate and amateur games featuring ice hockey teams from local and regional high schools, colleges, and universities, including the University of Connecticut, University of Massachusetts Amherst, University of New Hampshire, University of Maine, University of Vermont, University of Massachusetts Lowell, Northeastern University, Boston College, and Boston University, and is held during the first part of the event. After the completion of the hockey series, the rink was opened to the public for free ice skating. Fenway has hosted Gaelic games over the years. On June 6, 1937, the All-Ireland Football Champions from County Mayo defeated a Massachusetts team, 17–8, and on November 8, 1954, the All-Ireland Hurling champions County Cork beat an American line-up, 37–28. In more recent times, the Fenway Hurling Classic for the Players Champions Cup has been staged, first in November 2015 when Galway beat Dublin, and subsequently in November 2017 and November 2018. Fenway has been home to several concerts beginning in 1973 when Stevie Wonder and Ray Charles first played there. No further concerts were played there until 2003 when Bruce Springsteen and the E Street Band played a leg of their The Rising Tour. Since 2003, there has been at least one concert every year at Fenway by such artists as Bruce Springsteen and the E Street Band, Jimmy Buffett, Billy Joel, The Rolling Stones, Neil Diamond, The Police, Jason Aldean, Dave Matthews Band, Tom Petty & The Heartbreakers, Aerosmith, Phish, Roger Waters, Paul McCartney, James Taylor, Pearl Jam, Foo Fighters, and Dead & Company. In 2017, Lady Gaga brought her Joanne World Tour to the stadium, making her the first woman to headline a concert there. In 2019, The Who played their first ever show at the stadium with the Boston Symphony Orchestra. Polartec Big Air At Fenway is the first big air snowboarding and skiing competition that was held on February 11–12, 2016. This event was part of the U.S. Grand Prix Tour and the International Ski Federation's World Tour. Notable winter athletes that competed are Ty Walker, Sage Kotsenburg, and Joss Christensen. The big air jump was constructed to be about tall, standing above the lights of the stadium. Frank Fallon was the first public address (PA) announcer for the Red Sox, and held the job from 1953 to 1957. Fred Cusick, better known for his career of announcing Boston Bruins hockey games, joined him in 1956 and also left after 1957. Jay McMaster took over in 1958, until his replacement by Sherm Feller in 1967. Feller served as the announcer for 26 years until his death after the 1993 season. He was known for beginning his games by welcoming the fans with "Good afternoon, ladies and gentlemen, boys and girls. Welcome to Fenway Park", and ending them by saying "Thank you." Leslie Sterling took the job for the 1994 season, becoming the second female PA announcer in the history of Major League Baseball. Ed Brickley took over in 1997, and was replaced by Carl Beane in 2003. Beane was regarded as an "iconic" announcer, and served until his death in 2012, which was caused by a heart attack suffered while driving. Fenway used a series of guest announcers to finish the 2012 season before hiring its current announcers: Henry Mahegan, Bob Lobel, and Dick Flavin. There are eleven retired numbers above the right field grandstand. All of the numbers retired by the Red Sox are red on a white circle. Jackie Robinson's 42, which was retired by Major League Baseball, is blue on a white circle. The two are further delineated through the font difference; Boston numbers are in the same style as the Red Sox jerseys, while Robinson's number is in the more traditional "block" numbering found on the Dodgers jerseys. Until the late 1990s, the numbers originally hung on the right-field facade in the order in which they were retired: 9-4-1-8. It was pointed out that the numbers, when read as a date (9/4/18), marked the eve of the first game of the 1918 World Series, the last championship that the Red Sox won before 2004. After the facade was repainted, the numbers were rearranged in numerical order. The numbers remained in numerical order until the 2012 season, when the numbers were rearranged back into the order in which they were retired by the Red Sox. The Red Sox policy on retiring uniform numbers was once one of the most stringent in baseball—the player had to be elected to the National Baseball Hall of Fame, play at least 10 years with the team, and retire as a member of the Red Sox. The final requirement was waived for Carlton Fisk as he had finished his playing career with the Chicago White Sox. However, Fisk was assigned a Red Sox front office job and effectively "finished" his baseball career with the Red Sox in this manner. In 2008, the ownership relaxed the requirements further with the retirement of Johnny Pesky's number 6. Pesky has not been inducted into the Hall of Fame, but in light of his over 50 years of service to the club, the management made an exception. Pesky would have had 10 seasons, but he was credited with the three seasons he served as an Operations Officer in the U.S. Navy during World War II. The most recent number retired was 34, worn by 2013 World Series Most Valuable Player David "Big Papi" Ortiz. It is a misconception among fans that a fly ball that gets stuck in the ladder above the scoreboard on the left field wall is ruled a ground rule triple. There is no mention of it in the Red Sox ground rules list.
https://en.wikipedia.org/wiki?curid=11238
Flatulence Flatulence is defined in the medical literature as "flatus expelled through the anus" or the "quality or state of being flatulent", which is defined in turn as "marked by or affected with gases generated in the intestine or stomach; likely to cause digestive flatulence". The root of these words is from the Latin – "a blowing, a breaking wind". Flatus is also the medical word for gas generated in the stomach or bowels. Despite these standard definitions, a proportion of intestinal gas may be swallowed environmental air, and hence flatus is not totally generated in the stomach or bowels. The scientific study of this area of medicine is termed flatology. Flatus is brought to the rectum and pressurized by muscles in the intestines. It is normal to pass flatus, though volume and frequency vary greatly among individuals. It is also normal for intestinal gas to have a feculent odor, which may be intense. The noise commonly associated with flatulence ("blowing a raspberry") is produced by the anus and buttocks, which act together in a manner similar to that of an embouchure. Both the sound and the smell are sources of embarrassment, annoyance or comedy. There are five general symptoms related to intestinal gas: pain, bloating and abdominal distension, excessive flatus volume, excessive flatus smell and gas incontinence. Furthermore, eructation ("an act or instance of belching", colloquially known as "burping") is sometimes included under the topic of flatulence. When excessive or malodorous, flatus can be a sign of a health disorder, such as irritable bowel syndrome, celiac disease or lactose intolerance. Non-medical definitions of the term include "the uncomfortable condition of having gas in the stomach and bowels", or "a state of excessive gas in the alimentary canal". These definitions highlight that many people consider "bloating", abdominal distension or increased volume of intestinal gas, to be synonymous with the term flatulence (although this is technically inaccurate). Colloquially, flatulence may be referred to as "farting", "trumping", "blowing off", "tooting", "pooting", "passing gas", "breaking wind" or simply (in American English) "gas" or (British English) "wind". Derived terms include vaginal flatulence, otherwise known as a "queef". Generally speaking, there are four different types of complaints that relate to intestinal gas, which may present individually or in combination. Patients may complain of bloating as abdominal distension, discomfort and pain from "trapped wind". In the past, functional bowel disorders such as irritable bowel syndrome that produced symptoms of bloating were attributed to increased production of intestinal gas. However, three significant pieces of evidence refute this theory. First, in normal subjects, even very high rates of gas infusion into the small intestine (30mL/min) is tolerated without complaints of pain or bloating and harmlessly passed as flatus per rectum. Secondly, studies aiming to quantify the total volume of gas produced by patients with irritable bowel syndrome (some including gas emitted from the mouth by eructation) have consistently failed to demonstrate increased volumes compared to healthy subjects. The proportion of hydrogen produced may be increased in some patients with irritable bowel syndrome, but this does not affect the total volume. Thirdly, the volume of flatus produced by patients with irritable bowel syndrome who have pain and abdominal distension would be tolerated in normal subjects without any complaints of pain. Patients who complain of bloating frequently can be shown to have objective increases in abdominal girth, often increased throughout the day and then resolving during sleep. The increase in girth combined with the fact that the total volume of flatus is not increased led to studies aiming to image the distribution of intestinal gas in patients with bloating. They found that gas was not distributed normally in these patients: there was segmental gas pooling and focal distension. In conclusion, abdominal distension, pain and bloating symptoms are the result of abnormal intestinal gas dynamics rather than increased flatus production. As mentioned above, the normal range of volumes of flatus in normal individuals varies hugely (476–1,491 mL/24 h). All intestinal gas is either swallowed environmental air, present intrinsically in foods and beverages, or the result of gut fermentation. Swallowing small amounts of air occurs while eating and drinking. This is emitted from the mouth by eructation (burping) and is normal. Excessive swallowing of environmental air is called aerophagia, and has been shown in a few case reports to be responsible for increased flatus volume. This is, however, considered a rare cause of increased flatus volume. Gases contained in food and beverages are likewise emitted largely through eructation, e.g., carbonated beverages. Endogenously produced intestinal gases make up 74 percent of flatus in normal subjects. The volume of gas produced is partially dependent upon the composition of the intestinal microbiota, which is normally very resistant to change, but is also very different in different individuals. Some patients are predisposed to increased endogenous gas production by virtue of their gut microbiota composition. The greatest concentration of gut bacteria is in the colon, while the small intestine is normally nearly sterile. Fermentation occurs when unabsorbed food residues may arrive in the colon. Therefore, even more than the composition of the microbiota, diet is the primary factor that dictates the volume of flatus produced. Diets that aim to reduce the amount of undigested fermentable food residues arriving in the colon have been shown to significantly reduce the volume of flatus produced. Again, increased volume of intestinal gas will not cause bloating and pain in normal subjects. Abnormal intestinal gas dynamics will create pain, distension, and bloating, regardless of whether there is high or low total flatus volume. Although flatus possesses physiological smell, this may be abnormally increased in some patients and cause social distress to the patient. Increased smell of flatus presents a distinct clinical issue from other complaints related to intestinal gas. Some patients may exhibit over-sensitivity to bad flatus smell, and in extreme forms, olfactory reference syndrome may be diagnosed. "Gas incontinence" could be defined as loss of voluntary control over the passage of flatus. It is a recognised subtype of faecal incontinence, and is usually related to minor disruptions of the continence mechanisms. Some consider gas incontinence to be the first, sometimes only, symptom of faecal incontinence. Intestinal gas is composed of varying quantities of exogenous sources and endogenous sources. The exogenous gases are swallowed (aerophagia) when eating or drinking or increased swallowing during times of excessive salivation (as might occur when nauseated or as the result of gastroesophageal reflux disease). The endogenous gases are produced either as a by-product of digesting certain types of food, or of incomplete digestion, as is the case during steatorrhea. Anything that causes food to be incompletely digested by the stomach or small intestine may cause flatulence when the material arrives in the large intestine, due to fermentation by yeast or prokaryotes normally or abnormally present in the gastrointestinal tract. Flatulence-producing foods are typically high in certain polysaccharides, especially oligosaccharides such as inulin. Those foods include beans, lentils, dairy products, onions, garlic, spring onions, leeks, turnips, swedes, radishes, sweet potatoes, potatoes, cashews, Jerusalem artichokes, oats, wheat, and yeast in breads. Cauliflower, broccoli, cabbage, Brussels sprouts and other cruciferous vegetables that belong to the genus "Brassica" are commonly reputed to not only increase flatulence, but to increase the pungency of the flatus. In beans, endogenous gases seem to arise from complex oligosaccharides (carbohydrates) that are particularly resistant to digestion by mammals, but are readily digestible by gut flora – microorganisms (methane-producing archaea; "Methanobrevibacter smithii") that inhabit the digestive tract. These oligosaccharides pass through the upper intestine largely unchanged, and when they reach the lower intestine, bacteria feed on them, producing copious amounts of flatus. When excessive or malodorous, flatus can be a sign of a health disorder, such as irritable bowel syndrome, celiac disease, non-celiac gluten sensitivity or lactose intolerance. It can also be caused by certain medicines, such as ibuprofen, laxatives, antifungal medicines or statins. Some infections, such as giardiasis, are also associated with flatulence. Interest in the causes of flatulence was spurred by high-altitude flight and manned spaceflight; the low atmospheric pressure, confined conditions, and stresses peculiar to those endeavours were cause for concern. In the field of mountaineering, the phenomenon of high altitude flatus expulsion was first recorded over two hundred years ago. Flatus (intestinal gas) is mostly produced as a byproduct of bacterial fermentation in the gastrointestinal (GI) tract, especially the colon. There are reports of aerophagia (excessive air swallowing) causing excessive intestinal gas, but this is considered rare. Over 99% of the volume of flatus is composed of non-smelly gases. These include oxygen, nitrogen, carbon dioxide, hydrogen and methane. Nitrogen is not produced in the gut, but a component of environmental air. Patients who have excessive intestinal gas that is mostly composed of nitrogen have aerophagia. Hydrogen, carbon dioxide and methane are all produced in the gut and contribute 74% of the volume of flatus in normal subjects. Methane and hydrogen are flammable, and so flatus can be ignited if it contains adequate amounts of these components. Not all humans produce flatus that contains methane. For example, in one study of the faeces of nine adults, only five of the samples contained archaea capable of producing methane. The prevalence of methane over hydrogen in human farts may correlate with obesity, constipation and irritable bowel syndrome, as archaea that oxidise hydrogen into methane promote the metabolism's ability to absorb fatty acids from food. The remaining trace (<1% volume) compounds give flatus its smell. Historically, compounds such as indole, skatole, ammonia and short chain fatty acids were thought to cause the smell of flatus. More recent evidence proves that the major contribution to the smell of flatus comes from a combination of volatile sulfur compounds. Hydrogen sulfide, methyl mercaptan (also known as methanethiol), dimethyl sulfide, dimethyl disulfide and dimethyl trisulfide are present in flatus. The benzopyrrole volatiles indole and skatole have a mothball smell, and therefore probably do not contribute greatly to the characteristic smell of flatus. In one study, hydrogen sulfide concentration was shown to correlate convincingly with perceived bad smell of flatus, followed by methyl mercaptan and dimethyl sulfide. This is supported by the fact that hydrogen sulfide may be the most abundant volatile sulfur compound present. These results were generated from subjects who were eating a diet high in pinto beans to stimulate flatus production. Others report that methyl mercaptan was the greatest contributor to the smell of flatus in patients not under any specific dietary alterations. It has now been demonstrated that methyl mercaptan, dimethyl sulfide, and hydrogen sulfide (described as decomposing vegetables, unpleasantly sweet/wild radish and rotten eggs respectively) are all present in human flatus in concentrations above their smell perception thresholds. It is known that increased dietary sulfur-containing amino acids significantly increases the smell of flatus. It is therefore likely that the smell of flatus is created by a combination of volatile sulfur compounds, with minimal contribution from non-sulfur volatiles. Such smell can also be caused by the presence of large numbers of microflora bacteria or the presence of faeces in the rectum. Diets high in protein, especially sulfur-containing amino acids, have been demonstrated to significantly increase the smell of flatus. Normal flatus volume is 476 to 1491 mL per 24 hours. This variability between individuals is greatly dependent upon diet. Similarly, the number of flatus episodes per day is variable; the normal range is given as 8–20 per day. The volume of flatus associated with each flatulence event again varies (5–375 mL). The volume of the first flatulence upon waking in the morning is significantly larger than those during the day. This may be due to buildup of intestinal gas in the colon during sleep, the peak in peristaltic activity in the first few hours after waking or the strong prokinetic effect of rectal distension on the rate of transit of intestinal gas. It is now known that gas is moved along the gut independently of solids and liquids, and this transit is more efficient in the erect position compared to when supine. It is thought that large volumes of intestinal gas present low resistance, and can be propelled by subtle changes in gut tone, capacitance and proximal contraction and distal relaxation. This process is thought not to affect solid and liquid intra-lumenal contents. Researchers investigating the role of sensory nerve endings in the anal canal did not find them to be essential for retaining fluids in the anus, and instead speculate that their role may be to distinguish between flatus and faeces, thereby helping detect a need to defecate or to signal the end of defecation. The sound varies depending on the tightness of the sphincter muscle and velocity of the gas being propelled, as well as other factors, such as water and body fat. The auditory pitch (sound) of the flatulence outburst can also be affected by the anal embouchure. Among humans, flatulence occasionally happens accidentally, such as incidentally to coughing or sneezing or during orgasm; on other occasions, flatulence can be voluntarily elicited by tensing the rectum or "bearing down" on stomach or bowel muscles and subsequently relaxing the anal sphincter, resulting in the expulsion of flatus. Since problems involving intestinal gas present as different (but sometimes combined) complaints, the management is cause-related. While not affecting the production of the gases themselves, surfactants (agents that lower surface tension) can reduce the disagreeable sensations associated with flatulence, by aiding the dissolution of the gases into liquid and solid faecal matter. Preparations containing simethicone reportedly operate by promoting the coalescence of smaller bubbles into larger ones more easily passed from the body, either by burping or flatulence. Such preparations do not decrease the total amount of gas generated in or passed from the colon, but make the bubbles larger and thereby allowing them to be passed more easily. Other drugs including prokinetics, lubiprostone, antibiotics and probiotics are also used to treat bloating in patients with functional bowel disorders such as irritable bowel syndrome, and there is some evidence that these measures may reduce symptoms. A flexible tube, inserted into the rectum, can be used to collect intestinal gas in a flatus bag. This method is occasionally needed in a hospital setting, when the patient is unable to pass gas normally. One method of reducing the volume of flatus produced is dietary modification, reducing the amount of fermentable carbohydrates. This is the theory behind diets such as the low-FODMAP diet (low fermentable oligosaccharide, disacharide, monosaccharide and polyols). Certain spices have been reported to counteract the production of intestinal gas, most notably the closely related cumin, coriander, caraway, fennel and others such as ajwain, turmeric, asafoetida (hing), epazote, and kombu kelp (a Japanese seaweed). Most starches, including potatoes, corn, noodles, and wheat, produce gas as they are broken down in the large intestine. Intestinal gas can be reduced by fermenting the beans, and making them less gas-inducing, or by cooking them in the liquor from a previous batch. For example, the fermented bean product miso is less likely to produce as much intestinal gas. Some legumes also stand up to prolonged cooking, which can help break down the oligosaccharides into simple sugars. Fermentative lactic acid bacteria such as "Lactobacillus casei" and "Lactobacillus plantarum" reduce flatulence in the human intestinal tract. Probiotics (live yogurt, kefir, etc.) are reputed to reduce flatulence when used to restore balance to the normal intestinal flora. Live (bioactive) yogurt contains, among other lactic bacteria, "Lactobacillus acidophilus", which may be useful in reducing flatulence. "L. acidophilus" may make the intestinal environment more acidic, supporting a natural balance of the fermentative processes. "L. acidophilus" is available in supplements. Prebiotics, which generally are non-digestible oligosaccharides, such as fructooligosaccharide, generally increase flatulence in a similar way as described for lactose intolerance. Digestive enzyme supplements may significantly reduce the amount of flatulence caused by some components of foods not being digested by the body and thereby promoting the action of microbes in the small and large intestines. It has been suggested that alpha-galactosidase enzymes, which can digest certain complex sugars, are effective in reducing the volume and frequency of flatus. The enzymes alpha-galactosidase, lactase, amylase, lipase, protease, cellulase, glucoamylase, invertase, malt diastase, pectinase, and bromelain are available, either individually or in combination blends, in commercial products. The antibiotic rifaximin, often used to treat diarrhea caused by the microorganism "E. coli", may reduce both the production of intestinal gas and the frequency of flatus events. Bismuth Smell from flatulence is commonly treated with bismuth subgallate, available over-the-counter in the US as Devrom. Bismuth subgallate is commonly used by individuals who have had ostomy surgery, bariatric surgery, faecal incontinence and irritable bowel syndrome. Bismuth subsalicylate is a compound that binds hydrogen sulfide, and one study reported a dose of 524 mg four times a day for 3–7 days bismuth subsalicylate yielded a >95% reduction in faecal hydrogen sulfide release in both humans and rats. Another bismuth compound, bismuth subnitrate was also shown to bind to hydrogen sulfide. Another study showed that bismuth acted synergistically with various antibiotics to inhibit sulfate-reducing gut bacteria and sulfide production. Some authors proposed a theory that hydrogen sulfide was involved in the development of ulcerative colitis and that bismuth might be helpful in the management of this condition. However, bismuth administration in rats did not prevent them from developing ulcerative colitis despite reduced hydrogen sulfide production. Also, evidence suggests that colonic hydrogen sulfide is largely present in bound forms, probably sulfides of iron and other metals. Rarely, serious bismuth toxicity may occur with higher doses. Despite being an ancient treatment for various digestive complaints, activated charcoal did not produce reduction in both the total flatus volume nor the release of sulfur-containing gasses, and there was no reduction in abdominal symptoms (after 0.52g activated charcoal four times a day for one week). The authors suggested that saturation of charcoal binding sites during its passage through the gut was the reason for this. A further study concluded that activated charcoal (4g) does not influence gas formation in vitro or in vivo. Other authors reported that activated charcoal was effective. A study in 8 dogs concluded activated charcoal (unknown oral dose) reduced hydrogen sulfide levels by 71%. In combination with yucca schidigera, and zinc acetate, this was increased to an 86% reduction in hydrogen sulfide, although flatus volume and number was unchanged. An early study reported activated charcoal (unknown oral dose) prevented a large increase in the number of flatus events and increased breath hydrogen concentrations that normally occur following a gas-producing meal. Garments and external devices In 1998, Chester "Buck" Weimer of Pueblo, Colorado, received a patent for the first undergarment that contained a replaceable charcoal filter. The undergarments are air-tight and provide a pocketed escape hole in which a charcoal filter can be inserted. In 2001 Mr Weimer received the Ig Nobel Prize for Biology for his invention. A similar product was released in 2002, but rather than an entire undergarment, consumers are able to purchase an insert similar to a pantiliner that contains activated charcoal. The inventors, Myra and Brian Conant of Mililani, Hawaii, still claim on their website to have discovered the undergarment product in 2002 (four years after Chester Weimer filed for a patent for his product), but state that their tests "concluded" that they should release an insert instead. Flatus incontinence where there is involuntary passage of gas, is a type of faecal incontinence, and is managed similarly. In many cultures, flatulence in public is regarded as embarrassing, but, depending on context, can also be considered humorous. People will often strain to hold in the passing of gas when in polite company, or position themselves to silence the passing of gas. In other cultures, it may be no more embarrassing than coughing. While the act of passing flatus in said cultures is generally considered to be an unfortunate occurrence in public settings, flatulence may, in casual circumstances and especially among children, be used as either a humorous supplement to a joke ("pull my finger"), or as a comic activity in and of itself. The social acceptability of flatulence-based humour in entertainment and the mass media varies over the course of time and between cultures. Enough entertainers performed with their flatus that the term flatulist was coined. The whoopee cushion is a joking device invented in the early 20th century for simulating a fart. In 2008, a farting application for the iPhone earned nearly $10,000 in one day. A farting game named "Touch Wood" was documented by John Gregory Bourke in the 1890s. It existed under the name of "Safety" in the 20th century in the U.S., and has been found being played in 2011. In January 2011, the Malawi Minister of Justice, George Chaponda, said that Air Fouling Legislation would make public farting illegal in his country. When reporting the story, the media satirised Chaponda's statement with punning headlines. Later, the minister withdrew his statement. Flatulence is often blamed as a significant source of greenhouse gases, owing to the erroneous belief that the methane released by livestock is in the flatus. While livestock account for around 20% of global methane emissions, 90–95% of that is released by exhaling or burping. In cows, gas and burps are produced by methane-generating microbes called methanogens, that live inside the cow's digestive system. Proposals for reducing methane production in cows include the feeding of supplements such as oregano and seaweed, and the genetic engineering of gut biome microbes to produce less methane. Since New Zealand produces large amounts of agricultural products, it is in the unique position of having high methane emissions from livestock compared to other greenhouse gas sources. The New Zealand government is a signatory to the Kyoto Protocol and therefore attempts are being made to reduce greenhouse emissions. To achieve this, an agricultural emissions research levy was proposed, which promptly became known as a "fart tax" or "flatulence tax". It encountered opposition from farmers, farming lobby groups and opposition politicians. Historical comment on the ability to fart at will is observed as early as Saint Augustine's "The City of God" (5th century AD). Augustine mentions men who "have such command of their bowels, that they can break wind continuously at will, so as to produce the effect of singing". Intentional passing of gas and its use as entertainment for others appear to have been somewhat well known in pre-modern Europe, according to mentions of it in medieval and later literature, including Rabelais. Le Pétomane ("the Fartomaniac") was a famous French performer in the 19th century who, as well as many professional farters before him, did flatulence impressions and held shows. The performer Mr. Methane carries on le Pétomane's tradition today. Also, a 2002 film "Thunderpants" revolves around a boy named Patrick Smash who has an ongoing flatulence problem from the time of his birth. He eventually overcomes his problems and fulfills his dreams, including one of becoming an astronaut. In literature, farting features prominently in the novel "The Catcher in the Rye". Since the 1970s, farting has increasingly been featured in film, especially comedies such as "Blazing Saddles" and "Scooby-Doo". In Islam, flatulence invalidates "wudu" (ritual purity). Notes Bibliography
https://en.wikipedia.org/wiki?curid=11240
Final Fantasy: The Spirits Within Final Fantasy: The Spirits Within is a 2001 computer-animated science fiction film directed by Hironobu Sakaguchi, creator of the "Final Fantasy" franchise. It was the first photorealistic computer-animated feature film and was the most expensive video game-inspired film until the release of "" in 2010. It features the voices of Ming-Na Wen, Alec Baldwin, Donald Sutherland, James Woods, Ving Rhames, Peri Gilpin and Steve Buscemi. "The Spirits Within" follows scientists Aki Ross and Doctor Sid in their efforts to free a post-apocalyptic Earth from a mysterious and deadly alien race known as the Phantoms, which has driven the remnants of humanity into "barrier cities". Aki and Sid must fight against General Hein, who wishes to use more violent means to end the conflict. Square Pictures rendered the film using some of the most advanced processing capabilities available for film animation at the time. A render farm consisting of 960 workstations was tasked with rendering each of the film's 141,964 frames. It took a staff of 200 about four years to complete "The Spirits Within". Square intended to make the character of Aki Ross into the world's first photorealistic computer-animated actress, with plans for appearances in multiple films in different roles. "The Spirits Within" premiered in Los Angeles on July 2, 2001, and was theatrically released in the United States on July 11. It received a mixed critical reception, but was widely praised for the realism of the computer-animated characters. Due to rising costs, the film greatly exceeded its original budget towards the end of production, reaching a final cost of $137 million; it grossed only $85 million at the box office. The film has been called a box office bomb and is blamed for the demise of Square Pictures. In 2065, Earth is infested by alien life forms known as Phantoms. By physical contact Phantoms consume the Gaia spirit of living beings, killing them instantly, though a minor contact may only result in an infection. The surviving humans live in "barrier cities", areas protected by an energy shield that prevents Phantoms from entering, and are engaged in an ongoing struggle to free the planet. After being infected by a Phantom during one of her experiments, Aki Ross (Ming-Na) and her mentor, Doctor Sid (Donald Sutherland), uncover a means of defeating the Phantoms by gathering eight spirit signatures that, when joined, can negate the Phantoms. Aki is searching for the sixth spirit in the ruins of New York City when she is cornered by Phantoms but is rescued by Gray Edwards (Alec Baldwin) and his squad "Deep Eyes", consisting of Ryan Whittaker (Ving Rhames), Neil Fleming (Steve Buscemi) and Jane Proudfoot (Peri Gilpin). It is revealed that Gray was once romantically involved with Aki. Upon returning to her barrier city, Aki joins Sid and appears before the leadership council along with General Hein (James Woods), who is determined to use the powerful Zeus space cannon to destroy the Phantoms. Aki is concerned the cannon will damage Earth's Gaia (a spirit representing its ecosystem) and delays the use of it by revealing that she has been infected and the collected spirit signatures are keeping her infection stable, convincing the council that there may be another way to defeat the Phantoms. However, this revelation leads Hein to incorrectly conclude that she is being controlled by the Phantoms. Aki and the Deep Eyes squad succeed in finding the seventh spirit as Aki's infection begins to worsen and she slips into unconsciousness. Her dream reveals to her that the Phantoms are the spirits of dead aliens brought to Earth on a fragment of their destroyed planet. Sid uses the seventh spirit to bring Aki's infection back under control, reviving her. To scare the council into giving him clearance to fire the Zeus cannon, Hein lowers part of the barrier shield protecting the city. Though Hein intended that only a few Phantoms enter, his plan goes awry and legions of Phantoms invade the entire city. Aki, Sid and the Deep Eyes attempt to reach Aki's spaceship, their means of escape, but Ryan, Neil and Jane are killed by Phantoms. Hein escapes and boards the Zeus space-station where he finally receives authorisation to fire the cannon. Sid finds the eighth spirit at the crater site of the alien asteroid's impact on Earth. He lowers a shielded vehicle, with Aki and Gray aboard, into the crater to locate the final spirit. Just before they can reach it, Hein fires the Zeus cannon into the crater, not only destroying the eighth spirit but also revealing the Phantom Gaia. Aki has a vision of the Phantom home planet, where she is able to receive the eighth spirit from the alien particles in herself. When Aki awakens, she and Gray combine it with the other seven. Hein continues to fire the Zeus cannon despite overheating warnings and unintentionally destroys the cannon and himself. Gray sacrifices himself as a medium needed to physically transmit the completed spirit into the alien Gaia. The Earth's Gaia is returned to normal as the Phantoms ascend into space, finally at peace. Aki is pulled from the crater holding Gray's body, and is seen looking into the newly liberated world. Aki Ross's voice actor, Ming-Na Wen, was selected for a perceived fit between her personality and Aki's. Ming-Na, who found the role via her publicist, said she felt like she had given birth with her voice to the character. She gradually accustomed herself to the difficulty of working without the presence and spontaneity of real actors, and commented that the voice-acting work did not take much time, as she would just go into the studio "once or twice a month for about four months" with no need for make-up and costuming sessions. The workload was so light it did not interfere with her acting commitments in the television series "ER". Square accumulated four SGI Origin 2000 series servers, four Onyx2 systems, and 167 Octane workstations for the film's production. The basic film was rendered at a custom render farm created by Square in Hawaii. It housed 960 Pentium III-933 MHz workstations. Character movements were filmed using motion capture technology. Animator Matthew Hackett stated that while motion capture was effective for many of the scenes, in others animators still had to add movements manually. Hand and facial movements were all done manually. Some of General Hein's facial features and poses were based on Hackett. As animators did not want to use any actual photographs in the film, all backgrounds were done using matte paintings. 1,327 scenes in total needed to be filmed to animate the digital characters. The film consists of 141,964 frames, with each frame taking an average of 90 minutes to render. By the end of production Square had a total of 15 terabytes of artwork for the film. It is estimated that over the film's four-year production, approximately 200 people put in a combined 120 years of work on it. From early on, it had been decided that "The Spirits Within" would be filmed entirely in English. The original script, written by Sakaguchi, was titled "Gaia". The screenplay was written by Al Reinert and Jeff Vintar. The film was co-directed by Motonori Sakakibara, with Jun Aida and Chris Lee both serving as producers. Lee compared "The Spirits Within", the first full-length photorealistic animated film, to Walt Disney's "Snow White and the Seven Dwarfs", the first full-length cel animated film. In order to keep the film in line with Hironobu Sakaguchi's vision as director, several script rewrites took place, most in the initial stages of production. Sakaguchi stated he was pleased with the film's final cut, saying he would not have changed anything if given the chance. The film had high cost overruns towards the end of filming. New funds had to be sourced to cover the increasing production costs while maintaining staff salaries. The film's final cost of $137 million, which included about $30 million spent on marketing by the film's distributor Columbia Pictures, escalated from an original budget rumored to be around $70 million. $45 million alone was spent on the construction of Square's studio in Hawaii. It was Columbia Pictures' first animated feature film since "" in 1986. Director Sakaguchi named the main character after his mother, Aki, who died in an accident several years prior to the production of the film. Her death led Sakaguchi to reflect on what happened to the spirit after death, and these thoughts resurfaced while he was planning the film, eventually taking the form of the Gaia hypothesis. He later explained that the theme he wanted to convey was "more of a complex idea of life and death and spirit", believing that the best way to portray this would be to set the film on Earth. By comparison, "Final Fantasy" video games are set in fictional worlds. Dan Mayers from "Sight & Sound" stated the film followed the same theme typically found in "Final Fantasy" video games: "A party of heroes averts impending global holocaust by drawing on their individual skills, gaining knowledge through challenges and emerging victorious with new-found love and respect for themselves and their companions." Writing in the book "Robot Ghosts and Wired Dreams", Livia Monnet stated the film remediated "the notion of life in the neovitalistic, evolutionary biology of Lynn Margulis and in contemporary theories on artificial life", going on to state that the film's exploration of the Gaia hypothesis raised interesting questions regarding the life and death process of both cinema and digital media, as well as contemporary life sciences, cybernetics, philosophy and science fiction. The concept of artificial life and resurrection was also discussed, and compared to similar themes in the 1914 book "Locus Solus"; the Phantoms in "The Spirits Within" were considered to be brought to life by various forces: by the alien planet's red Gaia and then by human spiritual energy. Each character's base body model was built from more than 100,000 polygons, plus more than 300,000 for clothing alone. Aki's character model bears 60,000 hairs, each of which were separately and fully animated and rendered. In creating the characters, designers had to transition between using PowerAnimator, Autodesk Maya and RenderMan. Aki's appearance was conceived by the lead animator of the project, Roy Sato, who created several conceptual designs for Sakaguchi to consider, and then used the selected design as a guide for her character model. Sato perceived Aki's original look as a "supermodel", and subsequently removed her make-up and shortened her hair in order to give her a more intelligent look that would "convince people that she's a scientist." In an interview, Sato described actively trying to make her appear as realistic as possible, making her similar to himself in as many ways as he could in the animation, including elements of his personality through facial expressions. He concluded that Aki ended up being similar to him in almost every way, with the exception that "she's a lot cuter". The model for Aki was designed to closely follow human appearance, with Sakaguchi commenting in an interview "I think it's OK to look at Aki and be convinced that she's a human." While Square ruled out any chance of a sequel to "The Spirits Within" before it was even completed, Sakaguchi intended to position Aki as being the "main star" for Square Pictures, using her in later games and films by Square, and including the flexibility of being able to modify aspects such as her age for such appearances. Ming-Na stated that she would be willing if asked to continue voicing Aki. Aki only made one appearance outside of the film; in 2002 she appeared in a demonstration video that Square Pictures made to present to The Wachowskis before developing "Final Flight of the Osiris" for "The Animatrix". The short film, appearing in the DVD's bonus content and featuring her with a slightly modified design, shows her acrobatically dueling a robot from the "Matrix" setting. Shortly afterward, Square Pictures was closed and absorbed into Square and the company ceased use of the character. While the near lifelike appearance of the characters in the film was well received, some commentators felt the character renderings fell into the trap that many robotics scientists refer to as the "uncanny valley". This concept describes when a robot or animated character becomes very realistic, but subtly different enough from reality to feel "creepy". John Mangan from "The Age" cited the film as an example of this phenomenon. The soundtrack to the film was released on July 3, 2001 by Sony Music. Elliot Goldenthal composed the entire score, as well as the film's theme song, "The Dream Within", which had lyrics written by Richard Rudolf and vocals performed by Lara Fabian. Director Hironobu Sakaguchi opted for the acclaimed Goldenthal instead of Nobuo Uematsu, the composer of the "Final Fantasy" games' soundtracks, a decision met with mixed opinions as the former was completely unknown to many of the games' fans. The last song on the album and the second and final song to play during the film's credits (after "The Dream Within") is "Spirit Dreams Inside" by Japanese rock band L'Arc-en-Ciel. The film's score was performed by the London Symphony Orchestra with Belgian composer Dirk Brossé conducting. It was recorded in the United Kingdom at the Watford Coloseum and the London AIR Lyndhurst Hall and was mixed at the Manhattan Center Studios in the United States. In the liner notes to the album, Goldenthal describes the soundtrack as combining "orchestration techniques associated with the late 20th-century Polish avant-garde, as well as my own experiments from "Alien 3", and 19th-century Straussian brass and string instrumentation." In the film's 'Making of' featurette, Goldenthal states he used "ghostly choral" music when the Phantoms are emerging, in an attempt to give a celestial feeling, and focused on low brass clusters and taiko drum rhythms for violent scenes. When Aki talks about a dying girl, Goldenthal used a piano in order to give a domestic home-like feeling to a completely foreign environment, also choosing to use a flute each time Aki focusses on Gaia, as he believed it to be the most "human kind of instrument". The album was met with positive reviews. Neil Shurley from AllMusic, who gave the album 4 out of 5, stated the album would probably have been nominated for an Oscar if the film itself had been more popular, as did the reviewer from Soundtrack Express, who gave the soundtrack 5 out of 5. Christopher Coleman from Tracksounds gave the soundtrack 10 out of 10, stating the feel of the album was "expansive and majestic" and that the score elevated the viewing experience of the film. A review from "Filmtracks" gave the album 4 out of 5, calling it "an easy album to recommend", adding "parts of it will blow you out of your seat." Dan Goldwasser from Soundtrack.net also gave the soundtrack 4 out of 5, calling it a "must have". The album peaked at No. 19 on "Billboard"'s Top Soundtracks list and No. 193 on the "Billboard" 200 on July 28, 2001. The track "The Dream Within" was nominated for "Best Original Song Written for a Film" at the 2002 World Soundtrack Awards, but lost to "If I Didn't Have You" which was composed for "Monsters, Inc.". Before the film's release, there was already skepticism of its potential to be financially successful. Chris Taylor from "Time" magazine noted that video game adaptations had a poor track record at the box office and that it was Sakaguchi's first feature film. The film debuted on July 2, 2001 at the Mann Bruins Theater in Los Angeles, California, and was released in the United States on July 11, making $32 million in North America and going on to gross $85 million in worldwide box office receipts. In Japan, the film grossed . The film achieved average to poor results at the box office in most of Southeast Asia; however, it performed well in Australia, New Zealand and South Korea. In terms of box office admissions, the film sold 5,961,378 tickets in the United States, 1,456,523 tickets in France, 4,299,604 tickets in other European countries, 446,728 tickets in Brazil, and 160,100 tickets in Seoul City (South Korea). This adds up to 12,324,333 tickets sold in the United States, Europe, Brazil and Seoul. In 2006, Boston.com regarded it as the 4th biggest box office bomb, estimating the film's losses at the end of its cinema run at over $94 million. In March 2012, CNBC considered it to be the 9th biggest box office bomb, though "Time"s list of the ten biggest box office failures, which was released on the same day, did not include the film. The review aggregator website Rotten Tomatoes reported an approval rating of 45% based on 146 reviews, with an average rating of 5.3/10. The website's critical consensus reads, "The movie raises the bar for computer animated movies, but the story is dull and emotionally removed." Metacritic, which uses a weighted average, assigned a score of 49 out of 100 based on 28 critics, indicating "mixed or average reviews". Roger Ebert of the "Chicago Sun-Times" was a strong advocate of the film; he gave it 3½ stars out of four, praising it as a "technical milestone" while conceding that its "nuts and bolts" story lacked "the intelligence and daring of, say, Steven Spielberg's "A.I."" He noted that while he did not once feel convinced Aki Ross was an actual human being, she was "lifelike", stating her creators "dare us to admire their craft. If Aki is not as real as a human actress, she's about as human as a Playmate who has been retouched to glossy perfection." He also expressed a desire for the film to succeed in hopes of seeing more films made in its image, though he was skeptical of its ability to be accepted. Peter Bradshaw of "The Guardian" gave a more negative review, stating that while the animation was brilliant, the "solemnly realist human faces look shriekingly phoney precisely because they're almost there but not quite", concluding "The story is adequate, if familiar, but after half an hour relapses into cliche." Aki's appearance was received positively by critics, with praise for the finer details of the character model such as the rendering of her hair. "Entertainment Weekly" named Aki an "it girl", stating that "Calling this action heroine a cartoon would be like calling a Rembrandt a doodle." She was voted one of the sexiest women of 2001 by "Maxim" and its readers, ranking at No. 87 out of 100, becoming the first fictional woman to ever make the list, additionally appearing on the issue's cover in a purple bikini. The same image appeared in the "Babes: The Girls of Sci Fi" special issue of "SFX". Ruth La Ferla from "The New York Times" described her as having the "sinewy efficiency" of "Alien" franchise character Ellen Ripley and visual appeal of Julia Roberts' portrayal of Erin Brockovich. The book "Digital Shock: Confronting the New Reality" by Herve Fischer described her as a virtual actress having a "beauty that is 'really' impressive", comparing her to video game character Lara Croft. In contrast, Livia Monnet criticized her character as an example of the constantly kidnapped female in Japanese cinema, further "diluted" by her existence solely as a computer-generated character representing "an ideal, cinematic female character that has no real referent." Writing in the book "Action and Adventure Cinema", Marc O'Day described her as among the "least overtly eroticised" female characters in science fiction, though stated that Aki was "transformed in a variety of poses into an erotic fantasy machine" in a photo shoot that was included on the DVD's special features. The merger between Square and Enix, which had been under consideration since at least 2000 according to Yasuhiro Fukushima, Enix chairman at the time, was delayed because of the failure of the film and Enix's hesitation at merging with a company that had just lost a substantial amount of money. Square Pictures announced in late January 2002 that they were closing down, largely due to the commercial failure of "The Spirits Within". The film's CGI effects have been compared favourably with later CGI films such as James Cameron's 2009 film "Avatar". In 2011, BioWare art director Derek Watts cited "The Spirits Within" as a major influence on the successful "Mass Effect" series of action role-playing games. In the first episode of the Square Enix published 2015 video game "Life Is Strange", when the lead interacts with a TV, she mentions the idea of watching the film, and says "I don't care what anybody says, that's one of the best sci-fi films ever made." Although the film was loosely based on a video game series, there were never any plans for a game adaptation of the film itself. Sakaguchi indicated the reason for this was the lack of powerful gaming hardware at the time, feeling the graphics in any game adaptation would be far too much of a step down from the graphics in the film itself. A novelization was written by Dean Wesley Smith and published by "Pocket Books" in June 2001. "The Making of Final Fantasy: The Spirits Within", a companion book, was published by "BradyGames" in August 2001. Edited by Steven L. Kent, the 240 page color book contains a foreword by director Sakaguchi and extensive information on all aspects of the film's creation, including concept art, storyboards, sets and props, layout, motion capture and animation, as well as a draft of the full script. The film won the "Jury Prize" at the 2002 Japan Media Arts Festival. It was nominated for "Best Sound Editing – Animated Feature Film, Domestic and Foreign" at the 49th Golden Reel Awards as well as "Best Animated Feature" at the 5th Online Film Critics Society awards. Conversely, the film was also nominated in the worst screenplay category at the 2001 Stinkers Bad Movie Awards. The film's trailer was nominated for the "Golden Fleece" award at the 3rd Golden Trailer Awards. The DVD version of the film was released on October 23, 2001, with the Blu-ray edition released on August 7, 2007. Two weeks before it was released the DVD version was listed on Amazon.com as one of the most-anticipated releases, and it was expected to recoup some of the money lost on the film's disappointing box office performance. Both versions contained two full-length commentary tracks (one featuring Motonori Sakakibara, sequence supervisor Hiroyuki Hayashida, lead artist Tatsuro Maruyama, and creature supervisor Takoo Noguchi; the second featuring animation director Andy Jones, editor Chris S. Capp, and staging director Tani Kunitake) as well as an isolated score with commentary. They also contained a version of the film in its basic CGI and sketch form, with the option of pop-up comments on the film. An easter egg shows the cast of the film re-enacting the dance from "Michael Jackson's Thriller". Fifteen featurettes, including seven on character biographies, three on vehicle comparisons and an interactive "Making Of" featurette, were also included. Other features included Aki's dream viewable as a whole sequence, the film's original opening sequence, and intentional outtakes. Peter Bracke from "High-Def Digest" stated the DVD was "so packed with extras it was almost overwhelming", stating that Sony went "all-out" on the extra features in a likely attempt to boost DVD sales and recover losses. As of December 13, 2001, the film grossed in video rental revenue in the United States, equivalent to 83.4% of its box office gross in the country. The DVD was nominated for "Best DVD Special Edition Release" at the 28th Saturn Awards. Aaron Beierle from "DVD Talk" gave a positive review of the DVD, rating it 4½ out of 5 stars for audio quality, video quality and special features. Dustin Somner from "Blu-ray.com" gave the Blu-ray version 5 out of 5 stars for video quality and special features, and 4½ stars for audio quality. Peter Bracke gave the Blu-ray version 4 out of 5 stars overall. Bibliography Official website
https://en.wikipedia.org/wiki?curid=11242
Free Methodist Church The Free Methodist Church is a Methodist Christian denomination within the holiness movement. It is evangelical in nature and is Wesleyan-Arminian in theology. The Free Methodist Church has 77,000 members in the United States and 1,055,000 members worldwide in 82 nations. The "Light & Life Magazine" is their official publication. The Free Methodist Church World Ministries Center is in Indianapolis, Indiana. The Free Methodist Church was organized at Pekin, New York, in 1860. The founders had been members of the Methodist Episcopal Church but were excluded from its membership for too earnestly advocating what they saw as the doctrines and usages of authentic Wesleyan Methodism. Under the leadership of the Rev. Benjamin Titus (B. T.) Roberts, a graduate of Wesleyan University and an able and eloquent preacher, the movement spread rapidly. Societies were organized, churches built and the work established. At the 1910 session of the General Conference of the Methodist Church at Rochester, New York, a full acknowledgement was made of the wrong done to Roberts fifty years before, and the credentials unjustly taken from him were restored in a public meeting to his son, Rev. Benson Roberts. Before the founding of the church, Roberts began publication of a monthly journal, "The Earnest Christian". In 1868, "The Free Methodist" (now "Light & Life") was begun. A publishing house was established in 1886 to produce books, periodicals and Sunday school curriculum and literature. The name "Methodist" was retained for the newly organized church because the founders felt that their misfortunes (expulsion from the Methodist Episcopal Church) had come to them because of their adherence to doctrines and standards of Methodism. The word "Free" was suggested and adopted because the new church was anti-slavery. Next, pews were to be free to all rather than sold or rented (as was common), so as to provide full access to the poor. The new church hoped for the freedom of worship in the Holy Spirit, as opposed to a stifling formality. A fourth principle was "freedom" from secret and oath-bound societies (in particular the Masonic Lodge), so as to have full loyalty to Christ. Fifth was "freedom" from the abuse of ecclesiastical authority (due to the bishop's action in allowing expulsion of 120 clergy and lay). Finally was "freedom" to experience transformation in sanctification via the Holy Spirit due to personal consecration and faith versus just 'sin-management' or gradual growth following justification. Holiness Conservatives within the Free Methodist Church left to form the Reformed Free Methodist Church in 1932, the United Holiness Church in 1966 (which itself joined the Bible Methodist Connection of Churches in 1994) and also the Evangelical Wesleyan Church in 1963. Free Methodist headquarters were located in Winona Lake, Indiana, until 1990 when the denomination moved its headquarters to Indianapolis. The church has about 77,000 members in the United States. Worldwide its membership is over 1,000,000 with large segments of membership in East Central Africa (Rwanda, Burundi, D.R. of Congo) and India. In doctrine, Free Methodists’ beliefs are the standard beliefs of Wesleyan-Arminian Protestantism, with distinctive emphasis on the teaching of entire sanctification as held by John Wesley, to whom the Free Methodist Church traces its origins. The Free Methodist Church, along with the United Methodist Church, shares a common heritage linked to the Methodist revival in England during the 18th century. The Free Methodist Church itself arose within the context of holiness movement within 19th century Methodism. While some clergy in the Free Methodist Church wear the pulpit robe, others do not wear vestments of any sort during worship. The first general superintendent, B. T. Roberts, was in favor of ordaining women, but never saw it take place in his lifetime. Out of his own conviction he wrote "Ordaining Women: Biblical and Historical Insights". The impact of his writings eventually prevailed in the church. The Free Methodist Church affirmed the ordination of women in 1911. As of June 2008, women represented 11% of ordained clergy (216 of 2,011) and 26% of candidates for the ministry. Free Methodists recognize and license unordained persons for particular ministries. They mandate lay representation in numbers equal to clergy in the councils of the church. As a reaction to paid musicians in the Methodist Episcopal Church, early Free Methodists enjoyed a capella congregational hymns during worship. However, the General Conference of 1943 voted to allow each Conference to vote on whether or not their churches could have instrumental music. As a result, pianos and organs became common across most conferences. Currently, many churches have worship teams composed of vocalists, drums, keyboards, guitars, and other instruments. The Free Methodist Church's highest governing body is the World Conference, which is composed of representatives, both lay and clergy, from all countries with a Free Methodist General Conference. As the church in each country develops, its status progresses from Mission District to Annual Conference to General Conference. There are currently 13 General Conferences in the world, which are linked together through the articles of religion and common constitution of the first two chapters of the Book of Discipline, the World Conference, and the Council of Bishops. The USA branch of the Free Methodist Church is currently led by three bishops: Bishop Matthew Whitehead, Bishop Linda Adams, and Bishop Keith Cowart. All three were elected in 2019. International Child Care Ministries (ICCM), a child sponsorship initiative serves more than 21 000 children in 29 countries around the world. Through education, meals and medical care, children in need are given an opportunity for a better life. Each sponsored child is connected to a Free Methodist congregation or ministry at a local level. Sustainable Empowerment through Economic Development (Seed), a micro-enterprise and livelihood ministry of Free Methodist World Missions, facilitates self-sustaining businesses, training in business skills and Christian discipleship. Focused on economically vulnerable members of the Free Methodist world family, it provides an international market for products produced by Free Methodist artisans. Volunteers in Service Abroad (Visa) connects volunteers from the Free Methodist Church in the US and UK with Free Methodist World Missions for hands-on ministry internationally. The church currently has ministry in 82 countries around the world. These include: B. T. Roberts began what is now Roberts Wesleyan College in 1866. Spring Arbor College followed in 1873 (renamed Spring Arbor University in 2001), Seattle Pacific University in 1891 and Greenville College (renamed Greenville University in 2017) in 1892. Central College began in 1914, a continuation of Orleans Seminary begun in 1884. Azusa Pacific University was formed by a merger with Los Angeles Pacific College and other small denominational colleges. The following educational institutions are a part of the Association of Free Methodist Educational Institutions. The schools are not owned by the denomination but meet a set of requirements to maintain this relationship. In addition, the Free Methodist Church is one of several denominations supporting Azusa Pacific University (Azusa, CA). Wessington Springs College is a former, now closed institution which was located in South Dakota. Internationally, there is Osaka Christian College of the Japanese Free Methodist Church, Hope Africa University, a recently founded school in Bujumbura, Burundi, Haiti Providence University, and the Faculdade de Teologia Metodista Livre, São Paulo, Brazil. Through the John Wesley Seminary Foundation (JWSF) graduate students who are preparing for full-time ministry in the Free Methodist Church are provided a grant or loan at the following affiliated schools: Like John Wesley before him, B. T. Roberts recognized the Christian's responsibility for publishing. Before the founding of the church in 1860, B. T. Roberts began publication of a monthly journal, "The Earnest Christian". In 1868 "The Free Methodist" (now "Light & Life" Magazine) began. A publishing house was established in 1886 to produce books, periodicals and Sunday school curriculum and literature. Early leaders, T. B. Arnold and B. T. Roberts privately financed and produced several publications. The official publishing institution was established by the church at the 1886 General Conference. The church purchased the publishing business built by Rev. T. B. Arnold for $8,000. Arnold was named first publisher and B. T. Roberts was elected editor of "The Free Methodist". The Free Methodist Publishing House is recognized under its trade name Light and Life Press. The Free Methodist Publishing House operated at three locations in Chicago, Illinois. In February 1935, it moved along with Free Methodist Headquarters to Winona Lake, Indiana. During its history, the Free Methodist Publishing House built up a plant and accumulated property worth several hundred thousand dollars. It also contributed thousands of dollars out of its profits to other activities of the church. Over the years, as the ministry of the Free Methodist Church expanded, various departments of the general church gradually moved into Free Methodist Publishing House accommodations. This was provided at vast cost and without the investment of any capital by the general church. In 1960, the Free Methodist Publishing House board issued a deed in favor of the general church, whereby the church became the owner of the old property, plus nearly eight acres of land. For this the general church paid nothing, but agreed to make payments of $5,000 per year over a ten-year period to the Free Methodist Publishing House. In 1944 the Free Methodist Church began a weekly radio show called The Light and Life Radio Hour which featured hymns, sermons, prayer, and scripture reading. The show ran until 1980 and featured several different hosts over the years including Dr. Leroy Lowell, Myron F. Boyd, and Robert Andrews. In 2016 Josh Avery began The FMC Radio Show which was a spiritual successor to The Light and Life Radio Hour but embodied a very different focus. In a podcast format, the show is subtitled "your officially unofficial source for all things Free Methodist". Instead of worship and sermon, the show means to act as a uniting factor in the Free Methodist Church by informing listeners about different things that are happening in the denomination. "Arnold’s Commentary" was published from 1894–1980. In the late 1950s and early 60s the church pioneered fully graded church school materials. In 1960 the "Aldersgate Biblical Series" was developed as the only inductive curriculum of its time. A fully equipped printing area consisting of letterpresses, offset press, cutters, folders, bindery, linotypes etc. contributed toward making the church independent of commercial printers for the production for its printing needs at that time. Acting on the recommendation of its executive committee, the board voted in 1988 to phase out printing operations. This decision and the 1989 General Conference decision to move the Press and Headquarters from Winona Lake to Indianapolis in 1990 shifted the focus of the Press. Where formerly, the Press produced and published Sunday school curriculum, this venture is now carried on in cooperation with other holiness denominations. Beginning in 2008, the Wesleyan Publishing House, publishing arm of the Wesleyan Church, began serving the distribution and customer service needs of Light and Life Press. [https://web.archive.org/web/20120627011100/http://llcomm.org/about/ Light & Life Communications [LLCOMM]], the official publishing aim of the Free Methodist Church, is a not-for-profit corporation that exists to serve in partnership with its parent body, the Free Methodist Church. Its primary purpose is to publish and distribute materials that enable the church to fulfill its stated mission. Light & Life Communications also offers its services and materials to all who seek to make Christ known. Free Methodist Communications is the publishing division of the Free Methodist Church. Publications may also be printed or distributed under the name Light & Life Communications. [http://llcomm.org/llm/ Light & Life Magazine [LLM]] is the official magazine of the Free Methodist Church in the United States and is also available online. The magazine has a monthly circulation of 53,000 English copies. Each issue is also translated into Spanish and published concurrently as Revista Luz y Vida, which has a monthly circulation of 6,000 copies. Jeff Finley is the magazine's managing editor. "Free Methodist World Mission People" is a quarterly magazine about world missions offered free of charge upon request. Free Methodist Conversations is an online resource for discussing important values and issues.
https://en.wikipedia.org/wiki?curid=11244
John Falstaff Sir John Falstaff is a fictional character who appears in three plays by William Shakespeare and is eulogized in a fourth. His significance as a fully developed character is primarily formed in the plays "Henry IV, Part 1" and "Part 2", where he is a companion to Prince Hal, the future King Henry V of England. A notable eulogy for Falstaff is presented in Act II, Scene III of "Henry V", where Falstaff does not appear as a character on stage, as enacted by Mistress Quickly in terms that some scholars have ascribed to Plato's description of the death of Socrates after drinking hemlock. By comparison, Falstaff is presented as the buffoonish suitor of two married women in "The Merry Wives of Windsor". Though primarily a comic figure, Falstaff still embodies a kind of depth common to Shakespeare's major characters. A fat, vain, boastful, and cowardly knight, he spends most of his time drinking at the Boar's Head Inn with petty criminals, living on stolen or borrowed money. Falstaff leads the apparently wayward Prince Hal into trouble, and is ultimately repudiated after Hal becomes king. Falstaff has since appeared in other media, notably in operas by Giuseppe Verdi, Ralph Vaughan Williams, and Otto Nicolai, and in Orson Welles' 1966 film "Chimes at Midnight". The operas focus on his role in "The Merry Wives of Windsor", while the film adapts from the Henriad and "The Merry Wives". Welles, who played Falstaff in his film, considered the character to be "Shakespeare's greatest creation". Falstaff appears in three of Shakespeare's plays, "Henry IV, Part 1", "Henry IV, Part 2", and "The Merry Wives of Windsor". His death is mentioned in "Henry V" but he has no lines, nor is it directed that he appear on stage. However, many stage and film adaptations have seen it necessary to include Falstaff for the insight he provides into King Henry V's character. The most notable examples in cinema are Laurence Olivier's 1944 version and Kenneth Branagh's 1989 film, both of which draw additional material from the "Henry IV" plays. The character is known to have been very popular with audiences at the time, and for many years afterwards. According to Leonard Digges, writing shortly after Shakespeare's death, while many plays could not get good audiences, "let but Falstaff come, Hal, Poins, the rest, you scarce shall have a room". King Henry is troubled by the behaviour of his son and heir, the Prince of Wales. Hal (the future Henry V) has forsaken the Royal Court to waste his time in taverns with low companions. This makes him an object of scorn to the nobles and calls into question his royal worthiness. Hal's chief friend and foil in living the low life is Sir John Falstaff. Fat, old, drunk, and corrupt as he is, he has a charisma and a zest for life that captivates the Prince. Hal likes Falstaff but makes no pretense at being like him. He enjoys insulting his dissolute friend and makes sport of him by joining in Poins' plot to disguise themselves and rob and terrify Falstaff and three friends of loot they have stolen in a highway robbery, purely for the fun of watching Falstaff lie about it later, after which Hal returns the stolen money. Rather early in the play, in fact, Hal informs us that his riotous time will soon come to a close, and he will re-assume his rightful high place in affairs by showing himself worthy to his father and others through some (unspecified) noble exploits. Hal believes that this sudden change of manner will amount to a greater reward and acknowledgment of prince-ship, and in turn earn him respect from the members of the court. On the way to this climax, we are treated to Falstaff, who has "misused the King's press damnably", not only by taking money from able-bodied men who wished to evade service but by keeping the wages of the poor souls he brought instead who were killed in battle ("food for powder, food for powder"). Left on his own during Hal's battle with Hotspur, Falstaff dishonourably counterfeits death to avoid attack by Douglas. After Hal leaves Hotspur's body on the field, Falstaff revives in a mock miracle. Seeing he is alone, he stabs Hotspur's corpse in the thigh and claims credit for the kill. Though Hal knows better, he allows Falstaff his disreputable tricks. Soon after being given grace by Hal, Falstaff states that he wants to amend his life and begin "to live cleanly as a nobleman should do". The play focuses on Prince Hal's journey toward kingship, and his ultimate rejection of Falstaff. However, unlike "Part One", Hal's and Falstaff's stories are almost entirely separate, as the two characters meet only twice and very briefly. The tone of much of the play is elegiac, focusing on Falstaff's age and his closeness to death, which parallels that of the increasingly sick king. Falstaff is still drinking and engaging in petty criminality in the London underworld. He first appears, followed by a new character, a young page whom Prince Hal has assigned him as a joke. Falstaff enquires what the doctor has said about the analysis of his urine, and the page cryptically informs him that the urine is healthier than the patient. Falstaff delivers one of his most characteristic lines: "I am not only witty in myself, but the cause that wit is in other men." Falstaff promises to outfit the page in "vile apparel" (ragged clothing). He then complains of his insolvency, blaming it on "consumption of the purse." They go off, Falstaff vowing to find a wife "in the stews" (i.e., the local brothels). The Lord Chief Justice enters, looking for Falstaff. Falstaff at first feigns deafness in order to avoid conversing with him, and when this tactic fails pretends to mistake him for someone else. As the Chief Justice attempts to question Falstaff about a recent robbery, Falstaff insists on turning the subject of the conversation to the nature of the illness afflicting the King. He then adopts the pretense of being a much younger man than the Chief Justice: "You that are old consider not the capacities of us that are young." Finally, he asks the Chief Justice for one thousand pounds to help outfit a military expedition, but is denied. He has a relationship with Doll Tearsheet, a prostitute, who gets into a fight with Ancient Pistol, Falstaff's ensign. After Falstaff ejects Pistol, Doll asks him about the Prince. Falstaff is embarrassed when his derogatory remarks are overheard by Hal, who is present disguised as a musician. Falstaff tries to talk his way out of it, but Hal is unconvinced. When news of a second rebellion arrives, Falstaff joins the army again, and goes to the country to raise forces. There he encounters an old school friend, Justice Shallow, and they reminisce about their youthful follies. Shallow brings forward potential recruits for the loyalist army: Mouldy, Bullcalf, Feeble, Shadow and Wart, a motley collection of rustic yokels. Falstaff and his cronies accept bribes from two of them, Mouldy and Bullcalf, not to be conscripted. In the final scene, Falstaff, having learned from Pistol that Hal is now King, travels to London in expectation of great rewards. But Hal rejects him, saying that he has now changed, and can no longer associate with such people. The London lowlifes, expecting a paradise of thieves under Hal's governance, are instead purged and imprisoned by the authorities. Although Falstaff does not appear on stage in "Henry V", his death is the main subject of Act 2, Scene 3, in which Mistress Quickly delivers a memorable eulogy: Falstaff arrives in Windsor very short on money. To obtain financial advantage, he decides to court two wealthy married women, Mistress Ford and Mistress Page. Falstaff decides to send the women identical love letters and asks his servants – Pistol and Nym – to deliver them to the wives. When they refuse, Falstaff sacks them, and, in revenge, the men tell Ford and Page (the husbands) of Falstaff's intentions. Page is not concerned, but the jealous Ford persuades the Host of the Garter Inn to introduce him to Falstaff as a 'Master Brook' so that he can find out Falstaff's plans. When the women receive the letters, each goes to tell the other, and they quickly find that the letters are almost identical. The "merry wives" are not interested in the ageing, overweight Falstaff as a suitor; however, for the sake of their own amusement and to gain revenge for his indecent assumptions towards them both, they pretend to respond to his advances. This all results in great embarrassment for Falstaff. Mr. Ford poses as 'Mr. Brook' and says he is in love with Mistress Ford but cannot woo her as she is too virtuous. He offers to pay Falstaff to court her, saying that once she has lost her honour he will be able to tempt her himself. Falstaff cannot believe his luck, and tells 'Brook' he has already arranged to meet Mistress Ford while her husband is out. Falstaff leaves to keep his appointment and Ford soliloquises that he is right to suspect his wife and that the trusting Page is a fool. When Falstaff arrives to meet Mistress Ford, the merry wives trick him into hiding in a laundry basket ("buck basket") full of filthy, smelly clothes awaiting laundering. When the jealous Ford returns to try and catch his wife with the knight, the wives have the basket taken away and the contents (including Falstaff) dumped into the river. Although this affects Falstaff's pride, his ego is surprisingly resilient. He is convinced that the wives are just playing hard to get with him, so he continues his pursuit of sexual advancement, with its attendant capital and opportunities for blackmail. Again Falstaff goes to meet the women but Mistress Page comes back and warns Mistress Ford of her husband's approach again. They try to think of ways to hide him other than the laundry basket which he refuses to get into again. They trick him again, this time into disguising himself as Mistress Ford's maid's obese aunt, known as "the fat woman of Brentford". Ford tries once again to catch his wife with the knight but ends up beating the "old woman", whom he despises, and throwing her out of his house. Black and blue, Falstaff laments his bad luck. Eventually the wives tell their husbands about the series of jokes they have played on Falstaff, and together they devise one last trick which ends up with the Knight being humiliated in front of the whole town. They tell Falstaff to dress as "Herne, the Hunter" and meet them by an old oak tree in Windsor Forest (now part of Windsor Great Park). They then dress several of the local children as fairies and get them to pinch and burn Falstaff to punish him. The wives meet Falstaff, and almost immediately the "fairies" attack. After the chaos, the characters reveal their true identities to Falstaff. Although he is embarrassed, Falstaff takes the joke surprisingly well, as he sees it was what he deserved. Ford says he must pay back the 20 pounds 'Brook' gave him and takes the Knight's horses as recompense. Eventually they all leave together and Mistress Page even invites Falstaff to come with them: "let us every one go home, and laugh this sport o'er by a country fire; Sir John and all". Shakespeare originally named Falstaff "John Oldcastle". Lord Cobham, a descendant of the historical John Oldcastle, complained, forcing Shakespeare to change the name. Shakespeare's "Henry IV" plays and "Henry V" adapted and developed the material in an earlier play called "The Famous Victories of Henry V", in which Sir John "Jockey" Oldcastle appears as a dissolute companion of the young Henry. Prince Hal refers to Falstaff as "my old lad of the castle" in the first act of the play; the epilogue to "Henry IV, Part 2", moreover, explicitly disavows any connection between Falstaff and Oldcastle: "Oldcastle died a martyr, and this is not the man." The historical Oldcastle was a Lollard who was executed for heresy and rebellion, and he was respected by many Protestants as a martyr. In addition to the anonymous "The Famous Victories of Henry V", in which Oldcastle is Henry V's companion, Oldcastle's history is described in Raphael Holinshed's "Chronicles", Shakespeare's usual source for his histories. It is not clear, however, if Shakespeare characterised Falstaff as he did for dramatic purposes, or because of a specific desire to satirise Oldcastle or the Cobhams. Cobham was a common butt of veiled satire in Elizabethan popular literature; he figures in Ben Jonson's "Every Man in His Humour" and may have been part of the reason "The Isle of Dogs" was suppressed. Shakespeare's desire to burlesque a hero of early English Protestantism could indicate Roman Catholic sympathies, but Henry Brooke, 11th Baron Cobham was sufficiently sympathetic to Catholicism that in 1603, he was imprisoned as part of the Main Plot to place Arbella Stuart on the English throne, so if Shakespeare wished to use Oldcastle to embarrass the Cobhams, he seems unlikely to have done so on religious grounds. The Cobhams appear to have intervened while Shakespeare was in the process of writing either "The Merry Wives of Windsor" or the second part of "Henry IV". The first part of "Henry IV" was probably written and performed in 1596, and the name Oldcastle had almost certainly been allowed by Master of the Revels Edmund Tilney. William Brooke, 10th Baron Cobham may have become aware of the offensive representation after a public performance; he may also have learned of it while it was being prepared for a court performance (Cobham was at that time Lord Chamberlain). As father-in-law to the newly widowed Robert Cecil, Cobham certainly possessed the influence at court to get his complaint heard quickly. Shakespeare may have included a sly retaliation against the complaint in his play "The Merry Wives of Windsor" (published after the "Henry IV" series). In the play, the paranoid, jealous Master Ford uses the alias "Brook" to fool Falstaff, perhaps in reference to William Brooke. At any rate, the name is Falstaff in the "Henry IV, Part 1" quarto, of 1598, and the epilogue to the second part, published in 1600, contains this clarification: The new name "Falstaff" probably derived from the medieval knight Sir John Fastolf (who may also have been a Lollard). The historical John Fastolf fought at the Battle of Patay against Joan of Arc, which the English lost. Fastolf's previous actions as a soldier had earned him wide respect, but he seems to have become a scapegoat after the debacle. He was among the few English military leaders to avoid death or capture during the battle, and although there is no evidence that he acted with cowardice, he was temporarily stripped of his knighthood. Fastolf appears in "Henry VI, Part 1" in which he is portrayed as an abject coward. In the" First Folio "his name is spelled "Falstaffe", so Shakespeare may have directly appropriated the spelling of the name he used in the earlier play. In a further comic double meaning, the name implies impotence. It has been suggested that the dissolute writer Robert Greene may also have been an inspiration for the character of Falstaff. This theory was first proposed in 1930 and has recently been championed by Stephen Greenblatt. Notorious for a life of dissipation and debauchery somewhat similar to Falstaff, he was among the first to mention Shakespeare in his work (in "Greene's Groats-Worth of Wit"), suggesting to Greenblatt that the older writer may have influenced Shakespeare's characterisation. There are several works about Falstaff, inspired by Shakespeare's plays: All references to "Henry IV, Part 1", "Henry IV, Part 2", and "The Merry Wives of Windsor", unless otherwise specified, are taken from the Folger Shakespeare Library's "Folger Digital Editions" texts edited by Barbara Mowat, Paul Werstine, Michael Poston, and Rebecca Niles. Under their referencing system, 3.1.55 means act 3, scene 1, line 55. Prologues, epilogues, scene directions, and other parts of the play that are not a part of character speech in a scene, are referenced using Folger Through Line Number: a separate line numbering scheme that includes every line of text in the play.
https://en.wikipedia.org/wiki?curid=11246
Floorball Floorball is a type of floor hockey with five players and a goalkeeper in each team. Men and women play indoors with sticks and a plastic ball with holes. Matches are played in three twenty-minute periods. Floorball was included in the World Games for the first time in 2017 in Wroclaw, Poland. Sweden were the first World Games gold medal winners. The game was invented in Sweden in the late 1960s. The basic rules were established in 1979 when the first floorball club in the world, Sala IBK, from Sala, was founded in Sweden. Official rules for matches were first written down in 1981. The sport is organized internationally by the International Floorball Federation (IFF). As of 2014, there are over 300,133 registered floorball players worldwide. Events include an annual Euro Floorball Cup for club teams and the biennial World Floorball Championships with separate divisions for men and women. Norway, Czechia, Finland, Sweden, and Switzerland have consistently placed 1st, 2nd, or 3rd at the World Floorball Championships. Professional club leagues include Finland's Salibandyliiga, Sweden's Svenska Superligan and Czech's Superliga florbalu. While the IFF contains 74 members, floorball is most popular where it has been developed the longest, such as Czechia, Denmark, Estonia, Finland, Latvia, Norway, Sweden, and Switzerland. It is gaining popularity in Australia, New Zealand, India, Canada, Germany, Ireland, Japan, Singapore, Malaysia, the United States. and the United Kingdom. In various forms the game has been played since the early 20th Century in Canada as a recreational sport, especially in high school gymnasiums, as a playful variant of hockey, where the sticks got their form from the hockey game bandy. Most Canadian males born in the 1950s and before could attest to this. Similarly, during the 1950s and 1960s many public school systems within Michigan incorporated floorball into their primary and secondary school gym classes. Later, Americans claimed to have invented it, and held interstate tournaments in the 1960s. The game was formally organized as an international and more organised sport in the late 1970s in Gothenburg, Sweden. The sport began as something that was played for fun as a pastime in schools. After a decade or so, floorball began showing up in Nordic countries where the former schoolyard pastime was becoming a developed sport. Formal rules soon were developed, and clubs began to form. After some time, several countries developed national associations, and the IFF was founded in 1986. The game of floorball is also known by many other names, such as salibandy (in Finland), innebandy (in Sweden and Norway), and unihockey (in Switzerland and Ireland). The names "salibandy" and "innebandy" are derived from bandy; they translate to "hall bandy" and "indoor bandy" respectively. Unihockey is derived from "universal hockey" since it is meant to be a special and simplified hockey form. When the IFF was founded in 1986, the sport was played in mostly Nordic countries, Japan, and several parts of Europe. By 1990, floorball was recognized in 7 countries, and by the time of the first European Floorball Championships in 1994, that number had risen to 14. That number included the United States, who was the first country outside Europe and Asia to recognize floorball. By the time of the first men's world championships in 1996, 20 nations played floorball, with 12 of them participating at the tournament. As of 2009, the sport of floorball has been played in almost 80 countries. Of those, 58 have national floorball associations that are recognized by the IFF. With the addition of Sierra Leone, Africa's first floorball nation, the IFF has at least one national association on each continent of the world, with the exception of Antarctica. 10 years after the IFF was founded, the first world championships were played, with a sold out final of 15,106 people at the Globen in Stockholm, Sweden. In addition to that, the world's two largest floorball leagues, Finland's Salibandyliiga and Sweden's Svenska Superligan were formed, in 1986 and 1995 respectively. In December 2008, the IFF and the sport of floorball received recognition from the International Olympic Committee (IOC). In July 2011, the IOC officially welcomed the IFF into its family of Recognised International Sports Federations (ARISF). This will pave the way for floorball to enter the official sport programme. The IFF hopes that this recognition will help allow floorball to become a part of the 2020 Summer Olympics. In January 2009, the IFF and the sport of floorball received recognition from the Special Olympics. In addition to recognition by the International Olympic Committee and Special Olympics, the IFF is also a member of the Sport Accord – formerly known as General Association of International Sports Federations (GAISF), and co-operates with the International University Sports Federation (FISU). Floorball is now also member of IWGA which runs the World Games and floorball will be for the first time on programme in Wrocław 2017. The world floorball championships are an annual event where teams from across the world gather to play in a tournament in order to win the world championship. As of 2011, eight Men's, eight Women's, six Men's Under-19, and four Women's Under-19 World Floorball Championships have taken place. Czechia, Finland, Norway, Sweden, and Switzerland remain the only five countries to have ever captured a medal at a World Championship event. From 1996 to 2009, the IFF used a World Floorball Championship format where the last team in the A-Division was relegated to the B-Division, while the top team in the B-Division was promoted to the A-Division. This format caused much hardship for countries such as Australia, Canada, Slovakia, and Spain, who have all been trying to get to the B-Division from the C-Division since 2004. In 2010, the IFF adopted a FIFA-like continental qualification system, where teams must qualify to play at the world championships. Depending on the number of countries registered per continent or region, the IFF gives spots for the world championships. For example, Argentina, Brazil, Canada and the United States would need to play for one spot at the world championships in a continental qualification tournament for the Americas. Floorball is played indoors on a rink whose size can officially vary from wide to long, and which is surrounded by high enclosed boards with rounded corners. The goals are wide and high. Their depth is and they are from the end of the nearest boards. Face-off dots are marked on the center line. Dots are also marked from both sides of the rink on the goal lines imaginary extensions. The dots don't exceed in diameter. They do not have to be dots, they can also be crosses. Typical equipment for a floorball player consists of a stick, a pair of shorts, a shirt, socks, and indoor sport shoes. Players may wear shin guards, eye protectors and protective padding for vital areas although most do not. Protective eyewear is, in some countries, compulsory for junior players. A floorball stick is short compared with one for ice hockey; the maximum size for a stick is 114 cm. As a stick cannot weigh any more than 350 grams, floorball sticks are often made of carbon and composite materials. The blade of the stick can either be "right" or "left" which indicates which way stick is supposed to be held from the players point of view. A player who is right-handed will often use a "left" blade since this he/she will be holding the stick to left and the other way around for left-handed people. Goalkeepers wear limited protection provided by padded pants, a padded chest protector, knee pads and a helmet. Some goalkeepers like to wear gloves and/or wristbands The goalkeeper may also wear other protective equipment such as elbow pads and jock straps but bulky padding is not permitted. Goalkeepers do not use sticks and may use their hands to play the ball when they are within the goalkeeper's box. There, they are allowed to throw the ball out to their teammates provided that the ball touches the ground before the half court mark. When they are completely outside the box, goalkeepers are considered field players and are not allowed to touch the ball with their hands. A floor ball weighs and its diameter is . It has 26 holes in it, each of which are in diameter. Many of these balls now are made with aerodynamic technology, where the ball has over a thousand small dimples in it that reduce air resistance. There have been several times where a ball has been recorded to have traveled at a speed of approximately . Each team can field six players at a time on the court, one player being a goalkeeper. But the coach can take the goalkeeper off and switch him with a field player whenever they like. Although it usually only happens in the end to score with one more outfield player. This can make advantages for the attacking side of the team, and disadvantages when it comes to their own defense. The team is also allowed to change players any time in the game; usually, they change the whole team. Individual substitution happens sometimes, but it usually happens when a player is exhausted or is hurt. A floorball game is officially played over three periods lasting 20 minutes each (15 minutes for juniors). The time is stopped in the case of penalties, goals, time-outs and any situation where the ball is not considered to be in play. The signal of a timeout should be a triple sound. An intermission of 10 minutes (or maximum 15 minutes in some competitions) takes place between each period, where teams change ends and substitution areas. Each team is allowed one timeout of 30 seconds (which is often used late in matches). There are two referees to oversee the game, each with equal authority. If a game ends in a tie, teams play ten minutes extra, and the team that scores first wins.. If the game is still drawn after extra-time, a penalty shootout similar to ice-hockey decides over the winner. Checking is prohibited in floorball. Controlled shoulder-to-shoulder contact is allowed, but ice hockey-like checking is forbidden. Pushing players without the ball or competing for a loose ball is also disallowed, and many of these infractions lead to two-minute penalties. The best comparison in terms of legal physical contact is Association football (soccer), where checking is used to improve one's positioning in relation to the ball rather than to remove an opposing player from the play. In addition to checking, players cannot lift an opponent’s stick or perform any stick infractions in order to get to the ball. As well, players may not raise their stick or play the ball above their knee level, and a stick may not be placed between a player's legs. Passing the ball by foot is allowed, but only once, after that it the ball has to be touched by the stick. After stopping the ball by foot the ball has to be touched with the stick before it can be passed to a fellow player by foot (Rule change 2014). Passing by hand or head deliberately may result in a two minutes penalty for the offending player. A field player may not enter the marked goal area and playing without stick is prohibited. When a player commits a foul, or when the ball is deemed unplayable, the play is resumed from a free-hit or a face-off. A free hit involves a player from one team to start play from the place where the ball was last deemed unplayable. A comparison of this is a free kick in soccer. For many fouls, such as stick infractions, a free hit is the only discipline provided. However, at the referee's discretion, a penalty may be worth either two or five minutes. At that point, the player who committed the foul sits in the penalty area, and his team is shorthanded for the time of the penalty. If an 'extreme' foul is committed, such as physical contact or unsportsmanlike behavior, a player may receive a 10-minute penalty or even a match misconduct. Two-minute penalties can arise from a number of infractions and result in the offending player being sat on a penalty seat next to the scorers/timekeepers and away from the team benches. Each penalty has a specific code that is recorded on the official match record along with the time of the foul. The team of the offending player will play short-handed for the full length of the penalty. The codes are as follows; Two Minute Penalties 5 Minute Penalties Personal Fouls/Penalties Freebandy is a sport that developed in the 2000s from floorball fanatics who specialize in a technique called "zorro", which involves lifting the ball onto a stick and allowing air resistance and fast movements to keep the ball "stuck" to the stick. This technique is also referred to as "airhooking" or "skyhooking". In freebandy, the rules are very much the same of those of floorball, with the exception of high nets and no infractions for high sticking. As well, the sticks are slightly tweaked from those of a floorball variety to include a "pocket" where the ball can be placed. Floorball at the Special Olympics is slightly modified from the "regular" form of floorball. Matches are played 3-on-3 with a goaltender, on a smaller court that measures long by wide. This form of floorball was developed for the intellectually disabled, and has yet to be played at the Special Olympics. Floorball was played as a demonstration sport at the 2013 Special Olympics World Winter Games, and was played as an official sport at the games in 2017. A simplified less formal version of Floorball, played with smaller team numbers and shorter periods, and typically outdoors on various surfaces, including AstroTurf. In its most basic form, it is an informal pick up game amongst friends. However, a more formal version is played in Sweden, with the following structure: Swiss floorball called unihockey is a revised version of a floorball match. The match is played on a slightly smaller court and often involves only three field players playing on each side, in 3-on-3 floorball. This form of floorball is also slightly shorter, with only two periods of 15 to 20 minutes each played. In Switzerland this form of playing is called "smallcourt" (Kleinfeld), opposed to the usual style of playing on a bigger court, which is called "bigcourt" (Grossfeld). Originally developed for players with disabilities, wheelchair floorball is played with exactly the same rules as "regular" floorball. Players use the same stick and ball, and goaltenders are also allowed to play. The first ever IFF-sanctioned wheelchair floorball matches were played between the men's teams of Czechia and Sweden, during the 2008 Men's World Floorball Championships in Prague, Czechia. In addition to this, there is also an electric wheelchair variation. In addition to the Floorball World Championships, there are other IFF Events for club teams such as the Champions Cup which is for the national competition winners from the Top-4 ranked nations, and the EuroFloorball Cup for the national competition winners from the 5th and lower ranked nations. There are also many international floorball club competitions. The Asia Pacific Floorball Championships are played every single year in New Zealand, Australia, Singapore, or Japan. The event was created by the Singapore Floorball Association together with the cooperation of the Asia Oceania Floorball Confederation (AOFC). Members of the AOFC get together during this tournament to play for the Asia Pacific Floorball Championship every year. The most recent champions are Men's national floorball team of Australia. As of 2010, the Asia Pacific Floorball Championship is also the qualifying tournament for the World Floorball Championships. The Canada Cup is an international club tournament that is held every year in Toronto, Ontario, Canada. It is the largest floorball club tournament outside of Europe, and attracts 55+ clubs from worldwide, every year. The winners of the 2015 Canada Cup floorball tournament were the Ottawa Blizzard floorball club. The USA Cup is an international club tournament that is held annually in Dallas–Fort Worth metroplex, United States. It is a part of NAFC series with Canada Cup and Falls floorball classic. The NAFC - North American Floorball Championship is an international club tournament series currently including three tournaments: Canada Cup, Falls floorball classic and USA Cup. Teams must attend at least one tournament in Canada and one in the United States with a roster that is made up of at least 50% of the same players and using the same club name to be eligible. The world's largest club team tournament, the Czech Open is a traditional summer tournament held in Prague, Czechia. It is famous not only for its on-court activities, but also for those off-court. The tournament attracts 200+ clubs every year from 20 different countries. The most recent champions of the prestigious Czech Open are Sweden's IBF Falun. The Champions Cup was played for the first time in 2011. It is now the premier IFF event for Men's and Women's Club teams. The national championship winners from the Top-4 ranked nations receive automatic qualification, while the remaining two spots are made up by a team nominated by the Local Event organiser and the winner of the previous year's EuroFloorball Cup. In 2011, the winners were Men: SSV Helsinki (Finland) and Women: IF Djurgårdens IBF (Sweden). The EuroFloorball Cup (formerly European Cup) is an IFF-organised club event for both men's and women's teams. It has taken place every single year since 1993, and in 2000 it changed its format to a 2-year event (i.e. 2000–01). In 2008, the tournament switched back to its one-year format. In 2011 it underwent another change when the Champions Cup was introduced for the first time. The EuroFloorball Cup (EFC) is now for the national competition winners from the 5th and lower ranked nations. Qualification can be made via a number of processes. Firstly, the teams from the 5th, 6th & 7th ranked nations receive automatic qualification. A team nominated by the local event organiser also gets automatic qualification, and then the last two spots are determined by qualification tournaments. The winner of the EFC qualifies to play in the Champions Cup the following year. The reigning EuroFloorball Cup Champions as of October 2011, are Men: RTU/Inspecta (Latvia)and Women: RSU/Runway (Latvia).
https://en.wikipedia.org/wiki?curid=11247
Edinburgh Edinburgh (; ; ) is the capital of Scotland and one of its 32 council areas. Historically part of the county of Midlothian (interchangeably Edinburghshire before 1921), it is located in Lothian on the Firth of Forth's southern shore. Recognised as the capital of Scotland since at least the 15th century, Edinburgh is the seat of the Scottish Government, the Scottish Parliament and the supreme courts of Scotland. The city's Palace of Holyroodhouse is the official residence of the monarch in Scotland. The city has long been a centre of education, particularly in the fields of medicine, Scots law, literature, philosophy, the sciences and engineering. It is the second largest financial centre in the United Kingdom (after London) and the city's historical and cultural attractions have made it the United Kingdom's second most visited tourist destination attracting 4.9 million visits including 2.4 million from overseas in 2018. Edinburgh is Scotland's second most populous city (behind Glasgow) and the seventh most populous in the United Kingdom. The official population estimates are 488,050 (2016) for the Locality of Edinburgh (Edinburgh pre 1975 regionalisation plus Currie and Balerno), 518,500 (2018) for the City of Edinburgh, and 1,339,380 (2014) for the city region. Edinburgh lies at the heart of the Edinburgh and South East Scotland city region comprising East Lothian, Edinburgh, Fife, Midlothian, Scottish Borders and West Lothian. The city is the annual venue of the General Assembly of the Church of Scotland. It is home to national institutions such as the National Museum of Scotland, the National Library of Scotland and the Scottish National Gallery. The University of Edinburgh, founded in 1582 and now one of four in the city, is placed 20th in the QS World University Rankings for 2020. The city is also known for the Edinburgh International Festival and the Fringe, the latter being the world's largest annual international arts festival. Historic sites in Edinburgh include Edinburgh Castle, the Palace of Holyroodhouse, the churches of St. Giles, Greyfriars and the Canongate, and the extensive Georgian New Town built in the 18th/19th centuries. Edinburgh's Old Town and New Town together are listed as a UNESCO World Heritage site, which has been managed by Edinburgh World Heritage since 1999. "Edin", the root of the city's name, derives from "Eidyn", the name for this region in Cumbric, the Brittonic Celtic language formerly spoken there. The name's meaning is unknown. The district of Eidyn centred on the stronghold Din Eidyn, the dun or hillfort of Eidyn. This stronghold is believed to have been located at Castle Rock, now the site of Edinburgh Castle. Eidyn was conquered by the Angles of Bernicia in the 7th century and later by the Scots in the 10th century. As the language shifted to Old English, and subsequently to modern English and Scots, the Brittonic "din" in Din Eidyn was replaced by "burh", producing "Edinburgh". Similarly, "din" became "dùn" in Scottish Gaelic, producing "Dùn Èideann". The city is affectionately nicknamed "Auld Reekie", Scots for "Old Smoky", for the views from the country of the smoke-covered Old Town. A remark on a poem in an 1800 collection of the poems of Allan Ramsay said, "Auld Reeky. A name the country people give Edinburgh from the cloud of smoke or reek that is always impending over it." Thomas Carlyle said, "Smoke cloud hangs over old Edinburgh,—for, ever since Aeneas Silvius's time and earlier, the people have the art, very strange to Aeneas, of burning a certain sort of black stones, and Edinburgh with its chimneys is called 'Auld Reekie' by the country people." A character in Walter Scott's "The Abbot" says "... yonder stands Auld Reekie—you may see the smoke hover over her at twenty miles' distance." Robert Chambers who said that the sobriquet could not be traced before the reign of Charles II attributed the name to a Fife laird, Durham of Largo, who regulated the bedtime of his children by the smoke rising above Edinburgh from the fires of the tenements. "It's time now bairns, to tak' the beuks, and gang to our beds, for yonder's Auld Reekie, I see, putting on her nicht -cap!" Some have called Edinburgh the "Athens of the North" for a variety of reasons. The earliest comparison between the two cities showed that they had a similar topography, with the Castle Rock of Edinburgh performing a similar role to the Athenian Acropolis. Both of them had flatter, fertile agricultural land sloping down to a port several miles away (respectively Leith and Piraeus). Although this arrangement is common in Southern Europe, it is rare in Northern Europe. The 18th-century intellectual life, referred to as the Scottish Enlightenment, was a key influence in gaining the name. Such luminaries as David Hume and Adam Smith shone during this period. Having lost most of its political importance after the Union, some hoped that Edinburgh could gain a similar influence on London as Athens had on Rome. Also a contributing factor was the later neoclassical architecture, particularly that of William Henry Playfair, and the National Monument. Tom Stoppard's character Archie, of "Jumpers", said, perhaps playing on Reykjavík meaning "smoky bay", that the "Reykjavík of the South" would be more appropriate. The city has also been known by several Latin names such as "Edinburgum" while the adjectival forms "Edinburgensis" and "Edinensis" are used in educational and scientific contexts. "Edina" is a late 18th century poetical form used by the Scots poets Robert Fergusson and Robert Burns. "Embra" or "Embro" are colloquialisms from the same time, as in Robert Garioch's "Embro to the Ploy". Ben Jonson described it as "Britaine's other eye", and Sir Walter Scott referred to it as "yon Empress of the North". Robert Louis Stevenson, also a son of the city, wrote that Edinburgh "is what Paris ought to be." The earliest known human habitation in the Edinburgh area was at Cramond, where evidence was found of a Mesolithic camp site dated to c. 8500 BC. Traces of later Bronze Age and Iron Age settlements have been found on Castle Rock, Arthur's Seat, Craiglockhart Hill and the Pentland Hills. When the Romans arrived in Lothian at the end of the 1st century AD, they found a Brittonic Celtic tribe whose name they recorded as the Votadini. The Votadini transitioned into the Gododdin kingdom in the Early Middle Ages, with Eidyn serving as one of the kingdom's districts. During this period, the Castle Rock site, thought to have been the stronghold of Din Eidyn, emerged as the kingdom's major centre. The medieval poem "Y Gododdin" describes a war band from across the Brittonic world who gathered in Eidyn before a fateful raid; this may describe a historical event around AD 600. In 638, the Gododdin stronghold was besieged by forces loyal to King Oswald of Northumbria, and around this time control of Lothian passed to the Angles. Their influence continued for the next three centuries until around 950, when, during the reign of Indulf, son of Constantine II, the "burh" (fortress), named in the 10th-century "Pictish Chronicle" as "oppidum Eden", was abandoned to the Scots. It thenceforth remained under their jurisdiction. The royal burgh was founded by King David I in the early 12th century on land belonging to the Crown, though the date of its charter is unknown. The first documentary evidence of the medieval burgh is a royal charter, , by King David I granting a toft in to the Priory of Dunfermline. By the middle of the 14th century, the French chronicler Jean Froissart was describing it as the capital of Scotland (c. 1365), and James III (1451–88) referred to it in the 15th century as "the principal burgh of our kingdom". Despite the destruction caused by an English assault in 1544, the town slowly recovered, and was at the centre of events in the 16th-century Scottish Reformation and 17th-century Wars of the Covenant. In 1603, King James VI of Scotland succeeded to the English throne, uniting the crowns of Scotland and England in a personal union known as the Union of the Crowns, though Scotland remained, in all other respects, a separate kingdom. In 1638, King Charles I's attempt to introduce Anglican church forms in Scotland encountered stiff Presbyterian opposition culminating in the conflicts of the Wars of the Three Kingdoms. Subsequent Scottish support for Charles Stuart's restoration to the throne of England resulted in Edinburgh's occupation by Oliver Cromwell's Commonwealth of England forces – the New Model Army – in 1650. In the 17th century, Edinburgh's boundaries were still defined by the city's defensive town walls. As a result, the city's growing population was accommodated by increasing the height of the houses. Buildings of 11 storeys or more were common, and have been described as forerunners of the modern-day skyscraper. Most of these old structures were replaced by the predominantly Victorian buildings seen in today's Old Town. Following the Treaty of Union in 1706, the Parliaments of England and Scotland passed Acts of Union in 1706 and 1707 respectively, uniting the two kingdoms in the Kingdom of Great Britain effective from 1 May 1707. As a consequence, the Parliament of Scotland merged with the Parliament of England to form the Parliament of Great Britain, which sat at Westminster in London. The Union was opposed by many Scots, resulting in riots in the city. By the first half of the 18th century, Edinburgh was described as one of Europe's most densely populated, overcrowded and unsanitary towns. Visitors were struck by the fact that the various social classes shared the same urban space, even inhabiting the same tenement buildings; although here a form of social segregation did prevail, whereby shopkeepers and tradesmen tended to occupy the cheaper-to-rent cellars and garrets, while the more well-to-do professional classes occupied the more expensive middle storeys. During the Jacobite rising of 1745, Edinburgh was briefly occupied by the Jacobite "Highland Army" before its march into England. After its eventual defeat at Culloden, there followed a period of reprisals and pacification, largely directed at the rebellious clans. In Edinburgh, the Town Council, keen to emulate London by initiating city improvements and expansion to the north of the castle, reaffirmed its belief in the Union and loyalty to the Hanoverian monarch George III by its choice of names for the streets of the New Town: for example, Rose Street and Thistle Street; and for the royal family, George Street, Queen Street, Hanover Street, Frederick Street and Princes Street (in honour of George's two sons). In the second half of the century, the city was at the heart of the Scottish Enlightenment, when thinkers like David Hume, Adam Smith, James Hutton and Joseph Black were familiar figures in its streets. Edinburgh became a major intellectual centre, earning it the nickname "Athens of the North" because of its many neo-classical buildings and reputation for learning, recalling ancient Athens. In the 18th-century novel "The Expedition of Humphry Clinker" by Tobias Smollett one character describes Edinburgh as a "hotbed of genius". Edinburgh was also a major centre for the Scottish book trade. The highly successful London bookseller Andrew Millar was apprenticed there to James McEuen. From the 1770s onwards, the professional and business classes gradually deserted the Old Town in favour of the more elegant "one-family" residences of the New Town, a migration that changed the city's social character. According to the foremost historian of this development, "Unity of social feeling was one of the most valuable heritages of old Edinburgh, and its disappearance was widely and properly lamented." Although Edinburgh's traditional industries of printing, brewing and distilling continued to grow in the 19th century, and were joined by new rubber works and engineering works, there was little industrialisation compared with other cities in Britain. By 1821, Edinburgh had been overtaken by Glasgow as Scotland's largest city. The city centre between Princes Street and George Street became a major commercial and shopping district, a development partly stimulated by the arrival of railways in the 1840s. The Old Town became an increasingly dilapidated, overcrowded slum with high mortality rates. Improvements carried out under Lord Provost William Chambers in the 1860s began the transformation of the area into the predominantly Victorian Old Town seen today. More improvements followed in the early 20th century as a result of the work of Patrick Geddes, but relative economic stagnation during the two world wars and beyond saw the Old Town deteriorate further before major slum clearance in the 1960s and 1970s began to reverse the process. University building developments which transformed the George Square and Potterrow areas proved highly controversial. Since the 1990s a new "financial district", including the Edinburgh International Conference Centre, has grown mainly on demolished railway property to the west of the castle, stretching into Fountainbridge, a run-down 19th-century industrial suburb which has undergone radical change since the 1980s with the demise of industrial and brewery premises. This ongoing development has enabled Edinburgh to maintain its place as the United Kingdom's second largest financial and administrative centre after London. Financial services now account for a third of all commercial office space in the city. The development of Edinburgh Park, a new business and technology park covering , west of the city centre, has also contributed to the District Council's strategy for the city's major economic regeneration. In 1998, the Scotland Act, which came into force the following year, established a devolved Scottish Parliament and Scottish Executive (renamed the Scottish Government since September 2007). Both based in Edinburgh, they are responsible for governing Scotland while reserved matters such as defence, taxation and foreign affairs remain the responsibility of the Parliament of the United Kingdom in London. Situated in Scotland's Central Belt, Edinburgh lies on the Firth of Forth's southern shore. The city centre is southwest of the shoreline of Leith and inland, as the crow flies, from the east coast of Scotland and the North Sea at Dunbar. While the early burgh grew up near the prominent Castle Rock, the modern city is often said to be built on seven hills, namely Calton Hill, Corstorphine Hill, Craiglockhart Hill, Braid Hill, Blackford Hill, Arthur's Seat and the Castle Rock, giving rise to allusions to the seven hills of Rome. Occupying a narrow gap between the Firth of Forth to the north and the Pentland Hills and their outrunners to the south, the city sprawls over a landscape which is the product of early volcanic activity and later periods of intensive glaciation. Igneous activity between 350 and 400 million years ago, coupled with faulting, led to the creation of tough basalt volcanic plugs, which predominate over much of the area. One such example is the Castle Rock which forced the advancing ice sheet to divide, sheltering the softer rock and forming a tail of material to the east, thus creating a distinctive crag and tail formation. Glacial erosion on the north side of the crag gouged a deep valley later filled by the now drained Nor Loch. These features, along with another hollow on the rock's south side, formed an ideal natural strongpoint upon which Edinburgh Castle was built. Similarly, Arthur's Seat is the remains of a volcano dating from the Carboniferous period, which was eroded by a glacier moving west to east during the ice age. Erosive action such as plucking and abrasion exposed the rocky crags to the west before leaving a tail of deposited glacial material swept to the east. This process formed the distinctive Salisbury Crags, a series of teschenite cliffs between Arthur's Seat and the location of the early burgh. The residential areas of Marchmont and Bruntsfield are built along a series of drumlin ridges south of the city centre, which were deposited as the glacier receded. Other prominent landforms such as Calton Hill and Corstorphine Hill are also products of glacial erosion. The Braid Hills and Blackford Hill are a series of small summits to the south of the city centre that command expansive views looking northwards over the urban area to the Firth of Forth. Edinburgh is drained by the river named the Water of Leith, which rises at the Colzium Springs in the Pentland Hills and runs for through the south and west of the city, emptying into the Firth of Forth at Leith. The nearest the river gets to the city centre is at Dean Village on the north-western edge of the New Town, where a deep gorge is spanned by Thomas Telford's Dean Bridge, built in 1832 for the road to Queensferry. The Water of Leith Walkway is a mixed-use trail that follows the course of the river for from Balerno to Leith. Excepting the shoreline of the Firth of Forth, Edinburgh is encircled by a green belt, designated in 1957, which stretches from Dalmeny in the west to Prestongrange in the east. With an average width of the principal objectives of the green belt were to contain the outward expansion of the city and to prevent the agglomeration of urban areas. Expansion affecting the green belt is strictly controlled but developments such as Edinburgh Airport and the Royal Highland Showground at Ingliston lie within the zone. Similarly, suburbs such as Juniper Green and Balerno are situated on green belt land. One feature of the Edinburgh green belt is the inclusion of parcels of land within the city which are designated green belt, even though they do not connect with the peripheral ring. Examples of these independent wedges of green belt include Holyrood Park and Corstorphine Hill. Edinburgh includes former towns and villages that retain much of their original character as settlements in existence before they were absorbed into the expanding city of the nineteenth and twentieth centuries. Many areas, such as Dalry, contain residences that are multi-occupancy buildings known as tenements, although the more southern and western parts of the city have traditionally been less built-up with a greater number of detached and semi-detached villas. The historic centre of Edinburgh is divided in two by the broad green swathe of Princes Street Gardens. To the south, the view is dominated by Edinburgh Castle, built high on Castle Rock, and the long sweep of the Old Town descending towards Holyrood Palace. To the north lie Princes Street and the New Town. The West End includes the financial district, with insurance and banking offices as well as the Edinburgh International Conference Centre. Edinburgh's Old and New Towns were listed as a UNESCO World Heritage Site in 1995 in recognition of the unique character of the Old Town with its medieval street layout and the planned Georgian New Town, including the adjoining Dean Village and Calton Hill areas. There are over 4,500 listed buildings within the city, a higher proportion relative to area than any other city in the United Kingdom. The Royal Mile runs downhill and terminates at Holyrood Palace. Minor streets (called closes or wynds) lie on either side of the main spine forming a herringbone pattern. The street has several fine public buildings such as St Giles' Cathedral, the City Chambers and the Law Courts. Other places of historical interest nearby are Greyfriars Kirkyard and the Grassmarket. The street layout is typical of the old quarters of many northern European cities. The castle is perched on top of a rocky crag (the remnant of an extinct volcano) and the Royal Mile runs down the crest of a ridge from it. Due to space restrictions imposed by the narrowness of this landform, the Old Town became home to some of the earliest "high rise" residential buildings. Multi-storey dwellings known as "lands" were the norm from the 16th century onwards with ten and eleven storeys being typical and one even reaching fourteen or fifteen storeys. Numerous vaults below street level were inhabited to accommodate the influx of incomers, particularly Irish immigrants, during the Industrial Revolution. The New Town was an 18th-century solution to the problem of an increasingly crowded city which had been confined to the ridge sloping down from the castle. In 1766 a competition to design a "New Town" was won by James Craig, a 27-year-old architect. The plan was a rigid, ordered grid, which fitted in well with Enlightenment ideas of rationality. The principal street was to be George Street, running along the natural ridge to the north of what became known as the "Old Town". To either side of it are two other main streets: Princes Street and Queen Street. Princes Street has become Edinburgh's main shopping street and now has few of its Georgian buildings in their original state. The three main streets are connected by a series of streets running perpendicular to them. The east and west ends of George Street are terminated by St Andrew Square and Charlotte Square respectively. The latter, designed by Robert Adam, influenced the architectural style of the New Town into the early 19th century. Bute House, the official residence of the First Minister of Scotland, is on the north side of Charlotte Square. The hollow between the Old and New Towns was formerly the Nor Loch, which was created for the town's defence but came to be used by the inhabitants for dumping their sewage. It was drained by the 1820s as part of the city's northward expansion. Craig's original plan included an ornamental canal on the site of the loch, but this idea was abandoned. Soil excavated while laying the foundations of buildings in the New Town was dumped on the site of the loch to create the slope connecting the Old and New Towns known as The Mound. In the middle of the 19th century the National Gallery of Scotland and Royal Scottish Academy Building were built on The Mound, and tunnels for the railway line between Haymarket and Waverley stations were driven through it. The Southside is a residential part of the city, which includes the districts of St Leonards, Marchmont, Morningside, Newington, Sciennes, the Grange and Blackford. The Southside is broadly analogous to the area covered formerly by the Burgh Muir, and was developed as a residential area after the opening of the South Bridge in the 1780s. The Southside is particularly popular with families (many state and private schools are here), young professionals and students (the central University of Edinburgh campus is based around George Square just north of Marchmont and the Meadows), and Napier University (with major campuses around Merchiston and Morningside). The area is also well provided with hotel and "bed and breakfast" accommodation for visiting festival-goers. These districts often feature in works of fiction. For example, Church Hill in Morningside, was the home of Muriel Spark's Miss Jean Brodie, and Ian Rankin's Inspector Rebus lives in Marchmont and works in St Leonards. Leith was historically the port of Edinburgh, an arrangement of unknown date that was confirmed by the royal charter Robert the Bruce granted to the city in 1329. The port developed a separate identity from Edinburgh, which to some extent it still retains, and it was a matter of great resentment when the two burghs merged in 1920 into the City of Edinburgh. Even today the parliamentary seat is known as "Edinburgh North and Leith". The loss of traditional industries and commerce (the last shipyard closed in 1983) resulted in economic decline. The Edinburgh Waterfront development has transformed old dockland areas from Leith to Granton into residential areas with shopping and leisure facilities and helped rejuvenate the area. With the redevelopment, Edinburgh has gained the business of cruise liner companies which now provide cruises to Norway, Sweden, Denmark, Germany, and the Netherlands. The coastal suburb of Portobello is characterised by Georgian villas, Victorian tenements, a beach and promenade and cafés, bars, restaurants and independent shops. There are rowing and sailing clubs and a restored Victorian swimming pool, including Turkish baths. The urban area of Edinburgh is almost entirely within the City of Edinburgh Council boundary, merging with Musselburgh in East Lothian. Towns within easy reach of the city boundary include Haddington, Tranent, Prestonpans, Dalkeith, Bonnyrigg, Loanhead, Penicuik, Broxburn, Livingston and Dunfermline. Edinburgh lies at the heart of the Edinburgh & South East Scotland City region with a population in 2014 of 1,339,380. Like most of Scotland, Edinburgh has a temperate, maritime climate which, despite its northerly latitude, is milder than places which lie at similar latitudes such as Moscow and Labrador. The city's proximity to the sea mitigates any large variations in temperature or extremes of climate. Winter daytime temperatures rarely fall below freezing while summer temperatures are moderate, rarely exceeding . The highest temperature recorded in the city was on 25 July 2019 at Gogarbank, beating the previous record of on 4 August 1975 at Edinburgh Airport. The lowest temperature recorded in recent years was during December 2010 at Gogarbank. In an average year, the temperature will drop to a minimum of . Given Edinburgh's position between the coast and hills, it is renowned as "the windy city", with the prevailing wind direction coming from the south-west, which is often associated with warm, unstable air from the North Atlantic Current that can give rise to rainfall – although considerably less than cities to the west, such as Glasgow. Rainfall is distributed fairly evenly throughout the year. Winds from an easterly direction are usually drier but considerably colder, and may be accompanied by haar, a persistent coastal fog. Vigorous Atlantic depressions, known as European windstorms, can affect the city between October and May. Located slightly north of the city centre, the weather station at the Royal Botanic Garden Edinburgh (RBGE) has been an official weather station for the Met Office since 1956. The Met Office operates its own weather station at Gogarbank on the city's western outskirts, near Edinburgh Airport. This slightly inland station has a slightly wider temperature span between seasons, is cloudier and somewhat wetter, but differences are minor. Temperature and rainfall records have been kept at the Royal Observatory since 1764. In that time, the warmest month on record was July 1779, with an average temperature of , whereas the coldest was January 1814, with a mean temperature of . The warmest years on record are 1779 and 1846, both with mean temperatures of . The coldest year on record is 1879, with a mean temperature of . The wettest month on record was August 1948, with . The driest was February 1934, with . The wettest year on record was 1916, with . The driest year on record was 1826, with of rainfall. The most recent official population estimates are 512,150 (2016) for the Edinburgh settlement (includes Musselburgh) and 518,500 (2018) for the local authority area. Edinburgh has a high proportion of young adults, with 19.5% of the population in their 20s (exceeded only by Aberdeen) and 15.2% in their 30s which is the highest in Scotland. The proportion of Edinburgh's population born in the UK fell from 92% to 84% between 2001 and 2011, while the proportion of White Scottish-born fell from 78% to 70%. Of those Edinburgh residents born in the UK, 335,000 or 83% were born in Scotland, with 58,000 or 14% being born in England. Some 13,000 people or 2.7% of the city's population are of Polish descent. 39,500 people or 8.2% of Edinburgh's population class themselves as Non-White which is an increase from 4% in 2001. Of the Non-White population, the largest group by far are Asian, totalling 26,264 people. Within the Asian population, people of Chinese descent are now the largest sub-group, with 8,076 people, amounting to about 1.7% of the city's total population. The city's population of Indian descent amounts to 6,470 (1.4% of the total population), while there are some 5,858 of Pakistani descent (1.2% of the total population). Although they account for only 1,277 people or 0.3% of the city's population, Edinburgh has the highest number and proportion of people of Bangladeshi descent in Scotland. Over 7,000 people were born in African countries (1.6% of the total population) and nearly 7,000 in the Americas. With the notable exception of Inner London, Edinburgh has a higher number of people born in the United States (over 3,700) than any other city in the UK. The proportion of people born outside the UK was 15.9% compared with 8% in 2001. A census by the Edinburgh presbytery in 1592 recorded a population of 8,003 adults spread equally north and south of the High Street which runs along the spine of the ridge sloping down from the Castle. In the 18th and 19th centuries, the population expanded rapidly, rising from 49,000 in 1751 to 136,000 in 1831, primarily due to migration from rural areas. As the population grew, problems of overcrowding in the Old Town, particularly in the cramped tenements that lined the present day Royal Mile and the Cowgate, were exacerbated. Poor sanitary arrangements resulted in a high incidence of disease, with outbreaks of cholera occurring in 1832, 1848 and 1866. The construction of the New Town from 1767 onwards witnessed the migration of the professional and business classes from the difficult living conditions in the Old Town to the lower density, higher quality surroundings taking shape on land to the north. Expansion southwards from the Old Town saw more tenements being built in the 19th century, giving rise to Victorian suburbs such as Dalry, Newington, Marchmont and Bruntsfield. Early 20th-century population growth coincided with lower-density suburban development. As the city expanded to the south and west, detached and semi-detached villas with large gardens replaced tenements as the predominant building style. Nonetheless, the 2001 census revealed that over 55% of Edinburgh's population were still living in tenements or blocks of flats, a figure in line with other Scottish cities, but much higher than other British cities, and even central London. From the early to mid 20th century the growth in population, together with slum clearance in the Old Town and other areas, such as Dumbiedykes, Leith, and Fountainbridge, led to the creation of new estates such as Stenhouse and Saughton, Craigmillar and Niddrie, Pilton and Muirhouse, Piershill, and Sighthill. In 2018 the Church of Scotland had 20,956 members in 71 congregations in the Presbytery of Edinburgh.. Its most prominent church is St Giles' on the Royal Mile, first dedicated in 1243 but believed to date from before the 12th century. Saint Giles is historically the patron saint of Edinburgh. St Cuthbert's, situated at the west end of Princes Street Gardens in the shadow of Edinburgh Castle and St Giles' can lay claim to being the oldest Christian sites in the city, though the present St Cuthbert's, designed by Hippolyte Blanc, was dedicated in 1894. Other Church of Scotland churches include Greyfriars Kirk, the Canongate Kirk, St Andrew's and St George's West Church and the Barclay Church. The Church of Scotland Offices are in Edinburgh, as is the Assembly Hall where the annual General Assembly is held. The Roman Catholic Archdiocese of St Andrews and Edinburgh has 27 parishes across the city. The Archbishop of St Andrews and Edinburgh has his official residence in Greenhill, and the diocesan offices are in nearby Marchmont. The Diocese of Edinburgh of the Scottish Episcopal Church has over 50 churches, half of them in the city. Its centre is the late-19th-century Gothic style St Mary's Cathedral in the West End's Palmerston Place. Orthodox Christianity is represented by Pan, Romanian and Russian Orthodox churches. There are several independent churches in the city, both Catholic and Protestant, including Charlotte Chapel, Carrubbers Christian Centre, Bellevue Chapel and Sacred Heart. There are also churches belonging to Quakers, Christadelphians, Seventh-day Adventists, Church of Christ, Scientist, The Church of Jesus Christ of Latter-day Saints (LDS Church) and Elim Pentecostal Church. Muslims have several places of worship across the city. Edinburgh Central Mosque is the largest Islamic place of worship, is located in Potterrow, on the city's Southside, near Bristo Square. Construction was largely financed by a gift from King Fahd of Saudi Arabia and was completed in 1998. There is also an active presence of the Ahmadiyya Muslim community. The first recorded presence of a Jewish community in Edinburgh dates back to the late 18th century. Edinburgh's Orthodox synagogue, opened in 1932, is in Salisbury Road and can accommodate a congregation of 2000. A Liberal Jewish congregation also meets in the city. A Sikh gurdwara and a Hindu mandir are located in Leith. The city also has a Brahma Kumaris centre in the Polwarth area. The Edinburgh Buddhist Centre, run by the Triratna Buddhist Community, formerly situated in Melville Terrace, now runs sessions at the Healthy Life Centre, Bread Street. Other Buddhist traditions are represented by groups which meet in the capital: the Community of Interbeing (followers of Thich Nhat Hanh), Rigpa, Samye Dzong, Theravadin, Pure Land and Shambala. There is a Sōtō Zen Priory in Portobello and a Theravadin Thai Buddhist Monastery in Slateford Road. Edinburgh is home to an active Bahá'í Community, and a Theosophical Society meets in Great King Street. Edinburgh has an active Inter-Faith Association. Edinburgh has the strongest economy of any city in the United Kingdom outside London and the highest percentage of professionals in the UK with 43% of the population holding a degree-level or professional qualification. According to the Centre for International Competitiveness, it is the most competitive large city in the United Kingdom. It also has the highest gross value added per employee of any city in the UK outside London, measuring £57,594 in 2010. It was named European "Best Large City of the Future for Foreign Direct Investment" and "Best Large City for Foreign Direct Investment Strategy" in the" Financial Times "fDi magazine awards 2012/13. In the 19th century, Edinburgh's economy was known for banking, publishing and brewing. Today, its economy is based mainly on financial services, scientific research, higher education, and tourism. In March 2010, unemployment in Edinburgh was comparatively low at 3.6%, and it remains consistently below the Scottish average of 4.5%. Edinburgh is the 2nd most visited city by foreign visitors in the UK after London. Banking has been a mainstay of the Edinburgh economy for over 300 years, since the Bank of Scotland (now part of the Lloyds Banking Group) was established by an act of the Scottish Parliament in 1695. Today, the financial services industry, with its particularly strong insurance and investment sectors, and underpinned by Edinburgh-based firms such as Scottish Widows and Standard Life Aberdeen, accounts for the city being the UK's second financial centre after London and Europe's fourth in terms of equity assets. The Royal Bank of Scotland opened new global headquarters at Gogarburn in the west of the city in October 2005, and Edinburgh is home to the headquarters of Bank of Scotland, Sainsbury's Bank, Tesco Bank, and TSB Bank. Tourism is also an important element in the city's economy. As a World Heritage Site, tourists visit historical sites such as Edinburgh Castle, the Palace of Holyroodhouse and the Old and New Towns. Their numbers are augmented in August each year during the Edinburgh Festivals, which attracts 4.4 million visitors, and generates over £100m for the local economy. As the centre of Scotland's government and legal system, the public sector plays a central role in Edinburgh's economy. Many departments of the Scottish Government are in the city. Other major employers include NHS Scotland and local government administration. When the £1.3bn Edinburgh & South East Scotland City Region Deal was signed in 2018, the region's Gross Value Added (GVA) contribution to the Scottish economy was cited as £33bn, or 33% of the country's ouput. But the Deal's partners noted that prosperity was not evenly spread across the city region, citing 22.4% of children living in poverty and a shortage of affordable housing. The city hosts a series of festivals that run between the end of July and early September each year. The best known of these events are the Edinburgh Festival Fringe, the Edinburgh International Festival, the Edinburgh Military Tattoo, the Edinburgh Art Festival and the Edinburgh International Book Festival. The longest established of these festivals is the Edinburgh International Festival, which was first held in 1947 and consists mainly of a programme of high-profile theatre productions and classical music performances, featuring international directors, conductors, theatre companies and orchestras. This has since been overtaken in size by the Edinburgh Fringe which began as a programme of marginal acts alongside the "official" Festival and has become the world's largest performing arts festival. In 2017, nearly 3400 different shows were staged in 300 venues across the city. Comedy has become one of the mainstays of the Fringe, with numerous well-known comedians getting their first 'break' there, often by being chosen to receive the Edinburgh Comedy Award. The Edinburgh Military Tattoo, occupies the Castle Esplanade every night for three weeks each August, with massed pipe bands and military bands drawn from around the world. Performances end with a short fireworks display. As well as the various summer festivals, many other festivals are held during the rest of the year, including the Edinburgh International Film Festival and Edinburgh International Science Festival. The annual Edinburgh Hogmanay celebration was originally an informal street party focused on the Tron Kirk in the Old Town's High Street. Since 1993, it has been officially organised with the focus moved to Princes Street. In 1996, over 300,000 people attended, leading to ticketing of the main street party in later years up to a limit of 100,000 tickets. Hogmanay now covers four days of processions, concerts and fireworks, with the street party beginning on Hogmanay. Alternative tickets are available for entrance into the Princes Street Gardens concert and Cèilidh, where well-known artists perform and ticket holders can participate in traditional Scottish cèilidh dancing. The event attracts thousands of people from all over the world. On the night of 30 April the Beltane Fire Festival takes place on Calton Hill, involving a procession followed by scenes inspired by pagan old spring fertility celebrations. At the beginning of October each year the Dussehra Hindu Festival is also held on Calton Hill. Outside the Festival season, Edinburgh supports several theatres and production companies. The Royal Lyceum Theatre has its own company, while the King's Theatre, Edinburgh Festival Theatre and Edinburgh Playhouse stage large touring shows. The Traverse Theatre presents a more contemporary repertoire. Amateur theatre companies productions are staged at the Bedlam Theatre, Church Hill Theatre and King's Theatre among others. The Usher Hall is Edinburgh's premier venue for classical music, as well as occasional popular music concerts. It was the venue for the Eurovision Song Contest 1972. Other halls staging music and theatre include The Hub, the Assembly Rooms and the Queen's Hall. The Scottish Chamber Orchestra is based in Edinburgh. Edinburgh has two repertory cinemas, the Edinburgh Filmhouse and The Cameo, as well as the independent Dominion Cinema and a range of multiplexes. Edinburgh has a healthy popular music scene. Occasionally large concerts are staged at Murrayfield and Meadowbank, while mid-sized events take place at smaller venues such as 'The Corn Exchange', 'The Liquid Rooms' and 'The Bongo Club'. In 2010, PRS for Music listed Edinburgh among the UK's top ten 'most musical' cities. Several city pubs are well known for their live performances of folk music. They include 'Sandy Bell's' in Forrest Road, 'Captain's Bar' in South College Street and 'Whistlebinkies' in South Bridge. Like many other cities in the UK, numerous nightclub venues host Electronic dance music events. Edinburgh is home to a flourishing group of contemporary composers such as Nigel Osborne, Peter Nelson, Lyell Cresswell, Hafliði Hallgrímsson, Edward Harper, Robert Crawford, Robert Dow and John McLeod. McLeod's music is heard regularly on BBC Radio 3 and throughout the UK. The main local newspaper is the "Edinburgh Evening News". It is owned and published alongside its sister titles "The Scotsman" and "Scotland on Sunday" by JPIMedia. The city has two commercial radio stations: Forth 1, a station which broadcasts mainstream chart music, and Forth 2 on medium wave which plays classic hits. Capital Radio Scotland and Eklipse Sports Radio also have transmitters covering Edinburgh. Along with the UK national radio stations, Radio Scotland and the Gaelic language service BBC Radio nan Gàidheal are also broadcast. DAB digital radio is broadcast over two local multiplexes. BFBS Radio broadcasts from studios on the base at Dreghorn Barracks across the city on 98.5FM as part of its UK Bases network Television, along with most radio services, is broadcast to the city from the Craigkelly transmitting station situated in Fife on the opposite side of the Firth of Forth and the Black Hill transmitting station in North Lanarkshire to the west. There are currently no television stations based in the city. Edinburgh Television existed in the late 1990s to early 2003 and STV Edinburgh existed from 2015 to 2018. Edinburgh has many museums and libraries. These include the National Museum of Scotland, the National Library of Scotland, National War Museum, the Museum of Edinburgh, Surgeons' Hall Museum, the Writers' Museum, the Museum of Childhood and Dynamic Earth. The Museum on The Mound has exhibits on money and banking. Edinburgh Zoo, covering on Corstorphine Hill, is the second most visited paid tourist attraction in Scotland, and currently home to two giant pandas, Tian Tian and Yang Guang, on loan from the People's Republic of China. Edinburgh is also home to The Royal Yacht Britannia, decommissioned in 1997 and now a five-star visitor attraction and evening events venue permanently berthed at Ocean Terminal. Edinburgh contains Scotland's three National Galleries of Art as well as numerous smaller art galleries. The national collection is housed in the Scottish National Gallery, located on The Mound, comprising the linked National Gallery of Scotland building and the Royal Scottish Academy building. Contemporary collections are shown in the Scottish National Gallery of Modern Art which occupies a split site at Belford. The Scottish National Portrait Gallery on Queen Street focuses on portraits and photography. The council-owned City Art Centre in Market Street mounts regular art exhibitions. Across the road, The Fruitmarket Gallery offers world-class exhibitions of contemporary art, featuring work by British and international artists with both emerging and established international reputations. The city hosts several of Scotland's galleries and organisations dedicated to contemporary visual art. Significant strands of this infrastructure include: Creative Scotland, Edinburgh College of Art, Talbot Rice Gallery (University of Edinburgh), Collective (based at the City Observatory) and the Edinburgh Annuale. There are also many small private shops/galleries that provide space to showcase works from local artists. The locale around Princes Street is the main shopping area in the city centre, with souvenir shops, chain stores such as Boots the Chemist, Edinburgh Woollen Mill, H&M and Jenners. George Street, north of Princes Street, is the preferred location for some upmarket shops and independent stores. At the east end of Princes Street , the St. James Centre is currently being redeveloped while next to the Balmoral Hotel and Waverley Station is the Waverley Mall. Multrees Walk, adjacent to the St. James Centre, is a recent addition to the central shopping district, dominated by the presence of Harvey Nichols. Shops here include Louis Vuitton, Mulberry and Calvin Klein. Edinburgh also has substantial retail parks outside the city centre. These include The Gyle Shopping Centre and Hermiston Gait in the west of the city, Cameron Toll Shopping Centre, Straiton Retail Park (actually just outside the city, in Midlothian) and Fort Kinnaird in the south and east, and Ocean Terminal in the north on the Leith waterfront. Following local government reorganisation in 1996, The City of Edinburgh Council constitutes one of the 32 council areas of Scotland. Like all other local authorities of Scotland, the council has powers over most matters of local administration such as housing, planning, local transport, parks, economic development and regeneration. The council comprises 58 elected councillors, returned from 17 multi-member electoral wards in the city. Following the 2007 City of Edinburgh Council election the incumbent Labour Party lost majority control of the council after 23 years to a Liberal Democrat/SNP coalition. The 2012 City of Edinburgh Council election saw a Scottish Labour/SNP coalition. The 2017 City of Edinburgh Council election, saw a continuation of this administration, but with the SNP as the largest party. The city's coat of arms was registered by the Lord Lyon King of Arms in 1732. Edinburgh, like all of Scotland, is represented in the Scottish Parliament. For electoral purposes, the city is divided into six constituencies which, along with 3 seats outside of the city, form part of the Lothian region. Each constituency elects one Member of the Scottish Parliament (MSP) by the first past the post system of election, and the region elects seven additional MSPs to produce a result based on a form of proportional representation. As of the 2016 election, the Scottish National Party have three MSPs: Ash Denham for Edinburgh Eastern, Ben Macpherson for Edinburgh Northern and Leith and Gordon MacDonald for Edinburgh Pentlands constituencies. Alex Cole-Hamilton of the Scottish Liberal Democrats represents Edinburgh Western, Daniel Johnson of the Scottish Labour Party represents Edinburgh Southern constituency, and former Leader of the Scottish Conservative Party, Ruth Davidson currently represents the Edinburgh Central constituency. Edinburgh is also represented in the House of Commons of the United Kingdom by five Members of Parliament. The city is divided into Edinburgh North and Leith, Edinburgh East, Edinburgh South, Edinburgh South West, and Edinburgh West, each constituency electing one member by the first past the post system. Edinburgh is currently represented by three MPs affiliated with the Scottish National Party, one Liberal Democrat MP in Edinburgh West and one Labour MP in Edinburgh South. Edinburgh Airport is Scotland's busiest and biggest airport and the principal international gateway to the capital, handling over 12 million passengers in 2016. In anticipation of rising passenger numbers, the former operator of the airport BAA outlined a draft masterplan in 2011 to provide for the expansion of the airfield and the terminal building. In June 2012, Global Infrastructure Partners purchased the airport for £807 million. The possibility of building a second runway to cope with an increased number of aircraft movements has also been mooted. Travel in Edinburgh is undertaken predominantly by bus. Lothian Buses, the successor company to Edinburgh Corporation Transport Department, operate the majority of city bus services within the city and to surrounding suburbs, with the most routes running via Princes Street. Services further afield operate from the Edinburgh Bus Station off St Andrew Square and Waterloo Place and are operated mainly by Stagecoach East Scotland, Scottish Citylink, National Express Coaches and Borders Buses. Lothian Buses also operates all of the city's branded public tour buses, night bus service and airport bus link. In 2010, Lothian Buses recorded 109 million passenger journeys – a 1.9% rise on the previous year. Edinburgh Waverley is the second-busiest railway station in Scotland, with only Glasgow Central handling more passengers. On the evidence of passenger entries and exits between April 2015 and March 2016, Edinburgh Waverley is the fifth-busiest station outside London; it is also the UK's second biggest station in terms of the number of platforms and area size. Waverley is the terminus for most trains arriving from London King's Cross and the departure point for many rail services within Scotland operated by Abellio ScotRail. To the west of the city centre lies Haymarket Station which is an important commuter stop. Opened in 2003, Edinburgh Park station serves the Gyle business park in the west of the city and the nearby Gogarburn headquarters of the Royal Bank of Scotland. The Edinburgh Crossrail route connects Edinburgh Park with Haymarket, Edinburgh Waverley and the suburban stations of Brunstane and Newcraighall in the east of the city. There are also commuter lines to South Gyle and Dalmeny, the latter serving South Queensferry by the Forth Bridges, and to Wester Hailes and Curriehill in the south-west of the city. To tackle traffic congestion, Edinburgh is now served by six park and ride sites on the periphery of the city at Sheriffhall (in Midlothian), Ingliston, Riccarton, Inverkeithing (in Fife), Newcraighall and Straiton (in Midlothian). A referendum of Edinburgh residents in February 2005 rejected a proposal to introduce congestion charging in the city. Edinburgh Trams became operational on 31 May 2014. The city had been without a tram system since Edinburgh Corporation Tramways ceased on 16 November 1956. Following parliamentary approval in 2007, construction began in early 2008. The first stage of the project was expected to be completed by July 2011 but, following delays caused by extra utility work and a long-running contractual dispute between the Council and the main contractor, Bilfinger SE, the project was rescheduled. The cost of the project rose from the original projection of £545 million to £750 million in mid-2011 and some suggest it could eventually exceed £1 billion. The completed line is in length, running from Edinburgh Airport, west of the city, to its current terminus at York Place in the city centre's East End. It was originally planned to continue down Leith Walk to Ocean Terminal and terminate at Newhaven. Should the original plan be taken to completion, trams will also run from Haymarket through Ravelston and Craigleith to Granton Square on the Waterfront Edinburgh. Long-term proposals envisage a line running west from the airport to Ratho and Newbridge and another connecting Granton Square to Newhaven via Lower Granton Road, thus completing the Line 1 (North Edinburgh) loop. A further line serving the south of the city has also been suggested. Lothian Buses and Edinburgh Trams are both owned and operated by Transport for Edinburgh. There are three universities in Edinburgh, the University of Edinburgh, Heriot-Watt University and Edinburgh Napier University. Established by royal charter in 1583, the University of Edinburgh is one of Scotland's ancient universities and is the fourth oldest in the country after St Andrews, Glasgow and Aberdeen. Originally centred on Old College the university expanded to premises on The Mound, the Royal Mile and George Square. Today, the King's Buildings in the south of the city contain most of the schools within the College of Science and Engineering. In 2002, the medical school moved to purpose built accommodation adjacent to the new Royal Infirmary of Edinburgh at Little France. The University is placed 20th in the QS World University Rankings for 2020. Heriot-Watt University is based at the Riccarton campus in the west of Edinburgh. Originally established in 1821 as the world's first mechanics' institute it was granted university status by royal charter in 1966. It has other campuses in the Scottish Borders, Orkney, United Arab Emirates and Putrajaya in Malaysia. It takes the name Heriot-Watt from Scottish inventor James Watt and Scottish philanthropist and goldsmith George Heriot. Heriot-Watt University has been named International University of the Year by The Times and Sunday Times Good University Guide 2018. In the latest Research Excellence Framework, it was ranked overall in the Top 25% of UK universities and 1st in Scotland for research impact. Edinburgh Napier University was originally founded as the Napier College which was renamed Napier Polytechnic in 1986 and gained university status in 1992. Edinburgh Napier University has campuses in the south and west of the city, including the former Merchiston Tower and Craiglockhart Hydropathic. It is home to the Screen Academy Scotland. Queen Margaret University was located in Edinburgh before it moved to a new campus on the edge of Musselburgh in 2008. Until 2012 further education colleges in the city included Jewel and Esk College (incorporating Leith Nautical College founded in 1903), Telford College, opened in 1968, and Stevenson College, opened in 1970. These have now been amalgamated to form Edinburgh College. Scotland's Rural College also has a campus in south Edinburgh. Other institutions include the Royal College of Surgeons of Edinburgh and the Royal College of Physicians of Edinburgh which were established by royal charter in 1506 and 1681 respectively. The Trustees Drawing Academy of Edinburgh, founded in 1760, became the Edinburgh College of Art in 1907. There are 18 nursery, 94 primary and 23 secondary schools administered by the City of Edinburgh Council. Edinburgh is home to The Royal High School, one of the oldest schools in the country and the world. The city also has several independent, fee-paying schools including Edinburgh Academy, Fettes College, George Heriot's School, George Watson's College, Merchiston Castle School, Stewart's Melville College and The Mary Erskine School. In 2009, the proportion of pupils attending independent schools was 24.2%, far above the Scottish national average of just over 7% and higher than in any other region of Scotland. In August 2013, the City of Edinburgh Council opened the city's first stand-alone Gaelic primary school, Bun-sgoil Taobh na Pàirce. The main NHS Lothian hospitals serving the Edinburgh area are the Royal Infirmary of Edinburgh, which includes the University of Edinburgh Medical School, and the Western General Hospital, which has a large cancer treatment centre and nurse-led Minor Injuries Clinic. The Royal Edinburgh Hospital in Morningside specialises in mental health. The Royal Hospital for Sick Children, colloquially referred to as 'the Sick Kids', is a specialist paediatrics hospital. There are two private hospitals: Murrayfield Hospital in the west of the city and Shawfair Hospital in the south. Both are owned by Spire Healthcare. Edinburgh has three football clubs that play in the Scottish Professional Football League (SPFL): Heart of Midlothian, founded in 1874, Hibernian, founded in 1875 and Edinburgh City, founded in 1966. Heart of Midlothian and Hibernian are known locally as "Hearts" and "Hibs" respectively, both play in the Scottish Premiership. They are the oldest city rivals in Scotland and the Edinburgh derby is one of the oldest derby matches in world football. Both clubs have won the Scottish league championship four times. Hearts have won the Scottish Cup eight times and the Scottish League Cup four times. Hibs have won the Scottish Cup and the Scottish League Cup three times each. Edinburgh City were promoted to Scottish League Two in the 2015–16 season, becoming the first club to win promotion to the SPFL via the pyramid system playoffs. Edinburgh was also home to four other former Scottish Football League clubs: the original Edinburgh City, Leith Athletic, Meadowbank Thistle and St Bernard's. Meadowbank Thistle played at Meadowbank Stadium until 1995, when the club moved to Livingston and became Livingston F.C. The Scottish national team has very occasionally played at Easter Road and Tynecastle, although its normal home stadium is Hampden Park in Glasgow. St Bernard's' New Logie Green was used to host the 1896 Scottish Cup Final, the only time the match has been played outside Glasgow. The city also plays host to Lowland Football League clubs Civil Service Strollers, Edinburgh University and Spartans, as well as East of Scotland League clubs Craigroyston, Edinburgh United, Heriot-Watt University, Leith Athletic, Lothian Thistle Hutchison Vale, and Tynecastle. In women's football, Hearts, Hibs and Spartans play in the SWPL 1. Hutchison Vale play in the SWPL 2. The Scotland national rugby union team and the professional Edinburgh Rugby team play at Murrayfield Stadium, which is owned by the Scottish Rugby Union and also used for other events, including music concerts. It is the largest capacity stadium in Scotland, seating 67,144 spectators. Edinburgh is also home to Scottish Premiership teams Boroughmuir RFC, Currie RFC, the Edinburgh Academicals, Heriot's Rugby Club and Watsonians RFC. Rugby league is represented by the Edinburgh Eagles who play in the Rugby League Conference Scotland Division. Murrayfield Stadium has hosted the Magic Weekend where all Super League matches are played in the stadium over one weekend. The Scottish cricket team, which represents Scotland internationally, play their home matches at the Grange cricket club. The Murrayfield Racers are the latest of a succession of ice hockey clubs in the Scottish capital. Previously Edinburgh was represented by the Edinburgh Capitals "(who folded in 2018)", the original Murrayfield Racers "(who folded in 1996)" and the Edinburgh Racers. The club play their home games at the Murrayfield Ice Rink and have competed in the eleven-team professional Scottish National League (SNL) since the 2018–19 season. Next door to Murrayfield Ice Rink is a 7-sheeter dedicated curling facility where curling is played from October to March each season. Caledonia Pride are the only women's professional basketball team in Scotland. Established in 2016, the team compete in the UK wide Women's British Basketball League and play their home matches at the Oriam National Performance Centre. Edinburgh also has several men's basketball teams within the Scottish National League. Boroughmuir Blaze, City of Edinburgh Kings, Edinburgh Lions and Edinburgh University all compete in Division 1 of the National League, and Pleasance B.C. compete in Division 2. Boroughmuir won the league in 2016, and won the playoffs in the same year, beating the University in the final. The Edinburgh Diamond Devils is a baseball club which won its first Scottish Championship in 1991 as the "Reivers." 1992 saw the team repeat the achievement, becoming the first team to do so in league history. The same year saw the start of their first youth team, the Blue Jays. The club adopted its present name in 1999. Edinburgh has also hosted national and international sports events including the World Student Games, the 1970 British Commonwealth Games, the 1986 Commonwealth Games and the inaugural 2000 Commonwealth Youth Games. For the 1970 Games the city built Olympic standard venues and facilities including Meadowbank Stadium and the Royal Commonwealth Pool. The Pool underwent refurbishment in 2012 and hosted the Diving competition in the 2014 Commonwealth Games which were held in Glasgow. In American football, the Scottish Claymores played WLAF/NFL Europe games at Murrayfield, including their World Bowl 96 victory. From 1995 to 1997 they played all their games there, from 1998 to 2000 they split their home matches between Murrayfield and Glasgow's Hampden Park, then moved to Glasgow full-time, with one final Murrayfield appearance in 2002. The city's most successful non-professional team are the Edinburgh Wolves who play at Meadowbank Stadium. The Edinburgh Marathon has been held annually in the city since 2003 with more than 16,000 runners taking part on each occasion. Its organisers have called it "the fastest marathon in the UK" due to the elevation drop of . The city also organises a half-marathon, as well as 10 km () and 5 km () races, including a race on 1 January each year. Edinburgh has a speedway team, the Edinburgh Monarchs, which, since the loss of its stadium in the city, has raced at the Lothian Arena in Armadale, West Lothian. The Monarchs have won the Premier League championship five times in their history, in 2003 and again in 2008, 2010, 2014 and 2015. Edinburgh has a long literary tradition, which became especially evident during the Scottish Enlightenment. This heritage and the city's lively literary life in the present led to it being declared the first UNESCO City of Literature in 2004. Prominent authors who have lived in Edinburgh include the economist Adam Smith, born in Kirkcaldy and author of "The Wealth of Nations", Ian Rankin, author of the Inspector Rebus series of crime thrillers, Alexander McCall Smith, author of the No. 1 Ladies' Detective Agency series, and J. K. Rowling, creator of Harry Potter, who began her first book in an Edinburgh coffee shop and who lives in the Cramond area of the city. Scotland has a rich history of science and engineering, with Edinburgh producing a number of leading figures. John Napier, inventor of logarithms, was born in Merchiston Tower and lived and died in the city. His house now forms part of the original campus of Napier University which was named in his honour. He lies buried under St. Cuthbert's Church. James Clerk Maxwell, founder of the modern theory of electromagnetism, was born at 14 India Street (now the home of the James Clerk Maxwell Foundation) and educated at the Edinburgh Academy and the University of Edinburgh, as was the engineer and telephone pioneer Alexander Graham Bell. James Braidwood, who organised Britain's first municipal fire brigade, was also born in the city and began his career there. Other names connected with the city include Max Born, physicist and Nobel laureate; Charles Darwin, the biologist who propounded the theory of natural selection; David Hume, philosopher, economist and historian; James Hutton, regarded as the "Father of Geology"; Joseph Black, the chemist and one of the founders of thermodynamics; pioneering medical researchers Joseph Lister and James Young Simpson; chemist and discoverer of the element nitrogen Daniel Rutherford; Colin Maclaurin, mathematician and developer of the Maclaurin series, and Ian Wilmut, the geneticist involved in the cloning of Dolly the sheep just outside Edinburgh. The stuffed carcass of Dolly the sheep is now on display in the National Museum of Scotland. The latest in a long line of science celebrities associated with the city is theoretical physicist and Nobel Prizewinner Professor Emeritus Peter Higgs, born in Newcastle but resident in Edinburgh for most of his academic career, after whom the Higgs boson particle has been named. Edinburgh has been the birthplace of actors like Alastair Sim and Sir Sean Connery, known for being the first cinematic James Bond, the comedian and actor Ronnie Corbett, best known as one of The Two Ronnies, and the impressionist Rory Bremner. Famous artists from the city include the portrait painters Sir Henry Raeburn, Sir David Wilkie and Allan Ramsay. The city has produced or been home to some very successful musicians in recent decades, particularly Ian Anderson, front man of the band Jethro Tull, The Incredible String Band, the folk duo The Corries, Wattie Buchan, lead singer and founding member of punk band The Exploited, Shirley Manson, lead singer of the band Garbage, the Bay City Rollers, The Proclaimers, Boards of Canada and Idlewild. Edinburgh is the birthplace of former British Prime Minister Tony Blair who attended the city's Fettes College. Notorious criminals from Edinburgh's past include Deacon Brodie, head of a trades guild and Edinburgh city councillor by day but a burglar by night, who is said to have been the inspiration for Robert Louis Stevenson's story, the "Strange Case of Dr Jekyll and Mr Hyde", and murderers Burke and Hare who delivered fresh corpses for dissection to the famous anatomist Robert Knox. Another well-known Edinburgh resident was Greyfriars Bobby. The small Skye Terrier reputedly kept vigil over his dead master's grave in Greyfriars Kirkyard for 14 years in the 1860s and 1870s, giving rise to a story of canine devotion which plays a part in attracting visitors to the city. The City of Edinburgh has entered into 14 international twinning arrangements since 1954. Most of the arrangements are styled as 'Twin Cities' but the agreement with Kraków is designated as a 'Partner City', and the agreement with Kyoto Prefecture is officially styled as a 'Friendship Link', reflecting its status as the only region to be twinned with Edinburgh. For a list of consulates in Edinburgh, see List of diplomatic missions in Scotland.
https://en.wikipedia.org/wiki?curid=9602
Ernest Rutherford Ernest Rutherford, 1st Baron Rutherford of Nelson, (30 August 1871 – 19 October 1937) was a New Zealand-born British physicist who came to be known as the father of nuclear physics. "Encyclopædia Britannica" considers him to be the greatest experimentalist since Michael Faraday (1791–1867). In early work, Rutherford discovered the concept of radioactive half-life, the radioactive element radon, and differentiated and named alpha and beta radiation. This work was performed at McGill University in Montreal, Quebec, Canada. It is the basis for the Nobel Prize in Chemistry he was awarded in 1908 "for his investigations into the disintegration of the elements, and the chemistry of radioactive substances", for which he was the first Canadian and Oceanian Nobel laureate. Rutherford moved in 1907 to the Victoria University of Manchester (today University of Manchester) in the UK, where he and Thomas Royds proved that alpha radiation is helium nuclei. Rutherford performed his most famous work after he became a Nobel laureate. In 1911, although he could not prove that it was positive or negative, he theorized that atoms have their charge concentrated in a very small nucleus, and thereby pioneered the Rutherford model of the atom, through his discovery and interpretation of Rutherford scattering by the gold foil experiment of Hans Geiger and Ernest Marsden. He performed the first artificially induced nuclear reaction in 1917 in experiments where nitrogen nuclei were bombarded with alpha particles. As a result, he discovered the emission of a subatomic particle which, in 1919, he called the "hydrogen atom" but, in 1920, he more accurately named the proton. Rutherford became Director of the Cavendish Laboratory at the University of Cambridge in 1919. Under his leadership the neutron was discovered by James Chadwick in 1932 and in the same year the first experiment to split the nucleus in a fully controlled manner was performed by students working under his direction, John Cockcroft and Ernest Walton. After his death in 1937, he was buried in Westminster Abbey near Sir Isaac Newton's tomb. The chemical element rutherfordium (element 104) was named after him in 1997. Ernest Rutherford was the son of James Rutherford, a farmer, and his wife Martha Thompson, originally from Hornchurch, Essex, England. James had emigrated to New Zealand from Perth, Scotland, "to raise a little flax and a lot of children". Ernest was born at Brightwater, near Nelson, New Zealand. His first name was mistakenly spelled 'Earnest' when his birth was registered. Rutherford's mother Martha Thompson was a schoolteacher. He studied at Havelock School and then Nelson College and won a scholarship to study at Canterbury College, University of New Zealand, where he participated in the debating society and played rugby. After gaining his BA, MA and BSc, and doing two years of research during which he invented a new form of radio receiver, in 1895 Rutherford was awarded an 1851 Research Fellowship from the Royal Commission for the Exhibition of 1851, to travel to England for postgraduate study at the Cavendish Laboratory, University of Cambridge. He was among the first of the 'aliens' (those without a Cambridge degree) allowed to do research at the university, under the leadership of J. J. Thomson, which aroused jealousies from the more conservative members of the Cavendish fraternity. With Thomson's encouragement, he managed to detect radio waves at half a mile and briefly held the world record for the distance over which electromagnetic waves could be detected, though when he presented his results at the British Association meeting in 1896, he discovered he had been outdone by Guglielmo Marconi, who was also lecturing. In 1898, Thomson recommended Rutherford for a position at McGill University in Montreal, Canada. He was to replace Hugh Longbourne Callendar who held the chair of Macdonald Professor of physics and was coming to Cambridge. Rutherford was accepted, which meant that in 1900 he could marry Mary Georgina Newton (1876–1954) to whom he had become engaged before leaving New Zealand; they married at St Paul's Anglican Church, Papanui in Christchurch, they had one daughter, Eileen Mary (1901–1930), who married the physicist Ralph Fowler. In 1901, Rutherford gained a DSc from the University of New Zealand. In 1907, he returned to Britain to take the chair of physics at the Victoria University of Manchester. Rutherford was knighted in 1914. During World War I, he worked on a top secret project to solve the practical problems of submarine detection by sonar. In 1916, he was awarded the Hector Memorial Medal. In 1919, he returned to the Cavendish succeeding J. J. Thomson as the Cavendish professor and Director. Under him, Nobel Prizes were awarded to James Chadwick for discovering the neutron (in 1932), John Cockcroft and Ernest Walton for an experiment which was to be known as "splitting the atom" using a particle accelerator, and Edward Appleton for demonstrating the existence of the ionosphere. In 1925, Rutherford pushed calls to the Government of New Zealand to support education and research, which led to the formation of the Department of Scientific and Industrial Research (DSIR) in the following year. Between 1925 and 1930, he served as President of the Royal Society, and later as president of the Academic Assistance Council which helped almost 1,000 university refugees from Germany. He was appointed to the Order of Merit in the 1925 New Year Honours and raised to the peerage as Baron Rutherford of Nelson, of Cambridge in the County of Cambridge in 1931, a title that became extinct upon his unexpected death in 1937. In 1933, Rutherford was one of the two inaugural recipients of the T. K. Sidey Medal, set up by the Royal Society of New Zealand as an award for outstanding scientific research. For some time before his death, Rutherford had a small hernia, which he had neglected to have fixed, and it became strangulated, causing him to be violently ill. Despite an emergency operation in London, he died four days afterwards of what physicians termed "intestinal paralysis", at Cambridge. After cremation at Golders Green Crematorium, he was given the high honour of burial in Westminster Abbey, near Isaac Newton and other illustrious British scientists. At Cambridge, Rutherford started to work with J. J. Thomson on the conductive effects of X-rays on gases, work which led to the discovery of the electron which Thomson presented to the world in 1897. Hearing of Becquerel's experience with uranium, Rutherford started to explore its radioactivity, discovering two types that differed from X-rays in their penetrating power. Continuing his research in Canada, he coined the terms alpha ray and beta ray in 1899 to describe the two distinct types of radiation. He then discovered that thorium gave off a gas which produced an emanation which was itself radioactive and would coat other substances. He found that a sample of this radioactive material of any size invariably took the same amount of time for half the sample to decay – its "half-life" (11½ minutes in this case). From 1900 to 1903, he was joined at McGill by the young chemist Frederick Soddy (Nobel Prize in Chemistry, 1921) for whom he set the problem of identifying the thorium emanations. Once he had eliminated all the normal chemical reactions, Soddy suggested that it must be one of the inert gases, which they named thoron (later found to be an isotope of radon). They also found another type of thorium they called Thorium X, and kept on finding traces of helium. They also worked with samples of "Uranium X" from William Crookes and radium from Marie Curie. In 1903, they published their "Law of Radioactive Change," to account for all their experiments. Until then, atoms were assumed to be the indestructible basis of all matter and although Curie had suggested that radioactivity was an atomic phenomenon, the idea of the atoms of radioactive substances breaking up was a radically new idea. Rutherford and Soddy demonstrated that radioactivity involved the spontaneous disintegration of atoms into other, as yet, unidentified matter. The Nobel Prize in Chemistry 1908 was awarded to Ernest Rutherford "for his investigations into the disintegration of the elements, and the chemistry of radioactive substances". In 1903, Rutherford considered a type of radiation discovered (but not named) by French chemist Paul Villard in 1900, as an emission from radium, and realised that this observation must represent something different from his own alpha and beta rays, due to its very much greater penetrating power. Rutherford therefore gave this third type of radiation the name of gamma ray. All three of Rutherford's terms are in standard use today – other types of radioactive decay have since been discovered, but Rutherford's three types are among the most common. In Manchester, he continued to work with alpha radiation. In conjunction with Hans Geiger, he developed zinc sulfide scintillation screens and ionisation chambers to count alphas. By dividing the total charge they produced by the number counted, Rutherford decided that the charge on the alpha was two. In late 1907, Ernest Rutherford and Thomas Royds allowed alphas to penetrate a very thin window into an evacuated tube. As they sparked the tube into discharge, the spectrum obtained from it changed, as the alphas accumulated in the tube. Eventually, the clear spectrum of helium gas appeared, proving that alphas were at least ionised helium atoms, and probably helium nuclei. A long-standing myth existed, at least as early as 1948, running at least to 2017, that Rutherford was the first scientist to observe and report an artificial transmutation of a stable element into another element: nitrogen into oxygen. It was thought by many people to be one of Rutherford's greatest accomplishments. The New Zealand government even issued a commemorative stamp in the belief that the nitrogen-to-oxygen discovery belonged to Rutherford. Beginning in 2017, many scientific institutions corrected their versions of this history to indicate that the discovery credit for the reaction belongs to Patrick Blackett. Rutherford did detect the ejected proton in 1919 and interpreted it as evidence for disintegration of the nitrogen nucleus (to lighter nuclei). In 1925, Blackett showed that the actual product is oxygen and identified the true reaction as 14N + α → 17O + p. Rutherford therefore recognized "that the nucleus may increase rather than diminish in mass as the result of collisions in which the proton is expelled." Rutherford performed his most famous work after receiving the Nobel prize in 1908. Along with Hans Geiger and Ernest Marsden in 1909, he carried out the Geiger–Marsden experiment, which demonstrated the nuclear nature of atoms by deflecting alpha particles passing through a thin gold foil. Rutherford was inspired to ask Geiger and Marsden in this experiment to look for alpha particles with very high deflection angles, of a type not expected from any theory of matter at that time. Such deflections, though rare, were found, and proved to be a smooth but high-order function of the deflection angle. It was Rutherford's interpretation of this data that led him to formulate the Rutherford model of the atom in 1911that a very small charged nucleus, containing much of the atom's mass, was orbited by low-mass electrons. In 1919–1920, Rutherford found that nitrogen and other light elements ejected a proton, which he called a "hydrogen atom", when hit with α (alpha) particles. This result showed Rutherford that hydrogen nuclei were a part of nitrogen nuclei (and by inference, probably other nuclei as well). Such a construction had been suspected for many years on the basis of atomic weights which were whole numbers of that of hydrogen; see Prout's hypothesis. Hydrogen was known to be the lightest element, and its nuclei presumably the lightest nuclei. Now, because of all these considerations, Rutherford decided that a hydrogen nucleus was possibly a fundamental building block of all nuclei, and also possibly a new fundamental particle as well, since nothing was known from the nucleus that was lighter. Thus, confirming and extending the work of Wilhelm Wien who in 1898 discovered the proton in streams of ionized gas, Rutherford postulated the hydrogen nucleus to be a new particle in 1920, which he dubbed the "proton". In 1921, while working with Niels Bohr (who postulated that electrons moved in specific orbits), Rutherford theorized about the existence of neutrons, (which he had christened in his 1920 Bakerian Lecture), which could somehow compensate for the repelling effect of the positive charges of protons by causing an attractive nuclear force and thus keep the nuclei from flying apart from the repulsion between protons. The only alternative to neutrons was the existence of "nuclear electrons" which would counteract some of the proton charges in the nucleus, since by then it was known that nuclei had about twice the mass that could be accounted for if they were simply assembled from hydrogen nuclei (protons). But how these nuclear electrons could be trapped in the nucleus, was a mystery. Rutherford's theory of neutrons was proved in 1932 by his associate James Chadwick, who recognized neutrons immediately when they were produced by other scientists and later himself, in bombarding beryllium with alpha particles. In 1935, Chadwick was awarded the Nobel Prize in Physics for this discovery. Rutherford is considered to have been among the greatest scientists in history. At the opening session of the 1938 Indian Science Congress, which Rutherford had been expected to preside over before his death, astrophysicist James Jeans spoke in his place and deemed him "one of the greatest scientists of all time," saying: Rutherford's research, and work done under him as laboratory director, established the nuclear structure of the atom and the essential nature of radioactive decay as a nuclear process. Patrick Blackett, a research fellow working under Rutherford, using natural alpha particles, demonstrated "induced" nuclear transmutation. Rutherford's team later, using protons from an accelerator, demonstrated "artificially-induced" nuclear reactions and transmutation. He is known as the father of nuclear physics. Rutherford died too early to see Leó Szilárd's idea of controlled nuclear chain reactions come into being. However, a speech of Rutherford's about his artificially-induced transmutation in lithium, printed in 12 September 1933 London paper "The Times", was reported by Szilárd to have been his inspiration for thinking of the possibility of a controlled energy-producing nuclear chain reaction. Szilard had this idea while walking in London, on the same day. Rutherford's speech touched on the 1932 work of his students John Cockcroft and Ernest Walton in "splitting" lithium into alpha particles by bombardment with protons from a particle accelerator they had constructed. Rutherford realized that the energy released from the split lithium atoms was enormous, but he also realized that the energy needed for the accelerator, and its essential inefficiency in splitting atoms in this fashion, made the project an impossibility as a practical source of energy (accelerator-induced fission of light elements remains too inefficient to be used in this way, even today). Rutherford's speech in part, read:
https://en.wikipedia.org/wiki?curid=9603
Many-worlds interpretation The many-worlds interpretation (MWI) is an interpretation of quantum mechanics that asserts that the universal wavefunction is objectively real, and that there is no wavefunction collapse. This implies that all possible outcomes of quantum measurements are physically realized in some "world" or universe. In contrast to some other interpretations, such as the Copenhagen interpretation, the evolution of reality as a whole in MWI is rigidly deterministic. Many-worlds is also called the relative state formulation or the Everett interpretation, after physicist Hugh Everett, who first proposed it in 1957. Bryce DeWitt popularized the formulation and named it "many-worlds" in the 1960s and 1970s. In many-worlds, the subjective appearance of wavefunction collapse is explained by the mechanism of quantum decoherence. Decoherence approaches to interpreting quantum theory have been widely explored and developed since the 1970s, and have become quite popular. MWI is now considered a mainstream interpretation along with the other decoherence interpretations, collapse theories (including the Copenhagen interpretation), and hidden variable theories such as Bohmian mechanics. The many-worlds interpretation implies that there are very—perhaps infinitely—many universes. It is one of many multiverse hypotheses in physics and philosophy. MWI views time as a many-branched tree, wherein every possible quantum outcome is realised. This is intended to resolve some paradoxes of quantum theory, such as the EPR paradox and Schrödinger's cat, since every possible outcome of a quantum event exists in its own universe. In 1952 Erwin Schrödinger gave a lecture in Dublin in which at one point he jocularly warned his audience that what he was about to say might "seem lunatic". He went on to assert that while what the equation that won him a Nobel prize seems to be describing is several different histories, they are "not alternatives but all really happen simultaneously". This is the earliest known reference to many-worlds. MWI originated in Everett's Princeton Ph.D. thesis "The Theory of the Universal Wavefunction", developed under his thesis advisor John Archibald Wheeler, a shorter summary of which was published in 1957 under the title "Relative State Formulation of Quantum Mechanics" (Wheeler contributed the title "relative state"; Everett originally called his approach the "Correlation Interpretation", where "correlation" refers to quantum entanglement). The phrase "many-worlds" is due to Bryce DeWitt, who was responsible for the wider popularisation of Everett's theory, which was largely ignored for a decade after publication. The key idea of the many-worlds interpretation is that unitary quantum mechanics describes the whole universe. In particular, it describes a measurement as a unitary transformation, without using a collapse postulate, and describes observers as ordinary quantum-mechanical systems. This stands in sharp contrast to the Copenhagen interpretation, on which a measurement is a "primitive" concept, not describable by quantum mechanics, the universe is divided into a quantum and a classical domain, and the collapse postulate is central. MWI's main conclusion is that the universe (or multiverse in this context) is composed of a quantum superposition of an infinite or undefinable amount or number of increasingly divergent, non-communicating parallel universes or quantum worlds. The many-worlds interpretation makes essential use of decoherence to explain the measurement process and the emergence of a quasi-classical world. Wojciech H. Zurek, one of decoherence theory's pioneers, stated: "Under scrutiny of the environment, only pointer states remain unchanged. Other states decohere into mixtures of stable pointer states that can persist, and, in this sense, exist: They are einselected." Żurek emphasizes that his work does not depend on a particular interpretation. The many-worlds interpretation shares many similarities with the decoherent histories interpretation, which also uses decoherence to explain the process of measurement or wavefunction collapse. MWI treats the other histories or worlds as real since it regards the universal wavefunction as the "basic physical entity" or "the fundamental entity, obeying at all times a deterministic wave equation". Decoherent histories, on the other hand, needs only one of the histories (or worlds) to be real. Several authors, including Wheeler, Everett and Deutsch, call many-worlds a theory, rather than just an interpretation. Everett argued that it was the "only completely coherent approach to explaining both the contents of quantum mechanics and the appearance of the world." Deutsch dismissed the idea that many-worlds is an "interpretation", saying that to call it that "is like talking about dinosaurs as an 'interpretation' of fossil records." In Everett's formulation, a measuring apparatus M and an object system S form a composite system, each of which prior to measurement exists in well-defined (but time-dependent) states. Measurement is regarded as causing M and S to interact. After S interacts with M, it is no longer possible to describe either system by an independent state. According to Everett, the only meaningful descriptions of each system are relative states: for example the relative state of S given the state of M or the relative state of M given the state of S. In DeWitt's formulation, the state of S after a sequence of measurements is given by a quantum superposition of states, each one corresponding to an alternative measurement history of S. For example, consider the smallest possible truly quantum system S, as shown in the illustration. This describes for instance, the spin-state of an electron. Considering a specific axis (say the "z"-axis) the north pole represents spin "up" and the south pole, spin "down". The superposition states of the system are described by a sphere called the Bloch sphere. To perform a measurement on S, it is made to interact with another similar system M. After the interaction, the combined system can be regarded as a quantum superposition of two "alternative histories" of the original system S, one in which "up" was observed and the other in which "down" was observed. Each subsequent binary measurement (that is interaction with a system M) causes a similar split in the history tree. Thus after three measurements, the system can be regarded as a quantum superposition of 8 = 2 × 2 × 2 copies of the original system S. In his 1957 doctoral dissertation, Everett proposed that rather than modeling an isolated quantum system subject to external observation, one could mathematically model an object as well as its observers as purely physical systems within the mathematical framework developed by Paul Dirac, John von Neumann and others, discarding altogether the "ad hoc" mechanism of wave function collapse. Since Everett's original work, a number of similar formalisms have appeared in the literature. One is the relative state formulation. It makes two assumptions: first, the wavefunction is not simply a description of the object's state, but is entirely equivalent to the object—a claim it has in common with some other interpretations. Second, observation or measurement has no special laws or mechanics, unlike in the Copenhagen interpretation, which considers the wavefunction collapse a special kind of event that occurs as a result of observation. Instead, measurement in the relative state formulation is the consequence of a configuration change in an observer's memory described by the same basic wave physics as the object being modeled. The many-worlds interpretation is DeWitt's popularisation of Everett, who had referred to the combined observer–object system as split by an observation, each split corresponding to the different or multiple possible outcomes of an observation. These splits generate a tree, as shown in the graphic above. Subsequently, DeWitt introduced the term "world" to describe a complete measurement history of an observer, which corresponds roughly to a single branch of that tree. Under the many-worlds interpretation, the Schrödinger equation, or relativistic analog, holds all the time everywhere. An observation or measurement is modeled by applying the wave equation to the entire system comprising the observer "and" the object. One consequence is that every observation can be thought of as causing the combined observer–object's wavefunction to change into a quantum superposition of two or more non-interacting branches, or split into many "worlds". Since many observation-like events have happened and are constantly happening, there are an enormous and growing number of simultaneously existing states. If a system is composed of two or more subsystems, the system's state will be a superposition of products of the subsystems' states. Each product of subsystem states in the overall superposition evolves over time independently of other products. Once the subsystems interact, their states have become correlated or entangled and can no longer be considered independent. In Everett's terminology each subsystem state was now "correlated" with its "relative state", since each subsystem must now be considered relative to the other subsystems with which it has interacted. MWI removes the observer-dependent role in the quantum measurement process by replacing wavefunction collapse with quantum decoherence. Since the observer's role lies at the heart of most if not all "quantum paradoxes," this automatically resolves a number of problems, such as Schrödinger's cat thought experiment, the EPR paradox, von Neumann's "boundary problem", and even wave-particle duality. Since the Copenhagen interpretation requires the existence of a classical domain beyond the one described by quantum mechanics, it has been criticized as inadequate for the study of cosmology. MWI was developed with the explicit goal of allowing quantum mechanics to be applied to the universe as a whole, making quantum cosmology possible. MWI is a realist, deterministic, and local theory. It achieves this by removing wavefunction collapse, which is indeterministic and non-local, from the deterministic and local equations of quantum theory. MWI (like other, broader multiverse theories) provides a context for the anthropic principle, which may provide an explanation for the fine-tuned universe. MWI depends crucially on the linearity of quantum mechanics. If the final theory of everything is non-linear with respect to wavefunctions, then many-worlds is invalid.. While quantum gravity or string theory may be non-linear in this respect, there is no evidence of this as yet. As with the other interpretations of quantum mechanics, the many-worlds interpretation is motivated by behavior that can be illustrated by the double-slit experiment. When particles of light (or anything else) pass through the double slit, a calculation assuming wavelike behavior of light can be used to identify where the particles are likely to be observed. Yet when the particles are observed in this experiment, they appear as particles (i.e., at definite places) and not as non-localized waves. Some versions of the Copenhagen interpretation of quantum mechanics proposed a process of "collapse" in which an indeterminate quantum system would probabilistically collapse down onto, or select, just one determinate outcome to "explain" this phenomenon of observation. Wavefunction collapse was widely regarded as artificial and "ad hoc", so an alternative interpretation in which the behavior of measurement could be understood from more fundamental physical principles was considered desirable. Everett's Ph.D. work provided such an interpretation. He argued that for a composite system—such as a subject (the "observer" or measuring apparatus) observing an object (the "observed" system, such as a particle)—the claim that either the observer or the observed has a well-defined state is meaningless; in modern parlance, the observer and the observed have become entangled: we can only specify the state of one "relative" to the other, i.e., the state of the observer and the observed are correlated "after" the observation is made. This led Everett to derive from the unitary, deterministic dynamics alone (i.e., without assuming wavefunction collapse) the notion of a "relativity of states". Everett noticed that the unitary, deterministic dynamics alone entailed that after an observation is made each element of the quantum superposition of the combined subject–object wavefunction contains two "relative states": a "collapsed" object state and an associated observer who has observed the same collapsed outcome; what the observer sees and the state of the object have become correlated by the act of measurement or observation. The subsequent evolution of each pair of relative subject–object states proceeds with complete indifference as to the presence or absence of the other elements, "as if" wavefunction collapse has occurred, which has the consequence that later observations are always consistent with the earlier observations. Thus the "appearance" of the object's wavefunction's collapse has emerged from the unitary, deterministic theory itself. (This answered Einstein's early criticism of quantum theory, that the theory should define what is observed, not for the observables to define the theory.) Since the wavefunction merely appears to have collapsed then, Everett reasoned, there was no need to actually assume that it had collapsed. And so, invoking Occam's razor, he removed the postulate of wavefunction collapse from the theory. In 1985, David Deutsch proposed a variant of the Wigner's friend thought experiment as a test of many-worlds versus the Copenhagen interpretation. It consists of an experimenter (Wigner's friend) making a measurement on a quantum system in an isolated laboratory, and another experimenter (Wigner) who would make a measurement on the first one. According to the many-worlds theory, the first experimenter would end up in a macroscopic superposition of seeing one result of the measurement in one branch, and another result in another branch. The second experimenter could then interfere these two branches in order to test whether it is in fact in a macroscopic superposition or has collapsed into a single branch, as predicted by the Copenhagen interpretation. Since then Lockwood (1989), Vaidman and others have made similar proposals. These proposals require placing macroscopic objects in a coherent superposition and interfering them, a task now beyond experimental capability. Since the many-worlds interpretation's inception, physicists have been puzzled about the role of probability in it. As put by Wallace, there are two facets to the question: the "incoherence problem", which asks why should we assign probabilities at all to outcomes that are certain to occur in some worlds, and the "quantitative problem", which asks why the probabilities should be given by the Born rule. Everett tried to answer these questions in the paper that introduced many-worlds. To address the incoherence problem, he argued that an observer who makes a sequence of measurements on a quantum system will in general have an apparently random sequence of results in their memory, which justifies the use of probabilities to describe the measurement process. To address the quantitative problem, Everett proposed a derivation of the Born rule based on the properties that a measure on the branches of the wavefunction should have. His derivation has been criticized as relying on unmotivated assumptions. Since then several other derivations of the Born rule in the many-worlds framework have been proposed. There is no consensus on whether this has been successful. DeWitt and Graham and Farhi et al., among others, have proposed derivations of the Born rule based on a frequentist interpretation of probability. They try to show that in the limit of infinitely many measurements no worlds would have relative frequencies that didn't match the probabilities given by the Born rule, but these derivations have been shown to be mathematically incorrect. A decision-theoretic derivation of the Born rule was produced by David Deutsch (1999) and refined by Wallace (2002–2009) and Saunders (2004). They consider an agent who takes part in a quantum gamble: the agent makes a measurement on a quantum system, branches as a consequence, and each of the agent's future selves receives a reward that depends on the measurement result. The agent uses decision theory to evaluate the price they would pay to take part in such a gamble, and concludes that the price is given by the utility of the rewards weighted according to the Born rule. Some reviews have been positive, although these arguments remain highly controversial; some theoretical physicists have taken them as supporting the case for parallel universes. For example, a "New Scientist" story on a 2007 conference about Everettian interpretations quoted physicist Andy Albrecht as saying, "This work will go down as one of the most important developments in the history of science." In contrast, the philosopher Huw Price, also attending the conference, found the Deutsch–Wallace–Saunders approach fundamentally flawed. Żurek (2005) has produced a derivation of the Born rule based on the symmetries of entangled states; Schlosshauer and Fine argue that Żurek's derivation is not rigorous, as it does not define what probability is and has several unstated assumptions about how it should behave. Charles Sebens and Sean M. Carroll, building on work by Lev Vaidman, proposed a similar approach based on self-locating uncertainty. In this approach, decoherence creates multiple identical copies of observers, who can assign credences to being on different branches using the Born rule. The Sebens–Carroll approach has been criticized by Adrian Kent, and Vaidman himself does not find it satisfactory. As originally formulated by Everett and DeWitt, the many-worlds interpretation had a privileged role for measurements: they determined which basis of a quantum system would give rise to the eponymous worlds. Without this the theory was ambiguous, as a quantum state can equally well be described (e.g.) as having a well-defined position or as being a superposition of two delocalised states. The assumption that the preferred basis to use is the one from a measurement of position results in worlds having objects in well-defined positions, instead of worlds with delocalised objects (which would be grossly incompatible with experiment). This special role for measurements is problematic for the theory, as it contradicts Everett and DeWitt's goal of having a reductionist theory and undermines their criticism of the ill-defined measurement postulate of the Copenhagen interpretation. This is known today as the "preferred basis problem". The preferred basis problem has been solved, according to Saunders and Wallace, among others, by incorporating decoherence in the many-worlds theory. In this approach, the preferred basis does not have to be postulated, but rather is identified as the basis stable under environmental decoherence. In this way measurements no longer play a special role; rather, any interaction that causes decoherence causes the world to split. Since decoherence is never complete, there will always remain some infinitesimal overlap between two worlds, making it arbitrary whether a pair of worlds has split or not. Wallace argues that this is not problematic: it only shows that worlds are not a part of the fundamental ontology, but rather of the "emergent" ontology, where these approximate, effective descriptions are routine in the physical sciences. Since in this approach the worlds are derived, it follows that they must be present in any other interpretation of quantum mechanics that does not have a collapse mechanism, such as Bohmian mechanics. This approach to deriving the preferred basis has been criticized as creating a circularity with derivations of probability in the many-worlds interpretation, as decoherence theory depends on probability, and probability depends on the ontology derived from decoherence. Wallace contends that decoherence theory depends not on probability but only on the notion that one is allowed to do approximations in physics. MWI's initial reception was overwhelmingly negative, with the notable exception of DeWitt. Wheeler made considerable efforts to formulate the theory in a way that would be palatable to Bohr, visited Copenhagen in 1956 to discuss it with him, and convinced Everett to visit as well, which happened in 1959. Nevertheless, Bohr and his collaborators completely rejected the theory. Everett left academia in 1956, never to return, and Wheeler eventually disavowed the theory. One of MWI's strongest advocates is David Deutsch. According to Deutsch, the single photon interference pattern observed in the double slit experiment can be explained by interference of photons in multiple universes. Viewed this way, the single photon interference experiment is indistinguishable from the multiple photon interference experiment. In a more practical vein, in one of the earliest papers on quantum computing, he suggested that parallelism that results from MWI could lead to ""a method by which certain probabilistic tasks can be performed faster by a universal quantum computer than by any classical restriction of it"". Deutsch has also proposed that MWI will be testable (at least against "naive" Copenhagenism) when reversible computers become conscious via the reversible observation of spin. Asher Peres was an outspoken critic of MWI. A section of his 1993 textbook had the title "Everett's interpretation and other bizarre theories". Peres argued that the various many-worlds interpretations merely shift the arbitrariness or vagueness of the collapse postulate to the question of when "worlds" can be regarded as separate, and that no objective criterion for that separation can actually be formulated. Some consider MWI unfalsifiable and hence unscientific because the multiple parallel universes are non-communicating, in the sense that no information can be passed between them. Others claim MWI is directly testable. Victor J. Stenger remarks that Murray Gell-Mann's published work explicitly rejects the existence of simultaneous parallel universes. Collaborating with James Hartle, Gell-Mann had been, before his death, working toward the development a more "palatable" "post-Everett quantum mechanics". Stenger thinks it fair to say that most physicists dismiss the many-worlds interpretation as too extreme, while noting it "has merit in finding a place for the observer inside the system being analyzed and doing away with the troublesome notion of wave function collapse". Philosophers of science James Ladyman and Don Ross state that the MWI could be true, but that they do not embrace it. They note that no quantum theory is yet empirically adequate for describing all of reality, given its lack of unification with general relativity, and so they do not see a reason to regard any interpretation of quantum mechanics as the final word in metaphysics. They also suggest that the multiple branches may be an artifact of incomplete descriptions and of using quantum mechanics to represent the states of macroscopic objects. They argue that macroscopic objects are significantly different from microscopic objects in not being isolated from the environment, and that using quantum formalism to describe them lacks explanatory and descriptive power and accuracy. A poll of 72 "leading quantum cosmologists and other quantum field theorists" conducted before 1991 by L. David Raub showed 58% agreement with "Yes, I think MWI is true". Max Tegmark reports the result of a "highly unscientific" poll taken at a 1997 quantum mechanics workshop. According to Tegmark, "The many worlds interpretation (MWI) scored second, comfortably ahead of the consistent histories and Bohm interpretations." In response to Sean M. Carroll's statement "As crazy as it sounds, most working physicists buy into the many-worlds theory", Michael Nielsen counters: "at a quantum computing conference at Cambridge in 1998, a many-worlder surveyed the audience of approximately 200 people... Many-worlds did just fine, garnering support on a level comparable to, but somewhat below, Copenhagen and decoherence." But Nielsen notes that it seemed most attendees found it to be a waste of time: Peres "got a huge and sustained round of applause…when he got up at the end of the polling and asked 'And who here believes the laws of physics are decided by a democratic vote?'" A 2005 poll of fewer than 40 students and researchers taken after a course on the Interpretation of Quantum Mechanics at the Institute for Quantum Computing University of Waterloo found "Many Worlds (and decoherence)" to be the least favored. A 2011 poll of 33 participants at an Austrian conference found 6 endorsed MWI, 8 "Information-based/information-theoretical", and 14 Copenhagen; the authors remark that MWI received a similar percentage of votes as in Tegmark's 1997 poll. Everett believed in the literal reality of the other quantum worlds. His son reported that he "never wavered in his belief over his many-worlds theory". According to Martin Gardner, the "other" worlds of MWI have two different interpretations: real or unreal; he claimed that Stephen Hawking and Steven Weinberg both favour the unreal interpretation. Gardner also claimed that most physicists favour the nonreal interpretation, whereas the "realist" view is supported only by MWI experts such as Deutsch and DeWitt. Hawking has said that "according to Feynman's idea", all other histories are as "equally real" as our own, and Gardner reports Hawking saying that MWI is "trivially true". In a 1983 interview, Hawking also said he regarded MWI as "self-evidently correct" but was dismissive of questions about the interpretation of quantum mechanics, saying, "When I hear of Schrödinger's cat, I reach for my gun." In the same interview, he also said, "But, look: All that one does, really, is to calculate conditional probabilities—in other words, the probability of A happening, given B. I think that that's all the many worlds interpretation is. Some people overlay it with a lot of mysticism about the wave function splitting into different parts. But all that you're calculating is conditional probabilities." Elsewhere Hawking contrasted his attitude towards the "reality" of physical theories with that of his colleague Roger Penrose, saying, "He's a Platonist and I'm a positivist. He's worried that Schrödinger's cat is in a quantum state, where it is half alive and half dead. He feels that can't correspond to reality. But that doesn't bother me. I don't demand that a theory correspond to reality because I don't know what it is. Reality is not a quality you can test with litmus paper. All I'm concerned with is that the theory should predict the results of measurements. Quantum theory does this very successfully." For his own part, Penrose agrees with Hawking that QM applied to the universe implies MW, but he believes the lack of a successful theory of quantum gravity negates the claimed universality of conventional QM. "Quantum suicide" is a thought experiment in quantum mechanics and the philosophy of physics. Purportedly, it can distinguish between the Copenhagen interpretation of quantum mechanics and the many-worlds interpretation by means of a variation of the Schrödinger's cat thought experiment, from the cat's point of view. "Quantum immortality" refers to the subjective experience of surviving quantum suicide. Most experts believe that the experiment would not work in the real world, because the world with the surviving experimenter has a lower "measure" than the world prior to the experiment, making it less likely that the experimenter will go on to experience their survival. DeWitt has stated that "[Everett, Wheeler and Graham] do not in the end exclude any element of the superposition. All the worlds are there, even those in which everything goes wrong and all the statistical laws break down." Max Tegmark has affirmed that absurd or highly unlikely events are inevitable but rare under the MWI. To quote Tegmark, "Things inconsistent with the laws of physics will never happen—everything else will... it's important to keep track of the statistics, since even if everything conceivable happens somewhere, really freak events happen only exponentially rarely." Ladyman and Ross state that, in general, many of the unrealized possibilities that are discussed in other scientific fields will not have counterparts in other branches, because they are in fact incompatible with the universal wavefunction.
https://en.wikipedia.org/wiki?curid=9604
E-commerce E-commerce (electronic commerce) is the activity of electronically buying or selling of products on online services or over the Internet. Electronic commerce draws on technologies such as mobile commerce, electronic funds transfer, supply chain management, Internet marketing, online transaction processing, electronic data interchange (EDI), inventory management systems, and automated data collection systems. E-commerce is in turn driven by the technological advances of the semiconductor industry, and is the largest sector of the electronics industry. Modern electronic commerce typically uses the World Wide Web for at least one part of the transaction's life cycle although it may also use other technologies such as e-mail. Typical e-commerce transactions include the purchase of online books (such as Amazon) and music purchases (music download in the form of digital distribution such as iTunes Store), and to a less extent, customized/personalized online liquor store inventory services. There are three areas of e-commerce: online retailing, electronic markets, and online auctions. E-commerce is supported by electronic business. E-commerce businesses may also employ some or all of the followings: A timeline for the development of e-commerce: Some common applications related to electronic commerce are: In the United States, certain electronic commerce activities are regulated by the Federal Trade Commission (FTC). These activities include the use of commercial e-mails, online advertising and consumer privacy. The CAN-SPAM Act of 2003 establishes national standards for direct marketing over e-mail. The Federal Trade Commission Act regulates all forms of advertising, including online advertising, and states that advertising must be truthful and non-deceptive. Using its authority under Section 5 of the FTC Act, which prohibits unfair or deceptive practices, the FTC has brought a number of cases to enforce the promises in corporate privacy statements, including promises about the security of consumers' personal information. As a result, any corporate privacy policy related to e-commerce activity may be subject to enforcement by the FTC. The Ryan Haight Online Pharmacy Consumer Protection Act of 2008, which came into law in 2008, amends the Controlled Substances Act to address online pharmacies. Conflict of laws in cyberspace is a major hurdle for harmonization of legal framework for e-commerce around the world. In order to give a uniformity to e-commerce law around the world, many countries adopted the UNCITRAL Model Law on Electronic Commerce (1996). Internationally there is the International Consumer Protection and Enforcement Network (ICPEN), which was formed in 1991 from an informal network of government customer fair trade organisations. The purpose was stated as being to find ways of co-operating on tackling consumer problems connected with cross-border transactions in both goods and services, and to help ensure exchanges of information among the participants for mutual benefit and understanding. From this came Econsumer.gov, an ICPEN initiative since April 2001. It is a portal to report complaints about online and related transactions with foreign companies. There is also Asia Pacific Economic Cooperation (APEC) was established in 1989 with the vision of achieving stability, security and prosperity for the region through free and open trade and investment. APEC has an Electronic Commerce Steering Group as well as working on common privacy regulations throughout the APEC region. In Australia, Trade is covered under Australian Treasury Guidelines for electronic commerce and the Australian Competition and Consumer Commission regulates and offers advice on how to deal with businesses online, and offers specific advice on what happens if things go wrong. In the United Kingdom, The Financial Services Authority (FSA) was formerly the regulating authority for most aspects of the EU's Payment Services Directive (PSD), until its replacement in 2013 by the Prudential Regulation Authority and the Financial Conduct Authority. The UK implemented the PSD through the Payment Services Regulations 2009 (PSRs), which came into effect on 1 November 2009. The PSR affects firms providing payment services and their customers. These firms include banks, non-bank credit card issuers and non-bank merchant acquirers, e-money issuers, etc. The PSRs created a new class of regulated firms known as payment institutions (PIs), who are subject to prudential requirements. Article 87 of the PSD requires the European Commission to report on the implementation and impact of the PSD by 1 November 2012. In India, the Information Technology Act 2000 governs the basic applicability of e-commerce. In China, the Telecommunications Regulations of the People's Republic of China (promulgated on 25 September 2000), stipulated the Ministry of Industry and Information Technology (MIIT) as the government department regulating all telecommunications related activities, including electronic commerce. On the same day, The Administrative Measures on Internet Information Services released, is the first administrative regulation to address profit-generating activities conducted through the Internet, and lay the foundation for future regulations governing e-commerce in China. On 28 August 2004, the eleventh session of the tenth NPC Standing Committee adopted The Electronic Signature Law, which regulates data message, electronic signature authentication and legal liability issues. It is considered the first law in China's e-commerce legislation. It was a milestone in the course of improving China's electronic commerce legislation, and also marks the entering of China's rapid development stage for electronic commerce legislation. Contemporary electronic commerce can be classified into two categories. The first category is business based on types of goods sold (involves everything from ordering "digital" content for immediate online consumption, to ordering conventional goods and services, to "meta" services to facilitate other types of electronic commerce). The second category is based on the nature of the participant (B2B, B2C, C2B and C2C); On the institutional level, big corporations and financial institutions use the internet to exchange financial data to facilitate domestic and international business. Data integrity and security are pressing issues for electronic commerce. Aside from traditional e-commerce, the terms m-Commerce (mobile commerce) as well (around 2013) t-Commerce have also been used. In 2010, the United Kingdom had the highest per capita e-commerce spending in the world. As of 2013, the Czech Republic was the European country where e-commerce delivers the biggest contribution to the enterprises´ total revenue. Almost a quarter (24%) of the country's total turnover is generated via the online channel. Among emerging economies, China's e-commerce presence continues to expand every year. With 668 million Internet users, China's online shopping sales reached $253 billion in the first half of 2015, accounting for 10% of total Chinese consumer retail sales in that period. The Chinese retailers have been able to help consumers feel more comfortable shopping online. e-commerce transactions between China and other countries increased 32% to 2.3 trillion yuan ($375.8 billion) in 2012 and accounted for 9.6% of China's total international trade. In 2013, Alibaba had an e-commerce market share of 80% in China. In 2014, there were 600 million Internet users in China (twice as many as in the US), making it the world's biggest online market. China is also the largest e-commerce market in the world by value of sales, with an estimated in 2016. Recent research clearly indicates that electronic commerce, commonly referred to as e-commerce, presently shapes the manner in which people shop for products. The GCC countries have a rapidly growing market and are characterized by a population that becomes wealthier (Yuldashev). As such, retailers have launched Arabic-language websites as a means to target this population. Secondly, there are predictions of increased mobile purchases and an expanding internet audience (Yuldashev). The growth and development of the two aspects make the GCC countries to become larger players in the electronic commerce market with time progress. Specifically, research shows that e-commerce market is expected to grow to over $20 billion by the year 2020 among these GCC countries (Yuldashev). The e-commerce market has also gained much popularity among the western countries, and in particular Europe and the U.S. These countries have been highly characterized with consumer-packaged-goods (CPG) (Geisler, 34). However, trends show that there are future signs of a reverse. Similar to the GCC countries, there has been increased purchase of goods and services in online channels rather than offline channels. Activist investors are trying hard to consolidate and slash their overall cost and the governments in western countries continue to impose more regulation on CPG manufacturers (Geisler, 36). In these senses, CPG investors are being forced to adapt e-commerce as it is effective as a well as a means for them to thrive. In 2013, Brazil's e-commerce was growing quickly with retail e-commerce sales expected to grow at a double-digit pace through 2014. By 2016, eMarketer expected retail e-commerce sales in Brazil to reach $17.3 billion. India has an Internet user base of about 460 million as of December 2017. Despite being third largest user base in world, the penetration of Internet is low compared to markets like the United States, United Kingdom or France but is growing at a much faster rate, adding around 6 million new entrants every month. In India, cash on delivery is the most preferred payment method, accumulating 75% of the e-retail activities. The India retail market is expected to rise from 2.5% in 2016 to 5% in 2020. The future trends in the GCC countries will be similar with that of the western countries. Despite the forces that push business to adapt e-commerce as a means to sell goods and products, the manner in which customers make purchases is similar in countries from these two regions. For instance, there has been an increased usage of smartphones which comes in conjunction with an increase in the overall internet audience from the regions. Yuldashev writes that consumers are scaling up to more modern technology that allows for mobile marketing. However, the percentage of smartphone and internet users who make online purchases is expected to vary in the first few years. It will be independent on the willingness of the people to adopt this new trend (The Statistics Portal). For example, UAE has the greatest smartphone penetration of 73.8 percent and has 91.9 percent of its population has access to the internet. On the other hand, smartphone penetration in Europe has been reported to be at 64.7 percent (The Statistics Portal). Regardless, the disparity in percentage between these regions is expected to level out in future because e-commerce technology is expected to grow allowing for more users. The e-commerce business within these two regions will result in a competition. Government bodies at country level will enhance their measures and strategies to ensure sustainability and consumer protection (Krings, et al.). These increased measures will raise the environmental and social standards in the countries, factors that will determine the success of e-commerce market in these countries. For example, an adoption of tough sanctions will make it difficult for companies to enter the e-commerce market while lenient sanctions will allow ease of companies. As such, the future trends between GCC countries and the Western countries will be independent of these sanctions (Krings, et al.). These countries need to make rational conclusions in coming up with effective sanctions. The rate of growth of the number of internet users in the Arab countries has been rapid – 13.1% in 2015. A significant portion of the e-commerce market in the Middle East comprises people in the 30–34 year age group. Egypt has the largest number of internet users in the region, followed by Saudi Arabia and Morocco; these constitute 3/4th of the region's share. Yet, internet penetration is low: 35% in Egypt and 65% in Saudi Arabia. E-commerce has become an important tool for small and large businesses worldwide, not only to sell to customers, but also to engage them. In 2012, e-commerce sales topped $1 trillion for the first time in history. Mobile devices are playing an increasing role in the mix of e-commerce, this is also commonly called mobile commerce, or m-commerce. In 2014, one estimate saw purchases made on mobile devices making up 25% of the market by 2017. For traditional businesses, one research stated that information technology and cross-border e-commerce is a good opportunity for the rapid development and growth of enterprises. Many companies have invested enormous volume of investment in mobile applications. The DeLone and McLean Model stated that three perspectives contribute to a successful e-business: information system quality, service quality and users' satisfaction. There is no limit of time and space, there are more opportunities to reach out to customers around the world, and to cut down unnecessary intermediate links, thereby reducing the cost price, and can benefit from one on one large customer data analysis, to achieve a high degree of personal customization strategic plan, in order to fully enhance the core competitiveness of the products in company. Modern 3D graphics technologies, such as Facebook 3D Posts, are considered by some social media marketers and advertisers as a preferable way to promote consumer goods than static photos, and some brands like Sony are already paving the way for augmented reality commerce. Wayfair now lets you inspect a 3D version of its furniture in a home setting before buying. Logistics in e-commerce mainly concerns fulfillment. Online markets and retailers have to find the best possible way to fill orders and deliver products. Small companies usually control their own logistic operation because they do not have the ability to hire an outside company. Most large companies hire a fulfillment service that takes care of a company's logistic needs. Contrary to common misconception, there are significant barriers to entry in e-commerce. E-commerce markets are growing at noticeable rates. The online market is expected to grow by 56% in 2015–2020. In 2017, retail e-commerce sales worldwide amounted to 2.3 trillion US dollars and e-retail revenues are projected to grow to 4.88 trillion US dollars in 2021. Traditional markets are only expected 2% growth during the same time. Brick and mortar retailers are struggling because of online retailer's ability to offer lower prices and higher efficiency. Many larger retailers are able to maintain a presence offline and online by linking physical and online offerings. E-commerce allows customers to overcome geographical barriers and allows them to purchase products anytime and from anywhere. Online and traditional markets have different strategies for conducting business. Traditional retailers offer fewer assortment of products because of shelf space where, online retailers often hold no inventory but send customer orders directly to the manufacture. The pricing strategies are also different for traditional and online retailers. Traditional retailers base their prices on store traffic and the cost to keep inventory. Online retailers base prices on the speed of delivery. There are two ways for marketers to conduct business through e-commerce: fully online or online along with a brick and mortar store. Online marketers can offer lower prices, greater product selection, and high efficiency rates. Many customers prefer online markets if the products can be delivered quickly at relatively low price. However, online retailers cannot offer the physical experience that traditional retailers can. It can be difficult to judge the quality of a product without the physical experience, which may cause customers to experience product or seller uncertainty. Another issue regarding the online market is concerns about the security of online transactions. Many customers remain loyal to well-known retailers because of this issue. Security is a primary problem for e-commerce in developed and developing countries. E-commerce security is protecting business' websites and customers from unauthorized access, use, alteration, or destruction. The type of threats include: malicious codes, unwanted programs (ad ware, spyware), phishing, hacking, and cyber vandalism. E-commerce websites use different tools to avert security threats. These tools include firewalls, encryption software, digital certificates, and passwords. For a long time, companies had been troubled by the gap between the benefits which supply chain technology has and the solutions to deliver those benefits. However, the emergence of e-commerce has provided a more practical and effective way of delivering the benefits of the new supply chain technologies. E-commerce has the capability to integrate all inter-company and intra-company functions, meaning that the three flows (physical flow, financial flow and information flow) of the supply chain could be also affected by e-commerce. The affections on physical flows improved the way of product and inventory movement level for companies. For the information flows, e-commerce optimised the capacity of information processing than companies used to have, and for the financial flows, e-commerce allows companies to have more efficient payment and settlement solutions. In addition, e-commerce has a more sophisticated level of impact on supply chains: Firstly, the performance gap will be eliminated since companies can identify gaps between different levels of supply chains by electronic means of solutions; Secondly, as a result of e-commerce emergence, new capabilities such implementing ERP systems, like SAP ERP, Xero, or Megaventory, have helped companies to manage operations with customers and suppliers. Yet these new capabilities are still not fully exploited. Thirdly, technology companies would keep investing on new e-commerce software solutions as they are expecting investment return. Fourthly, e-commerce would help to solve many aspects of issues that companies may feel difficult to cope with, such as political barriers or cross-country changes. Finally, e-commerce provides companies a more efficient and effective way to collaborate with each other within the supply chain. E-commerce helps create new job opportunities due to information related services, software app and digital products. It also causes job losses. The areas with the greatest predicted job-loss are retail, postal, and travel agencies. The development of e-commerce will create jobs that require highly skilled workers to manage large amounts of information, customer demands, and production processes. In contrast, people with poor technical skills cannot enjoy the wages welfare. On the other hand, because e-commerce requires sufficient stocks that could be delivered to customers in time, the warehouse becomes an important element. Warehouse needs more staff to manage, supervise and organize, thus the condition of warehouse environment will be concerned by employees. E-commerce brings convenience for customers as they do not have to leave home and only need to browse website online, especially for buying the products which are not sold in nearby shops. It could help customers buy wider range of products and save customers’ time. Consumers also gain power through online shopping. They are able to research products and compare prices among retailers. Also, online shopping often provides sales promotion or discounts code, thus it is more price effective for customers. Moreover, e-commerce provides products’ detailed information; even the in-store staff cannot offer such detailed explanation. Customers can also review and track the order history online. E-commerce technologies cut transaction costs by allowing both manufactures and consumers to skip through the intermediaries. This is achieved through by extending the search area best price deals and by group purchase. The success of e-commerce in urban and regional levels depend on how the local firms and consumers have adopted to e-commerce. However, e-commerce lacks human interaction for customers, especially who prefer face-to-face connection. Customers are also concerned with the security of online transactions and tend to remain loyal to well-known retailers. In recent years, clothing retailers such as Tommy Hilfiger have started adding Virtual Fit platforms to their e-commerce sites to reduce the risk of customers buying the wrong sized clothes, although these vary greatly in their fit for purpose. When the customer regret the purchase of a product, it involves returning goods and refunding process. This process is inconvenient as customers need to pack and post the goods. If the products are expensive, large or fragile, it refers to safety issues. In 2018, E-commerce generated 1.3 million tons of container cardboard in North America, an increase from 1.1 million in 2017. Only 35 percent of North American cardboard manufacturing capacity is from recycled content. The recycling rate in Europe is 80 percent and Asia is 93 percent. Amazon, the largest user of boxes, has a strategy to cut back on packing material and has reduced packaging material used by 19 percent by weight since 2016. Amazon is requiring retailers to manufacture their product packaging in a way that doesn't require additional shipping packaging. Amazon also has an 85-person team researching ways to reduce and improve their packaging and shipping materials. E-commerce has been cited as a major force for the failure of major U.S. retailers in a trend frequently referred to as a "retail apocalypse." The rise of e-commerce outlets like Amazon has made it harder for traditional retailers to attract customers to their stores and forced companies to change their sales strategies. Many companies have turned to sales promotions and increased digital efforts to lure shoppers while shutting down brick-and-mortar locations. The trend has forced some traditional retailers to shutter its brick and mortar operations. E-commerce has grown in importance as companies have adopted pure-click and brick-and-click channel systems. We can distinguish pure-click and brick-and-click channel system adopted by companies. E-commerce may take place on retailers' Web sites or mobile apps, or those of e-commerce marketplaces such as on Amazon, or Tmall from AliBaba. Those channels may also be supported by conversational commerce, e.g. live chat or chatbots on Web sites. Conversational commerce may also be standalone such as live chat or chatbots on messaging apps and via voice assistants. The contemporary e-commerce trend recommends companies to shift the traditional business model where focus on "standardized products, homogeneous market and long product life cycle" to the new business model where focus on "varied and customized products". E-commerce requires the company to have the ability to satisfy multiple needs of different customers and provide them with wider range of products. With more choices of products, the information of products for customers to select and meet their needs become crucial. In order to address the mass customization principle to the company, the use of recommender system is suggested. This system helps recommend the proper products to the customers and helps customers make the decision during the purchasing process. The recommender system could be operated through the top sellers on the website, the demographics of customers or the consumers' buying behavior. However, there are 3 main ways of recommendations: recommending products to customers directly, providing detailed products' information and showing other buyers' opinions or critiques. It is benefit for consumer experience without physical shopping. In general, recommender system is used to contact customers online and assist finding the right products they want effectively and directly. In March 2020, global retail website traffic hit 14.3 billion visits signifying an unprecedented growth of e-commerce during the lockdown of 2020. Studies show that in the US, as many as 29% of surveyed shoppers state that they will never go back to shopping in person again; in the UK, 43% of consumers state that they expect to keep on shopping the same way even after the lockdown is over.
https://en.wikipedia.org/wiki?curid=9611
Euler's formula Euler's formula, named after Leonhard Euler, is a mathematical formula in complex analysis that establishes the fundamental relationship between the trigonometric functions and the complex exponential function. Euler's formula states that for any real number : where is the base of the natural logarithm, is the imaginary unit, and and are the trigonometric functions cosine and sine respectively. This complex exponential function is sometimes denoted ("cosine plus i sine"). The formula is still valid if is a complex number, and so some authors refer to the more general complex version as Euler's formula. Euler's formula is ubiquitous in mathematics, physics, and engineering. The physicist Richard Feynman called the equation "our jewel" and "the most remarkable formula in mathematics". When formula_2, Euler's formula evaluates to formula_3, which is known as Euler's identity. Johann Bernoulli noted that And since the above equation tells us something about complex logarithms by relating natural logarithms to imaginary (complex) numbers. Bernoulli, however, did not evaluate the integral. Bernoulli's correspondence with Euler (who also knew the above equation) shows that Bernoulli did not fully understand complex logarithms. Euler also suggested that the complex logarithms can have infinitely many values. Meanwhile, Roger Cotes in 1714 presented a geometrical argument that can be interpreted (after correcting a misplaced factor of formula_6) as:
https://en.wikipedia.org/wiki?curid=9613
Édouard Manet Édouard Manet (, , ; 23 January 1832 – 30 April 1883) was a French modernist painter. He was one of the first 19th-century artists to paint modern life, and a pivotal figure in the transition from Realism to Impressionism. Born into an upper-class household with strong political connections, Manet rejected the future originally envisioned for him, and became engrossed in the world of painting. His early masterworks, "The Luncheon on the Grass (Le déjeuner sur l'herbe)" and "Olympia", both 1863, caused great controversy and served as rallying points for the young painters who would create Impressionism. Today, these are considered watershed paintings that mark the start of modern art. The last 20 years of Manet's life saw him form bonds with other great artists of the time, and develop his own style that would be heralded as innovative and serve as a major influence for future painters. Édouard Manet was born in Paris on 23 January 1832, in the ancestral hôtel particulier (mansion) on the rue des Petits Augustins (now rue Bonaparte) to an affluent and well-connected family. His mother, Eugénie-Desirée Fournier, was the daughter of a diplomat and goddaughter of the Swedish crown prince Charles Bernadotte, from whom the Swedish monarchs are descended. His father, Auguste Manet, was a French judge who expected Édouard to pursue a career in law. His uncle, Edmond Fournier, encouraged him to pursue painting and took young Manet to the Louvre. In 1841 he enrolled at secondary school, the Collège Rollin. In 1845, at the advice of his uncle, Manet enrolled in a special course of drawing where he met Antonin Proust, future Minister of Fine Arts and subsequent lifelong friend. At his father's suggestion, in 1848 he sailed on a training vessel to Rio de Janeiro. After he twice failed the examination to join the Navy, his father relented to his wishes to pursue an art education. From 1850 to 1856, Manet studied under the academic painter Thomas Couture. In his spare time, Manet copied the Old Masters in the Louvre. From 1853 to 1856, Manet visited Germany, Italy, and the Netherlands, during which time he was influenced by the Dutch painter Frans Hals, and the Spanish artists Diego Velázquez and Francisco José de Goya. In 1856, Manet opened a studio. His style in this period was characterized by loose brush strokes, simplification of details and the suppression of transitional tones. Adopting the current style of realism initiated by Gustave Courbet, he painted "The Absinthe Drinker" (1858–59) and other contemporary subjects such as beggars, singers, Gypsies, people in cafés, and bullfights. After his early career, he rarely painted religious, mythological, or historical subjects; examples include his "", now in the Art Institute of Chicago, and "Christ with Angels", in the Metropolitan Museum of Art, New York. Manet had two canvases accepted at the Salon in 1861. A portrait of his mother and father, who at the time was paralysed and robbed of speech by a stroke, was ill-received by critics. The other, "The Spanish Singer", was admired by Theophile Gautier, and placed in a more conspicuous location as a result of its popularity with Salon-goers. Manet's work, which appeared "slightly slapdash" when compared with the meticulous style of so many other Salon paintings, intrigued some young artists. "The Spanish Singer", painted in a "strange new fashion [-] caused many painters' eyes to open and their jaws to drop." "Music in the Tuileries" is an early example of Manet's painterly style. Inspired by Hals and Velázquez, it is a harbinger of his lifelong interest in the subject of leisure. While the picture was regarded as unfinished by some, the suggested atmosphere imparts a sense of what the Tuileries gardens were like at the time; one may imagine the music and conversation. Here, Manet has depicted his friends, artists, authors, and musicians who take part, and he has included a self-portrait among the subjects. A major early work is "The Luncheon on the Grass (Le Déjeuner sur l'herbe)", originally "Le Bain". The Paris Salon rejected it for exhibition in 1863, but Manet agreed to exhibit it at the Salon des Refusés (Salon of the Rejected) which was a parallel exhibition to the official Salon, as an alternative exhibition in the Palais des Champs-Elysée. The "Salon des Refusés" was initiated by Emperor Napoleon III as a solution to a problematic situation which came about as the Selection Committee of the Salon that year rejected 2,783 paintings of the ca. 5000. Each painter could decide whether to take the opportunity to exhibit at the "Salon des Refusés", less than 500 of the rejected painters chose to do so. Manet employed model Victorine Meurent, his wife Suzanne, future brother-in-law Ferdinand Leenhoff, and one of his brothers to pose. Meurent also posed for several more of Manet's important paintings including "Olympia"; and by the mid-1870s she became an accomplished painter in her own right. The painting's juxtaposition of fully dressed men and a nude woman was controversial, as was its abbreviated, sketch-like handling, an innovation that distinguished Manet from Courbet. At the same time, Manet's composition reveals his study of the old masters, as the disposition of the main figures is derived from Marcantonio Raimondi's engraving of the "Judgement of Paris" (c. 1515) based on a drawing by Raphael. Two additional works cited by scholars as important precedents for "Le déjeuner sur l'herbe" are "Pastoral Concert" (c. 1510, The Louvre) and "The Tempest" (Gallerie dell'Accademia, Venice), both of which are attributed variously to Italian Renaissance masters Giorgione or Titian. "The Tempest" is an enigmatic painting featuring a fully dressed man and a nude woman in a rural setting. The man is standing to the left and gazing to the side, apparently at the woman, who is seated and breastfeeding a baby; the relationship between the two figures is unclear. In "Pastoral Concert", two clothed men and a nude woman are seated on the grass, engaged in music making, while a second nude woman stands beside them. As he had in "Luncheon on the Grass", Manet again paraphrased a respected work by a Renaissance artist in the painting "Olympia" (1863), a nude portrayed in a style reminiscent of early studio photographs, but whose pose was based on Titian's "Venus of Urbino" (1538). The painting is also reminiscent of Francisco Goya's painting "The Nude Maja" (1800). Manet embarked on the canvas after being challenged to give the Salon a nude painting to display. His uniquely frank depiction of a self-assured prostitute was accepted by the Paris Salon in 1865, where it created a scandal. According to Antonin Proust, "only the precautions taken by the administration prevented the painting being punctured and torn" by offended viewers. The painting was controversial partly because the nude is wearing some small items of clothing such as an orchid in her hair, a bracelet, a ribbon around her neck, and mule slippers, all of which accentuated her nakedness, sexuality, and comfortable courtesan lifestyle. The orchid, upswept hair, black cat, and bouquet of flowers were all recognized symbols of sexuality at the time. This modern Venus' body is thin, counter to prevailing standards; the painting's lack of idealism rankled viewers. The painting's flatness, inspired by Japanese wood block art, serves to make the nude more human and less voluptuous. A fully dressed black servant is featured, exploiting the then-current theory that black people were hyper-sexed. That she is wearing the clothing of a servant to a courtesan here furthers the sexual tension of the piece. Olympia's body as well as her gaze is unabashedly confrontational. She defiantly looks out as her servant offers flowers from one of her male suitors. Although her hand rests on her leg, hiding her pubic area, the reference to traditional female virtue is ironic; a notion of modesty is notoriously absent in this work. A contemporary critic denounced Olympia's "shamelessly flexed" left hand, which seemed to him a mockery of the relaxed, shielding hand of Titian's Venus. Likewise, the alert black cat at the foot of the bed strikes a sexually rebellious note in contrast to that of the sleeping dog in Titian's portrayal of the goddess in his "Venus of Urbino". "Olympia" was the subject of caricatures in the popular press, but was championed by the French avant-garde community, and the painting's significance was appreciated by artists such as Gustave Courbet, Paul Cézanne, Claude Monet, and later Paul Gauguin. As with "Luncheon on the Grass", the painting raised the issue of prostitution within contemporary France and the roles of women within society. After the death of his father in 1862, Manet married Suzanne Leenhoff in 1863. Leenhoff was a Dutch-born piano teacher two years Manet's senior with whom he had been romantically involved for approximately ten years. Leenhoff initially had been employed by Manet's father, Auguste, to teach Manet and his younger brother piano. She also may have been Auguste's mistress. In 1852, Leenhoff gave birth, out of wedlock, to a son, Leon Koella Leenhoff. Manet painted his wife in "The Reading", among other paintings. Her son, Leon Leenhoff, whose father may have been either of the Manets, posed often for Manet. Most famously, he is the subject of the "Boy Carrying a Sword" of 1861 (Metropolitan Museum of Art, New York). He also appears as the boy carrying a tray in the background of "The Balcony" (1868-69). Manet became friends with the Impressionists Edgar Degas, Claude Monet, Pierre-Auguste Renoir, Alfred Sisley, Paul Cézanne, and Camille Pissarro through another painter, Berthe Morisot, who was a member of the group and drew him into their activities. The supposed grand-niece of the painter Jean-Honoré Fragonard, Morisot had her first painting accepted in the Salon de Paris in 1864, and she continued to show in the salon for the next ten years. Manet became the friend and colleague of Morisot in 1868. She is credited with convincing Manet to attempt plein air painting, which she had been practicing since she was introduced to it by another friend of hers, Camille Corot. They had a reciprocating relationship and Manet incorporated some of her techniques into his paintings. In 1874, she became his sister-in-law when she married his brother, Eugène. Unlike the core Impressionist group, Manet maintained that modern artists should seek to exhibit at the Paris Salon rather than abandon it in favor of independent exhibitions. Nevertheless, when Manet was excluded from the International Exhibition of 1867, he set up his own exhibition. His mother worried that he would waste all his inheritance on this project, which was enormously expensive. While the exhibition earned poor reviews from the major critics, it also provided his first contacts with several future Impressionist painters, including Degas. Although his own work influenced and anticipated the Impressionist style, Manet resisted involvement in Impressionist exhibitions, partly because he did not wish to be seen as the representative of a group identity, and partly because he preferred to exhibit at the Salon. Eva Gonzalès, a daughter of the novelist Emmanuel Gonzalès, was his only formal student. He was influenced by the Impressionists, especially Monet and Morisot. Their influence is seen in Manet's use of lighter colors: after the early 1870s he made less use of dark backgrounds but retained his distinctive use of black, uncharacteristic of Impressionist painting. He painted many outdoor (plein air) pieces, but always returned to what he considered the serious work of the studio. Manet enjoyed a close friendship with composer Emmanuel Chabrier, painting two portraits of him; the musician owned 14 of Manet's paintings and dedicated his "Impromptu" to Manet's wife. One of Manet's frequent models at the beginning of the 1880s was the "semimondaine" Méry Laurent, who posed for seven portraits in pastel. Laurent's salons hosted many French (and even American) writers and painters of her time; Manet had connections and influence through such events. Throughout his life, although resisted by art critics, Manet could number as his champions Émile Zola, who supported him publicly in the press, Stéphane Mallarmé, and Charles Baudelaire, who challenged him to depict life as it was. Manet, in turn, drew or painted each of them. Manet's paintings of café scenes are observations of social life in 19th-century Paris. People are depicted drinking beer, listening to music, flirting, reading, or waiting. Many of these paintings were based on sketches executed on the spot. Manet often visited the Brasserie Reichshoffen on boulevard de Rochechourt, upon which he based "At the Cafe" in 1878. Several people are at the bar, and one woman confronts the viewer while others wait to be served. Such depictions represent the painted journal of a flâneur. These are painted in a style which is loose, referencing Hals and Velázquez, yet they capture the mood and feeling of Parisian night life. They are painted snapshots of bohemianism, urban working people, as well as some of the bourgeoisie. In "Corner of a Cafe Concert", a man smokes while behind him a waitress serves drinks. In "The Beer Drinkers" a woman enjoys her beer in the company of a friend. In "The Cafe Concert", shown at right, a sophisticated gentleman sits at a bar while a waitress stands resolutely in the background, sipping her drink. In "The Waitress", a serving woman pauses for a moment behind a seated customer smoking a pipe, while a ballet dancer, with arms extended as she is about to turn, is on stage in the background. Manet also sat at the restaurant on the Avenue de Clichy called Pere Lathuille's, which had a garden in addition to the dining area. One of the paintings he produced here was "Chez le père Lathuille" (At Pere Lathuille's), in which a man displays an unrequited interest in a woman dining near him. In "Le Bon Bock" (1873), a large, cheerful, bearded man sits with a pipe in one hand and a glass of beer in the other, looking straight at the viewer. Manet painted the upper class enjoying more formal social activities. In "Masked Ball at the Opera", Manet shows a lively crowd of people enjoying a party. Men stand with top hats and long black suits while talking to women with masks and costumes. He included portraits of his friends in this picture. His 1868 painting "The Luncheon" was posed in the dining room of the Manet house. Manet depicted other popular activities in his work. In "The Races at Longchamp", an unusual perspective is employed to underscore the furious energy of racehorses as they rush toward the viewer. In "Skating", Manet shows a well dressed woman in the foreground, while others skate behind her. Always there is the sense of active urban life continuing behind the subject, extending outside the frame of the canvas. In "View of the International Exhibition", soldiers relax, seated and standing, prosperous couples are talking. There is a gardener, a boy with a dog, a woman on horseback—in short, a sample of the classes and ages of the people of Paris. Manet's response to modern life included works devoted to war, in subjects that may be seen as updated interpretations of the genre of "history painting". The first such work was the "Battle of the Kearsarge and Alabama" (1864), a sea skirmish known as the "Battle of Cherbourg" from the American Civil War which took place off the French coast, and may have been witnessed by the artist. Of interest next was the French intervention in Mexico; from 1867 to 1869 Manet painted three versions of the "Execution of Emperor Maximilian", an event which raised concerns regarding French foreign and domestic policy. The several versions of the "Execution" are among Manet's largest paintings, which suggests that the theme was one which the painter regarded as most important. Its subject is the execution by Mexican firing squad of a Habsburg emperor who had been installed by Napoleon III. Neither the paintings nor a lithograph of the subject were permitted to be shown in France. As an indictment of formalized slaughter the paintings look back to Goya, and anticipate Picasso's "Guernica". In January 1871, Manet traveled to Oloron-Sainte-Marie in the Pyrenees. In his absence his friends added his name to the "Fédération des artistes" (see: Courbet) of the Paris Commune. Manet stayed away from Paris, perhaps, until after the "semaine sanglante": in a letter to Berthe Morisot at Cherbourg (10 June 1871) he writes, ""We came back to Paris a few days ago..."" (the semaine sanglante ended on 28 May). The prints and drawings collection of the Museum of Fine Arts (Budapest) has a watercolour/gouache by Manet, "The Barricade", depicting a summary execution of Communards by Versailles troops based on a lithograph of the execution of Maximilian. A similar piece, "The Barricade" (oil on plywood), is held by a private collector. On 18 March 1871, he wrote to his (confederate) friend Félix Bracquemond in Paris about his visit to Bordeaux, the provisory seat of the French National Assembly of the Third French Republic where Émile Zola introduced him to the sites: "I never imagined that France could be represented by such doddering old fools, not excepting that little twit Thiers..." If this could be interpreted as support of the Commune, a following letter to Bracquemond (21 March 1871) expressed his idea more clearly: "Only party hacks and the ambitious, the Henrys of this world following on the heels of the Milliéres, the grotesque imitators of the Commune of 1793..." He knew the communard Lucien Henry to have been a former painter's model and Millière, an insurance agent. "What an encouragement all these bloodthirsty caperings are for the arts! But there is at least one consolation in our misfortunes: that we're not politicians and have no desire to be elected as deputies". The public figure Manet admired most was the republican Léon Gambetta. In the heat of the "seize mai" coup in 1877, Manet opened up his atelier to a republican electoral meeting chaired by Gambetta's friend Eugène Spuller. Manet depicted many scenes of the streets of Paris in his works. "The Rue Mosnier Decked with Flags" depicts red, white, and blue pennants covering buildings on either side of the street; another painting of the same title features a one-legged man walking with crutches. Again depicting the same street, but this time in a different context, is "Rue Mosnier with Pavers", in which men repair the roadway while people and horses move past. "The Railway", widely known as "The Gare Saint-Lazare", was painted in 1873. The setting is the urban landscape of Paris in the late 19th century. Using his favorite model in his last painting of her, a fellow painter, Victorine Meurent, also the model for "Olympia" and the "Luncheon on the Grass", sits before an iron fence holding a sleeping puppy and an open book in her lap. Next to her is a little girl with her back to the painter, watching a train pass beneath them. Instead of choosing the traditional natural view as background for an outdoor scene, Manet opts for the iron grating which "boldly stretches across the canvas" The only evidence of the train is its white cloud of steam. In the distance, modern apartment buildings are seen. This arrangement compresses the foreground into a narrow focus. The traditional convention of deep space is ignored. Historian Isabelle Dervaux has described the reception this painting received when it was first exhibited at the official Paris Salon of 1874: "Visitors and critics found its subject baffling, its composition incoherent, and its execution sketchy. Caricaturists ridiculed Manet's picture, in which only a few recognized the symbol of modernity that it has become today". The painting is currently in the National Gallery of Art in Washington, D.C. Manet painted several boating subjects in 1874. "Boating", now in the Metropolitan Museum of Art, exemplifies in its conciseness the lessons Manet learned from Japanese prints, and the abrupt cropping by the frame of the boat and sail adds to the immediacy of the image. In 1875, a book-length French edition of Edgar Allan Poe's "The Raven" included lithographs by Manet and translation by Mallarmé. In 1881, with pressure from his friend Antonin Proust, the French government awarded Manet the Légion d'honneur. In his mid-forties Manet's health deteriorated, and he developed severe pain and partial paralysis in his legs. In 1879 he began receiving hydrotherapy treatments at a spa near Meudon intended to improve what he believed was a circulatory problem, but in reality he was suffering from locomotor ataxia, a known side-effect of syphilis. In 1880, he painted a portrait there of the opera singer Émilie Ambre as Carmen. Ambre and her lover Gaston de Beauplan had an estate in Meudon and had organized the first exhibition of Manet's "The Execution of Emperor Maximilian" in New York in December 1879. In his last years Manet painted many small-scale still lifes of fruits and vegetables, such as "Bunch of Asparagus" and "The Lemon" (both 1880). He completed his last major work, "A Bar at the Folies-Bergère (Un Bar aux Folies-Bergère)", in 1882 and it hung in the Salon that year. Afterwards he limited himself to small formats. His last paintings were of flowers in glass vases. In April 1883, his left foot was amputated because of gangrene, due to complications from syphilis and rheumatism. He died eleven days later on 30 April in Paris. He is buried in the Passy Cemetery in the city. Manet's public career lasted from 1861, the year of his first participation in the Salon, until his death in 1883. His known extant works, as catalogued in 1975 by Denis Rouart and Daniel Wildenstein, comprise 430 oil paintings, 89 pastels, and more than 400 works on paper. Although harshly condemned by critics who decried its lack of conventional finish, Manet's work had admirers from the beginning. One was Émile Zola, who wrote in 1867: "We are not accustomed to seeing such simple and direct translations of reality. Then, as I said, there is such a surprisingly elegant awkwardness ... it is a truly charming experience to contemplate this luminous and serious painting which interprets nature with a gentle brutality." The roughly painted style and photographic lighting in Manet's paintings was seen as specifically modern, and as a challenge to the Renaissance works he copied or used as source material. He rejected the technique he had learned in the studio of Thomas Couture – in which a painting was constructed using successive layers of paint on a dark-toned ground – in favor of a direct, "alla prima" method using opaque paint on a light ground. Novel at the time, this method made possible the completion of a painting in a single sitting. It was adopted by the Impressionists, and became the prevalent method of painting in oils for generations that followed. Manet's work is considered "early modern", partially because of the opaque flatness of his surfaces, the frequent sketchlike passages, and the black outlining of figures, all of which draw attention to the surface of the picture plane and the material quality of paint. The art historian Beatrice Farwell says Manet "has been universally regarded as the Father of Modernism. With Courbet he was among the first to take serious risks with the public whose favour he sought, the first to make "alla prima" painting the standard technique for oil painting and one of the first to take liberties with Renaissance perspective and to offer ‘pure painting’ as a source of aesthetic pleasure. He was a pioneer, again with Courbet, in the rejection of humanistic and historical subject-matter, and shared with Degas the establishment of modern urban life as acceptable material for high art." The late Manet painting, "Le Printemps" (1881), sold to the J. Paul Getty Museum for $65.1 million, setting a new auction record for Manet, exceeding its pre-sale estimate of $25–35 million at Christie's on 5 November 2014. The previous auction record was held by "Self-Portrait With Palette" which sold for $33.2 million at Sotheby's on 22 June 2010.
https://en.wikipedia.org/wiki?curid=9615
Evolutionarily stable strategy An evolutionarily stable strategy (ESS) is a strategy (or set of strategies) which, if adopted by a population in a given environment, is impenetrable, meaning that it cannot be invaded by any alternative strategy (or strategies) that are initially rare. It is relevant in game theory, behavioural ecology, and evolutionary psychology. An ESS is an equilibrium refinement of the Nash equilibrium. It is a Nash equilibrium that is "evolutionarily" stable: once it is fixed in a population, natural selection alone is sufficient to prevent alternative (mutant) strategies from invading successfully. The theory is not intended to deal with the possibility of gross external changes to the environment that bring new selective forces to bear. First published as a specific term in the 1972 book by John Maynard Smith, the ESS is widely used in behavioural ecology and economics, and has been used in anthropology, evolutionary psychology, philosophy, and political science. Evolutionarily stable strategies were defined and introduced by John Maynard Smith and George R. Price in a 1973 "Nature" paper. Such was the time taken in peer-reviewing the paper for "Nature" that this was preceded by a 1972 essay by Maynard Smith in a book of essays titled "On Evolution". The 1972 essay is sometimes cited instead of the 1973 paper, but university libraries are much more likely to have copies of "Nature". Papers in "Nature" are usually short; in 1974, Maynard Smith published a longer paper in the "Journal of Theoretical Biology". Maynard Smith explains further in his 1982 book "Evolution and the Theory of Games". Sometimes these are cited instead. In fact, the ESS has become so central to game theory that often no citation is given, as the reader is assumed to be familiar with it. Maynard Smith mathematically formalised a verbal argument made by Price, which he read while peer-reviewing Price's paper. When Maynard Smith realized that the somewhat disorganised Price was not ready to revise his article for publication, he offered to add Price as co-author. The concept was derived from R. H. MacArthur and W. D. Hamilton's work on sex ratios, derived from Fisher's principle, especially Hamilton's (1967) concept of an unbeatable strategy. Maynard Smith was jointly awarded the 1999 Crafoord Prize for his development of the concept of evolutionarily stable strategies and the application of game theory to the evolution of behaviour. Uses of ESS: The Nash equilibrium is the traditional solution concept in game theory. It depends on the cognitive abilities of the players. It is assumed that players are aware of the structure of the game and consciously try to predict the moves of their opponents and to maximize their own payoffs. In addition, it is presumed that all the players know this (see common knowledge). These assumptions are then used to explain why players choose Nash equilibrium strategies. Evolutionarily stable strategies are motivated entirely differently. Here, it is presumed that the players' strategies are biologically encoded and heritable. Individuals have no control over their strategy and need not be aware of the game. They reproduce and are subject to the forces of natural selection, with the payoffs of the game representing reproductive success (biological fitness). It is imagined that alternative strategies of the game occasionally occur, via a process like mutation. To be an ESS, a strategy must be resistant to these alternatives. Given the radically different motivating assumptions, it may come as a surprise that ESSes and Nash equilibria often coincide. In fact, every ESS corresponds to a Nash equilibrium, but some Nash equilibria are not ESSes. An ESS is a refined or modified form of a Nash equilibrium. (See the next section for examples which contrast the two.) In a Nash equilibrium, if all players adopt their respective parts, no player can "benefit" by switching to any alternative strategy. In a two player game, it is a strategy pair. Let E("S","T") represent the payoff for playing strategy "S" against strategy "T". The strategy pair ("S", "S") is a Nash equilibrium in a two player game if and only if this is true for both players and for all "T"≠"S": In this definition, strategy "T" can be a neutral alternative to "S" (scoring equally well, but not better). A Nash equilibrium is presumed to be stable even if "T" scores equally, on the assumption that there is no long-term incentive for players to adopt "T" instead of "S". This fact represents the point of departure of the ESS. Maynard Smith and Price specify two conditions for a strategy "S" to be an ESS. For all "T"≠"S", either The first condition is sometimes called a "strict" Nash equilibrium. The second is sometimes called "Maynard Smith's second condition". The second condition means that although strategy "T" is neutral with respect to the payoff against strategy "S", the population of players who continue to play strategy "S" has an advantage when playing against "T". There is also an alternative, stronger definition of ESS, due to Thomas. This places a different emphasis on the role of the Nash equilibrium concept in the ESS concept. Following the terminology given in the first definition above, this definition requires that for all "T"≠"S" In this formulation, the first condition specifies that the strategy is a Nash equilibrium, and the second specifies that Maynard Smith's second condition is met. Note that the two definitions are not precisely equivalent: for example, each pure strategy in the coordination game below is an ESS by the first definition but not the second. In words, this definition looks like this: The payoff of the first player when both players play strategy S is higher than (or equal to) the payoff of the first player when he changes to another strategy T and the second player keeps his strategy S "and" the payoff of the first player when only his opponent changes his strategy to T is higher than his payoff in case that both of players change their strategies to T. This formulation more clearly highlights the role of the Nash equilibrium condition in the ESS. It also allows for a natural definition of related concepts such as a weak ESS or an evolutionarily stable set. In most simple games, the ESSes and Nash equilibria coincide perfectly. For instance, in the prisoner's dilemma there is only one Nash equilibrium, and its strategy ("Defect") is also an ESS. Some games may have Nash equilibria that are not ESSes. For example, in harm thy neighbor (whose payoff matrix is shown here) both ("A", "A") and ("B", "B") are Nash equilibria, since players cannot do better by switching away from either. However, only "B" is an ESS (and a strong Nash). "A" is not an ESS, so "B" can neutrally invade a population of "A" strategists and predominate, because "B" scores higher against "B" than "A" does against "B". This dynamic is captured by Maynard Smith's second condition, since E("A", "A") = E("B", "A"), but it is not the case that E("A","B") > E("B","B"). Nash equilibria with equally scoring alternatives can be ESSes. For example, in the game "Harm everyone", "C" is an ESS because it satisfies Maynard Smith's second condition. "D" strategists may temporarily invade a population of "C" strategists by scoring equally well against "C", but they pay a price when they begin to play against each other; "C" scores better against "D" than does "D". So here although E("C", "C") = E("D", "C"), it is also the case that E("C","D") > E("D","D"). As a result, "C" is an ESS. Even if a game has pure strategy Nash equilibria, it might be that none of those pure strategies are ESS. Consider the Game of chicken. There are two pure strategy Nash equilibria in this game ("Swerve", "Stay") and ("Stay", "Swerve"). However, in the absence of an uncorrelated asymmetry, neither "Swerve" nor "Stay" are ESSes. There is a third Nash equilibrium, a mixed strategy which is an ESS for this game (see Hawk-dove game and Best response for explanation). This last example points to an important difference between Nash equilibria and ESS. Nash equilibria are defined on "strategy sets" (a specification of a strategy for each player), while ESS are defined in terms of strategies themselves. The equilibria defined by ESS must always be symmetric, and thus have fewer equilibrium points. In population biology, the two concepts of an "evolutionarily stable strategy" (ESS) and an "evolutionarily stable state" are closely linked but describe different situations. In an evolutionarily stable "strategy," if all the members of a population adopt it, no mutant strategy can invade. Once virtually all members of the population use this strategy, there is no 'rational' alternative. ESS is part of classical game theory. In an evolutionarily stable "state," a population's genetic composition is restored by selection after a disturbance, if the disturbance is not too large. An evolutionarily stable state is a dynamic property of a population that returns to using a strategy, or mix of strategies, if it is perturbed from that initial state. It is part of population genetics, dynamical system, or evolutionary game theory. This is now called convergent stability. B. Thomas (1984) applies the term ESS to an individual strategy which may be mixed, and evolutionarily stable population state to a population mixture of pure strategies which may be formally equivalent to the mixed ESS. Whether a population is evolutionarily stable does not relate to its genetic diversity: it can be genetically monomorphic or polymorphic. In the classic definition of an ESS, no mutant strategy can invade. In finite populations, any mutant could in principle invade, albeit at low probability, implying that no ESS can exist. In an infinite population, an ESS can instead be defined as a strategy which, should it become invaded by a new mutant strategy with probability p, would be able to counterinvade from a single starting individual with probability >p, as illustrated by the evolution of bet-hedging. A common model of altruism and social cooperation is the Prisoner's dilemma. Here a group of players would collectively be better off if they could play "Cooperate", but since "Defect" fares better each individual player has an incentive to play "Defect". One solution to this problem is to introduce the possibility of retaliation by having individuals play the game repeatedly against the same player. In the so-called "iterated" Prisoner's dilemma, the same two individuals play the prisoner's dilemma over and over. While the Prisoner's dilemma has only two strategies ("Cooperate" and "Defect"), the iterated Prisoner's dilemma has a huge number of possible strategies. Since an individual can have different contingency plan for each history and the game may be repeated an indefinite number of times, there may in fact be an infinite number of such contingency plans. Three simple contingency plans which have received substantial attention are "Always Defect", "Always Cooperate", and "Tit for Tat". The first two strategies do the same thing regardless of the other player's actions, while the latter responds on the next round by doing what was done to it on the previous round—it responds to "Cooperate" with "Cooperate" and "Defect" with "Defect". If the entire population plays "Tit-for-Tat" and a mutant arises who plays "Always Defect", "Tit-for-Tat" will outperform "Always Defect". If the population of the mutant becomes too large — the percentage of the mutant will be kept small. "Tit for Tat" is therefore an ESS, "with respect to only these two strategies". On the other hand, an island of "Always Defect" players will be stable against the invasion of a few "Tit-for-Tat" players, but not against a large number of them. If we introduce "Always Cooperate", a population of "Tit-for-Tat" is no longer an ESS. Since a population of "Tit-for-Tat" players always cooperates, the strategy "Always Cooperate" behaves identically in this population. As a result, a mutant who plays "Always Cooperate" will not be eliminated. However, even though a population of "Always Cooperate" and "Tit-for-Tat" can coexist, if there is a small percentage of the population that is "Always Defect", the selective pressure is against "Always Cooperate", and in favour of "Tit-for-Tat". This is due to the lower payoffs of cooperating than those of defecting in case the opponent defects. This demonstrates the difficulties in applying the formal definition of an ESS to games with large strategy spaces, and has motivated some to consider alternatives. The fields of sociobiology and evolutionary psychology attempt to explain animal and human behavior and social structures, largely in terms of evolutionarily stable strategies. Sociopathy (chronic antisocial or criminal behavior) may be a result of a combination of two such strategies. Evolutionarily stable strategies were originally considered for biological evolution, but they can apply to other contexts. In fact, there are stable states for a large class of adaptive dynamics. As a result, they can be used to explain human behaviours that lack any genetic influences.
https://en.wikipedia.org/wiki?curid=9616
Extremophile An extremophile (from Latin ' meaning "extreme" and Greek ' () meaning "love") is an organism with optimal growth in environmental conditions considered extreme in that it is challenging for a carbon-based life form with water as a solvent, such as all life on Earth, to survive. This is not the same as a more anthropocentric and non-scientific view which considers an extremophile to be an organism that lives in environments uncomfortable to humans. In contrast, organisms that live in more moderate environmental conditions, according to an anthropocentric view, may be termed mesophiles or neutrophiles. In the 1980s and 1990s, biologists found that microbial life has great flexibility for surviving in extreme environments—niches that are acidic or extraordinarily hot, for example—that would be completely inhospitable to complex organisms. Some scientists even concluded that life may have begun on Earth in hydrothermal vents far under the ocean's surface. According to astrophysicist Steinn Sigurdsson, "There are viable bacterial spores that have been found that are 40 million years old on Earth—and we know they're very hardened to radiation." Some bacteria were found living in the cold and dark in a lake buried a half-mile deep under the ice in Antarctica, and in the Marianas Trench, the deepest place in Earth's oceans. Some microorganisms have been found thriving inside rocks up to below the sea floor under of ocean off the coast of the northwestern United States. According to one of the researchers, "You can find microbes everywhere—they're extremely adaptable to conditions, and survive wherever they are." A key to extremophile adaptation is their amino acid composition, affecting their protein folding ability under particular conditions. Studying extreme environments on Earth can help researchers understand the limits of habitability on other worlds. Tom Gheysens from Ghent University in Belgium and some of his colleagues have presented research findings that show spores from a species of Bacillus bacteria survived and were still viable after being heated to temperatures of . There are many classes of extremophiles that range all around the globe; each corresponding to the way its environmental niche differs from mesophilic conditions. These classifications are not exclusive. Many extremophiles fall under multiple categories and are classified as polyextremophiles. For example, organisms living inside hot rocks deep under Earth's surface are thermophilic and piezophilic such as "Thermococcus barophilus". A polyextremophile living at the summit of a mountain in the Atacama Desert might be a radioresistant xerophile, a psychrophile, and an oligotroph. Polyextremophiles are well known for their ability to tolerate both high and low pH levels. Astrobiology is the study of the origin, evolution, distribution, and future of life in the universe: extraterrestrial life and life on Earth. Astrobiology makes use of physics, chemistry, astronomy, solar physics, biology, molecular biology, ecology, planetary science, geography, and geology to investigate the possibility of life on other worlds and help recognize biospheres that might be different from that on Earth. Astrobiologists are particularly interested in studying extremophiles, as it allows them to map what is known about the limits of life on Earth to potential extraterrestrial environments For example, analogous deserts of Antarctica are exposed to harmful UV radiation, low temperature, high salt concentration and low mineral concentration. These conditions are similar to those on Mars. Therefore, finding viable microbes in the subsurface of Antarctica suggests that there may be microbes surviving in endolithic communities and living under the Martian surface. Research indicates it is unlikely that Martian microbes exist on the surface or at shallow depths, but may be found at subsurface depths of around 100 meters. Recent research carried out on extremophiles in Japan involved a variety of bacteria including "Escherichia coli" and "Paracoccus denitrificans" being subject to conditions of extreme gravity. The bacteria were cultivated while being rotated in an ultracentrifuge at high speeds corresponding to 403,627 g (i.e. 403,627 times the gravity experienced on Earth). "Paracoccus denitrificans" was one of the bacteria which displayed not only survival but also robust cellular growth under these conditions of hyperacceleration which are usually found only in cosmic environments, such as on very massive stars or in the shock waves of supernovas. Analysis showed that the small size of prokaryotic cells is essential for successful growth under hypergravity. The research has implications on the feasibility of panspermia. On 26 April 2012, scientists reported that lichen survived and showed remarkable results on the adaptation capacity of photosynthetic activity within the simulation time of 34 days under Martian conditions in the Mars Simulation Laboratory (MSL) maintained by the German Aerospace Center (DLR). On 29 April 2013, scientists at Rensselaer Polytechnic Institute, funded by NASA, reported that, during spaceflight on the International Space Station, microbes seem to adapt to the space environment in ways "not observed on Earth" and in ways that "can lead to increases in growth and virulence". On 19 May 2014, scientists announced that numerous microbes, like "Tersicoccus phoenicis", may be resistant to methods usually used in spacecraft assembly clean rooms. It's not currently known if such resistant microbes could have withstood space travel and are present on the "Curiosity" rover now on the planet Mars. On 20 August 2014, scientists confirmed the existence of microorganisms living half a mile below the ice of Antarctica. On September 2015, scientists from CNR-National Research Council of Italy reported that "S.soflataricus" was able to survive under Martian radiation at a wavelength that was considered extremely lethal to most bacteria. This discovery is significant because it indicates that not only bacterial spores, but also growing cells can be remarkably resistant to strong UV radiation. On June 2016, scientists from Brigham Young University conclusively reported that endospores of "Bacillus subtilis" were able to survive high speed impacts up to 299±28 m/s, extreme shock, and extreme deceleration. They pointed out that this feature might allow endospores to survive and to be transferred between planets by traveling within meteorites or by experiencing atmosphere disruption. Moreover, they suggested that the landing of spacecraft may also result in interplanetary spore transfer, given that spores can survive high-velocity impact while ejected from the spacecraft onto the planet surface. This is the first study which reported that bacteria can survive in such high-velocity impact. However, the lethal impact speed is unknown, and further experiments should be done by introducing higher-velocity impact to bacterial endospores. New sub-types of -philes are identified frequently and the sub-category list for extremophiles is always growing. For example, microbial life lives in the liquid asphalt lake, Pitch Lake. Research indicates that extremophiles inhabit the asphalt lake in populations ranging between 106 to 107 cells/gram. Likewise, until recently boron tolerance was unknown but a strong borophile was discovered in bacteria. With the recent isolation of "Bacillus boroniphilus", borophiles came into discussion. Studying these borophiles may help illuminate the mechanisms of both boron toxicity and boron deficiency. In July 2019, a scientific study of Kidd Mine in Canada discovered sulfur-breathing organisms which live 7900 feet below the surface, and which breathe sulfur in order to survive. These organisms are also remarkable due to eating rocks such as pyrite as their regular food source. The thermoalkaliphilic catalase, which initiates the breakdown of hydrogen peroxide into oxygen and water, was isolated from an organism, "Thermus brockianus", found in Yellowstone National Park by Idaho National Laboratory researchers. The catalase operates over a temperature range from 30 °C to over 94 °C and a pH range from 6–10. This catalase is extremely stable compared to other catalases at high temperatures and pH. In a comparative study, the "T. brockianus" catalase exhibited a half life of 15 days at 80 °C and pH 10 while a catalase derived from "Aspergillus niger" had a half life of 15 seconds under the same conditions. The catalase will have applications for removal of hydrogen peroxide in industrial processes such as pulp and paper bleaching, textile bleaching, food pasteurization, and surface decontamination of food packaging. DNA modifying enzymes such as "Taq" DNA polymerase and some "Bacillus" enzymes used in clinical diagnostics and starch liquefaction are produced commercially by several biotechnology companies. Over 65 prokaryotic species are known to be naturally competent for genetic transformation, the ability to transfer DNA from one cell to another cell followed by integration of the donor DNA into the recipient cell's chromosome. Several extremophiles are able to carry out species-specific DNA transfer, as described below. However, it is not yet clear how common such a capability is among extremophiles. The bacterium "Deinococcus radiodurans" is one of the most radioresistant organisms known. This bacterium can also survive cold, dehydration, vacuum and acid and is thus known as a polyextremophile. "D. radiodurans" is competent to perform genetic transformation. Recipient cells are able to repair DNA damage in donor transforming DNA that had been UV irradiated as efficiently as they repair cellular DNA when the cells themselves are irradiated. The extreme thermophilic bacterium "Thermus thermophilus" and other related "Thermus" species are also capable of genetic transformation. "Halobacterium volcanii", an extreme halophilic (saline tolerant) archaeon, is capable of natural genetic transformation. Cytoplasmic bridges are formed between cells that appear to be used for DNA transfer from one cell to another in either direction. "Sulfolobus solfataricus" and "Sulfolobus acidocaldarius" are hyperthermophilic archaea. Exposure of these organisms to the DNA damaging agents UV irradiation, bleomycin or mitomycin C induces species-specific cellular aggregation. UV-induced cellular aggregation of "S. acidocaldarius" mediates chromosomal marker exchange with high frequency. Recombination rates exceed those of uninduced cultures by up to three orders of magnitude. Frols et al. and Ajon et al. hypothesized that cellular aggregation enhances species-specific DNA transfer between "Sulfolobus" cells in order to repair damaged DNA by means of homologous recombination. Van Wolferen et al. noted that this DNA exchange process may be crucial under DNA damaging conditions such as high temperatures. It has also been suggested that DNA transfer in "Sulfolobus" may be an early form of sexual interaction similar to the more well-studied bacterial transformation systems that involve species-specific DNA transfer leading to homologous recombinational repair of DNA damage (and see Transformation (genetics)). Extracellular membrane vesicles (MVs) might be involved in DNA transfer between different hyperthermophilic archaeal species. It has been shown that both plasmids and viral genomes can be transferred via MVs. Notably, a horizontal plasmid transfer has been documented between hyperthermophilic "Thermococcus" and "Methanocaldococcus" species, respectively belonging to the orders "Thermococcales" and "Methanococcales".
https://en.wikipedia.org/wiki?curid=9619
Education reform Education Reform in the United States is the name given to the goal of changing public education. Historically, reforms have taken different forms because the motivations of reformers have differed. However, since the 1980s, education reform has been focused on changing the existing system from one focused on inputs to one focused on outputs (i.e., student achievement). In the United States, education reform acknowledges and encourages public education as the primary source of K-12 education for American youth. The one constant for all forms of education reform includes the idea that small changes in education will have large social returns in citizen health, wealth and well-being. For example, a stated motivation has been to reduce cost to students and society. From ancient times until the 1800s, one goal was to reduce the expense of a classical education. Ideally, classical education is undertaken with a highly educated full-time (extremely expensive) personal tutor. Historically, this was available only to the most wealthy. Encyclopedias, public libraries and grammar schools are examples of innovations intended to lower the cost of a classical education. Related reforms attempted to develop similar classical results by concentrating on "why", and "which" questions neglected by classical education. Abstract, introspective answers to these questions can theoretically compress large numbers of facts into relatively few principles. This path was taken by some Transcendentalist educators, such as Amos Bronson Alcott. In the early modern age, Victorian schools were reformed to teach commercially useful topics, such as modern languages and mathematics, rather than classical subjects, such as Latin and Greek. Many reformers focused on reforming society by reforming education on more scientific, humanistic, pragmatic or democratic principles. John Dewey and Anton Makarenko are prominent examples of such reformers. Some reformers incorporated several motivations, e.g. Maria Montessori, who both "educated for peace" (a social goal), and to "meet the needs of the child" (A humanistic goal). In historic Prussia, an important motivation for the invention of Kindergarten was to foster national unity by teaching a national language while children were young enough that learning a language was easy. Proponents of evidence-based education call for the use of evidence in guiding education reform. Reform has taken many forms and directions. Throughout history and the present day, the meaning and methods of education have changed through debates over what content or experiences result in an educated individual or an educated society. Changes may be implemented by individual educators and/or by broad-based school organization and/or by curriculum changes with performance evaluations. Plato believed that children would never learn unless they wanted to learn. In "The Republic", he said, " ... compulsory learning never sticks in the mind." An educational debate in the time of the Roman Empire arose after Christianity had achieved broad acceptance. The question concerned the educational value of pre-Christian classical thought: "Given that the body of knowledge of the pre-Christian Romans was heathen in origin, was it safe to teach it to Christian children?" Though educational reform occurred on a local level at various points throughout history, the modern notion of education reform is tied with the spread of compulsory education. Education reforms did not become widespread until after organized schooling was sufficiently systematized to be 'reformed.' In the modern world, economic growth and the spread of democracy have raised the value of education and increased the importance of ensuring that all children and adults have access to high-quality, effective education. Modern education reforms are increasingly driven by a growing understanding of what works in education and how to go about successfully improving teaching and learning in schools. However, in some cases, the reformers' goals of "high-quality education" has meant "high-intensity education", with a narrow emphasis on teaching individual, test-friendly subskills quickly, regardless of long-term outcomes, developmental appropriateness, or broader educational goals. Western classical education as taught from the 18th to the 19th century has missing features that inspired reformers. Classical education is most concerned with answering the who, what, where, and when? questions that concern a majority of students. Unless carefully taught, group instruction naturally neglects the theoretical "why" and "which" questions that strongly concern fewer students. Classical education in this period also did not teach local (vernacular) languages and cultures. Instead it taught high-status ancient languages (Greek and Latin) and their cultures. This produced odd social effects in which an intellectual class might be more loyal to ancient cultures and institutions than to their native vernacular languages and their actual governing authorities. Before there were government-funded public schools, education of the lower classes was by the charity school, pioneered in the 19th century by Protestant organizations and adapted by the Roman Catholic Church and governments. Because these schools operated on very small budgets and attempted to serve as many needy children as possible, they were designed to be inexpensive. The basic program was to develop "grammar" schools. These taught only grammar and bookkeeping. This program permitted people to start businesses to make money, and gave them the skills to continue their education inexpensively from books. "Grammar" was the first third of the then-prevalent system of classical education. The ultimate development of the grammar school was by Joseph Lancaster and Andrew Bell who developed the monitorial system. Lancaster started as a poor Quaker in early 19th century London. Bell started the Madras School of India. The monitorial system uses slightly more-advanced students to teach less-advanced students, achieving student-teacher ratios as small as 2, while educating more than a thousand students per adult. Lancaster promoted his system in a piece called Improvements in Education that spread widely throughout the English-speaking world. Discipline and labor in a Lancaster school were provided by an economic system. Scrip, a form of money meaningless outside the school, was created at a fixed exchange rate from a student's tuition. Every job of the school was bid-for by students in scrip, with the largest bid winning. However, "any" student tutor could auction positions in his or her classes. Besides tutoring, students could use scrip to buy food, school supplies, books, and childish luxuries in a school store. The adult supervisors were paid from the bids on jobs. With fully developed internal economies, Lancaster schools provided a grammar-school education for a cost per student near $40 per year in 1999 U.S. dollars. The students were very clever at reducing their costs, and once invented, improvements were widely adopted in a school. For example, Lancaster students, motivated to save scrip, ultimately rented individual pages of textbooks from the school library, and read them in groups around music stands to reduce textbook costs. Students commonly exchanged tutoring, and paid for items and services with receipts from "down tutoring." Lancaster schools usually lacked sufficient adult supervision. As a result, the older children acting as disciplinary monitors tended to become brutal task masters. Also, the schools did not teach submission to orthodox Christian beliefs or government authorities. As a result, most English-speaking countries developed mandatory publicly paid education explicitly to keep public education in "responsible" hands. These elites said that Lancaster schools might become dishonest, provide poor education and were not accountable to established authorities. Lancaster's supporters responded that any schoolchild could avoid cheats, given the opportunity, and that the government was not paying for the education, and thus deserved no say in their composition. Lancaster, though motivated by charity, claimed in his pamphlets to be surprised to find that he lived well on the income of his school, even while the low costs made it available to the poorest street-children. Ironically, Lancaster lived on the charity of friends in his later life. The term "progressive" in education has been used somewhat indiscriminately; there are a number of kinds of educational progressivism, most of the historically significant kinds peaking in the period between the late 19th and the middle of the 20th centuries. Jean-Jacques Rousseau has been called the father of the child-study movement. It has been said that Rousseau "discovered" the child (as an object of study). Rousseau's principal work on education is "", in which he lays out an educational program for a hypothetical newborn's education to adulthood. Rousseau provided a dual critique of both the vision of education set forth in Plato's Republic and also of the society of his contemporary Europe and the educational methods he regarded as contributing to it; he held that a person can either be a man or a citizen, and that while Plato's plan could have brought the latter at the expense of the former, contemporary education failed at both tasks. He advocated a radical withdrawal of the child from society and an educational process that utilized the natural potential of the child and its curiosity, teaching it by confronting it with simulated real-life obstacles and conditioning it by experience rather than teaching it intellectually. His ideas were rarely implemented directly, but were influential on later thinkers, particularly Johann Heinrich Pestalozzi and Friedrich Wilhelm August Fröbel, the inventor of the kindergarten. In the United States, Horace Mann (1796 – 1859) of Massachusetts used his political base and role as Secretary of the Massachusetts State Board of Education to promote public education in his home state and nationwide. His crusading style attracted wide middle class support. Historian Ellwood P. Cubberley asserts: Education is often seen in Europe and Asia as an important system to maintain national, cultural and linguistic unity. Prussia instituted primary school reforms expressly to teach a unified version of the national language, "Hochdeutsch". One significant reform was kindergarten, whose purpose was to have the children spend time in supervised activities in the national language, when the children were young enough that they could easily learn new language skills. Since most modern schools copy the Prussian models, children start school at an age when their language skills remain plastic, and they find it easy to learn the national language. This was an intentional design on the part of the Prussians. In the U.S. over the last twenty years, more than 70% of non-English-speaking school-age immigrants have arrived in the U.S. before they were 6 years old. At this age, they could have been taught English in school, and achieved a proficiency indistinguishable from a native speaker. In other countries, such as the Soviet Union, France, Spain, and Germany this approach has dramatically improved reading and math test scores for linguistic minorities. John Dewey, a philosopher and educator based in Chicago and New York, helped conceptualize the role of American and international education during the first four decades of the 20th century. An important member of the American Pragmatist movement, he carried the subordination of knowledge to action into the educational world by arguing for experiential education that would enable children to learn theory and practice simultaneously; a well-known example is the practice of teaching elementary physics and biology to students while preparing a meal. He was a harsh critic of "dead" knowledge disconnected from practical human life. Dewey criticized the rigidity and volume of humanistic education, and the emotional idealizations of education based on the child-study movement that had been inspired by Rousseau and those who followed him. He presented his educational theories as a synthesis of the two views. His slogan was that schools should encourage children to "Learn by doing." He wanted people to realize that children are naturally active and curious. Dewey's understanding of logic is best presented in his "Logic, the Theory of Inquiry" (1938). His educational theories were presented in "My Pedagogic Creed", "The School and Society", "The Child and Curriculum", and "Democracy and Education" (1916). Bertrand Russell criticized Dewey's conception of logic, saying "What he calls "logic" does not seem to me to be part of logic at all; I should call it part of psychology." The question of the history of Deweyan educational practice is a difficult one. He was a widely known and influential thinker, but his views and suggestions were often misunderstood by those who sought to apply them, leading some historians to suggest that there was never an actual implementation on any considerable scale of Deweyan progressive education. The schools with which Dewey himself was most closely associated (though the most famous, the "Laboratory School", was really run by his wife) had considerable ups and downs, and Dewey left the University of Chicago in 1904 over issues relating to the Dewey School. Dewey's influence began to decline in the time after the Second World War and particularly in the Cold War era, as more conservative educational policies came to the fore. The form of educational progressivism which was most successful in having its policies implemented has been dubbed "administrative progressivism" by historians. This began to be implemented in the early 20th century. While influenced particularly in its rhetoric by Dewey and even more by his popularizers, administrative progressivism was in its practice much more influenced by the Industrial Revolution and the concept economies of scale. The administrative progressives are responsible for many features of modern American education, especially American high schools: counseling programs, the move from many small local high schools to large centralized high schools, curricular differentiation in the form of electives and tracking, curricular, professional, and other forms of standardization, and an increase in state and federal regulation and bureaucracy, with a corresponding reduction of local control at the school board level. (Cf. "State, federal, and local control of education in the United States", below) (Tyack and Cuban, pp. 17–26) These reforms have since become heavily entrenched, and many today who identify themselves as progressives are opposed to many of them, while conservative education reform during the Cold War embraced them as a framework for strengthening traditional curriculum and standards. In more recent times, groups such as the think tank Reform's education division, and S.E.R. have attempted to pressure the government of the U.K. into more modernist educational reform, though this has met with limited success. From the 1950s to the 1970s, many of the proposed and implemented reforms in U.S. education stemmed from the civil rights movement and related trends; examples include ending racial segregation, and busing for the purpose of desegregation, affirmative action, and banning of school prayer. In the 1980s, some of the momentum of education reform moved from the left to the right, with the release of "A Nation at Risk", Ronald Reagan's efforts to reduce or eliminate the United States Department of Education. "[T]he federal government and virtually all state governments, teacher training institutions, teachers' unions, major foundations, and the mass media have all pushed strenuously for higher standards, greater accountability, more "time on task," and more impressive academic results". This shift to the right caused many families to seek alternatives, including "charter schools, progressive schools, Montessori schools, Waldorf schools, Afrocentric schools, religious schools - or teaching them at home and in their communities." In the latter half of the decade, E. D. Hirsch put forth an influential attack on one or more versions of progressive education, advocating an emphasis on "cultural literacy"—the facts, phrases, and texts that Hirsch asserted every American had once known and that now only some knew, but was still essential for decoding basic texts and maintaining communication. Hirsch's ideas remain significant through the 1990s and into the 21st century, and are incorporated into classroom practice through textbooks and curricula published under his own imprint. Most states and districts in the 1990s adopted Outcome-Based Education (OBE) in some form or another. A state would create a committee to adopt standards, and choose a quantitative instrument to assess whether the students knew the required content or could perform the required tasks. The standards-based National Education Goals (Goals 2000) were set by the U.S. Congress in the 1990s. Many of these goals were based on the principles of outcomes-based education, and not all of the goals were attained by the year 2000 as was intended. The standards-based reform movement culminated in the No Child Left Behind Act of 2001, which is still an active nationwide mandate in the United States. OBE reforms usually had other disputed methods, such as constructivist mathematics and whole language, added onto them. Some proponents advocated replacing the traditional high school diploma with a Certificate of Initial Mastery. Other reform movements were school-to-work, which would require all students except those in a university track to spend substantial class time on a job site. See also Uncommon Schools. President Donald Trump relegated concerns in education to state governments. This began with the Every Student Succeeds Act (ESSA) which limits the role of the federal government in school liability. Giving states more authority can help prevent considerable discrepancies in educational performance across different states. ESSA was approved by former President Obama in 2015 which amended and empowered the Elementary and Secondary Education Act of 1965. The Department of Education has the choice to carry out measures in drawing attention to said differences by pinpointing lowest-performing state governments and supplying information on the condition and progress of each state on different educational parameters. It can also provide reasonable funding along with technical aid to help states with similar demographics collaborate in improving their public education programs. During his campaign, Trump criticized the 2010 Common Core States Standard and other cases of “federal government overreach.” His advocacy was to give state and local governments more responsibilities over education policies. Trump appointed Betsy DeVos as education secretary. She also supported the idea of leaving education to state governments under the new K-12 legislation. DeVos cited the interventionist approach of the federal government to education policy following the signing of the ESSA. The primary approach to that rule has not changed significantly. Her opinion was that the education movement's populist politics or populism. encouraged reformers to commit promises which were not very realistic and therefore difficult to deliver. Many opinion makers say the situation in all American social institutions is the same. These institutions which include government, higher education, healthcare, and mass media are still attuned with the traditional or original economic system. There is a need to upgrade to a digital information economy. More providers of higher education which include colleges and universities, non-traditional entities like school districts, libraries, and museums, and for-profit organizations will surface. All of these stakeholders will reach out to bigger audiences and use similar tools and technologies to achieve their goals. An article released by CBNC.com said a principal Senate Committee will take into account legislation that reauthorizes and modernizes the Carl D. Perkins Act. President George Bush approved this statute in 2006 on August 12, 2006. This new bill will emphasize the importance of federal funding for various Career and Technical (CTE) programs that will better provide learners with in-demand skills. Congress can provide more students with access to pertinent skills in education according to 21st century career opportunities. At present, there are many initiatives aimed at dealing with these concerns like innovative cooperation between federal and state governments, educators, and the business sector. One of these efforts is the Pathways to Technology Early College High School (P-TECH). This six-year program was launched in cooperation with IBM, educators from three cities in New York, Chicago, and Connecticut, and over 400 businesses. The program offers students high school and associate programs focusing on the STEM curriculum. The High School Involvement Partnership, private and public venture, was established through the help of Northrop Grumman, a global security firm. It has given assistance to some 7,000 high school students (juniors and seniors) since 1971 by means of one-on-one coaching as well as exposure to STEM areas and careers. In 2016, Time.com published an article mentioning that one way of reenergizing the United States economy is to provide quality education and training opportunities for American youngsters. There is a need to update funding streams for schools at the federal, state, and local levels such as Pell Grants addressing the requirements of college students. The Grant or specific amount of money is given by the government every school year for disadvantaged students who need to pay tuition fees in college. Higher education in the United States of America has always been regarded as exceptional worldwide although there are apprehensions regarding expensive and quality education, unimpressive completion rates, and increasing student debt. These issues raised doubts as to the effectiveness of the conventional approach to higher education. There have been numerous proposals for federal reforms to enhance the status of higher education in the US. Some of the recommendations included making institutions liable for students/ non-attendance or dropping out of school, changing the obsolete accreditation process in overseeing access to federal subsidies, and allowing access to free education. This uses a methodology that values purposeful engagement in activities that turn students into self-reliant and efficient learners. Holding on to the view that everyone possesses natural gifts that are unique to one's personality (e.g. computational aptitude, musical talent, visual arts abilities), it likewise upholds the idea that children, despite their inexperience and tender age, are capable of coping with anguish, able to survive hardships, and can rise above difficult times. President Donald Trump signed the Strengthening Career and Technical Education for the 21st Century Act (HR 2353) on July 31, 2018. This is the first law the American president signed that made meaningful amendments to the federal education system. It reauthorizes the Carl D. Perkins Career and Technical Education Act, a $1.2 billion program modified by the United States Congress in 2006. Legislators have repeatedly rebuffed the efforts of Trump and education secretary Betsy DeVos to implement school choice programs funded by the federal government. The move to change the Higher Education Act was also deferred. Business and education groups such as the Council of Chief State School Officers as well as the National Governors Association commended the US Congress for its prompt work during the past month. However, some advocacy organizations like Advanced CTE and Association for Career and Technical Education are apprehensive that said law can urge states to set passive laws for Career and Technical Education. The new legislation takes effect on July 1, 2019 and takes the place of the Carl D. Perkins Career and Technical Education (Perkins IV) Act of 2006. Stipulations in Perkins V enables school districts to make use of federal subsidies for all students' career search and development activities in the middle grades as well as comprehensive guidance and academic mentoring in the upper grades. At the same time, this law updates and magnifies the meaning of "special populations" to include homeless persons, foster youth, those who left the foster care system, and children with parents on active duty in the United States armed forces. In the first decade of the 21st century, several issues are salient in debates over further education reform: According to a 2005 report from the OECD, the United States is tied for first place with Switzerland when it comes to annual spending per student on its public schools, with each of those two countries spending more than $11,000 (in U.S. currency). Despite this high level of funding, U.S. public schools lag behind the schools of other rich countries in the areas of reading, math, and science. A further analysis of developed countries shows no correlation between per student spending and student performance, suggesting that there are other factors influencing education. Top performers include Singapore, Finland and Korea, all with relatively low spending on education, while high spenders including Norway and Luxembourg have relatively low performance. One possible factor is the distribution of the funding. In the US, schools in wealthy areas tend to be over-funded while schools in poorer areas tend to be underfunded. These differences in spending between schools or districts may accentuate inequalities, if they result in the best teachers moving to teach in the most wealthy areas. The inequality between districts and schools led to 23 states instituting school finance reform based on adequacy standards that aim to increase funding to low-income districts. A 2018 study found that between 1990 and 2012, these finance reforms led to an increase in funding and test scores in the low income districts; which suggests finance reform is effective at bridging inter-district performance inequalities. It has also been shown that the socioeconomic situation of the students family has the most influence in determining success; suggesting that even if increased funds in a low income area increase performance, they may still perform worse than their peers from wealthier districts. Starting in the early 1980s, a series of analyses by Eric Hanushek indicated that the amount spent on schools bore little relationship to student learning. This controversial argument, which focused attention on how money was spent instead of how much was spent, led to lengthy scholarly exchanges. In part the arguments fed into the class size debates and other discussions of "input policies." It also moved reform efforts towards issues of school accountability (including No Child Left Behind) and the use of merit pay and other incentives. There have been studies that show smaller class sizes and newer buildings (both of which require higher funding to implement) lead to academic improvements. It should also be noted that many of the reform ideas that stray from the traditional format require greater funding. It has been shown that some school districts do not use their funds in the most productive way. For example, according to a 2007 article in the "Washington Post", the Washington, D.C. public school district spends $12,979 per student per year. This is the third highest level of funding per student out of the 100 biggest school districts in the United States. Despite this high level of funding, the school district provides outcomes that are lower than the national average. In reading and math, the district's students score the lowest among 11 major school districts—even when poor children are compared only with other poor children. 33% of poor fourth graders in the United States lack basic skills in math, but in Washington, D.C., it's 62%. According to a 2006 study by the Goldwater Institute, Arizona's public schools spend 50% more per student than Arizona's private schools. The study also says that while teachers constitute 72% of the employees at private schools, they make up less than half of the staff at public schools. According to the study, if Arizona's public schools wanted to be like private schools, they would have to hire approximately 25,000 more teachers, and eliminate 21,210 administration employees. The study also said that public school teachers are paid about 50% more than private school teachers. In 1985 in Kansas City, Missouri, a judge ordered the school district to raise taxes and spend more money on public education. Spending was increased so much, that the school district was spending more money per student than any of the country's other 280 largest school districts. According to a 1999 article, William J. Bennett, former U.S. Secretary of Education, argued that increased levels of spending on public education have not made the schools better, citing the following statistics: In the United States, private schools (independent schools) have long been an alternative to public education for those with the ability to pay tuition. These include religious schools, preparatory and boarding schools, and schools based on alternative paradigms such as Montessori education. Over 4 million students, about one in twelve children attend religious schools in the United States, most of them Christian. Montessori pre- and primary school programs employ rigorously tested scientific theories of guided exploration which seek to embrace children's natural curiosity rather than, for instance, scolding them for falling out of rank. Home education is favored by a growing number of parents who take direct responsibility for their children's education rather than enrolling them in local public schools seen as not meeting expectations. Economists such as Nobel laureate Milton Friedman advocate school choice to promote excellence in education through competition and choice. A competitive "market" for schools eliminates the need to otherwise attempt a workable method of accountability for results. Public education vouchers permit guardians to select and pay any school, public or private, with public funds currently allocated to local public schools. The theory is that children's guardians will naturally shop for the best schools, much as is already done at college level. Though appealing in theory, many reforms based on school choice have led to slight to moderate improvements—which some teachers' union members see as insufficient to offset the decreased teacher pay and job security. For instance, New Zealand's landmark reform in 1989, during which schools were granted substantial autonomy, funding was devolved to schools, and parents were given a free choice of which school their children would attend, led to moderate improvements in most schools. It was argued that the associated increases in inequity and greater racial stratification in schools nullified the educational gains. Others, however, argued that the original system created more inequity (due to lower income students being required to attend poorer performing inner city schools and not being allowed school choice or better educations that are available to higher income inhabitants of suburbs). Instead, it was argued that the school choice promoted social mobility and increased test scores especially in the cases of low income students. Similar results have been found in other jurisdictions. Though discouraging, the merely slight improvements of some school choice policies often seems to reflect weaknesses in the way that choice is implemented rather than a failure of the basic principle itself. Critics of teacher tenure claim that the laws protect ineffective teachers from being fired, which can be detrimental to student success. Tenure laws vary from state to state, but generally they set a probationary period during which the teacher proves themselves worthy of the lifelong position. Probationary periods range from one to three years. Advocates for tenure reform often consider these periods too short to make such an important decision; especially when that decision is exceptionally hard to revoke. Due process restriction protect tenured teachers from being wrongfully fired; however these restrictions can also prevent administrators from removing ineffective or inappropriate teachers. A 2008 survey conducted by the US Department of Education found that, on average, only 2.1% of teachers are dismissed each year for poor performance. In October 2010 Apple Inc. CEO Steve Jobs had a consequential meeting with U.S. President Barack Obama to discuss U.S. competitiveness and the nation's education system. During the meeting Jobs recommended pursuing policies that would make it easier for school principals to hire and fire teachers based on merit. In 2012 tenure for school teachers was challenged in a California lawsuit called "Vergara v. California". The primary issue in the case was the impact of tenure on student outcomes and on equity in education. On June 10, 2014, the trial judge ruled that California's teacher tenure statute produced disparities that " shock the conscience" and violate the equal protection clause of the California Constitution. On July 7, 2014, U.S. Secretary of Education Arne Duncan commented on the "Vergara" decision during a meeting with President Barack Obama and representatives of teacher's unions. Duncan said that tenure for school teachers "should be earned through demonstrated effectiveness" and should not be granted too quickly. Specifically, he criticized the 18-month tenure period at the heart of the "Vergara" case as being too short to be a "meaningful bar." A study by the Fordham Institute found that some labor agreements with teachers' unions may restrict the ability of school systems to implement merit pay and other reforms. Contracts were more restrictive in districts with high concentrations of poor and minority students. The methodology and conclusions of the study have been criticized by teachers' unions. Another barrier to reform is assuming that schools are like businesses—when in fact they are very different. Legal barriers to reform are low in the United States compared to other countries: State and local governance of education creates "wiggle room for educational innovators" who can change local laws or move somewhere more favourable. Cultural barriers to reform are also relatively low, because the question of who should control education is still open. There are factors that can impede innovations in K-12 education. One could be “Site-Based Decision Making Councils” composed of teachers and some parents who vote on school rules and regulations, adoption of curriculum, hiring of new mentors, and other related matters. There are times attendance in meetings is not adequate or stakeholders are not represented properly. The belief is small meetings attended by a few individuals may not be ideal for innovation. Turnover of teachers is another possible hindrance to such innovations. The learning process is adversely affected because of frequent teacher resignations and replacements. Constant changing of mentors leads to waste of resources and dormant thinking influenced by policies, systems, and traditions. Education 2030 Agenda refers to the global commitment of the Education for All movement to ensure access to basic education for all. It is an essential part of the 2030 Agenda for Sustainable Development. The roadmap to achieve the Agenda is the Education 2030 Incheon Declaration and Framework for Action, which outlines how countries, working with UNESCO and global partners, can translate commitments into action. The United Nations, over 70 ministers, representatives of member-countries, bilateral and multilateral agencies, regional organizations, academic institutions, teachers, civil society, and the youth supported the Framework for Action of the Education 2030 platform. The Framework was described as the outcome of continuing consultation to provide guidance for countries in implementing this Agenda. At the same time, it mobilizes various stakeholders in the new education objectives, coordination, implementation process, funding, and review of Education 2030. In 1995, the minister of education, Sukavich Rangsitpol, launched a series of education reforms in 1995 with the intention of the education reform is to realize the potential of Thai people to develop themselves for a better quality of life and to develop the nation for a peaceful co-existence in the global community. According to UNESCO, Thailand education reform has led to the following results: World Bank report that after the 1997 Asian financial crisis Income in the northeast, the poorest part of Thailand, has risen by 46 percent from 1998 to 2006. Nationwide poverty fell from 21.3 to 11.3 percent. Education reform has been pursued for a variety of specific reasons, but generally most reforms aim at redressing some societal ills, such as poverty-, gender-, or class-based inequities, or perceived ineffectiveness. Current education trends in the United States represent multiple achievement gaps across ethnicities, income levels, and geographies. As McKinsey and Company reported in a 2009 analysis, “These educational gaps impose on the United States the economic equivalent of a permanent national recession.” Reforms are usually proposed by thinkers who aim to redress societal ills or institute societal changes, most often through a change in the education of the members of a class of people—the preparation of a ruling class to rule or a working class to work, the social hygiene of a lower or immigrant class, the preparation of citizens in a democracy or republic, etc. The idea that all children should be provided with a high level of education is a relatively recent idea, and has arisen largely in the context of Western democracy in the 20th century. The "beliefs" of school districts are optimistic that quite literally "all students will succeed", which in the context of high school graduation examination in the United States, all students in all groups, regardless of heritage or income will pass tests that in the introduction typically fall beyond the ability of all but the top 20 to 30 percent of students. The claims clearly renounce historical research that shows that all ethnic and income groups score differently on all standardized tests and standards based assessments and that students will achieve on a bell curve. Instead, education officials across the world believe that by setting clear, achievable, higher standards, aligning the curriculum, and assessing outcomes, learning can be increased for all students, and more students can succeed than the 50 percent who are defined to be above or below grade level by norm referenced standards. States have tried to use state schools to increase state power, especially to make better soldiers and workers. This strategy was first adopted to unify related linguistic groups in Europe, including France, Germany and Italy. Exact mechanisms are unclear, but it often fails in areas where populations are culturally segregated, as when the U.S. Indian school service failed to suppress Lakota and Navaho, or when a culture has widely respected autonomous cultural institutions, as when the Spanish failed to suppress Catalan. Many students of democracy have desired to improve education in order to improve the quality of governance in democratic societies; the necessity of good public education follows logically if one believes that the quality of democratic governance depends on the ability of citizens to make informed, intelligent choices, and that education can improve these abilities. Politically motivated educational reforms of the democratic type are recorded as far back as Plato in "The Republic". In the United States, this lineage of democratic education reform was continued by Thomas Jefferson, who advocated ambitious reforms partly along Platonic lines for public schooling in Virginia. Another motivation for reform is the desire to address socio-economic problems, which many people see as having significant roots in lack of education. Starting in the 20th century, people have attempted to argue that small improvements in education can have large returns in such areas as health, wealth and well-being. For example, in Kerala, India in the 1950s, increases in women's health were correlated with increases in female literacy rates. In Iran, increased primary education was correlated with increased farming efficiencies and income. In both cases some researchers have concluded these correlations as representing an underlying causal relationship: education causes socio-economic benefits. In the case of Iran, researchers concluded that the improvements were due to farmers gaining reliable access to national crop prices and scientific farming information. Reforms can be based on bringing education into alignment with a society's core values. Reforms that attempt to change a society's core values can connect alternative education initiatives with a network of other alternative institutions. Proponents of the evidence-based education movement call for the use of evidence in guiding education reform. Evidence-based education is the use of well designed scientific studies to determine which education methods work best. Evidence-based learning techniques such as spaced repetition have been shown to increase the rate at which students learn. The evidence-based education movement has its roots in the larger movement towards evidence-based-practices. The movement to use computers more in education naturally includes many unrelated ideas, methods, and pedagogies since there are many uses for digital computers. For example, the fact that computers are naturally good at math leads to the question of the use of calculators in math education. The Internet's communication capabilities make it potentially useful for collaboration, and foreign language learning. The computer's ability to simulate physical systems makes it potentially useful in teaching science. More often, however, debate of digital education reform centers around more general applications of computers to education, such as electronic test-taking and online classes. The idea of creating artificial intelligence led some computer scientists to believe that teachers could be replaced by computers, through something like an expert system; however, attempts to accomplish this have predictably proved inflexible. The computer is now more understood to be a tool or assistant for the teacher and students. Harnessing the richness of the Internet is another goal. In some cases classrooms have been moved entirely online, while in other instances the goal is more to learn how the Internet can be more than a classroom. Web-based international educational software is under development by students at New York University, based on the belief that current educational institutions are too rigid: effective teaching is not routine, students are not passive, and questions of practice are not predictable or standardized. The software allows for courses tailored to an individual's abilities through frequent and automatic multiple intelligences assessments. Ultimate goals include assisting students to be intrinsically motivated to educate themselves, and aiding the student in self-actualization. Courses typically taught only in college are being reformatted so that they can be taught to any level of student, whereby elementary school students may learn the foundations of any topic they desire. Such a program has the potential to remove the bureaucratic inefficiencies of education in modern countries, and with the decreasing digital divide, help developing nations rapidly achieve a similar quality of education. With an open format similar to Wikipedia, any teacher may upload their courses online and a feedback system will help students choose relevant courses of the highest quality. Teachers can provide links in their digital courses to webcast videos of their lectures. Students will have personal academic profiles and a forum will allow students to pose complex questions, while simpler questions will be automatically answered by the software, which will bring you to a solution by searching through the knowledge database, which includes all available courses and topics. The 21st century ushered in the acceptance and encouragement of internet research conducted on college and university campuses, in homes, and even in gathering areas of shopping centers. Addition of cyber cafes on campuses and coffee shops, loaning of communication devices from libraries, and availability of more portable technology devices, opened up a world of educational resources. Availability of knowledge to the elite had always been obvious, yet provision of networking devices, even wireless gadget sign-outs from libraries, made availability of information an expectation of most persons. Cassandra B. Whyte researched the future of computer use on higher education campuses focusing on student affairs. Though at first seen as a data collection and outcome reporting tool, the use of computer technology in the classrooms, meeting areas, and homes continued to unfold. The sole dependence on paper resources for subject information diminished and e-books and articles, as well as on-line courses, were anticipated to become increasingly staple and affordable choices provided by higher education institutions according to Whyte in a 2002 presentation. Digitally "flipping" classrooms is a trend in digital education that has gained significant momentum. Will Richardson, author and visionary for the digital education realm, points to the not-so-distant future and the seemingly infinite possibilities for digital communication linked to improved education. Education on the whole, as a stand-alone entity, has been slow to embrace these changes. The use of web tools such as wikis, blogs, and social networking sites is tied to increasing overall effectiveness of digital education in schools. Examples exist of teacher and student success stories where learning has transcended the classroom and has reached far out into society. Creativity is of the utmost importance when improving education. The "creative teachers" must have the confidence through training and availability of support and resources. These creative teachers are strongly encouraged to embrace a person-centered approach that develops the psychology of the educator ahead or in conjunction with the deployment of machines. Creative teachers have been also been inspired through Crowd-Accelerated Innovation. Crowd-Accelerated Innovation has pushed people to transition between media types and their understanding thereof at record-breaking paces. This process serves as a catalyst for creative direction and new methods of innovation. Innovation without desire and drive inevitably flat lines. Mainstream media continues to be both very influential and the medium where Crowd-Accelerated Innovation gains its leverage. Media is in direct competition with formal educational institutions in shaping the minds of today and those of tomorrow. [Buchanan, Rachel footnote] The media has been instrumental in pushing formal educational institutions to become savvier in their methods. Additionally, advertising has been (and continues to be) a vital force in shaping students and parents thought patterns. Technology is a dynamic entity that is constantly in flux. As time presses on, new technologies will continue to break paradigms that will reshape human thinking regarding technological innovation. This concept stresses a certain disconnect between teachers and learners and the growing chasm that started some time ago. Richardson asserts that traditional classroom's will essentially enter entropy unless teachers increase their comfort and proficiency with technology. Administrators are not exempt from the technological disconnect. They must recognize the existence of a younger generation of teachers who were born during the Digital Age and are very comfortable with technology. However, when old meets new, especially in a mentoring situation, conflict seems inevitable. Ironically, the answer to the outdated mentor may be digital collaboration with worldwide mentor webs; composed of individuals with creative ideas for the classroom. Another viable addition to digital education has been blended learning. In 2009, over 3 million K-12 students took an online course, compared to 2000 when 45,000 took an online course. Blended learning examples include pure online, blended, and traditional education. Research results show that the most effective learning takes place in a blended format. This allows children to view the lecture ahead of time and then spend class time practicing, refining, and applying what they have previously learned.
https://en.wikipedia.org/wiki?curid=9620
Ellensburg, Washington Ellensburg is a city in and county seat of Kittitas County, Washington, United States. The population was 21,111 in a July 2019 census estimate. It is located just east of the Cascade Range at the junction of Interstate 90 and Interstate 82. The city is located along the Yakima River in the Kittitas Valley, an agricultural region that extends east towards the Columbia River. The valley is a major producer of timothy hay, which is processed and shipped internationally. Ellensburg is also the home of Central Washington University (CWU). Ellensburg, originally named Ellensburgh for the wife of town founder John Alden Shoudy, was founded in 1871 and grew rapidly in the 1880s following the arrival of the Northern Pacific Railway. The city was once a leading candidate to become the state capital of Washington, but fell out of favor after a major fire in 1889. Ellensburg was officially incorporated on November 26, 1883. John Alden Shoudy came to the Kittitas Valley in 1871, and purchased a small trading post from Andrew Jackson "A.J." Splawn, called "Robber's Roost." Robber's Roost was the first business in the valley, other than the early trading that occurred among Native Americans, cattle drivers, trappers, and miners. Robber's Roost was located on the present-day 3rd Avenue, just west of Main Street near the alley. There is a placard on the wall commemorating the location, as well as a small stone monument against the wall on the sidewalk. Shoudy named the town after his wife, Mary Ellen Shoudy; thus officially began the city of Ellensburgh around 1872. Shoudy was not the first settler in the Kittitas Valley, nor was he the first businessperson, but he was responsible for platting the city of Ellensburgh in the 1870s, and he was the person who named the streets in the downtown district. The city was originally named Ellensburgh, until the final -"h" was dropped under standardization pressure from the United States Postal Service and Board of Geography Names in 1894. Ellensburg was among the first cities in Washington to have electrical service and was an early center of commerce. The city launched a bid to become the state capital in 1889, preparing a site in the Capital Hill neighborhood for government offices. On July 4, 1889, a major fire destroyed much of downtown Ellensburg and stalled the capital campaign, which resumed with a series of referendums, in which Washington voters chose Olympia. The state legislature selected Ellensburg as the location for the State Normal School (now Central Washington University) as a consolation prize for losing the capital bid. There were several early newspapers in Ellensburg. "The Daily Record", however, began in 1909 and is the name of the local newspaper today. Concerns over the state of Ellensburg's historic downtown led to the formation of the Ellensburg Downtown Association to work on revitalizing the area. The City of Ellensburg is home to a number of local art museums and galleries: Every first Friday of each month, Ellensburg hosts First Friday Art Walk from 5:00 to 7:00 pm. This downtown event showcases art of all forms. The local businesses, galleries and museums come alive with art, music, wine and people as they celebrate art in the community. According to the United States Census Bureau, the city has a total area of , of which is land and is water. Owing to the strong Cascade rain shadow, Ellensburg experiences a typical Intermountain cool semi-arid climate (Köppen "BSk"). As of the census of 2010, there were 18,174 people, 7,301 households, and 2,889 families living in the city. The population density was . There were 7,867 housing units at an average density of . The racial makeup of the city was 85.7% White, 1.5% African American, 1.0% Native American, 3.2% Asian, 0.2% Pacific Islander, 4.6% from other races, and 3.7% from two or more races. Hispanic or Latino of any race were 9.7% of the population. There were 7,301 households, of which 19.3% had children under the age of 18 living with them, 28.2% were married couples living together, 8.2% had a female householder with no husband present, 3.1% had a male householder with no wife present, and 60.4% were non-families. 35.1% of all households were made up of individuals and 9.6% had someone living alone who was 65 years of age or older. The average household size was 2.16 and the average family size was 2.86. The median age in the city was 23.5 years. 14.2% of residents were under the age of 18; 41.2% were between the ages of 18 and 24; 21.8% were from 25 to 44; 13.9% were from 45 to 64; and 8.9% were 65 years of age or older. The gender makeup of the city was 50.1% male and 49.9% female. As of the census of 2000, there were 15,414 people, 6,249 households, and 2,649 families living in the city. The population density was 2,338.9 people per square mile (903.1/km²). There were 6,732 housing units at an average density of 1,021.5 per square mile (394.4/km²). The racial makeup of the city was 88.07% White, 1.17% Black or African American, 0.95% Native American, 4.09% Asian, 0.16% Pacific Islander, 2.86% from other races, and 2.69% from two or more races. 6.33% of the population were Hispanic or Latino of any race. There were 6,249 households, of which 20.8% had children under the age of 18 living with them, 31.4% were married couples living together, 8.1% had a female householder with no husband present, and 57.6% were non-families. 35.5% of all households were made up of individuals and 9.1% had someone living alone who was 65 years of age or older. The average household size was 2.12 and the average family size was 2.84. In the city, the population was spread out with 15.8% under the age of 18, 39.3% from 18 to 24, 22.7% from 25 to 44, 12.8% from 45 to 64, and 9.4% who were 65 years of age or older. The median age was 24 years. For every 100 females, there were 95.0 males. For every 100 females age 18 and over, there were 93.1 males. The median income for a household in the city was $20,034, and the median income for a family was $37,625. Males had a median income of $31,022 versus $22,829 for females. The per capita income for the city was $13,662. About 18.8% of families and 34.3% of the population were below the poverty line, including 29.0% of those under age 18 and 11.2% of those age 65 or over. The City of Ellensburg uses the Manager/Council form of government with a City Manager hired by the City Council. The seven-member City Council is elected at large and serve 4-year terms. The City Council elects a Mayor and Deputy Mayor from the Council to serve 2-year terms. The Council meets the first and third Monday of each month, at 7:00 pm, in the City Council Chambers at City Hall. On the state legislative level, Ellensburg is in the 13th district. As of May, 2018, its state senator is Republican Judy Warnick, and its two state representatives are Republicans Matt Manweller and Tom Dent. On the congressional level, Ellensburg is located in Washington's 8th congressional district and is represented by Democrat Kim Schrier. Kittitas County is served by the "Daily Record", a newspaper published in Ellensburg five days a week. The city maintains its own public library, which opened on January 20, 1910, using funds donated by Andrew Carnegie. Public schools are operated by Ellensburg School District 401. The district includes one high school (Ellensburg High School), one middle school, and three elementary schools.
https://en.wikipedia.org/wiki?curid=9621
Eugene, Oregon Eugene ( ) is a city in the U.S. state of Oregon, in the Pacific Northwest. It is at the southern end of the verdant Willamette Valley, near the confluence of the McKenzie and Willamette Rivers, about east of the Oregon Coast. As of the 2010 census, Eugene had a population of 156,185; it is the county seat of Lane County and the state's third most populous city after Portland and Salem, though recent state estimates suggest its population may have surpassed Salem's. The Eugene-Springfield, Oregon metropolitan statistical area (MSA) is the 146th largest metropolitan statistical area in the US and the third-largest in the state, behind the Portland Metropolitan Area and the Salem Metropolitan Area. The city's population for 2019 was estimated to be 172,622 by the U.S. Census. Eugene is home to the University of Oregon, Northwest Christian University, and Lane Community College. Several spots within Eugene are believed to be inspiration for locations in "The Simpsons". This includes Max's Tavern and the Pioneer Statue on campus. The city is also noted for its natural environment, recreational opportunities (especially bicycling, running/jogging, rafting, and kayaking), and focus on the arts. Eugene's official slogan is "A Great City for the Arts and Outdoors". It is also referred to as the "Emerald City" and as "Track Town, USA". The Nike corporation had its beginnings in Eugene. In 2022, the city will host the 18th Track and Field World Championships. The first people to settle in the Eugene area were known as the Kalapuyans, also written Calapooia or Calapooya. They made "seasonal rounds," moving around the countryside to collect and preserve local foods, including acorns, the bulbs of the wapato and camas plants, and berries. They stored these foods in their permanent winter village. When crop activities waned, they returned to their winter villages and took up hunting, fishing, and trading. They were known as the Chifin Kalapuyans and called the Eugene area where they lived "Chifin", sometimes recorded as "Chafin" or "Chiffin". Other Kalapuyan tribes occupied villages that are also now within Eugene city limits. Pee-you or Mohawk Calapooians, Winefelly or Pleasant Hill Calapooians, and the Lungtum or Long Tom. They were close-neighbors to the Chifin, intermarried, and were political allies. Some authorities suggest the Brownsville Kalapuyans (Calapooia Kalapuyans) were related to the Pee-you. It is likely that since the Santiam had an alliance with the Brownsville Kalapuyans that the Santiam influence also went as far at Eugene. According to archeological evidence, the ancestors of the Kalapuyans may have been in Eugene for as long as 10,000 years. In the 1800s their traditional way of life faced significant changes due to devastating epidemics and settlement, first by French fur traders and later by an overwhelming number of American settlers French fur traders had settled seasonally in the Willamette Valley by the beginning of the 19th century. Their settlements were concentrated in the "French Prairie" community in Northern Marion County but may have extended south to the Eugene area. Having already developed relationships with Native communities through intermarriage and trade, they negotiated for land from the Kalapuyans. By 1828 to 1830 they and their Native wives began year-round occupation of the land, raising crops and tending animals. In this process, the mixed race families began to impact Native access to land, food supply, and traditional materials for trade and religious practices. In July 1830, "intermittent fever" struck the lower Columbia region and a year later, the Willamette Valley. Natives traced the arrival of the disease, then new to the Northwest, to the U.S. ship, Owyhee, captained by John Dominis. "Intermittent fever" is thought by researchers now to be malaria. According to Robert T. Boyd, an anthropologist at Portland State University, the first three years of the epidemic, "probably constitute the single most important epidemiological event in the recorded history of what would eventually become the state of Oregon". In his book "The Coming of the Spirit Pestilence" Boyd reports there was a 92% population loss for the Kalapuyans between 1830 and 1841. This catastrophic event shattered the social fabric of Kalapuyan society and altered the demographic balance in the Valley. This balance was further altered over the next few years by the arrival of Anglo-American settlers, beginning in 1840 with 13 people and growing steadily each year until within 20 years more than 11,000 American settlers, including Eugene Skinner, had arrived. As the demographic pressure from the settlers grew, the remaining Kalapuyans were forcibly removed to Indian reservations. Though some Natives escaped being swept into the reservation, most were moved to the Grand Ronde reservation in 1856. Strict racial segregation was enforced and mixed race people, known as Métis in French, had to make a choice between the reservation and Anglo society. Native Americans could not leave the reservation without traveling papers and white people could not enter the reservation. Eugene Franklin Skinner, after whom Eugene is named, arrived in the Willamette Valley in 1846 with 1,200 other settlers that year. Advised by the Kalapuyans to build on high ground to avoid flooding, he erected the first Anglo cabin on south or west slope of what the Kalapuyans called Ya-po-ah. The "isolated hill" is now known as Skinner's Butte. The cabin was used as a trading post and was registered as an official post office on January 8, 1850. At this time the settlement was known by Anglos as Skinner's Mudhole. It was relocated in 1853 and named Eugene City in 1853. Formally incorporated as a city in 1862, it was named simply Eugene in 1889. Skinner ran a ferry service across the Willamette River where the Ferry Street Bridge now stands. The first major educational institution in the area was Columbia College, founded a few years earlier than the University of Oregon. It fell victim to two major fires in four years, and after the second fire, the college decided not to rebuild again. The part of south Eugene known as College Hill was the former location of Columbia College. There is no college there today. The town raised the initial funding to start a public university, which later became the University of Oregon, with the hope of turning the small town into a center of learning. In 1872, the Legislative Assembly passed a bill creating the University of Oregon as a state institution. Eugene bested the nearby town of Albany in the competition for the state university. In 1873, community member J.H.D. Henderson donated the hilltop land for the campus, overlooking the city. The university first opened in 1876 with the regents electing the first faculty and naming John Wesley Johnson as president. The first students registered on October 16, 1876. The first building was completed in 1877; it was named Deady Hall in honor of the first Board of Regents President and community leader Judge Matthew P. Deady. Eugene grew rapidly throughout most of the twentieth century, with the exception being the early 1980s when a downturn in the timber industry caused high unemployment. By 1985, the industry had recovered and Eugene began to attract more high-tech industries, earning it the moniker the "Emerald Shire". The first Nike shoe was used in 1972 during the US Olympic Trials held in Eugene. According to the United States Census Bureau, the city has a total area of , of which is land and is water. Eugene is at an elevation of . To the north of downtown is Skinner Butte. Northeast of the city is the Coburg Hills. Spencer Butte is a prominent landmark south of the city. Mount Pisgah is southeast of Eugene and includes Mount Pisgah Arboretum and Howard Buford Recreation Area, a Lane County Park. Eugene is surrounded by foothills and forests to the south, east, and west, while to the north the land levels out into the Willamette Valley and consists of mostly farmland. The Willamette and McKenzie Rivers run through Eugene and its neighboring city, Springfield. Another important stream is Amazon Creek, whose headwaters are near Spencer Butte. The creek discharges west of the city into Fern Ridge Reservoir, maintained for winter flood control by the Army Corps of Engineers. The Eugene Yacht Club hosts a sailing school and sailing regattas at Fern Ridge during summer months. Eugene has 23 neighborhood associations: Like the rest of the Willamette Valley, Eugene lies in the Marine West Coast climate zone, with Mediterranean characteristics. Under the Köppen climate classification scheme, Eugene has a warm-summer Mediterranean climate (Köppen "Csb"). Temperatures can vary from cool to warm, with warm, dry summers and cool, wet winters. Spring and fall are also moist seasons, with light rain falling for long periods. The average rainfall is , with the wettest "rain year" being from July 1973 to June 1974 with and the driest from July 2000 to June 2001 with . Winter snowfall does occur, but it is sporadic and rarely accumulates in large amounts: the normal seasonal amount is , but the median is zero. The record snowfall was of accumulation due to a pineapple express on January 25–29, 1969. Ice storms, like snowfall, are rare, but occur sporadically. The hottest months are July and August, with a normal monthly mean temperature of , with an average of 16 days per year reaching . The coolest month is December, with a mean temperature of , and there are 53 mornings per year with a low at or below freezing, and 2.7 afternoons with highs not exceeding the freezing mark. Eugene's average annual temperature is , and annual precipitation at . Eugene is more wet and slightly cooler on average than Portland. Despite being about south and having only a slightly higher elevation, Eugene has a more continental climate, less subject to the maritime air that blows inland from the Pacific Ocean via the Columbia River. Eugene's normal annual mean minimum is , compared to in Portland; in August, the gap in the normal mean minimum widens to and for Eugene and Portland, respectively. Average winter temperatures (and summer high temperatures) are similar for the two cities. This disparity may be additionally caused by Portland's urban heat island, where the combination of black pavement and urban energy use raises nighttime temperatures. Extreme temperatures range from , recorded on December 8, 2013, to on August 9, 1981; the record cold daily maximum is , recorded on December 13, 1919, while, conversely, the record warm daily minimum is on July 22, 2006. Eugene is downwind of Willamette Valley grass seed farms. The combination of summer grass pollen and the confining shape of the hills around Eugene make it "the area of the highest grass pollen counts in the USA (>1,500 pollen grains/m3 of air)." These high pollen counts have led to difficulties for some track athletes who compete in Eugene. In the Olympic trials in 1972, "Jim Ryun won the 1,500 after being flown in by helicopter because he was allergic to Eugene's grass seed pollen." Further, six-time Olympian Maria Mutola abandoned Eugene as a training area "in part to avoid allergies". According to the 2010 census, Eugene's population was 156,185. The population density was 3,572.2 people per square mile. There were 69,951 housing units at an average density of 1,600 per square mile. Those age 18 and over accounted for 81.8% of the total population. The racial makeup of the city was 85.8% White, 4.0% Asian, 1.4% Black or African American, 1.0% Native American, 0.2% Pacific Islander, and 4.7% from other races. Hispanics and Latinos of any race accounted for 7.8% of the total population. Of the non-Hispanics, 82% were White, 1.3% Black or African American, 0.8% Native American, 4% Asian, 0.2% Pacific Islander, 0.2% some other race alone, and 3.4% were of two or more races. Females represented 51.1% of the total population, and males represented 48.9%. The median age in the city was 33.8 years. The census of 2000 showed there were 137,893 people, 58,110 households, and 31,321 families residing in the city of Eugene. The population density was 3,404.8 people per square mile (1,314.5/km²). There were 61,444 housing units at an average density of 1,516.4 per square mile (585.5/km²). The racial makeup of the city was 88.15% White, down from 99.5% in 1950, 3.57% Asian, 1.25% Black or African American, 0.93% Native American, 0.21% Pacific Islander, 2.18% from other races, and 3.72% from two or more races. 4.96% of the population were Hispanic or Latino of any race. There were 58,110 households, of which 25.8% had children under the age of 18 living with them, 40.6% were married couples living together, 9.7% had a female householder with no husband present, and 46.1% were non-families. 31.7% of all households were made up of individuals and 9.4% had someone living alone who was 65 years of age or older. The average household size was 2.27 and the average family size was 2.87. In the city, the population was 20.3% under the age of 18, 17.3% from 18 to 24, 28.5% from 25 to 44, 21.8% from 45 to 64, and 12.1% who were 65 years of age or older. The median age was 33 years. For every 100 females, there were 96.0 males. For every 100 females age 18 and over, there were 94.0 males. The median income for a household in the city was $35,850, and the median income for a family was $48,527. Males had a median income of $35,549 versus $26,721 for females. The per capita income for the city was $21,315. About 8.7% of families and 17.1% of the population were below the poverty line, including 14.8% of those under age 18 and 7.1% of those age 65 or over. Religious institutions of higher learning in Eugene include Northwest Christian University and New Hope Christian College. Northwest Christian University (formerly Northwest Christian College), founded in 1895, has ties with the Christian Church (Disciples of Christ). New Hope Christian College (formerly Eugene Bible College) originated with the Bible Standard Conference in 1915, which joined with Open Bible Evangelistic Association to create Open Bible Standard Churches in 1932. Eugene Bible College was started from this movement by Fred Hornshuh in 1925. There are two Eastern Orthodox Church parishes in Eugene: St John the Wonderworker Orthodox Christian Church in the Historic Whiteaker Neighborhood and Saint George Greek Orthodox Church. There are six Roman Catholic parishes in Eugene as well: St. Mary Catholic Church, St. Jude Catholic Church, St. Mark Catholic Church, St. Peter Catholic Church, St. Paul Catholic Church, and St. Thomas More Catholic Church. Eugene also has a Ukrainian Catholic Church named Nativity of the Mother of God. There is a mainline Protestant contingency in the city as well—such as the largest of the Lutheran Churches, Central Lutheran near the U of O Campus and the Episcopal Church of the Resurrection. The Eugene area has a sizeable LDS Church presence, with three stakes, consisting of 23 congregations (wards and branches). The Portland Oregon Temple is the nearest temple. The greater Eugene-Springfield area also has a Jehovah's Witnesses presence with five Kingdom Halls, several having multiple congregations in one Kingdom Hall. The Reconstructionist Temple Beth Israel is Eugene's largest Jewish congregation. It was also, for many decades, Eugene's only synagogue, until Orthodox members broke away in 1992 and formed "Congregation Ahavas Torah". Eugene has a community of some 140 Sikhs, who have established a Sikh temple. The 340-member congregation of the Unitarian Universalist Church in Eugene (UUCE) purchased the former Eugene Scottish Rite Temple in May 2010, renovated it, and began services there in September 2012. Saraha Nyingma Buddhist Temple in Eugene opened in 2012 in the former site of the Unitarian Universalist Church. Environmental activism In January 2006, the FBI conducted Operation Backfire, leading to federal indictment of eleven people, all members of a Eugene-based cell of the Earth Liberation Front (ELF). Operation Backfire was the largest investigation into radical underground environmental groups in United States history. Ongoing trials of accused eco-terrorists kept Eugene in the spotlight for a few years. Eugene's largest employers are PeaceHealth Medical Group, the University of Oregon, and the Eugene School District. Eugene's largest industries are wood products manufacturing and recreational vehicle manufacturing. Luckey's Club Cigar Store is one of the oldest bars in Oregon. Tad Luckey, Sr., purchased it in 1911, making it one of the oldest businesses in Eugene. The "Club Cigar", as it was called in the late 19th century, was for many years a men-only salon. It survived both the Great Depression and Prohibition, partly because Eugene was a dry town before the end of Prohibition. Corporate headquarters for the employee-owned Bi-Mart corporation and family-owned supermarket Market of Choice remain in Eugene. Eugene has quickly made a name for itself as a culinary hub in Oregon. The city has over 25 breweries, offers a variety of fine dining options with a local focus; the city is surrounded by award-winning wineries. The most notable fungi here is the truffle; Eugene hosts the annual Oregon Truffle Festival in January. Organically Grown Company, the largest distributor of organic fruits and vegetables in the northwest, started in Eugene in 1978 as a non-profit co-op for organic farmers. Notable local food processors, many of whom manufacture certified organic products, include Golden Temple (Yogi Tea), Merry Hempsters and Springfield Creamery (Nancy's Yogurt & owned by the Kesey Family), and Mountain Rose Herbs. Until July 2008, Hynix Semiconductor America had operated a large semiconductor plant in west Eugene. In late September 2009, Uni-Chem of South Korea announced its intention to purchase the Hynix site for solar cell manufacturing. However, this deal fell through and as of late 2012, is no longer planned. In 2015, semiconductor manufacturer Broadcom purchased the plant with plans to upgrade and reopen it. The company abandoned these plans and put it up for sale in November 2016. The footwear repair product Shoe Goo is manufactured by Eclectic Products, based in Eugene. Run Gum, an energy gum created for runners, also began its life in Eugene. Run Gum was created by track athlete Nick Symmonds and track and field coach Sam Lapray in 2014. Burley Design LLC produces bicycle trailers and was founded in Eugene by Alan Scholz out of a Saturday Market business in 1978. Eugene is also the birthplace and home of Bike Friday bicycle manufacturer Green Gear Cycling. Many multinational businesses were launched in Eugene. Some of the most famous include Nike, Taco Time, and Brøderbund Software. In 2012, the Eugene metro region was dubbed the Silicon Shire for its growing tech industry. According to Eugene's 2017 Comprehensive Annual Financial Report, the city's top employers are: Eugene has a significant population of people in pursuit of alternative ideas and a large original hippie population. Beginning in the 1960s, the countercultural ideas and viewpoints espoused by Ken Kesey became established as the seminal elements of the vibrant social tapestry that continue to define Eugene. The Merry Prankster, as Kesey was known, has arguably left the most indelible imprint of any cultural icon in his hometown. He is best known as the author of "One Flew Over the Cuckoo's Nest" and as the male protagonist in Tom Wolfe's "The Electric Kool-Aid Acid Test". In 2005, the city council unanimously approved a new slogan for the city: "World's Greatest City for the Arts & Outdoors". While Eugene has a vibrant arts community for a city its size, and is well situated near many outdoor opportunities, this slogan was frequently criticized by locals as embarrassing and ludicrous. In early 2010, the slogan was changed to "A Great City for the Arts & Outdoors." Eugene's Saturday Market, open every Saturday from April through November, was founded in 1970 as the first "Saturday Market" in the United States. It is adjacent to the Lane County Farmer's Market in downtown Eugene. All vendors must create or grow all their own products. The market reappears as the "Holiday Market" between Thanksgiving and New Year's in the Lane County Events Center at the fairgrounds. Eugene is noted for its "community inventiveness." Many U.S. trends in community development originated in Eugene. The University of Oregon's participatory planning process, known as The Oregon Experiment, was the result of student protests in the early 1970s. The book of the same name is a major document in modern enlightenment thinking in planning and architectural circles. The process, still used by the university in modified form, was created by Christopher Alexander, whose works also directly inspired the creation of the Wiki. Some research for the book "A Pattern Language", which inspired the Design Patterns movement and Extreme Programming, was done by Alexander in Eugene. Not coincidentally, those engineering movements also had origins here. Decades after its publication, "A Pattern Language" is still one of the best-selling books on urban design. In the 1970s, Eugene was packed with cooperative and community projects. It still has small natural food stores in many neighborhoods, some of the oldest student cooperatives in the country, and alternative schools have been part of the school district since 1971. The old Grower's Market, downtown near the Amtrak depot, is the only food cooperative in the U.S. with no employees. It is possible to see Eugene's trend-setting non-profit tendencies in much newer projects, such as the Tango Center and the Center for Appropriate Transport. In 2006, an initiative began to create a tenant-run development process for downtown Eugene. In the fall of 2003, neighbors noticed "an unassuming two-acre remnant orchard tucked into the Friendly Area Neighborhood" had been put up for sale by its owner, a resident of New York City. Learning a prospective buyer had plans to build several houses on the property, they formed a nonprofit organization called Madison Meadow in June 2004 in order to buy the property and "preserve it as undeveloped space in perpetuity." In 2007 their effort was named Third Best Community Effort by the "Eugene Weekly", and by the end of 2008 they had raised enough money to purchase the property. The City of Eugene has an active Neighborhood Program. Several neighborhoods are known for their green activism. Friendly Neighborhood has a highly popular neighborhood garden established on the right of way of a street never built. There are a number of community gardens on public property. Amazon Neighborhood has a former church turned into a community center. Whiteaker hosts a housing co-op that dates from the early 1970s that has re-purposed both their parking lots into food production and play space. An unusual eco-village with natural building techniques and large shared garden can be found in Jefferson Westside neighborhood. A several block area in the River Road Neighborhood is known as a permaculture hotspot with an increasing number of suburban homes trading grass for garden, installing rain water catchment systems, food producing landscapes and solar retrofits. Several sites have planted gardens by removing driveways. Citizen volunteers are working with the City of Eugene to restore a 65-tree filbert grove on public property. There are deepening social and economic networks in the neighborhood. Eugene museums include the University of Oregon's Jordan Schnitzer Museum of Art and Museum of Natural and Cultural History, the Oregon Air and Space Museum, Lane County History Museum, Maude Kerns Art Center, Shelton McMurphey Johnson House, and the Eugene Science Center. Eugene is home to numerous cultural organizations, including the Eugene Symphony, the Eugene Ballet, the Eugene Opera, the Eugene Concert Choir, the Northwest Christian University Community Choir, the Oregon Mozart Players, the Oregon Bach Festival, the Oregon Children's Choir, the Eugene-Springfield Youth Orchestras, Ballet Fantastique and Oregon Festival of American Music. Principal performing arts venues include the Hult Center for the Performing Arts, The John G. Shedd Institute for the Arts ("The Shedd"), Matthew Knight Arena, Beall Concert Hall and the Erb Memorial Union ballroom on the University of Oregon campus, the McDonald Theatre, and W.O.W. Hall. A number of live theater groups are based in Eugene, including Free Shakespeare in the Park, Oregon Contemporary Theatre, The Very Little Theatre, Actors Cabaret, LCC Theatre, Rose Children's Theatre, and University Theatre. Each has its own performance venue. Because of its status as a college town, Eugene has been home to many music genres, musicians and bands, ranging from electronic dance music such as dubstep and drum and bass to garage rock, hip hop, folk and heavy metal. Eugene also has a growing reggae and street-performing bluegrass and jug band scene. Multi-genre act the Cherry Poppin' Daddies became a prominent figure in Eugene's music scene and became the house band at Eugene's W.O.W. Hall. In the late 1990s, their contributions to the swing revival movement propelled them to national stardom. Rock band Floater originated in Eugene as did the Robert Cray blues band. Doom metal band YOB is among the leaders of the Eugene heavy music scene. Eugene is home to "Classical Gas" Composer and two-time Grammy award winner Mason Williams who spent his years as a youth living between his parents in Oakridge, Oregon and Oklahoma. Mason Williams puts on a yearly Christmas show at the Hult center for performing arts with a full orchestra produced by author, audio engineer and University of Oregon professor Don Latarski. Dick Hyman, noted jazz pianist and musical director for many of Woody Allen's films, designs and hosts the annual Now Hear This! jazz festival at the Oregon Festival of American Music (OFAM). OFAM and the Hult Center routinely draw major jazz talent for concerts. Eugene is also home to a large Zimbabwean music community. Kutsinhira Cultural Arts Center, which is "dedicated to the music and people of Zimbabwe," is based in Eugene. Eugene's visual arts community is supported by over 20 private art galleries and several organizations, including Maude Kerns Art Center, Lane Arts Council, DIVA (the Downtown Initiative for the Visual Arts) and the Eugene Glass School. In 2015 installations from a group of Eugene-based artists known as Light At Play were showcased in several events around the world as part of the International Year of Light, including displays at the Smithsonian and the National Academy of Sciences. The Eugene area has been used as a filming location for several Hollywood films, most famously for 1978's "National Lampoon's Animal House", which was also filmed in nearby Cottage Grove. John Belushi had the idea for the film "The Blues Brothers" during filming of "Animal House" when he happened to meet Curtis Salgado at what was then the Eugene Hotel. "Getting Straight", starring Elliott Gould and Candice Bergen, was filmed at Lane Community College in 1969. As the campus was still under construction at the time, the "occupation scenes" were easier to shoot. The "Chicken Salad on Toast" scene in the 1970 Jack Nicholson movie "Five Easy Pieces" was filmed at the Denny's restaurant at the southern I-5 freeway interchange near Glenwood. Nicholson directed the 1971 film "Drive, He Said" in Eugene. "How to Beat the High Co$t of Living", starring Jane Curtin, Jessica Lange and Susan St. James, was filmed in Eugene in the fall of 1979. Locations visible in the film include Valley River Center (which is a driving force in the plot), Skinner Butte and Ya-Po-Ah Terrace, the Willamette River and River Road Hardware. Several track and field movies have used Eugene as a setting and/or a filming location. "Personal Best", starring Mariel Hemingway, was filmed in Eugene in 1982. The film centered on a group of women who are trying to qualify for the Olympic track and field team. Two track and field movies about the life of Steve Prefontaine, "Prefontaine" and "Without Limits", were released within a year of each other in 1997–1998. Kenny Moore, Eugene-trained Olympic runner and co-star in "Prefontaine", co-wrote the screenplay for "Without Limits". "Prefontaine" was filmed in Washington because the "Without Limits" production bought out Hayward Field for the summer to prevent its competition from shooting there. Kenny Moore also wrote a biography of Bill Bowerman, played in "Without Limits" by Donald Sutherland back in Eugene 20 years after he had appeared in "Animal House". Moore had also had a role in "Personal Best". "Stealing Time", a 2003 independent film, was partially filmed in Eugene. When the film premiered in June 2001 at the Seattle International Film Festival, it was titled "Rennie's Landing" after a popular bar near the University of Oregon campus. The title was changed for its DVD release. "Zerophilia" was filmed in Eugene in 2006. Eugene's Oregon Ducks are part of the Pac-12 Conference (Pac-12). American football is especially popular, with intense rivalries between the Ducks and both the Oregon State University Beavers and the University of Washington Huskies. Autzen Stadium is home to Duck football, with a seating capacity of 54,000 but has had over 60,000 with standing room only. The basketball arena, McArthur Court, was built in 1926. The arena was replaced by the Matthew Knight Arena in late 2010. For nearly 40 years, Eugene has been the "Track and Field Capital of the World." Oregon's most famous track icon is the late world-class distance runner Steve Prefontaine, who was killed in a car crash in 1975. Eugene's jogging trails include Pre's Trail in Alton Baker Park, Rexius Trail, the Adidas Oregon Trail, and the Ridgeline Trail. Jogging was introduced to the U.S. through Eugene, brought from New Zealand by Bill Bowerman, who wrote the best-selling book "Jogging", and coached the champion University of Oregon track and cross country teams. During Bowerman's tenure, his "Men of Oregon" won 24 individual NCAA titles, including titles in 15 out of the 19 events contested. During Bowerman's 24 years at Oregon, his track teams finished in the top ten at the NCAA championships 16 times, including four team titles (1962, '64, '65, '70), and two second-place trophies. His teams also posted a dual meet record of 114–20. Bowerman also invented the waffle sole for running shoes in Eugene, and with Oregon alumnus Phil Knight founded shoe giant Nike. Eugene's miles of running trails, through its unusually large park system, are the most extensive in the U.S. The city has dozens of running clubs. The climate is cool and temperate, good both for jogging and record-setting. Eugene is home to the University of Oregon's Hayward Field track, which hosts numerous collegiate and amateur track and field meets throughout the year, most notably the Prefontaine Classic. Hayward Field was host to the 2004 AAU Junior Olympic Games, the 1989 World Masters Athletics Championships, the track and field events of the 1998 World Masters Games, the 2006 Pacific-10 track and field championships, the 1971, 1975, 1986, 1993, 1999, 2001, 2009, and 2011 USA Track & Field Outdoor Championships and the 1972, 1976, 1980, 2008, 2012, and 2016 U.S. Olympic trials. On April 16, 2015, it was announced by the IAAF that Eugene had been awarded the right to host the 2021 World Athletics Championships. The city bid for the 2019 event but lost narrowly to Doha, Qatar. Eugene is also home to the Eugene Emeralds, a short-season Class A minor-league baseball team. The "Ems" play their home games in PK Park, also the home of the University of Oregon baseball team. The Nationwide Tour's golfing event Oregon Classic takes place at Shadow Hills Country Club, just north of Eugene. The event has been played every year since 1998, except in 2001 when it was slated to begin the day after the 9/11 terrorist attacks. The top 20 players from the Nationwide Tour are promoted to the PGA Tour for the following year. The Eugene Jr. Generals, a Tier III Junior "A" ice hockey team belonging to the Northern Pacific Hockey League (NPHL) consisting of 8 teams throughout Oregon and Washington, plays at the Lane County Ice Center. Lane United FC, a soccer club that participates in the Northwest Division of USL League Two, was founded in 2013 and plays its home games at Civic Park. The following table lists some sports clubs in Eugene and their usual home venue: Spencer Butte Park at the southern edge of town provides access to Spencer Butte, a dominant feature of Eugene's skyline. Hendricks Park, situated on a knoll to the east of downtown, is known for its rhododendron garden and nearby memorial to Steve Prefontaine, known as Pre's Rock, where the legendary University of Oregon runner was killed in an auto accident. Alton Baker Park, next to the Willamette River, contains Pre's Trail. Also next to the Willamette are Skinner Butte Park and the Owen Memorial Rose Garden, which is home to more than 4,500 roses of over 400 varieties, as well as the 150-year-old Black Tartarian Cherry tree, an Oregon Heritage Tree. The city of Eugene maintains an urban forest. The University of Oregon campus is an arboretum, with over 500 species of trees. The city operates and maintains scenic hiking trails that pass through and across the ridges of a cluster of hills in the southern portion of the city, on the fringe of residential neighborhoods. Some trails allow biking, and others are for hikers and runners only. The nearest ski resort, Willamette Pass, is one hour from Eugene by car. On the way, along Oregon Route 58, are several reservoirs and lakes, the Oakridge mountain bike trails, hot springs, and waterfalls within Willamette National Forest. Eugene residents also frequent the Hoodoo and Mount Bachelor ski resorts. The Three Sisters Wilderness, the Oregon Dunes National Recreation Area, and Smith Rock are just a short drive away. In 1944, Eugene adopted a council–manager form of government, replacing the day-to-day management of city affairs by the part-time mayor and volunteer city council with a full-time professional city manager. The subsequent history of Eugene city government has largely been one of the dynamics—often contentious—between the city manager, the mayor and city council. The mayor of Eugene is Lucy Vinis, who has been in office since winning the popular vote in 2017. Recent mayors include Edwin Cone (1958–69), Les Anderson (1969–77) Gus Keller (1977–84), Brian Obie (1985–88), Jeff Miller (1989–92), Ruth Bascom (1993–96), Jim Torrey (1997–2004) and Kitty Piercy (2005-2017). Jon Ruiz has been the city manager since April 2008. Ten other people have held the city manager position. They were: Deane Seeger (1945–49), Oren King (1949–53), Robert Finlayson (1953–59), Hugh McKinley (1959–75), Charles Henry (1975–80), Mike Gleason (1981–96), Vicki Elmer (1996–98), Jim Johnson (1998–2002), Dennis Taylor (2002–2007), Angel Jones (2007–2008). Mayor: Lucy Vinis The Eugene Police Department (EPD) is the city's law enforcement and public safety agency. The Lane County Sheriff's Office also has its headquarters in Eugene. The University of Oregon is served by the University of Oregon Police Department (UOPD), and EPD has a police station in the West University District near campus. Lane Community College is served by the Lane Community College Public Safety Department. The Oregon State Police have a presence in the rural areas and highways around the Eugene metro area. The LTD downtown station, and the EmX lines are patrolled by LTD Transit Officers. Since 1989 the mental health crisis intervention non-governmental agency CAHOOTS has responded to Eugene's mental health 911 calls. Eugene City Hall was abandoned in 2012 for reasons of structural integrity, energy efficiency, and obsolete size. Various offices of city government became tenants in eight other buildings. Eugene is home to the University of Oregon. Other institutions of higher learning include Northwest Christian University, Lane Community College, New Hope Christian College, Gutenberg College, and Pacific University's Eugene campus. The Eugene School District includes four full-service high schools (Churchill, North Eugene, Sheldon, and South Eugene) and many alternative education programs, such as international schools and charter schools. Foreign language immersion programs in the district are available in Spanish, French, Chinese, and Japanese. The Bethel School District serves children in the Bethel neighborhood on the northwest edge of Eugene. The district is home to the traditional Willamette High School and the alternative Kalapuya High School. There are 11 schools in this district. Eugene also has several private schools, including the Eugene Waldorf School, the Outdoor High School, Eugene Montessori, Far Horizon Montessori, Eugene Sudbury School, Wellsprings Friends School, Oak Hill School, and The Little French School. Parochial schools in Eugene include Marist Catholic High School, O'Hara Catholic Elementary School, Eugene Christian School, and St. Paul Parish School. The largest library in Oregon is the University of Oregon's Knight Library, with collections totaling more than 3 million volumes and over 100,000 audio and video items. The Eugene Public Library moved into a new, larger building downtown in 2002. The four-story library is an increase from . There are also two branches of the Eugene Public Library, the Sheldon Branch Library in the neighborhood of Cal Young/Sheldon, and the Bethel Branch Library, in the neighborhood of Bethel. Eugene also has the Lane County Law Library. The largest newspaper serving the area is "The Register-Guard", a daily newspaper with a circulation of about 70,000, published independently by the Baker family of Eugene until 2018, before being acquired by GateHouse Media. Other newspapers serving the area include the "Eugene Weekly", the "Emerald", the student-run independent newspaper at the University of Oregon, now published on Mondays and Thursdays;"The Torch", the student-run newspaper at Lane Community College, the "Ignite", the newspaper at New Hope Christian College and "The Beacon Bolt," the student-run newspaper at Northwest Christian University. "Eugene Magazine", "Lifestyle Quarterly", "Eugene Living", and "Sustainable Home and Garden" magazines also serve the area. "Adelante Latino" is a Spanish language newspaper in Eugene that serves all of Lane County. Local television stations include KMTR (NBC), KVAL (CBS), KLSR-TV (Fox), KEVU-CD, KEZI (ABC), KEPB (PBS), and KTVC (independent). The local NPR affiliates are KOPB, and KLCC. Radio station KRVM-AM is an affiliate of Jefferson Public Radio, based at Southern Oregon University. The Pacifica Radio affiliate is the University of Oregon student-run radio station, KWVA. Additionally, the community supports two other radio stations: KWAX (classical) and KRVM-FM (alternative). AM Stations FM Stations Lane Transit District (LTD), a public transportation agency formed in 1970, covers of Lane County, including Creswell, Cottage Grove, Junction City, Veneta, and Blue River. Operating more than 90 buses during peak hours, LTD carries riders on 3.7 million trips every year. LTD also operates a bus rapid transit line that runs between Eugene and Springfield—Emerald Express (EmX)—much of which runs in its own lane, with stations providing for off-board fare payment. LTD's main terminus in Eugene is at the Eugene Station. LTD also offers paratransit. Greyhound Lines provides service between Los Angeles and Portland on the I-5 corridor. Cycling is popular in Eugene and many people commute via bicycle. Summertime events and festivals frequently have bike parking "corrals" that are often filled to capacity by three hundred or more bikes. Many people commute to work by bicycle every month of the year. Bike trails take commuting and recreational bikers along the Willamette River past a scenic rose garden, along Amazon Creek, through the downtown, and through the University of Oregon campus. In 2009, the League of American Bicyclists cited Eugene as 1 of 10 "Gold-level" cities in the U.S. because of its "remarkable commitments to bicycling." In 2010, "Bicycling" magazine named Eugene the 5th most bike-friendly city in America. The U.S. Census Bureau's annual American Community Survey reported that Eugene had a bicycle commuting mode share of 7.3% in 2011, the fifth highest percentage nationwide among U.S. cities with 65,000 people or more, and 13 times higher than the national average of 0.56%. The 1908 Amtrak depot downtown was restored in 2004; it is the southern terminus for two daily runs of the Amtrak "Cascades", and a stop along the route in each direction for the daily "Coast Starlight". Air travel is served by the Eugene Airport, also known as Mahlon Sweet Field, which is the fifth largest airport in the Northwest and second largest airport in Oregon. The Eugene Metro area also has numerous private airports. The Eugene Metro area also has several heliports, such as the Sacred Heart Medical Center Heliport and Mahlon Sweet Field Heliport, and many single helipads. Highways traveling within and through Eugene include: Eugene is the home of Oregon's largest publicly owned water and power utility, the Eugene Water & Electric Board (EWEB). EWEB got its start in the first decade of the 20th century, after an epidemic of typhoid found in the groundwater supply. The City of Eugene condemned Eugene's private water utility and began treating river water (first the Willamette; later the McKenzie) for domestic use. EWEB got into the electric business when power was needed for the water pumps. Excess electricity generated by the EWEB's hydropower plants was used for street lighting. Natural gas service is provided by NW Natural. Wastewater treatment services are provided by the Metropolitan Wastewater Management Commission, a partnership between the Cities of Eugene and Springfield and Lane County. Three hospitals serve the Eugene-Springfield area. Sacred Heart Medical Center University District is the only one within Eugene city limits. McKenzie-Willamette Medical Center and Sacred Heart Medical Center at RiverBend are in Springfield. Oregon Medical Group, a primary care based multi-specialty group, operates several clinics in Eugene, as does PeaceHealth Medical Group. White Bird Clinic provides a broad range of health and human services, including low-cost clinics. The Volunteers in Medicine & Occupy Medical clinics provide free medical and mental care to low-income adults without health insurance. Eugene is one of the few municipalities in the US that does not fluoridate its water supply. Eugene has four sister cities:
https://en.wikipedia.org/wiki?curid=9623
Elizabeth Barrett Browning Elizabeth Barrett Browning (née Moulton-Barrett, ; 6 March 1806 – 29 June 1861) was an English poet of the Victorian era, popular in Britain and the United States during her lifetime. Born in County Durham, the eldest of 12 children, Elizabeth Barrett wrote poetry from the age of eleven. Her mother's collection of her poems forms one of the largest extant collections of juvenilia by any English writer. At 15 she became ill, suffering intense head and spinal pain for the rest of her life. Later in life she also developed lung problems, possibly tuberculosis. She took laudanum for the pain from an early age, which is likely to have contributed to her frail health. In the 1840s Elizabeth was introduced to literary society through her cousin, John Kenyon. Her first adult collection of poems was published in 1838 and she wrote prolifically between 1841 and 1844, producing poetry, translation and prose. She campaigned for the abolition of slavery and her work helped influence reform in the child labour legislation. Her prolific output made her a rival to Tennyson as a candidate for poet laureate on the death of Wordsworth. Elizabeth's volume "Poems" (1844) brought her great success, attracting the admiration of the writer Robert Browning. Their correspondence, courtship and marriage were carried out in secret, for fear of her father's disapproval. Following the wedding she was indeed disinherited by her father. In 1846, the couple moved to Italy, where she would live for the rest of her life. They had one son, Robert Wiedeman Barrett Browning, whom they called Pen. She died in Florence in 1861. A collection of her last poems was published by her husband shortly after her death. Elizabeth's work had a major influence on prominent writers of the day, including the American poets Edgar Allan Poe and Emily Dickinson. She is remembered for such poems as "How Do I Love Thee?" (Sonnet 43, 1845) and "Aurora Leigh "(1856). Some of Elizabeth Barrett's family had lived in Jamaica since 1655. Their wealth derived mainly from Edward Barrett (1734–1798), owner of in the estates of Cinnamon Hill, Cornwall, Cambridge and Oxford in northern Jamaica. Elizabeth's maternal grandfather owned sugar plantations, mills, glassworks and ships that traded between Jamaica and Newcastle. Biographer Julia Markus states the poet "believed that she had African blood through her grandfather Charles Moulton", but there is no evidence of this – although other branches of her family had African blood through relationships between plantation owners and slaves. What the family believed to be their genealogy in relation to Jamaica is unclear. The family wished to hand down their name, stipulating that Barrett should always be held as a surname. In some cases inheritance was given on condition that the name was used by the beneficiary; the English gentry and "squirearchy" had long encouraged this sort of name changing. Given this strong tradition, Elizabeth used "Elizabeth Barrett Moulton Barrett" on legal documents and before she was married often signed herself "Elizabeth Barrett Barrett" or "EBB" (initials which she was able to keep after her wedding). Elizabeth's father chose to raise his family in England while his business enterprises - of which a majority was owning slaves - remained in Jamaica. The fortune of Elizabeth's mother's line, the Graham Clarke family, also derived in part from slave labour, and was considerable. Elizabeth Barrett Moulton-Barrett was born on 6 March 1806, in Coxhoe Hall, between the villages of Coxhoe and Kelloe in County Durham, England. Her parents were Edward Barrett Moulton Barrett and Mary Graham Clarke; Elizabeth was the eldest of 12 children (eight boys and four girls). Eleven lived to adulthood; one daughter died at the age of three, when Elizabeth was eight. The children all had nicknames: Elizabeth was "Ba". She rode her pony, went for family walks and picnics, socialised with other county families, and participated in home theatrical productions. But unlike her siblings, she immersed herself in books as often as she could get away from the social rituals of her family. She was baptized in 1809 at Kelloe parish church, although she had already been baptised by a family friend in her first week of life. In 1809, the family moved to Hope End, a estate near the Malvern Hills in Ledbury, Herefordshire. Her father converted the Georgian house into stables and built a new mansion of opulent Turkish design, which his wife described as something from the "Arabian Nights Entertainments". The interior's brass balustrades, mahogany doors inlaid with mother-of-pearl, and finely carved fireplaces were eventually complemented by lavish landscaping: ponds, grottos, kiosks, an ice house, a hothouse, and a subterranean passage from house to gardens. Her time at Hope End would inspire her in later life to write her most ambitious work, "Aurora Leigh" (1856), which went through more than 20 editions by 1900, but none between 1905 and 1978. She was educated at home and tutored by Daniel McSwiney with her oldest brother. She began writing verses at the age of four. During the Hope End period, she was an intensely studious, precocious child. She claimed that at the age of six she was reading novels, at eight entranced by Pope's translations of Homer, studying Greek at ten, and at eleven writing her own Homeric epic, "". In 1820 Mr Barrett privately published "The Battle of Marathon", an epic-style poem, though all copies remained within the family. Her mother compiled the child's poetry into collections of "Poems by Elizabeth B. Barrett". Her father called her the "Poet Laureate of Hope End" and encouraged her work. The result is one of the largest collections of juvenilia of any English writer. Mary Russell Mitford described the young Elizabeth at this time, as having "a slight, delicate figure, with a shower of dark curls falling on each side of a most expressive face; large, tender eyes, richly fringed by dark eyelashes, and a smile like a sunbeam." At about this time, Elizabeth began to battle with illness, which the medical science of the time was unable to diagnose. All three sisters came down with the syndrome although it lasted only with Elizabeth. She had intense head and spinal pain with loss of mobility. Various biographies link this to a riding accident at the time (she fell while trying to dismount a horse), but there is no evidence to support the link. Sent to recover at the Gloucester spa, she was treated – in the absence of symptoms supporting another diagnosis – for a spinal problem. Though this illness continued for the rest of her life, it is believed to be unrelated to the lung disease which she developed in 1837. She began to take opiates for the pain, laudanum (an opium concoction) followed by morphine, then commonly prescribed. She would become dependent on them for much of her adulthood; the use from an early age may well have contributed to her frail health. Biographers such as Alethea Hayter have suggested this may also have contributed to the wild vividness of her imagination and the poetry that it produced. By 1821 she had read Mary Wollstonecraft's "A Vindication of the Rights of Woman" (1792), and become a passionate supporter of Wollstonecraft's ideas. The child's intellectual fascination with the classics and metaphysics was reflected in a religious intensity which she later described as "not the deep persuasion of the mild Christian but the wild visions of an enthusiast." The Barretts attended services at the nearest Dissenting chapel, and Edward was active in Bible and Missionary societies. Elizabeth's mother died in 1828, and is buried at St Michael's Church, Ledbury, next to her daughter Mary. Sarah Graham-Clarke, Elizabeth's aunt, helped to care for the children, and she had clashes with Elizabeth's strong will. In 1831 Elizabeth's grandmother, Elizabeth Moulton, died. Following lawsuits and the abolition of slavery Mr Barrett incurred great financial and investment losses that forced him to sell Hope End. Although the family was never poor, the place was seized and put up for sale to satisfy creditors. Always secret in his financial dealings, he would not discuss his situation and the family was haunted by the idea that they might have to move to Jamaica. Between 1833 and 1835, she was living, with her family, at Belle Vue in Sidmouth. The site has now been renamed Cedar Shade and redeveloped. A blue plaque at the entrance to the site attests to this. In 1838, some years after the sale of Hope End, the family settled at 50 Wimpole Street. During 1837–38 the poet was struck with illness again, with symptoms today suggesting tuberculous ulceration of the lungs. That same year, at her physician's insistence, she moved from London to Torquay, on the Devonshire coast. Two tragedies then struck. In February 1840 her brother Samuel died of a fever in Jamaica. Then her favourite brother Edward ("Bro") was drowned in a sailing accident in Torquay in July. This had a serious effect on her already fragile health. She felt guilty as her father had disapproved of Edward's trip to Torquay. She wrote to Mitford, "That was a very near escape from madness, absolute hopeless madness". The family returned to Wimpole Street in 1841. At Wimpole Street Barrett Browning spent most of her time in her upstairs room. Her health began to improve, though she saw few people other than her immediate family. One of those was Kenyon, a wealthy friend of the family and patron of the arts. She received comfort from a spaniel named Flush, a gift from Mary Mitford. (Virginia Woolf later fictionalised the life of the dog, making him the protagonist of her 1933 novel ""). Between 1841 and 1844 Barrett Browning was prolific in poetry, translation and prose. The poem "The Cry of the Children", published in 1842 in "Blackwoods", condemned child labour and helped bring about child-labour reforms by raising support for Lord Shaftesbury's Ten Hours Bill (1844). At about the same time, she contributed critical prose pieces to Richard Henry Horne's "A New Spirit of the Age". In 1844 she published two volumes of "Poems", which included "A Drama of Exile", "A Vision of Poets", and "Lady Geraldine's Courtship" and two substantial critical essays for 1842 issues of "The Athenaeum". "Since she was not burdened with any domestic duties expected of her sisters, Barrett Browning could now devote herself entirely to the life of the mind, cultivating an enormous correspondence, reading widely". Her prolific output made her a rival to Tennyson as a candidate for poet laureate in 1850 on the death of Wordsworth. A Royal Society of Arts blue plaque now commemorates Elizabeth at 50 Wimpole Street. Her 1844 volume "Poems" made her one of the most popular writers in the country, and inspired Robert Browning to write to her. He wrote, "I love your verses with all my heart, dear Miss Barrett," praising their "fresh strange music, the affluent language, the exquisite pathos and true new brave thought." Kenyon arranged for Browning to meet Elizabeth on 20 May 1845, in her rooms, and so began one of the most famous courtships in literature. Elizabeth had already produced a large amount of work, but Browning had a great influence on her subsequent writing, as did she on his: two of Barrett's most famous pieces were written after she met Browning, "Sonnets from the Portuguese" and "Aurora Leigh." Robert's "Men and Women" is also a product of that time. Some critics state that her activity was, in some ways, in decay before she met Browning: "Until her relationship with Robert Browning began in 1845, Barrett's willingness to engage in public discourse about social issues and about aesthetic issues in poetry, which had been so strong in her youth, gradually diminished, as did her physical health. As an intellectual presence and a physical being, she was becoming a shadow of herself." The courtship and marriage between Robert Browning and Elizabeth were carried out secretly, as she knew her father would disapprove. After a private marriage at St Marylebone Parish Church, they honeymooned in Paris before moving to Italy, in September 1846, which became their home almost continuously until her death. Elizabeth's loyal nurse, Wilson, who witnessed the marriage, accompanied the couple to Italy. Mr Barrett disinherited Elizabeth, as he did each of his children who married. Elizabeth had foreseen her father's anger but had not anticipated her brothers' rejection. As Elizabeth had some money of her own, the couple were reasonably comfortable in Italy. The Brownings were well respected, and even famous. Elizabeth grew stronger and in 1849, at the age of 43, between four miscarriages, she gave birth to a son, Robert Wiedeman Barrett Browning, whom they called Pen. Their son later married, but had no legitimate children. At her husband's insistence, Elizabeth's second edition of "Poems" included her love sonnets; as a result, her popularity increased (as well as critical regard), and her artistic position was confirmed. The couple came to know a wide circle of artists and writers including William Makepeace Thackeray, sculptor Harriet Hosmer (who, she wrote, seemed to be the "perfectly emancipated female") and Harriet Beecher Stowe. In 1849 she met Margaret Fuller, and the female French novelist George Sand in 1852, whom she had long admired. Among her intimate friends in Florence was the writer Isa Blagden, whom she encouraged to write novels. They met Alfred Tennyson in Paris, and John Forster, Samuel Rogers and the Carlyles in London, later befriending Charles Kingsley and John Ruskin. After the death of an old friend, G. B. Hunter, and then of her father, Barrett Browning's health started to deteriorate. The Brownings moved from Florence to Siena, residing at the "Villa Alberti". Engrossed in Italian politics, she issued a small volume of political poems titled "Poems before Congress" (1860) "most of which were written to express her sympathy with the Italian cause after the outbreak of fighting in 1859". They caused a furore in England, and the conservative magazines "Blackwood's" and the "Saturday Review" labelled her a fanatic. She dedicated this book to her husband. Her last work was "A Musical Instrument", published posthumously. Barrett Browning's sister Henrietta died in November 1860. The couple spent the winter of 1860–61 in Rome where Barrett Browning's health further deteriorated and they returned to Florence in early June 1861. She became gradually weaker, using morphine to ease her pain. She died on 29 June 1861 in her husband's arms. Browning said that she died "smilingly, happily, and with a face like a girl's... Her last word was... "Beautiful". She was buried in the Protestant English Cemetery of Florence. "On Monday July 1 the shops in the area around Casa Guidi were closed, while Elizabeth was mourned with unusual demonstrations." The nature of her illness is still unclear. Some modern scientists speculate her illness may have been hypokalemic periodic paralysis, a genetic disorder that causes weakness and many of the other symptoms she described. Barrett Browning's first known poem was written at the age of six or eight, "On the Cruelty of Forcement to Man". The manuscript, which protests against impressment, is currently in the Berg Collection of the New York Public Library; the exact date is controversial because the "2" in the date 1812 is written over something else that is scratched out. Her first independent publication was "Stanzas Excited by Reflections on the Present State of Greece" in "The New Monthly Magazine" of May 1821; followed two months later by "Thoughts Awakened by Contemplating a Piece of the Palm which Grows on the Summit of the Acropolis at Athens". Her first collection of poems, "An Essay on Mind, with Other Poems," was published in 1826 and reflected her passion for Byron and Greek politics. Its publication drew the attention of a blind scholar of the Greek language, Hugh Stuart Boyd, and of another Greek scholar, Uvedale Price, with whom she maintained sustained correspondence. Among other neighbours was Mrs James Martin from Colwall, with whom she also corresponded throughout her life. Later, at Boyd's suggestion, she translated Aeschylus' "Prometheus Bound" (published in 1833; retranslated in 1850). During their friendship Barrett studied Greek literature, including Homer, Pindar and Aristophanes. Elizabeth opposed slavery and published two poems highlighting the barbarity of slavers and her support for the abolitionist cause: "The Runaway Slave at Pilgrim's Point"; and "A Curse for a Nation". In "Runaway" she describes a slave woman who is whipped, raped, and made pregnant as she curses the slavers. Elizabeth declared herself glad that the slaves were "virtually free" when the Emancipation Act abolishing slavery in British colonies was passed in 1833, despite the fact that her father believed that Abolitionism would ruin his business. The date of publication of these poems is in dispute, but her position on slavery in the poems is clear and may have led to a rift between Elizabeth and her father. She wrote to John Ruskin in 1855 "I belong to a family of West Indian slaveholders, and if I believed in curses, I should be afraid". After the Jamaican slave uprising of 1831–32, her father and uncle continued to treat the slaves humanely. In London, John Kenyon, a distant cousin, introduced Elizabeth to literary figures including William Wordsworth, Mary Russell Mitford, Samuel Taylor Coleridge, Alfred Tennyson and Thomas Carlyle. Elizabeth continued to write, contributing "The Romaunt of Margaret", "The Romaunt of the Page", "The Poet's Vow" and other pieces to various periodicals. She corresponded with other writers, including Mary Russell Mitford, who would become a close friend and who would support Elizabeth's literary ambitions. In 1838 "The Seraphim and Other Poems" appeared, the first volume of Elizabeth's mature poetry to appear under her own name. "Sonnets from the Portuguese" was published in 1850. There is debate about the origin of the title. Some say it refers to the series of sonnets of the 16th-century Portuguese poet Luís de Camões. However, "my little Portuguese" was a pet name that Browning had adopted for Elizabeth and this may have some connection. The verse-novel "Aurora Leigh," her most ambitious and perhaps the most popular of her longer poems, appeared in 1856. It is the story of a female writer making her way in life, balancing work and love, and based on Elizabeth's own experiences. "Aurora Leigh" was an important influence on Susan B. Anthony's thinking about the traditional roles of women, with regard to marriage versus independent individuality. The "North American Review" praised Elizabeth's poem: "Mrs. Browning's poems are, in all respects, the utterance of a woman — of a woman of great learning, rich experience, and powerful genius, uniting to her woman's nature the strength which is sometimes thought peculiar to a man." Much of Barrett Browning's work carries a religious theme. She had read and studied such works as Milton's "Paradise Lost" and Dante's "Inferno". She says in her writing, "We want the sense of the saturation of Christ's blood upon the souls of our poets, that it may cry through them in answer to the ceaseless wail of the Sphinx of our humanity, expounding agony into renovation. Something of this has been perceived in art when its glory was at the fullest. Something of a yearning after this may be seen among the Greek Christian poets, something which would have been much with a stronger faculty". She believed that "Christ's religion is essentially poetry – poetry glorified". She explored the religious aspect in many of her poems, especially in her early work, such as the sonnets. She was interested in theological debate, had learned Hebrew and read the Hebrew Bible. Her seminal "Aurora Leigh", for example, features religious imagery and allusion to the apocalypse. The critic Cynthia Scheinberg notes that female characters in "Aurora Leigh" and her earlier work "The Virgin Mary to the Child Jesus" allude to the female character Miriam from the Hebrew Bible. These allusions to Miriam in both poems mirror the way in which Barrett Browning herself drew from Jewish history, while distancing herself from it, in order to maintain the public appearance of a Christian woman poet of the Victorian Age. In the correspondence Barrett Browning kept with the Reverend William Merry from 1843 to 1844 on predestination and salvation by works, she identifies herself as a Congregationalist: "I am not a Baptist — but a Congregational Christian, — in the holding of my private opinions." In 1892, Ledbury, Herefordshire, held a design competition to build an Institute in honour of Barrett Browning. Brightwen Binyon beat 44 other designs. It was based on the timber-framed Market House, which was opposite the site. It was completed in 1896. However, Nikolaus Pevsner was not impressed by its style. In 1938, it became a public library. It has been Grade II-listed since 2007. Barrett Browning was widely popular in the United Kingdom and the United States during her lifetime. Edgar Allan Poe was inspired by her poem "Lady Geraldine's Courtship" and specifically borrowed the poem's metre for his poem "The Raven". Poe had reviewed Barrett Browning's work in the January 1845 issue of the "Broadway Journal", saying that "her poetic inspiration is the highest – we can conceive of nothing more august. Her sense of Art is pure in itself." In return, she praised "The Raven", and Poe dedicated his 1845 collection "The Raven and Other Poems" to her, referring to her as "the noblest of her sex". Barrett Browning's poetry greatly influenced Emily Dickinson, who admired her as a woman of achievement. Her popularity in the United States and Britain was further advanced by her stands against social injustice, including slavery in the United States, injustice toward Italian citizens by foreign rulers, and child labour. Lilian Whiting published a biography of Barrett Browning (1899) which describes her as "the most philosophical poet" and depicts her life as "a Gospel of applied Christianity". To Whiting, the term "art for art's sake" did not apply to Barrett Browning's work, as each poem, distinctively purposeful, was borne of a more "honest vision". In this critical analysis, Whiting portrays Barrett Browning as a poet who uses knowledge of Classical literature with an "intuitive gift of spiritual divination". In "Elizabeth Barrett Browning", Angela Leighton suggests that the portrayal of Barrett Browning as the "pious iconography of womanhood" has distracted us from her poetic achievements. Leighton cites the 1931 play by Rudolf Besier, "The Barretts of Wimpole Street", as evidence that 20th-century literary criticism of Barrett Browning's work has suffered more as a result of her popularity than poetic ineptitude. The play was popularized by actress Katharine Cornell, for whom it became a signature role. It was an enormous success, both artistically and commercially, and was revived several times and adapted twice into movies. Throughout the 20th century, literary criticism of Barrett Browning's poetry remained sparse until her poems were discovered by the women's movement. She once described herself as being inclined to reject several women's rights principles, suggesting in letters to Mary Russell Mitford and her husband that she believed that there was an inferiority of intellect in women. In "Aurora Leigh", however, she created a strong and independent woman who embraces both work and love. Leighton writes that because Elizabeth participates in the literary world, where voice and diction are dominated by perceived masculine superiority, she "is defined only in mysterious opposition to everything that distinguishes the male subject who writes..." A five-volume scholarly edition of her works was published in 2010, the first in over a century.
https://en.wikipedia.org/wiki?curid=9627
Enlil Enlil, later known as Elil, is an ancient Mesopotamian god associated with wind, air, earth, and storms. He is first attested as the chief deity of the Sumerian pantheon, but he was later worshipped by the Akkadians, Babylonians, Assyrians, and Hurrians. Enlil's primary center of worship was the Ekur temple in the city of Nippur, which was believed to have been built by Enlil himself and was regarded as the "mooring-rope" of heaven and earth. He is also sometimes referred to in Sumerian texts as Nunamnir. According to one Sumerian hymn, Enlil himself was so holy that not even the other gods could look upon him. Enlil rose to prominence during the twenty-fourth century BC with the rise of Nippur. His cult fell into decline after Nippur was sacked by the Elamites in 1230 BC and he was eventually supplanted as the chief god of the Mesopotamian pantheon by the Babylonian national god Marduk. The Babylonian god Bel was a syncretic deity of Enlil, Marduk, and the shepherd deity Dumuzid. Enlil plays a vital role in the Sumerian creation myth; he separates An (heaven) from Ki (earth), thus making the world habitable for humans. In the Sumerian flood myth, Enlil rewards Ziusudra with immortality for having survived the flood and, in the Babylonian flood myth, Enlil is the cause of the flood himself, having sent the flood to exterminate the human race, who made too much noise and prevented him from sleeping. The myth of "Enlil and Ninlil" is about Enlil's serial seduction of the goddess Ninlil in various guises, resulting in the conception of the moon-god Nanna and the Underworld deities Nergal, Ninazu, and Enbilulu. Enlil was regarded as the inventor of the mattock and the patron of agriculture. Enlil also features prominently in several myths involving his son Ninurta, including "Anzû and the Tablet of Destinies" and "Lugale". Enlil's name comes from ancient Sumerian EN (𒂗), meaning "lord" and LÍL (𒆤) meaning "wind". His name therefore literally translates as "Lord Wind". Enlil's name is not a genitive construction, indicating that Enlil was seen as the personification of the wind itself rather than merely the cause of wind. Enlil was the patron god of the Sumerian city-state of Nippur and his main center of worship was the Ekur temple located there. The name of the temple literally means "Mountain House" in ancient Sumerian. The Ekur was believed to have been built and established by Enlil himself. It was believed to be the "mooring-rope" of heaven and earth, meaning that it was seen as "a channel of communication between earth and heaven". A hymn written during the reign of Ur-Nammu, the founder of the Third Dynasty of Ur, describes the E-kur in great detail, stating that its gates were carved with scenes of Imdugud, a lesser deity sometimes shown as a giant bird, slaying a lion and an eagle snatching up a sinner. The Sumerians believed that the sole purpose of humanity's existence was to serve the gods. They thought that a god's statue was a physical embodiment of the god himself. As such, cult statues were given constant care and attention and a set of priests were assigned to tend to them. People worshipped Enlil by offering food and other human necessities to him. The food, which was ritually laid out before the god's cult statue in the form of a feast, was believed to be Enlil's daily meal, but, after the ritual, it would be distributed among his priests. These priests were also responsible for changing the cult statue's clothing. The Sumerians envisioned Enlil as a benevolent, fatherly deity, who watches over humanity and cares for their well-being. One Sumerian hymn describes Enlil as so glorious that even the other gods could not look upon him. The same hymn also states that, without Enlil, civilization could not exist. Enlil's epithets include titles such as "the Great Mountain" and "King of the Foreign Lands". Enlil is also sometimes described as a "raging storm", a "wild bull", and a "merchant". The Mesopotamians envisioned him as a creator, a father, a king, and the supreme lord of the universe. He was also known as "Nunamnir" and is referred to in at least one text as the "East Wind and North Wind". Kings regarded Enlil as a model ruler and sought to emulate his example. Enlil was said to be supremely just and intolerant towards evil. Rulers from all over Sumer would travel to Enlil's temple in Nippur to be legitimized. They would return Enlil's favor by devoting lands and precious objects to his temple as offerings. Nippur was the only Sumerian city-state that never built a palace; this was intended to symbolize the city's importance as the center of the cult of Enlil by showing that Enlil himself was the city's king. Even during the Babylonian Period, when Marduk had superseded Enlil as the supreme god, Babylonian kings still traveled to the holy city of Nippur to seek recognition of their right to rule. Enlil first rose to prominence during the twenty-fourth century BC, when the importance of the god An began to wane. During this time period, Enlil and An are frequently invoked together in inscriptions. Enlil remained the supreme god in Mesopotamia throughout the Amorite Period, with Amorite monarchs proclaiming Enlil as the source of their legitimacy. Enlil's importance began to wane after the Babylonian king Hammurabi conquered Sumer. The Babylonians worshipped Enlil under the name "Elil" and the Hurrians syncretized him with their own god Kumarbi. In one Hurrian ritual, Enlil and Apantu are invoked as "the father and mother of Išḫara". Enlil is also invoked alongside Ninlil as a member of "the mighty and firmly established gods". During the Kassite Period ( 1592 BC – 1155 BC), Nippur briefly managed to regain influence in the region and Enlil rose to prominence once again. From around 1300 BC onwards, Enlil was syncretized with the Assyrian national god Aššur, who was the most important deity in the Assyrian pantheon. Then, in 1230 BC, the Elamites attacked Nippur and the city fell into decline, taking the cult of Enlil along with it. Approximately one hundred years later, Enlil's role as the head of the pantheon was given to Marduk, the national god of the Babylonians. Enlil's importance in the pantheon significantly declined and he was sometimes assimilated as merely an aspect of Marduk. Nonetheless, his temples continued functioning throughout the Neo-Assyrian period (911 BC — 609 BC) and even the Babylonians saw Anu and Enlil as the ones who bestowed Marduk with his powers. During the first millennium BC, the Babylonians worshipped a deity under the title "Bel", meaning "lord", who was a syncretization of Enlil, Marduk, and the dying god Dumuzid. Bel held all the cultic titles of Enlil and his status in the Babylonian religion was largely the same. Eventually, Bel came to be seen as the god of order and destiny. Meanwhile, Aššur continued to be known as "the Assyrian Enlil" or "the Enlil of the gods". After the collapse of the Neo-Assyrian Empire, Enlil's statues were smashed and his temples were destroyed because he had become inextricably associated with the Assyrians, who many conquered peoples hated. Enlil continued to be venerated under the name of Marduk until around 141 BC, when the cult of Marduk fell into terminal decline, and was eventually largely forgotten. Enlil was not represented anthropomorphically in Mesopotamian iconography. Instead, he was represented by a horned cap, which consisted of up to seven superimposed pairs of ox-horns. Such crowns were an important symbol of divinity; gods had been shown wearing them ever since the third millennium BC. The horned cap remained consistent in form and meaning from the earliest days of Sumerian prehistory up until the time of the Persian conquest and beyond. The Sumerians had a complex numerological system, in which certain numbers were believed to hold special ritual significance. Within this system, Enlil was associated with the number fifty, which was considered sacred to him. Enlil was part of a triad of deities, which also included An and Enki. These three deities together were the embodiment of all the fixed stars in the night sky. An was identified with all the stars of the equatorial sky, Enlil with those of the northern sky, and Enki with those of the southern sky. The path of Enlil's celestial orbit was a continuous, symmetrical circle around the north celestial pole, but those of An and Enki were believed to intersect at various points. Enlil was associated with the constellation Boötes. The main source of information about the Sumerian creation myth is the prologue to the epic poem "Gilgamesh, Enkidu, and the Netherworld" (ETCSL 1.8.1.4), which briefly describes the process of creation: originally, there was only Nammu, the primeval sea. Then, Nammu gave birth to An, the sky, and Ki, the earth. An and Ki mated with each other, causing Ki to give birth to Enlil. Enlil separated An from Ki and carried off the earth as his domain, while An carried off the sky. "Enlil and Ninlil" (ETCSL 1.2.1) is a nearly complete 152-line Sumerian poem describing the affair between Enlil and the goddess Ninlil. First, Ninlil's mother Nunbarshegunu instructs Ninlil to go bathe in the river. Ninlil goes to the river, where Enlil seduces her and impregnates her with their son, the moon-god Nanna. Because of this, Enlil is banished to Kur, the Sumerian underworld. Ninlil follows Enlil to the underworld, where he impersonates the "man of the gate". Ninlil demands to know where Enlil has gone, but Enlil, still impersonating the gatekeeper, refuses to answer. He then seduces Ninlil and impregnates her with Nergal, the god of death. The same scenario repeats, only this time Enlil instead impersonates the "man of the river of the nether world, the man-devouring river"; once again, he seduces Ninlil and impregnates her with the god Ninazu. Finally, Enlil impersonates the "man of the boat"; once again, he seduces Ninlil and impregnates her with Enbilulu, the "inspector of the canals". The story of Enlil's courtship with Ninlil is primarily a genealogical myth invented to explain the origins of the moon-god Nanna, as well as the various gods of the Underworld, but it is also, to some extent, a coming-of-age story describing Enlil and Ninlil's emergence from adolescence into adulthood. The story also explains Ninlil's role as Enlil's consort; in the poem, Ninlil declares, "As Enlil is your master, so am I also your mistress!" The story is also historically significant because, if the current interpretation of it is correct, it is the oldest known myth in which a god changes shape. In the Sumerian version of the flood story (ETCSL 1.7.4), the causes of the flood are unclear because the portion of the tablet recording the beginning of the story has been destroyed. Somehow, a mortal known as Ziusudra manages to survive the flood, likely through the help of the god Enki. The tablet begins in the middle of the description of the flood. The flood lasts for seven days and seven nights before it subsides. Then, Utu, the god of the Sun, emerges. Ziusudra opens a window in the side of the boat and falls down prostrate before the god. Next, he sacrifices an ox and a sheep in honor of Utu. At this point, the text breaks off again. When it picks back up, Enlil and An are in the midst of declaring Ziusudra immortal as an honor for having managed to survive the flood. The remaining portion of the tablet after this point is destroyed. In the later Akkadian version of the flood story, recorded in the "Epic of Gilgamesh", Enlil actually causes the flood, seeking to annihilate every living thing on earth because the humans, who are vastly overpopulated, make too much noise and prevent him from sleeping. In this version of the story, the hero is Utnapishtim, who is warned ahead of time by Ea, the Babylonian equivalent of Enki, that the flood is coming. The flood lasts for seven days; when it ends, Ishtar, who had mourned the destruction of humanity, promises Utnapishtim that Enlil will never cause a flood again. When Enlil sees that Utnapishtim and his family have survived, he is outraged, but his son Ninurta speaks up in favor of humanity, arguing that, instead of causing floods, Enlil should simply ensure that humans never become overpopulated by reducing their numbers using wild animals and famines. Enlil goes into the boat; Utnapishtim and his wife bow before him. Enlil, now appeased, grants Utnapishtim immortality as a reward for his loyalty to the gods. A nearly complete 108-line poem from the Early Dynastic Period ( 2900 – 2350 BC) describes Enlil's invention of the mattock, a key agricultural pick, hoe, ax, or digging tool of the Sumerians. In the poem, Enlil conjures the mattock into existence and decrees its fate. The mattock is described as gloriously beautiful; it is made of pure gold and has a head carved from lapis lazuli. Enlil gives the tool over to the humans, who use it to build cities, subjugate their people, and pull up weeds. Enlil was believed to aid in the growth of plants. The Sumerian poem "Enlil Chooses the Farmer-God" (ETCSL 5.3.3) describes how Enlil, hoping "to establish abundance and prosperity", creates two gods Emesh and Enten, a shepherd and a farmer, respectively. The two gods argue and Emesh lays claim to Enten's position. They take the dispute before Enlil, who rules in favor of Enten; the two gods rejoice and reconcile. In the Sumerian poem "Lugale" (ETCSL 1.6.2), Enlil gives advice to his son, the god Ninurta, advising him on a strategy to slay the demon Asag. This advice is relayed to Ninurta by way of Sharur, his enchanted talking mace, which had been sent by Ninurta to the realm of the gods to seek counsel from Enlil directly. In the Old, Middle, and Late Babylonian myth of "Anzû and the Tablet of Destinies", the Anzû, a giant, monstrous bird, betrays Enlil and steals the Tablet of Destinies, a sacred clay tablet belonging to Enlil that grants him his authority, while Enlil is preparing for a bath. The rivers dry up and the gods are stripped of their powers. The gods send Adad, Gerra, and Shara to defeat the Anzû, but all of them fail. Finally, Ea proposes that the gods should send Ninurta, Enlil's son. Ninurta successfully defeats the Anzû and returns the Tablet of Destinies to his father. As a reward, Ninurta is a granted a prominent seat on the council of the gods. A badly damaged text from the Neo-Assyrian Period (911 — 612 BC) describes Marduk leading his army of Anunnaki into the sacred city of Nippur and causing a disturbance. The disturbance causes a flood, which forces the resident gods of Nippur under the leadership of Enlil to take shelter in the Eshumesha temple to Ninurta. Enlil is enraged at Marduk's transgression and orders the gods of Eshumesha to take Marduk and the other Anunnaki as prisoners. The Anunnaki are captured, but Marduk appoints his front-runner Mushteshirhablim to lead a revolt against the gods of Eshumesha and sends his messenger Neretagmil to alert Nabu, the god of literacy. When the Eshumesha gods hear Nabu speak, they come out of their temple to search for him. Marduk defeats the Eshumesha gods and takes 360 of them as prisoners of war, including Enlil himself. Enlil protests that the Eshumesha gods are innocent, so Marduk puts them on trial before the Anunnaki. The text ends with a warning from Damkianna (another name for Ninhursag) to the gods and to humanity, pleading them not to repeat the war between the Anunnaki and the gods of Eshumesha.
https://en.wikipedia.org/wiki?curid=9628
Ecology Ecology (from , "house", or "environment"; , "study of") is a branch of biology concerning interactions among organisms and their biophysical environment, which includes both biotic and abiotic components. Topics of interest include the biodiversity, distribution, biomass, and populations of organisms, as well as cooperation and competition within and between species. Ecosystems are dynamically interacting systems of organisms, the communities they make up, and the non-living components of their environment. Ecosystem processes, such as primary production, pedogenesis, nutrient cycling, and niche construction, regulate the flux of energy and matter through an environment. These processes are sustained by organisms with specific life history traits. Ecology is not synonymous with environmentalism, natural history, or environmental science. It overlaps with the closely related sciences of evolutionary biology, genetics, and ethology. An important focus for ecologists is to improve the understanding of how biodiversity affects ecological function. Ecologists seek to explain: Ecology has practical applications in conservation biology, wetland management, natural resource management (agroecology, agriculture, forestry, agroforestry, fisheries), city planning (urban ecology), community health, economics, basic and applied science, and human social interaction (human ecology). It is not treated as separate from humans. Organisms (including humans) and resources compose ecosystems which, in turn, maintain biophysical feedback mechanisms that moderate processes acting on living (biotic) and non-living (abiotic) components of the planet. Ecosystems sustain life-supporting functions and produce natural capital like biomass production (food, fuel, fiber, and medicine), the regulation of climate, global biogeochemical cycles, water filtration, soil formation, erosion control, flood protection, and many other natural features of scientific, historical, economic, or intrinsic value. The word "ecology" ("Ökologie") was coined in 1866 by the German scientist Ernst Haeckel. Ecological thought is derivative of established currents in philosophy, particularly from ethics and politics. Ancient Greek philosophers such as Hippocrates and Aristotle laid the foundations of ecology in their studies on natural history. Modern ecology became a much more rigorous science in the late 19th century. Evolutionary concepts relating to adaptation and natural selection became the cornerstones of modern ecological theory. The scope of ecology contains a wide array of interacting levels of organization spanning micro-level (e.g., cells) to a planetary scale (e.g., biosphere) phenomena. Ecosystems, for example, contain abiotic resources and interacting life forms (i.e., individual organisms that aggregate into populations which aggregate into distinct ecological communities). Ecosystems are dynamic, they do not always follow a linear successional path, but they are always changing, sometimes rapidly and sometimes so slowly that it can take thousands of years for ecological processes to bring about certain successional stages of a forest. An ecosystem's area can vary greatly, from tiny to vast. A single tree is of little consequence to the classification of a forest ecosystem, but critically relevant to organisms living in and on it. Several generations of an aphid population can exist over the lifespan of a single leaf. Each of those aphids, in turn, support diverse bacterial communities. The nature of connections in ecological communities cannot be explained by knowing the details of each species in isolation, because the emergent pattern is neither revealed nor predicted until the ecosystem is studied as an integrated whole. Some ecological principles, however, do exhibit collective properties where the sum of the components explain the properties of the whole, such as birth rates of a population being equal to the sum of individual births over a designated time frame. The main subdisciplines of ecology, population (or community) ecology and ecosystem ecology, exhibit a difference not only of scale, but also of two contrasting paradigms in the field. The former focus on organisms' distribution and abundance, while the later focus on materials and energy fluxes. The scale of ecological dynamics can operate like a closed system, such as aphids migrating on a single tree, while at the same time remain open with regard to broader scale influences, such as atmosphere or climate. Hence, ecologists classify ecosystems hierarchically by analyzing data collected from finer scale units, such as vegetation associations, climate, and soil types, and integrate this information to identify emergent patterns of uniform organization and processes that operate on local to regional, landscape, and chronological scales. To structure the study of ecology into a conceptually manageable framework, the biological world is organized into a nested hierarchy, ranging in scale from genes, to cells, to tissues, to organs, to organisms, to species, to populations, to communities, to ecosystems, to biomes, and up to the level of the biosphere. This framework forms a panarchy and exhibits non-linear behaviors; this means that "effect and cause are disproportionate, so that small changes to critical variables, such as the number of nitrogen fixers, can lead to disproportionate, perhaps irreversible, changes in the system properties." Biodiversity (an abbreviation of "biological diversity") describes the diversity of life from genes to ecosystems and spans every level of biological organization. The term has several interpretations, and there are many ways to index, measure, characterize, and represent its complex organization. Biodiversity includes species diversity, ecosystem diversity, and genetic diversity and scientists are interested in the way that this diversity affects the complex ecological processes operating at and among these respective levels. Biodiversity plays an important role in ecosystem services which by definition maintain and improve human quality of life. Conservation priorities and management techniques require different approaches and considerations to address the full ecological scope of biodiversity. Natural capital that supports populations is critical for maintaining ecosystem services and species migration (e.g., riverine fish runs and avian insect control) has been implicated as one mechanism by which those service losses are experienced. An understanding of biodiversity has practical applications for species and ecosystem-level conservation planners as they make management recommendations to consulting firms, governments, and industry. The habitat of a species describes the environment over which a species is known to occur and the type of community that is formed as a result. More specifically, "habitats can be defined as regions in environmental space that are composed of multiple dimensions, each representing a biotic or abiotic environmental variable; that is, any component or characteristic of the environment related directly (e.g. forage biomass and quality) or indirectly (e.g. elevation) to the use of a location by the animal." For example, a habitat might be an aquatic or terrestrial environment that can be further categorized as a montane or alpine ecosystem. Habitat shifts provide important evidence of competition in nature where one population changes relative to the habitats that most other individuals of the species occupy. For example, one population of a species of tropical lizard ("Tropidurus hispidus") has a flattened body relative to the main populations that live in open savanna. The population that lives in an isolated rock outcrop hides in crevasses where its flattened body offers a selective advantage. Habitat shifts also occur in the developmental life history of amphibians, and in insects that transition from aquatic to terrestrial habitats. Biotope and habitat are sometimes used interchangeably, but the former applies to a community's environment, whereas the latter applies to a species' environment. Definitions of the niche date back to 1917, but G. Evelyn Hutchinson made conceptual advances in 1957 by introducing a widely adopted definition: "the set of biotic and abiotic conditions in which a species is able to persist and maintain stable population sizes." The ecological niche is a central concept in the ecology of organisms and is sub-divided into the "fundamental" and the "realized" niche. The fundamental niche is the set of environmental conditions under which a species is able to persist. The realized niche is the set of environmental plus ecological conditions under which a species persists. The Hutchinsonian niche is defined more technically as a "Euclidean hyperspace whose "dimensions" are defined as environmental variables and whose "size" is a function of the number of values that the environmental values may assume for which an organism has "positive fitness"." Biogeographical patterns and range distributions are explained or predicted through knowledge of a species' traits and niche requirements. Species have functional traits that are uniquely adapted to the ecological niche. A trait is a measurable property, phenotype, or characteristic of an organism that may influence its survival. Genes play an important role in the interplay of development and environmental expression of traits. Resident species evolve traits that are fitted to the selection pressures of their local environment. This tends to afford them a competitive advantage and discourages similarly adapted species from having an overlapping geographic range. The competitive exclusion principle states that two species cannot coexist indefinitely by living off the same limiting resource; one will always out-compete the other. When similarly adapted species overlap geographically, closer inspection reveals subtle ecological differences in their habitat or dietary requirements. Some models and empirical studies, however, suggest that disturbances can stabilize the co-evolution and shared niche occupancy of similar species inhabiting species-rich communities. The habitat plus the niche is called the ecotope, which is defined as the full range of environmental and biological variables affecting an entire species. Organisms are subject to environmental pressures, but they also modify their habitats. The regulatory feedback between organisms and their environment can affect conditions from local (e.g., a beaver pond) to global scales, over time and even after death, such as decaying logs or silica skeleton deposits from marine organisms. The process and concept of ecosystem engineering is related to niche construction, but the former relates only to the physical modifications of the habitat whereas the latter also considers the evolutionary implications of physical changes to the environment and the feedback this causes on the process of natural selection. Ecosystem engineers are defined as: "organisms that directly or indirectly modulate the availability of resources to other species, by causing physical state changes in biotic or abiotic materials. In so doing they modify, maintain and create habitats." The ecosystem engineering concept has stimulated a new appreciation for the influence that organisms have on the ecosystem and evolutionary process. The term "niche construction" is more often used in reference to the under-appreciated feedback mechanisms of natural selection imparting forces on the abiotic niche. An example of natural selection through ecosystem engineering occurs in the nests of social insects, including ants, bees, wasps, and termites. There is an emergent homeostasis or homeorhesis in the structure of the nest that regulates, maintains and defends the physiology of the entire colony. Termite mounds, for example, maintain a constant internal temperature through the design of air-conditioning chimneys. The structure of the nests themselves are subject to the forces of natural selection. Moreover, a nest can survive over successive generations, so that progeny inherit both genetic material and a legacy niche that was constructed before their time. Biomes are larger units of organization that categorize regions of the Earth's ecosystems, mainly according to the structure and composition of vegetation. There are different methods to define the continental boundaries of biomes dominated by different functional types of vegetative communities that are limited in distribution by climate, precipitation, weather and other environmental variables. Biomes include tropical rainforest, temperate broadleaf and mixed forest, temperate deciduous forest, taiga, tundra, hot desert, and polar desert. Other researchers have recently categorized other biomes, such as the human and oceanic microbiomes. To a microbe, the human body is a habitat and a landscape. Microbiomes were discovered largely through advances in molecular genetics, which have revealed a hidden richness of microbial diversity on the planet. The oceanic microbiome plays a significant role in the ecological biogeochemistry of the planet's oceans. The largest scale of ecological organization is the biosphere: the total sum of ecosystems on the planet. Ecological relationships regulate the flux of energy, nutrients, and climate all the way up to the planetary scale. For example, the dynamic history of the planetary atmosphere's CO2 and O2 composition has been affected by the biogenic flux of gases coming from respiration and photosynthesis, with levels fluctuating over time in relation to the ecology and evolution of plants and animals. Ecological theory has also been used to explain self-emergent regulatory phenomena at the planetary scale: for example, the Gaia hypothesis is an example of holism applied in ecological theory. The Gaia hypothesis states that there is an emergent feedback loop generated by the metabolism of living organisms that maintains the core temperature of the Earth and atmospheric conditions within a narrow self-regulating range of tolerance. Population ecology studies the dynamics of species populations and how these populations interact with the wider environment. A population consists of individuals of the same species that live, interact, and migrate through the same niche and habitat. A primary law of population ecology is the Malthusian growth model which states, "a population will grow (or decline) exponentially as long as the environment experienced by all individuals in the population remains constant." Simplified population models usually start with four variables: death, birth, immigration, and emigration. An example of an introductory population model describes a closed population, such as on an island, where immigration and emigration does not take place. Hypotheses are evaluated with reference to a null hypothesis which states that random processes create the observed data. In these island models, the rate of population change is described by: where "N" is the total number of individuals in the population, "b" and "d" are the per capita rates of birth and death respectively, and "r" is the per capita rate of population change. Using these modelling techniques, Malthus' population principle of growth was later transformed into a model known as the logistic equation: where "N" is the number of individuals measured as biomass density, "a" is the maximum per-capita rate of change, and "K" is the carrying capacity of the population. The formula states that the rate of change in population size ("dN/dT") is equal to growth ("aN") that is limited by carrying capacity (1 – "N"/"K"). Population ecology builds upon these introductory models to further understand demographic processes in real study populations. Commonly used types of data include life history, fecundity, and survivorship, and these are analysed using mathematical techniques such as matrix algebra. The information is used for managing wildlife stocks and setting harvest quotas. In cases where basic models are insufficient, ecologists may adopt different kinds of statistical methods, such as the Akaike information criterion, or use models that can become mathematically complex as "several competing hypotheses are simultaneously confronted with the data." The concept of metapopulations was defined in 1969 as "a population of populations which go extinct locally and recolonize". Metapopulation ecology is another statistical approach that is often used in conservation research. Metapopulation models simplify the landscape into patches of varying levels of quality, and metapopulations are linked by the migratory behaviours of organisms. Animal migration is set apart from other kinds of movement; because, it involves the seasonal departure and return of individuals from a habitat. Migration is also a population-level phenomenon, as with the migration routes followed by plants as they occupied northern post-glacial environments. Plant ecologists use pollen records that accumulate and stratify in wetlands to reconstruct the timing of plant migration and dispersal relative to historic and contemporary climates. These migration routes involved an expansion of the range as plant populations expanded from one area to another. There is a larger taxonomy of movement, such as commuting, foraging, territorial behaviour, stasis, and ranging. Dispersal is usually distinguished from migration; because, it involves the one way permanent movement of individuals from their birth population into another population. In metapopulation terminology, migrating individuals are classed as emigrants (when they leave a region) or immigrants (when they enter a region), and sites are classed either as sources or sinks. A site is a generic term that refers to places where ecologists sample populations, such as ponds or defined sampling areas in a forest. Source patches are productive sites that generate a seasonal supply of juveniles that migrate to other patch locations. Sink patches are unproductive sites that only receive migrants; the population at the site will disappear unless rescued by an adjacent source patch or environmental conditions become more favourable. Metapopulation models examine patch dynamics over time to answer potential questions about spatial and demographic ecology. The ecology of metapopulations is a dynamic process of extinction and colonization. Small patches of lower quality (i.e., sinks) are maintained or rescued by a seasonal influx of new immigrants. A dynamic metapopulation structure evolves from year to year, where some patches are sinks in dry years and are sources when conditions are more favourable. Ecologists use a mixture of computer models and field studies to explain metapopulation structure. Community ecology is the study of the interactions among a collections of species that inhabit the same geographic area. Community ecologists study the determinants of patterns and processes for two or more interacting species. Research in community ecology might measure species diversity in grasslands in relation to soil fertility. It might also include the analysis of predator-prey dynamics, competition among similar plant species, or mutualistic interactions between crabs and corals. Ecosystems may be habitats within biomes that form an integrated whole and a dynamically responsive system having both physical and biological complexes. Ecosystem ecology is the science of determining the fluxes of materials (e.g. carbon, phosphorus) between different pools (e.g., tree biomass, soil organic material). Ecosystem ecologist attempt to determine the underlying causes of these fluxes. Research in ecosystem ecology might measure primary production (g C/m^2) in a wetland in relation to decomposition and consumption rates (g C/m^2/y). This requires an understanding of the community connections between plants (i.e., primary producers) and the decomposers (e.g., fungi and bacteria), The underlying concept of ecosystem can be traced back to 1864 in the published work of George Perkins Marsh ("Man and Nature"). Within an ecosystem, organisms are linked to the physical and biological components of their environment to which they are adapted. Ecosystems are complex adaptive systems where the interaction of life processes form self-organizing patterns across different scales of time and space. Ecosystems are broadly categorized as terrestrial, freshwater, atmospheric, or marine. Differences stem from the nature of the unique physical environments that shapes the biodiversity within each. A more recent addition to ecosystem ecology are technoecosystems, which are affected by or primarily the result of human activity. A food web is the archetypal ecological network. Plants capture solar energy and use it to synthesize simple sugars during photosynthesis. As plants grow, they accumulate nutrients and are eaten by grazing herbivores, and the energy is transferred through a chain of organisms by consumption. The simplified linear feeding pathways that move from a basal trophic species to a top consumer is called the food chain. The larger interlocking pattern of food chains in an ecological community creates a complex food web. Food webs are a type of concept map or a heuristic device that is used to illustrate and study pathways of energy and material flows. Food webs are often limited relative to the real world. Complete empirical measurements are generally restricted to a specific habitat, such as a cave or a pond, and principles gleaned from food web microcosm studies are extrapolated to larger systems. Feeding relations require extensive investigations into the gut contents of organisms, which can be difficult to decipher, or stable isotopes can be used to trace the flow of nutrient diets and energy through a food web. Despite these limitations, food webs remain a valuable tool in understanding community ecosystems. Food webs exhibit principles of ecological emergence through the nature of trophic relationships: some species have many weak feeding links (e.g., omnivores) while some are more specialized with fewer stronger feeding links (e.g., primary predators). Theoretical and empirical studies identify non-random emergent patterns of few strong and many weak linkages that explain how ecological communities remain stable over time. Food webs are composed of subgroups where members in a community are linked by strong interactions, and the weak interactions occur between these subgroups. This increases food web stability. Step by step lines or relations are drawn until a web of life is illustrated. A trophic level (from Greek "troph", τροφή, trophē, meaning "food" or "feeding") is "a group of organisms acquiring a considerable majority of its energy from the lower adjacent level (according to ecological pyramids) nearer the abiotic source." Links in food webs primarily connect feeding relations or trophism among species. Biodiversity within ecosystems can be organized into trophic pyramids, in which the vertical dimension represents feeding relations that become further removed from the base of the food chain up toward top predators, and the horizontal dimension represents the abundance or biomass at each level. When the relative abundance or biomass of each species is sorted into its respective trophic level, they naturally sort into a 'pyramid of numbers'. Species are broadly categorized as autotrophs (or primary producers), heterotrophs (or consumers), and Detritivores (or decomposers). Autotrophs are organisms that produce their own food (production is greater than respiration) by photosynthesis or chemosynthesis. Heterotrophs are organisms that must feed on others for nourishment and energy (respiration exceeds production). Heterotrophs can be further sub-divided into different functional groups, including primary consumers (strict herbivores), secondary consumers (carnivorous predators that feed exclusively on herbivores), and tertiary consumers (predators that feed on a mix of herbivores and predators). Omnivores do not fit neatly into a functional category because they eat both plant and animal tissues. It has been suggested that omnivores have a greater functional influence as predators, because compared to herbivores, they are relatively inefficient at grazing. Trophic levels are part of the holistic or complex systems view of ecosystems. Each trophic level contains unrelated species that are grouped together because they share common ecological functions, giving a macroscopic view of the system. While the notion of trophic levels provides insight into energy flow and top-down control within food webs, it is troubled by the prevalence of omnivory in real ecosystems. This has led some ecologists to "reiterate that the notion that species clearly aggregate into discrete, homogeneous trophic levels is fiction." Nonetheless, recent studies have shown that real trophic levels do exist, but "above the herbivore trophic level, food webs are better characterized as a tangled web of omnivores." A keystone species is a species that is connected to a disproportionately large number of other species in the food-web. Keystone species have lower levels of biomass in the trophic pyramid relative to the importance of their role. The many connections that a keystone species holds means that it maintains the organization and structure of entire communities. The loss of a keystone species results in a range of dramatic cascading effects that alters trophic dynamics, other food web connections, and can cause the extinction of other species. Sea otters ("Enhydra lutris") are commonly cited as an example of a keystone species; because, they limit the density of sea urchins that feed on kelp. If sea otters are removed from the system, the urchins graze until the kelp beds disappear, and this has a dramatic effect on community structure. Hunting of sea otters, for example, is thought to have led indirectly to the extinction of the Steller's sea cow ("Hydrodamalis gigas"). While the keystone species concept has been used extensively as a conservation tool, it has been criticized for being poorly defined from an operational stance. It is difficult to experimentally determine what species may hold a keystone role in each ecosystem. Furthermore, food web theory suggests that keystone species may not be common, so it is unclear how generally the keystone species model can be applied. Complexity is understood as a large computational effort needed to piece together numerous interacting parts exceeding the iterative memory capacity of the human mind. Global patterns of biological diversity are complex. This biocomplexity stems from the interplay among ecological processes that operate and influence patterns at different scales that grade into each other, such as transitional areas or ecotones spanning landscapes. Complexity stems from the interplay among levels of biological organization as energy, and matter is integrated into larger units that superimpose onto the smaller parts. "What were wholes on one level become parts on a higher one." Small scale patterns do not necessarily explain large scale phenomena, otherwise captured in the expression (coined by Aristotle) 'the sum is greater than the parts'. "Complexity in ecology is of at least six distinct types: spatial, temporal, structural, process, behavioral, and geometric." From these principles, ecologists have identified emergent and self-organizing phenomena that operate at different environmental scales of influence, ranging from molecular to planetary, and these require different explanations at each integrative level. Ecological complexity relates to the dynamic resilience of ecosystems that transition to multiple shifting steady-states directed by random fluctuations of history. Long-term ecological studies provide important track records to better understand the complexity and resilience of ecosystems over longer temporal and broader spatial scales. These studies are managed by the International Long Term Ecological Network (LTER). The longest experiment in existence is the Park Grass Experiment, which was initiated in 1856. Another example is the Hubbard Brook study, which has been in operation since 1960. Holism remains a critical part of the theoretical foundation in contemporary ecological studies. Holism addresses the biological organization of life that self-organizes into layers of emergent whole systems that function according to non-reducible properties. This means that higher order patterns of a whole functional system, such as an ecosystem, cannot be predicted or understood by a simple summation of the parts. "New properties emerge because the components interact, not because the basic nature of the components is changed." Ecological studies are necessarily holistic as opposed to reductionistic. Holism has three scientific meanings or uses that identify with ecology: 1) the mechanistic complexity of ecosystems, 2) the practical description of patterns in quantitative reductionist terms where correlations may be identified but nothing is understood about the causal relations without reference to the whole system, which leads to 3) a metaphysical hierarchy whereby the causal relations of larger systems are understood without reference to the smaller parts. Scientific holism differs from mysticism that has appropriated the same term. An example of metaphysical holism is identified in the trend of increased exterior thickness in shells of different species. The reason for a thickness increase can be understood through reference to principles of natural selection via predation without need to reference or understand the biomolecular properties of the exterior shells. Ecology and evolutionary biology are considered sister disciplines of the life sciences. Natural selection, life history, development, adaptation, populations, and inheritance are examples of concepts that thread equally into ecological and evolutionary theory. Morphological, behavioural, and genetic traits, for example, can be mapped onto evolutionary trees to study the historical development of a species in relation to their functions and roles in different ecological circumstances. In this framework, the analytical tools of ecologists and evolutionists overlap as they organize, classify, and investigate life through common systematic principles, such as phylogenetics or the Linnaean system of taxonomy. The two disciplines often appear together, such as in the title of the journal "Trends in Ecology and Evolution". There is no sharp boundary separating ecology from evolution, and they differ more in their areas of applied focus. Both disciplines discover and explain emergent and unique properties and processes operating across different spatial or temporal scales of organization. While the boundary between ecology and evolution is not always clear, ecologists study the abiotic and biotic factors that influence evolutionary processes, and evolution can be rapid, occurring on ecological timescales as short as one generation. All organisms can exhibit behaviours. Even plants express complex behaviour, including memory and communication. Behavioural ecology is the study of an organism's behaviour in its environment and its ecological and evolutionary implications. Ethology is the study of observable movement or behaviour in animals. This could include investigations of motile sperm of plants, mobile phytoplankton, zooplankton swimming toward the female egg, the cultivation of fungi by weevils, the mating dance of a salamander, or social gatherings of amoeba. Adaptation is the central unifying concept in behavioural ecology. Behaviours can be recorded as traits and inherited in much the same way that eye and hair colour can. Behaviours can evolve by means of natural selection as adaptive traits conferring functional utilities that increases reproductive fitness. Predator-prey interactions are an introductory concept into food-web studies as well as behavioural ecology. Prey species can exhibit different kinds of behavioural adaptations to predators, such as avoid, flee, or defend. Many prey species are faced with multiple predators that differ in the degree of danger posed. To be adapted to their environment and face predatory threats, organisms must balance their energy budgets as they invest in different aspects of their life history, such as growth, feeding, mating, socializing, or modifying their habitat. Hypotheses posited in behavioural ecology are generally based on adaptive principles of conservation, optimization, or efficiency. For example, "[t]he threat-sensitive predator avoidance hypothesis predicts that prey should assess the degree of threat posed by different predators and match their behaviour according to current levels of risk" or "[t]he optimal flight initiation distance occurs where expected postencounter fitness is maximized, which depends on the prey's initial fitness, benefits obtainable by not fleeing, energetic escape costs, and expected fitness loss due to predation risk." Elaborate sexual displays and posturing are encountered in the behavioural ecology of animals. The birds-of-paradise, for example, sing and display elaborate ornaments during courtship. These displays serve a dual purpose of signalling healthy or well-adapted individuals and desirable genes. The displays are driven by sexual selection as an advertisement of quality of traits among suitors. Cognitive ecology integrates theory and observations from evolutionary ecology and neurobiology, primarily cognitive science, in order to understand the effect that animal interaction with their habitat has on their cognitive systems and how those systems restrict behavior within an ecological and evolutionary framework. "Until recently, however, cognitive scientists have not paid sufficient attention to the fundamental fact that cognitive traits evolved under particular natural settings. With consideration of the selection pressure on cognition, cognitive ecology can contribute intellectual coherence to the multidisciplinary study of cognition." As a study involving the 'coupling' or interactions between organism and environment, cognitive ecology is closely related to enactivism, a field based upon the view that "...we must see the organism and environment as bound together in reciprocal specification and selection...". Social ecological behaviours are notable in the social insects, slime moulds, social spiders, human society, and naked mole-rats where eusocialism has evolved. Social behaviours include reciprocally beneficial behaviours among kin and nest mates and evolve from kin and group selection. Kin selection explains altruism through genetic relationships, whereby an altruistic behaviour leading to death is rewarded by the survival of genetic copies distributed among surviving relatives. The social insects, including ants, bees, and wasps are most famously studied for this type of relationship because the male drones are clones that share the same genetic make-up as every other male in the colony. In contrast, group selectionists find examples of altruism among non-genetic relatives and explain this through selection acting on the group; whereby, it becomes selectively advantageous for groups if their members express altruistic behaviours to one another. Groups with predominantly altruistic members survive better than groups with predominantly selfish members. Ecological interactions can be classified broadly into a host and an associate relationship. A host is any entity that harbours another that is called the associate. Relationships within a species that are mutually or reciprocally beneficial are called mutualisms. Examples of mutualism include fungus-growing ants employing agricultural symbiosis, bacteria living in the guts of insects and other organisms, the fig wasp and yucca moth pollination complex, lichens with fungi and photosynthetic algae, and corals with photosynthetic algae. If there is a physical connection between host and associate, the relationship is called symbiosis. Approximately 60% of all plants, for example, have a symbiotic relationship with arbuscular mycorrhizal fungi living in their roots forming an exchange network of carbohydrates for mineral nutrients. Indirect mutualisms occur where the organisms live apart. For example, trees living in the equatorial regions of the planet supply oxygen into the atmosphere that sustains species living in distant polar regions of the planet. This relationship is called commensalism; because, many others receive the benefits of clean air at no cost or harm to trees supplying the oxygen. If the associate benefits while the host suffers, the relationship is called parasitism. Although parasites impose a cost to their host (e.g., via damage to their reproductive organs or propagules, denying the services of a beneficial partner), their net effect on host fitness is not necessarily negative and, thus, becomes difficult to forecast. Co-evolution is also driven by competition among species or among members of the same species under the banner of reciprocal antagonism, such as grasses competing for growth space. The Red Queen Hypothesis, for example, posits that parasites track down and specialize on the locally common genetic defense systems of its host that drives the evolution of sexual reproduction to diversify the genetic constituency of populations responding to the antagonistic pressure. Biogeography (an amalgamation of "biology" and "geography") is the comparative study of the geographic distribution of organisms and the corresponding evolution of their traits in space and time. The "Journal of Biogeography" was established in 1974. Biogeography and ecology share many of their disciplinary roots. For example, the theory of island biogeography, published by the Robert MacArthur and Edward O. Wilson in 1967 is considered one of the fundamentals of ecological theory. Biogeography has a long history in the natural sciences concerning the spatial distribution of plants and animals. Ecology and evolution provide the explanatory context for biogeographical studies. Biogeographical patterns result from ecological processes that influence range distributions, such as migration and dispersal. and from historical processes that split populations or species into different areas. The biogeographic processes that result in the natural splitting of species explains much of the modern distribution of the Earth's biota. The splitting of lineages in a species is called vicariance biogeography and it is a sub-discipline of biogeography. There are also practical applications in the field of biogeography concerning ecological systems and processes. For example, the range and distribution of biodiversity and invasive species responding to climate change is a serious concern and active area of research in the context of global warming. A population ecology concept is r/K selection theory, one of the first predictive models in ecology used to explain life-history evolution. The premise behind the r/K selection model is that natural selection pressures change according to population density. For example, when an island is first colonized, density of individuals is low. The initial increase in population size is not limited by competition, leaving an abundance of available resources for rapid population growth. These early phases of population growth experience "density-independent" forces of natural selection, which is called "r"-selection. As the population becomes more crowded, it approaches the island's carrying capacity, thus forcing individuals to compete more heavily for fewer available resources. Under crowded conditions, the population experiences density-dependent forces of natural selection, called "K"-selection. In the "r/K"-selection model, the first variable "r" is the intrinsic rate of natural increase in population size and the second variable "K" is the carrying capacity of a population. Different species evolve different life-history strategies spanning a continuum between these two selective forces. An "r"-selected species is one that has high birth rates, low levels of parental investment, and high rates of mortality before individuals reach maturity. Evolution favours high rates of fecundity in "r"-selected species. Many kinds of insects and invasive species exhibit "r"-selected characteristics. In contrast, a "K"-selected species has low rates of fecundity, high levels of parental investment in the young, and low rates of mortality as individuals mature. Humans and elephants are examples of species exhibiting "K"-selected characteristics, including longevity and efficiency in the conversion of more resources into fewer offspring. The important relationship between ecology and genetic inheritance predates modern techniques for molecular analysis. Molecular ecological research became more feasible with the development of rapid and accessible genetic technologies, such as the polymerase chain reaction (PCR). The rise of molecular technologies and influx of research questions into this new ecological field resulted in the publication "Molecular Ecology" in 1992. Molecular ecology uses various analytical techniques to study genes in an evolutionary and ecological context. In 1994, John Avise also played a leading role in this area of science with the publication of his book, "Molecular Markers, Natural History and Evolution". Newer technologies opened a wave of genetic analysis into organisms once difficult to study from an ecological or evolutionary standpoint, such as bacteria, fungi, and nematodes. Molecular ecology engendered a new research paradigm for investigating ecological questions considered otherwise intractable. Molecular investigations revealed previously obscured details in the tiny intricacies of nature and improved resolution into probing questions about behavioural and biogeographical ecology. For example, molecular ecology revealed promiscuous sexual behaviour and multiple male partners in tree swallows previously thought to be socially monogamous. In a biogeographical context, the marriage between genetics, ecology, and evolution resulted in a new sub-discipline called phylogeography. Ecology is as much a biological science as it is a human science. Human ecology is an interdisciplinary investigation into the ecology of our species. "Human ecology may be defined: (1) from a bioecological standpoint as the study of man as the ecological dominant in plant and animal communities and systems; (2) from a bioecological standpoint as simply another animal affecting and being affected by his physical environment; and (3) as a human being, somehow different from animal life in general, interacting with physical and modified environments in a distinctive and creative way. A truly interdisciplinary human ecology will most likely address itself to all three." The term was formally introduced in 1921, but many sociologists, geographers, psychologists, and other disciplines were interested in human relations to natural systems centuries prior, especially in the late 19th century. The ecological complexities human beings are facing through the technological transformation of the planetary biome has brought on the Anthropocene. The unique set of circumstances has generated the need for a new unifying science called coupled human and natural systems that builds upon, but moves beyond the field of human ecology. Ecosystems tie into human societies through the critical and all encompassing life-supporting functions they sustain. In recognition of these functions and the incapability of traditional economic valuation methods to see the value in ecosystems, there has been a surge of interest in social-natural capital, which provides the means to put a value on the stock and use of information and materials stemming from ecosystem goods and services. Ecosystems produce, regulate, maintain, and supply services of critical necessity and beneficial to human health (cognitive and physiological), economies, and they even provide an information or reference function as a living library giving opportunities for science and cognitive development in children engaged in the complexity of the natural world. Ecosystems relate importantly to human ecology as they are the ultimate base foundation of global economics as every commodity, and the capacity for exchange ultimately stems from the ecosystems on Earth. Ecology is an employed science of restoration, repairing disturbed sites through human intervention, in natural resource management, and in environmental impact assessments. Edward O. Wilson predicted in 1992 that the 21st century "will be the era of restoration in ecology". Ecological science has boomed in the industrial investment of restoring ecosystems and their processes in abandoned sites after disturbance. Natural resource managers, in forestry, for example, employ ecologists to develop, adapt, and implement ecosystem based methods into the planning, operation, and restoration phases of land-use. Ecological science is used in the methods of sustainable harvesting, disease, and fire outbreak management, in fisheries stock management, for integrating land-use with protected areas and communities, and conservation in complex geo-political landscapes. The environment of ecosystems includes both physical parameters and biotic attributes. It is dynamically interlinked, and contains resources for organisms at any time throughout their life cycle. Like ecology, the term environment has different conceptual meanings and overlaps with the concept of nature. Environment "includes the physical world, the social world of human relations and the built world of human creation." The physical environment is external to the level of biological organization under investigation, including abiotic factors such as temperature, radiation, light, chemistry, climate and geology. The biotic environment includes genes, cells, organisms, members of the same species (conspecifics) and other species that share a habitat. The distinction between external and internal environments, however, is an abstraction parsing life and environment into units or facts that are inseparable in reality. There is an interpenetration of cause and effect between the environment and life. The laws of thermodynamics, for example, apply to ecology by means of its physical state. With an understanding of metabolic and thermodynamic principles, a complete accounting of energy and material flow can be traced through an ecosystem. In this way, the environmental and ecological relations are studied through reference to conceptually manageable and isolated material parts. After the effective environmental components are understood through reference to their causes; however, they conceptually link back together as an integrated whole, or "holocoenotic" system as it was once called. This is known as the dialectical approach to ecology. The dialectical approach examines the parts, but integrates the organism and the environment into a dynamic whole (or umwelt). Change in one ecological or environmental factor can concurrently affect the dynamic state of an entire ecosystem. Ecosystems are regularly confronted with natural environmental variations and disturbances over time and geographic space. A disturbance is any process that removes biomass from a community, such as a fire, flood, drought, or predation. Disturbances occur over vastly different ranges in terms of magnitudes as well as distances and time periods, and are both the cause and product of natural fluctuations in death rates, species assemblages, and biomass densities within an ecological community. These disturbances create places of renewal where new directions emerge from the patchwork of natural experimentation and opportunity. Ecological resilience is a cornerstone theory in ecosystem management. Biodiversity fuels the resilience of ecosystems acting as a kind of regenerative insurance. The Earth was formed approximately 4.5 billion years ago. As it cooled and a crust and oceans formed, its atmosphere transformed from being dominated by hydrogen to one composed mostly of methane and ammonia. Over the next billion years, the metabolic activity of life transformed the atmosphere into a mixture of carbon dioxide, nitrogen, and water vapor. These gases changed the way that light from the sun hit the Earth's surface and greenhouse effects trapped heat. There were untapped sources of free energy within the mixture of reducing and oxidizing gasses that set the stage for primitive ecosystems to evolve and, in turn, the atmosphere also evolved. Throughout history, the Earth's atmosphere and biogeochemical cycles have been in a dynamic equilibrium with planetary ecosystems. The history is characterized by periods of significant transformation followed by millions of years of stability. The evolution of the earliest organisms, likely anaerobic methanogen microbes, started the process by converting atmospheric hydrogen into methane (4H2 + CO2 → CH4 + 2H2O). Anoxygenic photosynthesis reduced hydrogen concentrations and increased atmospheric methane, by converting hydrogen sulfide into water or other sulfur compounds (for example, 2H2S + CO2 + h"v" → CH2O + H2O + 2S). Early forms of fermentation also increased levels of atmospheric methane. The transition to an oxygen-dominant atmosphere (the "Great Oxidation") did not begin until approximately 2.4–2.3 billion years ago, but photosynthetic processes started 0.3 to 1 billion years prior. The biology of life operates within a certain range of temperatures. Heat is a form of energy that regulates temperature. Heat affects growth rates, activity, behaviour, and primary production. Temperature is largely dependent on the incidence of solar radiation. The latitudinal and longitudinal spatial variation of temperature greatly affects climates and consequently the distribution of biodiversity and levels of primary production in different ecosystems or biomes across the planet. Heat and temperature relate importantly to metabolic activity. Poikilotherms, for example, have a body temperature that is largely regulated and dependent on the temperature of the external environment. In contrast, homeotherms regulate their internal body temperature by expending metabolic energy. There is a relationship between light, primary production, and ecological energy budgets. Sunlight is the primary input of energy into the planet's ecosystems. Light is composed of electromagnetic energy of different wavelengths. Radiant energy from the sun generates heat, provides photons of light measured as active energy in the chemical reactions of life, and also acts as a catalyst for genetic mutation. Plants, algae, and some bacteria absorb light and assimilate the energy through photosynthesis. Organisms capable of assimilating energy by photosynthesis or through inorganic fixation of H2S are autotrophs. Autotrophs—responsible for primary production—assimilate light energy which becomes metabolically stored as potential energy in the form of biochemical enthalpic bonds. Diffusion of carbon dioxide and oxygen is approximately 10,000 times slower in water than in air. When soils are flooded, they quickly lose oxygen, becoming hypoxic (an environment with O2 concentration below 2 mg/liter) and eventually completely anoxic where anaerobic bacteria thrive among the roots. Water also influences the intensity and spectral composition of light as it reflects off the water surface and submerged particles. Aquatic plants exhibit a wide variety of morphological and physiological adaptations that allow them to survive, compete, and diversify in these environments. For example, their roots and stems contain large air spaces (aerenchyma) that regulate the efficient transportation of gases (for example, CO2 and O2) used in respiration and photosynthesis. Salt water plants (halophytes) have additional specialized adaptations, such as the development of special organs for shedding salt and osmoregulating their internal salt (NaCl) concentrations, to live in estuarine, brackish, or oceanic environments. Anaerobic soil microorganisms in aquatic environments use nitrate, manganese ions, ferric ions, sulfate, carbon dioxide, and some organic compounds; other microorganisms are facultative anaerobes and use oxygen during respiration when the soil becomes drier. The activity of soil microorganisms and the chemistry of the water reduces the oxidation-reduction potentials of the water. Carbon dioxide, for example, is reduced to methane (CH4) by methanogenic bacteria. The physiology of fish is also specially adapted to compensate for environmental salt levels through osmoregulation. Their gills form electrochemical gradients that mediate salt excretion in salt water and uptake in fresh water. The shape and energy of the land is significantly affected by gravitational forces. On a large scale, the distribution of gravitational forces on the earth is uneven and influences the shape and movement of tectonic plates as well as influencing geomorphic processes such as orogeny and erosion. These forces govern many of the geophysical properties and distributions of ecological biomes across the Earth. On the organismal scale, gravitational forces provide directional cues for plant and fungal growth (gravitropism), orientation cues for animal migrations, and influence the biomechanics and size of animals. Ecological traits, such as allocation of biomass in trees during growth are subject to mechanical failure as gravitational forces influence the position and structure of branches and leaves. The cardiovascular systems of animals are functionally adapted to overcome pressure and gravitational forces that change according to the features of organisms (e.g., height, size, shape), their behaviour (e.g., diving, running, flying), and the habitat occupied (e.g., water, hot deserts, cold tundra). Climatic and osmotic pressure places physiological constraints on organisms, especially those that fly and respire at high altitudes, or dive to deep ocean depths. These constraints influence vertical limits of ecosystems in the biosphere, as organisms are physiologically sensitive and adapted to atmospheric and osmotic water pressure differences. For example, oxygen levels decrease with decreasing pressure and are a limiting factor for life at higher altitudes. Water transportation by plants is another important ecophysiological process affected by osmotic pressure gradients. Water pressure in the depths of oceans requires that organisms adapt to these conditions. For example, diving animals such as whales, dolphins, and seals are specially adapted to deal with changes in sound due to water pressure differences. Differences between hagfish species provide another example of adaptation to deep-sea pressure through specialized protein adaptations. Turbulent forces in air and water affect the environment and ecosystem distribution, form and dynamics. On a planetary scale, ecosystems are affected by circulation patterns in the global trade winds. Wind power and the turbulent forces it creates can influence heat, nutrient, and biochemical profiles of ecosystems. For example, wind running over the surface of a lake creates turbulence, mixing the water column and influencing the environmental profile to create thermally layered zones, affecting how fish, algae, and other parts of the aquatic ecosystem are structured. Wind speed and turbulence also influence evapotranspiration rates and energy budgets in plants and animals. Wind speed, temperature and moisture content can vary as winds travel across different land features and elevations. For example, the westerlies come into contact with the coastal and interior mountains of western North America to produce a rain shadow on the leeward side of the mountain. The air expands and moisture condenses as the winds increase in elevation; this is called orographic lift and can cause precipitation. This environmental process produces spatial divisions in biodiversity, as species adapted to wetter conditions are range-restricted to the coastal mountain valleys and unable to migrate across the xeric ecosystems (e.g., of the Columbia Basin in western North America) to intermix with sister lineages that are segregated to the interior mountain systems. Plants convert carbon dioxide into biomass and emit oxygen into the atmosphere. By approximately 350 million years ago (the end of the Devonian period), photosynthesis had brought the concentration of atmospheric oxygen above 17%, which allowed combustion to occur. Fire releases CO2 and converts fuel into ash and tar. Fire is a significant ecological parameter that raises many issues pertaining to its control and suppression. While the issue of fire in relation to ecology and plants has been recognized for a long time, Charles Cooper brought attention to the issue of forest fires in relation to the ecology of forest fire suppression and management in the 1960s. Native North Americans were among the first to influence fire regimes by controlling their spread near their homes or by lighting fires to stimulate the production of herbaceous foods and basketry materials. Fire creates a heterogeneous ecosystem age and canopy structure, and the altered soil nutrient supply and cleared canopy structure opens new ecological niches for seedling establishment. Most ecosystems are adapted to natural fire cycles. Plants, for example, are equipped with a variety of adaptations to deal with forest fires. Some species (e.g., "Pinus halepensis") cannot germinate until after their seeds have lived through a fire or been exposed to certain compounds from smoke. Environmentally triggered germination of seeds is called serotiny. Fire plays a major role in the persistence and resilience of ecosystems. Soil is the living top layer of mineral and organic dirt that covers the surface of the planet. It is the chief organizing centre of most ecosystem functions, and it is of critical importance in agricultural science and ecology. The decomposition of dead organic matter (for example, leaves on the forest floor), results in soils containing minerals and nutrients that feed into plant production. The whole of the planet's soil ecosystems is called the pedosphere where a large biomass of the Earth's biodiversity organizes into trophic levels. Invertebrates that feed and shred larger leaves, for example, create smaller bits for smaller organisms in the feeding chain. Collectively, these organisms are the detritivores that regulate soil formation. Tree roots, fungi, bacteria, worms, ants, beetles, centipedes, spiders, mammals, birds, reptiles, amphibians, and other less familiar creatures all work to create the trophic web of life in soil ecosystems. Soils form composite phenotypes where inorganic matter is enveloped into the physiology of a whole community. As organisms feed and migrate through soils they physically displace materials, an ecological process called bioturbation. This aerates soils and stimulates heterotrophic growth and production. Soil microorganisms are influenced by and feed back into the trophic dynamics of the ecosystem. No single axis of causality can be discerned to segregate the biological from geomorphological systems in soils. Paleoecological studies of soils places the origin for bioturbation to a time before the Cambrian period. Other events, such as the evolution of trees and the colonization of land in the Devonian period played a significant role in the early development of ecological trophism in soils. Ecologists study and measure nutrient budgets to understand how these materials are regulated, flow, and recycled through the environment. This research has led to an understanding that there is global feedback between ecosystems and the physical parameters of this planet, including minerals, soil, pH, ions, water, and atmospheric gases. Six major elements (hydrogen, carbon, nitrogen, oxygen, sulfur, and phosphorus; H, C, N, O, S, and P) form the constitution of all biological macromolecules and feed into the Earth's geochemical processes. From the smallest scale of biology, the combined effect of billions upon billions of ecological processes amplify and ultimately regulate the biogeochemical cycles of the Earth. Understanding the relations and cycles mediated between these elements and their ecological pathways has significant bearing toward understanding global biogeochemistry. The ecology of global carbon budgets gives one example of the linkage between biodiversity and biogeochemistry. It is estimated that the Earth's oceans hold 40,000 gigatonnes (Gt) of carbon, that vegetation and soil hold 2070 Gt, and that fossil fuel emissions are 6.3 Gt carbon per year. There have been major restructurings in these global carbon budgets during the Earth's history, regulated to a large extent by the ecology of the land. For example, through the early-mid Eocene volcanic outgassing, the oxidation of methane stored in wetlands, and seafloor gases increased atmospheric CO2 (carbon dioxide) concentrations to levels as high as 3500 ppm. In the Oligocene, from twenty-five to thirty-two million years ago, there was another significant restructuring of the global carbon cycle as grasses evolved a new mechanism of photosynthesis, C4 photosynthesis, and expanded their ranges. This new pathway evolved in response to the drop in atmospheric CO2 concentrations below 550 ppm. The relative abundance and distribution of biodiversity alters the dynamics between organisms and their environment such that ecosystems can be both cause and effect in relation to climate change. Human-driven modifications to the planet's ecosystems (e.g., disturbance, biodiversity loss, agriculture) contributes to rising atmospheric greenhouse gas levels. Transformation of the global carbon cycle in the next century is projected to raise planetary temperatures, lead to more extreme fluctuations in weather, alter species distributions, and increase extinction rates. The effect of global warming is already being registered in melting glaciers, melting mountain ice caps, and rising sea levels. Consequently, species distributions are changing along waterfronts and in continental areas where migration patterns and breeding grounds are tracking the prevailing shifts in climate. Large sections of permafrost are also melting to create a new mosaic of flooded areas having increased rates of soil decomposition activity that raises methane (CH4) emissions. There is concern over increases in atmospheric methane in the context of the global carbon cycle, because methane is a greenhouse gas that is 23 times more effective at absorbing long-wave radiation than CO2 on a 100-year time scale. Hence, there is a relationship between global warming, decomposition and respiration in soils and wetlands producing significant climate feedbacks and globally altered biogeochemical cycles. Ecology has a complex origin, due in large part to its interdisciplinary nature. Ancient Greek philosophers such as Hippocrates and Aristotle were among the first to record observations on natural history. However, they viewed life in terms of essentialism, where species were conceptualized as static unchanging things while varieties were seen as aberrations of an idealized type. This contrasts against the modern understanding of ecological theory where varieties are viewed as the real phenomena of interest and having a role in the origins of adaptations by means of natural selection. Early conceptions of ecology, such as a balance and regulation in nature can be traced to Herodotus (died "c". 425 BC), who described one of the earliest accounts of mutualism in his observation of "natural dentistry". Basking Nile crocodiles, he noted, would open their mouths to give sandpipers safe access to pluck leeches out, giving nutrition to the sandpiper and oral hygiene for the crocodile. Aristotle was an early influence on the philosophical development of ecology. He and his student Theophrastus made extensive observations on plant and animal migrations, biogeography, physiology, and on their behaviour, giving an early analogue to the modern concept of an ecological niche. Ernst Haeckel (left) and Eugenius Warming (right), two founders of ecology Ecological concepts such as food chains, population regulation, and productivity were first developed in the 1700s, through the published works of microscopist Antoni van Leeuwenhoek (1632–1723) and botanist Richard Bradley (1688?–1732). Biogeographer Alexander von Humboldt (1769–1859) was an early pioneer in ecological thinking and was among the first to recognize ecological gradients, where species are replaced or altered in form along environmental gradients, such as a cline forming along a rise in elevation. Humboldt drew inspiration from Isaac Newton as he developed a form of "terrestrial physics". In Newtonian fashion, he brought a scientific exactitude for measurement into natural history and even alluded to concepts that are the foundation of a modern ecological law on species-to-area relationships. Natural historians, such as Humboldt, James Hutton, and Jean-Baptiste Lamarck (among others) laid the foundations of the modern ecological sciences. The term "ecology" () was coined by Ernst Haeckel in his book "Generelle Morphologie der Organismen" (1866). Haeckel was a zoologist, artist, writer, and later in life a professor of comparative anatomy. Opinions differ on who was the founder of modern ecological theory. Some mark Haeckel's definition as the beginning; others say it was Eugenius Warming with the writing of Oecology of Plants: An Introduction to the Study of Plant Communities (1895), or Carl Linnaeus' principles on the economy of nature that matured in the early 18th century. Linnaeus founded an early branch of ecology that he called the economy of nature. His works influenced Charles Darwin, who adopted Linnaeus' phrase on the "economy or polity of nature" in "The Origin of Species". Linnaeus was the first to frame the balance of nature as a testable hypothesis. Haeckel, who admired Darwin's work, defined ecology in reference to the economy of nature, which has led some to question whether ecology and the economy of nature are synonymous. From Aristotle until Darwin, the natural world was predominantly considered static and unchanging. Prior to "The Origin of Species", there was little appreciation or understanding of the dynamic and reciprocal relations between organisms, their adaptations, and the environment. An exception is the 1789 publication "Natural History of Selborne" by Gilbert White (1720–1793), considered by some to be one of the earliest texts on ecology. While Charles Darwin is mainly noted for his treatise on evolution, he was one of the founders of soil ecology, and he made note of the first ecological experiment in "The Origin of Species". Evolutionary theory changed the way that researchers approached the ecological sciences. Modern ecology is a young science that first attracted substantial scientific attention toward the end of the 19th century (around the same time that evolutionary studies were gaining scientific interest). The scientist Ellen Swallow Richards may have first introduced the term "oekology" (which eventually morphed into home economics) in the U.S. as early 1892. In the early 20th century, ecology transitioned from a more descriptive form of natural history to a more analytical form of "scientific natural history". Frederic Clements published the first American ecology book in 1905, presenting the idea of plant communities as a superorganism. This publication launched a debate between ecological holism and individualism that lasted until the 1970s. Clements' superorganism concept proposed that ecosystems progress through regular and determined stages of seral development that are analogous to the developmental stages of an organism. The Clementsian paradigm was challenged by Henry Gleason, who stated that ecological communities develop from the unique and coincidental association of individual organisms. This perceptual shift placed the focus back onto the life histories of individual organisms and how this relates to the development of community associations. The Clementsian superorganism theory was an overextended application of an idealistic form of holism. The term "holism" was coined in 1926 by Jan Christiaan Smuts, a South African general and polarizing historical figure who was inspired by Clements' superorganism concept. Around the same time, Charles Elton pioneered the concept of food chains in his classical book "Animal Ecology". Elton defined ecological relations using concepts of food chains, food cycles, and food size, and described numerical relations among different functional groups and their relative abundance. Elton's 'food cycle' was replaced by 'food web' in a subsequent ecological text. Alfred J. Lotka brought in many theoretical concepts applying thermodynamic principles to ecology. In 1942, Raymond Lindeman wrote a landmark paper on the trophic dynamics of ecology, which was published posthumously after initially being rejected for its theoretical emphasis. Trophic dynamics became the foundation for much of the work to follow on energy and material flow through ecosystems. Robert MacArthur advanced mathematical theory, predictions, and tests in ecology in the 1950s, which inspired a resurgent school of theoretical mathematical ecologists. Ecology also has developed through contributions from other nations, including Russia's Vladimir Vernadsky and his founding of the biosphere concept in the 1920s and Japan's Kinji Imanishi and his concepts of harmony in nature and habitat segregation in the 1950s. Scientific recognition of contributions to ecology from non-English-speaking cultures is hampered by language and translation barriers. Ecology surged in popular and scientific interest during the 1960–1970s environmental movement. There are strong historical and scientific ties between ecology, environmental management, and protection. The historical emphasis and poetic naturalistic writings advocating the protection of wild places by notable ecologists in the history of conservation biology, such as Aldo Leopold and Arthur Tansley, have been seen as far removed from urban centres where, it is claimed, the concentration of pollution and environmental degradation is located. Palamar (2008) notes an overshadowing by mainstream environmentalism of pioneering women in the early 1900s who fought for urban health ecology (then called euthenics) and brought about changes in environmental legislation. Women such as Ellen Swallow Richards and Julia Lathrop, among others, were precursors to the more popularized environmental movements after the 1950s. In 1962, marine biologist and ecologist Rachel Carson's book "Silent Spring" helped to mobilize the environmental movement by alerting the public to toxic pesticides, such as DDT, bioaccumulating in the environment. Carson used ecological science to link the release of environmental toxins to human and ecosystem health. Since then, ecologists have worked to bridge their understanding of the degradation of the planet's ecosystems with environmental politics, law, restoration, and natural resources management.
https://en.wikipedia.org/wiki?curid=9630
Glossary of country dance terms An alphabetic list of modern country dance terminology: Active couple - for longways sets, the active couple is the couple nearest the head of the set within each "minor set." There are always exactly as many active couples as "minor sets." If the dance is "duple minor," this works out to every other couple, while in a "triple minor" it is every third couple. In older dances from the seventeenth and eighteenth centuries, the active couples do more complicated figures than the inactives, whence the name; however, this is not so usual in modern dances. Active couples may also be termed "first couple" or "the Ones," while inactives are "second couple/the Twos" and (only in a "triple minor" dance) "third couple/the Threes." Arm right (or left) - couples link right (or left) arms and move forward in a circle, returning to their starting positions. Back to back - facing another person, move forward "passing" right shoulders and "fall back" to place passing left. May also start by passing left and falling back right. Called a do si do in contra dance (and "dos-à-dos" in France). Balance - a "single," generally found in pairs, as "balance forward and back." Becket formation - a 20th-century variation of the duple minor "longways set." Each couple stands either on the men's line or the women's line, with the lady on the right. Within each minor set, one couple faces the left wall of the hall and the other the right wall, rather than facing the "head" or "foot." There are no active or inactive couples. "Progression" is accomplished by each couple moving to their own left along their line at the end of each iteration of the dance; thus the couples on the men's line go up, while those on the women's line go down. This was originally a contra dance form but can sometimes be found in English country dance. Both hands - two dancers face each other and give hands right to left and left to right. Cast - turn outward and dance up or down outside the set, as directed. The instruction "cast off" is frequently synonymous with "cast down". Changes of right and left - like the "circular hey", but dancers give hands as they pass (handing hey). The number of changes is given first (e.g. two changes, three changes, etc.). Chassé - slipping step to right or left as directed. Circular hey - dancers face partners or along the line and "pass" right and left alternating a stated number of changes. Usually done without hands, the circular hey may also be done by more than two couples facing alternately and moving in opposite directions - usually to their original places. This name for the figure was invented by Cecil Sharp and does not appear in sources pre-1900. Nonetheless, some early country dances calling for heys have been interpreted in modern times using circular heys. In early dances, where the hey is called a "double hey", it works to interpret this as an oval hey, like the modern circular hey but adapted to the straight sides of a longways formation. Clockwise - in a ring, move to one's left. In a "turn single" turn to the right. Contrary - your contrary is not your partner. In Playford's original notation, this term meant the same thing that "Corner" (or sometimes "Opposite") means today. Corner - in a two-couple "minor set," the dancer diagonally opposite one. The first man and the second woman are "first corners," while the first woman and second man are "second corners." In other dance formations, it has similar meanings. Counter-clockwise - the opposite of clockwise - in a ring, move right. In a "turn single", turn to the left. Cross hands - face and give left to left and right to right. Cross over or pass - change places with another dancer moving forward and passing by the right shoulder, unless otherwise directed. Cross and go below - cross as above and go outside below one couple, ending improper. Double - four steps forward or back, closing the feet on the 4th step (see "Single" below). Fall (back) - dance backwards. Figure of 8 - a weaving figure in which a moving couple crosses between a standing couple and casts around them in a figure 8 pattern. To do this once, ending in one's partner's place, is a half figure of 8; to do it twice, returning to one's own place, is a full figure of 8. The right of way in the cross has traditionally been given to the lady; some communities prefer to give it to whichever dancer is coming from the left-hand side. In a double figure of 8, the other couple does not stand still, but performs their own figure of 8 simultaneously; they begin with the cast and end with the cross to avoid collision. Forward - "lead" or move in the direction you are facing. Grand chain - a handing hey (changes of right and left) done in a circle of more than two couples. Gypsy - two dancers move around each other in a circular path while facing each other. Hands across - right or left hands are given to "corners", and dancers move in the direction they face. In contra dance, instead of taking one's corner's hand, one grasps the wrist of the next dancer. Also known as a star right/left. Hands three, four etc. - the designated number of dancers form a ring and move around in the direction indicated, usually first to the left and back to the right. Head and foot - the head of a "longways set" is the end with the music; the foot is the other end. Toward the head is "up," and toward the foot is "down." Hey - a weaving figure in which dancers move in single file along a set track, passing one another on alternating sides (see "circular hey" and "straight hey"). In Scottish country dance, the hey is known as the reel. "Hole in the Wall" cross - a type of "cross". In a regular cross, the dancers walk past each other and turn upon reaching the other line; in a "Hole in the Wall" cross, they meet in the middle, make a brief half-turn without hands, and back into one another's place, maintaining eye contact the while. Named for "Hole in the Wall," a dance in which it appears. Honour - couples step forward and right, close, shift weight, and curtsey or bow, then (usually) repeat to their left. In the time of Playford's original manual, a woman's curtsey was similar to the modern one, but a man's honour (or reverence) kept the upper body upright and involved sliding the left leg forward while bending the right knee Improper - see "proper". Ladies' chain - a figure in which ladies dance first with each other in the center of the set and then with the gentlemen on the sides. In its simplest form, two ladies begin in "second corner" positions (nearer the head on the women's line and nearer the foot on the men's line). The ladies pass each other by right hand and turn with the gentlemen by left hand, approximately once around, to end with the ladies in each other's place and the gentlemen where they began. The figure can be extended to more couples in a ring, as long as the dancers in the ring are alternating between gentlemen and ladies. If the gentlemen turn the ladies only by left hand, that is an open ladies' chain; if they also place their right hands on the ladies' backs during the turn, that is a closed ladies' chain. In English country dance, both closed and open ladies' chains are to be found, and the gentlemen make a short "cast" up or down the set to meet the ladies; in contra dance, only the closed ladies' chain is done, and the gentlemen sidestep to meet the ladies. The men's chain is a simple gender reversal, but is a much rarer figure. Lead - join inside hands and walk in a certain direction. To lead up or down is to walk toward or away from the head of the set; to lead out is to walk away from the other line of dancers. Link - see "set and link". Longways set - a line of couples dancing together. This is usually "longways for as many as will," indicating that any number of couples may join the longways set—although some dances require a three- or four-couple longways set. If the longways set is not restricted to three or four couples, it will be subdivided into "minor sets" of two or three couples each. "Mad Robin" figure - a figure in which one couple dances around their respective neighbours. Men take one step forward and then slide to the right passing in front of their neighbour, then step backward and slide left behind their neighbour. Conversely women take one step backward and then slide to the left passing behind their neighbour, then step forward and slide right in front of their neighbour. In one version, the dancer who is going outside the set at the moment casts out to begin that motion; in the other, the active couple maintains eye contact. The term "Mad Robin" comes from the name of the dance which originated the figure. A version involving all four dancers was developed for contra dancing and later readmitted into some modern English dances. Minor set - a "longways set" is subdivided into several minor sets. In a "duple minor" dance, every two couples form a minor set. In a "triple minor" dance, every three couples form a minor set. The "active couple" is always the couple in each minor set who are closest to the head. After every iteration of the dance, the "progression" will create new minor sets for the next iteration. Neighbour - the person you are standing beside, but not your partner. Opposite - the person you are facing, if you are not facing your partner. Poussette - two dancers face, give both hands and change places as a couple with two adjacent dancers. One pair moves a "double" toward one wall, the other toward the other wall; they shift up or down, respectively, and move into the other couple's place with another "double." This completes a half-poussette; it is repeated for a whole poussette. In a draw poussette, each couple turns instead of reversing direction, so that the same dancer in each couple is always in the lead. Progression - the process by which every couple will eventually dance with every other couple in the set, if the dance is repeated enough times. In a "duple minor" dance with five couples dancing, for example, the couples are initially in this order: Active (couple A)/Inactive (couple B)/Active (couple C)/Inactive (couple D)/Out (couple E). This represents two "minor sets" (couples A-B and couples C-D) and one couple (couple E) who are "standing out" due to having no one to dance with. After one iteration of the dance, every active couple will have moved below the inactive couple in their "minor set," which in the example would be thus: Inactive (couple B)/Active (couple A)/Inactive (couple D)/Active (couple C)/Out (couple E). For the next iteration, any inactive couple at the top (and any active couple at the bottom) will stand out, while any couple standing out will begin dancing as actives (if at the top) or inactives (if at the bottom). So the next iteration would begin as follows: Out (couple B)/Active (couple A)/Inactive (couple D)/Active (couple C)/Inactive (couple E). The "minor sets" now contain couples A-D and couples C-E, while couple B is "standing out." Dances in other forms progress differently, though the "triple minor" progression is quite similar. Progression, double or triple - a longways dance has a double progression if the arrangement of couples into minor sets advances twice during one iteration of the dance instead of just once. A triple-progression dance advances thrice during one iteration. Proper - with the man on the left and the woman on the right, from the perspective of someone facing the music. Improper is the opposite. The terms carry no value judgment, but only indicate whether one is on one's "home" side. A dance in duple-minor longways form is termed "improper" if the active couples are improper by default; this is the exception in English country dance, but the rule in contra dance. Right and left - see "changes of right and left." Set - a dancer steps right, closes with left foot and shifts weight to it, then steps back to the right foot (right-together-step); then repeats the process mirror-image (left-together-step). In some areas, such as the Society for Creative Anachronism, it is done starting to the left. It may be done in place or advancing. Often followed by a turn single. In Scottish country dance there are several variations; in contra dance its place is generally taken by a "balance right and left." Not to be confused with terms indicating groups of dancers, like "longways set" or "minor set." Set and link - a figure done by a pair of dancers and simultaneously by another pair of dancers who are facing them. Most commonly this means that the men do it facing the women, while the women do it facing the men. First, all dancers "set;" then the dancer on the left of each pair dances a "turn single right," while also moving to the right, to end in his or her neighbor's place. Meanwhile the dancer on the right of each pair "casts" to the left into his or her neighbor's place; thus the men have traded places with each other, and so have the women. This figure is most commonly found in Scottish country dance. Sicilian circle - a type of dance formation, roughly equivalent to a "longways set" rolled into a ring. Every couple stands along the line of a large circle, facing another couple; thus half of the couples face clockwise, while the other half face counterclockwise. Since, unlike the "longways set," the Sicilian circle has no place for dancers to "stand out," Sicilian circle dances must be done by an even number of couples. The "progression" is similar to that of a "duple minor," but since there is nowhere for couples to reverse direction, every clockwise couple will only dance with the counterclockwise couples (and vice versa). Siding - two dancers go forward in four counts to meet side by side, then back in four counts to where they started the figure. As depicted by Feuillet, this is done right side by right side the first time, left by left the second time. In Cecil Sharp's reconstruction, the dancers pass by left shoulder (in some versions holding hands), turn to face each other, then return along the same path, passing by right shoulder; this is then repeated. So-called "Cecil Sharp siding" is no longer considered historical, but is still used on its own merits. Standard siding is sometimes called "Pat Shaw siding" (after its reconstructor) to distinguish it from "Cecil Sharp siding." Single - two steps in any direction, closing feet on the second step. The second step tends to be interpreted as a closing action in which weight usually stays on the same foot as before, consistent with descriptions from Renaissance sources. Slipping circle (left or right) - dancers take hands in a circle (facing in) and chassé left or right. Star - see "hands across." Straight hey for four - dancers face alternately, the two in the middle facing out. Dancers pass right shoulders on either end and weave to the end opposite. If the last pass at the end is by the right, the dancer turns right and reenters the line by the same shoulder; vice versa if the last pass was to the left. Dancers end in their original places. Straight hey for three - the first dancer faces the other two and "passes" right shoulders with the second dancer, left shoulder with the third - the other dancers moving and passing the indicated shoulder. On making the last pass, each dancer makes a whole turn on the end, bearing right if the last pass was by the right shoulder or left if last pass was by the left, and reenters the figure returning to place. Each dancer describes a figure of eight pattern. Swing - a "turn" with two hands, but moving faster and making more than one revolution. Several variants exist, including the ballroom swing and the Welsh swing. Track figure - a generic term for any composite figure where the dancers involved travel within the set. An example track figure might be "Ones cast around the Twos, cross, cast around the Threes, and lead back up to place." The "figure of 8" would be considered a track figure if it were not common enough to have its own name. Turn both-hands - face, give "both hands", and make a complete circular, clockwise turn to place. Turn by right or left - dancers join right (or left) hands and turn around, separate, and "fall" to places. Turn single - dancers turn around in four steps. "Turn single right" is a clockwise turn; "turn single left" is a counterclockwise turn. May involve a backward motion, as after a "set advancing." Up a double and back - common combination in which dancers, usually having linked hands in a line, advance a double and then retire another double.
https://en.wikipedia.org/wiki?curid=9631
Ecosystem An ecosystem is a community of living organisms in conjunction with the nonliving components of their environment, interacting as a system. These biotic and abiotic components are linked together through nutrient cycles and energy flows. Energy enters the system through photosynthesis and is incorporated into plant tissue. By feeding on plants and on one another, animals play an important role in the movement of matter and energy through the system. They also influence the quantity of plant and microbial biomass present. By breaking down dead organic matter, decomposers release carbon back to the atmosphere and facilitate nutrient cycling by converting nutrients stored in dead biomass back to a form that can be readily used by plants and other microbes. Ecosystems are controlled by external and internal factors. External factors such as climate, parent material which forms the soil and topography, control the overall structure of an ecosystem but are not themselves influenced by the ecosystem. Unlike external factors, internal factors are controlled, for example, decomposition, root competition, shading, disturbance, succession, and the types of species present. Ecosystems are dynamic entities—they are subject to periodic disturbances and are in the process of recovering from some past disturbance. Ecosystems in similar environments that are located in different parts of the world can end up doing things very differently simply because they have different pools of species present. Internal factors not only control ecosystem processes but are also controlled by them and are often subject to feedback loops. Resource inputs are generally controlled by external processes like climate and parent material. Resource availability within the ecosystem is controlled by internal factors like decomposition, root competition or shading. Although humans operate within ecosystems, their cumulative effects are large enough to influence external factors like climate. Biodiversity affects ecosystem functioning, as do the processes of disturbance and succession. Ecosystems provide a variety of goods and services upon which people depend. The term ecosystem was first used in 1935 in a publication by British ecologist Arthur Tansley. Tansley devised the concept to draw attention to the importance of transfers of materials between organisms and their environment. He later refined the term, describing it as "The whole system, ... including not only the organism-complex, but also the whole complex of physical factors forming what we call the environment". Tansley regarded ecosystems not simply as natural units, but as "mental isolates". Tansley later defined the spatial extent of ecosystems using the term ecotope. G. Evelyn Hutchinson, a limnologist who was a contemporary of Tansley's, combined Charles Elton's ideas about trophic ecology with those of Russian geochemist Vladimir Vernadsky. As a result, he suggested that mineral nutrient availability in a lake limited algal production. This would, in turn, limit the abundance of animals that feed on algae. Raymond Lindeman took these ideas further to suggest that the flow of energy through a lake was the primary driver of the ecosystem. Hutchinson's students, brothers Howard T. Odum and Eugene P. Odum, further developed a "systems approach" to the study of ecosystems. This allowed them to study the flow of energy and material through ecological systems. Ecosystems are controlled both by external and internal factors. External factors, also called state factors, control the overall structure of an ecosystem and the way things work within it, but are not themselves influenced by the ecosystem. The most important of these is climate. Climate determines the biome in which the ecosystem is embedded. Rainfall patterns and seasonal temperatures influence photosynthesis and thereby determine the amount of water and energy available to the ecosystem. Parent material determines the nature of the soil in an ecosystem, and influences the supply of mineral nutrients. Topography also controls ecosystem processes by affecting things like microclimate, soil development and the movement of water through a system. For example, ecosystems can be quite different if situated in a small depression on the landscape, versus one present on an adjacent steep hillside. Other external factors that play an important role in ecosystem functioning include time and potential biota. Similarly, the set of organisms that can potentially be present in an area can also significantly affect ecosystems. Ecosystems in similar environments that are located in different parts of the world can end up doing things very differently simply because they have different pools of species present. The introduction of non-native species can cause substantial shifts in ecosystem function. Unlike external factors, internal factors in ecosystems not only control ecosystem processes but are also controlled by them. Consequently, they are often subject to feedback loops. While the resource inputs are generally controlled by external processes like climate and parent material, the availability of these resources within the ecosystem is controlled by internal factors like decomposition, root competition or shading. Other factors like disturbance, succession or the types of species present are also internal factors. Primary production is the production of organic matter from inorganic carbon sources. This mainly occurs through photosynthesis. The energy incorporated through this process supports life on earth, while the carbon makes up much of the organic matter in living and dead biomass, soil carbon and fossil fuels. It also drives the carbon cycle, which influences global climate via the greenhouse effect. Through the process of photosynthesis, plants capture energy from light and use it to combine carbon dioxide and water to produce carbohydrates and oxygen. The photosynthesis carried out by all the plants in an ecosystem is called the gross primary production (GPP). About half of the GPP is consumed in plant respiration. The remainder, that portion of GPP that is not used up by respiration, is known as the net primary production (NPP). Total photosynthesis is limited by a range of environmental factors. These include the amount of light available, the amount of leaf area a plant has to capture light (shading by other plants is a major limitation of photosynthesis), rate at which carbon dioxide can be supplied to the chloroplasts to support photosynthesis, the availability of water, and the availability of suitable temperatures for carrying out photosynthesis. Energy and carbon enter ecosystems through photosynthesis, are incorporated into living tissue, transferred to other organisms that feed on the living and dead plant matter, and eventually released through respiration. The carbon and energy incorporated into plant tissues (net primary production) is either consumed by animals while the plant is alive, or it remains uneaten when the plant tissue dies and becomes detritus. In terrestrial ecosystems, roughly 90% of the net primary production ends up being broken down by decomposers. The remainder is either consumed by animals while still alive and enters the plant-based trophic system, or it is consumed after it has died, and enters the detritus-based trophic system. In aquatic systems, the proportion of plant biomass that gets consumed by herbivores is much higher. In trophic systems photosynthetic organisms are the primary producers. The organisms that consume their tissues are called primary consumers or secondary producers—herbivores. Organisms which feed on microbes (bacteria and fungi) are termed microbivores. Animals that feed on primary consumers—carnivores—are secondary consumers. Each of these constitutes a trophic level. The sequence of consumption—from plant to herbivore, to carnivore—forms a food chain. Real systems are much more complex than this—organisms will generally feed on more than one form of food, and may feed at more than one trophic level. Carnivores may capture some prey which is part of a plant-based trophic system and others that are part of a detritus-based trophic system (a bird that feeds both on herbivorous grasshoppers and earthworms, which consume detritus). Real systems, with all these complexities, form food webs rather than food chains. The food chain usually consists of five levels of consumption which are producers, primary consumers, secondary consumers, tertiary consumers, and decomposers. The carbon and nutrients in dead organic matter are broken down by a group of processes known as decomposition. This releases nutrients that can then be re-used for plant and microbial production and returns carbon dioxide to the atmosphere (or water) where it can be used for photosynthesis. In the absence of decomposition, the dead organic matter would accumulate in an ecosystem, and nutrients and atmospheric carbon dioxide would be depleted. Approximately 90% of terrestrial net primary production goes directly from plant to decomposer. Decomposition processes can be separated into three categories—leaching, fragmentation and chemical alteration of dead material. As water moves through dead organic matter, it dissolves and carries with it the water-soluble components. These are then taken up by organisms in the soil, react with mineral soil, or are transported beyond the confines of the ecosystem (and are considered lost to it). Newly shed leaves and newly dead animals have high concentrations of water-soluble components and include sugars, amino acids and mineral nutrients. Leaching is more important in wet environments and much less important in dry ones. Fragmentation processes break organic material into smaller pieces, exposing new surfaces for colonization by microbes. Freshly shed leaf litter may be inaccessible due to an outer layer of cuticle or bark, and cell contents are protected by a cell wall. Newly dead animals may be covered by an exoskeleton. Fragmentation processes, which break through these protective layers, accelerate the rate of microbial decomposition. Animals fragment detritus as they hunt for food, as does passage through the gut. Freeze-thaw cycles and cycles of wetting and drying also fragment dead material. The chemical alteration of the dead organic matter is primarily achieved through bacterial and fungal action. Fungal hyphae produces enzymes that can break through the tough outer structures surrounding dead plant material. They also produce enzymes which break down lignin, which allows them access to both cell contents and the nitrogen in the lignin. Fungi can transfer carbon and nitrogen through their hyphal networks and thus, unlike bacteria, are not dependent solely on locally available resources. Decomposition rates vary among ecosystems. The rate of decomposition is governed by three sets of factors—the physical environment (temperature, moisture, and soil properties), the quantity and quality of the dead material available to decomposers, and the nature of the microbial community itself. Temperature controls the rate of microbial respiration; the higher the temperature, the faster the microbial decomposition occurs. It also affects soil moisture, which slows microbial growth and reduces leaching. Freeze-thaw cycles also affect decomposition—freezing temperatures kill soil microorganisms, which allows leaching to play a more important role in moving nutrients around. This can be especially important as the soil thaws in the spring, creating a pulse of nutrients which become available. Decomposition rates are low under very wet or very dry conditions. Decomposition rates are highest in wet, moist conditions with adequate levels of oxygen. Wet soils tend to become deficient in oxygen (this is especially true in wetlands), which slows microbial growth. In dry soils, decomposition slows as well, but bacteria continue to grow (albeit at a slower rate) even after soils become too dry to support plant growth. Ecosystems continually exchange energy and carbon with the wider environment. Mineral nutrients, on the other hand, are mostly cycled back and forth between plants, animals, microbes and the soil. Most nitrogen enters ecosystems through biological nitrogen fixation, is deposited through precipitation, dust, gases or is applied as fertilizer. Since most terrestrial ecosystems are nitrogen-limited, nitrogen cycling is an important control on ecosystem production. Until modern times, nitrogen fixation was the major source of nitrogen for ecosystems. Nitrogen-fixing bacteria either live symbiotically with plants or live freely in the soil. The energetic cost is high for plants that support nitrogen-fixing symbionts—as much as 25% of gross primary production when measured in controlled conditions. Many members of the legume plant family support nitrogen-fixing symbionts. Some cyanobacteria are also capable of nitrogen fixation. These are phototrophs, which carry out photosynthesis. Like other nitrogen-fixing bacteria, they can either be free-living or have symbiotic relationships with plants. Other sources of nitrogen include acid deposition produced through the combustion of fossil fuels, ammonia gas which evaporates from agricultural fields which have had fertilizers applied to them, and dust. Anthropogenic nitrogen inputs account for about 80% of all nitrogen fluxes in ecosystems. When plant tissues are shed or are eaten, the nitrogen in those tissues becomes available to animals and microbes. Microbial decomposition releases nitrogen compounds from dead organic matter in the soil, where plants, fungi, and bacteria compete for it. Some soil bacteria use organic nitrogen-containing compounds as a source of carbon, and release ammonium ions into the soil. This process is known as nitrogen mineralization. Others convert ammonium to nitrite and nitrate ions, a process known as nitrification. Nitric oxide and nitrous oxide are also produced during nitrification. Under nitrogen-rich and oxygen-poor conditions, nitrates and nitrites are converted to nitrogen gas, a process known as denitrification. Other important nutrients include phosphorus, sulfur, calcium, potassium, magnesium and manganese. Phosphorus enters ecosystems through weathering. As ecosystems age this supply diminishes, making phosphorus-limitation more common in older landscapes (especially in the tropics). Calcium and sulfur are also produced by weathering, but acid deposition is an important source of sulfur in many ecosystems. Although magnesium and manganese are produced by weathering, exchanges between soil organic matter and living cells account for a significant portion of ecosystem fluxes. Potassium is primarily cycled between living cells and soil organic matter. Biodiversity plays an important role in ecosystem functioning. The reason for this is that ecosystem processes are driven by the number of species in an ecosystem, the exact nature of each individual species, and the relative abundance organisms within these species. Ecosystem processes are broad generalizations that actually take place through the actions of individual organisms. The nature of the organisms—the species, functional groups and trophic levels to which they belong—dictates the sorts of actions these individuals are capable of carrying out and the relative efficiency with which they do so. Ecological theory suggests that in order to coexist, species must have some level of limiting similarity—they must be different from one another in some fundamental way, otherwise one species would competitively exclude the other. Despite this, the cumulative effect of additional species in an ecosystem is not linear—additional species may enhance nitrogen retention, for example, but beyond some level of species richness, additional species may have little additive effect. The addition (or loss) of species that are ecologically similar to those already present in an ecosystem tends to only have a small effect on ecosystem function. Ecologically distinct species, on the other hand, have a much larger effect. Similarly, dominant species have a large effect on ecosystem function, while rare species tend to have a small effect. Keystone species tend to have an effect on ecosystem function that is disproportionate to their abundance in an ecosystem. Similarly, an ecosystem engineer is any organism that creates, significantly modifies, maintains or destroys a habitat. Ecosystems are dynamic entities. They are subject to periodic disturbances and are in the process of recovering from some past disturbance. When a perturbation occurs, an ecosystem responds by moving away from its initial state. The tendency of an ecosystem to remain close to its equilibrium state, despite that disturbance, is termed its resistance. On the other hand, the speed with which it returns to its initial state after disturbance is called its resilience. Time plays a role in the development of soil from bare rock and the recovery of a community from disturbance. From one year to another, ecosystems experience variation in their biotic and abiotic environments. A drought, a colder than usual winter, and a pest outbreak all are short-term variability in environmental conditions. Animal populations vary from year to year, building up during resource-rich periods and crashing as they overshoot their food supply. These changes play out in changes in net primary production decomposition rates, and other ecosystem processes. Longer-term changes also shape ecosystem processes—the forests of eastern North America still show legacies of cultivation which ceased 200 years ago, while methane production in eastern Siberian lakes is controlled by organic matter which accumulated during the Pleistocene. Disturbance also plays an important role in ecological processes. F. Stuart Chapin and coauthors define disturbance as "a relatively discrete event in time and space that alters the structure of populations, communities, and ecosystems and causes changes in resources availability or the physical environment". This can range from tree falls and insect outbreaks to hurricanes and wildfires to volcanic eruptions. Such disturbances can cause large changes in plant, animal and microbe populations, as well as soil organic matter content. Disturbance is followed by succession, a "directional change in ecosystem structure and functioning resulting from biotically driven changes in resources supply." The frequency and severity of disturbance determine the way it affects ecosystem function. A major disturbance like a volcanic eruption or glacial advance and retreat leave behind soils that lack plants, animals or organic matter. Ecosystems that experience such disturbances undergo primary succession. A less severe disturbance like forest fires, hurricanes or cultivation result in secondary succession and a faster recovery. More severe disturbance and more frequent disturbance result in longer recovery times. Ecosystem ecology studies the processes and dynamics of ecosystems, and the way the flow of matter and energy through them structures natural systems. The study of ecosystems can cover 10 orders of magnitude, from the surface layers of rocks to the surface of the planet. There is no single definition of what constitutes an ecosystem. German ecologist Ernst-Detlef Schulze and coauthors defined an ecosystem as an area which is "uniform regarding the biological turnover, and contains all the fluxes above and below the ground area under consideration." They explicitly reject Gene Likens' use of entire river catchments as "too wide a demarcation" to be a single ecosystem, given the level of heterogeneity within such an area. Other authors have suggested that an ecosystem can encompass a much larger area, even the whole planet. Schulze and coauthors also rejected the idea that a single rotting log could be studied as an ecosystem because the size of the flows between the log and its surroundings are too large, relative to the proportion cycles within the log. Philosopher of science Mark Sagoff considers the failure to define "the kind of object it studies" to be an obstacle to the development of theory in ecosystem ecology. Ecosystems can be studied through a variety of approaches—theoretical studies, studies monitoring specific ecosystems over long periods of time, those that look at differences between ecosystems to elucidate how they work and direct manipulative experimentation. Studies can be carried out at a variety of scales, ranging from whole-ecosystem studies to studying or mesocosms (simplified representations of ecosystems). American ecologist Stephen R. Carpenter has argued that microcosm experiments can be "irrelevant and diversionary" if they are not carried out in conjunction with field studies done at the ecosystem scale. Microcosm experiments often fail to accurately predict ecosystem-level dynamics. The Hubbard Brook Ecosystem Study started in 1963 to study the White Mountains in New Hampshire. It was the first successful attempt to study an entire watershed as an ecosystem. The study used stream chemistry as a means of monitoring ecosystem properties, and developed a detailed biogeochemical model of the ecosystem. Long-term research at the site led to the discovery of acid rain in North America in 1972. Researchers documented the depletion of soil cations (especially calcium) over the next several decades. Human activities are important in almost all ecosystems. Although humans exist and operate within ecosystems, their cumulative effects are large enough to influence external factors like climate. Ecosystems provide a variety of goods and services upon which people depend. Ecosystem goods include the "tangible, material products" of ecosystem processes such as food, construction material, medicinal plants. They also include less tangible items like tourism and recreation, and genes from wild plants and animals that can be used to improve domestic species. Ecosystem services, on the other hand, are generally "improvements in the condition or location of things of value". These include things like the maintenance of hydrological cycles, cleaning air and water, the maintenance of oxygen in the atmosphere, crop pollination and even things like beauty, inspiration and opportunities for research. While material from the ecosystem had traditionally been recognized as being the basis for things of economic value, ecosystem services tend to be taken for granted. When natural resource management is applied to whole ecosystems, rather than single species, it is termed ecosystem management. Although definitions of ecosystem management abound, there is a common set of principles which underlie these definitions. A fundamental principle is the long-term sustainability of the production of goods and services by the ecosystem; "intergenerational sustainability [is] a precondition for management, not an afterthought". While ecosystem management can be used as part of a plan for wilderness conservation, it can also be used in intensively managed ecosystems (see, for example, agroecosystem and close to nature forestry). As human population and per capita consumption grow, so do the resource demands imposed on ecosystems and the effects of the human ecological footprint. Natural resources are vulnerable and limited. The environmental impacts of anthropogenic actions are becoming more apparent. Problems for all ecosystems include: environmental pollution, climate change and biodiversity loss. For terrestrial ecosystems further threats include air pollution, soil degradation, and deforestation. For aquatic ecosystems threats include also unsustainable exploitation of marine resources (for example overfishing of certain species), marine pollution, microplastics pollution, water pollution, the warming of oceans, and building on coastal areas. Society is increasingly becoming aware that ecosystem services are not only limited but also that they are threatened by human activities. The need to better consider long-term ecosystem health and its role in enabling human habitation and economic activity is urgent. To help inform decision-makers, many ecosystem services are being assigned economic values, often based on the cost of replacement with anthropogenic alternatives. The ongoing challenge of prescribing economic value to nature, for example through biodiversity banking, is prompting transdisciplinary shifts in how we recognize and manage the environment, social responsibility, business opportunities, and our future as a species.
https://en.wikipedia.org/wiki?curid=9632
E (mathematical constant) The number is a mathematical constant approximately equal to 2.71828 and is the base of the natural logarithm, that is the unique number whose natural logarithm equals one. It is the limit of as approaches infinity, an expression that arises in the study of compound interest. It can also be calculated as the sum of the infinite series The constant can be characterized in many ways. For example, it can be defined as the unique positive number such that the graph of the function has unit slope at . The (natural) exponential function is the unique function which is equal to its own derivative and has initial value , and one may define . The natural logarithm, or logarithm to base , is the inverse function to the natural exponential function. The natural logarithm of a number can be defined directly as the area under the curve between and , in which case is the value of "k" for which this area equals one (see image). There are alternative characterizations. The number has eminent importance in mathematics, alongside 0, 1, , and . All five of these numbers play important and recurring roles across mathematics, and these five constants appear in one formulation of Euler's identity. Like the constant , is also irrational (i.e. it cannot be represented as ratio of integers) and transcendental (i.e. it is not a root of any non-zero polynomial with rational coefficients). The numerical value of truncated to 50 decimal places is The first references to the constant were published in 1618 in the table of an appendix of a work on logarithms by John Napier. However, this did not contain the constant itself, but simply a list of logarithms calculated from the constant. It is assumed that the table was written by William Oughtred. The discovery of the constant itself is credited to Jacob Bernoulli in 1683, who attempted to find the value of the following expression (which is in fact ): The first known use of the constant, represented by the letter , was in correspondence from Gottfried Leibniz to Christiaan Huygens in 1690 and 1691. Leonhard Euler introduced the letter as the base for natural logarithms, writing in a letter to Christian Goldbach on 25 November 1731. Euler started to use the letter for the constant in 1727 or 1728, in an unpublished paper on explosive forces in cannons, and the first appearance of in a publication was in Euler's "Mechanica" (1736). While in the subsequent years some researchers used the letter , the letter was more common and eventually became standard. In mathematics, the standard is to typeset the constant as "", in italics; the ISO 80000-2:2009 standard recommends typesetting constants in an upright style, but this has not been validated by the scientific community. Jacob Bernoulli discovered this constant in 1683 by studying a question about compound interest: If the interest is credited twice in the year, the interest rate for each 6 months will be 50%, so the initial $1 is multiplied by 1.5 twice, yielding at the end of the year. Compounding quarterly yields , and compounding monthly yields If there are compounding intervals, the interest for each interval will be and the value at the end of the year will be $1.00×. Bernoulli noticed that this sequence approaches a limit (the force of interest) with larger and, thus, smaller compounding intervals. Compounding weekly () yields $2.692597..., while compounding daily () yields $2.714567..., just two cents more. The limit as grows large is the number that came to be known as ; with "continuous" compounding, the account value will reach $2.7182818... More generally, an account that starts at $1 and offers an annual interest rate of will, after years, yield dollars with continuous compounding. (Here is the decimal equivalent of the rate of interest expressed as a percentage, so for 5% interest, .) The number itself also has applications to probability theory, where it arises in a way not obviously related to exponential growth. Suppose that a gambler plays a slot machine that pays out with a probability of one in and plays it times. Then, for large (such as a million) the probability that the gambler will lose every bet is approximately . For it is already approximately 1/2.79. This is an example of a Bernoulli trial process. Each time the gambler plays the slots, there is a one in one million chance of winning. Playing one million times is modelled by the binomial distribution, which is closely related to the binomial theorem and Pascal's triangle. The probability of winning times out of a million trials is: In particular, the probability of winning zero times () is This is very close to the limit The normal distribution with zero mean and unit standard deviation is known as the "standard normal distribution", given by the probability density function The constraint of unit variance (and thus also unit standard deviation) results in the in the exponent, and the constraint of unit total area under the curve results in the factor formula_11.[proof] This function is symmetric around , where it attains its maximum value formula_11, and has inflection points at . Another application of , also discovered in part by Jacob Bernoulli along with Pierre Raymond de Montmort, is in the problem of derangements, also known as the "hat check problem": guests are invited to a party, and at the door, the guests all check their hats with the butler, who in turn places the hats into boxes, each labelled with the name of one guest. But the butler has not asked the identities of the guests, and so he puts the hats into boxes selected at random. The problem of de Montmort is to find the probability that "none" of the hats gets put into the right box. The answer is: As the number of guests tends to infinity, approaches . Furthermore, the number of ways the hats can be placed into the boxes so that none of the hats are in the right box is rounded to the nearest integer, for every positive . A stick of length is broken into equal parts. The value of that maximizes the product of the lengths is then either The stated result follows because the maximum value of formula_16 occurs at formula_17 (Steiner's problem, discussed below). The quantity formula_16 is a measure of information gleaned from an event occurring with probability formula_19, so that essentially the same optimal division appears in optimal planning problems like the secretary problem. The number occurs naturally in connection with many problems involving asymptotics. An example is Stirling's formula for the asymptotics of the factorial function, in which both the numbers and enter: As a consequence, The principal motivation for introducing the number , particularly in calculus, is to perform differential and integral calculus with exponential functions and logarithms. A general exponential has a derivative, given by a limit: The parenthesized limit on the right is independent of the it depends only on the When the base is set this limit is equal and so is symbolically defined by the equation: Consequently, the exponential function with base is particularly suited to doing calculus. as opposed to some other number, as the base of the exponential function makes calculations involving the derivative much simpler. Another motivation comes from considering the derivative of the base- logarithm, i.e., of for : where the substitution was made. The -logarithm of is 1, if equals . So symbolically, The logarithm with this special base is called the natural logarithm and is denoted as ; it behaves well under differentiation since there is no undetermined limit to carry through the calculations. There are thus two ways in which to select such special numbers . One way is to set the derivative of the exponential function equal to , and solve for . The other way is to set the derivative of the base logarithm to and solve for . In each case, one arrives at a convenient choice of base for doing calculus. It turns out that these two solutions for are actually "the same", the number . Other characterizations of are also possible: one is as the limit of a sequence, another is as the sum of an infinite series, and still others rely on integral calculus. So far, the following two (equivalent) properties have been introduced: The following four characterizations can be proven equivalent: where is the factorial of . (By convention formula_29.) As in the motivation, the exponential function is important in part because it is the unique nontrivial function (up to multiplication by a constant) which is its own derivative and therefore its own antiderivative as well: The number is the unique real number such that for all positive . Also, we have the inequality for all real , with equality if and only if . Furthermore, is the unique base of the exponential for which the inequality holds for all . This is a limiting case of Bernoulli's inequality. Steiner's problem asks to find the global maximum for the function This maximum occurs precisely at . For proof, the inequality formula_38, from above, evaluated at formula_39 and simplifying gives formula_40. So formula_41 for all positive "x". Similarly, is where the global minimum occurs for the function defined for positive . More generally, for the function the global maximum for positive occurs at for any ; and the global minimum occurs at for any . The infinite tetration converges if and only if (or approximately between 0.0660 and 1.4447), due to a theorem of Leonhard Euler. The real number is irrational. Euler proved this by showing that its simple continued fraction expansion is infinite. (See also Fourier's proof that is irrational.) Furthermore, by the Lindemann–Weierstrass theorem, is transcendental, meaning that it is not a solution of any non-constant polynomial equation with rational coefficients. It was the first number to be proved transcendental without having been specifically constructed for this purpose (compare with Liouville number); the proof was given by Charles Hermite in 1873. It is conjectured that is normal, meaning that when is expressed in any base the possible digits in that base are uniformly distributed (occur with equal probability in any sequence of given length). The exponential function may be written as a Taylor series Because this series is convergent for every complex value of , it is commonly used to extend the definition of to the complex numbers. This, with the Taylor series for and, allows one to derive Euler's formula: which holds for every complex . The special case with is Euler's identity: from which it follows that, in the principal branch of the logarithm, Furthermore, using the laws for exponentiation, which is de Moivre's formula. The expression is sometimes referred to as . The expressions of formula_52 and formula_53 in terms of the exponential function can be deduced: The family of functions where is any real number, is the solution to the differential equation The number can be represented as a real number in a variety of ways: as an infinite series, an infinite product, a continued fraction, or a limit of a sequence. The chief among these representations, particularly in introductory calculus courses is the limit given above, as well as the series given by evaluating the above power series for at . Less common is the continued fraction which written out looks like This continued fraction for converges three times as quickly: Many other series, sequence, continued fraction, and infinite product representations of have been developed. In addition to exact analytical expressions for representation of , there are stochastic techniques for estimating . One such approach begins with an infinite sequence of independent random variables , ..., drawn from the uniform distribution on [0, 1]. Let be the least number such that the sum of the first observations exceeds 1: Then the expected value of is : . The number of known digits of has increased substantially during the last decades. This is due both to the increased performance of computers and to algorithmic improvements. Since around 2010, the proliferation of modern high-speed desktop computers has made it feasible for most amateurs to compute trillions of digits of "e" within acceptable amounts of time. It currently has been calculated to 8 trillion digits. During the emergence of internet culture, individuals and organizations sometimes paid homage to the number . In an early example, the computer scientist Donald Knuth let the version numbers of his program Metafont approach . The versions are 2, 2.7, 2.71, 2.718, and so forth. In another instance, the IPO filing for Google in 2004, rather than a typical round-number amount of money, the company announced its intention to raise 2,718,281,828 USD, which is billion dollars rounded to the nearest dollar. Google was also responsible for a billboard that appeared in the heart of Silicon Valley, and later in Cambridge, Massachusetts; Seattle, Washington; and Austin, Texas. It read "{first 10-digit prime found in consecutive digits of }.com". Solving this problem and visiting the advertised (now defunct) web site led to an even more difficult problem to solve, which in turn led to Google Labs where the visitor was invited to submit a résumé. The first 10-digit prime in is 7427466391, which starts at the 99th digit.
https://en.wikipedia.org/wiki?curid=9633
Euler–Maclaurin formula In mathematics, the Euler–Maclaurin formula is a formula for the difference between an integral and a closely related sum. It can be used to approximate integrals by finite sums, or conversely to evaluate finite sums and infinite series using integrals and the machinery of calculus. For example, many asymptotic expansions are derived from the formula, and Faulhaber's formula for the sum of powers is an immediate consequence. The formula was discovered independently by Leonhard Euler and Colin Maclaurin around 1735. Euler needed it to compute slowly converging infinite series while Maclaurin used it to calculate integrals. It was later generalized to Darboux's formula. If formula_1 and formula_2 are natural numbers and formula_3 is a real or complex valued continuous function for real numbers formula_4 in the interval formula_5, then the integral can be approximated by the sum (or vice versa) (see rectangle method). The Euler–Maclaurin formula provides expressions for the difference between the sum and the integral in terms of the higher derivatives formula_8 evaluated at the endpoints of the interval, that is to say when formula_9 and formula_10. Explicitly, for formula_11 a positive integer and a function formula_3 that is formula_11 times continuously differentiable on the interval formula_5, we have where formula_16 is the formula_17th Bernoulli number (with formula_18) and formula_19 is an error term which depends on formula_2, formula_1, formula_11, and formula_23 and is usually small for suitable values of formula_11. The formula is often written with the subscript taking only even values, since the odd Bernoulli numbers are zero except for formula_25. In this case we have or alternatively The remainder term arises because the integral is usually not exactly equal to the sum. The formula may be derived by applying repeated integration by parts to successive intervals formula_28 for formula_29. The boundary terms in these integrations lead to the main terms of the formula, and the leftover integrals form the remainder term. The remainder term has an exact expression in terms of the periodized Bernoulli functions formula_30. The Bernoulli polynomials may be defined recursively by formula_31 and, for formula_32, The periodized Bernoulli functions are defined as where formula_35 denotes the largest integer less than or equal to formula_4 (so that formula_37 always lies in the interval formula_38). With this notation, the remainder term formula_19 equals When formula_41, it can be shown that where formula_43 denotes the Riemann zeta function; one approach to prove this inequality is to obtain the Fourier series for the polynomials formula_44. The bound is achieved for even formula_17 when formula_4 is zero. The term formula_47 may be omitted for odd formula_17 but the proof in this case is more complex (see Lehmer). Using this inequality, the size of the remainder term can be estimated as The Bernoulli numbers from formula_25 to formula_51 are formula_52 Therefore the low-order cases of the Euler-Maclaurin formula are: The Basel problem is to determine the sum Euler computed this sum to 20 decimal places with only a few terms of the Euler–Maclaurin formula in 1735. This probably convinced him that the sum equals formula_55, which he proved in the same year. If formula_23 is a polynomial and formula_11 is big enough, then the remainder term vanishes. For instance, if formula_58, we can choose formula_59 to obtain, after simplification, The formula provides a means of approximating a finite integral. Let formula_61 be the endpoints of the interval of integration. Fix formula_62, the number of points to use in the approximation, and denote the corresponding step size by formula_63. Set formula_64, so that formula_65 and formula_66. Then: This may be viewed as an extension of the trapezoid rule by the inclusion of correction terms. Note that this asymptotic expansion is usually not convergent; there is some formula_11, depending upon formula_23 and formula_70, such that the terms past order formula_11 increase rapidly. Thus, the remainder term generally demands close attention. The Euler–Maclaurin formula is also used for detailed error analysis in numerical quadrature. It explains the superior performance of the trapezoidal rule on smooth periodic functions and is used in certain extrapolation methods. Clenshaw–Curtis quadrature is essentially a change of variables to cast an arbitrary integral in terms of integrals of periodic functions where the Euler–Maclaurin approach is very accurate (in that particular case the Euler–Maclaurin formula takes the form of a discrete cosine transform). This technique is known as a periodizing transformation. In the context of computing asymptotic expansions of sums and series, usually the most useful form of the Euler–Maclaurin formula is where formula_73 and formula_74 are integers. Often the expansion remains valid even after taking the limits formula_75 or formula_76 or both. In many cases the integral on the right-hand side can be evaluated in closed form in terms of elementary functions even though the sum on the left-hand side cannot. Then all the terms in the asymptotic series can be expressed in terms of elementary functions. For example, Here the left-hand side is equal to formula_78, namely the first-order polygamma function defined by formula_79; the gamma function formula_80 is equal to formula_81 if formula_82 is a positive integer. This results in an asymptotic expansion for formula_78. That expansion, in turn, serves as the starting point for one of the derivations of precise error estimates for Stirling's approximation of the factorial function. If is an integer greater than 1 we have: Collecting the constants into a value of the Riemann zeta function, we can write an asymptotic expansion: For equal to 2 this simplifies to or When , the corresponding technique gives an asymptotic expansion for the harmonic numbers: where formula_89 is the Euler–Mascheroni constant. We outline the argument given in Apostol. The Bernoulli polynomials and the periodic Bernoulli functions for were introduced above. The first several Bernoulli polynomials are The values are the Bernoulli numbers . Notice that for we have and for , The functions agree with the Bernoulli polynomials on the interval and are periodic with period 1. Furthermore, except when , they are also continuous. Thus, Let be an integer, and consider the integral where Integrating by parts, we get Using formula_97, formula_98, and summing the above from to , we get Adding ("f"("n") − "f"(0))/2 to both sides and rearranging, we have This is the case of the summation formula. To continue the induction, we apply integration by parts to the error term: where The result of integrating by parts is Summing from to and substituting this for the lower order error term results in the case of the formula, This process can be iterated. In this way we get a proof of the Euler–Maclaurin summation formula which can be formalized by mathematical induction, in which the induction step relies on integration by parts and on identities for periodic Bernoulli functions.
https://en.wikipedia.org/wiki?curid=9637
Epimenides paradox The Epimenides paradox reveals a problem with self-reference in logic. It is named after the Cretan philosopher Epimenides of Knossos (alive circa 600 BC) who is credited with the original statement. A typical description of the problem is given in the book "Gödel, Escher, Bach", by Douglas Hofstadter: A paradox of self-reference arises when one considers whether it is possible for Epimenides to have spoken the truth. Thomas Fowler (1869) states the paradox as follows: "Epimenides the Cretan says, 'that all the Cretans are liars,' but Epimenides is himself a Cretan; therefore he is himself a liar. But if he is a liar, what he says is untrue, and consequently, the Cretans are veracious; but Epimenides is a Cretan, and therefore what he says is true; saying the Cretans are liars, Epimenides is himself a liar, and what he says is untrue. Thus we may go on alternately proving that Epimenides and the Cretans are truthful and untruthful." The Epimenides paradox in this form, however, can be solved. There are two options: it is either true or false. First, assume that it is true, but then Epimenides, being a Cretan, would be a liar, and making the assumption that liars only make false statements, the statement is false. So, assuming the statement is true leads us to conclude that the statement is false. This is a contradiction, so the option of the statement being true is not possible. This leaves the second option: that it is false. If we assume the statement is false and that Epimenides is lying about all Cretans being liars, then there must exist at least one Cretan who is honest. This does not lead to a contradiction since it is not required that this Cretan be Epimenides. This means that Epimenides can say the false statement that all Cretans are liars while knowing at least one honest Cretan and lying about this particular Cretan. Hence, from the assumption that the statement is false, it does not follow that the statement is true. So we can avoid a paradox as seeing the statement "all Cretans are liars" as a false statement, which is made by a lying Cretan, Epimenides. The mistake made by Thomas Fowler (and many other people) above is to think that the negation of "all Cretans are liars" is "all Cretans are honest" (a paradox) when in fact the negation is "there exists a Cretan who is honest", or "not all Cretans are liars". The Epimenides paradox can be slightly modified as to not allow the kind of solution described above, as it was in the first paradox of Eubulides but instead leading to a non-avoidable self-contradiction. Paradoxical versions of the Epimenides problem are closely related to a class of more difficult logical problems, including the liar paradox, Socratic paradox, and the Burali-Forti paradox, all of which have self-reference in common with Epimenides. Indeed, the Epimenides paradox is usually classified as a variation on the liar paradox, and sometimes the two are not distinguished. The study of self-reference led to important developments in logic and mathematics in the twentieth century. In other words, it is not a paradox once one realizes "All Cretans are liars" being untrue only means "Not all Cretans are liars" instead of the assumption that "All Cretans are honest". Perhaps better put, for "All Cretans are liars" to be a true statement, it does not mean that all Cretans must lie all the time. In fact, Cretans could tell the truth quite often, but still all be liars in the sense that liars are people prone to deception for dishonest gain. Considering that “All Cretans are liars” has been seen as a paradox only since the 19th century, this seems to resolve the alleged paradox. Of course, if ‘all Cretans are continuous liars’ is actually true, then asking a Cretan if they are honest would always elicit the dishonest answer ‘yes’. So arguably the original proposition is not so much paradoxical as invalid. A contextual reading of the contradiction may also provide an answer to the paradox. The original phrase, "The Cretans, always liars, evil beasts, idle bellies!" asserts not an intrinsic paradox, but rather an opinion of the Cretans from Epimenides. A stereotyping of his people not intended to be an absolute statement about the people as a whole. Rather it is a claim made about their position regarding their religious beliefs and socio-cultural attitudes. Within the context of his poem the phrase is specific to a certain belief, a context that Callimachus repeats in his poem regarding Zeus. Further, a more poignant answer to the paradox is simply that to be a "liar" is to state falsehoods, nothing in the statement asserts everything said is false, but rather they're "always" lying. This is not an absolute statement of fact and thus we cannot conclude there's a true contradiction made by Epimenides with this statement. Epimenides was a 6th-century BC philosopher and religious prophet who, against the general sentiment of Crete, proposed that Zeus was immortal, as in the following poem: Denying the immortality of Zeus, then, was the lie of the Cretans. The phrase "Cretans, always liars" was quoted by the poet Callimachus in his "Hymn to Zeus", with the same theological intent as Epimenides: The logical inconsistency of a Cretan asserting all Cretans are always liars may not have occurred to Epimenides, nor to Callimachus, who both used the phrase to emphasize their point, without irony, perhaps meaning that all Cretans lie routinely, but not exclusively. In the 1st or 2nd century AD, the quote is mentioned by the author of the Epistle to Titus as having been spoken truly by "one of their own prophets." Clement of Alexandria, in the late 2nd century AD, fails to indicate that the concept of logical paradox is an issue: During the early 4th century, Saint Augustine restates the closely related liar paradox in "Against the Academicians" (III.13.29), but without mentioning Epimenides. In the Middle Ages, many forms of the liar paradox were studied under the heading of insolubilia, but these were not explicitly associated with Epimenides. Finally, in 1740, the second volume of Pierre Bayle's "Dictionnaire Historique et Critique" explicitly connects Epimenides with the paradox, though Bayle labels the paradox a "sophisme". All of the works of Epimenides are now lost, and known only through quotations by other authors. The quotation from the "Cretica" of Epimenides is given by R.N. Longenecker, "Acts of the Apostles", in volume 9 of "The Expositor's Bible Commentary", Frank E. Gaebelein, editor (Grand Rapids, Michigan: Zondervan Corporation, 1976–1984), page 476. Longenecker in turn cites M.D. Gibson, "Horae Semiticae X" (Cambridge: Cambridge University Press, 1913), page 40, "in Syriac". Longenecker states the following in a footnote: An oblique reference to Epimenides in the context of logic appears in "The Logical Calculus" by W. E. Johnson, "Mind" (New Series), volume 1, number 2 (April, 1892), pages 235–250. Johnson writes in a footnote, The Epimenides paradox appears explicitly in "Mathematical Logic as Based on the Theory of Types", by Bertrand Russell, in the "American Journal of Mathematics", volume 30, number 3 (July, 1908), pages 222–262, which opens with the following: In that article, Russell uses the Epimenides paradox as the point of departure for discussions of other problems, including the Burali-Forti paradox and the paradox now called Russell's paradox. Since Russell, the Epimenides paradox has been referenced repeatedly in logic. Typical of these references is "Gödel, Escher, Bach" by Douglas Hofstadter, which accords the paradox a prominent place in a discussion of self-reference.
https://en.wikipedia.org/wiki?curid=9638
Engine An engine or motor is a machine designed to convert one form of energy into mechanical energy. Heat engines, like the internal combustion engine, burn a fuel to create heat which is then used to do work. Electric motors convert electrical energy into mechanical motion, pneumatic motors use compressed air, and clockwork motors in wind-up toys use elastic energy. In biological systems, molecular motors, like myosins in muscles, use chemical energy to create forces and ultimately motion. The word "engine" derives from Old French "engin", from the Latin "ingenium"–the root of the word "ingenious". Pre-industrial weapons of war, such as catapults, trebuchets and battering rams, were called "siege engines", and knowledge of how to construct them was often treated as a military secret. The word "gin", as in "cotton gin", is short for "engine". Most mechanical devices invented during the industrial revolution were described as engines—the steam engine being a notable example. However, the original steam engines, such as those by Thomas Savery, were not mechanical engines but pumps. In this manner, a fire engine in its original form was merely a water pump, with the engine being transported to the fire by horses. In modern usage, the term "engine" typically describes devices, like steam engines and internal combustion engines, that burn or otherwise consume fuel to perform mechanical work by exerting a torque or linear force (usually in the form of thrust). Devices converting heat energy into motion are commonly referred to simply as "engines". Examples of engines which exert a torque include the familiar automobile gasoline and diesel engines, as well as turboshafts. Examples of engines which produce thrust include turbofans and rockets. When the internal combustion engine was invented, the term "motor" was initially used to distinguish it from the steam engine—which was in wide use at the time, powering locomotives and other vehicles such as steam rollers. The term "motor" derives from the Latin verb "" which means to set in motion, or maintain motion. Thus a motor is a device that imparts motion. "Motor" and "engine" are interchangeable in standard English.
https://en.wikipedia.org/wiki?curid=9640
European Environment Agency The European Environment Agency (EEA) is the agency of the European Union (EU) which provides independent information on the environment. The European Environment Agency (EEA) is the agency of the European Union (EU) which provides independent information on the environment. Its goal is to help those involved in developing, implementing and evaluating environmental policy, and to inform the general public. The EEA was established by the European Economic Community (EEC) Regulation 1210/1990 (amended by EEC Regulation 933/1999 and EC Regulation 401/2009) and became operational in 1994, headquartered in Copenhagen, Denmark. The agency is governed by a management board composed of representatives of the governments of its 33 member states, a European Commission representative and two scientists appointed by the European Parliament, assisted by a committee of scientists. The current Executive Director of the agency is Professor Hans Bruyninckx, who has been appointed for a five-year term. He is the successor of Professor Jacqueline McGlade. The member states of the union are members; however other states may become members of it by means of agreements concluded between them and the EU. It was the first EU body to open its membership to the 13 candidate countries (pre-2004 enlargement). The EEA has 33 member countries and six cooperating countries. The 33 member countries include the 27 European Union member states together with Iceland, Liechtenstein, Norway, Switzerland and Turkey. Since Brexit in 2020, the UK is not a member of the EU anymore and therefore not a member state of the EEA The six Western Balkan countries are cooperating countries: Albania, Bosnia and Herzegovina, Montenegro, North Macedonia, Serbia as well as Kosovo under the UN Security Council Resolution 1244/99. These cooperation activities are integrated into Eionet and are supported by the EU under the "Instrument for Pre-Accession Assistance". The EEA is an active member of the EPA Network. The European Environment Agency (EEA) reported in 2017 that climate-related extreme events accounted ca €400 billion ($430 billion) of economic losses in EEA area from 1980 to 2013, and were responsible for 85,000 deaths during 1980-2013. The European environment information and observation network (Eionet) is a partnership network of the EEA and the countries. The EEA is responsible for developing the network and coordinating its activities. To do so, the EEA works closely together with national focal points (NFP´s), typically national environment agencies or environment ministries which are responsible for coordinating national networksof the National Reference Centres (NRCs) involving many institutions (about 350 in all). Apart from the NFPs and NRCs, Eionet covers six European Topic Centres (ETCs) in the areas of air and climate change, biological diversity, climate change impacts, vulnerability and adaptation, water, land use and spatial information and analysis and sustainable consumption and production. In February 2012, the European Parliament's Committee on Budgetary Control published a draft report, identifying areas of concern in the use of funds and its influence for the 2010 budget such as a 26% budget increase from 2009 to 2010 to €50 600 000. and questioned that maximum competition and value-for-money principles were honored in hiring, also possible fictitious employees. The EEA's Executive Director refuted allegations of irregularities in a public hearing. On 27 March 2012 Members of the European Parliament (MEPs) voted on the report and commended the cooperation between the Agency and NGOs working in the environmental area. On 23 October 2012, the European Parliament voted and granted the discharge to the European Environment Agency for its 2010 budget. In April 2013, the MEPs voted and granted the discharge to the EEA for its 2011 budget. In addition to its 33 members and six Balkan cooperating countries, the EEA also cooperates and fosters partnerships with its neighbours and other countries and regions, mostly in the context of the European Neighbourhood Policy: Additionally the EEA cooperates with multiple international organizations and the corresponding agencies of the following countries: The 26 official languages used by the EEA are: Bulgarian, Czech, Croatian, Danish, German, Greek, English, Spanish, Estonian, Finnish, French, Hungarian, Icelandic, Italian, Lithuanian, Latvian, Malti, Dutch, Norwegian, Polish, Portuguese, Romanian, Slovak, Slovene, Swedish and Turkish.
https://en.wikipedia.org/wiki?curid=9644
Erlang (programming language) Erlang ( ) is a general-purpose, concurrent, functional programming language, and a garbage-collected runtime system. The term Erlang is used interchangeably with Erlang/OTP, or Open Telecom Platform (OTP), which consists of the Erlang runtime system, several ready-to-use components (OTP) mainly written in Erlang, and a set of design principles for Erlang programs. The Erlang runtime system is designed for systems with these traits: The Erlang programming language has immutable data, pattern matching, and functional programming. The sequential subset of the Erlang language supports eager evaluation, single assignment, and dynamic typing. It was originally proprietary software within Ericsson, developed by Joe Armstrong, Robert Virding, and Mike Williams in 1986, but was released as free and open-source software in 1998. Erlang/OTP is supported and maintained by the Open Telecom Platform (OTP) product unit at Ericsson. The name "Erlang", attributed to Bjarne Däcker, has been presumed by those working on the telephony switches (for whom the language was designed) to be a reference to Danish mathematician and engineer Agner Krarup Erlang and a syllabic abbreviation of "Ericsson Language". Erlang was designed with the aim of improving the development of telephony applications. The initial version of Erlang was implemented in Prolog and was influenced by the programming language PLEX used in earlier Ericsson exchanges. By 1988 Erlang had proven that it was suitable for prototyping telephone exchanges, but the Prolog interpreter was far too slow. One group within Ericsson estimated that it would need to be 40 times faster to be suitable for production use. In 1992, work began on the BEAM virtual machine (VM) which compiles Erlang to C using a mix of natively compiled code and threaded code to strike a balance between performance and disk space. According to Armstrong, the language went from lab product to real applications following the collapse of the next-generation AXE telephone exchange named in 1995. As a result, Erlang was chosen for the next asynchronous transfer mode (ATM) exchange "AXD". In 1998 Ericsson announced the AXD301 switch, containing over a million lines of Erlang and reported to achieve a high availability of nine "9"s. Shortly thereafter, Ericsson Radio Systems banned the in-house use of Erlang for new products, citing a preference for non-proprietary languages. The ban caused Armstrong and others to leave Ericsson. The implementation was open-sourced at the end of the year. Ericsson eventually lifted the ban and re-hired Armstrong in 2004. In 2006, native symmetric multiprocessing support was added to the runtime system and VM. Joe Armstrong, co-inventor of Erlang, summarized the principles of processes in his PhD thesis: Joe Armstrong remarked in an interview with Rackspace in 2013: "If Java is 'write once, run anywhere', then Erlang is 'write once, run forever'.” In 2014, Ericsson reported Erlang was being used in its support nodes, and in GPRS, 3G and LTE mobile networks worldwide and also by Nortel and T-Mobile. As Tim Bray, director of Web Technologies at Sun Microsystems, expressed in his keynote at O'Reilly Open Source Convention (OSCON) in July 2008: Erlang is the programming language used to code WhatsApp. Since being released as open source, Erlang has been spreading beyond Telecoms, establishing itself in other verticals such as FinTech, Gaming, Healthcare, Automotive, IoT and Blockchain. Apart from WhatsApp there are other companies listed as Erlang’s success stories: Vocalink (a MasterCard company), Goldman Sachs, Nintendo, AdRoll, Grindr, BT Mobile, Samsung, OpenX, SITA. A factorial algorithm implemented in Erlang: -module(fact). % This is the file 'fact.erl', the module and the filename must match -export([fac/1]). % This exports the function 'fac' of arity 1 (1 parameter, no type, no name) fac(0) -> 1; % If 0, then return 1, otherwise (note the semicolon ; meaning 'else') fac(N) when N > 0, is_integer(N) -> N * fac(N-1). % Recursively determine, then return the result % (note the period . meaning 'endif' or 'function end') %% This function will crash if anything other than a nonnegative integer is given. %% It illustrates the "Let it crash" philosophy of Erlang. A tail recursive algorithm that produces the Fibonacci sequence: %% The module declaration must match the file name "series.erl" -module(series). %% The export statement contains a list of all those functions that form %% the module's public API. In this case, this module exposes a single %% function called fib that takes 1 argument (I.E. has an arity of 1) %% The general syntax for -export is a list containing the name and %% arity of each public function -export([fib/1]). %% Public API %% Handle cases in which fib/1 receives specific values %% The order in which these function signatures are declared is a vital %% part of this module's functionality %% If fib/1 is passed precisely the integer 0, then return 0 fib(0) -> 0; %% If fib/1 receives a negative number, then return the atom err_neg_val %% Normally, such defensive coding is discouraged due to Erlang's 'Let %% it Crash' philosophy; however, in this case we should explicitly %% prevent a situation that will crash Erlang's runtime engine fib(N) when N < 0 -> err_neg_val; %% If fib/1 is passed an integer less than 3, then return 1 %% The preceding two function signatures handle all cases where N < 1, %% so this function signature handles cases where N = 1 or N = 2 fib(N) when N < 3 -> 1; %% For all other values, call the private function fib_int/3 to perform %% the calculation fib(N) -> fib_int(N, 0, 1). %% Private API %% If fib_int/3 receives a 1 as its first argument, then we're done, so %% return the value in argument B. Since we are not interested in the %% value of the second argument, we denote this using _ to indicate a %% "don't care" value fib_int(1, _, B) -> B; %% For all other argument combinations, recursively call fib_int/3 %% where each call does the following: %% - decrement counter N %% - Take the previous fibonacci value in argument B and pass it as %% argument A %% - Calculate the value of the current fibonacci number and pass it %% as argument B fib_int(N, A, B) -> fib_int(N-1, B, A+B). Here's the same program without the explanatory comments: -module(series). -export([fib/1]). fib(0) -> 0; fib(N) when N < 0 -> err_neg_val; fib(N) when N < 3 -> 1; fib(N) -> fib_int(N, 0, 1). fib_int(1, _, B) -> B; fib_int(N, A, B) -> fib_int(N-1, B, A+B). Quicksort in Erlang, using list comprehension: %% qsort:qsort(List) %% Sort a list of items -module(qsort). % This is the file 'qsort.erl' -export([qsort/1]). % A function 'qsort' with 1 parameter is exported (no type, no name) qsort([]) -> []; % If the list [] is empty, return an empty list (nothing to sort) qsort([Pivot|Rest]) -> The above example recursively invokes the function codice_1 until nothing remains to be sorted. The expression codice_2 is a list comprehension, meaning "Construct a list of elements codice_3 such that codice_3 is a member of codice_5, and codice_3 is less than codice_7." codice_8 is the list concatenation operator. A comparison function can be used for more complicated structures for the sake of readability. The following code would sort lists according to length: % This is file 'listsort.erl' (the compiler is made this way) -module(listsort). % Export 'by_length' with 1 parameter (don't care about the type and name) -export([by_length/1]). by_length(Lists) -> % Use 'qsort/2' and provides an anonymous function as a parameter qsort([], _)-> []; % If list is empty, return an empty list (ignore the second parameter) qsort([Pivot|Rest], Smaller) -> A codice_7 is taken from the first parameter given to codice_10 and the rest of codice_11 is named codice_5. Note that the expression [X || X length(A) < length(B) end The anonymous function is named codice_16 in the parameter list of the second definition of codice_1 so that it can be referenced by that name within that function. It is not named in the first definition of codice_1, which deals with the base case of an empty list and thus has no need of this function, let alone a name for it. Erlang has eight primitive data types: And three compound data types: Two forms of syntactic sugar are provided: Erlang has no method to define classes, although there are external libraries available. In most other programming languages, software crashes have always been (and often still are) considered highly undesirable situations that must be avoided at all costs. Consequently, elaborate exception handling mechanisms exist to trap these situations and then mitigate their effects. This design philosophy exists because many of the foundational principles of software design were defined at a time when computers were single processor machines. Under these conditions, software crashes were indeed fatal. Given this basic constraint, it was perfectly natural therefore to develop programming styles in which a large proportion of the code was dedicated to detecting and then handling error situations. This in turn, led directly to the still widely popular coding style known as defensive programming. However, the designers of Erlang realised that in spite of their undesirable effects, software crashes are much like death and taxes - quite unavoidable. Therefore, rather than treating a crash as a crisis situation that temporarily suspends all normal work until a solution is found, they reasoned it would make far greater sense to treat a crash in exactly the same manner as any other normal runtime event. Consequently, when an Erlang process crashes, this situation is reported as just another type of message arriving in a process' mail box. This realisation led to Erlang's designers building a language with the following fundamental features: The "Let it Crash" style of coding is therefore the practical consequence of working in a language that operates on these principles. A typical Erlang application is written in the form of a supervisor tree. This architecture is based on a hierarchy of processes in which the top level process is known as a "supervisor". The supervisor then spawns multiple child processes that act either as workers or more, lower level supervisors. Such hierarchies can exist to arbitrary depths and have proven to provide a highly scalable and fault-tolerant environment within which application functionality can be implemented. Within a supervisor tree, all supervisor processes are responsible for managing the lifecycle of their child processes, and this includes handling situations in which those child processes crash. Any process can become a supervisor by first spawning a child process, then calling codice_35 on that process. If the monitored process then crashes, the supervisor will receive a message containing a tuple whose first member is the atom codice_36. The supervisor is responsible firstly for listening for such messages and secondly, for taking the appropriate action to correct the error condition. In addition, "Let it Crash" results in a style of coding that contains little defensive code, resulting in smaller applications. Erlang's main strength is support for concurrency. It has a small but powerful set of primitives to create processes and communicate among them. Erlang is conceptually similar to the language occam, though it recasts the ideas of communicating sequential processes (CSP) in a functional framework and uses asynchronous message passing. Processes are the primary means to structure an Erlang application. They are neither operating system processes nor threads, but lightweight processes that are scheduled by BEAM. Like operating system processes (but unlike operating system threads), they share no state with each other. The estimated minimal overhead for each is 300 words. Thus, many processes can be created without degrading performance. In 2005, a benchmark with 20 million processes was successfully performed with 64-bit Erlang on a machine with 16 GB random-access memory (RAM; total 800 bytes/process). Erlang has supported symmetric multiprocessing since release R11B of May 2006. While threads need external library support in most languages, Erlang provides language-level features to create and manage processes with the goal of simplifying concurrent programming. Though all concurrency is explicit in Erlang, processes communicate using message passing instead of shared variables, which removes the need for explicit locks (a locking scheme is still used internally by the VM). Inter-process communication works via a shared-nothing asynchronous message passing system: every process has a "mailbox", a queue of messages that have been sent by other processes and not yet consumed. A process uses the codice_37 primitive to retrieve messages that match desired patterns. A message-handling routine tests messages in turn against each pattern, until one of them matches. When the message is consumed and removed from the mailbox the process resumes execution. A message may comprise any Erlang structure, including primitives (integers, floats, characters, atoms), tuples, lists, and functions. The code example below shows the built-in support for distributed processes: As the example shows, processes may be created on remote nodes, and communication with them is transparent in the sense that communication with remote processes works exactly as communication with local processes. Concurrency supports the primary method of error-handling in Erlang. When a process crashes, it neatly exits and sends a message to the controlling process which can then take action, such as for instance starting a new process that takes over the old process's task. The official reference implementation of Erlang uses BEAM. BEAM is included in the official distribution of Erlang, called Erlang/OTP. BEAM executes bytecode which is converted to threaded code at load time. It also includes a native code compiler on most platforms, developed by the High Performance Erlang Project (HiPE) at Uppsala University. Since October 2001 the HiPE system is fully integrated in Ericsson's Open Source Erlang/OTP system. It also supports interpreting, directly from source code via abstract syntax tree, via script as of R11B-5 release of Erlang. Erlang supports language-level Dynamic Software Updating. To implement this, code is loaded and managed as "module" units; the module is a compilation unit. The system can keep two versions of a module in memory at the same time, and processes can concurrently run code from each. The versions are referred to as the "new" and the "old" version. A process will not move into the new version until it makes an external call to its module. An example of the mechanism of hot code loading: For the second version, we add the possibility to reset the count to zero. Only when receiving a message consisting of the atom codice_38 will the loop execute an external call to codeswitch/1 (codice_39 is a preprocessor macro for the current module). If there is a new version of the "counter" module in memory, then its codeswitch/1 function will be called. The practice of having a specific entry-point into a new version allows the programmer to transform state to what is needed in the newer version. In the example, the state is kept as an integer. In practice, systems are built up using design principles from the Open Telecom Platform, which leads to more code upgradable designs. Successful hot code loading is exacting. Code must be written with care to make use of Erlang's facilities. In 1998, Ericsson released Erlang as free and open-source software to ensure its independence from a single vendor and to increase awareness of the language. Erlang, together with libraries and the real-time distributed database Mnesia, forms the OTP collection of libraries. Ericsson and a few other companies support Erlang commercially. Since the open source release, Erlang has been used by several firms worldwide, including Nortel and T-Mobile. Although Erlang was designed to fill a niche and has remained an obscure language for most of its existence, its popularity is growing due to demand for concurrent services. Erlang has found some use in fielding massively multiplayer online role-playing game (MMORPG) servers.
https://en.wikipedia.org/wiki?curid=9646
Euphoria (programming language) Euphoria is a programming language originally created by Robert Craig of Rapid Deployment Software in Toronto, Ontario, Canada. Initially developed (though not publicly released) on the Atari ST, the first commercial release was for the 16-bit DOS platform and was proprietary. In 2006, with the release of version 3, Euphoria became open-source software. The openEuphoria Group continues to administer and develop the project. In December 2010, the openEuphoria Group released version 4 of openEuphoria along with a new identity and mascot for the project. OpenEuphoria is currently available for Windows, Linux, macOS and three flavors of *BSD. Euphoria is a general-purpose high-level imperative-procedural interpreted language. A translator generates C source code and the GNU compiler collection (GCC) and Open Watcom compilers are supported. Alternatively, Euphoria programs may be bound with the interpreter to create stand-alone executables. A number of graphical user interface (GUI) libraries are supported including Win32lib and wrappers for wxWidgets, GTK+ and IUP. Euphoria has a simple built-in database and wrappers for a variety of other databases. The Euphoria language is a general purpose procedural language that focuses on simplicity, legibility, rapid development and performance via several means. Developed as a personal project to invent a programming language from scratch, Euphoria was created by Robert Craig on an Atari Mega-ST. Many design ideas for the language came from Craig's Master's thesis in computer science at the University of Toronto. Craig's thesis was heavily influenced by the work of John Backus on functional programming (FP) languages. Craig ported his original Atari implementation to the 16-bit DOS platform and Euphoria was first released, version 1.0, in July 1993 under a proprietary licence. The original Atari implementation is described by Craig as "primitive" and has not been publicly released. Euphoria continued to be developed and released by Craig via his company Rapid Deployment Software (RDS) and website rapideuphoria.com. In October 2006 RDS released version 3 of Euphoria and announced that henceforth Euphoria would be freely distributed under an open-source software licence. RDS continued to develop Euphoria, culminating with the release of version 3.1.1 in August, 2007. Subsequently, RDS ceased unilateral development of Euphoria and the openEuphoria Group took over ongoing development. The openEuphoria Group released version 4 in December, 2010 along with a new logo and mascot for the openEuphoria project. Version 3.1.1 remains an important milestone release, being the last version of Euphoria which supports the DOS platform. Euphoria is an acronym for "End-User Programming with Hierarchical Objects for Robust Interpreted Applications" although there is some suspicion that this is a backronym. The Euphoria interpreter was originally written in C. With the release of version 2.5 in November 2004 the Euphoria interpreter was split into two parts: a front-end parser, and a back-end interpreter. The front-end is now written in Euphoria (and used with the Euphoria-to-C translator and the Binder). The main back-end and run time library are written in C. Euphoria was conceived and developed with the following design goals and features: Euphoria is designed to readily facilitate handling of dynamic sets of data of varying types and is particularly useful for string and image processing. Euphoria has been used in artificial intelligence experiments, the study of mathematics, for teaching programming, and to implement fonts involving thousands of characters. A large part of the Euphoria interpreter is written in Euphoria. Euphoria has two basic data types: Euphoria has two additional data types predefined: There is no character string data type. Strings are represented by a "sequence" of "integer" values. However, because literal strings are so commonly used in programming, Euphoria interprets double-quote enclosed characters as a sequence of integers. Thus is seen as if the coder had written: which is the same as: puts(1, "Hello, World!\n") Program comments start with a double hyphen codice_1 and go through the end of line. The following code looks for an old item in a group of items. If found, it removes it by concatenating all the elements before it with all the elements after it. Note that the first element in a sequence has the index one [1] and that $ refers to the length (i.e., total number of elements) of the sequence. The following modification to the above example replaces an old item with a new item. As the variables "old" and "new" have been defined as objects, they could be "atoms" or "sequences". Type checking is not needed as the function will work with any sequence of data of any type and needs no external libraries. Furthermore, no pointers are involved and subscripts are automatically checked. Thus the function cannot access memory out-of-bounds. There is no need to allocate or deallocate memory explicitly and no chance of a memory leak. The line shows some of the "sequence" handling facilities. A "sequence" may contain a set of any types, and this can be sliced (to take a subset of the data in a "sequence") and concatenated in expressions with no need for special functions. Arguments to routines are always passed by value; there is no pass-by-reference facility. However, parameters are allowed to be modified "locally" (i.e., within the callee) which is implemented very efficiently as sequences have automatic copy-on-write semantics. In other words, when you pass a sequence to a routine, initially only a reference to it is passed, but at the point the routine modifies this sequence parameter the sequence is copied and the routine updates only a copy of the original. Free downloads of Euphoria for the various platforms, packages, Windows IDE, Windows API libraries, a cross-platform GTK3 wrapper for Linux and Windows, graphics libraries (DOS, OpenGL, etc.).
https://en.wikipedia.org/wiki?curid=9647
Energy In physics, energy is the quantitative property that must be transferred to an object in order to perform work on, or to heat, the object. Energy is a conserved quantity; the law of conservation of energy states that energy can be converted in form, but not created or destroyed. The SI unit of energy is the joule, which is the energy transferred to an object by the work of moving it a distance of 1 metre against a force of 1 newton. Common forms of energy include the kinetic energy of a moving object, the potential energy stored by an object's position in a force field (gravitational, electric or magnetic), the elastic energy stored by stretching solid objects, the chemical energy released when a fuel burns, the radiant energy carried by light, and the thermal energy due to an object's temperature. Mass and energy are closely related. Due to mass–energy equivalence, any object that has mass when stationary (called rest mass) also has an equivalent amount of energy whose form is called rest energy, and any additional energy (of any form) acquired by the object above that rest energy will increase the object's total mass just as it increases its total energy. For example, after heating an object, its increase in energy could be measured as a small increase in mass, with a sensitive enough scale. Living organisms require energy to stay alive, such as the energy humans get from food. Human civilization requires energy to function, which it gets from energy resources such as fossil fuels, nuclear fuel, or renewable energy. The processes of Earth's climate and ecosystem are driven by the radiant energy Earth receives from the sun and the geothermal energy contained within the earth. The total energy of a system can be subdivided and classified into potential energy, kinetic energy, or combinations of the two in various ways. Kinetic energy is determined by the movement of an object – or the composite motion of the components of an object – and potential energy reflects the potential of an object to have motion, and generally is a function of the position of an object within a field or may be stored in the field itself. While these two categories are sufficient to describe all forms of energy, it is often convenient to refer to particular combinations of potential and kinetic energy as its own form. For example, macroscopic mechanical energy is the sum of translational and rotational kinetic and potential energy in a system neglects the kinetic energy due to temperature, and nuclear energy which combines utilize potentials from the nuclear force and the weak force), among others. The word "energy" derives from the , which possibly appears for the first time in the work of Aristotle in the 4th century BC. In contrast to the modern definition, energeia was a qualitative philosophical concept, broad enough to include ideas such as happiness and pleasure. In the late 17th century, Gottfried Leibniz proposed the idea of the , or living force, which defined as the product of the mass of an object and its velocity squared; he believed that total "vis viva" was conserved. To account for slowing due to friction, Leibniz theorized that thermal energy consisted of the random motion of the constituent parts of matter, although it would be more than a century until this was generally accepted. The modern analog of this property, kinetic energy, differs from "vis viva" only by a factor of two. In 1807, Thomas Young was possibly the first to use the term "energy" instead of "vis viva", in its modern sense. Gustave-Gaspard Coriolis described "kinetic energy" in 1829 in its modern sense, and in 1853, William Rankine coined the term "potential energy". The law of conservation of energy was also first postulated in the early 19th century, and applies to any isolated system. It was argued for some years whether heat was a physical substance, dubbed the caloric, or merely a physical quantity, such as momentum. In 1845 James Prescott Joule discovered the link between mechanical work and the generation of heat. These developments led to the theory of conservation of energy, formalized largely by William Thomson (Lord Kelvin) as the field of thermodynamics. Thermodynamics aided the rapid development of explanations of chemical processes by Rudolf Clausius, Josiah Willard Gibbs, and Walther Nernst. It also led to a mathematical formulation of the concept of entropy by Clausius and to the introduction of laws of radiant energy by Jožef Stefan. According to Noether's theorem, the conservation of energy is a consequence of the fact that the laws of physics do not change over time. Thus, since 1918, theorists have understood that the law of conservation of energy is the direct mathematical consequence of the translational symmetry of the quantity conjugate to energy, namely time. In 1843, Joule independently discovered the mechanical equivalent in a series of experiments. The most famous of them used the "Joule apparatus": a descending weight, attached to a string, caused rotation of a paddle immersed in water, practically insulated from heat transfer. It showed that the gravitational potential energy lost by the weight in descending was equal to the internal energy gained by the water through friction with the paddle. In the International System of Units (SI), the unit of energy is the joule, named after James Prescott Joule. It is a derived unit. It is equal to the energy expended (or work done) in applying a force of one newton through a distance of one metre. However energy is also expressed in many other units not part of the SI, such as ergs, calories, British Thermal Units, kilowatt-hours and kilocalories, which require a conversion factor when expressed in SI units. The SI unit of energy rate (energy per unit time) is the watt, which is a joule per second. Thus, one joule is one watt-second, and 3600 joules equal one watt-hour. The CGS energy unit is the erg and the imperial and US customary unit is the foot pound. Other energy units such as the electronvolt, food calorie or thermodynamic kcal (based on the temperature change of water in a heating process), and BTU are used in specific areas of science and commerce. In classical mechanics, energy is a conceptually and mathematically useful property, as it is a conserved quantity. Several formulations of mechanics have been developed using energy as a core concept. Work, a function of energy, is force times distance. This says that the work (formula_2) is equal to the line integral of the force F along a path "C"; for details see the mechanical work article. Work and thus energy is frame dependent. For example, consider a ball being hit by a bat. In the center-of-mass reference frame, the bat does no work on the ball. But, in the reference frame of the person swinging the bat, considerable work is done on the ball. The total energy of a system is sometimes called the Hamiltonian, after William Rowan Hamilton. The classical equations of motion can be written in terms of the Hamiltonian, even for highly complex or abstract systems. These classical equations have remarkably direct analogs in nonrelativistic quantum mechanics. Another energy-related concept is called the Lagrangian, after Joseph-Louis Lagrange. This formalism is as fundamental as the Hamiltonian, and both can be used to derive the equations of motion or be derived from them. It was invented in the context of classical mechanics, but is generally useful in modern physics. The Lagrangian is defined as the kinetic energy "minus" the potential energy. Usually, the Lagrange formalism is mathematically more convenient than the Hamiltonian for non-conservative systems (such as systems with friction). Noether's theorem (1918) states that any differentiable symmetry of the action of a physical system has a corresponding conservation law. Noether's theorem has become a fundamental tool of modern theoretical physics and the calculus of variations. A generalisation of the seminal formulations on constants of motion in Lagrangian and Hamiltonian mechanics (1788 and 1833, respectively), it does not apply to systems that cannot be modeled with a Lagrangian; for example, dissipative systems with continuous symmetries need not have a corresponding conservation law. In the context of chemistry, energy is an attribute of a substance as a consequence of its atomic, molecular or aggregate structure. Since a chemical transformation is accompanied by a change in one or more of these kinds of structure, it is invariably accompanied by an increase or decrease of energy of the substances involved. Some energy is transferred between the surroundings and the reactants of the reaction in the form of heat or light; thus the products of a reaction may have more or less energy than the reactants. A reaction is said to be exothermic or exergonic if the final state is lower on the energy scale than the initial state; in the case of endothermic reactions the situation is the reverse. Chemical reactions are almost invariably not possible unless the reactants surmount an energy barrier known as the activation energy. The "speed" of a chemical reaction (at given temperature "T") is related to the activation energy "E" by the Boltzmann's population factor e−"E"/"kT"that is the probability of molecule to have energy greater than or equal to "E" at the given temperature "T". This exponential dependence of a reaction rate on temperature is known as the Arrhenius equation. The activation energy necessary for a chemical reaction can be provided in the form of thermal energy. In biology, energy is an attribute of all biological systems from the biosphere to the smallest living organism. Within an organism it is responsible for growth and development of a biological cell or an organelle of a biological organism. Energy used in respiration is mostly stored in molecular oxygen and can be unlocked by reactions with molecules of substances such as carbohydrates (including sugars), lipids, and proteins stored by cells. In human terms, the human equivalent (H-e) (Human energy conversion) indicates, for a given amount of energy expenditure, the relative quantity of energy needed for human metabolism, assuming an average human energy expenditure of 12,500 kJ per day and a basal metabolic rate of 80 watts. For example, if our bodies run (on average) at 80 watts, then a light bulb running at 100 watts is running at 1.25 human equivalents (100 ÷ 80) i.e. 1.25 H-e. For a difficult task of only a few seconds' duration, a person can put out thousands of watts, many times the 746 watts in one official horsepower. For tasks lasting a few minutes, a fit human can generate perhaps 1,000 watts. For an activity that must be sustained for an hour, output drops to around 300; for an activity kept up all day, 150 watts is about the maximum. The human equivalent assists understanding of energy flows in physical and biological systems by expressing energy units in human terms: it provides a "feel" for the use of a given amount of energy. Sunlight's radiant energy is also captured by plants as "chemical potential energy" in photosynthesis, when carbon dioxide and water (two low-energy compounds) are converted into carbohydrates, lipids, and proteins and high-energy compounds like oxygen and ATP. Carbohydrates, lipids, and proteins can release the energy of oxygen, which is utilized by living organisms as an electron acceptor. Release of the energy stored during photosynthesis as heat or light may be triggered suddenly by a spark, in a forest fire, or it may be made available more slowly for animal or human metabolism, when organic molecules are ingested, and catabolism is triggered by enzyme action. Any living organism relies on an external source of energy – radiant energy from the Sun in the case of green plants, chemical energy in some form in the case of animals – to be able to grow and reproduce. The daily 1500–2000 Calories (6–8 MJ) recommended for a human adult are taken as a combination of oxygen and food molecules, the latter mostly carbohydrates and fats, of which glucose (C6H12O6) and stearin (C57H110O6) are convenient examples. The food molecules are oxidised to carbon dioxide and water in the mitochondria and some of the energy is used to convert ADP into ATP. The rest of the chemical energy in O2 and the carbohydrate or fat is converted into heat: the ATP is used as a sort of "energy currency", and some of the chemical energy it contains is used for other metabolism when ATP reacts with OH groups and eventually splits into ADP and phosphate (at each stage of a metabolic pathway, some chemical energy is converted into heat). Only a tiny fraction of the original chemical energy is used for work: It would appear that living organisms are remarkably inefficient (in the physical sense) in their use of the energy they receive (chemical or radiant energy), and it is true that most real machines manage higher efficiencies. In growing organisms the energy that is converted to heat serves a vital purpose, as it allows the organism tissue to be highly ordered with regard to the molecules it is built from. The second law of thermodynamics states that energy (and matter) tends to become more evenly spread out across the universe: to concentrate energy (or matter) in one specific place, it is necessary to spread out a greater amount of energy (as heat) across the remainder of the universe ("the surroundings"). Simpler organisms can achieve higher energy efficiencies than more complex ones, but the complex organisms can occupy ecological niches that are not available to their simpler brethren. The conversion of a portion of the chemical energy to heat at each step in a metabolic pathway is the physical reason behind the pyramid of biomass observed in ecology: to take just the first step in the food chain, of the estimated 124.7 Pg/a of carbon that is fixed by photosynthesis, 64.3 Pg/a (52%) are used for the metabolism of green plants, i.e. reconverted into carbon dioxide and heat. In geology, continental drift, mountain ranges, volcanoes, and earthquakes are phenomena that can be explained in terms of energy transformations in the Earth's interior, while meteorological phenomena like wind, rain, hail, snow, lightning, tornadoes and hurricanes are all a result of energy transformations brought about by solar energy on the atmosphere of the planet Earth. Sunlight may be stored as gravitational potential energy after it strikes the Earth, as (for example) water evaporates from oceans and is deposited upon mountains (where, after being released at a hydroelectric dam, it can be used to drive turbines or generators to produce electricity). Sunlight also drives many weather phenomena, save those generated by volcanic events. An example of a solar-mediated weather event is a hurricane, which occurs when large unstable areas of warm ocean, heated over months, give up some of their thermal energy suddenly to power a few days of violent air movement. In a slower process, radioactive decay of atoms in the core of the Earth releases heat. This thermal energy drives plate tectonics and may lift mountains, via orogenesis. This slow lifting represents a kind of gravitational potential energy storage of the thermal energy, which may be later released to active kinetic energy in landslides, after a triggering event. Earthquakes also release stored elastic potential energy in rocks, a store that has been produced ultimately from the same radioactive heat sources. Thus, according to present understanding, familiar events such as landslides and earthquakes release energy that has been stored as potential energy in the Earth's gravitational field or elastic strain (mechanical potential energy) in rocks. Prior to this, they represent release of energy that has been stored in heavy atoms since the collapse of long-destroyed supernova stars created these atoms. In cosmology and astronomy the phenomena of stars, nova, supernova, quasars and gamma-ray bursts are the universe's highest-output energy transformations of matter. All stellar phenomena (including solar activity) are driven by various kinds of energy transformations. Energy in such transformations is either from gravitational collapse of matter (usually molecular hydrogen) into various classes of astronomical objects (stars, black holes, etc.), or from nuclear fusion (of lighter elements, primarily hydrogen). The nuclear fusion of hydrogen in the Sun also releases another store of potential energy which was created at the time of the Big Bang. At that time, according to theory, space expanded and the universe cooled too rapidly for hydrogen to completely fuse into heavier elements. This meant that hydrogen represents a store of potential energy that can be released by fusion. Such a fusion process is triggered by heat and pressure generated from gravitational collapse of hydrogen clouds when they produce stars, and some of the fusion energy is then transformed into sunlight. In quantum mechanics, energy is defined in terms of the energy operator as a time derivative of the wave function. The Schrödinger equation equates the energy operator to the full energy of a particle or a system. Its results can be considered as a definition of measurement of energy in quantum mechanics. The Schrödinger equation describes the space- and time-dependence of a slowly changing (non-relativistic) wave function of quantum systems. The solution of this equation for a bound system is discrete (a set of permitted states, each characterized by an energy level) which results in the concept of quanta. In the solution of the Schrödinger equation for any oscillator (vibrator) and for electromagnetic waves in a vacuum, the resulting energy states are related to the frequency by Planck's relation: formula_3 (where formula_4 is Planck's constant and formula_5 the frequency). In the case of an electromagnetic wave these energy states are called quanta of light or photons. When calculating kinetic energy (work to accelerate a massive body from zero speed to some finite speed) relativistically – using Lorentz transformations instead of Newtonian mechanics – Einstein discovered an unexpected by-product of these calculations to be an energy term which does not vanish at zero speed. He called it rest energy: energy which every massive body must possess even when being at rest. The amount of energy is directly proportional to the mass of the body: where For example, consider electron–positron annihilation, in which the rest energy of these two individual particles (equivalent to their rest mass) is converted to the radiant energy of the photons produced in the process. In this system the matter and antimatter (electrons and positrons) are destroyed and changed to non-matter (the photons). However, the total mass and total energy do not change during this interaction. The photons each have no rest mass but nonetheless have radiant energy which exhibits the same inertia as did the two original particles. This is a reversible process – the inverse process is called pair creation – in which the rest mass of particles is created from the radiant energy of two (or more) annihilating photons. In general relativity, the stress–energy tensor serves as the source term for the gravitational field, in rough analogy to the way mass serves as the source term in the non-relativistic Newtonian approximation. Energy and mass are manifestations of one and the same underlying physical property of a system. This property is responsible for the inertia and strength of gravitational interaction of the system ("mass manifestations"), and is also responsible for the potential ability of the system to perform work or heating ("energy manifestations"), subject to the limitations of other physical laws. In classical physics, energy is a scalar quantity, the canonical conjugate to time. In special relativity energy is also a scalar (although not a Lorentz scalar but a time component of the energy–momentum 4-vector). In other words, energy is invariant with respect to rotations of space, but not invariant with respect to rotations of space-time (= boosts). Energy may be transformed between different forms at various efficiencies. Items that transform between these forms are called transducers. Examples of transducers include a battery, from chemical energy to electric energy; a dam: gravitational potential energy to kinetic energy of moving water (and the blades of a turbine) and ultimately to electric energy through an electric generator; or a heat engine, from heat to work. Examples of energy transformation include generating electric energy from heat energy via a steam turbine, or lifting an object against gravity using electrical energy driving a crane motor. Lifting against gravity performs mechanical work on the object and stores gravitational potential energy in the object. If the object falls to the ground, gravity does mechanical work on the object which transforms the potential energy in the gravitational field to the kinetic energy released as heat on impact with the ground. Our Sun transforms nuclear potential energy to other forms of energy; its total mass does not decrease due to that in itself (since it still contains the same total energy even if in different forms), but its mass does decrease when the energy escapes out to its surroundings, largely as radiant energy. There are strict limits to how efficiently heat can be converted into work in a cyclic process, e.g. in a heat engine, as described by Carnot's theorem and the second law of thermodynamics. However, some energy transformations can be quite efficient. The direction of transformations in energy (what kind of energy is transformed to what other kind) is often determined by entropy (equal energy spread among all available degrees of freedom) considerations. In practice all energy transformations are permitted on a small scale, but certain larger transformations are not permitted because it is statistically unlikely that energy or matter will randomly move into more concentrated forms or smaller spaces. Energy transformations in the universe over time are characterized by various kinds of potential energy that has been available since the Big Bang later being "released" (transformed to more active types of energy such as kinetic or radiant energy) when a triggering mechanism is available. Familiar examples of such processes include nuclear decay, in which energy is released that was originally "stored" in heavy isotopes (such as uranium and thorium), by nucleosynthesis, a process ultimately using the gravitational potential energy released from the gravitational collapse of supernovae, to store energy in the creation of these heavy elements before they were incorporated into the solar system and the Earth. This energy is triggered and released in nuclear fission bombs or in civil nuclear power generation. Similarly, in the case of a chemical explosion, chemical potential energy is transformed to kinetic energy and thermal energy in a very short time. Yet another example is that of a pendulum. At its highest points the kinetic energy is zero and the gravitational potential energy is at maximum. At its lowest point the kinetic energy is at maximum and is equal to the decrease of potential energy. If one (unrealistically) assumes that there is no friction or other losses, the conversion of energy between these processes would be perfect, and the pendulum would continue swinging forever. Energy is also transferred from potential energy (formula_8) to kinetic energy (formula_9) and then back to potential energy constantly. This is referred to as conservation of energy. In this closed system, energy cannot be created or destroyed; therefore, the initial energy and the final energy will be equal to each other. This can be demonstrated by the following: The equation can then be simplified further since formula_10 (mass times acceleration due to gravity times the height) and formula_11 (half mass times velocity squared). Then the total amount of energy can be found by adding formula_12. Energy gives rise to weight when it is trapped in a system with zero momentum, where it can be weighed. It is also equivalent to mass, and this mass is always associated with it. Mass is also equivalent to a certain amount of energy, and likewise always appears associated with it, as described in mass-energy equivalence. The formula "E" = "mc"², derived by Albert Einstein (1905) quantifies the relationship between rest-mass and rest-energy within the concept of special relativity. In different theoretical frameworks, similar formulas were derived by J.J. Thomson (1881), Henri Poincaré (1900), Friedrich Hasenöhrl (1904) and others (see Mass-energy equivalence#History for further information). Part of the rest energy (equivalent to rest mass) of matter may be converted to other forms of energy (still exhibiting mass), but neither energy nor mass can be destroyed; rather, both remain constant during any process. However, since formula_13 is extremely large relative to ordinary human scales, the conversion of an everyday amount of rest mass (for example, 1 kg) from rest energy to other forms of energy (such as kinetic energy, thermal energy, or the radiant energy carried by light and other radiation) can liberate tremendous amounts of energy (~formula_14 joules = 21 megatons of TNT), as can be seen in nuclear reactors and nuclear weapons. Conversely, the mass equivalent of an everyday amount energy is minuscule, which is why a loss of energy (loss of mass) from most systems is difficult to measure on a weighing scale, unless the energy loss is very large. Examples of large transformations between rest energy (of matter) and other forms of energy (e.g., kinetic energy into particles with rest mass) are found in nuclear physics and particle physics. Thermodynamics divides energy transformation into two kinds: reversible processes and irreversible processes. An irreversible process is one in which energy is dissipated (spread) into empty energy states available in a volume, from which it cannot be recovered into more concentrated forms (fewer quantum states), without degradation of even more energy. A reversible process is one in which this sort of dissipation does not happen. For example, conversion of energy from one type of potential field to another, is reversible, as in the pendulum system described above. In processes where heat is generated, quantum states of lower energy, present as possible excitations in fields between atoms, act as a reservoir for part of the energy, from which it cannot be recovered, in order to be converted with 100% efficiency into other forms of energy. In this case, the energy must partly stay as heat, and cannot be completely recovered as usable energy, except at the price of an increase in some other kind of heat-like increase in disorder in quantum states, in the universe (such as an expansion of matter, or a randomisation in a crystal). As the universe evolves in time, more and more of its energy becomes trapped in irreversible states (i.e., as heat or other kinds of increases in disorder). This has been referred to as the inevitable thermodynamic heat death of the universe. In this heat death the energy of the universe does not change, but the fraction of energy which is available to do work through a heat engine, or be transformed to other usable forms of energy (through the use of generators attached to heat engines), grows less and less. The fact that energy can be neither created nor be destroyed is called the law of conservation of energy. In the form of the first law of thermodynamics, this states that a closed system's energy is constant unless energy is transferred in or out by work or heat, and that no energy is lost in transfer. The total inflow of energy into a system must equal the total outflow of energy from the system, plus the change in the energy contained within the system. Whenever one measures (or calculates) the total energy of a system of particles whose interactions do not depend explicitly on time, it is found that the total energy of the system always remains constant. While heat can always be fully converted into work in a reversible isothermal expansion of an ideal gas, for cyclic processes of practical interest in heat engines the second law of thermodynamics states that the system doing work always loses some energy as waste heat. This creates a limit to the amount of heat energy that can do work in a cyclic process, a limit called the available energy. Mechanical and other forms of energy can be transformed in the other direction into thermal energy without such limitations. The total energy of a system can be calculated by adding up all forms of energy in the system. Richard Feynman said during a 1961 lecture: Most kinds of energy (with gravitational energy being a notable exception) are subject to strict local conservation laws as well. In this case, energy can only be exchanged between adjacent regions of space, and all observers agree as to the volumetric density of energy in any given space. There is also a global law of conservation of energy, stating that the total energy of the universe cannot change; this is a corollary of the local law, but not vice versa. This law is a fundamental principle of physics. As shown rigorously by Noether's theorem, the conservation of energy is a mathematical consequence of translational symmetry of time, a property of most phenomena below the cosmic scale that makes them independent of their locations on the time coordinate. Put differently, yesterday, today, and tomorrow are physically indistinguishable. This is because energy is the quantity which is canonical conjugate to time. This mathematical entanglement of energy and time also results in the uncertainty principle - it is impossible to define the exact amount of energy during any definite time interval. The uncertainty principle should not be confused with energy conservation - rather it provides mathematical limits to which energy can in principle be defined and measured. Each of the basic forces of nature is associated with a different type of potential energy, and all types of potential energy (like all other types of energy) appears as system mass, whenever present. For example, a compressed spring will be slightly more massive than before it was compressed. Likewise, whenever energy is transferred between systems by any mechanism, an associated mass is transferred with it. In quantum mechanics energy is expressed using the Hamiltonian operator. On any time scales, the uncertainty in the energy is by which is similar in form to the Heisenberg Uncertainty Principle (but not really mathematically equivalent thereto, since "H" and "t" are not dynamically conjugate variables, neither in classical nor in quantum mechanics). In particle physics, this inequality permits a qualitative understanding of virtual particles which carry momentum, exchange by which and with real particles, is responsible for the creation of all known fundamental forces (more accurately known as fundamental interactions). Virtual photons (which are simply lowest quantum mechanical energy state of photons) are also responsible for electrostatic interaction between electric charges (which results in Coulomb law), for spontaneous radiative decay of exited atomic and nuclear states, for the Casimir force, for van der Waals bond forces and some other observable phenomena. Energy transfer can be considered for the special case of systems which are closed to transfers of matter. The portion of the energy which is transferred by conservative forces over a distance is measured as the work the source system does on the receiving system. The portion of the energy which does not do work during the transfer is called heat. Energy can be transferred between systems in a variety of ways. Examples include the transmission of electromagnetic energy via photons, physical collisions which transfer kinetic energy, and the conductive transfer of thermal energy. Energy is strictly conserved and is also locally conserved wherever it can be defined. In thermodynamics, for closed systems, the process of energy transfer is described by the first law: where formula_16 is the amount of energy transferred, formula_2  represents the work done on the system, and formula_18 represents the heat flow into the system. As a simplification, the heat term, formula_18, is sometimes ignored, especially when the thermal efficiency of the transfer is high. This simplified equation is the one used to define the joule, for example. Beyond the constraints of closed systems, open systems can gain or lose energy in association with matter transfer (both of these process are illustrated by fueling an auto, a system which gains in energy thereby, without addition of either work or heat). Denoting this energy by formula_16, one may write Internal energy is the sum of all microscopic forms of energy of a system. It is the energy needed to create the system. It is related to the potential energy, e.g., molecular structure, crystal structure, and other geometric aspects, as well as the motion of the particles, in form of kinetic energy. Thermodynamics is chiefly concerned with changes in internal energy and not its absolute value, which is impossible to determine with thermodynamics alone. The first law of thermodynamics asserts that energy (but not necessarily thermodynamic free energy) is always conserved and that heat flow is a form of energy transfer. For homogeneous systems, with a well-defined temperature and pressure, a commonly used corollary of the first law is that, for a system subject only to pressure forces and heat transfer (e.g., a cylinder-full of gas) without chemical changes, the differential change in the internal energy of the system (with a "gain" in energy signified by a positive quantity) is given as where the first term on the right is the heat transferred into the system, expressed in terms of temperature "T" and entropy "S" (in which entropy increases and the change d"S" is positive when the system is heated), and the last term on the right hand side is identified as work done on the system, where pressure is "P" and volume "V" (the negative sign results since compression of the system requires work to be done on it and so the volume change, d"V", is negative when work is done on the system). This equation is highly specific, ignoring all chemical, electrical, nuclear, and gravitational forces, effects such as advection of any form of energy other than heat and pV-work. The general formulation of the first law (i.e., conservation of energy) is valid even in situations in which the system is not homogeneous. For these cases the change in internal energy of a "closed" system is expressed in a general form by where formula_23 is the heat supplied to the system and formula_24 is the work applied to the system. The energy of a mechanical harmonic oscillator (a mass on a spring) is alternatively kinetic and potential energy. At two points in the oscillation cycle it is entirely kinetic, and at two points it is entirely potential. Over the whole cycle, or over many cycles, net energy is thus equally split between kinetic and potential. This is called equipartition principle; total energy of a system with many degrees of freedom is equally split among all available degrees of freedom. This principle is vitally important to understanding the behaviour of a quantity closely related to energy, called entropy. Entropy is a measure of evenness of a distribution of energy between parts of a system. When an isolated system is given more degrees of freedom (i.e., given new available energy states that are the same as existing states), then total energy spreads over all available degrees equally without distinction between "new" and "old" degrees. This mathematical result is called the second law of thermodynamics. The second law of thermodynamics is valid only for systems which are near or in equilibrium state. For non-equilibrium systems, the laws governing system's behavior are still debatable. One of the guiding principles for these systems is the principle of maximum entropy production. It states that nonequilibrium systems behave in such a way to maximize its entropy production.
https://en.wikipedia.org/wiki?curid=9649
Expected value In probability theory, the expected value of a random variable is a generalization of the weighted average and intuitively is the arithmetic mean of a large number of independent realizations of that variable. The expected value is also known as the expectation, mathematical expectation, mean, average, or first moment. By definition, the expected value of a constant random variable formula_1 is formula_2. The expected value of a random variable formula_3 with equiprobable outcomes formula_4 is defined as the arithmetic mean of the terms formula_5 If some of the probabilities formula_6 of an individual outcome formula_7 are unequal, then the expected value is defined to be the probability-weighted average of the formula_7s, i.e. the sum of the formula_9 products formula_10. Expected value of a general random variable involves integration in the sense of Lebesgue. The idea of the expected value originated in the middle of the 17th century from the study of the so-called problem of points, which seeks to divide the stakes "in a fair way" between two players who have to end their game before it's properly finished. This problem had been debated for centuries, and many conflicting proposals and solutions had been suggested over the years, when it was posed in 1654 to Blaise Pascal by French writer and amateur mathematician Chevalier de Méré. Méré claimed that this problem couldn't be solved and that it showed just how flawed mathematics was when it came to its application to the real world. Pascal, being a mathematician, was provoked and determined to solve the problem once and for all. He began to discuss the problem in a now famous series of letters to Pierre de Fermat. Soon enough they both independently came up with a solution. They solved the problem in different computational ways but their results were identical because their computations were based on the same fundamental principle. The principle is that the value of a future gain should be directly proportional to the chance of getting it. This principle seemed to have come naturally to both of them. They were very pleased by the fact that they had found essentially the same solution and this in turn made them absolutely convinced they had solved the problem conclusively; however, they did not publish their findings. They only informed a small circle of mutual scientific friends in Paris about it. Three years later, in 1657, a Dutch mathematician Christiaan Huygens, who had just visited Paris, published a treatise (see ) ""De ratiociniis in ludo aleæ"" on probability theory. In this book he considered the problem of points and presented a solution based on the same principle as the solutions of Pascal and Fermat. Huygens also extended the concept of expectation by adding rules for how to calculate expectations in more complicated situations than the original problem (e.g., for three or more players). In this sense this book can be seen as the first successful attempt at laying down the foundations of the theory of probability. In the foreword to his book, Huygens wrote: Thus, Huygens learned about de Méré's Problem in 1655 during his visit to France; later on in 1656 from his correspondence with Carcavi he learned that his method was essentially the same as Pascal's; so that before his book went to press in 1657 he knew about Pascal's priority in this subject. Neither Pascal nor Huygens used the term "expectation" in its modern sense. In particular, Huygens writes: More than a hundred years later, in 1814, Pierre-Simon Laplace published his tract ""Théorie analytique des probabilités"", where the concept of expected value was defined explicitly: The use of the letter formula_11 to denote expected value goes back to W. A. Whitworth in 1901. The symbol has become popular since for English writers. In German, formula_11 stands for "Erwartungswert", in Spanish for "Esperanza matemática", and in French for "Espérance mathématique". Another popular notation is formula_13, where as formula_14 is commonly used in physics, and formula_15 in Russian-language literature. Let formula_3 be a random variable with a finite number of finite outcomes formula_17 occurring with probabilities formula_18 respectively. The expectation of formula_3 is defined as Since the sum of all probabilities formula_21 is 1 (formula_22), the expected value is the weighted average of the formula_23 values, with the formula_21 values being the weights. If all outcomes formula_23 are equiprobable (that is, formula_26), then the weighted average turns into the simple average. If the outcomes formula_23 are not equiprobable, then the simple average must be replaced with the weighted average, which takes into account the fact that some outcomes are more likely than the others. Intuitively, the expectation of a random variable taking values in a countable set of outcomes is defined analogously as the weighted sum of the outcome values, where the weights correspond to the probabilities of realizing that value. However, convergence issues associated with the infinite sum necessitate a more careful definition. A rigorous definition first defines expectation of a non-negative random variable, and then adapts it to general random variables. Let formula_3 be a non-negative random variable with a countable set of outcomes formula_40 occurring with probabilities formula_41 respectively. Analogous to the discrete case, the expected value of formula_3 is then defined as the series Note that since formula_44, the infinite sum is well-defined and does not depend on the order in which it is computed. Unlike the finite case, the expectation here can be equal to infinity, if the infinite sum above increases without bound. For a general (not necessarily non-negative) random variable formula_3 with a countable number of outcomes, set formula_46 and formula_47. By definition, Like with non-negative random variables, formula_49 can, once again, be finite or infinite. The third option here is that formula_49 is no longer guaranteed to be well defined at all. The latter happens whenever formula_51. If formula_3 is a random variable with a probability density function of formula_79, then the expected value is defined as the Lebesgue integral where the values on both sides are well defined or not well defined simultaneously. Example. A random variable that has the Cauchy distribution has a density function, but the expected value is undefined since the distribution has large "tails". In general, if formula_3 is a random variable defined on a probability space formula_82, then the expected value of formula_3, denoted by formula_49, is defined as the Lebesgue integral For multidimensional random variables, their expected value is defined per component, i.e. and, for a random matrix formula_3 with elements formula_88, formula_89 The basic properties below (and their names in bold) replicate or follow immediately from those of Lebesgue integral. Note that the letters "a.s." stand for "almost surely" that is a central property of the Lebesgue integral. Basically, one says that an inequality like formula_90 is true almost surely when the probability measure attributes zero-mass to the complementary event formula_91. The expectation of a random variable plays an important role in a variety of contexts. For example, in decision theory, an agent making an optimal choice in the context of incomplete information is often assumed to maximize the expected value of their utility function. For a different example, in statistics, where one seeks estimates for unknown parameters based on available data, the estimate itself is a random variable. In such settings, a desirable criterion for a "good" estimator is that it is unbiased – that is, the expected value of the estimate is equal to the true value of the underlying parameter. It is possible to construct an expected value equal to the probability of an event by taking the expectation of an indicator function that is one if the event has occurred and zero otherwise. This relationship can be used to translate properties of expected values into properties of probabilities, e.g. using the law of large numbers to justify estimating probabilities by frequencies. The expected values of the powers of "X" are called the moments of "X"; the moments about the mean of "X" are expected values of powers of "X" − E["X"]. The moments of some random variables can be used to specify their distributions, via their moment generating functions. To empirically estimate the expected value of a random variable, one repeatedly measures observations of the variable and computes the arithmetic mean of the results. If the expected value exists, this procedure estimates the true expected value in an unbiased manner and has the property of minimizing the sum of the squares of the residuals (the sum of the squared differences between the observations and the estimate). The law of large numbers demonstrates (under fairly mild conditions) that, as the size of the sample gets larger, the variance of this estimate gets smaller. This property is often exploited in a wide variety of applications, including general problems of statistical estimation and machine learning, to estimate (probabilistic) quantities of interest via Monte Carlo methods, since most quantities of interest can be written in terms of expectation, e.g. formula_163, where formula_164 is the indicator function of the set formula_165. Expected values can also be used to compute the variance, by means of the computational formula for the variance A very important application of the expectation value is in the field of quantum mechanics. The expectation value of a quantum mechanical operator formula_167 operating on a quantum state vector formula_168 is written as formula_169. The uncertainty in formula_167 can be calculated using the formula formula_171. In general, it is not the case that formula_172 despite formula_173 pointwise. Thus, one cannot interchange limits and expectation, without additional conditions on the random variables. To see this, let formula_174 be a random variable distributed uniformly on formula_175. For formula_176 define a sequence of random variables with formula_178 being the indicator function of the event formula_100. Then, it follows that formula_180 (a.s). But, formula_181 for each formula_9. Hence, formula_183 Analogously, for general sequence of random variables formula_184, the expected value operator is not formula_185-additive, i.e. An example is easily obtained by setting formula_187 and formula_188 for formula_189, where formula_190 is as in the previous example. A number of convergence results specify exact conditions which allow one to interchange limits and expectations, as specified below. There are a number of inequalities involving the expected values of functions of random variables. The following list includes some of the more basic ones. The probability density function formula_249 of a scalar random variable formula_3 is related to its characteristic function formula_251 by the inversion formula: For the expected value of formula_128 (where formula_254 is a Borel function), we can use this inversion formula to obtain If formula_256 is finite, changing the order of integration, we get, in accordance with Fubini–Tonelli theorem, where is the Fourier transform of formula_259 The expression for formula_256 also follows directly from Plancherel theorem.
https://en.wikipedia.org/wiki?curid=9653
Electric light An electric light is a device that produces visible light from electric current. It is the most common form of artificial lighting and is essential to modern society, providing interior lighting for buildings and exterior light for evening and nighttime activities. In technical usage, a replaceable component that produces light from electricity is called a lamp. Lamps are commonly called light bulbs; for example, the incandescent light bulb. Lamps usually have a base made of ceramic, metal, glass, or plastic, which secures the lamp in the socket of a light fixture. The electrical connection to the socket may be made with a screw-thread base, two metal pins, two metal caps or a bayonet cap. The three main categories of electric lights are incandescent lamps, which produce light by a filament heated white-hot by electric current, gas-discharge lamps, which produce light by means of an electric arc through a gas, and LED lamps, which produce light by a flow of electrons across a band gap in a semiconductor. Before electric lighting became common in the early 20th century, people used candles, gas lights, oil lamps, and fires. English chemist Humphry Davy developed the first incandescent light in 1802, followed by the first practical electric arc light in 1806. By the 1870s, Davy's arc lamp had been successfully commercialized, and was used to light many public spaces. Efforts by Swan and Edison led to commercial incandescent light bulbs becoming widely available in the 1880s, and by the early twentieth century these had completely replaced arc lamps. The energy efficiency of electric lighting has increased radically since the first demonstration of arc lamps and the incandescent light bulb of the 19th century. Modern electric light sources come in a profusion of types and sizes adapted to many applications. Most modern electric lighting is powered by centrally generated electric power, but lighting may also be powered by mobile or standby electric generators or battery systems. Battery-powered light is often reserved for when and where stationary lights fail, often in the form of flashlights or electric lanterns, as well as in vehicles. Types of electric lighting include: Different types of lights have vastly differing efficacies and color of light. *Color temperature is defined as the temperature of a black body emitting a similar spectrum; these spectra are quite different from those of black bodies. The most efficient source of electric light is the low-pressure sodium lamp. It produces, for all practical purposes, a monochromatic orange-yellow light, which gives a similarly monochromatic perception of any illuminated scene. For this reason, it is generally reserved for outdoor public lighting applications. Low-pressure sodium lights are favoured for public lighting by astronomers, since the light pollution that they generate can be easily filtered, contrary to broadband or continuous spectra. The modern incandescent light bulb, with a coiled filament of tungsten, and commercialized in the 1920s, developed from the carbon filament lamp introduced about 1880. Less than 3% of the input energy is converted into usable light. Nearly all of the input energy ends up as heat that, in warm climates, must then be removed from the building by ventilation or air conditioning, often resulting in more energy consumption. In colder climates where heating and lighting is required during the cold and dark winter months, the heat byproduct has some value. Incandescent bulbs are being phased out in many countries due to their low energy efficiency. As well as bulbs for normal illumination, there is a very wide range, including low voltage, low-power types often used as components in equipment, but now largely displaced by LEDs. Halogen lamps are usually much smaller than standard incandescent lamps, because for successful operation a bulb temperature over 200 °C is generally necessary. For this reason, most have a bulb of fused silica (quartz) or aluminosilicate glass. This is often sealed inside an additional layer of glass. The outer glass is a safety precaution, to reduce ultraviolet emission and to contain hot glass shards should the inner envelope explode during operation. Oily residue from fingerprints may cause a hot quartz envelope to shatter due to excessive heat buildup at the contamination site. The risk of burns or fire is also greater with bare bulbs, leading to their prohibition in some places, unless enclosed by the luminaire. Those designed for 12- or 24-volt operation have compact filaments, useful for good optical control. Also, they have higher efficacies (lumens per watt) and better lives than non-halogen types. The light output remains almost constant throughout their life. Fluorescent lamps consist of a glass tube that contains mercury vapour or argon under low pressure. Electricity flowing through the tube causes the gases to give off ultraviolet energy. The inside of the tubes are coated with phosphors that give off visible light when struck by ultraviolet photons. They have much higher efficiency than incandescent lamps. For the same amount of light generated, they typically use around one-quarter to one-third the power of an incandescent. The typical luminous efficacy of fluorescent lighting systems is 50–100 lumens per watt, several times the efficacy of incandescent bulbs with comparable light output. Fluorescent lamp fixtures are more costly than incandescent lamps, because they require a ballast to regulate the current through the lamp, but the lower energy cost typically offsets the higher initial cost. Compact fluorescent lamps are available in the same popular sizes as incandescent lamps and are used as an energy-saving alternative in homes. Because they contain mercury, many fluorescent lamps are classified as hazardous waste. The United States Environmental Protection Agency recommends that fluorescent lamps be segregated from general waste for recycling or safe disposal, and some jurisdictions require recycling of them. The solid-state light-emitting diode (LED) has been popular as an indicator light in consumer electronics and professional audio gear since the 1970s. In the 2000s, efficacy and output have risen to the point where LEDs are now being used in lighting applications such as car headlights and brake lights, in flashlights and bicycle lights, as well as in decorative applications, such as holiday lighting. Indicator LEDs are known for their extremely long life, up to 100,000 hours, but lighting LEDs are operated much less conservatively, and consequently have shorter lives. LED technology is useful for lighting designers, because of its low power consumption, low heat generation, instantaneous on/off control, and in the case of single color LEDs, continuity of color throughout the life of the diode and relatively low cost of manufacture. LED lifetime depends strongly on the temperature of the diode. Operating an LED lamp in conditions that increase the internal temperature can greatly shorten the lamp's life. Carbon arc lamps consist of two carbon rod electrodes in open air, supplied by a current-limiting ballast. The electric arc is struck by touching the rod tips then separating them. The ensuing arc produces a white-hot plasma between the rod tips. These lamps have higher efficacy than filament lamps, but the carbon rods are short-lived and require constant adjustment in use, as the intense heat of the arc erodes them. The lamps produce significant ultraviolet output, they require ventilation when used indoors, and due to their intensity they need protection from direct sight. Invented by Humphry Davy around 1805, the carbon arc was the first practical electric light. It was used commercially beginning in the 1870s for large building and street lighting until it was superseded in the early 20th century by the incandescent light. Carbon arc lamps operate at high power and produce high intensity white light. They also are a point source of light. They remained in use in limited applications that required these properties, such as movie projectors, stage lighting, and searchlights, until after World War II. A discharge lamp has a glass or silica envelope containing two metal electrodes separated by a gas. Gases used include, neon, argon, xenon, sodium, metal halide, and mercury. The core operating principle is much the same as the carbon arc lamp, but the term "arc lamp" normally refers to carbon arc lamps, with more modern types of gas discharge lamp normally called discharge lamps. With some discharge lamps, very high voltage is used to strike the arc. This requires an electrical circuit called an igniter, which is part of the electrical ballast circuitry. After the arc is struck, the internal resistance of the lamp drops to a low level, and the ballast limits the current to the operating current. Without a ballast, excess current would flow, causing rapid destruction of the lamp. Some lamp types contain a little neon, which permits striking at normal running voltage, with no external ignition circuitry. Low pressure sodium lamps operate this way. The simplest ballasts are just an inductor, and are chosen where cost is the deciding factor, such as street lighting. More advanced electronic ballasts may be designed to maintain constant light output over the life of the lamp, may drive the lamp with a square wave to maintain completely flicker-free output, and shut down in the event of certain faults. Many lamp units, or light bulbs, are specified in standardized shape codes and socket names. Incandescent bulbs and their retrofit replacements are often specified as "A19/A60 E26/E27", a common size for these kind of light bulbs. In this example, the "A" parameters describe the bulb size and shape while the "E" parameters describe the Edison screw base size and thread characteristics. Life expectancy for many types of lamp is defined as the number of hours of operation at which 50% of them fail, that is the median life of the lamps. Production tolerances as low as 1% can create a variance of 25% in lamp life, so in general some lamps will fail well before the rated life expectancy, and some will last much longer. For LEDs, lamp life is defined as the operation time at which 50% of lamps have experienced a 70% decrease in light output. Some types of lamp are also sensitive to switching cycles. Rooms with frequent switching, such as bathrooms, can expect much shorter lamp life than what is printed on the box. Compact fluorescent lamps are particularly sensitive to switching cycles. The total amount of artificial light (especially from street light) is sufficient for cities to be easily visible at night from the air, and from space. This light is the source of light pollution that burdens astronomers and others. Electric lamps can be used as heat sources, for example in incubators, as infrared lamps in fast food restaurants and toys such as the Kenner Easy-Bake Oven. Due to their nonlinear resistance characteristics, tungsten filament lamps have long been used as fast-acting thermistors in electronic circuits. Popular uses have included: A stylized depiction of a light bulb features as the logo of the Turkish AK Party. In circuit diagrams lamps usually are shown as symbols. There are two main types of symbols, these are:
https://en.wikipedia.org/wiki?curid=9656
Edgar Rice Burroughs Edgar Rice Burroughs (September 1, 1875 – March 19, 1950) was an American speculative fiction writer, best known for his prolific output in the adventure, science fiction and fantasy genres. Among his most well-known creations include Tarzan of the Apes, John Carter of Mars ("Barsoom" series), and Carson Napier of Venus ("Amtor" series). He is also known for the hollow Earth-themed "Pellucidar" series, beginning with "At the Earth's Core" (1914); and the lost world-themed "Caspak" trilogy, beginning with "The Land that Time Forgot" (1918). Burroughs' California ranch is now the center of the Tarzana neighborhood in Los Angeles. Burroughs was born on September 1, 1875, in Chicago (he later lived for many years in the suburb of Oak Park), the fourth son of Major George Tyler Burroughs (1833–1913), a businessman and Civil- War veteran, and his wife, Mary Evaline (Zieger) Burroughs (1840–1920). His middle name is from his paternal grandmother, Mary Coleman Rice Burroughs (1802–1889). He was of almost entirely English ancestry, with a family line that had been in North America since the Colonial era. Through his Rice grandmother, Burroughs was descended from settler Edmund Rice, one of the English Puritans who moved to Massachusetts Bay Colony in the early 17th Century. He once remarked, "I can trace my ancestry back to Deacon Edmund Rice." The Burroughs side of the family was also of English origin and also emigrated to Massachusetts around the same time. Many of his ancestors fought in the American Revolution. Some of his ancestors settled in Virginia during the colonial period, and Burroughs often emphasized his connection with that side of his family, seeing it as romantic and warlike, and, in fact, could have counted among his close cousins no less than seven signers of the U.S. Declaration of Independence, including his third cousin, four times removed, 2nd President of the United States John Adams. Burroughs was educated at a number of local schools. He then attended Phillips Academy, in Andover, Massachusetts, and then the Michigan Military Academy. Graduating in 1895, and failing the entrance exam for the United States Military Academy at West Point, he became an enlisted soldier with the 7th U.S. Cavalry in Fort Grant, Arizona Territory. After being diagnosed with a heart problem and thus ineligible to serve, he was discharged in 1897. After his discharge Burroughs worked a number of different jobs. During the Chicago influenza epidemic of 1891, he spent half a year at his brother's ranch on the Raft River in Idaho, as a cowboy, drifted somewhat afterward, then worked at his father's Chicago battery factory in 1899, marrying his childhood sweetheart, Emma Hulbert (1876–1944), in January 1900. In 1903, Burroughs joined his brothers, Yale graduates George and Harry, who were, by then, prominent Pocatello area ranchers in southern Idaho, and partners in the Sweetser-Burroughs Mining Company, where he took on managing their ill-fated Snake River gold dredge, a classic bucket-line dredge. The Burroughs brothers were also the sixth cousins, once removed, of famed miner Kate Rice, a brilliant and statuesque Maths professor who, in 1914, became the first female prospector in the Canadian North. Journalist and publisher C. Allen Thorndike Rice was also his third cousin. When the new mine proved unsuccessful, the brothers secured for Burroughs a position with the Oregon Short Line Railroad in Salt Lake City. Burroughs resigned from the railroad in October 1904. By 1911, after seven years of low wages as a pencil-sharpener wholesaler; Burroughs began to write fiction. By this time, Emma and he had two children, Joan (1908–1972), and Hulbert (1909–1991). During this period, he had copious spare time and began reading pulp-fiction magazines. In 1929, he recalled thinking that In 1913, Burroughs and Emma had their third and last child, John Coleman Burroughs (1913–1979), later known for his illustrations of his father's books. In the 1920s, Burroughs became a pilot, purchased a Security Airster S-1, and encouraged his family to learn to fly. Daughter Joan married "Tarzan" film actor, James Pierce, starring with her husband, as the voice of "Jane", during 1932–34 for the "Tarzan" radio series. The pair were wed for more than forty years, until her death, in 1972. Burroughs divorced Emma in 1934 and, in 1935, married the former actress Florence Gilbert Dearholt, who was the former wife of his friend (who was then himself remarrying), Ashton Dearholt, with whom he had co-founded Burroughs-Tarzan Enterprises while filming "The New Adventures of Tarzan". Burroughs adopted the Dearholts' two children. He and Florence divorced in 1942. Burroughs was in his late 60s and was in Honolulu at the time of the Japanese attack on Pearl Harbor. Despite his age, he applied for and received permission to become a war correspondent, becoming one of the oldest U.S. war correspondents during World War II. This period of his life is mentioned in William Brinkley's bestselling novel "Don't Go Near the Water". After the war ended, Burroughs moved back to Encino, California, where after many health problems, he died of a heart attack on March 19, 1950, having written almost 80 novels. He is buried at Tarzana, California, US. When he died, he was believed to have been the writer who had made the most from films, earning over $2 million in royalties from 27 Tarzan pictures. The Science Fiction Hall of Fame inducted Burroughs in 2003. Aiming his work at the pulps—under the name "Norman Bean" to protect his reputation—Burroughs had his first story, "Under the Moons of Mars", serialized by Frank Munsey in the February to July 1912 issues of "The All-Story". "Under the Moons of Mars" inaugurated the "Barsoom" series and earned Burroughs ($ today). It was first published as a book by A. C. McClurg of Chicago in 1917, entitled "A Princess of Mars", after three Barsoom sequels had appeared as serials and McClurg had published the first four serial Tarzan novels as books. Burroughs soon took up writing full-time, and by the time the run of "Under the Moons of Mars" had finished, he had completed two novels, including "Tarzan of the Apes", published from October 1912 and one of his most successful series. Burroughs also wrote popular science fiction and fantasy stories involving adventurers from Earth transported to various planets (notably Barsoom, Burroughs's fictional name for Mars, and Amtor, his fictional name for Venus), lost islands, and into the interior of the Hollow Earth in his "Pellucidar" stories. He also wrote Westerns and historical romances. Besides those published in "All-Story", many of his stories were published in "The Argosy" magazine. Tarzan was a cultural sensation when introduced. Burroughs was determined to capitalize on Tarzan's popularity in every way possible. He planned to exploit Tarzan through several different media including a syndicated Tarzan comic strip, movies, and merchandise. Experts in the field advised against this course of action, stating that the different media would just end up competing against each other. Burroughs went ahead, however, and proved the experts wrong – the public wanted Tarzan in whatever fashion he was offered. Tarzan remains one of the most successful fictional characters to this day and is a cultural icon. In either 1915 or 1919, Burroughs purchased a large ranch north of Los Angeles, California, which he named "Tarzana". The citizens of the community that sprang up around the ranch voted to adopt that name when their community, Tarzana, California, was formed in 1927. Also, the unincorporated community of Tarzan, Texas, was formally named in 1927 when the US Postal Service accepted the name, reputedly coming from the popularity of the first (silent) "Tarzan of the Apes" film, starring Elmo Lincoln, and an early "Tarzan" comic strip. In 1923, Burroughs set up his own company, Edgar Rice Burroughs, Inc., and began printing his own books through the 1930s. Because of the part Burroughs's science fiction played in inspiring real exploration of Mars, an impact crater on Mars was named in his honor after his death. In a "Paris Review" interview, Ray Bradbury said of Burroughs that "Edgar Rice Burroughs never would have looked upon himself as a social mover and shaker with social obligations. But as it turns out – and I love to say it because it upsets everyone terribly – Burroughs is probably the most influential writer in the entire history of the world." Bradbury continued that "By giving romance and adventure to a whole generation of boys, Burroughs caused them to go out and decide to become special." In "Something of Myself" (published posthumously in 1937) Rudyard Kipling wrote: "My "Jungle Books" begat Zoos of [imitators]. But the genius of all the genii was one who wrote a series called "Tarzan of the Apes". I read it, but regret I never saw it on the films, where it rages most successfully. He had 'jazzed' the motif of the "Jungle Books" and, I imagine, had thoroughly enjoyed himself. He was reported to have said that he wanted to find out how bad a book he could write and 'get away with', which is a legitimate ambition." By 1963, Floyd C. Gale of "Galaxy Science Fiction" wrote when discussing reprints of several Burroughs novels by Ace Books, "an entire generation has grown up inexplicably Burroughs-less". He stated that most of the author's books had been out of print for years and that only the "occasional laughable Tarzan film" reminded public of his fiction. Gale reported his surprise that after two decades his books were again available, with Canaveral Press, Dover Publications, and Ballantine Books also reprinting them. Few critical books have been written about Burroughs. From an academic standpoint, the most helpful are Erling Holtsmark's two books: "Tarzan and Tradition" and "Edgar Rice Burroughs"; Stan Galloway's "The Teenage Tarzan: A Literary Analysis of Edgar Rice Burroughs' "Jungle Tales of Tarzan; and Richard Lupoff's two books: "Master of Adventure: Edgar Rice Burroughs" and "Barsoom: Edgar Rice Burroughs and the Martian Vision". Galloway was identified by James Edwin Gunn as "one of the half-dozen finest Burroughs scholars in the world"; Galloway called Holtsmark his "most important predecessor." These three texts have been published by various houses in one or two volumes. Adding to the confusion, some editions have the original (significantly longer) introduction to Part I from the first publication as a magazine serial, and others have the shorter version from the first book publication, which included all three parts under the title "The Moon Maid".
https://en.wikipedia.org/wiki?curid=9657
Eugène Viollet-le-Duc Eugène Emmanuel Viollet-le-Duc (; 27 January 181417 September 1879) was a French architect and author who restored many prominent medieval landmarks in France, including those which had been damaged or abandoned during the French Revolution. His major restoration projects included Notre-Dame de Paris, the Basilica of Saint Denis, Mont Saint-Michel, Sainte-Chapelle, and the medieval walls of the city of Carcassonne. His later writings on the relationship between form and function in architecture had a notable influence on a new generation of architects, including Antoni Gaudí, Victor Horta, and Louis Sullivan. Viollet-le-Duc was born in Paris in 1814, in the last year of the Empire of Napoleon Bonaparte. His grandfather was an architect, and his father was a high-ranking civil servant, who in 1816 became the overseer of the royal residences of Louis XVIII. His uncle Étienne-Jean Delécluze was a painter, a former student of Jacques-Louis David, an art critic and hosted a literary salon, which was attended by Stendhal and Sainte-Beuve. His mother hosted her own salon, which women could attend as well as men. There, in 1822 or 1823, Eugène met Prosper Mérimée, a writer who would play a decisive role in his career. In 1825 he began his education at the Pension Moran, in Fontenay-aux-Roses. He returned to Paris in 1829 as a student at the College de Bourbon (now the Lycée Condorcet). He passed his baccalaureate examination in 1830. His uncle urged him to enter the École des Beaux-Arts, which had been created in 1806, but the "École " had an extremely rigid system, based entirely on copying classical models, and Eugène was not interested. Instead he decided to get practical experience in the architectural offices of Jacques-Marie Huvé and Achille Leclère, while devoting much of his time to drawing medieval churches and monuments around Paris. He participated in the July 1830 revolution which overthrew Charles X, building a barricade, his first known construction project. Following the revolution, which brought Louis Philippe to power, his father became chief of the bureau of royal residences. The new government created, for the first time, the position of Inspector General of Historic Monuments. Eugène's uncle Delécluze agreed to take Eugène on a long tour of France to see monuments. They travelled from July to October 1831 throughout the south of France, and he returned with a large collection of detailed paintings and watercolours of churches and monuments. On his return to Paris, he moved with his family into the Tuileries Palace, where his father was now governor of royal residences. His family again urged him to attend the École des Beaux-Arts, but he still refused. He wrote in his journal in December 1831, "the "École" is just a mould for architects. they all come out practically identical." He was a talented and meticulous artist; he travelled around France to visit monuments, cathedrals, and other medieval architecture, made detailed drawings and watercolours, which he sometimes sold at a high price to members of the Court. On May 3, 1834, at age twenty, he married Élisabeth Templier, and in the same year he was named an associate professor of ornamental decoration at the Royal School of Decorative Arts, which gave him a more regular income. With the money from the sale of his drawings and paintings, the couple set off on a long tour of the monuments of Italy, visiting Rome, Venice, Florence and other sites, drawing and painting. His reaction to the Leaning Tower of Pisa was characteristic: "It was extremely disagreeable to see", he wrote, "it would have been infinitely better if it had been straight." In 1838, he presented several of his drawings at the Paris Salon, and began making a travel book, "Picturesque and romantic images of the old France", for which, between 1838 and 1844, he made nearly three hundred engravings. In October 1838, with the recommendation of Achille Leclère, the architect with whom he had trained, he was named deputy inspector of the enlargement of the Hôtel Soubise, the new home of the French National Archives. His uncle, Delécluze, then recommended him to the new Commission of Historic Monuments of France, led by Prosper Mérimée, who had just published a book on medieval French monuments. Though he was just twenty-four years old and had no degree in architecture, he was asked to go to Narbonne to propose a plan for the completion of the cathedral there. He made his first plan, which included not only the completion but also the restoration of the oldest parts of the structure. His first project was rejected by the local authorities as too ambitious and too expensive. His next project was a restoration of the Vézelay Abbey, the church of a Benedictine monastery founded in the 12th century to house the reputed relics of Mary Magdalene. The church had been sacked by the Huguenots in 1569, and during the French Revolution, the facade and statuary on the facade were destroyed. The vaults of the roof were weakened, and many of the stones had been carried off for other projects. When Mérimée visited to inspect the structure he heard stones falling around him. In February 1840 Mérimée gave Viollet-le-Duc the mission of restoring and reconstructing the church so it would not collapse, while "respecting exactly in his project of restoration all the ancient dispositions of the church". The task was all the more difficult because up until that time no scientific studies had been made of medieval building techniques, and there were no schools of restoration. He had no plans for the original building to work from. Viollet-le-Duc had to discover the flaws of construction that had caused the building to start to collapse in the first place and to construct a more solid and stable structure. He lightened the roof and built new arches to stabilize the structure, and slightly changed the shape of the vaults and arches. He was criticized for these modifications in the 1960s, though, as his defenders argued, without them the roof would have collapsed under its own weight. Mérimée's deputy, Lenormant, inspected the construction and reported to Mérimée: "The young Leduc seems entirely worthy of your confidence. He needed a magnificent audacity to take charge of such a desperate enterprise; it's certain that he arrived just in time, and if we had waited only ten years the church would have been a pile of stones." Viollet-le-Duc's work at Vezelay led to a series of larger projects. In 1840, in collaboration with his friend the architect Jean-Baptiste Lassus he began the restoration of Sainte-Chapelle in Paris, which had been turned into a storage depot after the Revolution. His role in this project was relatively minor, with Lassus taking the lead. In February 1843, King Louis Philippe sent him to the Château of Amboise, to restore the stained glass windows in the chapel holding the tomb of Leonardo da Vinci. The windows were unfortunately destroyed in 1940 during World War II. In 1843, Mérimée took Viollet-le-Duc with him to Burgundy and the south of France, on one of his long inspection tours of possible monuments. The two men shared the same passion for the gothic. Viollet-le-Duc made drawings of the buildings and wrote detailed accounts of each site, illustrated with his drawing, which were published in architectural journals. These articles were later turned into books; he became the most prominent academic scholar on French medieval architecture. In 1844, with the backing of Mérimée, Viollet-le-Duc, just thirty years old, and Lassus, then thirty-seven, won a competition for the restoration of Notre-Dame Cathedral. Their project involved primarily the facade, where many of the statues over the portals had been beheaded or smashed during the Revolution. They proposed two major changes to the interior: rebuilding two of the bays to their original medieval height of four storeys, and removing the marble neoclassical structures and decoration which had been added to the choir during the reign of Louis XIV. Mérimée warned them to be careful: "In such a project, one cannot act with too much prudence or discretion...A restoration may be more disastrous for a monument than the ravages of centuries." The Commission on Historical Monuments approved most of Viollet-le-Duc's plans, but rejected his proposal to remove the choir built under Louis XIV. Viollet-le-Duc himself turned down a proposal to add two new spires atop the towers, arguing that such a monument "would be remarkable but would not be Notre-Dame de Paris". Instead, he proposed to rebuild the original medieval spire and bell tower over the transept, which had been removed in 1786 because it was unstable in the wind. Once the project was approved, Viollet-le-Duc made drawings and photographs of the existing decorative elements; then they were removed and a stream of sculptors began making new statues of saints, gargoyles, chimeras and other architectural elements in a workshop he established, working from his drawings and photographs of similar works in other cathedrals of the same period. Other craftsmen made stained glass windows in Gothic grisaille patterns designed by Viollet-le-Duc to replace the destroyed medieval windows in the chapels of the ground floor of the nave of the cathedral. He also designed a new treasury in the Gothic style to serve as the museum of the cathedral, replacing the residence of the Archbishop, which had been destroyed in a riot in 1831. The bells in the two towers had been taken out in 1791 and melted down to make cannons. Viollet-le-Duc had new bells cast for the north tower and a new structure built inside to support them. Viollet-le-Duc and Lassus also rebuilt the sacristy, on the south side of the church, which had been built in 1756, but had been burned by rioters during the July Revolution of 1830. The new spire was completed, taller and more strongly built to withstand the weather; it was decorated with statues of the apostles, and the face of Saint Thomas bore a noticeable resemblance to Viollet-le-Duc. The spire was destroyed on 15 April 2019, as a result of the Notre-Dame de Paris fire. The restoration of Notre-Dame continued in this slow and methodical manner for twenty-five years. When not engaged in Paris, Viollet-le-Duc continued his long tours into the French provinces, inspecting, drawing, and making recommendations, and checking the progress of more than twenty different restoration projects that were under his control, including seven in Burgundy alone. His new projects included the Basilica of Saint-Sernin, Toulouse, and the Basilica of Saint-Denis just outside Paris. Saint-Denis had undergone a restoration by a different architect, Francois Debret, who had rebuilt one of the two towers. However, in 1846, the new tower, overloaded with masonry, began to crack, and Viollet-le-Duc was called in. He found no way the building could be saved; he had to oversee the demolition of the tower, saving the stones. He concentrated on restoring the interior of the church, and was able to substantially restore the original burial chamber of the Kings of France. In May 1849, he was named the architect for the restoration of Amiens Cathedral, one of the largest in France, which had been built over many centuries in a variety of different styles. He wrote, "his goal should be to save in each part of the monument its own character, and yet to make it so that the united parts don't conflict with each other; and that can be maintained in a state that is durable and simple." For his restorations of churches and cathedrals, Viollet-le-Duc designed not only architecture, but new altars and furnishings. His new furnishings were installed in the sacristy of Notre-Dame, and his neo-Gothic altar was placed in the restored Cathedral of Clermont-Ferrand. Thanks largely to Viollet-le-Duc, the neo-Gothic became the standard style for church furnishing throughout France. The French coup d'état of 1851 had transformed France from a republic to an empire and brought Napoleon III to power. The coup accelerated some of Viollet-le-Duc's projects. His patron and supervisor, Prosper Mérimée, had introduced the Emperor to the new Empress, and brought Viollet-le-Duc into close contact with the Emperor. The Emperor married the Empress Eugénie in Notre-Dame, and the government voted additional funds to advance the restoration. He moved forward with the slow work of restoration of the Cathedral of Reims and Cathedral of Amiens. In Amiens, he cleared the interior of the French classical decoration added under Louis XIV, and proposed to make it resolutely Gothic. He gave the Emperor and Empress a tour of his project in September 1853; the Empress immediately offered to pay two-thirds of the cost of the restoration. In the same year he undertook the restoration of the Château de Vincennes, long occupied by the military, along with its chapel, similar to Sainte-Chapelle. A devotee of the pure Gothic, he described the chapel as "one of the finest specimens of Gothic in decline". In November 1853, he provided the costs and plans for the medieval ramparts of Carcassonne, which he had first begun planning in 1849. The first fortifications had been built by the Visigoths; on top of these, in the Middle Ages Louis XI and then Philip the Bold had built a formidable series of towers, galleries, walls, gates and interlocking defences that resisted all sieges until 1355. The fortifications were largely intact, since the surroundings of the city were still a military defensive zone in the 19th century, but the towers were without tops and a large number of structures had been built up against the old walls. Once he obtained funding and made his plans, he began demolishing all structures which had been attached to ramparts over the centuries, and restored the gates, walls and towers to their original form, including the defence platforms, roofs on the towers and shelters for archers that would have been used during a siege. He found many of the original mountings for weapons still in place. To accompany his work, he published a detailed history of the city and its fortifications, with his drawings. Carcassonne became the best example of medieval military architecture in France, and also an important tourist attraction. Napoleon III provided additional funding for the continued restoration of Notre-Dame. Viollet-le-Duc was also to replace the great bestiary of mythical beasts and animals which had decorated the cathedral in the 18th century. In 1856, using examples from other medieval churches and debris from Notre-Dame as his model, his workshop produced dragons, chimeras, grotesques, and gargoyles, as well as an assortment of picturesque pinnacles and fleurons. He engaged in a new project for restoration of the Cathedral of Clermont-Ferrand, a project which continued for ten years. He also undertook an unusual project for Napoleon III; the design and construction of six railway coaches with neo-Gothic interior décor for the Emperor and his entourage. Two of the cars still exist; the salon of honour car, with a fresco on the ceiling, is at the Château de Compiègne, and the dining car, with a massive golden eagle as the centrepiece of the décor, is at the Railroad Museum of Mulhouse. Napoleon III asked Viollet-le-Duc if he could restore a medieval chateau for the Emperor's own use near Compiègne, where the Emperor traditionally passed September and October. Viollet-le-Duc first studied a restoration of the Château de Coucy, which had the highest medieval tower in France, later destroyed. When this proved too complicated, he settled upon Château de Pierrefonds, a castle begun by Louis of Orleans in 1396, then dismantled in 1617 after several sieges by Louis XIII of France. Napoleon bought the ruin for 5000 francs in 1812, and Mérimée declared it an historic monument in 1848. In 1857 Viollet-le-Duc began designing an entirely new chateau on the ruins. This structure was not designed to recreate anything exactly that had existed, but a castle which recaptured the spirit of the gothic, with lavish neo-gothic decoration and 19th-century comforts. While most of his attention was devoted to restorations, Viollet-le-Duc designed and built a number of private residences and new buildings in Paris. He also participated in the most important competition of the period, for the new Paris Opera. There were one hundred seventy-one projects proposed in the original competition, presented the 1855 Paris Universal Exposition. A jury of noted architects narrowed it down to five, including projects from Viollet-le-Duc, Charles Rohault de Fleury and Charles Garnier, age thirty-five. The favorites of the Emperor and Empress were de Fleury and Viollet-le-Duc, but both were eliminated in the next round. Viollet-le-Duc was not a good loser, and he dismissed Garnier's style. Garnier wrote of his rival in 1869: "Monsieur Viollet-le-Duc has produced much, but his best works without doubt are his restorations...One hesitates to appreciate his personal works. You cannot find any personality in them, only compromise. He is broken by archeology and crushed by the weight of the past. If it is difficult to learn, it is even more difficult to forget." Napoleon III called upon Viollet-le-Duc for a wide variety of archeological and architectural tasks. When he wished to put up a monument to mark the Battle of Alesia, where Julius Caesar defeated the Gauls, a battle whose actual site was disputed by historians, he asked Viollet-le-Duc to locate the exact battlefield. Viollet-le-Duc conducted excavations at various purported sites, and finally found vestiges of the walls that Caesar had built. He also designed the metal frame for the six-metre-high statue that would be placed on the site. He later designed a similar frame for a much larger statue, the Statue of Liberty, but died before that statue was finished. In 1863, Viollet-le-Duc was named a professor at the École des Beaux-Arts, the school where he had refused to become a student, and the fortress of neoclassical Beaux-Arts architecture. This launched him on a new academic career as an architectural theorist, where he would have as much influence as he did as an architect of restorations. There was much resistance from the traditional faculty, but he attracted two hundred students to his course, who applauded his lecture at the end. He had already published the first volumes of his first major work, "A Reasoned Dictionary of French Architecture". This series eventually included ten volumes, published between 1854 and 1868. But while he had many supporters, the faculty and many of the students were strongly against him. His critics complained that, aside from having little formal architectural training himself, he had only built a handful of new buildings. He tired of the confrontations and resigned on 16 May 1863, and continued his writing and teaching outside the Beaux-Arts. In the beginning of 1864, he celebrated the conclusion of his most important project, the restoration of Notre-Dame. In January of the same year he completed the first phase of the restoration of the Cathedral of Saint Sernin in Toulouse, one of the landmarks of French Romanesque architecture. Napoleon III invited Viollet-le-Duc to study possible restorations overseas, including in Algeria, Corsica, and in Mexico, where Napoleon had installed a new Emperor, Maximilien, under French sponsorship. He also saw the consecration of the third church that he had designed, the neo-Gothic Church of Saint-Denis de l'Estree, in the Paris suburb of Saint-Denis. Between 1866 and 1870, his major project was the ongoing transformation of Pierrefonds from a ruin into a royal residence. His plans for the metal framework he had designed for Pierrefonds were displayed at the Paris Universal Exposition of 1867. He also completed the tenth and final volume of his monumental dictionary of medieval architecture. He also began a new area of study, researching the geology and geography of the region around Mont Blanc in the Alps. While on his mapping excursion in the Alps in July 1870, he learned that war had been declared between Prussia and France. As the Franco-Prussian War commenced, Viollet-le-Duc hurried back to Paris, and offered his services as a military engineer; he was put into service as a colonel of engineers, preparing the defenses of Paris. In September, the Emperor was captured at the Battle of Sedan, a new Republican government took power, and the Empress Eugénie fled into exile, as Germans marched as far as Paris and put it under siege. At the same time, on September 23, Viollet-le-Duc's primary patron and supporter, Prosper Mérimée, died peacefully in the south of France. Viollet-le-Duc supervised the construction of new defensive works outside Paris. On 14 December 1870, he wrote in his journal, "Disorganization is everywhere. The officers have no confidence in the troops, and the troops have no confidence in the officers. Each day, new orders and new projects which contravene those of the day before." He fought with the French army against the Germans at Buzenval on 24 January 1871. The battle was lost, and the French capitulated on 28 January. Viollet-le-Duc wrote to his wife on February 28, "I don't know what will become of me, but I do not want to return any more to administration. I am disgusted by it forever, and want nothing more than to pass the years that remain to me in study and in the most modest possible life." In May 1871 he left his home in Paris just before national guardsmen arrived to draft him into the armed force of the Paris Commune. He returned to Pierrefonds, where he had a small apartment. Always the scholar, he wrote a detailed study of the effectiveness and deficiencies of the fortifications of Paris during the siege. He returned to the city shortly after the Commune was suppressed in May, 1871, and saw the ruins of most of the public buildings of the city, burned by the Commune in its last days. He received his only commission from the new government of the French Third Republic; Jules Simon, the new Minister of Culture and Public Instruction, asked him to design a plaque to be placed before Notre-Dame to honor the hostages killed by the Paris Commune in its final days. The new government of the French Third Republic made little use of his expertise in the restoration of the major government buildings which had been burned by the Paris Commune, including the Tuileries Palace, the Palace of the Legion of Honor, the Palais Royale, the library of the Louvre, the Ministry of Justice and the Ministry of Finance. The only reconstruction on which he was consulted was that of the Hotel de Ville. The writer Edmond de Goncourt called for leaving the ruin of the Hotel de Ville exactly as it was, "a ruin of a magical palace, A marvel of the picturesque. The country should not condemn it without appeal to restoration by Viollet-le-Duc." The government asked Viollet-le-Duc to organize a competition. He presented two options; to either restore the building to its original state, with its historic interior; or to demolish it and build a new city hall. In July 1872 the government decided to preserve the Renaissance facade, but otherwise to completely demolish and rebuild the building. In his later years he devoted most of his time to writing about architectural history. He consulted on many of his earlier projects around France, which were still underway. He also continued his explorations of the Alps around Mount Blanc, making a detailed map and a series of thirty-two drawings of the alpine scenery. He passed by Lausanne, where he was asked to prepare a plan for the restoration of the cathedral, which he did. In 1872, he completed the second volume of his major theoretical work, "Entretiens sur l'architecture". In his "Entretiens sur l'architecture" he concentrated in particular on the use of iron and other new materials, and the importance of designing buildings whose architecture was adapted to their function, rather than to a particular style. The book was translated into English in 1881 and won a large following in the United States. The Chicago architect Louis Sullivan, one of the inventors of the skyscraper, often invoked the phrase, "Form follows function." The Lausanne cathedral was his final major restoration project; it was rebuilt following his plans between 1873 and 1876. Work continued after his death. His reconstruction of the bell tower was later criticized; he eliminated the original octagonal base and added a new spire, which rested on the walls, and not on the vaulting, like the original spire. He also added new decoration, crowning the spire at mid-height with gables, another original element, and removing the original tiles. He was also criticized for the materials and ornaments he added to the towers, including gargoyles. His structural design was preserved, but in 1925 his gargoyles and original ornamentation were removed, and the spire was recovered with tiles. His reputation had reached outside of France. The spire and roof of Strasbourg Cathedral had been damaged by German artillery during the Franco-Prussian War, and the city was now part of Germany. The German government invited Viollet-le-Duc to comment on their plans for the restoration, which involved a more grandiose Romanesque tower. Viollet-le-Duc informed the German architect that his planned new tower was completely out of character with the original facade and style of the cathedral. His advice was accepted, and the church was restored to its original form. In 1872 Viollet-le-Duc was engaged in the reconstruction of the Château d'Amboise, owned by the descendants of the former King, Louis-Philippe. The chateau had been confiscated by Napoleon III in 1848 but was returned to the family in 1872. It was a massive project to turn it into a residence, involving at times three hundred workers. Viollet-le-Duc designed all the work to the finest details, including the floor tiles, the gas lights in the salons, the ovens in the kitchen, and the electric bells for summoning servants. In 1874 Viollet-le-Duc resigned as diocesan architect of Paris and was succeeded by his contemporary, Paul Abadie. In his final years, he continued to supervise the restoration projects that were underway for the Commission of Historical Monuments. He engaged in polemics about architecture in the press, and was elected to the Paris municipal council. He became engaged in the planning and construction of the Paris Universal Exposition of 1878. He proposed to the Minister of Education, Jules Ferry, that the Palais de Trocadero, the main building of the Exposition on the hilltop of Chaillot, be transformed after the Exposition into a museum of French monuments, displaying models of architecture and sculpture from landmarks around France. This idea was accepted. The National Museum of French Monuments opened in 1882, after his death. The Palais was reconstructed into the Palais de Chaillot in 1937, but the Museum of French Monuments was preserved and can be seen there today. In his final years his son Eugène-Louis became the head of the Commission of Historic Monuments. He took on just one new project, the restoration of the cloister of the Augustines at Toulouse. He completed his series of dictionaries of architectural periods, designed for a general audience. He also devoted more time to studying the geography of the Alps around Mont-Blanc. He spent his summers hiking in the mountains and writing articles about his travels. He launched a public campaign for the re-forestation of the Alps, and published a detailed map of the area in 1876. He spent more and more time at "La Vedette", the villa he constructed in Lausanne, a house on the model of a Savoyard chalet, but with a minimum of decoration, illustrating his new doctrine of form following function. He made one last visit to inspect Carcassonne, whose work was now under his son's direction. After an exhausting summer of hiking in the Alps in 1879, he became ill and died in Lausanne on 17 September 1879. He was buried in the cemetery of La Sallaz in Lausanne. In 1946 his grave and monument were transferred to the Cemetery of Bois-le-Vaux (Section XVIII) in Lausanne. Viollet-le-Duc married Elisabeth Tempier in Paris on 3 May 1834. The couple had two children, but separated a few years after marriage, and spent little time together; he was continually on the road. The writer Geneviève Viollet-le-Duc (winner of the prix Broquette-Gonin in 1978) was his great-granddaughter. Viollet-le-Duc famously defined restoration in volume eight of his "Dictionnaire raisonné de l'architecture française du XI au XVI siecle" of 1858: "To restore a building is not to maintain it, repair it or remake it: it is to re-establish it in a complete state which may never have existed at any given moment." He then explained that it had to meet four conditions: (1) The "re-establishment" had to be scientifically documented with plans and photographs and archeological records, which would guarantee exactness. (2) The restoration had to involve not just the appearance of the monument, or the effect that it produced, but also its structure; it had to use the most efficient means to assure the long life of the building, including using more solid materials, used more wisely. (3) the restoration had to exclude any modification contrary to obvious evidence; but the structure could be adapted to conform to more modern or rational uses and practices, which meant alterations to the original plan; and (4) The restoration should preserve older modifications made to the building, with the exception of those which compromised its stability or its conservation, or those which gravely violated the value of its historical presence. He drew conclusions from medieval architecture that he applied to modern architecture. He noted that it was sometimes necessary to employ an iron frame in restoration to avoid the danger of fires, as long as the new structure was not heavier than the original, and kept the original balance of forces found in medieval structures. "The monuments of the Middle Ages were carefully calculated, and their organism is delicate. There is nothing in excess in their works, nothing useless. If you change one of the conditions of these organisms, you change all the others. Many people consider this a fault; for us, this is a quality which we too often neglect in our modern construction...Why should we build expensive walls two meters thick, if walls fifty centimeters thick [with reinforced supports], offer sufficient stability? In the structure of the Middle Ages, every portion of a work fulfilled a function and possessed an action." During the entire career of Viollet-le-Duc, he was engaged in a dispute with the doctrines of the École des Beaux-Arts, the leading architectural school of France, which he refused to attend as a student, and where he taught briefly as a professor, before being pressured to depart. In 1846 he engaged in a fervent exchange in print with Quatremère de Quincy, the Perpetual Secretary of the French Academy, on the question, "Is it suitable, in the 19th century, to build churches in the gothic style?" De Quincy and his followers denounced the gothic style as incoherent, disorderly, unintelligent, decadent and without taste. Viollet-le-Duc responded, "What we want, "messieurs", is the return of an art which was born in our country...Leave to Rome what belongs to Rome, and to Athens what belongs to Athens. Rome didn't want our Gothic (and was perhaps the only one in Europe to reject it) and they were right, because when one has the good fortune to possess a national architecture, the best thing is to keep it." "If you study for a moment a church of the 13th century", he wrote, "you see that all of the construction is carried out according to an invariable system. All the forces and the weights are thrust out to the exterior, a disposition which gives the interior the greatest open space possible. The flying buttresses and contreforts alone support the entire structure, and always have an aspect of resistance, of force and stability which reassures the eye and the spirit; The vaults, built with materials that are easy to mount and to place at a great height, are combined in a easy that places the totality of their weight on the piles; that the most simple means are always employed...and that all the parts of these constructions, independent of each other, even as they rely on each other, present an elasticity and a lightness needed in a building of such great dimensions. We can still see (and this is only found in gothic architecture) that human proportions are the one fixed rule." Viollet-le-Duc was accused by his critics, in his own time and later, of pursuing the spirit of the gothic style in some of his restorations instead of strict historical accuracy. The British historian and architectural writer John Ruskin wrote in "The Seven Lamps of Architecture" in 1849: "Neither the public, nor those who are responsible for the maintenance of public monuments, understand the true meaning of 'restoration'. It signifies the most complete destruction that an edifice can suffer; a destruction from which not a single vestige can be recovered; a destruction that comes from the false description of the thing destroyed. It is impossible, as impossible as it is to bring the dead back to life, to restore whatever might have been grand or beautiful in architecture...the enterprise is a lie from the beginning to the end." Despite this scepticism, Ruskin expressed his admiration for the historical research and writings of Viollet-le-Duc. Viollet-le-Duc's restorations sometimes involved non-historical additions, either to assure the stability of the building, or sometimes simply to maintain the harmony of the design. The flèche, or spire of Notre-Dame de Paris, which had been constructed in about 1250, was removed in 1786 after it was damaged by the wind. Viollet-le-Duc designed and constructed a new spire, ornamented with statuary, which was taller than the original and modified to resist the weather, but in harmony with the rest of the design. In the 20th century, his flèche was a target for critics. He was also criticized later for his modifications of the choir of Notre-Dame, which had been rebuilt in the Louis XIV style during the reign of that king. Viollet-le-Duc took out the old choir, including the altar where Napoleon Bonaparte had been crowned Emperor and replaced them with a gothic altar and decoration which he designed. When he modified the choir, he also constructed new bays with small gothic rose windows modelled on those in the church of Chars, in the Oise Valley. Some historians condemned these restorations as non-historical invention. His defenders pointed out that Viollet-le-Duc did not make any decisions on the restoration of Notre-Dame by himself; all of his plans were approved by Prosper Mérimée, the Inspector of Historical Monuments, and by the Commission of historic monuments. He was criticized for the abundance of gothic gargoyles, chimeras, fleurons, and pinnacles which he added to Notre-Dame Cathedral. These decorations had existed in the Middle Ages but had largely been removed during the reign of Louis XIV. The last original gargoyles had been taken down in 1813. He modelled the new gargoyles and monsters on examples from other cathedrals of the period. He was later criticized also for the stained glass windows he designed and had made for the chapels around the ground level of the cathedral, which feature intricate gothic designs in grisaille, which allow more light into the church. The contemporary view of the controversy of his restoration is summarized on a descriptive panel near the altar of the cathedral: "The great restoration, carried to fruition by Viollet-le-Duc following the death of Lassus, supplied new radiance to the Cathedral – whatever reservations one might have about the choices that were made. The work of the nineteenth century is now as much a part of the architectural history of Notre-Dame as that undertaken in previous centuries." The restoration of ramparts of Carcassonne was also criticized in the 20th century. His critics pointed out that the pointed caps of the towers he constructed were more typical of northern France, not the region where Carcassonne was located, near the Spanish border. Similarly he added roofs of northern slate tiles rather than southern clay tiles, a choice that has been reversed in more recent restorations. His critics also claimed that Viollet-le-Duc sought a "condition of completeness" which never actually existed at any given time. The principal counter-argument made by Viollet-le-Duc's defenders was that, without his prompt restorations, many of the buildings that he restored would have been lost, and that he did the best that he could with the knowledge that was then available. Restoration of the Château de Pierrefonds, reinterpreted by Viollet-le-Duc for Napoleon III, was interrupted by the departure of the Emperor in 1870. Throughout his career Viollet-le-Duc made notes and drawings, not only for the buildings he was working on but also on Romanesque, Gothic and Renaissance buildings that were to be soon demolished. His notes were useful when preparing his published works. His study of medieval and Renaissance periods was not limited to architecture but extended also to such areas as furniture, clothing, musical instruments, armament, and geology. His work was published, first in serial form, and then as full-scale books, as: Viollet-le-Duc is considered by many to be the first theorist of modern architecture. Sir John Summerson wrote that "there have been two supremely eminent theorists in the history of European architecture – Leon Battista Alberti and Eugène Viollet-le-Duc." His architectural theory was largely based on finding the ideal forms for specific materials and using these forms to create buildings. His writings centered on the idea that materials should be used "honestly". He believed that the outward appearance of a building should reflect the rational construction of the building. In "Entretiens sur l'architecture", Viollet-le-Duc praised the Greek temple for its rational representation of its construction. For him, "Greek architecture served as a model for the correspondence of structure and appearance." There is speculation that this philosophy was heavily influenced by the writings of John Ruskin, who championed honesty of materials as one of the seven main emphases of architecture. Another component in Viollet-le-Duc's theory was how the design of a building should start from its program and the plan, and end with its decorations. If this resulted in an asymmetrical exterior, so be it. He dismissed the symmetry of classicist buildings as vain, caring too much about appearances at the expense of practicality and convenience for the inhabitants of the house. In several unbuilt projects for new buildings, Viollet-le-Duc applied the lessons he had derived from Gothic architecture, applying its rational structural systems to modern building materials such as cast iron. For inspiration, he also examined organic structures, such as leaves and animal skeletons. He was especially interested in the wings of bats, an influence represented by his Assembly Hall project. Viollet-le-Duc's drawings of iron trusswork were innovative for the time. Many of his designs emphasizing iron would later influence the Art Nouveau movement, most noticeably in the work of Hector Guimard, Victor Horta, Antoni Gaudí and Hendrik Petrus Berlage. His writings inspired several American architects, including Frank Furness, John Wellborn Root, Louis Sullivan, and Frank Lloyd Wright. Viollet-le-Duc had a second career in the military, primarily in the defense of Paris during the Franco-Prussian War (1870–71). He was so influenced by the conflict that during his later years he described the idealized defense of France by the analogy of the military history of Le Roche-Pont, an imaginary castle, in his work "Histoire d'une Forteresse" ("Annals of a Fortress", twice translated into English). Accessible and well researched, it is partly fictional. "Annals of a Fortress" strongly influenced French military defensive thinking. Viollet-le-Duc's critique of the effect of artillery (applying his practical knowledge from the 1870–1871 war) is so complete that it accurately describes the principles applied to the defense of France until World War II. The physical results of his theories are present in the fortification of Verdun prior to World War I and the Maginot Line prior to World War II. His theories are also represented by the French military theory of "Deliberate Advance", which stresses that artillery and a strong system of fortresses in the rear of an army are essential. The English architect Benjamin Bucknall (1833–95) was a devotee of Viollet-le-Duc and during 1874 to 1881 translated several of his publications into English to popularise his principles in Great Britain. The later works of the English designer and architect William Burges were greatly influenced by Viollet-le-Duc, most strongly in Burges's designs for his own home, The Tower House in London's Holland Park district, and Burges's designs for Castell Coch near Cardiff, Wales. An exhibition, "Eugène Viollet-le-Duc 1814–1879" was presented in Paris in 1965, and there was a larger, centennial exhibition in 1980. Viollet-le-Duc was the subject of a Google Doodle on January 27, 2014.
https://en.wikipedia.org/wiki?curid=9658
Endocarditis Endocarditis is an inflammation of the inner layer of the heart, the endocardium. It usually involves the heart valves. Other structures that may be involved include the interventricular septum, the chordae tendineae, the mural endocardium, or the surfaces of intracardiac devices. Endocarditis is characterized by lesions, known as "vegetations", which is a mass of platelets, fibrin, microcolonies of microorganisms, and scant inflammatory cells. In the subacute form of infective endocarditis, the vegetation may also include a center of granulomatous tissue, which may fibrose or calcify. There are several ways to classify endocarditis. The simplest classification is based on cause: either "infective" or "non-infective", depending on whether a microorganism is the source of the inflammation or not. Regardless, the diagnosis of endocarditis is based on clinical features, investigations such as an echocardiogram, and blood cultures demonstrating the presence of endocarditis-causing microorganisms. Signs and symptoms include fever, chills, sweating, malaise, weakness, anorexia, weight loss, splenomegaly, flu-like feeling, cardiac murmur, heart failure, petechia (red spots on the skin), Osler's nodes (subcutaneous nodules found on hands and feet), Janeway lesions (nodular lesions on palms and soles), and Roth's spots (retinal hemorrhages). Infective endocarditis is an infection of the inner surface of the heart, usually the valves. Symptoms may include fever, small areas of bleeding into the skin, heart murmur, feeling tired, and low red blood cells. Complications may include valvular insufficiency, heart failure, stroke, and kidney failure. The cause is typically a bacterial infection and less commonly a fungal infection. Risk factors include valvular heart disease including rheumatic disease, congenital heart disease, artificial valves, hemodialysis, intravenous drug use, and electronic pacemakers. The bacterial most commonly involved are streptococci or staphylococci. Diagnosis is suspected based on symptoms and supported by blood cultures or ultrasound. The usefulness of antibiotics following dental procedures for prevention is unclear. Some recommend them in those at high risk. Treatment is generally with intravenous antibiotics. The choice of antibiotics is based on the blood cultures. Occasionally heart surgery is required. The number of people affected is about 5 per 100,000 per year. Rates, however, vary between regions of the world. Males are affected more often than females. The risk of death among those infected is about 25%. Without treatment it is almost universally fatal. Nonbacterial thrombotic endocarditis (NBTE) is most commonly found on previously undamaged valves. As opposed to infective endocarditis, the vegetations in NBTE are small, sterile, and tend to aggregate along the edges of the valve or the cusps. Also unlike infective endocarditis, NBTE does not cause an inflammation response from the body. NBTE usually occurs during a hypercoagulable state such as system-wide bacterial infection, or pregnancy, though it is also sometimes seen in patients with venous catheters. NBTE may also occur in patients with cancers, particularly mucinous adenocarcinoma where Trousseau syndrome can be encountered. Typically NBTE does not cause many problems on its own, but parts of the vegetations may break off and embolize to the heart or brain, or they may serve as a focus where bacteria can lodge, thus causing infective endocarditis. Another form of sterile endocarditis is termed Libman–Sacks endocarditis; this form occurs more often in patients with lupus erythematosus and is thought to be due to the deposition of immune complexes. Like NBTE, Libman-Sacks endocarditis involves small vegetations, while infective endocarditis is composed of large vegetations. These immune complexes precipitate an inflammation reaction, which helps to differentiate it from NBTE. Also unlike NBTE, Libman-Sacks endocarditis does not seem to have a preferred location of deposition and may form on the undersurfaces of the valves or even on the endocardium.
https://en.wikipedia.org/wiki?curid=9659
Euler's sum of powers conjecture Euler's conjecture is a disproved conjecture in mathematics related to Fermat's last theorem. It was proposed by Leonhard Euler in 1769. It states that for all integers and greater than 1, if the sum of th powers of positive integers is itself a th power, then is greater than or equal to : The conjecture represents an attempt to generalize Fermat's last theorem, which is the special case : if , then . Although the conjecture holds for the case (which follows from Fermat's last theorem for the third powers), it was disproved for and . It is unknown whether the conjecture fails or holds for any value . Euler was aware of the equality involving sums of four fourth powers; this however is not a counterexample because no term is isolated on one side of the equation. He also provided a complete solution to the four cubes problem as in Plato's number or the taxicab number 1729. The general solution of the equation is where and are any integers. Euler's conjecture was disproven by L. J. Lander and T. R. Parkin in 1966 when, through a direct computer search on a CDC 6600, they found a counterexample for . This was published in a paper comprising just two sentences. A total of three primitive (that is, in which the summands do not all have a common factor) counterexamples are known: In 1986, Noam Elkies found a method to construct an infinite series of counterexamples for the case. His smallest counterexample was A particular case of Elkies' solutions can be reduced to the identity where This is an elliptic curve with a rational point at . From this initial rational point, one can compute an infinite collection of others. Substituting into the identity and removing common factors gives the numerical example cited above. In 1988, Roger Frye found the smallest possible counterexample for by a direct computer search using techniques suggested by Elkies. This solution is the only one with values of the variables below 1,000,000. In 1967, L. J. Lander, T. R. Parkin, and John Selfridge conjectured that if where are positive integers for all and , then . In the special case , the conjecture states that if (under the conditions given above) then . The special case may be described as the problem of giving a partition of a perfect power into few like powers. For and or , there are many known solutions. Some of these are listed below. As of 2002, there are no solutions for formula_6 whose final term is ≤ 730000. This is the smallest solution to the problem by R. Norrie.
https://en.wikipedia.org/wiki?curid=9660
Book of Exodus The Book of Exodus is the second book of the Bible, It describes the Israelites' deliverance from slavery in Egypt through the hand of Yahweh, the revelations at biblical Mount Sinai, and the subsequent "divine indwelling" of God with Israel. Exodus is traditionally ascribed to Moses, but modern scholars see its initial composition as a product of the Babylonian exile (6th century BCE), with final revisions in the Persian post-exilic period (5th century BCE). Carol Meyers, in her commentary on Exodus, suggests that it is arguably the most important book in the Bible, as it presents the defining features of Israel's identity—memories of a past marked by hardship and escape, a binding covenant with God, who chooses Israel, and the establishment of the life of the community and the guidelines for sustaining it. The English name "Exodus" comes from the , "éxodos", meaning "going out", from ἐξ- (ex, “out”) + ὁδός (hodós, “path, road”). In Hebrew the book's title is שְׁמוֹת, "shemōt", "Names", from the beginning words of the text: "These are the names of the sons of Israel" (). There is no unanimous agreement among scholars on the structure of Exodus. One strong possibility is that it is a diptych (i.e., divided into two parts), with the division between parts 1 and 2 at the crossing of the Red Sea or at the beginning of the theophany (appearance of God) in chapter 19. On this plan, the first part tells of God's rescue of his people from Egypt and their journey under his care to Sinai (chapters 1–19) and the second tells of the covenant between them (chapters 20–40). Jacob's sons and their families join their brother Joseph in Egypt, where they begin to grow in number. Egypt's new Pharaoh who doesn't remember how Joseph had saved Egypt from famine years ago, is fearful that the Israelites could become a fifth column so he forces the Israelites into slavery and orders the throwing of all newborn boys into the Nile. A Levite woman (Jochebed, according to other sources) saves her baby by setting him adrift on the river Nile in an ark of bulrushes. Pharaoh's daughter finds the child, names him Moses, and brings him up as her own. Aware of his origins, an adult Moses kills an Egyptian overseer beating a Hebrew slave and flees into Midian to escape punishment. There he marries Zipporah, daughter of Midianite priest Jethro, and encounters God in a burning bush. Moses asks God for his name, to which God replies: "I Am that I Am," the book's explanation for the origins of the name Yahweh, as God is thereafter known. God tells Moses to return to Egypt and lead the Hebrews into Canaan, the land promised to Abraham. On the way to Egypt God seeks to kill Moses, but Zipporah saves his life. Moses returns to Egypt and fails to convince the Pharaoh to release the Israelites. God smites the Egyptians with ten terrible plagues, among them a , an outbreak of frogs, and the death of all firstborn sons. Moses leads the Israelites out of bondage, prompting a final pursuit through the Red Sea by the forces of Pharaoh when Pharaoh reneges on his coerced consent. As desert life proves arduous, the Israelites complain and long for Egypt, but God miraculously provides manna for them to eat and water to drink. The Israelites arrive at the mountain of God, where Moses's father-in-law Jethro visits Moses; at his suggestion Moses appoints judges over Israel. God asks whether they will agree to be his people. They accept. The people gather at the foot of the mountain, and with thunder and lightning, fire and clouds of smoke, the sound of trumpets, and the trembling of the mountain, God appears on the peak, and the people see the cloud and hear the voice (or possibly sound) of God. God tells Moses to ascend the mountain. God pronounces the Ten Commandments (the Ethical Decalogue) in the hearing of all Israel. Moses goes up the mountain into the presence of God, who pronounces the Covenant Code of ritual and civil law and promises Canaan to them if they obey. Moses comes down the mountain and writes down God's words, and the people agree to keep them. God calls Moses up the mountain, where he remains for forty days and forty nights, at the conclusion of which he returns, bearing the set of stone tablets. God gives Moses instructions for the construction of the tabernacle so that God may dwell permanently among his chosen people, along with instructions for the priestly vestments, the altar and its appurtenances, procedures for ordination of priests, and the daily sacrifice offerings. Aaron becomes the first hereditary high priest. God gives Moses the two tablets of stone containing the words of the ten commandments, written with the "finger of God". While Moses is with God, Aaron casts a golden calf, which the people worship. God informs Moses of their apostasy and threatens to kill them all, but relents when Moses pleads for them. Moses comes down from the mountain, smashes the stone tablets in anger, and commands the Levites to massacre the unfaithful Israelites. God commands Moses to construct two new tablets. Moses ascends the mountain again, where God dictates the Ten Commandments for Moses to write on the tablets. Moses descends from the mountain with a transformed face; from that time onwards he must hide his face with a veil. Moses assembles the Hebrews and repeats to them the commandments he has received from God, which are to keep the Sabbath and to construct the Tabernacle. The Israelites do as they are commanded. From that time God dwells in the Tabernacle and orders the travels of the Hebrews. Jewish and Christian tradition viewed Moses as the author of Exodus and the entire Torah, but by the end of the 19th century the increasing awareness of discrepancies, inconsistencies, repetitions and other features of the Pentateuch had led scholars to abandon this idea. In approximate round dates, the process which produced Exodus and the Pentateuch probably began around 600 BCE when existing oral and written traditions were brought together to form books recognisable as those we know, reaching their final form as unchangeable sacred texts around 400 BCE. Although mythical elements are not so prominent in Exodus as in Genesis, ancient legends may have an influence on the book's form or content: for example, the story of the infant Moses's salvation from the Nile is argued to be based on an earlier legend of king Sargon of Akkad, while the story of the parting of the Red Sea may trade on Mesopotamian creation mythology. Similarly, the Covenant Code (the law code in Exodus 20:22–23:33) has some similarities in both content and structure with the Laws of Hammurabi. These influences serve to reinforce the conclusion that the Book of Exodus originated in the exiled Jewish community of 6th-century BCE Babylon, but not all the sources are Mesopotamian: the story of Moses's flight to Midian following the murder of the Egyptian overseer may draw on the Egyptian "Story of Sinuhe". Biblical scholars describe the Bible's theologically-motivated history writing as "salvation history", meaning a history of God's saving actions that give identity to Israel – the promise of offspring and land to the ancestors, the Exodus from Egypt (in which God saves Israel from slavery), the wilderness wandering, the revelation at Sinai, and the hope for the future life in the promised land. A theophany is a manifestation (appearance) of a god – in the Bible, an appearance of the God of Israel, accompanied by storms – the earth trembles, the mountains quake, the heavens pour rain, thunder peals and lightning flashes. The theophany in Exodus begins "the third day" from their arrival at Sinai in chapter 19: Yahweh and the people meet at the mountain, God appears in the storm and converses with Moses, giving him the Ten Commandments while the people listen. The theophany is therefore a public experience of divine law. The second half of Exodus marks the point at which, and describes the process through which, God's theophany becomes a permanent presence for Israel via the Tabernacle. That so much of the book (chapters 25–31, 35–40) describes the plans of the Tabernacle demonstrates the importance it played in the perception of Second Temple Judaism at the time of the text's redaction by the Priestly writers: the Tabernacle is the place where God is physically present, where, through the priesthood, Israel could be in direct, literal communion with him. The heart of Exodus is the Sinaitic covenant. A covenant is a legal document binding two parties to take on certain obligations towards each other. There are several covenants in the Bible, and in each case they exhibit at least some of the elements in real-life treaties of the ancient Middle East: a preamble, historical prologue, stipulations, deposition and reading, list of witnesses, blessings and curses, and ratification by animal sacrifice. Biblical covenants, in contrast to Eastern covenants in general, are between a god, Yahweh, and a people, Israel, instead of between a strong ruler and a weaker vassal. God elects Israel for salvation because the "sons of Israel" are "the firstborn son" of the God of Israel, descended through Shem and Abraham to the chosen line of Jacob whose name is changed to Israel. The goal of the divine plan in Exodus is a return to humanity's state in Eden, so that God can dwell with the Israelites as he had with Adam and Eve through the Ark and Tabernacle, which together form a model of the universe; in later Abrahamic religions Israel becomes the guardian of God's plan for humanity, to bring "God's creation blessing to mankind" begun in Adam. The overwhelming consensus among scholars is that the Exodus story is best understood as a myth and does not accurately describe historical events. While archaeology has found traces left by even small bands of hunter-gatherers in the Sinai, there is no evidence at all for the large body of people described in the Exodus story: "The conclusion – that Exodus did not happen at the time and in the manner described in the Bible – seems irrefutable [...] repeated excavations and surveys throughout the entire area have not provided even the slightest evidence" (Finkelstein and Silberman, 2001). Instead, modern archaeology suggests continuity between Canaanite and Israelite settlement, indicating a primarily Canaanite origin for Israel. While all but the most conservative scholars reject the biblical account of the Exodus, a majority still believes that the story has some historical basis, with Kenton Sparks referring to it as "mythologized history." Some scholars, associated with the interpretative school of Biblical minimalism, show more skepticism towards ascribing any historicity to the Exodus, with some arguing that the myth has its origins in the exilic and post-exilic Jewish community and has little to no historical basis.
https://en.wikipedia.org/wiki?curid=9662
Electronics Electronics comprises the physics, engineering, technology and applications that deal with the emission, flow and control of electrons in vacuum and matter. It uses active devices to control electron flow by amplification and rectification, which distinguishes it from classical electrical engineering which uses passive effects such as resistance, capacitance and inductance to control current flow. Electronics has had a major effect on the development of modern society. The identification of the electron in 1897, along with the subsequent invention of the vacuum tube which could amplify and rectify small electrical signals, inaugurated the field of electronics and the electron age. This distinction started around 1906 with the invention by Lee De Forest of the triode, which made electrical amplification of weak radio signals and audio signals possible with a non-mechanical device. Until 1950, this field was called "radio technology" because its principal application was the design and theory of radio transmitters, receivers, and vacuum tubes. The term "solid-state electronics" emerged after the first working transistor was invented by William Shockley, Walter Houser Brattain and John Bardeen at Bell Labs in 1947. The MOSFET (MOS transistor) was later invented by Mohamed Atalla and Dawon Kahng at Bell Labs in 1959. The MOSFET was the first truly compact transistor that could be miniaturised and mass-produced for a wide range of uses, revolutionizing the electronics industry, and playing a central role in the microelectronics revolution and Digital Revolution. The MOSFET has since become the basic element in most modern electronic equipment, and is the most widely used electronic device in the world. Electronics is widely used in information processing, telecommunication, and signal processing. The ability of electronic devices to act as switches makes digital information-processing possible. Interconnection technologies such as circuit boards, electronics packaging technology, and other varied forms of communication infrastructure complete circuit functionality and transform the mixed electronic components into a regular working system, called an electronic system; examples are computers or control systems. An electronic system may be a component of another engineered system or a standalone device. most electronic devices use semiconductor components to perform electron control. Commonly, electronic devices contain circuitry consisting of active semiconductors supplemented with passive elements; such a circuit is described as an electronic circuit. Electronics deals with electrical circuits that involve active electrical components such as vacuum tubes, transistors, diodes, integrated circuits, optoelectronics, and sensors, associated passive electrical components, and interconnection technologies. The nonlinear behaviour of active components and their ability to control electron flows makes amplification of weak signals possible. The study of semiconductor devices and related technology is considered a branch of solid-state physics, whereas the design and construction of electronic circuits to solve practical problems come under electronics engineering. This article focuses on engineering aspects of electronics. Electronics has branches as follows: An electronic component is any physical entity in an electronic system used to affect the electrons or their associated fields in a manner consistent with the intended function of the electronic system. Components are generally intended to be connected together, usually by being soldered to a printed circuit board (PCB), to create an electronic circuit with a particular function (for example an amplifier, radio receiver, or oscillator). Components may be packaged singly, or in more complex groups as integrated circuits. Some common electronic components are capacitors, inductors, resistors, diodes, transistors, etc. Components are often categorized as active (e.g. transistors and thyristors) or passive (e.g. resistors, diodes, inductors and capacitors). Vacuum tubes (Thermionic valves) were among the earliest electronic components. They were almost solely responsible for the electronics revolution of the first half of the twentieth century. They allowed for vastly more complicated systems and gave us radio, television, phonographs, radar, long-distance telephony and much more. They played a leading role in the field of microwave and high power transmission as well as television receivers until the middle of the 1980s. Since that time, solid-state devices have all but completely taken over. Vacuum tubes are still used in some specialist applications such as high power RF amplifiers, cathode ray tubes, specialist audio equipment, guitar amplifiers and some microwave devices. The first working point-contact transistor was invented by John Bardeen and Walter Houser Brattain at Bell Labs in 1947. In April 1955, the IBM 608 was the first IBM product to use transistor circuits without any vacuum tubes and is believed to be the first all-transistorized calculator to be manufactured for the commercial market. The 608 contained more than 3,000 germanium transistors. Thomas J. Watson Jr. ordered all future IBM products to use transistors in their design. From that time on transistors were almost exclusively used for computer logic and peripherals. However, early junction transistors were relatively bulky devices that were difficult to manufacture on a mass-production basis, which limited them to a number of specialised applications. The MOSFET (MOS transistor) was invented by Mohamed Atalla and Dawon Kahng at Bell Labs in 1959. The MOSFET was the first truly compact transistor that could be miniaturised and mass-produced for a wide range of uses. Its advantages include high scalability, affordability, low power consumption, and high density. It revolutionized the electronics industry, becoming the most widely used electronic device in the world. The MOSFET is the basic element in most modern electronic equipment, and has been central to the electronics revolution, the microelectronics revolution, and the Digital Revolution. The MOSFET has thus been credited as the birth of modern electronics, and possibly the most important invention in electronics. Circuits and components can be divided into two groups: analog and digital. A particular device may consist of circuitry that has one or the other or a mix of the two types. Most analog electronic appliances, such as radio receivers, are constructed from combinations of a few types of basic circuits. Analog circuits use a continuous range of voltage or current as opposed to discrete levels as in digital circuits. The number of different analog circuits so far devised is huge, especially because a 'circuit' can be defined as anything from a single component, to systems containing thousands of components. Analog circuits are sometimes called linear circuits although many non-linear effects are used in analog circuits such as mixers, modulators, etc. Good examples of analog circuits include vacuum tube and transistor amplifiers, operational amplifiers and oscillators. One rarely finds modern circuits that are entirely analog. These days analog circuitry may use digital or even microprocessor techniques to improve performance. This type of circuit is usually called "mixed signal" rather than analog or digital. Sometimes it may be difficult to differentiate between analog and digital circuits as they have elements of both linear and non-linear operation. An example is the comparator which takes in a continuous range of voltage but only outputs one of two levels as in a digital circuit. Similarly, an overdriven transistor amplifier can take on the characteristics of a controlled switch having essentially two levels of output. In fact, many digital circuits are actually implemented as variations of analog circuits similar to this example – after all, all aspects of the real physical world are essentially analog, so digital effects are only realized by constraining analog behavior. Digital circuits are electric circuits based on a number of discrete voltage levels. Digital circuits are the most common physical representation of Boolean algebra, and are the basis of all digital computers. To most engineers, the terms "digital circuit", "digital system" and "logic" are interchangeable in the context of digital circuits. Most digital circuits use a binary system with two voltage levels labeled "0" and "1". Often logic "0" will be a lower voltage and referred to as "Low" while logic "1" is referred to as "High". However, some systems use the reverse definition ("0" is "High") or are current based. Quite often the logic designer may reverse these definitions from one circuit to the next as he sees fit to facilitate his design. The definition of the levels as "0" or "1" is arbitrary. Ternary (with three states) logic has been studied, and some prototype computers made. Computers, electronic clocks, and programmable logic controllers (used to control industrial processes) are constructed of digital circuits. Digital signal processors are another example. Building blocks: Highly integrated devices: Heat generated by electronic circuitry must be dissipated to prevent immediate failure and improve long term reliability. Heat dissipation is mostly achieved by passive conduction/convection. Means to achieve greater dissipation include heat sinks and fans for air cooling, and other forms of computer cooling such as water cooling. These techniques use convection, conduction, and radiation of heat energy. Electronic noise is defined as unwanted disturbances superposed on a useful signal that tend to obscure its information content. Noise is not the same as signal distortion caused by a circuit. Noise is associated with all electronic circuits. Noise may be electromagnetically or thermally generated, which can be decreased by lowering the operating temperature of the circuit. Other types of noise, such as shot noise cannot be removed as they are due to limitations in physical properties. Mathematical methods are integral to the study of electronics. To become proficient in electronics it is also necessary to become proficient in the mathematics of circuit analysis. Circuit analysis is the study of methods of solving generally linear systems for unknown variables such as the voltage at a certain node or the current through a certain branch of a network. A common analytical tool for this is the SPICE circuit simulator. Also important to electronics is the study and understanding of electromagnetic field theory. Due to the complex nature of electronics theory, laboratory experimentation is an important part of the development of electronic devices. These experiments are used to test or verify the engineer's design and detect errors. Historically, electronics labs have consisted of electronics devices and equipment located in a physical space, although in more recent years the trend has been towards electronics lab simulation software, such as CircuitLogix, Multisim, and PSpice. Today's electronics engineers have the ability to design circuits using premanufactured building blocks such as power supplies, semiconductors (i.e. semiconductor devices, such as transistors), and integrated circuits. Electronic design automation software programs include schematic capture programs and printed circuit board design programs. Popular names in the EDA software world are NI Multisim, Cadence (ORCAD), EAGLE PCB and Schematic, Mentor (PADS PCB and LOGIC Schematic), Altium (Protel), LabCentre Electronics (Proteus), gEDA, KiCad and many others. Many different methods of connecting components have been used over the years. For instance, early electronics often used point to point wiring with components attached to wooden breadboards to construct circuits. Cordwood construction and wire wrap were other methods used. Most modern day electronics now use printed circuit boards made of materials such as FR4, or the cheaper (and less hard-wearing) Synthetic Resin Bonded Paper (SRBP, also known as Paxoline/Paxolin (trade marks) and FR2) – characterised by its brown colour. Health and environmental concerns associated with electronics assembly have gained increased attention in recent years, especially for products destined to the European Union, with its Restriction of Hazardous Substances Directive (RoHS) and Waste Electrical and Electronic Equipment Directive (WEEE), which went into force in July 2006. Electronic systems design deals with the multi-disciplinary design issues of complex electronic devices and systems, such as mobile phones and computers. The subject covers a broad spectrum, from the design and development of an electronic system (new product development) to assuring its proper function, service life and disposal. Electronic systems design is therefore the process of defining and developing complex electronic devices to satisfy specified requirements of the user. Electrical components are generally mounted in the following ways: The electronics industry consists of various sectors. The central driving force behind the entire electronics industry is the semiconductor industry sector, which has annual sales of over as of 2018. The largest industry sector is e-commerce, which generated over in 2017. The most widely manufactured electronic device is the metal-oxide-semiconductor field-effect transistor (MOSFET), with an estimated 13sextillion MOSFETs having been manufactured between 1960 and 2018.
https://en.wikipedia.org/wiki?curid=9663
Erewhon Erewhon: or, Over the Range () is a novel by Samuel Butler which was first published anonymously in 1872, set in a fictional country discovered and explored by the protagonist. Butler meant the title to be understood as the word "nowhere" backwards even though the letters "h" and "w" are transposed. The book is a satire on Victorian society. The first few chapters of the novel dealing with the discovery of Erewhon are in fact based on Butler's own experiences in New Zealand, where, as a young man, he worked as a sheep farmer on Mesopotamia Station for about four years (1860–64), and explored parts of the interior of the South Island and which he wrote about in his "A First Year in Canterbury Settlement" (1863). The greater part of the book consists of a description of Erewhon. The nature of this nation is intended to be ambiguous. At first glance, Erewhon appears to be a Utopia, yet it soon becomes clear that this is far from the case. Yet for all the failings of Erewhon, it is also clearly not a dystopia, such as that depicted in George Orwell's "Nineteen Eighty-Four". As a satirical utopia, "Erewhon" has sometimes been compared to "Gulliver's Travels" (1726), a classic novel by Jonathan Swift; the image of Utopia in this latter case also bears strong parallels with the self-view of the British Empire at the time. It can also be compared to the William Morris novel, "News from Nowhere". "Erewhon" satirises various aspects of Victorian society, including criminal punishment, religion and anthropocentrism. For example, according to Erewhonian law, offenders are treated as if they were ill, whereas ill people are looked upon as criminals. Another feature of Erewhon is the absence of machines; this is due to the widely shared perception by the Erewhonians that they are potentially dangerous. This last aspect of "Erewhon" reveals the influence of Charles Darwin's evolution theory; Butler had read "On the Origin of Species" soon after it was published in 1859. Butler developed the three chapters of "Erewhon" that make up "The Book of the Machines" from a number of articles that he had contributed to "The Press", which had just begun publication in Christchurch, New Zealand, beginning with "Darwin among the Machines" (1863). Butler was the first to write about the possibility that machines might develop consciousness by natural selection. Many dismissed this as a joke; but, in his preface to the second edition, Butler wrote, "I regret that reviewers have in some cases been inclined to treat the chapters on Machines as an attempt to reduce Mr. Darwin's theory to an absurdity. Nothing could be further from my intention, and few things would be more distasteful to me than any attempt to laugh at Mr. Darwin." After its first release, this book sold far better than any of Butler's other works, perhaps because the British public assumed that the anonymous author was some better-known figure (the favourite being Edward Bulwer-Lytton, who had published "The Coming Race" two years previously). In a 1945 broadcast, George Orwell praised the book and said that when Butler wrote "Erewhon" it needed "imagination of a very high order to see that machinery could be dangerous as well as useful." He recommended the novel, though not its sequel, "Erewhon Revisited". The French philosopher Gilles Deleuze used ideas from Butler's book at various points in the development of his philosophy of difference. In "Difference and Repetition" (1968), Deleuze refers to what he calls "Ideas" as "Erewhon". "Ideas are not concepts", he argues, but rather "a form of eternally positive differential multiplicity, distinguished from the identity of concepts." "Erewhon" refers to the "nomadic distributions" that pertain to simulacra, which "are not universals like the categories, nor are they the "hic et nunc" or "nowhere", the diversity to which categories apply in representation." "Erewhon", in this reading, is "not only a disguised "no-where" but a rearranged "now-here"." In his collaboration with Félix Guattari, "Anti-Oedipus" (1972), Deleuze draws on Butler's "The Book of the Machines" to "go beyond" the "usual polemic between vitalism and mechanism" as it relates to their concept of "desiring-machines": In 1994, a group of ex-Yugoslavian writers in Amsterdam, who had established the PEN centre of Yugoslav Writers in Exile, published a single issue of a literary journal "Erewhon". New Zealand sound art organisation, the Audio Foundation, published in 2012 an anthology edited by Bruce Russell named "Erewhon Calling" after Butler's book. In 2014, New Zealand artist Gavin Hipkins released his first feature film, titled "Erewhon" and based on Butler's book. It premiered at the New Zealand International Film Festival and the Edinburgh Art Festival. In "Smile", the second episode of the 2017 season of "Doctor Who", the Doctor and Bill explore a spaceship named "Erehwon". Despite the slightly different spelling, the episode writer Frank Cottrell-Boyce confirmed that this was a reference to Butler's novel. 'The Butlerian Jihad' is the name of the crusade to wipeout 'thinking machines' in the novel, Dune, by Frank Herbert. 'Erewhon' is the name of Los Angeles based natural foods grocery store. 'Erewhon' is also the name of an independent speculative fiction publishing company founded in 2018 by Liz Gorinsky.
https://en.wikipedia.org/wiki?curid=9664
Entorhinal cortex The entorhinal cortex (EC) is an area of the brain located in the medial temporal lobe and functions as a hub in a widespread network for memory, navigation and the perception of time. The EC is the main interface between the hippocampus and neocortex. The EC-hippocampus system plays an important role in declarative (autobiographical/episodic/semantic) memories and in particular spatial memories including memory formation, memory consolidation, and memory optimization in sleep. The EC is also responsible for the pre-processing (familiarity) of the input signals in the reflex nictitating membrane response of classical trace conditioning; the association of impulses from the eye and the ear occurs in the entorhinal cortex. In rodents, the EC is located at the caudal end of the temporal lobe. In primates it is located at the rostral end of the temporal lobe and stretches dorsolaterally. It is usually divided into medial and lateral regions with three bands with distinct properties and connectivity running perpendicular across the whole area. A distinguishing characteristic of the EC is the lack of cell bodies where layer IV should be; this layer is called the "lamina dissecans". The superficial layers – layers II and III – of EC project to the dentate gyrus and hippocampus: Layer II projects primarily to dentate gyrus and hippocampal region CA3; layer III projects primarily to hippocampal region CA1 and the subiculum. These layers receive input from other cortical areas, especially associational, perirhinal, and parahippocampal cortices, as well as prefrontal cortex. EC as a whole, therefore, receives highly processed input from every sensory modality, as well as input relating to ongoing cognitive processes, though it should be stressed that, within EC, this information remains at least partially segregated. The deep layers, especially layer V, receive one of the three main outputs of the hippocampus and, in turn, reciprocate connections from other cortical areas that project to superficial EC. The rodent entorhinal cortex shows a modular organization, with different properties and connections in different areas. In 2005, it was discovered that entorhinal cortex contains a neural map of the spatial environment in rats. In 2014, John O'Keefe, May-Britt Moser and Edvard Moser received the Nobel Prize in Physiology or Medicine, partly because of this discovery. Neurons in the lateral entorhinal cortex exhibit little spatial selectivity, whereas neurons of the medial entorhinal cortex (MEC), exhibit multiple "place fields" that are arranged in a hexagonal pattern, and are, therefore, called "grid cells". These fields and spacing between fields increase from the dorso-lateral MEA to the ventro-medial MEA. The same group of researchers have found speed cells in the medial entorhinal cortex of rats. The speed of movement is translated from proprioceptive information and is represented as firing rates in these cells. The cells are known to fire in correlation to future speed of the rodent. Single-unit recording of neurons in humans playing video games find path cells in the EC, the activity of which indicates whether a person is taking a clockwise or counterclockwise path. Such EC "direction" path cells show this directional activity irrespective of the location of where a person experiences themselves, which contrasts them to place cells in the hippocampus, which are activated by specific locations. EC neurons process general information such as directional activity in the environment, which contrasts to that of the hippocampal neurons, which usually encode information about specific places. This suggests that EC encodes general properties about current contexts that are then used by hippocampus to create unique representations from combinations of these properties. Research generally highlights a useful distinction in which the medial entorhinal cortex (MEC) mainly supports processing of space, whereas the lateral entorhinal cortex (LEC) mainly supports the processing of time. Individual variation in the volume of EC is linked to taste perception. People with a larger EC in the left hemisphere found quinine, the source of bitterness in tonic water, less bitter. The entorhinal cortex is the first area of the brain to be affected in Alzheimer's disease; a recent functional magnetic resonance imaging study has localised the area to the lateral entorhinal cortex. Lopez "et al." have shown, in a multimodal study, that there are differences in the volume of the left entorhinal cortex between progressing (to Alzheimer's disease) and stable mild cognitive impairment patients. These authors also found that the volume of the left entorhinal cortex inversely correlates with the level of alpha band phase synchronization between the right anterior cingulate and temporo-occipital regions. In 2012, neuroscientists at UCLA conducted an experiment using a virtual taxi video game connected to seven epilepsy patients with electrodes already implanted in their brains, allowing the researchers to monitor neuronal activity whenever memories were being formed. As the researchers stimulated the nerve fibers of each of the patients' entorhinal cortex as they were learning, they were then able to better navigate themselves through various routes and recognize landmarks more quickly. This signified an improvement in the patients' spatial memory. A study finds that regardless of gender, young adults who have greater aerobic fitness also have greater volume of their entorhinal cortex. It suggests that aerobic exercise may have a positive effect on the medial temporal lobe memory system (which includes the entorhinal cortex) in healthy young adults. This also suggests that exercise training, when designed to increase aerobic fitness, might have a positive effect on the brain in healthy young adults For delineating the Entorhinal cortex, see Desikan RS, Ségonne F, Fischl B, Quinn BT, Dickerson BC, Blacker D, Buckner RL, Dale AM, Maguire RP, Hyman BT, Albert MS, Killiany RJ. An automated labeling system for subdividing the human cerebral cortex on MRI scans into gyral based regions of interest. Neuroimage. 2006 Jul 1;31(3):968-80.
https://en.wikipedia.org/wiki?curid=9667
Ernst Haeckel Ernst Heinrich Philipp August Haeckel (; 16 February 1834 – 9 August 1919) was a German zoologist and an influential promoter of the scientific racism that led to murderous results during the 20th century. Haeckel promoted and popularized Charles Darwin's work in Germany but he embraced a misinterpretation of Darwin called Social Darwinism that promoted the idea that "survival of the fittest" was a natural law that could be applied to human societies, and that this struggle resulted in "improvement" of some races of human beings. He advocated for the killing of 200,000 mentally and congenitally ill people by a medical control board. Aspects of Haeckel's other work, in zoology, resulted in much more benign, helpful, and productive results. He was able to describe and name thousands of species, map a genealogical tree relating to many life forms, and to invent a number of biological terms, including "ecology", and "phylum". Ernst Haeckel was born on 16 February 1834, in Potsdam (then part of the Kingdom of Prussia). In 1852 Haeckel completed studies at the "Domgymnasium", the cathedral high-school of Merseburg. He then studied medicine in Berlin and Würzburg, particularly with Albert von Kölliker, Franz Leydig, Rudolf Virchow (with whom he later worked briefly as assistant), and with the anatomist-physiologist Johannes Peter Müller (1801–1858). Together with Hermann Steudner he attended botany lectures in Würzburg. In 1857 Haeckel attained a doctorate in medicine, and afterwards he received the license to practice medicine. The occupation of physician appeared less worthwhile to Haeckel after contact with suffering patients. Ernst Haeckel studied under Karl Gegenbaur at the University of Jena for three years, earning a habilitation in comparative anatomy in 1861, before becoming a professor of zoology at Jena, where he remained for 47 years, from 1862 to 1909. Between 1859 and 1866 Haeckel worked on many phyla, such as radiolarians, poriferans (sponges) and annelids (segmented worms). During a trip to the Mediterranean, Haeckel named nearly 150 new species of radiolarians. From 1866 to 1867 Haeckel made an extended journey to the Canary Islands with Hermann Fol. During this period, he met with Charles Darwin (in 1866 at Down House in Kent), Thomas Huxley and Charles Lyell. In 1867 he married Agnes Huschke. Their son Walter was born in 1868, their daughters Elizabeth in 1871 and Emma in 1873. In 1869 he traveled as a researcher to Norway, in 1871 to Croatia (where he lived on the island of Hvar in a monastery), and in 1873 to Egypt, Turkey, and Greece. In 1907 he had a museum built in Jena to teach the public about evolution. Haeckel retired from teaching in 1909, and in 1910 he withdrew from the Evangelical Church of Prussia. On the occasion of his 80th birthday celebration he was presented with a two-volume work entitled "Was wir Ernst Haeckel verdanken (What We Owe to Ernst Haeckel)", edited at the request of the German Monistenbund by Heinrich Schmidt of Jena. Haeckel's wife, Agnes, died in 1915, and he became substantially frailer, breaking his leg and arm. He sold his "Villa Medusa" in Jena in 1918 to the Carl Zeiss foundation, which preserved his library. Haeckel died on 9 August 1919. Haeckel became the most famous proponent of Monism in Germany. Haeckel's affinity for the German Romantic movement, coupled with his acceptance of a form of Lamarckism, influenced his political beliefs. Rather than being a strict Darwinian, Haeckel believed that the characteristics of an organism were acquired through interactions with the environment and that ontogeny reflected phylogeny. He saw the social sciences as instances of "applied biology", and that phrase was picked up and used for Nazi propaganda. In 1906 Haeckel founded a group called the Monist League () to promote his religious and political beliefs. This group lasted until 1933 and included such notable members as Wilhelm Ostwald, Georg von Arco (1869-1940), Helene Stöcker and Walter Arthur Berendsohn. He was the first person to use the term "first world war". Many historians have discussed both the direct and the divergent ways in which Haeckel's social Darwinism influenced National Socialist ideology. Haeckel was a racist and an influential promoter of scientific racism. He was a social Darwinist who believed that "survival of the fittest" was a natural law, and that struggle somehow led to some vaguely or romantically conceived "improvement" of a so-called "race" of human beings. He advocated for the murder of 200,000 mentally and congenitally ill by a medical control board. This idea was later put into practice by the Third Reich, as part of the Aktion T4 program. Alfred Ploetz, founder of the German Society for Racial Hygiene, praised Haeckel repeatedly, and invited him to become an honorary member, an invitation which Haeckel accepted. Haeckel also believed that Germany should be governed by an authoritarian political system, and that inequalities both within and between societies were an inevitable product of evolutionary law. Haeckel was also an extreme German nationalist who believed strongly in the superiority of German culture. Nonetheless, in the racial hierarchies Haeckel constructed, Jews tended to appear closer to the top, rather than closer to the bottom as in National Socialist racial thought. The Nazis directly discussed Haeckel's influence. SS captain and biologist Heinz Brücher wrote a biography of Haeckel in 1936, in which he praised Haeckel as a "pioneer in biological state thinking". In 1934, the journal, "Der Biologie" celebrated Haeckel's 100th birthday with several essays acclaiming him as a pioneering thinker of National Socialism. Other Nazis kept their distance from Haeckel. Nazi propaganda guidelines issued in 1935 listed books which popularized Darwin and evolution on an "expunged list". Haeckel was included by name as a forbidden author. Gunther Hecht, a member of the Nazi Department of Race Politics, also issued a memorandum rejecting Haeckel as a forerunner of National Socialism. Kurt Hildebrandt, a Nazi political philosopher, also rejected Haeckel. Eventually Nazis rejected Haeckel because his evolutionary ideas could not be reconciled with Nazi ideology. In Haeckel's work as a zoologist, he noted that species constantly evolved into new species that seemed to retain few consistent features among themselves and therefore few features that distinguished them as a group ("a self-contained unity"). ""Wohl aber ist eine solche reale und vollkommen abgeschlossene Einheit die Summe aller Species, welche aus einer und derselben gemeinschaftlichen Stammform allmählig sich entwickelt haben, wie z. B. alle Wirbelthiere. Diese Summe nennen wir Stamm (Phylon)."" (However, perhaps such a real and completely self-contained unity is the aggregate of all species which have gradually evolved from one and the same common original form, as, for example, all vertebrates. We name this aggregate [a] "Stamm" [i.e., race] ("Phylon").) "phylogeny", and "Protista." Although Haeckel's ideas are important to the history of evolutionary theory, and although he was a competent invertebrate anatomist most famous for his work on radiolaria, many speculative concepts that he championed are now considered incorrect. For example, Haeckel described and named hypothetical ancestral microorganisms that have never been found. He developed the influential but no longer widely held recapitulation theory ("ontogeny recapitulates phylogeny") claiming that an individual organism's biological development, or ontogeny, parallels and summarises its species' evolutionary development, or phylogeny. He was one of the first to consider psychology as a branch of physiology. He also proposed the kingdom "Protista" in 1866. His chief interests lay in evolution and life development processes in general, including development of nonrandom form, which culminated in the beautifully illustrated "Kunstformen der Natur" ("Art forms of nature"). Haeckel did not support natural selection, rather believing in Lamarckism. Haeckel advanced a version of the earlier recapitulation theory previously set out by Étienne Serres in the 1820s and supported by followers of Étienne Geoffroy Saint-Hilaire including Robert Edmond Grant. It proposed a link between ontogeny (development of form) and phylogeny (evolutionary descent), summed up by Haeckel in the phrase "ontogeny recapitulates phylogeny". His concept of recapitulation has been refuted in the form he gave it (now called "strong recapitulation"), in favour of the ideas first advanced by Karl Ernst von Baer. The strong recapitulation hypothesis views ontogeny as repeating forms of adult ancestors, while weak recapitulation means that what is repeated (and built upon) is the ancestral embryonic development process. Haeckel supported the theory with embryo drawings that have since been shown to be oversimplified and in part inaccurate, and the theory is now considered an oversimplification of quite complicated relationships, however comparison of embryos remains a powerful way to demonstrate that all animals are related. Haeckel introduced the concept of heterochrony, the change in timing of embryonic development over the course of evolution. Haeckel was a flamboyant figure, who sometimes took great, non-scientific leaps from available evidence. For example, at the time when Darwin published "On the Origin of Species by Means of Natural Selection" (1859), Haeckel postulated that evidence of human evolution would be found in the Dutch East Indies (now Indonesia). At that time, no remains of human ancestors had yet been identified. He described these theoretical remains in great detail and even named the as-yet unfound species, "Pithecanthropus alalus", and instructed his students such as Richard and Oskar Hertwig to go and find it. One student did find some remains: a Dutchman named Eugène Dubois searched the East Indies from 1887 to 1895, discovering the remains of Java Man in 1891, consisting of a skullcap, thighbone, and a few teeth. These remains are among the oldest hominid remains ever found. Dubois classified Java Man with Haeckel's "Pithecanthropus" label, though they were later reclassified as "Homo erectus". Some scientists of the day suggested Dubois' Java Man as a potential intermediate form between modern humans and the common ancestor we share with the other great apes. The current consensus of anthropologists is that the direct ancestors of modern humans were African populations of "Homo erectus" (possibly "Homo ergaster"), rather than the Asian populations exemplified by Java Man and Peking Man. (Ironically, a new human species, Homo floresiensis, a dwarf human type, has recently been discovered in the island of Flores). The creationist polygenism of Samuel George Morton and Louis Agassiz, which presented human races as separately created species, was rejected by Charles Darwin, who argued for the monogenesis of the human species and the African origin of modern humans. In contrast to most of Darwin's supporters, Haeckel put forward a doctrine of evolutionary polygenism based on the ideas of the linguist August Schleicher, in which several different language groups had arisen separately from speechless prehuman "Urmenschen" (), which themselves had evolved from simian ancestors. These separate languages had completed the transition from animals to man, and under the influence of each main branch of languages, humans had evolved – in a kind of Lamarckian use-inheritance – as separate species, which could be subdivided into races. From this, Haeckel drew the implication that languages with the most potential yield the human races with the most potential, led by the Semitic and Indo-Germanic groups, with Berber, Jewish, Greco-Roman and Germanic varieties to the fore. As Haeckel stated: Haeckel's view can be seen as a forerunner of the views of Carleton Coon, who also believed that human races evolved independently and in parallel with each other. These ideas eventually fell from favour. Haeckel also applied the hypothesis of polygenism to the modern diversity of human groups. He became a key figure in social darwinism and leading proponent of scientific racism, stating for instance: Haeckel divided human beings into ten races, of which the Caucasian was the highest and the primitives were doomed to extinction. Haeckel claimed that '[t]he Negro' had stronger and more freely movable toes than any other race, which, he argued, was evidence of their being less evolved, and which led him to compare them to four-handed" Apes'. Haeckel also believed 'Negroes' were savages and that Whites were the most civilised. In his "Ontogeny and Phylogeny" Harvard paleontologist Stephen Jay Gould wrote: "[Haeckel's] evolutionary racism; his call to the German people for racial purity and unflinching devotion to a 'just' state; his belief that harsh, inexorable laws of evolution ruled human civilization and nature alike, conferring upon favored races the right to dominate others ... all contributed to the rise of Nazism." In his introduction to the Nazi party ideologue Alfred Rosenberg's 1930 book, "The Myth of the Twentieth Century", Peter Peel affirms that Rosenberg had indeed read Haeckel. In the same line of thought, historian Daniel Gasman states that Haeckel's ideology stimulated the birth of Fascist ideology in Italy and France. However, Robert J. Richards notes: "Haeckel, on his travels to Ceylon and Indonesia, often formed closer and more intimate relations with natives, even members of the untouchable classes, than with the European colonials." and says the Nazis rejected Haeckel, since he opposed antisemitism, while supporting ideas they disliked (for instance atheism, feminism, internationalism, pacifism etc.). Haeckel claimed the origin of humanity was to be found in Asia: he believed that Hindustan (Indian subcontinent) was the actual location where the first humans had evolved. Haeckel argued that humans were closely related to the primates of Southeast Asia and rejected Darwin's hypothesis of Africa. Haeckel later claimed that the missing link was to be found on the lost continent of Lemuria located in the Indian Ocean. He believed that Lemuria was the home of the first humans and that Asia was the home of many of the earliest primates; he thus supported that Asia was the cradle of hominid evolution. Haeckel also claimed that Lemuria connected Asia and Africa, which allowed the migration of humans to the rest of the world. In Haeckel’s book "The History of Creation" (1884) he included migration routes which he thought the first humans had used outside of Lemuria. When Haeckel was a student in the 1850s he showed great interest in embryology, attending the rather unpopular lectures twice and in his notes sketched the visual aids: textbooks had few illustrations, and large format plates were used to show students how to see the tiny forms under a reflecting microscope, with the translucent tissues seen against a black background. Developmental series were used to show stages within a species, but inconsistent views and stages made it even more difficult to compare different species. It was agreed by all European evolutionists that all vertebrates looked very similar at an early stage, in what was thought of as a common ideal type, but there was a continuing debate from the 1820s between the Romantic recapitulation theory that human embryos developed through stages of the forms of all the major groups of adult animals, literally manifesting a sequence of organisms on a linear chain of being, and Karl Ernst von Baer's opposing view, stated in von Baer's laws of embryology, that the early general forms diverged into four major groups of specialised forms without ever resembling the adult of another species, showing affinity to an archetype but no relation to other types or any transmutation of species. By the time Haeckel was teaching he was able to use a textbook with woodcut illustrations written by his own teacher Albert von Kölliker, which purported to explain human development while also using other mammalian embryos to claim a coherent sequence. Despite the significance to ideas of transformism, this was not really polite enough for the new popular science writing, and was a matter for medical institutions and for experts who could make their own comparisons. Darwin's "On the Origin of Species", which made a powerful impression on Haeckel when he read it in 1864, was very cautious about the possibility of ever reconstructing the history of life, but did include a section reinterpreting von Baer's embryology and revolutionising the field of study, concluding that "Embryology rises greatly in interest, when we thus look at the embryo as a picture, more or less obscured, of the common parent-form of each great class of animals." It mentioned von Baer's 1828 anecdote (misattributing it to Louis Agassiz) that at an early stage embryos were so similar that it could be impossible to tell whether an unlabelled specimen was of a mammal, a bird, or of a reptile, and Darwin's own research using embryonic stages of barnacles to show that they are crustaceans, while cautioning against the idea that one organism or embryonic stage is "higher" or "lower", or more or less evolved. Haeckel disregarded such caution, and in a year wrote his massive and ambitious "Generelle Morphologie", published in 1866, presenting a revolutionary new synthesis of Darwin's ideas with the German tradition of "Naturphilosophie" going back to Goethe and with the progressive evolutionism of Lamarck in what he called "Darwinismus". He used morphology to reconstruct the evolutionary history of life, in the absence of fossil evidence using embryology as evidence of ancestral relationships. He invented new terms, including ontogeny and phylogeny, to present his evolutionised recapitulation theory that "ontogeny recapitulated phylogeny". The two massive volumes sold poorly, and were heavy going: with his limited understanding of German, Darwin found them impossible to read. Haeckel's publisher turned down a proposal for a "strictly scholarly and objective" second edition. Haeckel's aim was a reformed morphology with evolution as the organising principle of a cosmic synthesis unifying science, religion, and art. He was giving successful "popular lectures" on his ideas to students and townspeople in Jena, in an approach pioneered by his teacher Rudolf Virchow. To meet his publisher's need for a popular work he used a student's transcript of his lectures as the basis of his "Natürliche Schöpfungsgeschichte" of 1868, presenting a comprehensive presentation of evolution. In the Spring of that year he drew figures for the book, synthesising his views of specimens in Jena and published pictures to represent types. After publication he told a colleague that the images "are completely exact, partly copied from nature, partly assembled from all illustrations of these early stages that have hitherto become known". There were various styles of embryological drawings at that time, ranging from more schematic representations to "naturalistic" illustrations of specific specimens. Haeckel believed privately that his figures were both exact and synthetic, and in public asserted that they were schematic like most figures used in teaching. The images were reworked to match in size and orientation, and though displaying Haeckel's own views of essential features, they support von Baer's concept that vertebrate embryos begin similarly and then diverge. Relating different images on a grid conveyed a powerful evolutionary message. As a book for the general public, it followed the common practice of not citing sources. The book sold very well, and while some anatomical experts hostile to Haeckel's evolutionary views expressed some private concerns that certain figures had been drawn rather freely, the figures showed what they already knew about similarities in embryos. The first published concerns came from Ludwig Rütimeyer, a professor of zoology and comparative anatomy at the University of Basel who had placed fossil mammals in an evolutionary lineage early in the 1860s and had been sent a complimentary copy. At the end of 1868 his review in the "Archiv für Anthropologie" wondered about the claim that the work was "popular and scholarly", doubting whether the second was true, and expressed horror about such public discussion of man's place in nature with illustrations such as the evolutionary trees being shown to non-experts. Though he made no suggestion that embryo illustrations should be directly based on specimens, to him the subject demanded the utmost "scrupulosity and conscientiousness" and an artist must "not arbitrarily model or generalise his originals for speculative purposes" which he considered proved by comparison with works by other authors. In particular, "one and the same, moreover incorrectly interpreted woodcut, is presented to the reader three times in a row and with three different captions as [the] embryo of the dog, the chick, [and] the turtle". He accused Haeckel of "playing fast and loose with the public and with science", and failing to live up to the obligation to the truth of every serious researcher. Haeckel responded with angry accusations of bowing to religious prejudice, but in the second (1870) edition changed the duplicated embryo images to a single image captioned "embryo of a mammal or bird". Duplication using galvanoplastic stereotypes (clichés) was a common technique in textbooks, but not on the same page to represent different eggs or embryos. In 1891 Haeckel made the excuse that this "extremely rash foolishness" had occurred in undue haste but was "bona fide", and since repetition of incidental details was obvious on close inspection, it is unlikely to have been intentional deception. The revised 1870 second edition of 1,500 copies attracted more attention, being quickly followed by further revised editions with larger print runs as the book became a prominent part of the optimistic, nationalist, anticlerical "culture of progress" in Otto von Bismarck's new German Empire. The similarity of early vertebrate embryos became common knowledge, and the illustrations were praised by experts such as Michael Foster of the University of Cambridge. In the introduction to his 1871 "The Descent of Man, and Selection in Relation to Sex", Darwin gave particular praise to Haeckel, writing that if "Natürliche Schöpfungsgeschichte" "had appeared before my essay had been written, I should probably never have completed it". The first chapter included an illustration: "As some of my readers may never have seen a drawing of an embryo, I have given one of man and another of a dog, at about the same early stage of development, carefully copied from two works of undoubted accuracy" with a footnote citing the sources and noting that "Häckel has also given analogous drawings in his "Schöpfungsgeschichte."" The fifth edition of Haeckel's book appeared in 1874, with its frontispiece a heroic portrait of Haeckel himself, replacing the previous controversial image of the heads of apes and humans. Later in 1874, Haeckel's simplified embryology textbook "Anthropogenie" made the subject into a battleground over Darwinism aligned with Bismarck's "Kulturkampf" ("culture struggle") against the Catholic Church. Haeckel took particular care over the illustrations, changing to the leading zoological publisher Wilhelm Engelmann of Leipzig and obtaining from them use of illustrations from their other textbooks as well as preparing his own drawings including a dramatic double page illustration showing "early", "somewhat later" and "still later" stages of 8 different vertebrates. Though Haeckel's views had attracted continuing controversy, there had been little dispute about the embryos and he had many expert supporters, but Wilhelm His revived the earlier criticisms and introduced new attacks on the 1874 illustrations. Others joined in: both expert anatomists and Catholic priests and supporters were politically opposed to Haeckel's views. While it has been widely claimed that Haeckel was charged with fraud by five professors and convicted by a university court at Jena, there does not appear to be an independently verifiable source for this claim. Recent analyses (Richardson 1998, Richardson and Keuck 2002) have found that some of the criticisms of Haeckel's embryo drawings were legitimate, but others were unfounded. There were multiple versions of the embryo drawings, and Haeckel rejected the claims of fraud. It was later said that "there is evidence of sleight of hand" on both sides of the feud between Haeckel and Wilhelm His. Robert J. Richards, in a paper published in 2008, defends the case for Haeckel, shedding doubt against the fraud accusations based on the material used for comparison with what Haeckel could access at the time. Haeckel was awarded the title of Excellency by Kaiser Wilhelm II in 1907 and the Linnean Society of London's prestigious Darwin-Wallace Medal in 1908. In the United States, "Mount Haeckel", a summit in the Eastern Sierra Nevada, overlooking the Evolution Basin, is named in his honour, as is another "Mount Haeckel", a summit in New Zealand; and the asteroid 12323 Haeckel. In Jena he is remembered with a monument at Herrenberg (erected in 1969), an exhibition at Ernst-Haeckel-Haus, and at the Jena Phyletic Museum, which continues to teach about evolution and share his work to this day. Darwin's 1859 book "On the Origin of Species" had immense popular influence, but although its sales exceeded its publisher's hopes it was a technical book rather than a work of popular science: long, difficult and with few illustrations. One of Haeckel's books did a great deal to explain his version of "Darwinism" to the world. It was a bestselling, provocatively illustrated book in German, titled "Natürliche Schöpfungsgeschichte", published in Berlin in 1868, and translated into English as "The History of Creation" in 1876. It was frequently reprinted until 1926. Haeckel argued that human evolution consisted of precisely 22 phases, the 21st – the "missing link" – being a halfway step between apes and humans. He even formally named this missing link "Pithecanthropus alalus", translated as "ape man without speech". Haeckel's literary output was extensive, including many books, scientific papers, and illustrations. The published artwork of Haeckel includes over 100 detailed, multi-colour illustrations of animals and sea creatures, collected in his "Kunstformen der Natur" ("Art Forms of Nature"). As a philosopher, Ernst Haeckel wrote "Die Welträthsel" (1895–1899; in English: "The Riddle of the Universe", 1901), the genesis for the term "world riddle" ("Welträtsel"); and "Freedom in Science and Teaching" to support teaching evolution. For a fuller list of works of and about Haeckel, see his entry in the .
https://en.wikipedia.org/wiki?curid=9668
Evolutionism Evolutionism is a term used (often derogatorily) to denote the theory of evolution. Its exact meaning has changed over time as the study of evolution has progressed. In the 19th century, it was used to describe the belief that organisms deliberately improved themselves through progressive inherited change (orthogenesis). The teleological belief went on to include cultural evolution and social evolution. In the 1970s the term Neo-Evolutionism was used to describe the idea "that human beings sought to preserve a familiar style of life unless change was forced on them by factors that were beyond their control". The term is most often used by creationists to describe adherence to the scientific consensus on evolution as equivalent to a secular religion. The term is very seldom used within the scientific community, since the scientific position on evolution is accepted by the overwhelming majority of scientists. Because evolutionary biology is the default scientific position, it is assumed that "scientists" or "biologists" are "evolutionists" unless specifically noted otherwise. In the creation–evolution controversy, creationists often call those who accept the validity of the modern evolutionary synthesis "evolutionists" and the theory itself "evolutionism". Before its use to describe biological evolution, the term "evolution" was originally used to refer to any orderly sequence of events with the outcome somehow contained at the start. The first five editions of Darwin's in "Origin of Species" used the word "evolved", but the word "evolution" was only used in its sixth edition in 1872. By then, Herbert Spencer had developed the concept theory that organisms strive to evolve due to an internal "driving force" (orthogenesis) in 1862. Edward B. Tylor and Lewis H Morgan brought the term "evolution" to anthropology though they tended toward the older pre-Spencerian definition helping to form the concept of unilineal (social) evolution used during the later part of what Trigger calls the Antiquarianism-Imperial Synthesis period (c1770-c1900). The term evolutionism subsequently came to be used for the now discredited theory that evolution contained a deliberate component, rather than the selection of beneficial traits from random variation by differential survival. The term "evolution" is widely used, but the term "evolutionism" is not used in the scientific community to refer to evolutionary biology as it is redundant and anachronistic. However, the term has been used by creationists in discussing the creation–evolution controversy. For example, the Institute for Creation Research, in order to imply placement of evolution in the category of 'religions', including atheism, fascism, humanism and occultism, commonly uses the words "evolutionism" and "evolutionist" to describe the consensus of mainstream science and the scientists subscribing to it, thus implying through language that the issue is a matter of religious belief. The BioLogos Foundation, an organization that promotes the idea of theistic evolution, uses the term "evolutionism" to describe "the atheistic worldview that so often accompanies the acceptance of biological evolution in public discourse." It views this as a subset of scientism.
https://en.wikipedia.org/wiki?curid=9670
Entscheidungsproblem In mathematics and computer science, the (, German for "decision problem") is a challenge posed by David Hilbert and Wilhelm Ackermann in 1928.
https://en.wikipedia.org/wiki?curid=9672
Einhard Einhard (also Eginhard or Einhart; ; 775 – March 14, 840) was a Frankish scholar and courtier. Einhard was a dedicated servant of Charlemagne and his son Louis the Pious; his main work is a biography of Charlemagne, the "Vita Karoli Magni", "one of the most precious literary bequests of the early Middle Ages." Einhard was from the eastern German-speaking part of the Frankish Kingdom. Born into a family of landowners of some importance, his parents sent him to be educated by the monks of Fulda, one of the most impressive centers of learning in the Frank lands. Perhaps due to his small stature, which restricted his riding and sword-fighting ability, Einhard concentrated his energies on scholarship, especially the mastering of Latin. He was accepted into the hugely wealthy court of Charlemagne around 791 or 792. Charlemagne actively sought to amass scholarly men around him and established a royal school led by the Northumbrian scholar Alcuin. Einhard evidently was a talented builder and construction manager, because Charlemagne put him in charge of the completion of several palace complexes including Aachen and Ingelheim. Despite the fact that Einhard was on intimate terms with Charlemagne, he never achieved office in his reign. In 814, on Charlemagne's death, his son Louis the Pious made Einhard his private secretary. Einhard retired from court during the time of the disputes between Louis and his sons in the spring of 830. He died at Seligenstadt in 840. Einhard was married to Emma, of whom little is known. There is a possibility that their marriage bore a son, Vussin. Their marriage also appears to have been exceptionally liberal for the period, with Emma being as active as Einhard, if not more so, in the handling of their property. It is said that in the later years of their marriage Emma and Einhard abstained from sexual relations, choosing instead to focus their attentions on their many religious commitments. Though he was undoubtedly devoted to her, Einhard wrote nothing of his wife until after her death on 13 December 835, when he wrote to a friend that he was reminded of her loss in ‘every day, in every action, in every undertaking, in all the administration of the house and household, in everything needing to be decided upon and sorted out in my religious and earthly responsibilities’. Einhard made numerous references to himself as a "sinner" according to his strong Christian faith. He erected churches at both of his estates in Michelstadt and Mulinheim. In Michelstadt, he also saw fit to build a basilica completed in 827 and then sent a servant, Ratleic, to Rome with an end to find relics for the new building. Once in Rome, Ratleic robbed a catacomb of the bones of the Martyrs Marcellinus and Peter and had them translated to Michelstadt. Once there, the relics made it known they were unhappy with their new tomb and thus had to be moved again to Mulinheim. Once established there, they proved to be miracle workers. Although unsure as to why these saints should choose such a "sinner" as their patron, Einhard nonetheless set about ensuring they continued to receive a resting place fitting of their honour. Between 831 and 834 he founded a Benedictine Monastery and, after the death of his wife, served as its Abbot until his own death in 840. Local lore from Seligenstadt portrays Einhard as the lover of Emma, one of Charlemagne's daughters, and has the couple elope from court. Charlemagne found them at Seligenstadt (then called Obermühlheim) and forgave them. This account is used to explain the name "Seligenstadt" by folk etymology. Einhard and his wife were originally buried in one sarcophagus in the choir of the church in Seligenstadt, but in 1810 the sarcophagus was presented by the Grand Duke of Hesse to the count of Erbach, who claims descent from Einhard as the husband of Imma, the reputed daughter of Charlemagne. The count put it in the famous chapel of his castle at Erbach in the Odenwald. The most famous of Einhard's works is his biography of Charlemagne, the "Vita Karoli Magni", "The Life of Charlemagne" (c. 817–836), which provides much direct information about Charlemagne's life and character, written sometime between 817 and 830. In composing this he relied heavily upon the Royal Frankish Annals. Einhard's literary model was the classical work of the Roman historian Suetonius, the "Lives of the Caesars", though it is important to stress that the work is very much Einhard's own, that is to say he adapts the models and sources for his own purposes. His work was written as a praise of Charlemagne, whom he regarded as a foster-father ("nutritor") and to whom he was a debtor "in life and death". The work thus contains an understandable degree of bias, Einhard taking care to exculpate Charlemagne in some matters, not mention others, and to gloss over certain issues which would be of embarrassment to Charlemagne, such as the morality of his daughters; by contrast, other issues are curiously not glossed over, like his concubines. Einhard is also responsible for three other extant works: a collection of letters, "On the Translations and the Miracles of SS. Marcellinus and Petrus", and "On the Adoration of the Cross". The latter dates from ca. 830 and was not rediscovered until 1885, when Ernst Dümmler identified a text in a manuscript in Vienna as the missing "Libellus de adoranda cruce", which Einhard had dedicated to his pupil Lupus Servatus.
https://en.wikipedia.org/wiki?curid=9674
Ester In chemistry, an ester is a chemical compound derived from an acid (organic or inorganic) in which at least one –OH (hydroxyl) group is replaced by an –O–alkyl (alkoxy) group. Usually, esters are derived from a carboxylic acid and an alcohol. Glycerides, which are fatty acid esters of glycerol, are important esters in biology, being one of the main classes of lipids, and making up the bulk of animal fats and vegetable oils. Esters with low molecular weight are commonly used as fragrances and found in essential oils and pheromones. Phosphoesters form the backbone of DNA molecules. Nitrate esters, such as nitroglycerin, are known for their explosive properties, while polyesters are important plastics, with monomers linked by ester moieties. Esters usually have a sweet smell and are considered high-quality solvents for a broad array of plastics, plasticizers, resins, and lacquers. They are also one of the largest classes of synthetic lubricants on the commercial market. The word "ester" was coined in 1848 by a German chemist Leopold Gmelin, probably as a contraction of the German "Essigäther", "acetic ether". Ester names are derived from the parent alcohol and the parent acid, where the latter may be organic or inorganic. Esters derived from the simplest carboxylic acids are commonly named according to the more traditional, so-called "trivial names" e.g. as formate, acetate, propionate, and butyrate, as opposed to the IUPAC nomenclature methanoate, ethanoate, propanoate and butanoate. Esters derived from more complex carboxylic acids are, on the other hand, more frequently named using the systematic IUPAC name, based on the name for the acid followed by the suffix "-oate". For example, the ester hexyl octanoate, also known under the trivial name hexyl caprylate, has the formula CH3(CH2)6CO2(CH2)5CH3. The chemical formulas of organic esters usually take the form RCO2R′, where R and R′ are the hydrocarbon parts of the carboxylic acid and the alcohol, respectively. For example, butyl acetate (systematically butyl ethanoate), derived from butanol and acetic acid (systematically ethanoic acid) would be written CH3CO2C4H9. Alternative presentations are common including BuOAc and CH3COOC4H9. Cyclic esters are called lactones, regardless of whether they are derived from an organic or an inorganic acid. One example of an organic lactone is γ-valerolactone. An uncommon class of organic esters are the orthoesters, which have the formula RC(OR′)3. Triethylorthoformate (HC(OC2H5)3) is derived, in terms of its name (but not its synthesis) from orthoformic acid (HC(OH)3) and ethanol. Esters can also be derived from inorganic acids. Inorganic acids that exist as tautomers form diverse esters Inorganic acids that are unstable or elusive form stable esters. In principle, all metal and metalloid alkoxides, of which many hundreds are known, could be classified as esters of the hypothetical acids. Esters contain a carbonyl center, which gives rise to 120° C–C–O and O–C–O angles. Unlike amides, esters are structurally flexible functional groups because rotation about the C–O–C bonds has a low barrier. Their flexibility and low polarity is manifested in their physical properties; they tend to be less rigid (lower melting point) and more volatile (lower boiling point) than the corresponding amides. The p"K"a of the alpha-hydrogens on esters is around 25. Many esters have the potential for conformational isomerism, but they tend to adopt an "s"-cis (or Z) conformation rather than the "s"-trans (or E) alternative, due to a combination of hyperconjugation and dipole minimization effects. The preference for the Z conformation is influenced by the nature of the substituents and solvent, if present. Lactones with small rings are restricted to the "s"-trans (i.e. E) conformation due to their cyclic structure. Esters are more polar than ethers but less polar than alcohols. They participate in hydrogen bonds as hydrogen-bond acceptors, but cannot act as hydrogen-bond donors, unlike their parent alcohols. This ability to participate in hydrogen bonding confers some water-solubility. Because of their lack of hydrogen-bond-donating ability, esters do not self-associate. Consequently, esters are more volatile than carboxylic acids of similar molecular weight. Esters are generally identified by gas chromatography, taking advantage of their volatility. IR spectra for esters feature an intense sharp band in the range 1730–1750 cm−1 assigned to "ν"C=O. This peak changes depending on the functional groups attached to the carbonyl. For example, a benzene ring or double bond in conjugation with the carbonyl will bring the wavenumber down about 30 cm−1. Esters are widespread in nature and are widely used in industry. In nature, fats are in general triesters derived from glycerol and fatty acids. Esters are responsible for the aroma of many fruits, including apples, durians, pears, bananas, pineapples, and strawberries. Several billion kilograms of polyesters are produced industrially annually, important products being polyethylene terephthalate, acrylate esters, and cellulose acetate. Esterification is the general name for a chemical reaction in which two reactants (typically an alcohol and an acid) form an ester as the reaction product. Esters are common in organic chemistry and biological materials, and often have a pleasant characteristic, fruity odor. This leads to their extensive use in the fragrance and flavor industry. Ester bonds are also found in many polymers. The classic synthesis is the Fischer esterification, which involves treating a carboxylic acid with an alcohol in the presence of a dehydrating agent: The equilibrium constant for such reactions is about 5 for typical esters, e.g., ethyl acetate. The reaction is slow in the absence of a catalyst. Sulfuric acid is a typical catalyst for this reaction. Many other acids are also used such as polymeric sulfonic acids. Since esterification is highly reversible, the yield of the ester can be improved using Le Chatelier's principle: Reagents are known that drive the dehydration of mixtures of alcohols and carboxylic acids. One example is the Steglich esterification, which is a method of forming esters under mild conditions. The method is popular in peptide synthesis, where the substrates are sensitive to harsh conditions like high heat. DCC (dicyclohexylcarbodiimide) is used to activate the carboxylic acid to further reaction. 4-Dimethylaminopyridine (DMAP) is used as an acyl-transfer catalyst. Another method for the dehydration of mixtures of alcohols and carboxylic acids is the Mitsunobu reaction: Carboxylic acids can be esterified using diazomethane: Using this diazomethane, mixtures of carboxylic acids can be converted to their methyl esters in near quantitative yields, e.g., for analysis by gas chromatography. The method is useful in specialized organic synthetic operations but is considered too hazardous and expensive for large-scale applications. Carboxylic acids are esterified by treatment with epoxides, giving β-hydroxyesters: This reaction is employed in the production of vinyl ester resin resins from acrylic acid. Alcohols react with acyl chlorides and acid anhydrides to give esters: The reactions are irreversible simplifying work-up. Since acyl chlorides and acid anhydrides also react with water, anhydrous conditions are preferred. The analogous acylations of amines to give amides are less sensitive because amines are stronger nucleophiles and react more rapidly than does water. This method is employed only for laboratory-scale procedures, as it is expensive. Although not widely employed for esterifications, salts of carboxylate anions can be alkylating agent with alkyl halides to give esters. In the case that an alkyl chloride is used, an iodide salt can catalyze the reaction (Finkelstein reaction). The carboxylate salt is often generated "in situ". In difficult cases, the silver carboxylate may be used, since the silver ion coordinates to the halide aiding its departure and improving the reaction rate. This reaction can suffer from anion availability problems and, therefore, can benefit from the addition of phase transfer catalysts or highly polar aprotic solvents such as DMF. Transesterification, which involves changing one ester into another one, is widely practiced: Like the hydrolysation, transesterification is catalysed by acids and bases. The reaction is widely used for degrading triglycerides, e.g. in the production of fatty acid esters and alcohols. Poly(ethylene terephthalate) is produced by the transesterification of dimethyl terephthalate and ethylene glycol: A subset of transesterification is the alcoholysis of diketene. This reaction affords 2-ketoesters. Alkenes undergo "hydroesterification" in the presence of metal carbonyl catalysts. Esters of propionic acid are produced commercially by this method: A preparaton of methyl propionate is one illustrative example. The carbonylation of methanol yields methyl formate, which is the main commercial source of formic acid. The reaction is catalyzed by sodium methoxide: In the presence of palladium-based catalysts, ethylene, acetic acid, and oxygen react to give vinyl acetate: Direct routes to this same ester are not possible because vinyl alcohol is unstable. Carboxylic acids also add across alkynes to give the same products. Silicotungstic acid is used to manufacture ethyl acetate by the alkylation of acetic acid by ethylene: The Tishchenko reaction involve disproportionation of an aldehyde in the presence of an anhydrous base to give an ester. Catalysts are aluminium alkoxides or sodium alkoxides. Benzaldehyde reacts with sodium benzyloxide (generated from sodium and benzyl alcohol) to generate benzyl benzoate. The method is used in the production of ethyl acetate from acetaldehyde. Esters react with nucleophiles at the carbonyl carbon. The carbonyl is weakly electrophilic but is attacked by strong nucleophiles (amines, alkoxides, hydride sources, organolithium compounds, etc.). The C–H bonds adjacent to the carbonyl are weakly acidic but undergo deprotonation with strong bases. This process is the one that usually initiates condensation reactions. The carbonyl oxygen in esters is weakly basic, less so than the carbonyl oxygen in amides due to resonance donation of an electron pair from nitrogen in amides, but forms adducts. Esterification is a reversible reaction. Esters undergo hydrolysis under acid and basic conditions. Under acidic conditions, the reaction is the reverse reaction of the Fischer esterification. Under basic conditions, hydroxide acts as a nucleophile, while an alkoxide is the leaving group. This reaction, saponification, is the basis of soap making. The alkoxide group may also be displaced by stronger nucleophiles such as ammonia or primary or secondary amines to give amides: (ammonolysis reaction) This reaction is not usually reversible. Hydrazines and hydroxylamine can be used in place of amines. Esters can be converted to isocyanates through intermediate hydroxamic acids in the Lossen rearrangement. Sources of carbon nucleophiles, e.g., Grignard reagents and organolithium compounds, add readily to the carbonyl. Compared to ketones and aldehydes, esters are relatively resistant to reduction. The introduction of catalytic hydrogenation in the early part of the 20th century was a breakthrough; esters of fatty acids are hydrogenated to fatty alcohols. A typical catalyst is copper chromite. Prior to the development of catalytic hydrogenation, esters were reduced on a large scale using the Bouveault–Blanc reduction. This method, which is largely obsolete, uses sodium in the presence of proton sources. Especially for fine chemical syntheses, lithium aluminium hydride is used to reduce esters to two primary alcohols. The related reagent sodium borohydride is slow in this reaction. DIBAH reduces esters to aldehydes. Direct reduction to give the corresponding ether is difficult as the intermediate hemiacetal tends to decompose to give an alcohol and an aldehyde (which is rapidly reduced to give a second alcohol). The reaction can be achieved using triethylsilane with a variety of Lewis acids. As for aldehydes, the hydrogen atoms on the carbon adjacent ("α to") the carboxyl group in esters are sufficiently acidic to undergo deprotonation, which in turn leads to a variety of useful reactions. Deprotonation requires relatively strong bases, such as alkoxides. Deprotonation gives a nucleophilic enolate, which can further react, e.g., the Claisen condensation and its intramolecular equivalent, the Dieckmann condensation. This conversion is exploited in the malonic ester synthesis, wherein the diester of malonic acid reacts with an electrophile (e.g., alkyl halide), and is subsequently decarboxylated. Another variation is the Fráter–Seebach alkylation. As a class, esters serve as protecting groups for carboxylic acids. Protecting a carboxylic acid is useful in peptide synthesis, to prevent self-reactions of the bifunctional amino acids. Methyl and ethyl esters are commonly available for many amino acids; the "t"-butyl ester tends to be more expensive. However, "t"-butyl esters are particularly useful because, under strongly acidic conditions, the "t"-butyl esters undergo elimination to give the carboxylic acid and isobutylene, simplifying work-up. Many esters have distinctive fruit-like odors, and many occur naturally in the essential oils of plants. This has also led to their common use in artificial flavorings and fragrances which aim to mimic those odors.
https://en.wikipedia.org/wiki?curid=9675
Endosymbiont An endosymbiont or endobiont is any organism that lives within the body or cells of another organism most often, though not always, in a mutualistic relationship. (The term endosymbiosis is from the Greek: ἔνδον "endon" "within", σύν "syn" "together" and βίωσις "biosis" "living".) Examples are nitrogen-fixing bacteria (called rhizobia), which live in the root nodules of legumes; single-cell algae inside reef-building corals, and bacterial endosymbionts that provide essential nutrients to about 10–15% of insects. There are two types of symbiont transmissions. In horizontal transmission, each new generation acquires free living symbionts from the environment. An example is the nitrogen-fixing bacteria in certain plant roots. Vertical transmission takes place when the symbiont is transferred directly from parent to offspring. There is also a combination of these types, where symbionts are transferred vertically for some generation before a switch of host occurs and new symbionts are horizontally acquired from the environment. In vertical transmissions, the symbionts often have a reduced genome and are no longer able to survive on their own. As a result, the symbiont depends on the host, resulting in a highly intimate co-dependent relationship. For instance, pea aphid symbionts have lost genes for essential molecules, now relying on the host to supply them with nutrients. In return, the symbionts synthesize essential amino acids for the aphid host . Other examples include "Wigglesworthia" nutritional symbionts of tse-tse flies, or in sponges. When a symbiont reaches this stage, it begins to resemble a cellular organelle, similar to mitochondria or chloroplasts. Many instances of endosymbiosis are obligate; that is, either the endosymbiont or the host cannot survive without the other, such as the gutless marine worms of the genus "Riftia", which get nutrition from their endosymbiotic bacteria. The most common examples of obligate endosymbioses are mitochondria and chloroplasts. Some human parasites, e.g. "Wuchereria bancrofti" and "Mansonella perstans", thrive in their intermediate insect hosts because of an obligate endosymbiosis with "Wolbachia spp". They can both be eliminated from said hosts by treatments that target this bacterium. However, not all endosymbioses are obligate and some endosymbioses can be harmful to either of the organisms involved. Two major types of organelle in eukaryotic cells, mitochondria and plastids such as chloroplasts, are considered to be bacterial endosymbionts. This process is commonly referred to as symbiogenesis. Symbiogenesis explains the origins of eukaryotes, whose cells contain two major kinds of organelle: mitochondria and chloroplasts. The theory proposes that these organelles evolved from certain types of bacteria that eukaryotic cells engulfed through phagocytosis. These cells and the bacteria trapped inside them entered an endosymbiotic relationship, meaning that the bacteria took up residence and began living exclusively within the eukaryotic cells. Numerous insect species have endosymbionts at different stages of symbiogenesis. A common theme of symbiogenesis involves the reduction of the genome to only essential genes for the host and symbiont collective genome. A remarkable example of this is the fractionation of the "Hodgkinia" genome of "Magicicada" cicadas. Because the cicada life cycle takes years underground, natural selection on endosymbiont populations is relaxed for many bacterial generations. This allows the symbiont genomes to diversify within the host for years with only punctuated periods of selection when the cicadas reproduce. As a result, the ancestral "Hodgkinia" genome has split into three groups of primary endosymbiont, each encoding only a fraction of the essential genes for the symbiosis. The host now requires all three sub-groups of symbiont, each with degraded genomes lacking most essential genes for bacterial viability. The best-studied examples of endosymbiosis are known from invertebrates. These symbioses affect organisms with global impact, including "symbiodinium" of corals, or "Wolbachia" of insects. Many insect agricultural pests and human disease vectors have intimate relationships with primary endosymbionts. Scientists classify insect endosymbionts in two broad categories, 'Primary' and 'Secondary'. Primary endosymbionts (sometimes referred to as P-endosymbionts) have been associated with their insect hosts for many millions of years (from 10 to several hundred million years in some cases). They form obligate associations (see below), and display cospeciation with their insect hosts. Secondary endosymbionts exhibit a more recently developed association, are sometimes horizontally transferred between hosts, live in the hemolymph of the insects (not specialized bacteriocytes, see below), and are not obligate. Among primary endosymbionts of insects, the best-studied are the pea aphid ("Acyrthosiphon pisum") and its endosymbiont "Buchnera sp." APS, the tsetse fly "Glossina morsitans morsitans" and its endosymbiont "Wigglesworthia glossinidia brevipalpis" and the endosymbiotic protists in lower termites. As with endosymbiosis in other insects, the symbiosis is obligate in that neither the bacteria nor the insect is viable without the other. Scientists have been unable to cultivate the bacteria in lab conditions outside of the insect. With special nutritionally-enhanced diets, the insects can survive, but are unhealthy, and at best survive only a few generations. In some insect groups, these endosymbionts live in specialized insect cells called bacteriocytes (also called "mycetocytes"), and are maternally-transmitted, i.e. the mother transmits her endosymbionts to her offspring. In some cases, the bacteria are transmitted in the egg, as in "Buchnera"; in others like "Wigglesworthia", they are transmitted via milk to the developing insect embryo. In termites, the endosymbionts reside within the hindguts and are transmitted through trophallaxis among colony members. The primary endosymbionts are thought to help the host either by providing nutrients that the host cannot obtain itself or by metabolizing insect waste products into safer forms. For example, the putative primary role of "Buchnera" is to synthesize essential amino acids that the aphid cannot acquire from its natural diet of plant sap. Likewise, the primary role of "Wigglesworthia", it is presumed, is to synthesize vitamins that the tsetse fly does not get from the blood that it eats. In lower termites, the endosymbiotic protists play a major role in the digestion of lignocellulosic materials that constitute a bulk of the termites' diet. Bacteria benefit from the reduced exposure to predators and competition from other bacterial species, the ample supply of nutrients and relative environmental stability inside the host. Genome sequencing reveals that obligate bacterial endosymbionts of insects have among the smallest of known bacterial genomes and have lost many genes that are commonly found in closely related bacteria. Several theories have been put forth to explain the loss of genes. It is presumed that some of these genes are not needed in the environment of the host insect cell. A complementary theory suggests that the relatively small numbers of bacteria inside each insect decrease the efficiency of natural selection in 'purging' deleterious mutations and small mutations from the population, resulting in a loss of genes over many millions of years. Research in which a parallel phylogeny of bacteria and insects was inferred supports the belief that the primary endosymbionts are transferred only vertically (i.e., from the mother), and not horizontally (i.e., by escaping the host and entering a new host). Attacking obligate bacterial endosymbionts may present a way to control their insect hosts, many of which are pests or carriers of human disease. For example, aphids are crop pests and the tsetse fly carries the organism "Trypanosoma brucei" that causes African sleeping sickness. Other motivations for their study involve understanding the origins of symbioses in general, as a proxy for understanding e.g. how chloroplasts or mitochondria came to be obligate symbionts of eukaryotes or plants. The pea aphid ("Acyrthosiphon pisum") is known to contain at least three secondary endosymbionts, "Hamiltonella defensa", "Regiella insecticola", and "Serratia symbiotica". "Hamiltonella defensa" defends its aphid host from parasitoid wasps. This defensive symbiosis improves the survival of aphids, which have lost some elements of the insect immune response. One of the best-understood defensive symbionts is the spiral bacteria "Spiroplasma poulsonii". "Spiroplasma sp." can be reproductive manipulators, but also defensive symbionts of "Drosophila" flies. In "Drosophila neotestacea", "S. poulsonii" has spread across North America owing to its ability to defend its fly host against nematode parasites. This defence is mediated by toxins called "ribosome-inactivating proteins" that attack the molecular machinery of invading parasites. These "Spiroplasma" toxins represent one of the first examples of a defensive symbiosis with a mechanistic understanding for defensive symbiosis between an insect endosymbiont and its host. "Sodalis glossinidius" is a secondary endosymbiont of tsetse flies that lives inter- and intracellularly in various host tissues, including the midgut and hemolymph. Phylogenetic studies have not indicated a correlation between evolution of "Sodalis" and tsetse. Unlike tsetse's primary symbiont "Wigglesworthia", though, "Sodalis" has been cultured "in vitro". Many other insects have secondary endosymbionts not reviewed here. Extracellular endosymbionts are also represented in all four extant classes of Echinodermata (Crinoidea, Ophiuroidea, Echinoidea, and Holothuroidea). Little is known of the nature of the association (mode of infection, transmission, metabolic requirements, etc.) but phylogenetic analysis indicates that these symbionts belong to the alpha group of the class Proteobacteria, relating them to "Rhizobium" and "Thiobacillus". Other studies indicate that these subcuticular bacteria may be both abundant within their hosts and widely distributed among the Echinoderms in general. Some marine oligochaeta (e.g., "Olavius algarvensis" and "Inanidrillus spp.") have obligate extracellular endosymbionts that fill the entire body of their host. These marine worms are nutritionally dependent on their symbiotic chemoautotrophic bacteria lacking any digestive or excretory system (no gut, mouth, or nephridia). The sea slug "Elysia chlorotica" lives in endosymbiotic relationship with the algae "Vaucheria litorea", and the jellyfish "Mastigias" have a similar relationship with an algae. Dinoflagellate endosymbionts of the genus "Symbiodinium", commonly known as zooxanthellae, are found in corals, mollusks (esp. giant clams, the "Tridacna"), sponges, and foraminifera. These endosymbionts drive the formation of coral reefs by capturing sunlight and providing their hosts with energy for carbonate deposition. Previously thought to be a single species, molecular phylogenetic evidence over the past couple decades has shown there to be great diversity in "Symbiodinium". In some cases, there is specificity between host and "Symbiodinium" clade. More often, however, there is an ecological distribution of "Symbiodinium", the symbionts switching between hosts with apparent ease. When reefs become environmentally stressed, this distribution of symbionts is related to the observed pattern of coral bleaching and recovery. Thus, the distribution of "Symbiodinium" on coral reefs and its role in coral bleaching presents one of the most complex and interesting current problems in reef ecology. In marine environments, bacterial endosymbionts have more recently been discovered. These endosymbiotic relationships are especially prevalent in oligotrophic or nutrient-poor regions of the ocean like that of the North Atlantic.  In these oligotrophic waters, cell growth of larger phytoplankton like that of diatoms is limited by low nitrate concentrations.  Endosymbiotic bacteria fix nitrogen for their diatom hosts and in turn receive organic carbon from photosynthesis. These symbioses play an important role in global carbon cycling in oligotrophic regions. One known symbiosis between the diatom "Hemialus" spp. and the cyanobacterium "Richelia intracellularis" has been found in the North Atlantic, Mediterranean, and Pacific Ocean. The "Richelia" endosymbiont is found within the diatom frustule of "Hemiaulus" spp., and has a reduced genome likely losing genes related to pathways the host now provides.  Research by Foster et al. (2011) measured nitrogen fixation by the cyanobacterial host "Richelia intracellularis" well above intracellular requirements, and found the cyanobacterium was likely fixing excess nitrogen for Hemiaulus host cells.  Additionally, both host and symbiont cell growth were much greater than free-living "Richelia intracellularis" or symbiont-free "Hemiaulus" spp.  The "Hemaiulus"-"Richelia" symbiosis is not obligatory especially in areas with excess nitrogen (nitrogen replete). "Richelia intracellularis" is also found in "Rhizosolenia" spp., a diatom found in oligotrophic oceans. Compared to the "Hemaiulus" host, the endosymbiosis with "Rhizosolenia" is much more consistent, and "Richelia intracellularis" is generally found in "Rhizosolenia". There are some asymbiotic (occurs without an endosymbiont) Rhizosolenia, however there appears to be mechanisms limiting growth of these organisms in low nutrient conditions. Cell division for both the diatom host and cyanobacterial symbiont can be uncoupled and mechanisms for passing bacterial symbionts to daughter cells during cell division are still relatively unknown. Other endosymbiosis with nitrogen fixers in open oceans include Calothrix in Chaetocerous spp. and UNCY-A in prymnesiophyte microalga.  The Chaetocerous-Calothrix endosymbiosis is hypothesized to be more recent, as the Calothrix genome is generally intact. While other species like that of the UNCY-A symbiont and Richelia have reduced genomes.  This reduction in genome size occurs within nitrogen metabolism pathways indicating endosymbiont species are generating nitrogen for their hosts and losing the ability to use this nitrogen independently. This endosymbiont reduction in genome size, might be a step that occurred in the evolution of organelles (above). "Mixotricha paradoxa" is a protozoan that lacks mitochondria. However, spherical bacteria live inside the cell and serve the function of the mitochondria. "Mixotricha" also has three other species of symbionts that live on the surface of the cell. "Paramecium bursaria", a species of ciliate, has a mutualistic symbiotic relationship with green alga called Zoochlorella. The algae live inside the cell, in the cytoplasm. "Paulinella chromatophora" is a freshwater amoeboid which has recently (evolutionarily speaking) taken on a cyanobacterium as an endosymbiont. Many foraminifera are hosts to several types of algae, such as red algae, diatoms, dinoflagellates and chlorophyta. These endosymbionts can be transmitted vertically to the next generation via asexual reproduction of the host, but because the endosymbionts are larger than the foraminiferal gametes, they need to acquire new algae again after sexual reproduction. Several species of radiolaria have photosynthetic symbionts. In some species the host will sometimes digest algae to keep their population at a constant level. Hatena arenicola is a flagellate protist with a complicated feeding apparaturs that feed on other microbes. But when it engulf a green alga from the genus Nephroselmis, the feeding apparatus disappears and it becomes photosyntethic. During mitosis the algae is transferred to only one of the two cells, and the cell without the algae needs to start the cycle all over again. In 1966, biologist Kwang W. Jeon found that a lab strain of Amoeba proteus had been infected by bacteria that lived inside the cytoplasmic vacuoles. This infection killed all the protists except from a few individuals. After the equivalent of 40 host generations, the two organisms gradually became mutually interdependent. Over many years of study, it has been confirmed that a genetic exchenge between the prokaryotes and protists has occurred. The spotted salamander ("Ambystoma maculatum") lives in a relationship with the algae "Oophila amblystomatis", which grows in the egg cases. Chloroplasts are primary endosymbionts of plants that provide energy to the plant by generating sugars. Of all the plants, "Azolla" has the most intimate relationship with a symbiont, as its cyanobacterium symbiont "Anabaena" is passed on directly from one generation to the next. The human genome project found several thousand endogenous retroviruses, endogenous viral elements in the genome that closely resemble and can be derived from retroviruses, organized into 24 families.
https://en.wikipedia.org/wiki?curid=9677
Exponential function In mathematics, an exponential function is a function of the form where is a positive real number, and in which the argument occurs as an exponent. For real numbers and a function of the form formula_1 is also an exponential function, as it can be rewritten as As functions of a real variable, exponential functions are uniquely characterized by the fact that the growth rate of such a function (that is, its derivative) is directly proportional to the value of the function. The constant of proportionality of this relationship is the natural logarithm of the base : For , the function formula_3 is increasing (as depicted for and ), because formula_4 makes the derivative always positive; while for , the function is decreasing (as depicted for ); and for the function is constant. The constant is the unique base for which the constant of proportionality is 1, so that the function is its own derivative: This is called the "natural exponential function", or simply "the exponential function". Since any exponential function can be written in terms of the natural exponential as formula_5, it is computationally and conceptually convenient to reduce the study of exponential functions to this particular one. The natural exponential is denoted by The former notation is commonly used for simpler exponents, while the latter is preferred when the exponent is a complicated expression. The graph of formula_6 is upward-sloping, and increases faster as increases. The graph always lies above the -axis but becomes arbitrarily close to it for large negative ; thus, the -axis is a horizontal asymptote. The equation formula_7 means that the slope of the tangent to the graph at each point is equal to its -coordinate at that point. Its inverse function is the natural logarithm, denoted formula_8 formula_9 or formula_10 because of this, some old texts refer to the exponential function as the antilogarithm. The exponential function satisfies the fundamental multiplicative identity (This identity extends to complex-valued exponents.) It can be shown that every continuous, nonzero solution of the functional equation formula_11 is an exponential function, formula_12 with formula_13 The multiplicative identity, along with the definition formula_14, shows that formula_15 for positive integers , relating the exponential function to the elementary notion of exponentiation. The argument of the exponential function can be any real or complex number or even an entirely different kind of mathematical object (for example, a matrix). Its ubiquitous occurrence in pure and applied mathematics has led mathematician W. Rudin to opine that the exponential function is "the most important function in mathematics". In applied settings, exponential functions model a relationship in which a constant change in the independent variable gives the same proportional change (that is, percentage increase or decrease) in the dependent variable. This occurs widely in the natural and social sciences, as for a self-reproducing population; a fund accruing compound interest; or a growing body of manufacturing expertise. Thus, the exponential function also appears in a variety of contexts within physics, chemistry, engineering, mathematical biology, and economics. The real exponential function formula_16 can be characterized in a variety of equivalent ways. It is commonly defined by the following power series: Since the radius of convergence of this power series is infinite, this definition is, in fact, applicable to all complex numbers formula_18 (see below for the extension of formula_19 to the complex plane). The constant can then be defined as formula_20 The term-by-term differentiation of this power series reveals that formula_21 for all real , leading to another common characterization of formula_19 as the unique solution of the differential equation satisfying the initial condition formula_24 Based on this characterization, the chain rule shows that its inverse function, the natural logarithm, satisfies formula_25 for formula_26 or formula_27 This relationship leads to a less common definition of the real exponential function formula_19 as the solution formula_29 to the equation By way of the binomial theorem and the power series definition, the exponential function can also be defined as the following limit: The exponential function arises whenever a quantity grows or decays at a rate proportional to its current value. One such situation is continuously compounded interest, and in fact it was this observation that led Jacob Bernoulli in 1683 to the number now known as . Later, in 1697, Johann Bernoulli studied the calculus of the exponential function. If a principal amount of 1 earns interest at an annual rate of compounded monthly, then the interest earned each month is times the current value, so each month the total value is multiplied by , and the value at the end of the year is . If instead interest is compounded daily, this becomes . Letting the number of time intervals per year grow without bound leads to the limit definition of the exponential function, first given by Leonhard Euler. This is one of a number of characterizations of the exponential function; others involve series or differential equations. From any of these definitions it can be shown that the exponential function obeys the basic exponentiation identity, which justifies the notation for . The derivative (rate of change) of the exponential function is the exponential function itself. More generally, a function with a rate of change "proportional" to the function itself (rather than equal to it) is expressible in terms of the exponential function. This function property leads to exponential growth or exponential decay. The exponential function extends to an entire function on the complex plane. Euler's formula relates its values at purely imaginary arguments to trigonometric functions. The exponential function also has analogues for which the argument is a matrix, or even an element of a Banach algebra or a Lie algebra. The importance of the exponential function in mathematics and the sciences stems mainly from its property as the unique function which is equal to its derivative and is equal to 1 when . That is, Functions of the form for constant are the only functions that are equal to their derivative (by the Picard–Lindelöf theorem). Other ways of saying the same thing include: If a variable's growth or decay rate is proportional to its size—as is the case in unlimited population growth (see Malthusian catastrophe), continuously compounded interest, or radioactive decay—then the variable can be written as a constant times an exponential function of time. Explicitly for any real constant , a function satisfies if and only if for some constant . The constant "k" is called the decay constant, disintegration constant, rate constant, or transformation constant. Furthermore, for any differentiable function , we find, by the chain rule: A continued fraction for can be obtained via an identity of Euler: The following generalized continued fraction for converges more quickly: or, by applying the substitution : with a special case for : This formula also converges, though more slowly, for . For example: As in the real case, the exponential function can be defined on the complex plane in several equivalent forms. The most common definition of the complex exponential function parallels the power series definition for real arguments, where the real variable is replaced by a complex one: Term-wise multiplication of two copies of these power series in the Cauchy sense, permitted by Mertens' theorem, shows that the defining multiplicative property of exponential functions continues to hold for all complex arguments: The definition of the complex exponential function in turn leads to the appropriate definitions extending the trigonometric functions to complex arguments. In particular, when formula_45 (formula_46 real), the series definition yields the expansion In this expansion, the rearrangement of the terms into real and imaginary parts is justified by the absolute convergence of the series. The real and imaginary parts of the above expression in fact correspond to the series expansions of and , respectively. This correspondence provides motivation for cosine and sine for all complex arguments in terms of formula_48 and the equivalent power series: for all formula_50 The functions , , and so defined have infinite radii of convergence by the ratio test and are therefore entire functions ("i.e.", holomorphic on formula_51). The range of the exponential function is formula_52, while the ranges of the complex sine and cosine functions are both formula_51 in its entirety, in accord with Picard's theorem, which asserts that the range of a nonconstant entire function is either all of formula_51, or formula_51 excluding one lacunary value. These definitions for the exponential and trigonometric functions lead trivially to Euler's formula: We could alternatively define the complex exponential function based on this relationship. If formula_58, where formula_59 and formula_29 are both real, then we could define its exponential as where , , and on the right-hand side of the definition sign are to be interpreted as functions of a real variable, previously defined by other means. For formula_62, the relationship formula_63 holds, so that formula_64 for real formula_46 and formula_66 maps the real line (mod formula_67) to the unit circle. Based on the relationship between formula_68 and the unit circle, it is easy to see that, restricted to real arguments, the definitions of sine and cosine given above coincide with their more elementary definitions based on geometric notions. The complex exponential function is periodic with period formula_69 and formula_70 holds for all formula_71. When its domain is extended from the real line to the complex plane, the exponential function retains the following properties: for all formula_44. Extending the natural logarithm to complex arguments yields the complex logarithm , which is a multivalued function. We can then define a more general exponentiation: for all complex numbers and . This is also a multivalued function, even when is real. This distinction is problematic, as the multivalued functions and are easily confused with their single-valued equivalents when substituting a real number for . The rule about multiplying exponents for the case of positive real numbers must be modified in a multivalued context: See failure of power and logarithm identities for more about problems with combining powers. The exponential function maps any line in the complex plane to a logarithmic spiral in the complex plane with the center at the origin. Two special cases exist: when the original line is parallel to the real axis, the resulting spiral never closes in on itself; when the original line is parallel to the imaginary axis, the resulting spiral is a circle of some radius. Considering the complex exponential function as a function involving four real variables: the graph of the exponential function is a two-dimensional surface curving through four dimensions. Starting with a color-coded portion of the formula_80 domain, the following are depictions of the graph as variously projected into two or three dimensions. The second image shows how the domain complex plane is mapped into the range complex plane: The third and fourth images show how the graph in the second image extends into one of the other two dimensions not shown in the second image. The third image shows the graph extended along the real formula_59 axis. It shows the graph is a surface of revolution about the formula_59 axis of the graph of the real exponential function, producing a horn or funnel shape. The fourth image shows the graph extended along the imaginary formula_29 axis. It shows that the graph's surface for positive and negative formula_29 values doesn't really meet along the negative real formula_82 axis, but instead forms a spiral surface about the formula_29 axis. Because its formula_29 values have been extended to ±2π, this image also better depicts the 2π periodicity in the imaginary formula_29 value. Complex exponentiation can be defined by converting to polar coordinates and using the identity : However, when is not an integer, this function is multivalued, because is not unique (see failure of power and logarithm identities). The power series definition of the exponential function makes sense for square matrices (for which the function is called the matrix exponential) and more generally in any unital Banach algebra . In this setting, , and is invertible with inverse for any in . If , then , but this identity can fail for noncommuting and . Some alternative definitions lead to the same function. For instance, can be defined as Or can be defined as , where is the solution to the differential equation , with initial condition ; it follows that for every in Given a Lie group and its associated Lie algebra formula_94, the exponential map is a map formula_94 satisfying similar properties. In fact, since is the Lie algebra of the Lie group of all positive real numbers under multiplication, the ordinary exponential function for real arguments is a special case of the Lie algebra situation. Similarly, since the Lie group of invertible matrices has as Lie algebra , the space of all matrices, the exponential function for square matrices is a special case of the Lie algebra exponential map. The identity can fail for Lie algebra elements and that do not commute; the Baker–Campbell–Hausdorff formula supplies the necessary correction terms. The function is not in (i.e., is not the quotient of two polynomials with complex coefficients). For distinct complex numbers }, the set } is linearly independent over . The function is transcendental over . When computing (an approximation of) the exponential function near the argument , the result will be close to 1, and computing the value of the difference formula_96 with floating-point arithmetic may lead to the loss of (possibly all) significant figures, producing a large calculation error, possibly even a meaningless result. Following a proposal by William Kahan, it may thus be useful to have a dedicated routine, often called codice_1, for computing directly, bypassing computation of . For example, if the exponential is computed by using its Taylor series one may use the Taylor series of formula_98 This was first implemented in 1979 in the Hewlett-Packard HP-41C calculator, and provided by several calculators, operating systems (for example Berkeley UNIX 4.3BSD), computer algebra systems, and programming languages (for example C99). In addition to base , the IEEE 754-2008 standard defines similar exponential functions near 0 for base 2 and 10: formula_100 and formula_101. A similar approach has been used for the logarithm (see lnp1). An identity in terms of the hyperbolic tangent, gives a high-precision value for small values of on systems that do not implement .
https://en.wikipedia.org/wiki?curid=9678
Prince Eugene of Savoy Prince Eugene Francis of Savoy–Carignano (18 October 1663 – 21 April 1736) was a field marshal in the army of the Holy Roman Empire and of the Austrian Habsburg dynasty during the 17th and 18th centuries. He was one of the most successful military commanders of his time, and rose to the highest offices of state at the Imperial court in Vienna. Born in Paris, Eugene grew up around the court of King Louis XIV of France. Based on his poor physique and bearing, the Prince was initially prepared for a clerical career, but by the age of 19 he had determined on a military career. Following a scandal involving his mother Olympe, he was rejected by Louis XIV for service in the French army. Eugene moved to Austria and transferred his loyalty to the Habsburg Monarchy. Spanning six decades, Eugene served three Holy Roman emperors: Leopold I, Joseph I, and Charles VI. He first saw action against the Ottoman Turks at the Siege of Vienna in 1683 and the subsequent War of the Holy League, before serving in the Nine Years' War, fighting alongside his cousin, the Duke of Savoy. However, the Prince's fame was secured with his decisive victory against the Ottomans at the Battle of Zenta in 1697, earning him Europe-wide fame. Eugene enhanced his standing during the War of the Spanish Succession, where his partnership with the Duke of Marlborough secured victories against the French on the fields of Blenheim (1704), Oudenarde (1708), and Malplaquet (1709); he gained further success in the war as Imperial commander in northern Italy, most notably at the Battle of Turin (1706). Renewed hostilities against the Ottomans in the Austro-Turkish War consolidated his reputation, with victories at the battles of Petrovaradin (1716), and the decisive encounter at Belgrade (1717). Throughout the late 1720s, Eugene's influence and skillful diplomacy managed to secure the Emperor powerful allies in his dynastic struggles with the Bourbon powers, but physically and mentally fragile in his later years, Eugene enjoyed less success as commander-in-chief of the army during his final conflict, the War of the Polish Succession. Nevertheless, in Austria, Eugene's reputation remains unrivalled. Although opinions differ as to his character, there is no dispute over his great achievements: he helped to save the Habsburg Empire from French conquest; he broke the westward thrust of the Ottomans, liberating parts of Europe after a century and a half of Turkish occupation; and he was one of the great patrons of the arts whose building legacy can still be seen in Vienna today. Eugene died in his sleep at his home on 21 April 1736, aged 72. Prince Eugene was born in the Hôtel de Soissons in Paris on 18 October 1663. His mother, Olympia Mancini, was one of Cardinal Mazarin's nieces whom he had brought to Paris from Rome in 1647 to further his, and, to a lesser extent, their ambitions. The Mancinis were raised at the Palais-Royal along with the young Louis XIV, with whom Olympia formed an intimate relationship. Yet to her great disappointment, her chance to become queen passed by, and in 1657, Olympia married Eugene Maurice, Count of Soissons, Count of Dreux and Prince of Savoy. Together they had had five sons (Eugene being the youngest) and three daughters, but neither parent spent much time with the children: his father, a brave, unglamorous French soldier, spent much of his time away campaigning, while Olympia's passion for court intrigue meant the children received little attention from her. The King remained strongly attached to Olympia, so much so that many believed them to be lovers; but her scheming eventually led to her downfall. After falling out of favour at court, Olympia turned to Catherine Deshayes (known as "La Voisin"), and the arts of black magic and astrology. It was a fatal relationship. Embroiled in the "affaire des poisons", suspicions now abounded of her involvement in her husband's premature death in 1673, and even implicated her in a plot to kill the King himself. Whatever the truth, Olympia, rather than face trial, subsequently fled France for Brussels in January 1680, leaving Eugene in the care of his father's mother, Marie de Bourbon, and her daughter, Hereditary Princess of Baden, mother of Prince Louis of Baden. From the age of ten, Eugene had been brought up for a career in the church; a personal choice of the King, basing the decision on the young Prince's poor physique and bearing. Certainly Eugene's appearance was not impressive — "He was never good-looking …" wrote the Duchess of Orléans, "It is true that his eyes are not ugly, but his nose ruins his face; he has two large teeth which are visible at all times." In February 1683, to the surprise of his family, Eugene declared his intention of joining the army. Now 19 years old, Eugene applied directly to Louis XIV for command of a company in French service, but the King—who had shown no compassion for Olympia's children since her disgrace—refused him out of hand. "The request was modest, not so the petitioner," he remarked. "No one else ever presumed to stare me out so insolently." Whatever the case, Louis XIV's choice would cost him dearly twenty years later, for it would be precisely Eugene, in collaboration with the Duke of Marlborough, who would defeat the French army at Blenheim, a decisive battle which checked French military supremacy and political power. Denied a military career in France, Eugene decided to seek service abroad. One of Eugene's brothers, Louis Julius, had entered Imperial service the previous year, but he had been immediately killed fighting the Ottoman Turks in 1683. When news of his death reached Paris, Eugene decided to travel to Austria in the hope of taking over his brother's command. It was not an unnatural decision: his cousin, Louis of Baden, was already a leading general in the Imperial army, as was a more distant cousin, Maximilian II Emanuel, Elector of Bavaria. On the night of July 26, 1683, Eugene left Paris and headed east. By May 1683, the Ottoman threat to Emperor Leopold I's capital, Vienna, was very real. The Grand Vizier, Kara Mustafa Pasha—encouraged by Imre Thököly's Magyar rebellion—had invaded Hungary with between 100,000–200,000 men; within two months approximately 90,000 were beneath Vienna's walls. With the 'Turks at the gates', the Emperor fled for the safe refuge of Passau up the Danube, a more distant and secure part of his dominion. It was at Leopold I's camp that Eugene arrived in mid-August. Although Eugene was not of Austrian extraction, he did have Habsburg antecedents. His grandfather, Thomas Francis, founder of the Carignano line of the House of Savoy, was the son of Catherine Michelle—a daughter of Philip II of Spain—and the great-grandson of the Emperor Charles V. But of more immediate consequence to Leopold I was the fact that Eugene was the second cousin of Victor Amadeus, the Duke of Savoy, a connection that the Emperor hoped might prove useful in any future confrontation with France. These ties, together with his ascetic manner and appearance (a positive advantage to him at the sombre court of Leopold I), ensured the refugee from the hated French king a warm welcome at Passau, and a position in Imperial service. Though French was his favored language, he communicated with Leopold in Italian, as the Emperor (though he knew it perfectly) disliked French. But Eugene also had a reasonable command of German, which he understood very easily, something that helped him much in the military. Eugene was in no doubt where his new allegiance lay—"I will devote all my strength, all my courage, and if need be, my last drop of blood, to the service of your Imperial Majesty." This loyalty was immediately put to the test. By September, the Imperial forces under the Duke of Lorraine, together with a powerful Polish army under King John III Sobieski, were poised to strike the Sultan's army. On the morning of 12 September, the Christian forces drew up in line of battle on the south-eastern slopes of the Vienna Woods, looking down on the massed enemy camp. The day-long Battle of Vienna resulted in the lifting of the 60-day siege, and the Sultan's forces were routed and in retreat. Serving under Baden, Eugene distinguished himself in the battle, earning commendation from Lorraine and the Emperor; he later received the nomination for the colonelcy of the Dragoon Regiment Kufstein. In March 1684, Leopold I formed the Holy League with Poland and Venice to counter the Ottoman threat. For the next two years, Eugene continued to perform with distinction on campaign and establish himself as a dedicated, professional soldier; by the end of 1685, still only 22 years old, he was made a Major-General. However, little is known of Eugene's life during these early campaigns. Contemporary observers make only passing comments of his actions, and his own surviving correspondence, largely to his cousin Victor Amadeus, are typically reticent about his own feelings and experiences. Nevertheless, it is clear that Baden was impressed with Eugene's qualities—"This young man will, with time, occupy the place of those whom the world regards as great leaders of armies." In June 1686, the Duke of Lorraine besieged Buda (Budapest), the centre of the Ottoman occupation in Hungary. After resisting for 78 days, the city fell on 2 September, and Turkish resistance collapsed throughout the region as far away as Transylvania and Serbia. Further success followed in 1687, where, commanding a cavalry brigade, Eugene made an important contribution to the victory at the Battle of Mohács on 12 August. Such was the scale of their defeat that the Ottoman army mutinied—a revolt which spread to Constantinople. The Grand Vizier, Suluieman Pasha, was executed and Sultan Mehmed IV, deposed. Once again, Eugene's courage earned him recognition from his superiors, who granted him the honour of personally conveying the news of victory to the Emperor in Vienna. For his services, Eugene was promoted to Lieutenant-General in November 1687. He was also gaining wider recognition. King Charles II of Spain bestowed upon him the Order of the Golden Fleece, while his cousin, Victor Amadeus, provided him with money and two profitable abbeys in Piedmont. However, Eugene's military career suffered a temporary setback in 1688 when, on 6 September, the Prince suffered a severe wound to his knee by a musket ball during the Siege of Belgrade. It was not until January 1689 that he could return to active service. Just as Belgrade was falling to Imperial forces under Max Emmanuel in the east, French troops in the west were crossing the Rhine into the Holy Roman Empire. Louis XIV had hoped that a show of force would lead to a quick resolution to his dynastic and territorial disputes with the princes of the Empire along his eastern border, but his intimidatory moves only strengthened German resolve, and in May 1689, Leopold I and the Dutch signed an offensive compact aimed at repelling French aggression. The Nine Years' War was professionally and personally frustrating for the Prince. Initially fighting on the Rhine with Max Emmanuel—receiving a slight head wound at the Siege of Mainz in 1689—Eugene subsequently transferred himself to Piedmont after Victor Amadeus joined the Alliance against France in 1690. Promoted to general of cavalry, he arrived in Turin with his friend the Prince of Commercy; but it proved an inauspicious start. Against Eugene's advice, Amadeus insisted on engaging the French at Staffarda and suffered a serious defeat—only Eugene's handling of the Savoyard cavalry in retreat saved his cousin from disaster. Eugene remained unimpressed with the men and their commanders throughout the war in Italy. "The enemy would long ago have been beaten," he wrote to Vienna, "if everyone had done their duty." So contemptuous was he of the Imperial commander, Count Caraffa, he threatened to leave Imperial service. In Vienna, Eugene's attitude was dismissed as the arrogance of a young upstart, but so impressed was the Emperor by his passion for the Imperial cause, he promoted him to Field-Marshal in 1693. When Caraffa's replacement, Count Caprara, was himself transferred in 1694, it seemed that Eugene's chance for command and decisive action had finally arrived. But Amadeus, doubtful of victory and now more fearful of Habsburg influence in Italy than he was of French, had begun secret dealings with Louis XIV aimed at extricating himself from the war. By 1696, the deal was done, and Amadeus transferred his troops and his loyalty to the enemy. Eugene was never to fully trust his cousin again; although he continued to pay due reverence to the Duke as head of his family, their relationship would forever after remain strained. Military honours in Italy undoubtedly belonged to the French commander Marshal Catinat, but Eugene, the one Allied general determined on action and decisive results, did well to emerge from the Nine Years' War with an enhanced reputation. With the signing of the Treaty of Ryswick in September/October 1697, the desultory war in the west was finally brought to an inconclusive end, and Leopold I could once again devote all his martial energies into defeating the Ottoman Turks in the east. The distractions of the war against Louis XIV had enabled the Turks to recapture Belgrade in 1690. In August 1691, the Austrians, under Louis of Baden, regained the advantage by heavily defeating the Turks at the Battle of Slankamen on the Danube, securing Habsburg possession of Hungary and Transylvania. However, when Baden was transferred west to fight the French in 1692, his successors, first Caprara, then from 1696, Frederick Augustus, the Elector of Saxony, proved incapable of delivering the final blow. On the advice of the President of the Imperial War Council, Rüdiger Starhemberg, Eugene was offered supreme command of Imperial forces in April 1697. This was Eugene's first truly independent command—no longer need he suffer under the excessively cautious generalship of Caprara and Caraffa, or be thwarted by the deviations of Victor Amadeus. But on joining his army, he found it in a state of 'indescribable misery'. Confident and self-assured, the Prince of Savoy (ably assisted by Commercy and Guido Starhemberg) set about restoring order and discipline. Leopold I had warned Eugene to act cautiously, but when the Imperial commander learnt of Sultan Mustafa II's march on Transylvania, Eugene abandoned all ideas of a defensive campaign and moved to intercept the Turks as they crossed the River Tisza at Zenta on 11 September 1697. It was late in the day before the Imperial army struck. The Turkish cavalry had already crossed the river so Eugene decided to attack immediately, arranging his men in a half-moon formation. The vigour of the assault wrought terror and confusion amongst the Turks, and by nightfall, the battle was won. For the loss of some 2,000 dead and wounded, Eugene had inflicted approximately 25,000 casualties on his enemy—including the Grand Vizier, Elmas Mehmed Pasha—annihilating the Turkish army. Although the Ottomans lacked western organisation and training, the Savoyard prince had revealed his tactical skill, his capacity for bold decision, and his ability to inspire his men to excel in battle against a dangerous foe. After a brief terror-raid into Ottoman-held Bosnia, culminating in the sack of Sarajevo, Eugene returned to Vienna in November to a triumphal reception. His victory at Zenta had turned him into a European hero, and with victory came reward. Land in Hungary, given him by the Emperor, yielded a good income, enabling the Prince to cultivate his newly acquired tastes in art and architecture (see below); but for all his new-found wealth and property, he was, nevertheless, without personal ties or family commitments. Of his four brothers, only one was still alive at this time. His fourth brother, Emmanuel, had died aged 14 in 1676; his third, Louis Julius (already mentioned) had died on active service in 1683, and his second brother, Philippe, died of smallpox in 1693. Eugene's remaining brother, Louis Thomas—ostracised for incurring the displeasure of Louis XIV—travelled Europe in search of a career, before arriving in Vienna in 1699. With Eugene's help, Louis found employment in the Imperial army, only to be killed in action against the French in 1702. Of Eugene's sisters, the youngest had died in childhood. The other two, Marie Jeanne-Baptiste and Louise Philiberte, led dissolute lives. Expelled from France, Marie joined her mother in Brussels, before eloping with a renegade priest to Geneva, living with him unhappily until her premature death in 1705. Of Louise, little is known after her early salacious life in Paris, but in due course, she lived for a time in a convent in Savoy before her death in 1726. The Battle of Zenta proved to be the decisive victory in the long war against the Turks. With Leopold I's interests now focused on Spain and the imminent death of Charles II, the Emperor terminated the conflict with the Sultan, and signed the Treaty of Karlowitz on 26 January 1699. With the death of the infirm and childless Charles II of Spain on 1 November 1700, the succession of the Spanish throne and subsequent control over her empire once again embroiled Europe in war—the War of the Spanish Succession. On his deathbed Charles II had bequeathed the entire Spanish inheritance to Louis XIV's grandson, Philip, Duke of Anjou. This threatened to unite the Spanish and French kingdoms under the House of Bourbon—something unacceptable to England, the Dutch Republic, and Leopold I, who had himself a claim to the Spanish throne. From the beginning, the Emperor had refused to accept the will of Charles II, and he did not wait for England and the Dutch Republic to begin hostilities. Before a new Grand Alliance could be concluded Leopold I prepared to send an expedition to seize the Spanish lands in Italy. Eugene crossed the Alps with some 30,000 men in May/June 1701. After a series of brilliant manoeuvres the Imperial commander defeated Catinat at the Battle of Carpi on 9 July. "I have warned you that you are dealing with an enterprising young prince," wrote Louis XIV to his commander, "he does not tie himself down to the rules of war." On 1 September Eugene defeated Catinat's successor, Marshal Villeroi, at the Battle of Chiari, in a clash as destructive as any in the Italian theatre. But as so often throughout his career the Prince faced war on two fronts—the enemy in the field and the government in Vienna. Starved of supplies, money and men, Eugene was forced into unconventional means against the vastly superior enemy. During a daring raid on Cremona on the night of 31 January/1 February 1702 Eugene captured the French commander-in-chief. Yet the coup was less successful than hoped: Cremona remained in French hands, and the Duke of Vendôme, whose talents far exceeded Villeroi's, became the theatre's new commander. Villeroi's capture caused a sensation in Europe, and had a galvanising effect on English public opinion. "The surprise at Cremona," wrote the diarist John Evelyn, "… was the greate discourse of this weeke"; but appeals for succour from Vienna remained unheeded, forcing Eugene to seek battle and gain a 'lucky hitt'. The resulting Battle of Luzzara on 15 August proved inconclusive. Although Eugene's forces inflicted double the number of casualties on the French the battle settled little except to deter Vendôme trying an all-out assault on Imperial forces that year, enabling Eugene to hold on south of the Alps. With his army rotting away, and personally grieving for his long standing friend Prince Commercy who had died at Luzzara, Eugene returned to Vienna in January 1703. Eugene's European reputation was growing (Cremona and Luzzara had been celebrated as victories throughout the Allied capitals), yet because of the condition and morale of his troops the 1702 campaign had not been a success. Austria itself was now facing the direct threat of invasion from across the border in Bavaria where the state's Elector, Maximilian Emanuel, had declared for the Bourbons in August the previous year. Meanwhile, in Hungary a small-scale revolt had broken out in May and was fast gaining momentum. With the monarchy at the point of complete financial breakdown Leopold I was at last persuaded to change the government. At the end of June 1703 Gundaker Starhemberg replaced Gotthard Salaburg as President of the Treasury, and Prince Eugene succeeded Henry Mansfeld as the new President of the Imperial War Council ("Hofkriegsratspräsident"). As head of the war council Eugene was now part of the Emperor's inner circle, and the first president since Montecuccoli to remain an active commander. Immediate steps were taken to improve efficiency within the army: encouragement and, where possible, money, was sent to the commanders in the field; promotion and honours were distributed according to service rather than influence; and discipline improved. But the Austrian monarchy faced severe peril on several fronts in 1703: by June the Duke of Villars had reinforced the Elector of Bavaria on the Danube thus posing a direct threat to Vienna, while Vendôme remained at the head of a large army in northern Italy opposing Guido Starhemberg's weak Imperial force. Of equal alarm was Francis II Rákóczi's revolt which, by the end of the year, had reached as far as Moravia and Lower Austria. Dissension between Villars and the Elector of Bavaria had prevented an assault on Vienna in 1703, but in the Courts of Versailles and Madrid, ministers confidently anticipated the city's fall. The Imperial ambassador in London, Count Wratislaw, had pressed for Anglo-Dutch assistance on the Danube as early as February 1703, but the crisis in southern Europe seemed remote from the Court of St. James's where colonial and commercial considerations were more to the fore of men's minds. Only a handful of statesmen in England or the Dutch Republic realised the true implications of Austria's peril; foremost amongst these was the English Captain-General, the Duke of Marlborough. By early 1704 Marlborough had resolved to march south and rescue the situation in southern Germany and on the Danube, personally requesting the presence of Eugene on campaign so as to have "a supporter of his zeal and experience". The Allied commanders met for the first time at the small village of Mundelsheim on 10 June, and immediately formed a close rapport—the two men becoming, in the words of Thomas Lediard, 'Twin constellations in glory'. This professional and personal bond ensured mutual support on the battlefield, enabling many successes during the Spanish Succession war. The first of these victories, and the most celebrated, came on 13 August 1704 at the Battle of Blenheim. Eugene commanded the right wing of the Allied army, holding the Elector of Bavaria's and Marshal Marsin's superior forces, while Marlborough broke through the Marshal Tallard's center, inflicting over 30,000 casualties. The battle proved decisive: Vienna was saved and Bavaria was knocked out of the war. Both Allied commanders were full of praise for each other's performance. Eugene's holding operation, and his pressure for action leading up to the battle, proved crucial for the Allied success. In Europe Blenheim is regarded as much a victory for Eugene as it is for Marlborough, a sentiment echoed by Sir Winston Churchill (Marlborough's descendant and biographer), who pays tribute to "the glory of Prince Eugene, whose fire and spirit had exhorted the wonderful exertions of his troops." France now faced the real danger of invasion, but Leopold I in Vienna was still under severe strain: Rákóczi's revolt was a major threat; and Guido Starhemberg and Victor Amadeus (who had once again switched loyalties and rejoined the Grand Alliance in 1703) had been unable to halt the French under Vendôme in northern Italy. Only Amadeus' capital, Turin, held on. Eugene returned to Italy in April 1705, but his attempts to move west towards Turin were thwarted by Vendôme's skillful manoeuvres. Lacking boats and bridging materials, and with desertion and sickness rife within his army, the outnumbered Imperial commander was helpless. Leopold I's assurances of money and men had proved illusory, but desperate appeals from Amadeus and criticism from Vienna goaded the Prince into action, resulting in the Imperialists' bloody defeat at the Battle of Cassano on 16 August. However, following Leopold I's death and the accession of Joseph I to the Imperial throne in May 1705, Eugene at last began to receive the personal backing he desired. Joseph I proved to be a strong supporter of Eugene's supremacy in military affairs; he was the most effective emperor the Prince served and the one he was happiest under. Promising support, Joseph I persuaded Eugene to return to Italy and restore Habsburg honour. The Imperial commander arrived in theatre in mid-April 1706, just in time to organise an orderly retreat of what was left of Count Reventlow's inferior army following his defeat by Vendôme at the Battle of Calcinato on 19 April. Vendôme now prepared to defend the lines along the river Adige, determined to keep Eugene cooped to the east while the Marquis of La Feuillade threatened Turin. However, feigning attacks along the Adige, Eugene descended south across the river Po in mid-July, outmanoeuvring the French commander and gaining a favourable position from which he could at last move west towards Piedmont and relieve Savoy's capital. Events elsewhere were now to have major consequences for the war in Italy. With Villeroi's crushing defeat by Marlborough at the Battle of Ramillies on 23 May, Louis XIV recalled Vendôme north to take command of French forces in Flanders. It was a transfer that Saint-Simon considered something of a deliverance for the French commander who was "now beginning to feel the unlikelihood of success [in Italy] … for Prince Eugene, with the reinforcements that had joined him after the Battle of Calcinato, had entirely changed the outlook in that theatre of the war." The Duke of Orléans, under the direction of Marsin, replaced Vendôme, but indecision and disorder in the French camp led to their undoing. After uniting his forces with Victor Amadeus at Villastellone in early September, Eugene attacked, overwhelmed, and decisively defeated the French forces besieging Turin on 7 September. Eugene's success broke the French hold on northern Italy, and the whole Po valley fell under Allied control. Eugene had gained a victory as signal as his colleague had at Ramillies—"It is impossible for me to express the joy it has given me;" wrote Marlborough, "for I not only esteem but I really love the prince. This glorious action must bring France so low, that if our friends could but be persuaded to carry on the war with vigour one year longer, we cannot fail, with the blessing of God, to have such a peace as will give us quiet for all our days." The Imperial victory in Italy marked the beginning of Austrian rule in Lombardy, and earned Eugene the Governorship of Milan. But the following year was to prove a disappointment for the Prince and the Grand Alliance as a whole. The Emperor and Eugene (whose main goal after Turin was to take Naples and Sicily from Philip duc d'Anjou's supporters), reluctantly agreed to Marlborough's plan for an attack on Toulon—the seat of French naval power in the Mediterranean. However, disunion between the Allied commanders—Victor Amadeus, Eugene, and the English Admiral Shovell—doomed the Toulon enterprise to failure. Although Eugene favoured some sort of attack on France's south-eastern border it was clear he felt the expedition impractical, and had shown none of the "alacrity which he had displayed on other occasions." Substantial French reinforcements finally brought an end to the venture, and on 22 August 1707 the Imperial army began its retirement. The subsequent capture of Susa could not compensate for the total collapse of the Toulon expedition and with it any hope of an Allied war-winning blow that year. At the beginning of 1708 Eugene successfully evaded calls for him to take charge in Spain (in the end Guido Starhemberg was sent), thus enabling him to take command of the Imperial army on the Moselle and once again unite with Marlborough in the Spanish Netherlands. Eugene (without his army) arrived at the Allied camp at Assche, west of Brussels, in early July, providing a welcome boost to morale after the early defection of Bruges and Ghent to the French. " … our affairs improved through God's support and Eugene's aid," wrote the Prussian General Natzmer, "whose timely arrival raised the spirits of the army again and consoled us." Heartened by the Prince's confidence the Allied commanders devised a bold plan to engage the French army under Vendôme and the Duke of Burgundy. On 10 July the Anglo-Dutch army made a forced march to surprise the French, reaching the river Scheldt just as the enemy were crossing to the north. The ensuing battle on 11 July—more a contact action rather than a set-piece engagement—ended in a resounding success for the Allies, aided by the dissension of the two French commanders. While Marlborough remained in overall command, Eugene had led the crucial right flank and centre. Once again the Allied commanders had co-operated remarkably well. "Prince Eugene and I," wrote the Duke, "shall never differ about our share of the laurels." Marlborough now favoured a bold advance along the coast to bypass the major French fortresses, followed by a march on Paris. But fearful of unprotected supply-lines, the Dutch and Eugene favoured a more cautious approach. Marlborough acquiesced and resolved upon the siege of Vauban's great fortress, Lille. While the Duke commanded the covering force, Eugene oversaw the siege of the town which surrendered on 22 October; however, it was not until 10 December that the resolute Marshal Boufflers yielded the citadel. Yet for all the difficulties of the siege (Eugene was badly wounded above his left eye by a musket ball, and even survived an attempt to poison him), the campaign of 1708 had been a remarkable success. The French were driven out of almost all the Spanish Netherlands. "He who has not seen this," wrote Eugene, "has seen nothing." The recent defeats, together with the severe winter of 1708–09, had caused extreme famine and privation in France. Louis XIV was close to accepting Allied terms, but the conditions demanded by the leading Allied negotiators, Anthonie Heinsius, Charles Townshend, Marlborough, and Eugene—principally that Louis XIV should use his own troops to force Philip V off the Spanish throne—proved unacceptable to the French. Neither Eugene nor Marlborough had objected to the Allied demands at the time, but neither wanted the war with France to continue, and would have preferred further talks to deal with the Spanish issue. But the French King offered no further proposals. Lamenting the collapse of the negotiations, and aware of the vagaries of war, Eugene wrote to the Emperor in mid-June 1709. "There can be no doubt that the next battle will be the biggest and bloodiest that has yet been fought." After the fall of Tournai on 3 September (itself a major undertaking), the Allied generals turned their attention towards Mons. Marshal Villars, recently joined by Boufflers, moved his army south-west of the town and began to fortify his position. Marlborough and Eugene favoured an engagement before Villars could render his position impregnable; but they also agreed to wait for reinforcements from Tournai which did not arrive until the following night, thus giving the French further opportunity to prepare their defences. Notwithstanding the difficulties of the attack, however, the Allied generals did not shrink from their original determination. The subsequent Battle of Malplaquet, fought on 11 September 1709, was the bloodiest engagement of the war. On the left flank, the Prince of Orange led his Dutch infantry in desperate charges only to have it cut to pieces; on the other flank, Eugene attacked and suffered almost as severely. But sustained pressure on his extremities forced Villars to weaken his centre, thus enabling Marlborough to breakthrough and claim victory. Villars was unable to save Mons, which subsequently capitulated on 21 October, but his resolute defence at Malplaquet—inflicting up to 25% casualties on the Allies—may have saved France from destruction. In August 1709 Eugene's chief political opponent and critic in Vienna, Prince Salm, retired as court chamberlain. Eugene and Wratislaw were now the undisputed leaders of the Austrian government: all major departments of state were in their hands or those of their political allies. However, another attempt at a negotiated settlement at Geertruidenberg in April 1710 failed, largely because the English Whigs still felt strong enough to refuse concessions, while Louis XIV saw little reason to accept what he had refused the previous year. Eugene and Marlborough could not be accused of wrecking the negotiations, but neither showed regret at the breakdown of the talks. There was no alternative but to continue the war, and in June the Allied commanders captured Douai. This success was followed by a series of minor sieges, and by the close of 1710 the Allies had cleared much of France's protective ring of fortresses. Yet there had been no final, decisive breakthrough, and this was to be the last year that Eugene and Marlborough would work together. Following the death of Joseph I on 17 April 1711 his brother, Charles, the pretender to the Spanish throne, became emperor. In England the new Tory government (the 'peace party' who had deposed the Whigs in October 1710) declared their unwillingness to see Charles VI become Emperor as well as King of Spain, and had already begun secret negotiations with the French. In January 1712 Eugene arrived in England hoping to divert the government away from its peace policy, but despite the social success the visit was a political failure: Queen Anne and her ministers remained determined to end the war regardless of the Allies. Eugene had also arrived too late to save Marlborough who, seen by the Tories as the main obstacle to peace, had already been dismissed on charges of embezzlement. Elsewhere, however, the Austrians had made some progress—the Hungarian revolt had finally came to end. Although Eugene would have preferred to crush the rebels the Emperor had offered lenient conditions, leading to the signing of the Treaty of Szatmár on 30 April 1711. Hoping to influence public opinion in England and force the French into making substantial concessions, Eugene prepared for a major campaign. However, on 21 May 1712—when the Tories felt they had secured favourable terms with their unilateral talks with the French—the Duke of Ormonde (Marlborough's successor) received the so-called 'restraining orders', forbidding him to take part in any military action. Eugene took the fortress of Le Quesnoy in early July, before besieging Landrecies, but Villars, taking advantage of Allied disunity, outmanoeuvred Eugene and defeated the Earl of Albermarle's Dutch garrison at Denain on 24 July. The French followed the victory by seizing the Allies' main supply magazine at Marchiennes, before reversing their earlier losses at Douai, Le Quesnoy and Bouchain. In one summer the whole forward Allied position laboriously built up over the years to act as the springboard into France had been precipitously abandoned. With the death in December of his friend and close political ally, Count Wratislaw, Eugene became undisputed 'first minister' in Vienna. His position was built on his military successes, but his actual power was expressed through his role as president of the war council, and as "de facto" president of the conference which dealt with foreign policy. In this position of influence Eugene took the lead in pressing Charles VI towards peace. The government had come to accept that further war in the Netherlands or Spain was impossible without the aid of the Maritime Powers; yet the Emperor, still hoping that somehow he could place himself on the throne in Spain, refused to make peace at the Utrecht conference along with the other Allies. Reluctantly, Eugene prepared for another campaign, but lacking troops, finance, and supplies his prospects in 1713 were poor. Villars, with superior numbers, was able to keep Eugene guessing as to his true intent. Through successful feints and stratagems Landau fell to the French commander in August, followed in November by Freiburg. Eugene was reluctant to carry on the war, and wrote to the Emperor in June that a bad peace would be better than being 'ruined equally by friend and foe'. With Austrian finances exhausted and the German states reluctant to continue the war, Charles VI was compelled to enter into negotiations. Eugene and Villars (who had been old friends since the Turkish campaigns of the 1680s) initiated talks on 26 November. Eugene proved an astute and determined negotiator, and gained favourable terms by the Treaty of Rastatt signed on 7 March 1714 and the Treaty of Baden signed on 7 September 1714. Despite the failed campaign in 1713 the Prince was able to declare that, "in spite of the military superiority of our enemies and the defection of our Allies, the conditions of peace will be more advantageous and more glorious than those we would have obtained at Utrecht." Eugene's main reason for desiring peace in the west was the growing danger posed by the Turks in the east. Turkish military ambitions had revived after 1711 when they had mauled Peter the Great's army on the river Pruth: in December 1714 Sultan Ahmed III's forces attacked the Venetians in the Morea. To Vienna it was clear that the Turks intended to attack Hungary and undo the whole Karlowitz settlement of 1699. After the Porte rejected an offer of mediation in April 1716, Charles VI despatched Eugene to Hungary to lead his relatively small but professional army. Of all Eugene's wars this was the one in which he exercised most direct control; it was also a war which, for the most part, Austria fought and won on her own. Eugene left Vienna in early June 1716 with a field army of between 80,000–90,000 men. By early August 1716 the Ottoman Turks, some 200,000 men under the sultan's son-in-law, the Grand Vizier Damat Ali Pasha, were marching from Belgrade towards Eugene's position west of the fortress of Petrovaradin on the north bank of the Danube. The Grand Vizier had intended to seize the fortress; but Eugene gave him no chance to do so. After resisting calls for caution and forgoing a council of war, the Prince decided to attack immediately on the morning of 5 August with approximately 70,000 men. The Turkish janissaries had some initial success, but after an Imperial cavalry attack on their flank, Ali Pasha's forces fell into confusion. Although the Imperials lost almost 5,000 dead or wounded, the Turks, who retreated in disorder to Belgrade, seem to have lost double that amount, including the Grand Vizier himself who had entered the mêlée and subsequently died of his wounds. Eugene proceeded to take the Banat fortress of Timișoara (Temeswar in German) in mid-October 1716 (thus ending 164 years of Turkish rule), before turning his attention to the next campaign and to what he considered the main goal of the war, Belgrade. Situated at the confluence of the Rivers Danube and Sava, Belgrade held a garrison of 30,000 men under Mustapha Pasha. Imperial troops besieged the place in mid-June 1717, and by the end of July large parts of the city had been destroyed by artillery fire. By the first days of August, however, a huge Turkish field army (150,000–200,000 strong), under the new Grand Vizier, Halil Pasha, had arrived on the plateau east of the city to relieve the garrison. News spread through Europe of Eugene's imminent destruction; but he had no intention of lifting the siege. With his men suffering from dysentery, and continuous bombardment from the plateau, Eugene, aware that a decisive victory alone could extricate his army, decided to attack the relief force. On the morning of 16 August 40,000 Imperial troops marched through the fog, caught the Turks unaware, and routed Halil Pasha's army; a week later Belgrade surrendered, effectively bringing an end to the war. The victory was the crowning point of Eugene's military career and had confirmed him as the leading European general. His ability to snatch victory at the moment of defeat had shown the Prince at his best. The principal objectives of the war had been achieved: the task Eugene had begun at Zenta was complete, and the Karlowitz settlement secured. By the terms of the Treaty of Passarowitz, signed on 21 July 1718, the Turks surrendered the Banat of Temeswar, along with Belgrade and most of Serbia, although they regained the Morea from the Venetians. The war had dispelled the immediate Turkish threat to Hungary, and was a triumph for the Empire and for Eugene personally. While Eugene fought the Turks in the east, unresolved issues following the Utrecht/Rastatt settlements led to hostilities between the Emperor and Philip V of Spain in the west. Charles VI had refused to recognise Philip V as King of Spain, a title which he himself claimed; in return, Philip V had refused to renounce his claims to Naples, Milan, and the Netherlands, all of which had transferred to the House of Austria following the Spanish Succession war. Philip V was roused by his influential wife, Elisabeth Farnese, daughter of the Hereditary Prince of Parma, who personally held dynastic claims in the name of her son, Don Charles, to the duchies of Tuscany, Parma and Piacenza. Representatives from a newly formed Anglo-French alliance—who were desirous of European peace for their own dynastic securities and trade opportunities—called on both parties to recognise each other's sovereignty. Yet Philip V remained intractable, and on 22 August 1717 his chief minister, Alberoni, effected the invasion of Austrian Sardinia in what seemed like the beginning of the reconquest of Spain's former Italian empire. Eugene returned to Vienna from his recent victory at Belgrade (before the conclusion of the Turkish war) determined to prevent an escalation of the conflict, complaining that, "two wars cannot be waged with one army"; only reluctantly did the Prince release some troops from the Balkans for the Italian campaign. Rejecting all diplomatic overtures Philip V unleashed another assault in June 1718, this time against Savoyard Sicily as a preliminary to attacking the Italian mainland. Realising that only the British fleet could prevent further Spanish landings, and that pro-Spanish groups in France might push the regent, Duke of Orléans, into war against Austria, Charles VI had no option but to sign the Quadruple Alliance on 2 August 1718, and formally renounce his claim to Spain. Despite the Spanish fleet's destruction off Cape Passaro, Philip V and Elisabeth remained resolute, and rejected the treaty. Although Eugene could have gone south after the conclusion of the Turkish war, he chose instead to conduct operations from Vienna; but Austria's military effort in Sicily proved derisory, and Eugene's chosen commanders, Zum Jungen, and later Count Mercy, performed poorly. It was only from pressure exerted by the French army advancing into the Basque provinces of northern Spain in April 1719, and the British Navy's attacks on the Spanish fleet and shipping, that compelled Philip V and Elisabeth to dismiss Alberoni and join the Quadruple Alliance on 25 January 1720. Nevertheless, the Spanish attacks had strained Charles VI's government, causing tension between the Emperor and his Spanish Council on the one hand, and the conference, headed by Eugene, on the other. Despite Charles VI's own personal ambitions in the Mediterranean it was clear to the Emperor that Eugene had put the safeguarding of his conquests in Hungary before everything else, and that military failure in Sicily also had to rest on Eugene. Consequently, the Prince's influence over the Emperor declined considerably. Eugene had become governor of the Southern Netherlands—then the Austrian Netherlands—in June 1716, but he was an absent ruler, directing policy from Vienna through his chosen representative the Marquis of Prié. Prié proved unpopular with the local population and the guilds who, following the Barrier Treaty of 1715, were obliged to meet the financial demands of the administration and the Dutch barrier garrisons; with Eugene's backing and encouragement, civil disturbances in Antwerp and Brussels were forcibly suppressed. After displeasing the Emperor over his initial opposition to the formation of the Ostend Company, Prié also lost the support of the native nobility from within his own council of state in Brussels, particularly from the Marquis de Mérode-Westerloo. One of Eugene's former favourites, General Bonneval, also joined the noblemen in opposition to Prié, further undermining the Prince. When Prié's position became untenable, Eugene felt compelled to resign his post as governor of the Southern Netherlands on 16 November 1724. As compensation, Charles VI conferred on him the honorary position as vicar-general of Italy, worth 140,000 gulden a year, and an estate at Siebenbrunn in Lower Austria said to be worth double that amount. But his resignation distressed him, and to compound his concerns Eugene caught a severe bout of influenza that Christmas, marking the beginning of permanent bronchitis and acute infections every winter for the remaining twelve years of his life. The 1720s saw rapidly changing alliances between the European powers and almost constant diplomatic confrontation, largely over unsolved issues regarding the Quadruple Alliance. The Emperor and the Spanish King continued to use each other's titles, and Charles VI still refused to remove the remaining legal obstacles to Don Charles' eventual succession to the duchies of Parma and Tuscany. Yet in a surprise move Spain and Austria moved closer with the signing of the Treaty of Vienna in April/May 1725. In response Britain, France, and Prussia joined together in the Alliance of Hanover to counter the danger to Europe of an Austro-Spanish hegemony. For the next three years there was the continual threat of war between the Hanover Treaty powers and the Austro-Spanish bloc. From 1726 Eugene gradually began to regain his political influence. With his many contacts throughout Europe Eugene, backed by Gundaker Starhemberg and Count Schönborn, the Imperial vice-chancellor, managed to secure powerful allies and strengthen the Emperor's position—his skill in managing the vast secret diplomatic network over the coming years was the main reason why Charles VI once again came to depend upon him. In August 1726 Russia acceded to the Austro-Spanish alliance, and in October Frederick William of Prussia followed suit by defecting from the Allies with the signing of a mutual defensive treaty with the Emperor. Despite the conclusion of the brief Anglo-Spanish conflict, war between the European powers persisted throughout 1727–28. However, in 1729 Elisabeth Farnese abandoned the Austro-Spanish alliance. Realizing that Charles VI could not be drawn into the marriage pact she wanted, Elisabeth concluded that the best way to secure her son's succession to Parma and Tuscany now lay with Britain and France. To Eugene it was 'an event that which is seldom to be found in history'. Following the Prince's determined lead to resist all pressure, Charles VI sent troops into Italy to prevent the entry of Spanish garrisons into the contested duchies. By the beginning of 1730 Eugene, who had remained bellicose throughout the whole period, was again in control of Austrian policy. In Britain there now emerged a new political re-alignment as the Anglo-French "entente" became increasingly defunct. Believing that a resurgent France now posed the greatest danger to their security British ministers, headed by Robert Walpole, moved to reform the Anglo-Austrian alliance, leading to the signing of the Second Treaty of Vienna on 16 March 1731. Eugene had been the Austrian minister most responsible for the alliance, believing once again it would provide security against France and Spain. The treaty compelled Charles VI to sacrifice the Ostend Company and accept, unequivocally, the accession of Don Charles to Parma and Tuscany. In return King George II as King of Great Britain and Elector of Hanover guaranteed the Pragmatic Sanction, the device to secure the rights of the Emperor's daughter, Maria Theresa, to the entire Habsburg inheritance. It was largely through Eugene's diplomacy that in January 1732 the Imperial diet also guaranteed the Pragmatic Sanction which, together with the Treaties with Britain, Russia, and Prussia, marked the culmination of the Prince's diplomacy. But the Treaty of Vienna had infuriated the court of King Louis XV: the French had been ignored and the Pragmatic Sanction guaranteed, thus increasing Habsburg influence and confirming Austria's vast territorial size. The Emperor also intended Maria Theresa to marry Francis Stephen of Lorraine which would present an unacceptable threat on France's border. By the beginning of 1733 the French army was ready for war: all that was needed was the excuse. In 1733 the Polish King and Elector of Saxony, Augustus the Strong, died. There were two candidates for his successor: first, Stanisław Leszczyński, the father-in-law of Louis XV; second, the Elector of Saxony's son, Augustus, supported by Russia, Austria, and Prussia. The Polish succession had afforded Louis XV's chief minister, Fleury, the opportunity to attack Austria and take Lorraine from Francis Stephen. In order to gain Spanish support France backed the succession of Elisabeth Farnese's sons to further Italian lands. Eugene entered the War of the Polish Succession as President of the Imperial War Council and commander-in-chief of the army, but he was severely handicapped by the quality of his troops and the shortage of funds; now in his seventies, the Prince was also burdened by rapidly declining physical and mental powers. France declared war on Austria on 10 October 1733, but without the funds from the Maritime Powers—who, despite the Vienna treaty, remained neutral throughout the war—Austria could not hire the necessary troops to wage an offensive campaign. "The danger to the monarchy," wrote Eugene to the Emperor in October, "cannot be exaggerated". By the end of the year Franco-Spanish forces had seized Lorraine and Milan; by early 1734 Spanish troops had taken Sicily. Eugene took command on the Rhine in April 1734, but vastly outnumbered he was forced onto the defensive. In June Eugene set out to relieve Philippsburg, yet his former drive and energy was now gone. Accompanying Eugene was a young Frederick the Great, sent by his father to learn the art of war. Frederick gained considerable knowledge from Eugene, recalling in later life his great debt to his Austrian mentor, but the Prussian prince was aghast at Eugene's condition, writing later, "his body was still there but his soul had gone." Eugene conducted another cautious campaign in 1735, once again pursuing a sensible defensive strategy on limited resources; but his short-term memory was by now practically non-existent, and his political influence disappeared completely—Gundaker Starhemberg and Johann Christoph von Bartenstein now dominated the conference in his place. However, fortunately for Charles VI Fleury was determined to limit the scope of the war, and in October 1735 he granted generous peace preliminaries to the Emperor. During the last 20 years of his life Eugene was particularly close to Countess Eleonora Batthyány, daughter of Count Theodor von Strattman. Much about their acquaintance remains speculative (Eugene never mentions her in any of his surviving letters), and there is certainly no suggestion of a sexual relationship, but although they lived apart most foreign diplomats regarded Eleonora as his "official lover". Eugene and Eleonora were constant companions, meeting for dinner, receptions and card games almost every day till his death. But their surviving correspondence does not indicate any real intimacy in the relationship. Eugene's other friends such as the papal nuncio, Passionei, made up for the family he still lacked. For his only surviving nephew, Emmanuel, the son of his brother Louis Thomas, Eugene arranged a marriage with one of the daughters of Prince Liechtenstein, but Emmanuel died of smallpox in 1729. With the death of Emmanuel's son in 1734, no close male relatives remained to succeed the Prince. His closest relative, therefore, was Louis Thomas's unmarried daughter, Princess Maria Anna Victoria of Savoy, whom Eugene had never met and, as he had heard nothing but bad of her, made no effort to do so. Eugene returned to Vienna from the War of the Polish Succession in October 1735, weak and feeble; when Maria Theresa and Francis Stephen married in February 1736 Eugene was too ill to attend. After playing cards at Countess Batthyány's on the evening of 20 April he returned to his bed at the Stadtpalais. When his servants arrived to wake him the next morning, 21 April 1736, they found Prince Eugene dead after choking from phlegm in his throat, presumably after suffering from pneumonia. Eugene's heart was buried with those of others of his family in Turin. His remains were carried in a long procession to St. Stephen's Cathedral, where the body was interred in the "Kreuzkapelle". Countess Batthyány expressed in a letter dated 23 December 1720, that at the "Kreuzkapelle" a solemn requiem would be held annually. She dedicated for this purpose two thousand guilders. Despite being one of the richest and most celebrated men of his age, Eugene never married and the suggestion is that he was predominantly homosexual. History knows little of his life before 1683. In his early boyhood in Paris "he belonged to a small, effeminate set that included such unabashed perverts as the young abbé de Choisy who was invariably dressed as a girl" wrote historian Nicholas Henderson. The Duchess of Orléans, who had known Eugene from those days, would later write to her aunt, Princess Sophia of Hanover, describing Eugene's antics with lackeys and pages. He was "a vulgar whore" along with the Prince of Turenne, and "often played the woman with young people" with the nickname of 'Madame Simone' or 'Madam l'Ancienne'. He preferred a "couple of fine page boys" to any woman, and was refused an ecclesiastical benefice due to his "depravity". Of related interest is a popular soldier's song which parodied an imaginary voyage by Eugene and the Marquis de la Moussaye on the Rhine. A storm breaks and the general fears the worst, but the Marquis consoles him: "Our lives are safe/ For we are sodomites/ Destined to perish only by fire/ We shall land." Riess suggests that a comment made by Johann Matthias von der Schulenburg in 1709, who had served under Eugene that the prince enjoyed "la petite débauche et la p[ine] au-delà de tout," is another testament to sodomy. Eugene's rewards for his victories, his share of booty, his revenues from his abbeys in Savoy, and a steady income from his Imperial offices and governorships, enabled him to contribute to the landscape of Baroque architecture. Eugene spent most of his life in Vienna at his Winter Palace, the Stadtpalais, built by Fischer von Erlach. The palace acted as his official residence and home, but for reasons that remain speculative the Prince's association with Fischer ended before the building was complete, favouring instead Johann Lukas von Hildebrandt as his chief architect. Eugene first employed Hildebrandt to finish the Stadtpalais before commissioning him to prepare plans for a palace (Savoy Castle) on his Danubian island at Ráckeve. Begun in 1701 the single-story building took twenty years to complete; yet, probably because of the Rákóczi revolt, the Prince seems to have visited it only once—after the siege of Belgrade in 1717. Of more importance was the grandiose complex of the two Belvedere palaces in Vienna. The single-storey Lower Belvedere, with its exotic gardens and zoo, was completed in 1716. The Upper Belvedere, completed between 1720 and 1722, is a more substantial building; with sparkling white stucco walls and copper roof it became a wonder of Europe. Eugene and Hildebrandt also converted an existing structure on his Marchfeld estate into a country seat, the Schlosshof, situated between the Rivers Danube and Morava. The building, completed in 1729, was far less elaborate than his other projects but it was strong enough to serve as a fortress in case of need. Eugene spent much of his spare time there in his last years accommodating large hunting parties. In the years following the Peace of Rastatt Eugene became acquainted with a large number of scholarly men. Given his position and responsiveness they were keen to meet him: few could exist without patronage and this was probably the main reason for Gottfried Leibniz's association with him in 1714. Eugene also befriended the French writer Jean-Baptiste Rousseau who, by 1716, was receiving financial support from Eugene. Rousseau stayed on attached to the Prince's household, probably helping in the library, until he left for the Netherlands in 1722. Another acquaintance, Montesquieu, already famous for his "Persian Letters" when he arrived in Vienna in 1728, favourably recalled his time spent at the Prince's table. Nevertheless, Eugene had no literary pretensions of his own, and was not tempted like Maurice de Saxe or Marshal Villars to write his memoirs or books on the art of war. He did, however, become a collector on the grandest scale: his picture galleries were filled with 16th- and 17th-century Italian, Dutch and Flemish art; his library at the Stadtpalais crammed with over 15,000 books, 237 manuscripts as well as a huge collection of prints (of particular interest were books on natural history and geography). "It is hardly believable," wrote Rousseau, "that a man who carries on his shoulders the burden of almost all the affairs of Europe … should find as much time to read as though he had nothing else to do." At Eugene's death his possessions and estates, except those in Hungary which the crown reclaimed, went to his niece, Princess Maria Anna Victoria, who at once decided to sell everything. The artwork was bought by Charles Emmanuel III of Sardinia. Eugene's library, prints and drawings were purchased by the Emperor in 1737 and have since passed into Austrian national collections. Napoleon considered Eugene one of the seven greatest commanders of history. Although later military critics have disagreed with that assessment, Eugene was undoubtedly the greatest Austrian general. He was no military innovator, but he had the ability to make an inadequate system work. He was equally adept as an organizer, strategist, and tactician, believing in the primacy of battle and his ability to seize the opportune moment to launch a successful attack. "The important thing," wrote Maurice de Saxe in his "Reveries", "is to see the opportunity and to know how to use it. Prince Eugene possessed this quality which is the greatest in the art of war and which is the test of the most elevated genius." This fluidity was key to his battlefield successes in Italy and in his wars against the Turks. Nevertheless, in the Low Countries, particularly after the battle of Oudenarde in 1708, Eugene, like his cousin Louis of Baden, tended to play safe and become bogged down in a conservative strategy of sieges and defending supply lines. After the attempt on Toulon in 1707, he also became very wary of combined land/sea operations. To historian Derek McKay, however, the main criticism of him as a general is his legacy—he left no school of officers nor an army able to function without him. Eugene was a disciplinarian—when ordinary soldiers disobeyed orders he was prepared to shoot them himself—but he rejected blind brutality, writing "you should only be harsh when, as often happens, kindness proves useless". On the battlefield Eugene demanded courage in his subordinates, and expected his men to fight where and when he wanted; his criteria for promotion were based primarily on obedience to orders and courage on the battlefield rather than social position. On the whole his men responded because he was willing to push himself as hard as them. However, his position as President of the Imperial War Council proved less successful. Following the long period of peace after the Austro-Turkish War, the idea of creating a separate field army or providing garrison troops with effective training for them to be turned into such an army quickly was never considered by Eugene. By the time of the War of the Polish Succession, therefore, the Austrians were outclassed by a better prepared French force. For this Eugene was largely to blame—in his view (unlike the drilling and manoeuvres carried out by the Prussians which to Eugene seemed irrelevant to real warfare) the time to create actual fighting men was when war came. But although Frederick the Great had been struck by the muddle of the Austrian army and its poor organisation during the Polish Succession war, he later amended his initial harsh judgements. "If I understand anything of my trade," commented Frederick in 1758, "especially in the more difficult aspects, I owe that advantage to Prince Eugene. From him I learnt to hold grand objectives constantly in view, and direct all my resources to those ends." To historian Christopher Duffy it was this awareness of the 'grand strategy' that was Eugene's legacy to Frederick. To his responsibilities Eugene attached his own personal values—physical courage, loyalty to his sovereign, honesty, self-control in all things—and he expected these qualities from his commanders. Eugene's approach was dictatorial, but he was willing to co-operate with someone he regarded as his equal, such as Baden or Marlborough. Yet the contrast to his co-commander of the Spanish Succession war were stark. "Marlborough," wrote Churchill, "was the model husband and father, concerned with building up a home, founding a family, and gathering a fortune to sustain it"; whereas Eugene, the bachelor, was "disdainful of money, content with his bright sword and his lifelong animosities against Louis XIV". The result was an austere figure, inspiring respect and admiration rather than affection. The huge equestrian statue in the centre of Vienna commemorates Eugene's achievements. It is inscribed on one side, 'To the wise counsellor of three Emperors', and on the other, 'To the glorious conqueror of Austria's enemies'. Several ships have been named in Eugene's honour: , an Austrian battleship; an Austrian central battery ship with the same name; , a Royal Navy monitor; , an Italian light cruiser and the (later USS "Prinz Eugen"), a World War II heavy cruiser. The 7th SS Volunteer Mountain Division Prinz Eugen, a German mountain infantry division of the Waffen-SS formed in 1941 from "Volksdeutsche" volunteers and conscripts from the Banat, Independent State of Croatia, Hungary and Romania, was named after him.
https://en.wikipedia.org/wiki?curid=9679
Emanuel Leutze Emanuel Gottlieb Leutze (May 24, 1816July 18, 1868) was a German American history painter best known for his 1851 painting "Washington Crossing the Delaware". He is associated with the Düsseldorf school of painting. Leutze was born in Schwäbisch Gmünd, Württemberg, Germany, and was brought to the United States as a child in 1825. His parents settled first in Fredericksburg, Virginia, and then at Philadelphia. The first development of his artistic talent occurred while he was attending the sickbed of his father, when he attempted drawing to occupy the long hours of waiting. His father died in 1831. At 14, he was painting portraits for $5 apiece. Through such work, he supported himself after the death of his father. In 1834, he received his first instruction in art at the classes of John Rubens Smith, a portrait painter in Philadelphia. He soon became skilled, and promoted a plan for publishing, in Washington, portraits of eminent American statesmen; however, he was met with slight encouragement. In 1840, one of his paintings attracted attention and gave him several orders, which enabled him to attend the Kunstakademie Düsseldorf in his native Germany. Due to his anti-academic attitude, he studied only one year at the academy, in the class of Director Schadow. Leutze was mostly influenced by the painter Karl Friedrich Lessing. In 1842 he went to Munich, studying the works of Cornelius and Kaulbach, and, while there, finished his "Columbus before the Queen". The following year he visited Venice and Rome, making studies from Titian and Michelangelo. His first work, "Columbus before the Council of Salamanca" (1841) was purchased by the Düsseldorf Art Union. A companion picture, "Columbus in Chains", procured him the gold medal of the Brussels Art Exhibition, and was subsequently purchased by the Art Union in New York; it was the basis of the 1893 $2 Columbian Issue stamp. In 1845, after a tour in Italy, he returned to Düsseldorf, marrying Juliane Lottner and making his home there for 14 years. During his years in Düsseldorf, he was a resource for visiting Americans: he found them places to live and work, provided introductions, and gave them emotional and even financial support. For many years, he was the president of the Düsseldorf Artists' Association; in 1848, he was an early promoter of the "Malkasten" art association; and in 1857, he led the call for a gathering of artists which originated the founding of the Allgemeine deutsche Kunstgenossenschaft. A strong supporter of Europe's Revolutions of 1848, Leutze decided to paint an image that would encourage Europe's liberal reformers with the example of the American Revolution. Using American tourists and art students as models and assistants, Leutze finished a first version of "Washington Crossing the Delaware" in 1850. Just after it was completed, the first version was damaged by fire in his studio, subsequently restored, and acquired by the Kunsthalle Bremen. On September 5, 1942, during World War II, it was destroyed in a bombing raid by the Allied forces. The second painting, a replica of the first, only larger, was ordered in 1850 by the Parisian art trader Adolphe Goupil for his New York branch and placed on exhibition on Broadway in October 1851. It is now owned by the Metropolitan Museum of Art in New York. In 1854, Leutze finished his depiction of the Battle of Monmouth, "Washington rallying the troops at Monmouth," commissioned by an important patron, the banker David Leavitt of New York City and Great Barrington, Massachusetts. In 1859, Leutze returned to the United States and opened a studio in New York City. He divided his time between New York City and Washington, D.C. In 1859, he painted a portrait of Chief Justice Roger Brooke Taney, which hangs in the Harvard Law School. In a 1992 opinion, Justice Antonin Scalia described the portrait of Taney, made two years after Taney's infamous decision in "Dred Scott v. Sandford", as showing Taney "in black, sitting in a shadowed red armchair, left hand resting upon a pad of paper in his lap, right hand hanging limply, almost lifelessly, beside the inner arm of the chair. He sits facing the viewer and staring straight out. There seems to be on his face, and in his deep-set eyes, an expression of profound sadness and disillusionment." Leutze also executed other portraits, including one of fellow painter William Morris Hunt. That portrait was owned by Hunt's brother Leavitt Hunt, a New York attorney and sometime Vermont resident, and was shown at an exhibition devoted to William Morris Hunt's work at the Museum of Fine Arts, Boston in 1878. In 1860 Leutze was commissioned by the U.S. Congress to decorate a stairway in the Capitol Building in Washington, DC, for which he painted a large composition, "Westward the Course of Empire Takes Its Way", which is also commonly known as "Westward Ho!". Late in life, he became a member of the National Academy of Design. He was also a member of the Union League Club of New York, which has a number of his paintings. At age 52, he died in Washington, D.C. of heat stroke. He was interred at Glenwood Cemetery. At the time of his death, a painting, "The Emancipation of the Slaves", was in preparation. Leutze's portraits are known for their artistic quality and their patriotic romanticism. "Washington Crossing the Delaware" firmly ranks among the American national iconography. Additional References:
https://en.wikipedia.org/wiki?curid=9683
Earley parser In computer science, the Earley parser is an algorithm for parsing strings that belong to a given context-free language, though (depending on the variant) it may suffer problems with certain nullable grammars. The algorithm, named after its inventor, Jay Earley, is a chart parser that uses dynamic programming; it is mainly used for parsing in computational linguistics. It was first introduced in his dissertation in 1968 (and later appeared in an abbreviated, more legible, form in a journal). Earley parsers are appealing because they can parse all context-free languages, unlike LR parsers and LL parsers, which are more typically used in compilers but which can only handle restricted classes of languages. The Earley parser executes in cubic time in the general case formula_1, where "n" is the length of the parsed string, quadratic time for unambiguous grammars formula_2, and linear time for all deterministic context-free grammars. It performs particularly well when the rules are written left-recursively. The following algorithm describes the Earley recogniser. The recogniser can be easily modified to create a parse tree as it recognises, and in that way can be turned into a parser. In the following descriptions, α, β, and γ represent any string of terminals/nonterminals (including the empty string), X and Y represent single nonterminals, and "a" represents a terminal symbol. Earley's algorithm is a top-down dynamic programming algorithm. In the following, we use Earley's dot notation: given a production X → αβ, the notation X → α • β represents a condition in which α has already been parsed and β is expected. Input position 0 is the position prior to input. Input position "n" is the position after accepting the "n"th token. (Informally, input positions can be thought of as locations at token boundaries.) For every input position, the parser generates a "state set". Each state is a tuple (X → α • β, "i"), consisting of The state set at input position "k" is called S("k"). The parser is seeded with S(0) consisting of only the top-level rule. The parser then repeatedly executes three operations: "prediction", "scanning", and "completion". Duplicate states are not added to the state set, only new ones. These three operations are repeated until no new states can be added to the set. The set is generally implemented as a queue of states to process, with the operation to be performed depending on what kind of state it is. The algorithm accepts if (X → γ •, 0) ends up in S("n"), where (X → γ) is the top level-rule and "n" the input length, otherwise it rejects. Adapted from Speech and Language Processing by Daniel Jurafsky and James H. Martin, DECLARE ARRAY S; function INIT(words) function EARLEY-PARSE(words, grammar) procedure PREDICTOR((A → α•Bβ, j), k, grammar) procedure SCANNER((A → α•aβ, j), k, words) procedure COMPLETER((B → γ•, x), k) Consider the following simple grammar for arithmetic expressions: ::= # the start rule ::= "+" | ::= "*" | ::= "1" | "2" | "3" | "4" With the input: This is the sequence of state sets: The state (P → S •, 0) represents a completed parse. This state also appears in S(3) and S(1), which are complete sentences. Earley's dissertation briefly describes an algorithm for constructing parse trees by adding a set of pointers from each non-terminal in an Earley item back to the items that caused it to be recognized. But Tomita noticed that this does not take into account the relations between symbols, so if we consider the grammar S → SS | b and the string bbb, it only notes that each S can match one or two b's, and thus produces spurious derivations for bb and bbbb as well as the two correct derivations for bbb. Another method is to build the parse forest as you go, augmenting each Earley item with a pointer to a shared packed parse forest (SPPF) node labelled with a triple (s, i, j) where s is a symbol or an LR(0) item (production rule with dot), and i and j give the section of the input string derived by this node. A node's contents are either a pair of child pointers giving a single derivation, or a list of "packed" nodes each containing a pair of pointers and representing one derivation. SPPF nodes are unique (there is only one with a given label), but may contain more than one derivation for ambiguous parses. So even if an operation does not add an Earley item (because it already exists), it may still add a derivation to the item's parse forest. SPPF nodes are never labeled with a completed LR(0) item: instead they are labelled with the symbol that is produced so that all derivations are combined under one node regardless of which alternative production they come from.
https://en.wikipedia.org/wiki?curid=9685
Ethiopian cuisine Ethiopian cuisine () characteristically consists of vegetable and often very spicy meat dishes. This is usually in the form of "wat," a thick stew, served atop "injera", a large sourdough flatbread, which is about in diameter and made out of fermented teff flour. Ethiopians eat most of the time with their right hands, using pieces of "injera" to pick up bites of entrées and side dishes. The Ethiopian Orthodox Tewahedo Church prescribes a number of fasting periods ("tsom", "ṣōm") from any kind of animal products (including dairy products and eggs) on Wednesdays, Fridays, and the entire Lenten season, so Ethiopian cuisine contains many dishes that are vegan. A typical dish consists of injera accompanied by a spicy stew, which frequently includes beef, lamb, vegetables and various types of legumes, such as lentils. The cuisines of the Southern Nations, Nationalities and People's Region and the Sidama people also make use of the false banana plant ("enset", Ge'ez: እንሰት "ïnset"), a type of ensete. The plant is pulverized and fermented to make various foods, including a bread-like food called "qocho" or "kocho" (Ge'ez: ቆጮ "ḳōč̣ō"), which is eaten with kitfo. The root of this plant may be powdered and prepared as a hot drink called "bulla" (Ge'ez: ቡላ "būlā"), which is often given to those who are tired or ill. Another typical Gurage preparation is coffee with butter ("kebbeh"). "Kita" herb bread is also baked. Due in part to the brief Italian occupation, pasta is popular and frequently available throughout Ethiopia, including rural areas. Coffee is also a large part of Ethiopian culture and cuisine. After every meal, a coffee ceremony is enacted and coffee is served. Ethiopian Orthodox Christians, Ethiopian Jews and Ethiopian Muslims avoid eating pork or shellfish, for religious reasons. Pork is considered unclean in Ethiopian Orthodox Christianity, Judaism and Islam. Many Ethiopians abstain from eating certain meats, and mostly eat vegetarian and vegan foods. "Berbere", a combination of powdered chili pepper and other spices (somewhat analogous to Southwestern American chili powder), is an important ingredient used in many dishes. Also essential is "niter kibbeh", a clarified butter infused with ginger, garlic, and several spices. "Mitmita" (, ) is a powdered seasoning mix used in Ethiopian cuisine. It is orange-red in color and contains ground birdseye chili peppers (piri piri), cardamom seed, cloves and salt. It occasionally has other spices including cinnamon, cumin and ginger. In their adherence to strict fasting, Ethiopian cooks have developed a rich array of cooking oil sources—besides sesame and safflower—for use as a substitute for animal fats which are forbidden during fasting periods. Ethiopian cuisine also uses "nug" (also spelled "noog", also known as "niger seed"). "Wat" begins with a large amount of chopped red onion, which is simmered or sauteed in a pot. Once the onions have softened, "niter kebbeh" (or, in the case of vegan dishes, vegetable oil) is added. Following this, "berbere" is added to make a spicy "keiy wat" or "keyyih tsebhi". Turmeric is used instead of "berbere" for a milder "alicha wat" or both spices are omitted when making vegetable stews, such as atkilt wat. Meat such as beef ("ሥጋ", "səga"), chicken ("ዶሮ", "doro" or "derho"), fish ("ዓሣ", "asa"), goat or lamb ("በግ", "beg" or "beggi") is also added. Legumes such as split peas ("ክክ", "kək" or "kikki") and lentils ("ምስር", "məsər" or "birsin"); or vegetables such as potatoes ("ድንች", "Dənəch"), carrots and chard (ቆስጣ) are also used instead in vegan dishes. Each variation is named by appending the main ingredient to the type of wat (e.g. "kek alicha wat"). However, the word "keiy" is usually not necessary, as the spicy variety is assumed when it is omitted (e.g. "doro wat"). The term "atkilt wat" is sometimes used to refer to all vegetable dishes, but a more specific name can also be used (as in "dinich'na caroht wat", which translates to "potatoes and carrots stew"; but notice the word "atkilt" is usually omitted when using the more specific term). Meat along with vegetables are sautéed to make "tibs" (also "tebs", "t'ibs", "tibbs", etc., Ge'ez: ጥብስ "ṭïbs"). Tibs is served in a variety of manners, and can range from hot to mild or contain little to no vegetables. There are many variations of the delicacy, depending on type, size or shape of the cuts of meat used. Beef, mutton, and goat are the most common meats used in the preparation of tibs. The mid-18th century European visitor to Ethiopia describes "tibs" as a portion of grilled meat served "to pay a particular compliment or show especial respect to someone." It may still be seen this way; today the dish is prepared to commemorate special events and holidays. Kinche (Qinch’e) is a very common Ethiopian breakfast, its equivalent of oatmeal. It is incredibly simple, inexpensive, and nutritious. It is made from cracked wheat, Ethiopian oats, barley or a mixture of those. It can be boiled in either milk or water. The flavor of the Kinche comes from the nit'ir qibe, which is a spiced butter. Another distinctively Ethiopian dish is "kitfo" (frequently spelled "ketfo"). It consists of raw (or rare) beef mince marinated in "mitmita" (Ge'ez: ሚጥሚጣ "mīṭmīṭā" a very spicy chili powder similar to the "berbere") and "niter kibbeh". "Gored gored" is very similar to "kitfo", but uses cubed rather than ground beef. "Ayibe" is a cottage cheese that is mild and crumbly. It is much closer in texture to crumbled feta. Although not quite pressed, the whey has been drained and squeezed out. It is often served as a side dish to soften the effect of very spicy food. It has little to no distinct taste of its own. However, when served separately, ayibe is often mixed with a variety of mild or hot spices typical of Gurage cuisine. "Gomen kitfo" is another typical Gurage dish. Collard greens (ጎመን "gōmen") are boiled, dried and then finely chopped and served with butter, chili and spices. It is a dish specially prepared for the occasion of Meskel, a very popular holiday marking the discovery of the True Cross. It is served along with "ayibe" or sometimes even "kitfo" in this tradition called "dengesa". The enset plant (called "wesse" in the Sidamo language) is central to Sidama cuisine and after grinding and fermenting the root to produce "wassa", it is used in the preparation of several foods. "Amulcho" is an enset flatbread used similarly to injera to eat wats made from beef, mushrooms, beans, gomen, and pumpkin. "Borasaame" is a cooked mixture of wassa and butter sometimes eaten with Ethiopian mustard greens. It is traditionally eaten by hand using a false banana leaf and is served in a "shafeta", a vase-like ceramic vessel. A common variant of borasaame uses maize flour instead of wassa and is called "badela borasaame". Borasaame is typically paired with a seasoned yogurt drink called wätät. Both are common foods for funerals and the celebration of "Fichee Chambalaalla", the Sidama new year. Gomen ba siga (ጎመን በስጋ, Amharic: "cabbage with meat") is a stewed mixture of beef and Ethiopian mustard served under a layer of amulcho. A commonly grown crop in Sidama, maize (badela in Sidaamu) is often eaten as a snack with coffee. It can be ground into flour to make bread, roasted on the cob, or the kernels can be picked off to make "bokolo", which is served either boiled or roasted. "Fit-fit" or "fir-fir" is a common breakfast dish. It is made from shredded "injera" or "kitcha" stir-fried with spices or wat. Another popular breakfast food is "fatira". The delicacy consists of a large fried pancake made with flour, often with a layer of egg. It is eaten with honey. "Chechebsa" (or "kita firfir") resembles a pancake covered with "berbere" and "niter kibbeh", or other spices, and may be eaten with a spoon. "Genfo" is a kind of porridge, which is another common breakfast dish. It is usually served in a large bowl with a dug-out made in the middle of the genfo and filled with spiced "niter kibbeh". A variation of "ful", a fava bean stew with condiments, served with baked rolls instead of "injera", is also common for breakfast. Typical Ethiopian snacks are Dabo Kolo (small pieces of baked bread that are similar to pretzels) or "kolo" (roasted barley sometimes mixed with other local grains). "Kolo" made from roasted and spiced barley, safflower kernels, chickpeas and/or peanuts are often sold by kiosks and street vendors, wrapped in a paper cone. Snacking on popcorn is also common. A "gursha" (var. "gorsha", "goorsha") is an act of friendship and love. When eating injera, a person uses his or her right hand to strip off a piece, wraps it around some "wat" or "kitfo", and then puts it into his or her mouth. During a meal with friends or family, it is a common custom to feed others in the group with one's hand by putting the rolled injera or a spoon full of other dishes into another's mouth. This is called a "gursha", and the larger the gursha, the stronger the friendship or bond (only surpassed by the brewing of Tej together). This tradition was featured in "The Food Wife," an episode of "The Simpsons" that uses Ethiopian cuisine as a plot point. There are many different traditional alcoholic drinks which are home made and of natural ingredients. Tella is a home-brewed beer served in "tella bet" ("tella houses") which specialize in serving only tella. Tella is the most common beverage made and served in households during holidays. It is an alcoholic drink which is prepared from bikil (barley) as main ingredient and gesho ("Rhamnus prinoides") for fermentation purpose. In Oromiffaa the drink is called "farso" and in Tigrinya "siwa". Tej is a potent honey wine. It is similar to mead, and is frequently served in bars, particularly in a "tej bet" or "tej house". It is prepared from honey and gesho. It has a sweet taste and the alcoholic content is relatively higher than tella. This drink can be stored for a long time; the longer it is stored, the higher the alcohol content, and the stronger the taste. Areki, also known as katikala, is probably the strongest alcoholic drink of Ethiopia. Ethiopians have diverse traditional non-alcoholic drinks which include natural and healthy ingredients. Kenetto, also known as keribo, is a non-alcoholic traditional drink. It is mostly used as substitute for tella for those who don't drink alcohol. Borde is a cereal-based traditional fermented beverage famous in southern Ethiopia. Just like the rest of the world Ethiopians also enjoy several locally manufactured beers, wine and non alcoholic products like Coca-Cola and other similar products. Ambo Mineral Water or "Ambo wuha" is a bottled carbonated mineral water, sourced from the springs in Ambo Senkele near the town of Ambo. "Atmet" is a barley and oat-flour based drink that is cooked with water, sugar and "kibe" (Ethiopian clarified butter) until the ingredients have combined to create a consistency slightly thicker than egg-nog. Though this drink is often given to women who are nursing, the sweetness and smooth texture make it a comfort drink for anyone who enjoys its flavor. According to some sources, drinking of coffee ("buna") is likely to have originated in Ethiopia. A key national beverage, it is an important part of local commerce. The coffee ceremony is the traditional serving of coffee, usually after a big meal. It often involves the use of a "jebena" (ጀበና), a clay coffee pot in which the coffee is boiled. The preparer roasts the coffee beans in front of guests, then walks around wafting the smoke throughout the room so participants may sample the scent of coffee. Then the preparer grinds the coffee beans in a traditional tool called a "mokecha". The coffee is put into the jebena, boiled with water, and then served in small cups called "si'ni". Coffee is usually served with sugar, but is also served with salt in many parts of Ethiopia. In some parts of the country, "niter kibbeh" is added instead of sugar or salt. Snacks, such as popcorn or toasted barley (or "kolo"), are often served with the coffee. In most homes, a dedicated coffee area is surrounded by fresh grass, with special furniture for the coffee maker. A complete ceremony has three rounds of coffee (abol, tona and bereka) and is accompanied by the burning of frankincense. Tea will most likely be served if coffee is declined.
https://en.wikipedia.org/wiki?curid=9686
Epistle of James The Epistle of James, the Letter of James, or simply James (), is one of the 21 epistles (didactic letters) in the New Testament. The author identifies himself as "James, a servant [or slave] of God and of the Lord Jesus Christ" who is writing to "the twelve tribes scattered abroad". The epistle is traditionally attributed to James the brother of Jesus (James the Just), and the audience is generally considered to be Jewish Christians, who were dispersed outside Palestine. Framing his letter within an overall theme of patient perseverance during trials and temptations, James writes to encourage his readers to live consistently with what they have learned in Christ. He wants his readers to mature in their faith in Christ by living what they say they believe. He condemns various sins, including pride, hypocrisy, favouritism, and slander. He encourages and implores believers to humbly live by godly, rather than worldly wisdom and to pray in all situations. For the most part, until the late 20th century, the epistle of James was relegated to benign disregard – though it was shunned by many early theologians and scholars due to its advocacy of Torah observance and good works. Famously, Luther disliked the epistle due to its lack of Christology and its focus on Torah observance, and sidelined it to an appendix. The epistle is now widely acknowledged by many scholars as a second-century text that emanates from traditions associated with the historical James who was part of an active mission to the Jews, is a strict defender and upholder of the Law, and aims at a wide Jewish audience. During the last decades, the epistle of James has attracted increasing scholarly interest due to a surge in the quest for the historical James, his role within the Jesus movement, his beliefs, and his relationships and views. This James revival is also associated with the growing awareness about the Jewish grounding of the epistle and of the early Jesus movement. According to Baukham, as Christian scholarship and theology have moved to a more comfortable embrace of the Jewish grounding of the early Jesus movement, and of the diversity of early belief in Jesus, interest in the earliest layers of the tradition has been on the rise. The debate about the authorship of James is inconclusive and shadows debates about Christology, and about historical accuracy. Direct authorship by James is now a minority view. The text’s affinity to Torah observance, good deeds, and the traditions that emanated from the historical James are acknowledged by a majority of scholars. The current scholarship regarding the authorship of James: According to Robert J. Foster, "there is little consensus as to the genre, structure, dating, and authorship of the book of James." There are four "commonly espoused" views concerning authorship and dating of the Epistle of James: The writer refers to himself only as "James, a servant of God and of the Lord Jesus Christ" (). Jesus had two apostles named James: James, the son of Zebedee and James, the son of Alphaeus, but it is unlikely that either of these wrote the letter. According to the Book of Acts, James, the brother of John, was killed by Herod Agrippa I (). James, the son of Alphaeus is a more viable candidate for authorship, although he is not prominent in the scriptural record, and relatively little is known about him. Hippolytus, writing in the early third century, asserted in his work "On the 12 Apostles": The similarity of his alleged martyrdom to the stoning of James the Just has led some scholars, such as Robert Eisenman and James Tabor, to assume that these "two Jameses" were one and the same. This identification of James of Alphaeus with James the Just (as well as James the Less) has long been asserted, as evidenced by their conflation in Jacobus de Voragine's medieval hagiography the "Golden Legend". Some have said the authorship of this epistle points to James, the brother of Jesus, to whom Jesus evidently had made a special appearance after his resurrection described in the New Testament as this James was prominent among the disciples. James the brother of Jesus was not a follower of Jesus before Jesus died according to John 7:2-5, which states that during Jesus' life "not even his brothers believed in him". From the middle of the 3rd century, patristic authors cited the epistle as written by James, the brother of Jesus and a leader of the Jerusalem church. Not numbered among the Twelve Apostles unless he is identified as James the Less, James was nonetheless a very important figure: Paul the Apostle described him as "the brother of the Lord" in Galatians 1:19 and as one of the three "pillars of the Church" in 2:9."There is no doubt that James became a much more important person in the early Christian movement than a casual reader of the New Testament is likely to imagine." The James believers are acquainted with emerges out of ; and Acts 12,15,21. We also have accounts about James in Josephus, Eusebius, Origen, the Gospel of Thomas, the Apocalypses of James, the Gospel of the Hebrews and the Pseudo-Clementine literature - most of whom cast him as righteous and as the undisputed leader of the Jewish camp. "His influence is central and palpable in Jerusalem and in Antioch, despite the fact that he did not minister at Antioch. Although we are dependent on sources dominated by the Pauline perspective… the role and influence of James overshadow all others at Antioch." Pseudonymous authorship (3 above) implies that the person named "James" is respected and doubtless well known. Moreover, this James, brother of Jesus, is honored by the epistle written and distributed after the lifetime of James, the brother of Jesus. John Calvin and others suggested that the author was the James, son of Alphaeus, who is referred to as James the Less. The Protestant reformer Martin Luther denied it was the work of an apostle and termed it an "epistle of straw". The Holy Tradition of the Orthodox Church teaches that the Book of James was "written not by either of the apostles, but by the 'brother of the Lord' who was the first bishop of the Church in Jerusalem (see , )." Koester H. (1965) and Kloppenborg J. (1987) are widely recognized for bringing about the pivot from the above (traditional) emphasis on James as wisdom and ethics literature, to focus on the apocalyptic and pre-Gentile (Jewish) context of James. Later studies strengthened this recent appreciation for the pre-Gentile foundations of Q, M, and James. In addition to james, traces of the Jewish followers of Jesus are to be found in the extra-canonical Jewish Gospels (Nazoraeans, Ebionites), in the Didacheand in the Pseudo-Clementine literature,texts not focused on Jesus’ death and resurrection and either advocate, or seem to advocate, Torah observance. Many scholars consider the epistle to be pseudepigrapha: Scholars, such as Luke Timothy Johnson, suggest an early dating for the Epistle of James: The Letter of James also, according to the majority of scholars who have carefully worked through its text in the past two centuries, is among the earliest of New Testament compositions. It contains no reference to the events in Jesus' life, but it bears striking testimony to Jesus' words. Jesus' sayings are embedded in James' exhortations in a form that is clearly not dependent on the written Gospels. If written by James the brother of Jesus, it would have been written sometime before AD 69 (or AD 62), when he was martyred. However, the mention of elders and the role of elders in the church suggests a late first century composition, after the first generation of disciples. The epistle also assumes that Paul's writings were well known, even to the extent of being misused. Other factors, including it's history of use in the church point to a much later date. The earliest extant manuscripts of James usually date to the mid-to-late 3rd century. James is considered New Testament wisdom literature: "like Proverbs and Sirach, it consists largely of moral exhortations and precepts of a traditional and eclectic nature." The content of James is directly parallel, in many instances, to sayings of Jesus found in the gospels of Luke and Matthew, i.e., those attributed to the hypothetical Q source. Compare, e.g., "Do not swear at all, either by heaven...or by the earth...Let your word be 'Yes, Yes' or 'No, No'; anything more than this comes from the evil one" (Matthew 5:34, 37) and "...do not swear either by heaven or by earth or by any other oath, but let your 'Yes' be yes and your 'No' be no, so that you may not fall under condemnation" (James 5:12). According to James Tabor, the epistle of James contains "no fewer than thirty direct references, echoes, and allusions to the teachings of Jesus found in the Q source." Some view the epistle as having no overarching outline: "James may have simply grouped together small 'thematic essays' without having more linear, Greco-Roman structures in mind." That view is generally supported by those who believe that the epistle may not be a true piece of correspondence between specific parties but an example of wisdom literature, formulated as a letter for circulation. The "Catholic Encyclopedia" says, "the subjects treated of in the Epistle are many and various; moreover, St. James not infrequently, whilst elucidating a certain point, passes abruptly to another, and presently resumes once more his former argument." Others view the letter as having only broad topical or thematic structure. They generally organize James under three (Ralph Martin) to seven (Luke Johnson) general key themes or segments. A third group believes that James was more purposeful in structuring his letter, linking each paragraph theologically and thematically: The third view of the structuring of James is a historical approach that is supported by scholars who are not content with leaving the book as "New Testament wisdom literature, like a small book of proverbs" or "like a loose collection of random pearls dropped in no particular order onto a piece of string." A fourth group uses modern discourse analysis or Greco-Roman rhetorical structures to describe the structure of James. The United Bible Societies' "Greek New Testament" divides the letter into the following sections: A 2013 article in the "Evangelical Quarterly" explores a violent historical background behind the epistle and offers the suggestion that it was indeed written by James, the brother of Jesus, and it was written before AD 62, the year he was killed. The 50s saw the growth of turmoil and violence in Roman Judea, as Jews became more and more frustrated with corruption, injustice and poverty. It continued into the 60s, four years before James was killed. War broke out with Rome and would lead to the destruction of Jerusalem and the scattering of the people. The epistle is renowned for exhortations on fighting poverty and caring for the poor in practical ways (1:26–27; 2:1-4; 2:14-19; 5:1-6), standing up for the oppressed (2:1-4; 5:1-6) and not being "like the world" in the way one responds to evil in the world (1:26-27; 2:11; 3:13-18; 4:1-10). Worldly wisdom is rejected and people are exhorted to embrace heavenly wisdom, which includes peacemaking and pursuing righteousness and justice (3:13-18). This approach sees the epistle as a real letter with a real immediate purpose: to encourage Christian Jews not to revert to violence in their response to injustice and poverty but to stay focused on doing good, staying holy and to embrace the wisdom of heaven, not that of the world. The epistle contains the following famous passage concerning salvation and justification: That passage has been cited in Christian theological debates, especially regarding the doctrine of justification. Gaius Marius Victorinus (4th century) associated James's teaching on works with the heretical Symmachian sect, followers of Symmachus the Ebionite, and openly questioned whether James' teachings were heretical. This passage has also been contrasted with the teachings of Paul the Apostle on justification. Some scholars even believe that the passage is a response to Paul. One issue in the debate is the meaning of the Greek word δικαιόω (dikaiόō) 'render righteous or such as he ought to be', with some among the participants taking the view that James is responding to a misunderstanding of Paul. Roman Catholicism and Eastern Orthodoxy have historically argued that the passage disproves the doctrine of justification by faith alone ("sola fide"). The early (and many modern) Protestants resolve the apparent conflict between James and Paul regarding faith and works in alternate ways from the Catholics and Orthodox: According to Ben Witherington III, differences exist between the Apostle Paul and James, but both used the law of Moses, the teachings of Jesus and other Jewish and non-Jewish sources, and "Paul was not anti-law any more than James was a legalist". The epistle is also the chief biblical text for the Anointing of the Sick. James wrote: G. A. Wells suggested that the passage was evidence of late authorship of the epistle, on the grounds that the healing of the sick being done through an official body of presbyters (elders) indicated a considerable development of ecclesiastical organisation "whereas in Paul's day to heal and work miracles pertained to believers indiscriminately (I Corinthians, XII:9)." James and the M material in Matthew are unique in the canon in their stand against the rejection of works and deeds. According to Sanders, traditional Christian theology wrongly divested the term "works" of its ethical grounding, part of the effort to characterize Judaism as legalistic. However, for James and for all Jews, faith is alive only through Torah observance. In other words, belief demonstrates itself through practice and manifestation. For James, claims about belief are empty, unless they are alive in action, works and deeds. The Epistle of James was first explicitly referred to and quoted by Origen of Alexandria, and possibly a bit earlier by Irenaeus of Lyons as well as Clement of Alexandria in a lost work according to Eusebius, although it was not mentioned by Tertullian, who was writing at the end of the Second century. It is also absent from the Muratorian fragment, the earliest known list of New Testament books. The Epistle of James was included among the twenty-seven New Testament books first listed by Athanasius of Alexandria in his "Thirty-Ninth Festal Epistle" (AD 367) and was confirmed as a canonical epistle of the New Testament by a series of councils in the fourth century. In the first centuries of the Church the authenticity of the Epistle was doubted by some, including Theodore of Mopsuestia in the mid-fifth century. Because of the silence of several of the western churches regarding it, Eusebius classes it among the Antilegomena or contested writings ("Historia ecclesiae", 3.25; 2.23). Jerome gives a similar appraisal but adds that with time it had been universally admitted. Gaius Marius Victorinus, in his commentary on the Epistle to the Galatians, openly questioned whether the teachings of James were heretical. Its late recognition in the Church, especially in the West, may be explained by the fact that it was written for or by Jewish Christians, and therefore not widely circulated among the Gentile Churches. There is some indication that a few groups distrusted the book because of its doctrine. In Reformation times a few theologians, most notably Martin Luther in his early ministry, argued that this epistle should not be part of the canonical New Testament. Martin Luther's description of the Epistle of James varies. In some cases, Luther argues that it was not written by an apostle; but in other cases, he describes James as the work of an apostle. He even cites it as authoritative teaching from God and describes James as "a good book, because it sets up no doctrines of men but "vigorously promulgates the law of God"." Lutherans hold that the Epistle is rightly part of the New Testament, citing its authority in the Book of Concord; however, it remains part of the Lutheran antilegomena.
https://en.wikipedia.org/wiki?curid=9688
Eusebius Amort Eusebius Amort (November 15, 1692February 5, 1775) was a German Roman Catholic theologian. Amort was born at Bibermuhle, near Tolz, in Upper Bavaria. He studied at Munich, and at an early age joined the Canons Regular at Polling, where, shortly after his ordination in 1717, he taught theology and philosophy. The Parnassus Boicus learned society was based on a plan started in 1720 by three Augustinian fathers: Eusebius Amort, Gelasius Hieber (1671–1731), a famous preacher in the German language and Agnellus Kandler (1692–1745), a genealogist and librarian. The initial plans fell through, but in 1722 they issued the first volume of the "Parnassus Boicus" journal, communicating interesting information from the arts and sciences. In 1733 Amort went to Rome as theologian to Cardinal Niccolo Maria Lercari (died 1757). He returned to Polling in 1735 and devoted the rest of his life to the revival of learning in Bavaria. He died at Polling in 1775. Amort, who had the reputation of being the most learned man of his age, was a voluminous writer on every conceivable subject, from poetry to astronomy, from dogmatic theology to mysticism. His best known works are: The list of his other works, including his three erudite contributions to the question of authorship of the "Imitatio Christi", will be found in C. Toussaint's scholarly article in Alfred Vacant's "Dictionnaire de theologie" (1900, cols 1115-1117). Citations Sources
https://en.wikipedia.org/wiki?curid=9692
Episcopus vagans The "Oxford Dictionary of the Christian Church" mentions as the main lines of succession deriving from "" in the 20th century those founded by Arnold Mathew, Joseph René Vilatte and Leon Chechemian. Others that could be added are those derived from Aftimios Ofiesh, Carlos Duarte Costa, Paolo Miraglia-Gulotti, Emmanuel Milingo, Pierre Martin Ngô Đình Thục and Marcel Lefebvre. In Western Christianity it has traditionally been taught, since as far back as the time of the Donatist controversy of the fourth and fifth centuries, that any bishop can consecrate any other baptised man as a bishop provided the bishop observes the minimum requirements for the sacramental validity of the ceremony. This means that the consecration is considered valid even if it flouts certain ecclesiastical laws, and even if the participants are schismatics or heretics. According to a theological view affirmed, for instance, by the International Bishops' Conference of the Old Catholic Church with regard to ordinations by Arnold Mathew, an episcopal ordination is for service within a specific Christian church, and an ordination ceremony that concerns only the individual himself does not make him truly a bishop. The Holy See has not commented on the validity of this theory, but has declared with regard to ordinations of this kind carried out, for example, by Emmanuel Milingo, that the Church "does not recognize and does not intend to recognize in the future those ordinations or any of the ordinations derived from them and therefore the canonical state of the alleged bishops remains that in which they were before the ordination conferred by Mr Milingo". Other theologians, notably those of the Eastern Orthodox Church, dispute the notion that such ordinations have effect, a notion that opens up the possibility of valid but irregular consecrations proliferating outside the structures of the "official" denominations. A Catholic ordained to the episcopacy without a mandate from the Pope is automatically excommunicated and is thereby forbidden to celebrate the sacraments, according to canon law. Vlassios Pheidas, on an official Church of Greece site, uses the canonical language of the Orthodox tradition, to describe the conditions in ecclesial praxis when sacraments, including Holy Orders, are real, valid, and efficacious. He notes language is itself part of the ecclesiological problem. This applies to the validity and efficacy of the ordination of bishops and the other sacraments, not only of the Independent Catholic Churches, but also of all other Christian churches, including the Roman Catholic Church, Oriental Orthodoxy and the Assyrian Church of the East. Anglican bishop Colin Buchanan, in the "Historical Dictionary of Anglicanism", says that the Anglican Communion has held an Augustinian view of orders, by which "the validity of Episcopal ordinations (to whichever order) is based solely upon the historic succession in which the ordaining bishop stands, irrespective of their contemporary ecclesial context". He describes the circumstances of Archbishop Matthew Parker's consecration as one of the reasons why this theory is "generally held". Parker was chosen by Queen Elizabeth I of England to be the first Church of England Archbishop of Canterbury after the death of the previous office holder, Cardinal Reginald Pole, the last Roman Catholic Archbishop of Canterbury. Buchanan notes the Roman Catholic Church also focuses on issues of intention and not just breaks in historical succession. He does not explain whether intention has an ecclesiological role, for Anglicans, in conferring or receiving sacraments. According to Buchanan, "the real rise of the problem" happened in the 19th century, in the "wake of the Anglo-Catholic movement", "through mischievous activities of a tiny number of independently acting bishops". They exist worldwide, he writes, "mostly without congregations", and "many in different stages of delusion and fantasy, not least in the Episcopal titles they confer on themselves"; "the distinguishing mark" to "specifically identif[y]" an "episcopus vagans" is "the lack of a true see or the lack of a real church life to oversee". Paul Halsall, on the Internet History Sourcebooks Project, did not list a single church edifice of independent bishops, in a 1996–1998 New York City building architecture survey of religious communities, which maintain bishops claiming apostolic succession and claim cathedral status but noted there "are now literally hundreds of these ", of lesser or greater spiritual probity. They seem to have a tendency to call living room sanctuaries 'cathedrals';" those buildings were not perceived as cultural symbols and did not meet the survey criteria. David V. Barrett wrote, in "A Brief Guide to Secret Religions", that "one hallmark of such bishops is that they often collect as many lineages as they can to strengthen their Episcopal legitimacy—at least in their own eyes" and their groups have more clergy than members. Many "episcopi vagantes" claim succession from the Old Catholic See of Utrecht, or from Eastern Orthodox, Oriental Orthodox, or Eastern Catholic Churches. A few others derive their orders from Roman Catholic bishops who have consecrated their own bishops after disputes with the Holy See. Barrett wrote that leaders "of some esoteric movements, are also priests or bishops in small non-mainstream Christian Churches"; he explains, this type of "independent or autocephalous" group has "little in common with the Church it developed from, the Old Catholic Church, and even less in common with the Roman Catholic Church" but still claims its authority from Apostolic succession. Many, if not most, "episcopi vagantes" are associated with Independent Catholic Churches. They may be very liberal or very conservative. "Episcopi vagantes" may also include some conservative "Continuing Anglicans" who have broken with the Anglican Communion over various issues such as Prayer Book revision, the ordination of women and the ordination of unmarried, non-celibate individuals (including homosexuals). Buchanan writes that based the criteria of having "a true see" or having "a real church life to oversee", the bishops of most forms of the Continuing Anglican movement are not necessarily classified as vagantes, but "are always in danger of becoming such". Arnold Mathew, according to Buchanan, "lapsed into the vagaries of an """ Stephen Edmonds, in the "Oxford Dictionary of National Biography", wrote that in 1910 Mathew's wife separated from him; that same year, he declared himself and his church seceded from the Union of Utrecht. Within a few months, on 2 November 1911, he was excommunicated by the Roman Catholic Church; sued "The Times" for libel based on the words "pseudo-bishop" used to describe him in the newspaper's translation from the Latin text """"; and, lost his case in 1913. Henry R.T. Brandreth wrote, in "Episcopi Vagantes and the Anglican Church", "[o]ne of the most regrettable features of Mathew's episcopate was the founding of the Order of Corporate Reunion (OCR) in 1908. This claimed to be a revival of Frederick George Lee's movement, but was in fact unconnected with it." Brandreth thought it "seems still to exist in a shadowy underground way" in 1947, but disconnected. Colin Holden, in "Ritualist on a Tricycle", places Mathew and his into perspective, he wrote Mathew was an "", lived in a cottage provided for him, and performed his conditional acts, sometimes called according to Holden "bedroom ordinations", in his cottage. Mathew questioned the validity of Anglican ordinations and became involved with the OCR, in 1911 according to Edmonds, and he openly advertised his offer to reordain Anglican clergy who requested it. This angered the Church of England. In 1912, D. J. Scannell O'Neill wrote in "The Fortnightly Review" that London "seems to have more than her due share of bishops" and enumerates what he refers to as "these hireling shepherds". He also announces that one of them, Mathew, revived the OCR and published "The Torch", a monthly review, advocating the reconstruction of Western Christianity and reunion with Eastern Christianity. "The Torch" stated "that the ordinations of the Church of England are not recognized by any church claiming to be Catholic" so the promoters involved Mathew to conditionally ordain group members who are "clergy of the Established Church" and "sign a profession of the Catholic Faith". It stipulated Mathew's services were not a system of simony and given without simoniac expectations. The group sought to enroll "earnest-minded Catholics who sincerely desire to help forward the work of [c]orporate [r]eunion with the Holy See". Nigel Yates, in "Anglican Ritualism in Victorian Britain, 1830-1910", described it as "an even more bizarre scheme to promote a Catholic Uniate Church in Britain" than Lee and Ambrose Lisle March Phillipps de Lisle's Association for the Promotion of the Unity of Christendom. It was editorialized by O'Neill that the "most charitable construction to be placed on this latest move of Mathew is that he is not mentally sound. Being an Irishman, it is strange that he has not sufficient humor to see the absurdity of falling away from the Catholic Church in order to assist others to unite with the Holy See." Edmonds reports that "anything between 4 and 265 was suggested" as to how many took up his offer of reordination. When it declared devoid of canonical effect the consecration ceremony conducted by Archbishop Pierre Martin Ngô Đình Thục for the Carmelite Order of the Holy Face group at midnight of 31 December 1975, the Holy See refrained from pronouncing on its validity. It also made the same statement with regard to later ordinations by those bishops, saying that, "as for those who have already thus unlawfully received ordination or any who may yet accept ordination from these, whatever may be the validity of the orders ("quidquid sit de ordinum validitate"), the Church does not and will not recognise their ordination ("ipsorum ordinationem"), and will consider them, for all legal effects, as still in the state in which they were before, except that the ... penalties remain until they repent". A similar declaration was issued with regard to Archbishop Emmanuel Milingo's conferring of episcopal ordination on four men - all of whom, by virtue of previous Independent Catholic consecrations, claimed already to be bishops - on 24 September 2006: the Holy See, as well as stating that, in accordance with Canon 1382 of the Code of Canon Law, all five men involved incurred automatic (""latae sententiae"") excommunication through their actions, declared that "the Church does not recognise and does not intend in the future to recognise these ordinations or any ordinations derived from them, and she holds that the canonical state of the four alleged bishops is the same as it was prior to the ordination." In contrast, the Holy See has not questioned the validity of the consecrations that the late Archbishop Marcel Lefebvre performed in 1988 for the service of the relatively numerous followers of the Traditionalist Catholic Society of St. Pius X that he had founded, and of the bishops who, under pressure from the Chinese Catholic Patriotic Association, "have been ordained without the Pontifical mandate and who have not asked for, or have not yet obtained, the necessary legitimation", and who consequently, Pope Benedict XVI declared, "are to be considered illegitimate, but validly ordained". Victor LaValle, in the novel "Big Machine" (2009), included three "" as part of his character's childhood involvement with an independent church: Calvin Baker, in his novel "Dominion" (2006), includes an "" as one of his characters: Intertextuality of ' language. Jim Higgins saw, in "More Years for the Locust", "similarities between Marxist obscurantism and an addiction to Christian arcana" and used ' pejoratively as his example of "the ever-growing proliferation of sects, sectlets and insects claiming direct descent from the master" with "fissiparous tendencies". He saw humor in the ludicrous characters and farce in their titles.
https://en.wikipedia.org/wiki?curid=9693
Elizabeth Garrett Anderson Elizabeth Garrett Anderson (9 June 1836 – 17 December 1917) was an English physician and suffragist. She was the first woman to qualify in Britain as a physician and surgeon. She was the co-founder of the first hospital staffed by women, the first dean of a British medical school, the first woman in Britain to be elected to a school board and, as Mayor of Aldeburgh, the first female mayor in Britain. Garrett was born in Whitechapel, London, the second of eleven children of Newson Garrett (1812–1893), from Leiston, Suffolk, and his wife, Louisa (born Dunnell; 1813–1903), from London. The Garrett ancestors had been ironworkers in East Suffolk since the early seventeenth century. Newson was the youngest of three sons and not academically inclined, although he possessed the family's entrepreneurial spirit. When he finished school, the town of Leiston offered little to Newson, so he left for London to make his fortune. There, he fell in love with his brother's sister-in-law, Louisa Dunnell, the daughter of an innkeeper of Suffolk origin. After their wedding, the couple went to live in a pawnbroker's shop at 1 Commercial Road, Whitechapel. The Garretts had their first three children in quick succession: Louie, Elizabeth and their brother (Dunnell Newson) who died at the age of six months. While Louisa grieved the loss of her third child, it was not easy to raise their two daughters in the city of London at that time. When Garrett was three years old, the family moved to 142 Long Acre, where they lived for two years, whilst one more child was born and her father moved up in the world, becoming not only the manager of a larger pawnbroker's shop, but also a silversmith. Garrett's grandfather, owner of the family engineering works, Richard Garrett & Sons, had died in 1837, leaving the business to his eldest son, Garrett's uncle. Despite his lack of capital, Newson was determined to be successful and in 1841, at the age of 29, he moved his family to Suffolk, where he bought a barley and coal merchants business in Snape, constructing Snape Maltings, a fine range of buildings for malting barley. The Garretts lived in a square Georgian house opposite the church in Aldeburgh until 1852. Newson's malting business expanded and more children were born, Edmund (1840), Alice (1842), Agnes (1845), Millicent (1847), who was to become a leader in the constitutional campaign for women's suffrage, Sam (1850), Josephine (1853) and George (1854). By 1850, Newson was a prosperous businessman and was able to build Alde House, a mansion on a hill behind Aldeburgh. A "by-product of the industrial revolution", Garrett grew up in an atmosphere of "triumphant economic pioneering" and the Garrett children were to grow up to become achievers in the professional classes of late-Victorian England. Garrett was encouraged to take an interest in local politics and, contrary to practices at the time, was allowed the freedom to explore the town with its nearby salt-marshes, beach and the small port of Slaughden with its boatbuilders' yards and sailmakers' lofts. There was no school in Aldeburgh so Garrett learned the three Rs from her mother. When she was 10 years old, a governess, Miss Edgeworth, a poor gentlewoman, was employed to educate Garrett and her sister. Mornings were spent in the schoolroom; there were regimented afternoon walks; educating the young ladies continued at mealtimes when Edgeworth ate with the family; at night, the governess slept in a curtained off area in the girls' bedroom. Garrett despised her governess and sought to outwit the teacher in the classroom. When Garrett was 13 and her sister 15, they were sent to a private school, the Boarding School for Ladies in Blackheath, London, which was run by the step aunts of the poet Robert Browning. There, English literature, French, Italian and German as well as deportment, were taught. Later in life, Garrett recalled the stupidity of her teachers there, though her schooling there did help establish a love of reading. Her main complaint about the school was the lack of science and mathematics instruction. Her reading matter included Tennyson, Wordsworth, Milton, Coleridge, Trollope, Thackeray and George Eliot. Elizabeth and Louie were known as "the bathing Garretts", as their father had insisted they be allowed a hot bath once a week. However, they made what were to be lifelong friends there. When they finished in 1851, they were sent on a short tour abroad, ending with a memorable visit to the Great Exhibition in Hyde Park, London. After this formal education, Garrett spent the next nine years tending to domestic duties, but she continued to study Latin and arithmetic in the mornings and also read widely. Her sister Millicent recalled Garrett's weekly lectures, "Talks on Things in General", when her younger siblings would gather while she discussed politics and current affairs from Garibaldi to Macaulay's "History of England". In 1854, when she was eighteen, Garrett and her sister went on a long visit to their school friends, Jane and Anne Crow, in Gateshead where she met Emily Davies, the early feminist and future co-founder of Girton College, Cambridge. Davies was to be a lifelong friend and confidante, always ready to give sound advice during the important decisions of Garrett's career. It may have been in the "English Woman's Journal", first issued in 1858, that Garrett first read of Elizabeth Blackwell, who had become the first female doctor in the United States in 1849. When Blackwell visited London in 1859, Garrett travelled to the capital. By then, her sister Louie was married and living in London. Garrett joined the Society for Promoting the Employment of Women, which organised Blackwell's lectures on "Medicine as a Profession for Ladies" and set up a private meeting between Garrett and the doctor. It is said that during a visit to Alde House around 1860, one evening while sitting by the fireside, Garrett and Davies selected careers for advancing the frontiers of women's rights; Garrett was to open the medical profession to women, Davies the doors to a university education for women, while 13-year-old Millicent was allocated politics and votes for women. At first Newson was opposed to the radical idea of his daughter becoming a physician but came round and agreed to do all in his power, both financially and otherwise, to support Garrett. After an initial unsuccessful visit to leading doctors in Harley Street, Garrett decided to first spend six months as a surgery nurse at Middlesex Hospital, London in August 1860. On proving to be a good nurse, she was allowed to attend an outpatients' clinic, then her first operation. She unsuccessfully attempted to enroll in the hospital's Medical School but was allowed to attend private tuition in Latin, Greek and "materia medica" with the hospital's apothecary, while continuing her work as a nurse. She also employed a tutor to study anatomy and physiology three evenings a week. Eventually she was allowed into the dissecting room and the chemistry lectures. Gradually, Garrett became an unwelcome presence among the male students, who in 1861 presented a memorial to the school against her admittance as a fellow student, despite the support she enjoyed from the administration. She was obliged to leave the Middlesex Hospital but she did so with an honours certificate in chemistry and "materia medica". Garrett then applied to several medical schools, including Oxford, Cambridge, Glasgow, Edinburgh, St Andrews and the Royal College of Surgeons, all of which refused her admittance. A companion to her in this struggle was the lesser known Dr. Sophia Jex-Blake. While both are considered "outstanding" medical figures of the late 19th century, Garrett was able to obtain her credentials by way of a "side door" through a loophole in admissions at the Worshipful Society of Apothecaries. Having privately obtained a certificate in anatomy and physiology, she was admitted in 1862 by the Society of Apothecaries who, as a condition of their charter, could not legally exclude her on account of her sex. She continued her battle to qualify by studying privately with various professors, including some at the University of St Andrews, the Edinburgh Royal Maternity and the London Hospital Medical School. In 1865, she finally took her exam and obtained a licence (LSA) from the Society of Apothecaries to practise medicine, the first woman qualified in Britain to do so openly (previously there was Dr James Barry who was born and raised female but presented as male from the age of 20, and lived his adult life as a man). On the day, three out of seven candidates passed the exam, Garrett with the highest marks. The Society of Apothecaries immediately amended its regulations to prevent other women obtaining a licence meaning that Jex-Blake however could not follow this same path; the new rule disallowed privately educated women to be eligible for examination. It was not until 1876 that the new Medical Act (39 and 40 Vict, Ch. 41) passed which allowed British medical authorities to license all qualified applicants whatever their gender. Though she was now a licentiate of the Society of Apothecaries, as a woman, Garrett could not take up a medical post in any hospital. So in late 1865, Garrett opened her own practice at 20 Upper Berkeley Street, London. At first patients were scarce, but the practice gradually grew. After six months in practice, she wished to open an outpatients dispensary, to enable poor women to obtain medical help from a qualified practitioner of their own gender. In 1865, there was an outbreak of cholera in Britain, affecting both rich and poor, and in their panic, some people forgot any prejudices they had in relation to a female physician. The first death due to cholera occurred in 1866, but by then Garrett had already opened St Mary's Dispensary for Women and Children, at 69 Seymour Place. In the first year, she tended to 3,000 new patients, who made 9,300 outpatient visits to the dispensary. On hearing that the Dean of the faculty of medicine at the University of Sorbonne, Paris was in favour of admitting women as medical students, Garrett studied French so that she could apply for a medical degree, which she obtained in 1870 after some difficulty. The same year she was elected to the first London School Board, an office newly opened to women; Garrett's was the highest vote among all the candidates. Also in that year, she was made one of the visiting physicians of the East London Hospital for Children (later the Queen Elizabeth Hospital for Children), becoming the first woman in Britain to be appointed to a medical post, but she found the duties of these two positions to be incompatible with her principal work in her private practice and the dispensary, as well as her role as a new mother, so she resigned from these posts by 1873. In 1872, the dispensary became the New Hospital for Women and Children, treating women from all over London for gynaecological conditions; the hospital moved to new premises in Marylebone Street in 1874. Around this time, Garrett also entered into discussion with male medical views regarding women. In 1874, Henry Maudsley’s article on Sex and Mind in Education appeared, which argued that education for women caused over-exertion and thus reduced their reproductive capacity, sometimes causing "nervous and even mental disorders". Garrett's counter-argument was that the real danger for women was not education but boredom and that fresh air and exercise were preferable to sitting by the fire with a novel. In the same year, she co-founded London School of Medicine for Women with Sophia Jex-Blake and became a lecturer in what was the only teaching hospital in Britain to offer courses for women. She continued to work there for the rest of her career and was dean of the school from 1883 to 1902. This school was later called the Royal Free Hospital of Medicine, which later became part of what is now the medical school of University College London. In 1873 she gained membership of the British Medical Association (BMA). In 1878 a motion was proposed to exclude women following the election of Garrett Anderson and Frances Hoggan. The motion was opposed by Dr Norman Kerr who maintained the equal rights of members. This was "one of several instances where Garrett, uniquely, was able to enter a hitherto all male medical institution which subsequently moved formally to exclude any women who might seek to follow her." In 1892, women were again admitted to the British Medical Association. In 1897, Garrett Anderson was elected president of the East Anglian branch of the BMA. Garrett Anderson worked steadily at the development of the New Hospital for Women, and (from 1874) at the creation of the London School of Medicine for Women, where she served as its dean. Both institutions were handsomely and suitably housed and equipped. The New Hospital for Women was able to commission a building in the Euston Road; the architect was J. M. Brydon, who took into his employment at this time Anderson's sister Agnes Garrett and her cousin Rhoda Garrett, who contributed to its design. The hospital was for many years worked entirely by medical women. The schools (in Hunter Street, WC1) had over 200 students, most of them preparing for the medical degree of London University (the present-day University College London), which was opened to women in 1877. Garrett Anderson was also active in the women's suffrage movement. In 1866, Garrett Anderson and Davies presented petitions signed by more than 1,500 asking that female heads of household be given the vote. That year, Garrett Anderson joined the first British Women's Suffrage Committee. She was not as active as her sister, Millicent Garrett Fawcett, though Garrett Anderson became a member of the Central Committee of the National Society for Women's Suffrage in 1889. After her husband's death in 1907, she became more active. As mayor of Aldeburgh, she gave speeches for suffrage, before the increasing militant activity in the movement led to her withdrawal in 1911. Her daughter Louisa, also a physician, was more active and more militant, spending time in prison in 1912 for her suffrage activities. Elizabeth Garrett Anderson once remarked that "a doctor leads two lives, the professional and the private, and the boundaries between the two are never traversed". In 1871, she married James George Skelton Anderson (died 1907) of the Orient Steamship Company co-owned by his uncle Arthur Anderson, but she did not give up her medical practice. She had three children, Louisa (1873–1943), Margaret (1874–1875), who died of meningitis, and Alan (1877–1952). Louisa also became a pioneering doctor of medicine and feminist activist. They retired to Aldeburgh in 1902, moving to Alde House in 1903, after the death of Elizabeth's mother. Skelton died of a stroke in 1907. She enjoyed a happy marriage and in later life, devoted time to Alde House, gardening, and travelling with younger members of the extended family. On 9 November 1908, she was elected mayor of Aldeburgh, the first female mayor in England. Her father had been mayor in 1889. She died in 1917 and is buried in the churchyard of St Peter and St Paul's Church, Aldeburgh. The New Hospital for Women was renamed the Elizabeth Garrett Anderson Hospital in 1918 and amalgamated with the Obstetric Hospital in 2001 to form the Elizabeth Garrett Anderson and Obstetric Hospital before relocating to become the University College Hospital Elizabeth Garrett Anderson Wing at UCH. The former Elizabeth Garrett Anderson Hospital buildings are incorporated into the new National Headquarters for the public service trade union UNISON. The Elizabeth Garrett Anderson Gallery, a permanent installation set within the restored hospital building, uses a variety of media to set the story of Garrett Anderson, her hospital, and women's struggle to achieve equality in the field of medicine within the wider framework of 19th and 20th century social history. The critical care centre at Ipswich Hospital was named the Garrett Anderson Centre in her honour, in recognition of her connection to the county of Suffolk. Elizabeth Garrett Anderson School, a secondary school for girls in Islington, London, is named after her. The archives of Elizabeth Garrett Anderson are held at the Women's Library at the London School of Economics. The archives of the Elizabeth Garrett Anderson Hospital (formerly the New Hospital for Women) are held at the London Metropolitan Archives. On 9 June 2016, Google Doodle commemorated her 180th birthday. The Elizabeth Garrett Anderson programme of the NHS Leadership Academy is a master's degree in leadership and management.
https://en.wikipedia.org/wiki?curid=9695
Erosion In earth science, erosion is the action of surface processes (such as water flow or wind) that removes soil, rock, or dissolved material from one location on the Earth's crust, and then transports it to another location (not to be confused with weathering which involves no movement). This natural process is caused by the dynamic activity of erosive agents, that is, water, ice (glaciers), snow, air (wind), plants, animals, and humans. In accordance with these agents, erosion is sometimes divided into water erosion, glacial erosion, snow erosion, wind (aeolic) erosion, zoogenic erosion, and anthropogenic erosion. The particulate breakdown of rock or soil into clastic sediment is referred to as "physical" or "mechanical" erosion; this contrasts with "chemical" erosion, where soil or rock material is removed from an area by its dissolving into a solvent (typically water), followed by the flow away of that solution. Eroded sediment or solutes may be transported just a few millimetres, or for thousands of kilometres. Natural rates of erosion are controlled by the action of geological weathering geomorphic drivers, such as rainfall; bedrock wear in rivers; coastal erosion by the sea and waves; glacial plucking, abrasion, and scour; areal flooding; wind abrasion; groundwater processes; and mass movement processes in steep landscapes like landslides and debris flows. The rates at which such processes act control how fast a surface is eroded. Typically, physical erosion proceeds fastest on steeply sloping surfaces, and rates may also be sensitive to some climatically-controlled properties including amounts of water supplied (e.g., by rain), storminess, wind speed, wave fetch, or atmospheric temperature (especially for some ice-related processes). Feedbacks are also possible between rates of erosion and the amount of eroded material that is already carried by, for example, a river or glacier. Processes of erosion that produce sediment or solutes from a place contrast with those of deposition, which control the arrival and emplacement of material at a new location. While erosion is a natural process, human activities have increased by 10-40 times the rate at which erosion is occurring globally. At well-known agriculture sites such as the Appalachian Mountains, intensive farming practices have caused erosion up to 100x the speed of the natural rate of erosion in the region. Excessive (or accelerated) erosion causes both "on-site" and "off-site" problems. On-site impacts include decreases in agricultural productivity and (on natural landscapes) ecological collapse, both because of loss of the nutrient-rich upper soil layers. In some cases, the eventual end result is desertification. Off-site effects include sedimentation of waterways and eutrophication of water bodies, as well as sediment-related damage to roads and houses. Water and wind erosion are the two primary causes of land degradation; combined, they are responsible for about 84% of the global extent of degraded land, making excessive erosion one of the most significant environmental problems worldwide. Intensive agriculture, deforestation, roads, anthropogenic climate change and urban sprawl are amongst the most significant human activities in regard to their effect on stimulating erosion. However, there are many prevention and remediation practices that can curtail or limit erosion of vulnerable soils. Rainfall, and the surface runoff which may result from rainfall, produces four main types of soil erosion: "splash erosion", "sheet erosion", "rill erosion", and "gully erosion". Splash erosion is generally seen as the first and least severe stage in the soil erosion process, which is followed by sheet erosion, then rill erosion and finally gully erosion (the most severe of the four). In "splash erosion", the impact of a falling raindrop creates a small crater in the soil, ejecting soil particles. The distance these soil particles travel can be as much as 0.6 m (two feet) vertically and 1.5 m (five feet) horizontally on level ground. If the soil is saturated, or if the rainfall rate is greater than the rate at which water can infiltrate into the soil, surface runoff occurs. If the runoff has sufficient flow energy, it will transport loosened soil particles (sediment) down the slope. "Sheet erosion" is the transport of loosened soil particles by overland flow. "Rill erosion" refers to the development of small, ephemeral concentrated flow paths which function as both sediment source and sediment delivery systems for erosion on hillslopes. Generally, where water erosion rates on disturbed upland areas are greatest, rills are active. Flow depths in rills are typically of the order of a few centimetres (about an inch) or less and along-channel slopes may be quite steep. This means that rills exhibit hydraulic physics very different from water flowing through the deeper, wider channels of streams and rivers. "Gully erosion" occurs when runoff water accumulates and rapidly flows in narrow channels during or immediately after heavy rains or melting snow, removing soil to a considerable depth. "Valley" or "stream erosion" occurs with continued water flow along a linear feature. The erosion is both downward, deepening the valley, and headward, extending the valley into the hillside, creating head cuts and steep banks. In the earliest stage of stream erosion, the erosive activity is dominantly vertical, the valleys have a typical V cross-section and the stream gradient is relatively steep. When some base level is reached, the erosive activity switches to lateral erosion, which widens the valley floor and creates a narrow floodplain. The stream gradient becomes nearly flat, and lateral deposition of sediments becomes important as the stream meanders across the valley floor. In all stages of stream erosion, by far the most erosion occurs during times of flood when more and faster-moving water is available to carry a larger sediment load. In such processes, it is not the water alone that erodes: suspended abrasive particles, pebbles, and boulders can also act erosively as they traverse a surface, in a process known as "traction". "Bank erosion" is the wearing away of the banks of a stream or river. This is distinguished from changes on the bed of the watercourse, which is referred to as "scour". Erosion and changes in the form of river banks may be measured by inserting metal rods into the bank and marking the position of the bank surface along the rods at different times. "Thermal erosion" is the result of melting and weakening permafrost due to moving water. It can occur both along rivers and at the coast. Rapid river channel migration observed in the Lena River of Siberia is due to thermal erosion, as these portions of the banks are composed of permafrost-cemented non-cohesive materials. Much of this erosion occurs as the weakened banks fail in large slumps. Thermal erosion also affects the Arctic coast, where wave action and near-shore temperatures combine to undercut permafrost bluffs along the shoreline and cause them to fail. Annual erosion rates along a segment of the Beaufort Sea shoreline averaged per year from 1955 to 2002. Most river erosion happens nearer to the mouth of a river. On a river bend, the longest least sharp side has slower moving water. Here deposits build up. On the narrowest sharpest side of the bend, there is faster moving water so this side tends to erode away mostly. Shoreline erosion, which occurs on both exposed and sheltered coasts, primarily occurs through the action of currents and waves but sea level (tidal) change can also play a role. "Hydraulic action" takes place when the air in a joint is suddenly compressed by a wave closing the entrance of the joint. This then cracks it. "Wave pounding" is when the sheer energy of the wave hitting the cliff or rock breaks pieces off. "Abrasion" or "corrasion" is caused by waves launching sea load at the cliff. It is the most effective and rapid form of shoreline erosion (not to be confused with "corrosion"). "Corrosion" is the dissolving of rock by carbonic acid in sea water. Limestone cliffs are particularly vulnerable to this kind of erosion. "Attrition" is where particles/sea load carried by the waves are worn down as they hit each other and the cliffs. This then makes the material easier to wash away. The material ends up as shingle and sand. Another significant source of erosion, particularly on carbonate coastlines, is boring, scraping and grinding of organisms, a process termed "bioerosion". Sediment is transported along the coast in the direction of the prevailing current (longshore drift). When the upcurrent amount of sediment is less than the amount being carried away, erosion occurs. When the upcurrent amount of sediment is greater, sand or gravel banks will tend to form as a result of deposition. These banks may slowly migrate along the coast in the direction of the longshore drift, alternately protecting and exposing parts of the coastline. Where there is a bend in the coastline, quite often a buildup of eroded material occurs forming a long narrow bank (a spit). Armoured beaches and submerged offshore sandbanks may also protect parts of a coastline from erosion. Over the years, as the shoals gradually shift, the erosion may be redirected to attack different parts of the shore. Chemical erosion is the loss of matter in a landscape in the form of solutes. Chemical erosion is usually calculated from the solutes found in streams. Anders Rapp pioneered the study of chemical erosion in his work about Kärkevagge published in 1960. Glaciers erode predominantly by three different processes: abrasion/scouring, plucking, and ice thrusting. In an abrasion process, debris in the basal ice scrapes along the bed, polishing and gouging the underlying rocks, similar to sandpaper on wood. Scientists have shown that, in addition to the role of temperature played in valley-deepening, other glaciological processes, such as erosion also control cross-valley variations. In a homogeneous bedrock erosion pattern, curved channel cross-section beneath the ice is created. Though the glacier continues to incise vertically, the shape of the channel beneath the ice eventually remain constant, reaching a U-shaped parabolic steady-state shape as we now see in glaciated valleys. Scientists also provide a numerical estimate of the time required for the ultimate formation of a steady-shaped U-shaped valley—approximately 100,000 years. In a weak bedrock (containing material more erodible than the surrounding rocks) erosion pattern, on the contrary, the amount of over deepening is limited because ice velocities and erosion rates are reduced. Glaciers can also cause pieces of bedrock to crack off in the process of plucking. In ice thrusting, the glacier freezes to its bed, then as it surges forward, it moves large sheets of frozen sediment at the base along with the glacier. This method produced some of the many thousands of lake basins that dot the edge of the Canadian Shield. Differences in the height of mountain ranges are not only being the result tectonic forces, such as rock uplift, but also local climate variations. Scientists use global analysis of topography to show that glacial erosion controls the maximum height of mountains, as the relief between mountain peaks and the snow line are generally confined to altitudes less than 1500 m. The erosion caused by glaciers worldwide erodes mountains so effectively that the term "glacial buzzsaw" has become widely used, which describes the limiting effect of glaciers on the height of mountain ranges. As mountains grow higher, they generally allow for more glacial activity (especially in the accumulation zone above the glacial equilibrium line altitude), which causes increased rates of erosion of the mountain, decreasing mass faster than isostatic rebound can add to the mountain. This provides a good example of a negative feedback loop. Ongoing research is showing that while glaciers tend to decrease mountain size, in some areas, glaciers can actually reduce the rate of erosion, acting as a "glacial armor". Ice can not only erode mountains but also protect them from erosion. Depending on glacier regime, even steep alpine lands can be preserved through time with the help of ice. Scientists have proved this theory by sampling eight summits of northwestern Svalbard using Be10 and Al26, showing that northwestern Svalbard transformed from a glacier-erosion state under relatively mild glacial maxima temperature, to a glacier-armor state occupied by cold-based, protective ice during much colder glacial maxima temperatures as the Quaternary ice age progressed. These processes, combined with erosion and transport by the water network beneath the glacier, leave behind glacial landforms such as moraines, drumlins, ground moraine (till), kames, kame deltas, moulins, and glacial erratics in their wake, typically at the terminus or during glacier retreat. The best-developed glacial valley morphology appears to be restricted to landscapes with low rock uplift rates (less than or equal to 2 mm per year) and high relief, leading to long-turnover times. Where rock uplift rates exceed 2 mm per year, glacial valley morphology has generally been significantly modified in postglacial time. Interplay of glacial erosion and tectonic forcing governs the morphologic impact of glaciations on active orogens, by both influencing their height, and by altering the patterns of erosion during subsequent glacial periods via a link between rock uplift and valley cross-sectional shape. At extremely high flows, kolks, or vortices are formed by large volumes of rapidly rushing water. Kolks cause extreme local erosion, plucking bedrock and creating pothole-type geographical features called rock-cut basins. Examples can be seen in the flood regions result from glacial Lake Missoula, which created the channeled scablands in the Columbia Basin region of eastern Washington. Wind erosion is a major geomorphological force, especially in arid and semi-arid regions. It is also a major source of land degradation, evaporation, desertification, harmful airborne dust, and crop damage—especially after being increased far above natural rates by human activities such as deforestation, urbanization, and agriculture. Wind erosion is of two primary varieties: "deflation", where the wind picks up and carries away loose particles; and "abrasion", where surfaces are worn down as they are struck by airborne particles carried by wind. Deflation is divided into three categories: (1) "surface creep", where larger, heavier particles slide or roll along the ground; (2) "saltation", where particles are lifted a short height into the air, and bounce and saltate across the surface of the soil; and (3) "suspension", where very small and light particles are lifted into the air by the wind, and are often carried for long distances. Saltation is responsible for the majority (50-70%) of wind erosion, followed by suspension (30-40%), and then surface creep (5-25%). Wind erosion is much more severe in arid areas and during times of drought. For example, in the Great Plains, it is estimated that soil loss due to wind erosion can be as much as 6100 times greater in drought years than in wet years. "Mass movement" is the downward and outward movement of rock and sediments on a sloped surface, mainly due to the force of gravity. Mass movement is an important part of the erosional process and is often the first stage in the breakdown and transport of weathered materials in mountainous areas. It moves material from higher elevations to lower elevations where other eroding agents such as streams and glaciers can then pick up the material and move it to even lower elevations. Mass-movement processes are always occurring continuously on all slopes; some mass-movement processes act very slowly; others occur very suddenly, often with disastrous results. Any perceptible down-slope movement of rock or sediment is often referred to in general terms as a landslide. However, landslides can be classified in a much more detailed way that reflects the mechanisms responsible for the movement and the velocity at which the movement occurs. One of the visible topographical manifestations of a very slow form of such activity is a scree slope. "Slumping" happens on steep hillsides, occurring along distinct fracture zones, often within materials like clay that, once released, may move quite rapidly downhill. They will often show a spoon-shaped isostatic depression, in which the material has begun to slide downhill. In some cases, the slump is caused by water beneath the slope weakening it. In many cases it is simply the result of poor engineering along highways where it is a regular occurrence. "Surface creep" is the slow movement of soil and rock debris by gravity which is usually not perceptible except through extended observation. However, the term can also describe the rolling of dislodged soil particles in diameter by wind along the soil surface. The amount and intensity of precipitation is the main climatic factor governing soil erosion by water. The relationship is particularly strong if heavy rainfall occurs at times when, or in locations where, the soil's surface is not well protected by vegetation. This might be during periods when agricultural activities leave the soil bare, or in semi-arid regions where vegetation is naturally sparse. Wind erosion requires strong winds, particularly during times of drought when vegetation is sparse and soil is dry (and so is more erodible). Other climatic factors such as average temperature and temperature range may also affect erosion, via their effects on vegetation and soil properties. In general, given similar vegetation and ecosystems, areas with more precipitation (especially high-intensity rainfall), more wind, or more storms are expected to have more erosion. In some areas of the world (e.g. the mid-western USA), rainfall intensity is the primary determinant of erosivity (for a definition of "erosivity" check,) with higher intensity rainfall generally resulting in more soil erosion by water. The size and velocity of rain drops is also an important factor. Larger and higher-velocity rain drops have greater kinetic energy, and thus their impact will displace soil particles by larger distances than smaller, slower-moving rain drops. In other regions of the world (e.g. western Europe), runoff and erosion result from relatively low intensities of stratiform rainfall falling onto the previously saturated soil. In such situations, rainfall amount rather than intensity is the main factor determining the severity of soil erosion by water. In Taiwan, where typhoon frequency increased significantly in the 21st century, a strong link has been drawn between the increase in storm frequency with an increase in sediment load in rivers and reservoirs, highlighting the impacts climate change can have on erosion. Vegetation acts as an interface between the atmosphere and the soil. It increases the permeability of the soil to rainwater, thus decreasing runoff. It shelters the soil from winds, which results in decreased wind erosion, as well as advantageous changes in microclimate. The roots of the plants bind the soil together, and interweave with other roots, forming a more solid mass that is less susceptible to both water and wind erosion. The removal of vegetation increases the rate of surface erosion. The topography of the land determines the velocity at which surface runoff will flow, which in turn determines the erosivity of the runoff. Longer, steeper slopes (especially those without adequate vegetative cover) are more susceptible to very high rates of erosion during heavy rains than shorter, less steep slopes. Steeper terrain is also more prone to mudslides, landslides, and other forms of gravitational erosion processes. Tectonic processes control rates and distributions of erosion at the Earth's surface. If the tectonic action causes part of the Earth's surface (e.g., a mountain range) to be raised or lowered relative to surrounding areas, this must necessarily change the gradient of the land surface. Because erosion rates are almost always sensitive to the local slope (see above), this will change the rates of erosion in the uplifted area. Active tectonics also brings fresh, unweathered rock towards the surface, where it is exposed to the action of erosion. However, erosion can also affect tectonic processes. The removal by erosion of large amounts of rock from a particular region, and its deposition elsewhere, can result in a lightening of the load on the lower crust and mantle. Because tectonic processes are driven by gradients in the stress field developed in the crust, this unloading can in turn cause tectonic or isostatic uplift in the region. In some cases, it has been hypothesised that these twin feedbacks can act to localize and enhance zones of very rapid exhumation of deep crustal rocks beneath places on the Earth's surface with extremely high erosion rates, for example, beneath the extremely steep terrain of Nanga Parbat in the western Himalayas. Such a place has been called a "tectonic aneurysm". Human land development, in forms including agricultural and urban development, is considered a significant factor in erosion and sediment transport. In Taiwan, increases in sediment load in the northern, central, and southern regions of the island can be tracked with the timeline of development for each region throughout the 20th century. Mountain ranges are known to take many millions of years to erode to the degree they effectively cease to exist. Scholars Pitman and Golovchenko estimate that it takes probably more than 450 million years to erode a mountain mass similar to the Himalaya into an almost-flat peneplain if there are no major sea-level changes. Erosion of mountains massifs can create a pattern of equally high summits called summit accordance. It has been argued that extension during post-orogenic collapse is a more effective mechanism of lowering the height of orogenic mountains than erosion. Examples of heavily eroded mountain ranges include the Timanides of Northern Russia. Erosion of this orogen has produced sediments that are now found in the East European Platform, including the Cambrian Sablya Formation near Lake Ladoga. Studies of these sediments indicate that it is likely that the erosion of the orogen began in the Cambrian and then intensified in the Ordovician. If the rate of erosion is higher than the rate of soil formation the soils are being destroyed by erosion. Where soil is not destroyed by erosion, erosion can in some cases prevent the formation of soil features that form slowly. Inceptisols are common soils that form in areas of fast erosion. While erosion of soils is a natural process, human activities have increased by 10-40 times the rate at which erosion is occurring globally. Excessive (or accelerated) erosion causes both "on-site" and "off-site" problems. On-site impacts include decreases in agricultural productivity and (on natural landscapes) ecological collapse, both because of loss of the nutrient-rich upper soil layers. In some cases, the eventual end result is desertification. Off-site effects include sedimentation of waterways and eutrophication of water bodies, as well as sediment-related damage to roads and houses. Water and wind erosion are the two primary causes of land degradation; combined, they are responsible for about 84% of the global extent of degraded land, making excessive erosion one of the most significant environmental problems.
https://en.wikipedia.org/wiki?curid=9696
Cnidaria Cnidaria () is a phylum under kingdom Animalia containing over 11,000 species of aquatic animals found both in freshwater and marine environments: they are predominantly marine. Their distinguishing feature is cnidocytes, specialized cells that they use mainly for capturing prey. Their bodies consist of mesoglea, a non-living jelly-like substance, sandwiched between two layers of epithelium that are mostly one cell thick. They have two basic body forms: swimming medusae and sessile polyps, both of which are radially symmetrical with mouths surrounded by tentacles that bear cnidocytes. Both forms have a single orifice and body cavity that are used for digestion and respiration. Many cnidarian species produce colonies that are single organisms composed of medusa-like or polyp-like zooids, or both (hence they are trimorphic). Cnidarians' activities are coordinated by a decentralized nerve net and simple receptors. Several free-swimming species of Cubozoa and Scyphozoa possess balance-sensing statocysts, and some have simple eyes. Not all cnidarians reproduce sexually, with many species having complex life cycles of asexual polyp stages and sexual medusae. Some, however, omit either the polyp or the medusa stage. Cnidarians were formerly grouped with ctenophores in the phylum Coelenterata, but increasing awareness of their differences caused them to be placed in separate phyla. Cnidarians are classified into four main groups: the almost wholly sessile Anthozoa (sea anemones, corals, sea pens); swimming Scyphozoa (jellyfish); Cubozoa (box jellies); and Hydrozoa (a diverse group that includes all the freshwater cnidarians as well as many marine forms, and has both sessile members, such as "Hydra", and colonial swimmers, such as the Portuguese Man o' War). Staurozoa have recently been recognised as a class in their own right rather than a sub-group of Scyphozoa, and the parasitic Myxozoa and Polypodiozoa were firmly recognized as cnidarians in 2007. Most cnidarians prey on organisms ranging in size from plankton to animals several times larger than themselves, but many obtain much of their nutrition from dinoflagellates, and a few are parasites. Many are preyed on by other animals including starfish, sea slugs, fish, turtles, and even other cnidarians. Many scleractinian corals—which form the structural foundation for coral reefs—possess polyps that are filled with symbiotic photo-synthetic zooxanthellae. While reef-forming corals are almost entirely restricted to warm and shallow marine waters, other cnidarians can be found at great depths, in polar regions, and in freshwater. Recent phylogenetic analyses support monophyly of cnidarians, as well as the position of cnidarians as the sister group of bilaterians. Fossil cnidarians have been found in rocks formed about , and other fossils show that corals may have been present shortly before and diversified a few million years later. However, molecular clock analysis of mitochondrial genes suggests a much older age for the crown group of cnidarians, estimated around , almost 200 million years before the Cambrian period as well as any fossils. Cnidarians form a phylum of animal that are more complex than sponges, about as complex as ctenophores (comb jellies), and less complex than bilaterians, which include almost all other animals. Both cnidarians and ctenophores are more complex than sponges as they have: cells bound by inter-cell connections and carpet-like basement membranes; muscles; nervous systems; and some have sensory organs. Cnidarians are distinguished from all other animals by having cnidocytes that fire harpoon like structures and are usually used mainly to capture prey. In some species, cnidocytes can also be used as anchors.Cnidarians are also distinguished by the fact that they have only one opening in their body for ingestion and excretion i.e. they don't have a separate mouth and anus. Like sponges and ctenophores, cnidarians have two main layers of cells that sandwich a middle layer of jelly-like material, which is called the mesoglea in cnidarians; more complex animals have three main cell layers and no intermediate jelly-like layer. Hence, cnidarians and ctenophores have traditionally been labelled diploblastic, along with sponges. However, both cnidarians and ctenophores have a type of muscle that, in more complex animals, arises from the middle cell layer. As a result, some recent text books classify ctenophores as triploblastic, and it has been suggested that cnidarians evolved from triploblastic ancestors. Most adult cnidarians appear as either free-swimming medusae or sessile polyps, and many hydrozoans species are known to alternate between the two forms. Both are radially symmetrical, like a wheel and a tube respectively. Since these animals have no heads, their ends are described as "oral" (nearest the mouth) and "aboral" (furthest from the mouth). Most have fringes of tentacles equipped with cnidocytes around their edges, and medusae generally have an inner ring of tentacles around the mouth. Some hydroids may consist of colonies of zooids that serve different purposes, such as defense, reproduction and catching prey. The mesoglea of polyps is usually thin and often soft, but that of medusae is usually thick and springy, so that it returns to its original shape after muscles around the edge have contracted to squeeze water out, enabling medusae to swim by a sort of jet propulsion. In medusae the only supporting structure is the mesoglea. "Hydra" and most sea anemones close their mouths when they are not feeding, and the water in the digestive cavity then acts as a hydrostatic skeleton, rather like a water-filled balloon. Other polyps such as "Tubularia" use columns of water-filled cells for support. Sea pens stiffen the mesoglea with calcium carbonate spicules and tough fibrous proteins, rather like sponges. In some colonial polyps, a chitinous periderm gives support and some protection to the connecting sections and to the lower parts of individual polyps. Stony corals secrete massive calcium carbonate exoskeletons. A few polyps collect materials such as sand grains and shell fragments, which they attach to their outsides. Some colonial sea anemones stiffen the mesoglea with sediment particles. Cnidaria are diploblastic animals; in other words, they have two main cell layers, while more complex animals are triploblasts having three main layers. The two main cell layers of cnidarians form epithelia that are mostly one cell thick, and are attached to a fibrous basement membrane, which they secrete. They also secrete the jelly-like mesoglea that separates the layers. The layer that faces outwards, known as the ectoderm ("outside skin"), generally contains the following types of cells: In addition to epitheliomuscular, nerve and interstitial cells, the inward-facing gastroderm ("stomach skin") contains gland cells that secrete digestive enzymes. In some species it also contains low concentrations of cnidocytes, which are used to subdue prey that is still struggling. The mesoglea contains small numbers of amoeba-like cells, and muscle cells in some species. However, the number of middle-layer cells and types are much lower than in sponges. Polymorphism refers to the occurrence of structurally and functionally more than two different types of individuals within the same organism. It is a characteristic feature of Cnidarians, particularly the polyp and medusa forms, or of zooids within colonial organisms like those in Hydrozoa. In Hydrozoans, colonial individuals arising from individuals zooids will take on separate tasks. For example, in "Obelia" there are feeding individuals, the gastrozooids; the individuals capable of asexual reproduction only, the gonozooids, blastostyles and free-living or sexually reproducing individuals, the medusae. These "nettle cells" function as harpoons, since their s remain connected to the bodies of the cells by threads. Three types of cnidocytes are known: The main components of a cnidocyte are: It is difficult to study the firing mechanisms of cnidocytes as these structures are small but very complex. At least four hypotheses have been proposed: Cnidocytes can only fire once, and about 25% of a hydra's nematocysts are lost from its tentacles when capturing a brine shrimp. Used cnidocytes have to be replaced, which takes about 48 hours. To minimise wasteful firing, two types of stimulus are generally required to trigger cnidocytes: nearby sensory cells detect chemicals in the water, and their cilia respond to contact. This combination prevents them from firing at distant or non-living objects. Groups of cnidocytes are usually connected by nerves and, if one fires, the rest of the group requires a weaker minimum stimulus than the cells that fire first. Medusae swim by a form of jet propulsion: muscles, especially inside the rim of the bell, squeeze water out of the cavity inside the bell, and the springiness of the mesoglea powers the recovery stroke. Since the tissue layers are very thin, they provide too little power to swim against currents and just enough to control movement within currents. Hydras and some sea anemones can move slowly over rocks and sea or stream beds by various means: creeping like snails, crawling like inchworms, or by somersaulting. A few can swim clumsily by waggling their bases. Cnidarians are generally thought to have no brains or even central nervous systems. However, they do have integrative areas of neural tissue that could be considered some form of centralization. Most of their bodies are innervated by decentralized nerve nets that control their swimming musculature and connect with sensory structures, though each clade has slightly different structures. These sensory structures, usually called rhopalia, can generate signals in response to various types of stimuli such as light, pressure, and much more. Medusa usually have several of them around the margin of the bell that work together to control the motor nerve net, that directly innervates the swimming muscles. Most Cnidarians also have a parallel system. In scyphozoans, this takes the form of a diffuse nerve net, which has modulatory effects on the nervous system. As well as forming the "signal cables" between sensory neurons and motoneurons, intermediate neurons in the nerve net can also form ganglia that act as local coordination centers. Communication between nerve cells can occur by chemical synapses or gap junctions in hydrozoans, though gap junctions are not present in all groups. Cnidarians have many of the same neurotransmitters as many animals, including chemicals such as glutamate, GABA, and acetylcholine. This structure ensures that the musculature is excited rapidly and simultaneously, and can be directly stimulated from any point on the body, and it also is better able to recover after injury. Medusae and complex swimming colonies such as siphonophores and chondrophores sense tilt and acceleration by means of statocysts, chambers lined with hairs which detect the movements of internal mineral grains called statoliths. If the body tilts in the wrong direction, the animal rights itself by increasing the strength of the swimming movements on the side that is too low. Most species have ocelli ("simple eyes"), which can detect sources of light. However, the agile box jellyfish are unique among Medusae because they possess four kinds of true eyes that have retinas, corneas and lenses. Although the eyes probably do not form images, Cubozoa can clearly distinguish the direction from which light is coming as well as negotiate around solid-colored objects. Cnidarians feed in several ways: predation, absorbing dissolved organic chemicals, filtering food particles out of the water, obtaining nutrients from symbiotic algae within their cells, and parasitism. Most obtain the majority of their food from predation but some, including the corals "Hetroxenia" and "Leptogorgia", depend almost completely on their endosymbionts and on absorbing dissolved nutrients. Cnidaria give their symbiotic algae carbon dioxide, some nutrients, a place in the sun and protection against predators. Predatory species use their cnidocytes to poison or entangle prey, and those with venomous nematocysts may start digestion by injecting digestive enzymes. The "smell" of fluids from wounded prey makes the tentacles fold inwards and wipe the prey off into the mouth. In medusae the tentacles round the edge of the bell are often short and most of the prey capture is done by "oral arms", which are extensions of the edge of the mouth and are often frilled and sometimes branched to increase their surface area. Medusae often trap prey or suspended food particles by swimming upwards, spreading their tentacles and oral arms and then sinking. In species for which suspended food particles are important, the tentacles and oral arms often have rows of cilia whose beating creates currents that flow towards the mouth, and some produce nets of mucus to trap particles. Their digestion is both intra and extracellular. Once the food is in the digestive cavity, gland cells in the gastroderm release enzymes that reduce the prey to slurry, usually within a few hours. This circulates through the digestive cavity and, in colonial cnidarians, through the connecting tunnels, so that gastroderm cells can absorb the nutrients. Absorption may take a few hours, and digestion within the cells may take a few days. The circulation of nutrients is driven by water currents produced by cilia in the gastroderm or by muscular movements or both, so that nutrients reach all parts of the digestive cavity. Nutrients reach the outer cell layer by diffusion or, for animals or zooids such as medusae which have thick mesogleas, are transported by mobile cells in the mesoglea. Indigestible remains of prey are expelled through the mouth. The main waste product of cells' internal processes is ammonia, which is removed by the external and internal water currents. There are no respiratory organs, and both cell layers absorb oxygen from and expel carbon dioxide into the surrounding water. When the water in the digestive cavity becomes stale it must be replaced, and nutrients that have not been absorbed will be expelled with it. Some Anthozoa have ciliated grooves on their tentacles, allowing them to pump water out of and into the digestive cavity without opening the mouth. This improves respiration after feeding and allows these animals, which use the cavity as a hydrostatic skeleton, to control the water pressure in the cavity without expelling undigested food. Cnidaria that carry photosynthetic symbionts may have the opposite problem, an excess of oxygen, which may prove toxic. The animals produce large quantities of antioxidants to neutralize the excess oxygen. All cnidarians can regenerate, allowing them to recover from injury and to reproduce asexually. Medusae have limited ability to regenerate, but polyps can do so from small pieces or even collections of separated cells. This enables corals to recover even after apparently being destroyed by predators. Cnidarian sexual reproduction often involves a complex life cycle with both polyp and medusa stages. For example, in Scyphozoa (jellyfish) and Cubozoa (box jellies) a larva swims until it finds a good site, and then becomes a polyp. This grows normally but then absorbs its tentacles and splits horizontally into a series of disks that become juvenile medusae, a process called strobilation. The juveniles swim off and slowly grow to maturity, while the polyp re-grows and may continue strobilating periodically. The adults have gonads in the gastroderm, and these release ova and sperm into the water in the breeding season. This phenomenon of succession of differently organized generations (one asexually reproducing, sessile polyp, followed by a free-swimming medusa or a sessile polyp that reproduces sexually) is sometimes called "alternation of asexual and sexual phases" or "metagenesis", but should not be confused with the alternation of generations as found in plants. Shortened forms of this life cycle are common, for example some oceanic scyphozoans omit the polyp stage completely, and cubozoan polyps produce only one medusa. Hydrozoa have a variety of life cycles. Some have no polyp stages and some (e.g. "hydra") have no medusae. In some species, the medusae remain attached to the polyp and are responsible for sexual reproduction; in extreme cases these reproductive zooids may not look much like medusae. Meanwhile, life cycle reversal, in which polyps are formed directly from medusae without the involvement of sexual reproduction process, was observed in both Hydrozoa ("Turritopsis dohrnii" and "Laodicea undulata") and Scyphozoa ("Aurelia" sp.1). Anthozoa have no medusa stage at all and the polyps are responsible for sexual reproduction. Spawning is generally driven by environmental factors such as changes in the water temperature, and their release is triggered by lighting conditions such as sunrise, sunset or the phase of the moon. Many species of Cnidaria may spawn simultaneously in the same location, so that there are too many ova and sperm for predators to eat more than a tiny percentage — one famous example is the Great Barrier Reef, where at least 110 corals and a few non-cnidarian invertebrates produce enough gametes to turn the water cloudy. These mass spawnings may produce hybrids, some of which can settle and form polyps, but it is not known how long these can survive. In some species the ova release chemicals that attract sperm of the same species. The fertilized eggs develop into larvae by dividing until there are enough cells to form a hollow sphere (blastula) and then a depression forms at one end (gastrulation) and eventually becomes the digestive cavity. However, in cnidarians the depression forms at the end further from the yolk (at the animal pole), while in bilaterians it forms at the other end (vegetal pole). The larvae, called planulae, swim or crawl by means of cilia. They are cigar-shaped but slightly broader at the "front" end, which is the aboral, vegetal-pole end and eventually attaches to a substrate if the species has a polyp stage. Anthozoan larvae either have large yolks or are capable of feeding on plankton, and some already have endosymbiotic algae that help to feed them. Since the parents are immobile, these feeding capabilities extend the larvae's range and avoid overcrowding of sites. Scyphozoan and hydrozoan larvae have little yolk and most lack endosymbiotic algae, and therefore have to settle quickly and metamorphose into polyps. Instead, these species rely on their medusae to extend their ranges. All known cnidaria can reproduce asexually by various means, in addition to regenerating after being fragmented. Hydrozoan polyps only bud, while the medusae of some hydrozoans can divide down the middle. Scyphozoan polyps can both bud and split down the middle. In addition to both of these methods, Anthozoa can split horizontally just above the base. Asexual reproduction makes the daughter cnidarian a clone of the adult. Cnidarians were for a long time grouped with Ctenophores in the phylum Coelenterata, but increasing awareness of their differences caused them to be placed in separate phyla. Modern cnidarians are generally classified into four main classes: sessile Anthozoa (sea anemones, corals, sea pens); swimming Scyphozoa (jellyfish) and Cubozoa (box jellies); and Hydrozoa, a diverse group that includes all the freshwater cnidarians as well as many marine forms, and has both sessile members such as "Hydra" and colonial swimmers such as the Portuguese Man o' War. Staurozoa have recently been recognised as a class in their own right rather than a sub-group of Scyphozoa, and the parasitic Myxozoa and Polypodiozoa are now recognized as highly derived cnidarians rather than more closely related to the bilaterians. Stauromedusae, small sessile cnidarians with stalks and no medusa stage, have traditionally been classified as members of the Scyphozoa, but recent research suggests they should be regarded as a separate class, Staurozoa. The Myxozoa, microscopic parasites, were first classified as protozoans. Research then found that "Polypodium hydriforme", a non-Myxozoan parasite "within" the egg cells of sturgeon, is closely related to the Myxozoa and suggested that both "Polypodium" and the Myxozoa were intermediate between cnidarians and bilaterian animals. More recent research demonstrates that the previous identification of bilaterian genes reflected contamination of the Myxozoan samples by material from their host organism, and they are now firmly identified as heavily derived cnidarians, and more closely related to Hydrozoa and Scyphozoa than to Anthozoa. Some researchers classify the extinct conulariids as cnidarians, while others propose that they form a completely separate phylum. Current classification according to the World Register of Marine Species: Many cnidarians are limited to shallow waters because they depend on endosymbiotic algae for much of their nutrients. The life cycles of most have polyp stages, which are limited to locations that offer stable substrates. Nevertheless, major cnidarian groups contain species that have escaped these limitations. Hydrozoans have a worldwide range: some, such as "Hydra", live in freshwater; "Obelia" appears in the coastal waters of all the oceans; and "Liriope" can form large shoals near the surface in mid-ocean. Among anthozoans, a few scleractinian corals, sea pens and sea fans live in deep, cold waters, and some sea anemones inhabit polar seabeds while others live near hydrothermal vents over below sea-level. Reef-building corals are limited to tropical seas between 30°N and 30°S with a maximum depth of , temperatures between , high salinity, and low carbon dioxide levels. Stauromedusae, although usually classified as jellyfish, are stalked, sessile animals that live in cool to Arctic waters. Cnidarians range in size from a mere handful of cells for the parasitic myxozoans through "Hydra"'s length of , to the Lion's mane jellyfish, which may exceed in diameter and in length. Prey of cnidarians ranges from plankton to animals several times larger than themselves. Some cnidarians are parasites, mainly on jellyfish but a few are major pests of fish. Others obtain most of their nourishment from endosymbiotic algae or dissolved nutrients. Predators of cnidarians include: sea slugs, which can incorporate nematocysts into their own bodies for self-defense; starfish, notably the crown of thorns starfish, which can devastate corals; butterfly fish and parrot fish, which eat corals; and marine turtles, which eat jellyfish. Some sea anemones and jellyfish have a symbiotic relationship with some fish; for example clown fish live among the tentacles of sea anemones, and each partner protects the other against predators. Coral reefs form some of the world's most productive ecosystems. Common coral reef cnidarians include both Anthozoans (hard corals, octocorals, anemones) and Hydrozoans (fire corals, lace corals). The endosymbiotic algae of many cnidarian species are very effective primary producers, in other words converters of inorganic chemicals into organic ones that other organisms can use, and their coral hosts use these organic chemicals very efficiently. In addition, reefs provide complex and varied habitats that support a wide range of other organisms. Fringing reefs just below low-tide level also have a mutually beneficial relationship with mangrove forests at high-tide level and seagrass meadows in between: the reefs protect the mangroves and seagrass from strong currents and waves that would damage them or erode the sediments in which they are rooted, while the mangroves and seagrass protect the coral from large influxes of silt, fresh water and pollutants. This additional level of variety in the environment is beneficial to many types of coral reef animals, which for example may feed in the sea grass and use the reefs for protection or breeding. The earliest widely accepted animal fossils are rather modern-looking cnidarians, possibly from around , although fossils from the Doushantuo Formation can only be dated approximately. The identification of some of these as embryos of animals has been contested, but other fossils from these rocks strongly resemble tubes and other mineralized structures made by corals. Their presence implies that the cnidarian and bilaterian lineages had already diverged. Although the Ediacaran fossil "Charnia" used to be classified as a jellyfish or sea pen, more recent study of growth patterns in "Charnia" and modern cnidarians has cast doubt on this hypothesis, leaving only the Canadian polyp, "Haootia", as the only bona-fide cnidarian body fossil from the Ediacaran. Few fossils of cnidarians without mineralized skeletons are known from more recent rocks, except in lagerstätten that preserved soft-bodied animals. A few mineralized fossils that resemble corals have been found in rocks from the Cambrian period, and corals diversified in the Early Ordovician. These corals, which were wiped out in the Permian-Triassic extinction about , did not dominate reef construction since sponges and algae also played a major part. During the Mesozoic era rudist bivalves were the main reef-builders, but they were wiped out in the Cretaceous–Paleogene extinction event , and since then the main reef-builders have been scleractinian corals. It is difficult to reconstruct the early stages in the evolutionary "family tree" of animals using only morphology (their shapes and structures), because the large differences between Porifera (sponges), Cnidaria plus Ctenophora (comb jellies), Placozoa and Bilateria (all the more complex animals) make comparisons difficult. Hence reconstructions now rely largely or entirely on molecular phylogenetics, which groups organisms according to similarities and differences in their biochemistry, usually in their DNA or RNA. It is now generally thought that the Calcarea (sponges with calcium carbonate spicules) are more closely related to Cnidaria, Ctenophora (comb jellies) and Bilateria (all the more complex animals) than they are to the other groups of sponges. In 1866 it was proposed that Cnidaria and Ctenophora were more closely related to each other than to Bilateria and formed a group called Coelenterata ("hollow guts"), because Cnidaria and Ctenophora both rely on the flow of water in and out of a single cavity for feeding, excretion and respiration. In 1881, it was proposed that Ctenophora and Bilateria were more closely related to each other, since they shared features that Cnidaria lack, for example muscles in the middle layer (mesoglea in Ctenophora, mesoderm in Bilateria). However more recent analyses indicate that these similarities are rather vague, and the current view, based on molecular phylogenetics, is that Cnidaria and Bilateria are more closely related to each other than either is to Ctenophora. This grouping of Cnidaria and Bilateria has been labelled "Planulozoa" because it suggests that the earliest Bilateria were similar to the planula larvae of Cnidaria. Within the Cnidaria, the Anthozoa (sea anemones and corals) are regarded as the sister-group of the rest, which suggests that the earliest cnidarians were sessile polyps with no medusa stage. However, it is unclear how the other groups acquired the medusa stage, since Hydrozoa form medusae by budding from the side of the polyp while the other Medusozoa do so by splitting them off from the tip of the polyp. The traditional grouping of Scyphozoa included the Staurozoa, but morphology and molecular phylogenetics indicate that Staurozoa are more closely related to Cubozoa (box jellies) than to other "Scyphozoa". Similarities in the double body walls of Staurozoa and the extinct Conulariida suggest that they are closely related. The position of Anthozoa nearest the beginning of the cnidarian family tree also implies that Anthozoa are the cnidarians most closely related to Bilateria, and this is supported by the fact that Anthozoa and Bilateria share some genes that determine the main axes of the body. However, in 2005 Katja Seipel and Volker Schmid suggested that cnidarians and ctenophores are simplified descendants of triploblastic animals, since ctenophores and the medusa stage of some cnidarians have striated muscle, which in bilaterians arises from the mesoderm. They did not commit themselves on whether bilaterians evolved from early cnidarians or from the hypothesized triploblastic ancestors of cnidarians. In molecular phylogenetics analyses from 2005 onwards, important groups of developmental genes show the same variety in cnidarians as in chordates. In fact cnidarians, and especially anthozoans (sea anemones and corals), retain some genes that are present in bacteria, protists, plants and fungi but not in bilaterians. The mitochondrial genome in the medusozoan cnidarians, unlike those in other animals, is linear with fragmented genes. The reason for this difference is unknown. Jellyfish stings killed about 1,500 people in the 20th century, and cubozoans are particularly dangerous. On the other hand, some large jellyfish are considered a delicacy in East and Southeast Asia. Coral reefs have long been economically important as providers of fishing grounds, protectors of shore buildings against currents and tides, and more recently as centers of tourism. However, they are vulnerable to over-fishing, mining for construction materials, pollution, and damage caused by tourism. Beaches protected from tides and storms by coral reefs are often the best places for housing in tropical countries. Reefs are an important food source for low-technology fishing, both on the reefs themselves and in the adjacent seas. However, despite their great productivity, reefs are vulnerable to over-fishing, because much of the organic carbon they produce is exhaled as carbon dioxide by organisms at the middle levels of the food chain and never reaches the larger species that are of interest to fishermen. Tourism centered on reefs provides much of the income of some tropical islands, attracting photographers, divers and sports fishermen. However, human activities damage reefs in several ways: mining for construction materials; pollution, including large influxes of fresh water from storm drains; commercial fishing, including the use of dynamite to stun fish and the capture of young fish for aquariums; and tourist damage caused by boat anchors and the cumulative effect of walking on the reefs. Coral, mainly from the Pacific Ocean has long been used in jewellery, and demand rose sharply in the 1980s. Some large jellyfish species of the Rhizostomae order are commonly consumed in Japan, Korea and Southeast Asia. In parts of the range, fishing industry is restricted to daylight hours and calm conditions in two short seasons, from March to May and August to November. The commercial value of jellyfish food products depends on the skill with which they are prepared, and "Jellyfish Masters" guard their trade secrets carefully. Jellyfish is very low in cholesterol and sugars, but cheap preparation can introduce undesirable amounts of heavy metals. The "sea wasp" "Chironex fleckeri" has been described as the world's most venomous jellyfish and is held responsible for 67 deaths, although it is difficult to identify the animal as it is almost transparent. Most stingings by "C. fleckeri" cause only mild symptoms. Seven other box jellies can cause a set of symptoms called Irukandji syndrome, which takes about 30 minutes to develop, and from a few hours to two weeks to disappear. Hospital treatment is usually required, and there have been a few deaths. A number of the parasitic Myxozoans are commercially important pathogens in salmonid aquaculture.
https://en.wikipedia.org/wiki?curid=6621
Conservative Judaism Conservative Judaism (known as Masorti Judaism outside North America) is a Jewish religious movement that regards the authority of Jewish law and tradition as emanating primarily from the assent of the people and the community through the generations, more than from divine revelation. It therefore views Jewish law, or "halakha", as both binding and subject to historical development. The Conservative rabbinate employs modern historical-critical research, rather than only traditional methods and sources, and lends great weight to its constituency when determining its stance on matters of practice. The movement considers its approach as the authentic and most appropriate continuation of "halakhic" discourse, maintaining both fealty to received forms and flexibility in their interpretation. It also eschews strict theological definitions, lacking a consensus in matters of faith and allowing great pluralism. While regarding itself as the heir of Rabbi Zecharias Frankel's 19th-century Positive-Historical School in Europe, Conservative Judaism fully institutionalized only in the United States during the mid-20th century. Its largest center today is in North America, where its main congregational arm is the United Synagogue of Conservative Judaism and the New York–based Jewish Theological Seminary of America operates as its largest rabbinic seminary. Globally, affiliated communities are united within the umbrella organization Masorti Olami. Conservative Judaism is the third-largest Jewish religious movement worldwide, estimated to represent close to 1.1 million people, both over 600,000 registered adult congregants and many non-member identifiers. Conservative Judaism, from its earliest stages, was marked by ambivalence and ambiguity in all matters theological. Rabbi Zecharias Frankel, considered its intellectual progenitor, believed the very notion of theology was alien to traditional Judaism. He was often accused of obscurity on the subject by his opponents, both Reform and Orthodox. The American movement largely espoused a similar approach, and its leaders mostly avoided the field. Only in 1985 did a course about Conservative theology open in the Jewish Theological Seminary of America (JTS). The hitherto sole major attempt to define a clear credo was made in 1988, with the Statement of Principles "Emet ve-Emunah" (Truth and Belief), formulated and issued by the Leadership Council of Conservative Judaism. The introduction stated that "lack of definition was useful" in the past but a need to articulate one now arose. The platform provided many statements citing key concepts such as God, revelation and Election, but also acknowledged that a variety of positions and convictions existed within its ranks, eschewing strict delineation of principles and often expressing conflicting views. In a 1999 special edition of "Conservative Judaism" dedicated to the matter, leading rabbis Elliot N. Dorff and Gordon Tucker clarified that "the great diversity" within the movement "makes the creation of a theological vision shared by all neither possible nor desirable". Conservative Judaism largely upholds the theistic notion of a personal God. "Emet ve-Emunah" stated that "we affirm our faith in God as the Creator and Governor of the universe. His power called the world into being; His wisdom and goodness guide its destiny." Concurrently, the platform also noted that His nature was "elusive" and subject to many options of belief. A naturalistic conception of divinity, regarding it as inseparable from the mundane world, once had an important place within the movement, especially represented by Mordecai Kaplan. After Kaplan's Reconstructionism fully coalesced into an independent movement, these views were marginalized. A similarly inconclusive position is expressed toward other precepts. Most theologians adhere to the Immortality of the Soul, but while references to the Resurrection of the Dead are maintained, English translations of the prayers obscure the issue. In "Emet", it was stated that death is not tantamount to the end of one's personality. Relating to the Messianic ideal, the movement rephrased most petitions for the restoration of the Sacrifices into past tense, rejecting a renewal of animal offerings, though not opposing a Return to Zion and even a new Temple. The 1988 platform announced that "some" believe in classic eschathology, but dogmatism in this matter was "philosophically unjustified". The notions of Election of Israel and God's covenant with it were basically retained as well. Conservative conception of Revelation encompasses an extensive spectrum. Zecharias Frankel himself applied critical-scientific methods to analyze the stages in the development of the Oral Torah, pioneering modern study of the Mishnah. He regarded the Beatified Sages as innovators who added their own, original contribution to the canon, not merely as expounders and interpreters of a legal system given in its entirety to Moses on Mount Sinai. Yet he also vehemently rejected utilizing these disciplines on the Pentateuch, maintaining it was beyond human reach and wholly celestial in origin. Frankel never elucidated his beliefs, and the exact correlation between human and divine in his thought is still subject to scholarly debate. A similar negative approach toward Higher Criticism, while accepting an evolutionary understanding of Oral Law, defined Rabbi Alexander Kohut, Solomon Schechter and the early generation of American Conservative Judaism. When JTS faculty began to embrace Biblical criticism in the 1920s, they adapted a theological view consistent with it: an original, verbal revelation did occur at Sinai, but the text itself was composed by later authors. The latter, classified by Dorff as a relatively moderate metamorphosis of the old one, is still espoused by few traditionalist right-wing Conservative rabbis, though it is marginalized among senior leadership. A small but influential segment within the JTS and the movement adhered, from the 1930s, to Mordecai Kaplan's philosophy that denied any form of revelation but viewed all scripture as a purely human product. Along with other Reconstructionist tenets, it dwindled as the latter consolidated into a separate group. Kaplan's views and the permeation of Higher Criticism gradually swayed most Conservative thinkers towards a non-verbal understanding of theophany, which has become dominant in the 1970s. This was en sync with the wider trend of lowering rates of Americans who accepted the Bible as the Word of God. Dorff categorized the proponents of this into two schools. One maintains that God projected some form of message which inspired the human authors of the Pentateuch to record what they perceived. The other is often strongly influenced by Franz Rosenzweig and other existentialists, but also attracted many Objectivists who consider human reason paramount. The second school states that God conferred merely his presence on those he influenced, without any communication, and the experience drove them to spiritual creativity. While they differ in the theoretical level surrounding revelation, both practically regard all scripture and religious tradition as a human product with certain divine inspiration—providing an understanding that recognizes Biblical Criticism and also justifies major innovation in religious conduct. The first doctrine, advocated by such leaders as rabbis Ben-Zion Bokser and Robert Gordis, largely imparted that some elements within Judaism are fully divine but determining which would be impractical, and therefore received forms of interpretation should be basically upheld. Exponents of the latter view, among them rabbis Louis Jacobs and Neil Gillman, also emphasized the encounter of God with the Jews as a collective and the role of religious authorities through the generations in determining what it implied. The stress on the supremacy of community and tradition, rather than individual consciousness, defines the entire spectrum of Conservative thought. The Conservative mainstay was the adoption of the historical-critical method in understanding Judaism and setting its future course. In accepting an evolutionary approach to the religion, as something that developed over time and absorbed considerable external influences, the movement distinguished between the original meaning implied in traditional sources and the manner they were grasped by successive generations, rejecting belief in an unbroken chain of interpretation from God's original Revelation, immune to any major extraneous effects. This evolutionary perception of religion, while relatively moderate in comparison with more radical modernizers—the scholarship of the Positive-Historical school, for example, sought to demonstrate the continuity and cohesiveness of Judaism along the years—still challenged Conservative leaders. They regarded tradition and received mores with reverence, especially the continued adherence to the mechanism of Religious Law ("Halakha"), opposing indiscriminate modification, and emphasized they should be changed only with care and caution, and remain observed by the people. Rabbi Louis Ginzberg, summarizing his movement's position, wrote: This discrepancy between scientific criticism and insistence on heritage had to be compensated by a conviction that would forestall either deviation from accepted norms or laxity and apathy. A key doctrine which was to fulfill this capacity was the collective will of the Jewish people. Conservatives lent it great weight in determining religious practice, both in historical precedent and as a means to shape present conduct. Zecharias Frankel pioneered this approach; as Michael A. Meyer commented, "the extraordinary status which he ascribed to the ingrained beliefs and practices of the community is probably the most original element of his thought." He turned it into a source of legitimacy for both change and preservation, but mostly the latter. The basic moderation and traditionalism of the majority among the people were to guarantee a sense of continuity and unity, restraining the guiding rabbis and scholars who at his age were intent on reform, but also allowing them maneuverability in adopting or discarding certain elements. Solomon Schechter espoused a similar position. He turned the old rabbinic concept of "K'lal Yisrael", which he translated as "Catholic Israel", into a comprehensive worldview. For him, the details of divine Revelation were of secondary significance, as historical change dictated its interpretation through the ages notwithstanding: "the center of authority is actually removed from the Bible", he surmised, "and placed in some living body... in touch with the ideal aspirations and the religious needs of the age, best able to determine... This living body, however, is not represented by... priesthood, or Rabbihood, but by the collective conscience of Catholic Israel." The scope, limits and role of this corpus were a matter for contention in Conservative ranks. Schechter himself used it to oppose any major break with either traditionalist or progressive elements within American Jewry of his day, while some of his successors argued that the idea became obsolete due to the great alienation of many from received forms, that had to be countered by innovative measures to draw them back. The Conservative rabbinate often vacillated on to which degree may the non-practicing, religiously apathetic strata be included as a factor within Catholic Israel, providing impulse for them in determining religious questions; even avant-garde leaders acquiesced that the majority could not serve that function. Right-wing critics often charged that the movement allowed its uncommitted laity an exaggerated role, conceding to its demands and successively stretching "halakhic" boundaries beyond any limit. The Conservative leadership had limited success in imparting their worldview to the general public. While the rabbinate perceived itself as bearing a unique, original conception of Judaism, the masses lacked much interest, regarding it mainly as a compromise offering a channel for religious identification that was more traditional than Reform Judaism yet less strict than Orthodoxy. Only a low percentage of Conservative congregants actively pursue an observant lifestyle: in the mid-1980s, Charles Liebman and Daniel J. Elazar calculated that barely 3 to 4 per cent held to one quite thoroughly. This gap between principle and the public, more pronounced than in any other Jewish movement, is often credited at explaining the decline of the Conservative movement. While some 41 per cent of American Jews identified with it in the 1970s, it had shrunk to an estimated 18 per cent (and 11 per cent among those under 30) in 2013. Fidelity and commitment to "Halakha", while subject to criticism as disingenuous both from within and without, were and remain a cornerstone doctrine of Conservative Judaism: The movement views the legalistic system as normative and binding, and believes Jews must practically observe its precepts, like Sabbath, dietary ordinances, ritual purity, daily prayer with phylacteries and the like. Concurrently, examining Jewish history and rabbinic literature through the lens of academic criticism, it maintained that these laws were always subject to considerable evolution, and must continue to do so. "Emet ve-Emunah" titled its chapter on the subject with "The Indispensability of Halakha", stating that ""Halakha" in its developing form is an indispensable element of a traditional Judaism which is vital and modern." Conservative Judaism regards itself as the authentic inheritor of a flexible legalistic tradition, charging the Orthodox with petrifying the process and Reform with abandoning it. The tension between "tradition and change"—which were also the motto adopted by the movement since the 1950s—and the need to balance them were always a topic of intense debate within Conservative Judaism. In its early stages, the leadership opposed pronounced innovation, mostly adopting a relatively rigid position. Mordecai Kaplan's Reconstructionism raised the demand for thoroughgoing modification without much regard for the past or "halakhic" considerations, but senior rabbis opposed him vigorously. Even in the 1940s and 1950s, when Kaplan's influence grew, his superiors rabbis Ginzberg, Louis Finkelstein and Saul Lieberman espoused a very conservative line. Since the 1970s, with the strengthening of the liberal wing within the movemement, the majority in the Rabbinic Assembly opted for quite radical reformulations in religious conduct, but rejected the Reconstructionist non-"halakhic" approach, insisting that the legalistic method be maintained. The "halakhic" commitment of Conservative Judaism has been subject to much criticism, from within and without. Right-wing discontents, including the Union for Traditional Judaism which seceded in protest of the 1983 resolution to ordain women rabbis—adopted at an open vote, where all JTS faculty regardless of qualification were counted—contested the validity of this description, as well as progressives like Rabbi Neil Gillman, who exhorted the movement to cease describing itself as "halakhic" in 2005, stating that after repeated concessions, "our original claim has died a death by a thousand qualifications... It has lost all factual meaning." The main body entrusted with formulating rulings, responsa and statues is the Committee on Jewish Law and Standards (CJLS), a panel with 25 voting legalistic specialists and further 11 observers. There is also the smaller "Va'ad ha-Halakha" (Law Committee) of Israel's Masorti Movement. Every responsa must receive a minimum of six voters to be considered an official position of the CJLS. Conservative Judaism explicitly acknowledges the principle of "halakhic" pluralism, enabling the panel to adopt more than one resolution in any given subject. The final authority in each Conservative community is the local rabbi, the "mara d'atra" (Lord of the Locality, in traditional terms), enfranchised to adopt either minority or majority opinions from the CJLS or maintain local practice. Thus, on the issue of admitting openly Homosexual rabbinic candidates, the Committee approved two resolutions, one in favour and one against; the JTS took the lenient position, while the Seminario Rabinico Latinoamericano still adheres to the latter. Likewise, while most Conservative synagogues approved of egalitarianism for women in religious life, some still maintain traditional gender roles and do not count females for prayer quorums. The Conservative treatment of "Halakha" is defined by several features, though the entire range of its "Halakhic" discourse cannot be sharply distinguished from either the traditional or Orthodox one. Rabbi David Golinkin, who attempted to classify its parameters, stressed that quite often rulings merely reiterate conclusions reached in older sources or even Orthodox ones. for example, in the details of preparing Sabbath ritual enclosures, it draws directly on the opinions of the "Shulchan Aruch" and Rabbi Hayim David HaLevi. Another tendency prevalent among the movement's rabbis, yet again not particular to it, is the adoption of the more lenient positions on the matters at question—though this is not universal, and responsa also took stringent ones not infrequently. A more distinctive characterization is a greater proclivity to base rulings on earlier sources, in the "Rishonim" or before them, as far back as the Talmud. Conservative decisors frequently resort to less canonical sources, isolated responsa or minority opinions. They demonstrate more fluidity in regards to established precedent and continuum in rabbinic literature, mainly those by the later authorities, and lay little stress on the perceived hierarchy between major and minor legalists of the past. They are far more inclined to contend ("machloket") with old rulings, to be flexible towards custom or to wholly disregard it. This is especially expressed in less hesitancy to rule against or notwithstanding the major codifications of Jewish Law, like "Mishne Torah", "Arba'ah Turim" and especially the "Shulchan Aruch" with its Isserles Gloss and later commentaries. Conservative authorities, while often relying on the "Shulchan Aruch" themselves, criticize the Orthodox for relatively rarely venturing beyond it and overly canonizing Rabbi Joseph Karo's work. In several occasions, Conservative rabbis discerned that the "Shulchan Aruch" ruled without firm precedent, sometimes deriving his conclusions from the "Kabbalah". An important example is the ruling of Rabbi Golinkin—contrary to the majority consensus among the "Acharonim" and the more prominent "Rishonim", but based on many opinions of the lesser "Rishonim" which is derived from a minority view in the Talmud—that the Sabbatical Year is not obligatory in present times at all (neither "de'Oraita" nor "de'Rabanan") but rather an act of piety. Ethical considerations and the weight due to them in determining "halakhic" issues, mainly to what degree may modern sensibilities shape the outcome, are subject to much discourse. Right-wing decisors, like Rabbi Joel Roth, maintained that such elements are naturally a factor in formulating conclusions, but may not alone serve as a justification for adopting a position. The majority, however, basically subscribed to the opinion evinced already by Rabbi Seymour Siegel in the 1960s, that the cultural and ethical norms of the community, the contemporary equivalents of Talmudic "Aggadah", should supersede the legalistic forms when the two came into conflict and there was a pivotal ethical concern. Rabbi Elliot Dorff concluded that in contrast to the Orthodox, Conservative Judaism maintains that the juridical details and processes mainly serve higher moral purposes and could be modified if they no longer do so: "in other words, the "Aggadah" should control the "Halakha"." The liberal Rabbi Gordon Tucker, along with Gillman and other progressives, supported a far-reaching implementation of this approach, making Conservative Judaism much more "Aggadic" and allowing moral priorities an overriding authority at all occasions. This idea became very popular among the young generation, but it was not fully embraced either. In the 2006 resolution on homosexuals, the CJLS chose a middle path: they agreed that the ethical consideration of human dignity was of supreme importance, but not sufficient to uproot the express Biblical prohibition on not to lie with mankind as with womankind (traditionally understood as banning full anal intercourse). All other limitations, including on other forms of sexual relations, were lifted. A similar approach is manifest in the great weight ascribed to sociological changes in deciding religious policy. The CJLS and the Rabbinical Assembly members frequently state that circumstances were profoundly transformed in modern times, fulfilling the criteria mandating new rulings in various fields (based on general talmudic principles like "Shinui ha-I'ttim", "Change of Times"). This, along with the ethical aspect, was a main argument for revolutionizing the role of women in religious life and embracing egalitarianism. The most distinctive feature of Conservative legalistic discourse, in which it is conspicuously and sharply different from Orthodoxy, is the incorporation of critical-scientific methods into the process. Deliberations almost always delineate the historical development of the specific issue at hand, from the earliest known mentions until modern times. This approach enables a thorough analysis of the manner in which it was practiced, accepted, rejected or modified in various periods, not necessarily en sync with the received rabbinic understanding. Archaeology, philology and Judaic Studies are employed; rabbis use comparative compendiums of religious manuscripts, sometimes discerning that sentences were only added later or include spelling, grammar and transcription errors, changing the entire understanding of certain passages. This critical approach is central to the movement, for its historicist underpinning stresses that all religious literature has an original meaning relevant in the context of its formulation. This meaning may be analyzed and discerned, and is distinct from the later interpretations ascribed by traditional commentators. Decisors are also far more prone to include references to external scientific sources in relevant fields, like veterinarian publications in "halakhic" matters concerning livestock. Conservative authorities, as part of their promulgation of a dynamic "Halakha", often cite the manner in which the sages of old used rabbinic statues ("Takkanah") that enabled to bypass prohibitions in the Pentateuch, like the "Prozbul" or "Heter I'ska". In 1948, when employing those was first debated, Rabbi Isaac Klein argued that since there was no consensus on leadership within Catholic Israel, formulation of significant "takkanot" should be avoided. Another proposal, to ratify them only with a two-thirds majority in the RA, was rejected. New statues require a simple majority, 13 supporters among the 25 members of the CJLS. In the 1950s and 1960s, such drastic measures—as Rabbi Arnold M. Goodman cited in a 1996 writ allowing members of the priestly caste to marry divorcees, "Later authorities were reluctant to assume such unilateral authority... fear that invoking this principle would create the proverbial slippery slope, thereby weakening the entire "halakhic" structure... thus imposed severe limitations on the conditions and situations where it would be appropriate"—were carefully drafted as temporal, emergency ordinances ("Horaat Sha'ah"), grounded on the need the avoid a total rift of many nonobservant Jews. Later on, these ordinances became accepted and permanent on the practical level. The Conservative movement issued a wide range of new, thoroughgoing statues, from the famous 1950 responsum that allowed driving to the synagogue on the Sabbath and up to the 2000 decision to ban rabbis from inquiring about whether someone was a Bastard, de facto abolishing this legal category. The RA and CJLS reached many decisions through the years, shaping a distinctive profile for Conservative practice and worship. In the 1940s, when the public demanded mixed seating of both sexes in synagogue, some rabbis argued there was no precedent but obliged on the ground of dire need (Eth la'asot); others noted that archaeological research showed no partitions in ancient synagogues. Mixed seating became commonplace in almost all congregations. In 1950, it was ruled that using electricity (that is, closure of an electrical circuit) did not constitute kindling a fire unto itself, not even in incandescent bulbs, and therefore was not a forbidden labour and could be done on the Sabbath. On that basis, while performing banned labours is of course forbidden—for example, video recording is still constituted as writing—switching lights and other functions are allowed, though the RA strongly urges adherents to keep the sanctity of the Sabbath (refraining from doing anything that may imitate the atmosphere of weekdays, like loud noise reminiscent of work). The need to encourage arrival at synagogue also motivated the CJLS, during the same year, to issue a temporal statue allowing driving on that day, for that purpose alone; it was supported by decreeing that the combustion of fuel did not serve any of the acts prohibited during the construction of the Tabernacle, and could therefore be classified, according to their interpretation of the Tosafists' opinion, as "redundant labour" ("Sh’eina Tzricha L’gufa") and be permitted. The validity of this argument was heavily disputed within the movement. In 1952, members of the priestly caste were allowed to marry divorcees, conditioned on forfeiture of their privileges, as termination of marriage became widespread and women who underwent it could not be suspected of unsavory acts. In 1967, the ban on priests marrying converts was also lifted. In 1954, the issue of "agunot" (women refused divorce by their husbands) was largely settled by adding a clause to the prenuptial contract under which men had to pay alimony as long as they did not concede. In 1968, this mechanism was replaced by a retroactive expropriation of the bride price, rendering the marriage void. In 1955, more girls were celebrating Bat Mitzvah and demanded to be allowed ascents to the Torah, the CJLS agreed that the ordinance under which women were banned from this due to respect for the congregation ("Kvod ha'Tzibur") was no longer relevant. In 1972 it was decreed that rennet, even if derived from unclean animals, was so transformed that it constituted a wholly new item ("Panim Chadashot ba'u l'Khan") and therefore all hard cheese could be considered kosher. The 1970s and 1980s saw the emergence of women's rights on the main agenda. Growing pressure led the CJLS to adopt a motion that females may be counted as part of a quorum, based on the argument that only the "Shulchan Aruch" explicitly stated that it consist of men. While accepted, this was very controversial in the Committee and heavily disputed. A more complete solution was offered in 1983 by Rabbi Joel Roth, and was also enacted to allow women rabbinic ordination. Roth noted that some decisors of old acknowledged that women may bless when performing positive time-bound commandments (from which they are exempted, and therefore unable to fulfill the obligation for others), especially citing the manner in which they assumed upon themselves the Counting of the Omer. He suggested that women voluntarily commit to pray thrice a day et cetera, and his responsa was adopted. Since then, female rabbis were ordained at JTS and other seminaries. In 1994, the movement accepted Judith Hauptman's principally egalitarian argument, according to which equal prayer obligations for women were never banned explicitly and it was only their inferior status that hindered participation. In 2006, openly gay rabbinic candidates were also to be admitted into the JTS. In 2012, a commitment ceremony for same-sex couples was devised, though not defined as "kiddushin". In 2016, the rabbis passed a resolution supporting transgender rights. Conservative Judaism in the United States held a relatively strict policy regarding intermarriage. Propositions for acknowledging Jews by patrilineal descent, as in the Reform movement, were overwhelmingly dismissed. Unconverted spouses were largely barred from community membership and participation in rituals; clergy are banned from any involvement in interfaith marriage on pain of dismissal. However, as the rate of such unions rose dramatically, Conservative congregations began describing gentile family members as "K'rov Yisrael" (Kin of Israel) and be more open toward them. The Leadership Council of Conservative Judaism stated in 1995: "we want to encourage the Jewish partner to maintain his/her Jewish identity, and raise their children as Jews." Despite the centralization of legal deliberation on matters of Jewish law in the CJLS individual synagogues and communities must, in the end, depend on their local decision-makers. The rabbi in his or her or their community is regarded as the Mara D'atra, or the local halakhic decisor. Rabbis trained in the reading practices of Conservative Jewish approaches, historical evaluation of Jewish law and interpretation of Biblical and Rabbinic texts may align directly with the CJLS decisions or themselves opine on matters based on precedents or readings of text that shine light on congregants' questions. So, for instance, a rabbi may or may not choose to permit video streaming on Shabbat despite a majority ruling that allows for use of electronics. A local mara d'atra may rely on the reasoning found in the majority or minority opinions of the CJLS or have other textual and halakhic grounds, i.e., prioritizing Jewish values or legal concepts, to rule one way or another on matters of ritual, family life or sacred pursuits. This balance between a centralization of halakhic authority and maintaining the authority of local rabbis reflects the commitment to pluralism at the heart of the Movement. The term "Conservative Judaism" was used, still generically and not yet as a specific label, already in the 1887 dedication speech of the Jewish Theological Seminary of America by Rabbi Alexander Kohut. By 1901, the JTS alumni formed the Rabbinical Assembly, of which all ordained Conservative clergy in the world are members. As of 2010, there were 1,648 rabbis in the RA. In 1913, the United Synagogue of America, renamed the United Synagogue of Conservative Judaism in 1991, was founded as a congregational arm of the RA. The movement established the World Council of Conservative Synagogues in 1957. Offshoots outside North America mostly adopted the Hebrew name "Masorti", traditional', as did the Israeli Masorti Movement, founded in 1979, and the British Assembly of Masorti Synagogues, formed in 1985. The World Council eventually changed its name to "Masorti Olami", Masorti International. Besides the RA, the international Cantors Assembly supplies prayer leaders for congregations worldwide. The United Synagogue of Conservative Judaism, covering the United States, Canada and Mexico, is by far the largest constituent of Masorti Olami. While most congregations defining themselves as "Conservative" are affiliated with the USCJ, some are independent. While accurate information of Canada is scant, it is estimated that some third of religiously affiliated Canadian Jews are Conservative. In 2008, the more traditional Canadian Council of Conservative Synagogues seceded from the parent organization. It numbered seven communities as of 2014. According to the Pew Research Center survey in 2013, 18 per cent of Jews in the United States identified with the movement, making it the second largest in the country. Steven M. Cohen calculated that as of 2013, 962,000 U.S. Jewish adults considered themselves Conservative: 570,000 were registered congregants and further 392,000 were not members in a synagogue but identified. In addition, Cohen assumed in 2006 that 57,000 unconverted non-Jewish spouses were also registered (12 per cent of member households had one at the time): 40 per cent of members intermarry. Conservatives are also the most aged group: among those aged under 30 only 11 per cent identified as such, and there are three people over 55 for every single one aged between 35 and 44. As of November 2015, the USCJ had 580 member congregations (a sharp decline from 630 two years prior), 19 in Canada and the remainder in the United States. In 2011 the USCJ initiated a plan to reinvigorate the movement. Beyond North America, the movement has little presence—in 2011, Rela Mintz Geffen appraised there were only 100,000 members outside the U.S. (and the former figure including Canada). "Masorti AmLat", the MO branch in Latin America, is the largest with 35 communities in Argentina, 7 in Brazil, 6 in Chile and further 11 in the other countries. The British Assembly of Masorti Synagogues has 13 communities and estimates its membership at over 4,000. More than 20 communities are spread across Europe, and there are 3 in Australia and 2 in Africa. The Masorti Movement in Israel incorporates some 70 communities and prayer groups with several thousand full members. In addition, while Hungarian Neolog Judaism, with a few thousands of adherents and forty partially active synagogues, is not officially affiliated with Masorti Olami, Conservative Judaism regards it as a fraternal, "non-Orthodox but halakhic" movement. In New York, the JTS serves as the movement's original seminary and legacy institution, along with the Ziegler School of Rabbinic Studies at the American Jewish University in Los Angeles; the Marshall T. Meyer Latin American Rabbinical Seminary (Spanish: "Seminario Rabínico Latinoamericano Marshall T. Meyer"), in Buenos Aires, Argentina; and the Schechter Institute of Jewish Studies in Jerusalem. A Conservative institution that does not grant rabbinic ordination but which runs along the lines of a traditional yeshiva is the Conservative Yeshiva, located in Jerusalem. The Neolog Budapest University of Jewish Studies also maintains connections with Conservative Judaism. The current chancellor of the JTS is Rabbi Arnold Eisen, in office since 2008. The current dean of the Ziegler School of Rabbinic Studies is Bradley Shavit Artson. The Committee on Jewish Law and Standards is chaired by Rabbi Elliot N. Dorff, serving since 2007. The Rabbinical Assembly is headed by President Rabbi Debra Newman Kamin, as of 2019, and managed by executive vice-president Rabbi Julie Schonfeld. The USCJ is directed by President Ned Gladstein. In South America, Rabbi Ariel Stofenmacher serves as chancellor in the Seminary and Rabbi Marcelo Rittner as president of Masorti AmLat. In Britain, the Masorti Assembly is chaired by Senior Rabbi Jonathan Wittenberg. In Israel, the Masorti movement's executive director is Yizhar Hess and chair Sophie Fellman Rafalovitz. The global youth movement is known as NOAM, an acronym for No'ar Masorti; its North American chapter is called the United Synagogue Youth. Marom Israel is the Masorti movement's organization for students and young adults, providing activities based on religious pluralism and Jewish content. The Women's League for Conservative Judaism is also active in North America. The USCJ maintains the Solomon Schechter Day Schools, comprising 76 day schools in 17 American states and 2 Canadian provinces serving Jewish children. Many other "community day schools" that are not affiliated with Schechter take a generally Conservative approach, but unlike these, generally have "no barriers to enrollment based on the faith of the parents or on religious practices in the home". During the first decade of the 21st century, a number of schools that were part of the Schechter network transformed themselves into non-affiliated community day schools. The USCJ also maintains the Camp Ramah system, where children and adolescents spend summers in an observant environment. The rise of modern, centralized states in Europe by the early 19th century hearkened the end of Jewish judicial autonomy and social seclusion. Their communal corporate rights were abolished, and the process of emancipation and acculturation that followed quickly transformed the values and norms of the public. Estrangement and apathy toward Judaism were rampant. The process of communal, educational and civil reform could not be restricted from affecting the core tenets of the faith. The new academic, critical study of Judaism ("Wissenschaft des Judentums") soon became a source of controversy. Rabbis and scholars argued to what degree, if at all, its findings could be used to determine present conduct. The modernized Orthodox in Germany, like rabbis Isaac Bernays and Azriel Hildesheimer, were content to cautiously study it while stringently adhering to the sanctity of holy texts and refusing to grant "Wissenschaft" any say in religious matters. On the other extreme were Rabbi Abraham Geiger, who would emerge as the founding father of Reform Judaism, and his supporters. They opposed any limit on critical research or its practical application, laying more weight on the need for change than on continuity. The Prague-born Rabbi Zecharias Frankel, appointed chief rabbi of the Kingdom of Saxony in 1836, gradually rose to become the leader of those who stood at the middle. Besides working for the civic betterment of local Jews and educational reform, he displayed keen interest in "Wissenschaft". But Frankel was always cautious and deeply reverent towards tradition, privately writing in 1836 that "the means must be applied with such care and discretion... that forward progress will be reached unnoticed, and seem inconsequential to the average spectator." He soon found himself embroiled in the great disputes of the 1840s. In 1842, during the second Hamburg Temple controversy, he opposed the new Reform prayerbook, arguing the elimination of petitions for a future Return to Zion led by the Messiah was a violation of an ancient tenet. But he also opposed the ban placed on the tome by Rabbi Bernays, stating this was a primitive behaviour. In the same year, he and the moderate conservative S.L. Rapoport were the only ones of nineteen respondents who negatively answered the Breslau community's enquiry on whether the deeply unorthodox Geiger could serve there. In 1843, Frankel clashed with the radical Reform rabbi Samuel Holdheim, who argued that the act of marriage in Judaism was a civic ("memonot") rather than sanctified ("issurim") matter and could be subject to the Law of the Land. In December 1843 Frankel launched the magazine "Zeitschrift für die Religiösen Interessen des Judenthums". In the preamble, he attempted to present his approach to the present plight: "the further development of Judaism cannot be done through Reform that would lead to total dissipation... But must be involved in its study... pursued via scientific research, on a "positive, historical" basis." The term Positive-Historical became associated with him and his middle way. The "Zeitschrift" was, along the convictions of its publisher, neither dogmatically orthodox nor overly polemic, wholly opposing Biblical criticism and arguing for the antiquity of custom and practice. In 1844, Geiger and like-minded allies arranged a conference in Braunschweig that was to have enough authority (since 1826, Rabbi Aaron Chorin called for the convocation of a new "Sanhedrin") to debate and enact thoroughgoing revisions. Frankel was willing to agree only to a meeting without any practical results, and refused the invitation. When the protocols, which contained many radical statements, were published, he denounced the assembly for "applying the scalpel of criticism" and favouring the spirit of the age over tradition. However, he later agreed to attend the second conference, held in Frankfurt am Main on 15 July 1845—in spite of warnings from Rapoport, who cautioned that compromise with Geiger was impossible and he would only damage his reputation among the traditionalists. On the 16th, the issue of Hebrew in the liturgy arose. Most present were inclined to retain it, but with more German segments. A small majority adopted a resolution stating there were subjective, but no objective, imperatives to keep it as the language of service. Frankel then astounded his peers by vehemently protesting, stating it was a breach with the past and that Hebrew was of dire importance and great sentimental value. The others immediately began quoting all passages in rabbinic literature allowing prayer in the vernacular. Frankel could not contend with the "halakhic" validity of their decision, but he perceived it as a sign of profound differences between them. On the 17th he formally withdrew, publishing a lambasting critique of the procedures. "Opponents of the conference, who feared he went to the other side," noted historian Michael A. Meyer, "now felt reassured of his loyalty". The rabbi of Saxony had many sympathizers, who supported a similarly moderate approach and change only on the basis of the authority of the Talmud. When Geiger began preparing a third conference in Breslau, Hirsch Bär Fassel convinced Frankel to organize one of his own in protest. Frankel invited colleagues to an assembly in Dresden, which was to be held on 21 October 1846. He announced that one measure he was willing to countenance was the possible abolition of the second day of festivals, though only if a broad consensus will be reached and not before thorough deliberation. Attendants were to include Rapoport, Fassel, Adolf Jellinek, Leopold Löw, Michael Sachs, Abraham Kohn and others. However, the Dresden assembly soon drew heated Orthodox resistance, especially from Rabbi Jacob Ettlinger, and was postponed indefinitely. In 1854, Frankel was appointed chancellor in the new Jewish Theological Seminary of Breslau, the first modern rabbinical seminary in Germany. His opponents on both flanks were incensed. Geiger and the Reform camp long accused him of theological ambiguity, hypocrisy and attachment to stagnant remnants, and now protested the "medieval" atmosphere in the seminary, which was mainly concerned with teaching Jewish Law. The hardline Orthodox Samson Raphael Hirsch, who fiercely opposed "Wissenschaft" and emphasized the divine origin of the entire "halakhic" system in the Theophany at Sinai, was deeply suspicious of Frankel's beliefs, use of science and constant assertions that Jewish Law was flexible and evolving. The final schism between Frankel and the Orthodox occurred after the 1859 publication of his "Darke ha-Mishna" (Ways of the Mishna). He heaved praise on the Beatified Sages, presenting them as bold innovators, but not once affirmed the divinity of the Oral Torah. On the ordinances classified as Law given to Moses at Sinai, he quoted Asher ben Jehiel that stated several of those were only apocryphally dubbed as such; he applied the latter's conclusion to all, noting they were "so evident "as if" given at Sinai". Hirsch branded Frankel a heretic, demanding he announce whether he believed that both the Oral and Written Torah were of celestial origin. Rabbis Benjamin Hirsch Auerbach, Solomon Klein and others published more complaisant tracts, but also requested an explanation. Rapoport marshaled to Frankel's aid, assuring that his words were merely reiterating ben Jehiel's and that he would soon release a statement that will belie Hirsch's accusations. But then the Chancellor of Breslau issued an ambiguous defence, writing that his book was not concerned with theology and avoiding giving any clear answer. Now even Rapoport joined his critics. Hirsch succeeded, severely tarnishing his Frankel's reputation among most concerned. Along with fellow Orthodox Rabbi Azriel Hildesheimer, Hirsch launched a protracted public campaign through the 1860s. They ceaselessly stressed the chasm between an Orthodox understanding of "Halakha" as derived and revealed, applied differently to different circumstances and subject to human judgement and possibly error, yet unchanging and divine in principle—as opposed to an evolutionary, historicist and non-dogmatic approach in which past authorities were not just elaborating but consciously innovating, as taught by Frankel. Hildesheimer often repeated that this issue utterly overshadowed any specific technical argument with the Breslau School (the students of which were often more lenient on matters of headcovering for women, Chalav Yisrael and other issues). Hildesheimer was concerned that Jewish public opinion perceived no practical difference between them; though he cared to distinguish the observant acolytes of Frankel from the Reform camp, he noted in his diary: "how meager is the principal difference between the Breslau School, who don silk gloves at their work, and Geiger who wields a sledgehammer." In 1863, when Breslau faculty member Heinrich Graetz published an article where he appeared to doubt the Messianic belief, Hildesheimer immediately seized upon the occasion to prove once more the dogmatic, rather than practical, divide. He denounced Graetz as a heretic. The Positive-Historical School was influential, but never institutionalized itself as thoroughly as its opponents. Apart from the many graduates of Breslau, Isaac Noah Mannheimer, Adolf Jellinek and Rabbi Moritz Güdemann led the central congregation in Vienna along a similar path. In Jellinek's local seminary, Meir Friedmann and Isaac Hirsch Weiss followed Frankel's moderate approach to critical research. The rabbinate of the liberal Neolog public in Hungary, which formally separated from the Orthodox, was also permeated with the "Breslsu spirit". Many of its members studied there, and its Jewish Theological Seminary of Budapest was modeled after it, though the assimilationist congregants cared little for rabbinic opinion. In Germany itself, Breslau alumni founded in 1868 a short-lived society, the Jüdisch-Theologische Verein. It was dissolved within a year, boycotted by both Reform and Orthodox. Michael Sachs led the Berlin congregation in a very conservative style, eventually resigning when an organ was introduced in services. Manuel Joël, another of the Frankelist party, succeeded Geiger in Breslau. He maintained his predecessor's truncated German translation of the liturgy for the sake of compromise, but restored the full Hebrew text. The Breslau Seminary and the Reform Hochschule für die Wissenschaft des Judentums maintained very different approaches; but on the communal level, the former's alumni failure to organize or articulate a coherent agenda, coupled with the declining prestige of Breslau and the conservatism of the Hochschule's alumni—a necessity in heterogeneous communities which remained unified, especially after the Orthodox gained the right to secede in 1876—imposed a rather uniform and mild character on what was known in Germany as "Liberal Judaism". In 1909, 63 rabbis associated with the Breslau approach founded the Freie jüdische Vereinigung, another brief attempt at institutionalization, but it too failed soon. Only in 1925 did the Religiöse Mittelpartei für Frieden und Einheit succeed in driving the same agenda. It won several seats in communal elections, but was small and of little influence. Jewish immigration to the United States bred an amalgam of loose communities, lacking strong tradition or stable structures. In this free-spirited environment, a multitude of forces was at work. As early as 1866, Rabbi Jonas Bondi of New York wrote that a Judaism of the "golden middleway, which was termed Orthodox by the left and heterodox or reformer by the right" developed in the new country. The rapid ascendancy of Reform Judaism by the 1880s left few who opposed it: merely a handful of congregations and ministers remained outside the Union of American Hebrew Congregations. These included Sabato Morais and Rabbi Henry Pereira Mendes of the elitist Sephardi congregations, along with rabbis Bernard Drachman (ordained at Breslau, though he regarded himself as Orthodox) and Henry Schneeberger. While spearheaded by radical and principled Reformers like Rabbi Kaufmann Kohler, the UAHC was also home to more conservative elements. President Isaac Meyer Wise, a pragmatist intent on compromise, hoped to forge a broad consensus that would turn a moderate version of Reform to dominant in America. He kept the dietary laws at home and attempted to assuage traditionalists. On 11 July 1883, apparently due to negligence by the Jewish caterer, non-kosher dishes were served to UAHC rabbis in Wise's presence. Known to posterity as the "trefa banquet", it purportedly made some guests abandon the hall in disgust, but little is factually known about the incident. In 1885, the traditionalist forces were bolstered upon the arrival of Rabbi Alexander Kohut, an adherent of Frankel. He publicly excoriated Reform for disdaining ritual and received forms, triggering a heated polemic with Kohler. The debate was one of the main factors which motivated the latter to compose the Pittsburgh Platform, which unambiguously declared the principles of Reform Judaism: "to-day we accept as binding only the moral laws, and maintain only such ceremonies as elevate and sanctify our lives." The explicit wording alienated a handful of conservative UAHC ministers: Henry Hochheimer, Frederick de Sola Mendes, Aaron Wise, Marcus Jastrow, and Benjamin Szold. They joined Kohut, Morais and the others in seeking to establish a traditional rabbinic seminary that would serve as a counterweight to Hebrew Union College. In 1886, they founded the Jewish Theological Seminary of America in New York City. Kohut, professor of Talmud who held to the Positive-Historical ideal, was the main educational influence in the early years, prominent among the founders who encompassed the entire spectrum from progressive Orthodox to the brink of Reform; to describe what the seminary intended to espouse, he used the term "Conservative Judaism", which had no independent meaning at the time and was only in relation to Reform. In 1898, Pereira Mendes, Schneeberger and Drachman also founded the Orthodox Union, which maintained close ties with the seminary. The JTS was a small, fledgling institution with financial difficulties, and was ordaining merely a rabbi per year. But soon after Chancellor Morais' death in 1897, its fortunes turned. Since 1881, a wave of Jewish immigration from Eastern Europe was inundating the country—by 1920, 2.5 million of them had arrived, increasing American Jewry tenfold. They came from regions where civil equality or emancipation were never granted, while acculturation and modernization made little headway. Whether devout or irreligious, they mostly retained strong traditional sentiments in matters of faith, accustomed to old-style rabbinate; the hardline Agudas HaRabbanim, founded by emigrant clergy, opposed secular education or vernacular sermons, and its members spoke almost only Yiddish. The Eastern Europeans were alienated by the local Jews, who were all assimilated in comparison, and especially aghast by the mores of Reform. The need to find a religious framework that would both accommodate and Americanize them motivated Jacob Schiff and other rich philanthropists, all Reform and of German descent, to donate $500,000 to the JTS. The contribution was solicited by Professor Cyrus Adler. It was conditioned on the appointment of Solomon Schechter as Chancellor. In 1901, the Rabbinical Assembly was established as the fraternity of JTS alumni. Schechter arrived in 1902, and at once reorganized the faculty, dismissing both Pereira Mendes and Drachman for lack of academic merit. Under his aegis, the institute began to draw famous scholars, becoming a center of learning on par with HUC. Schechter was both traditional in sentiment and quite unorthodox in conviction. He maintained that theology was of little importance and it was practice that must be preserved. He aspired to solicit unity in American Judaism, denouncing sectarianism and not perceiving himself as leading a new denomination: "not to create a new party, but to consolidate an old one". The need to raise funds convinced him that a congregational arm for the Rabbinical Assembly and the JTS was required. On 23 February 1913, he founded the United Synagogue of America (since 1991: United Synagogue of Conservative Judaism), which then consisted of 22 communities. He and Mendes first came to major disagreement; Schechter insisted that any alumnus could be appointed to the USoA's managerial board, and not just to serve as communal rabbi, including several the latter did not consider sufficiently devout, or who tolerated mixed seating in their synagogues (though some of those he still regarded as Orthodox). Mendes, president of the Orthodox Union, therefore refused to join. He began to distinguish between the "Modern Orthodoxy" of himself and his peers in the OU, and "Conservatives" who tolerated what was beyond the pale for him. However, this first sign of institutionalization and separation was far from conclusive. Mendes himself could not clearly differentiate between the two groups, and many he viewed as Orthodox were members of the USoA. The epithets "Conservative" and "Orthodox" remained interchangeable for decades to come. JTS graduates served in OU congregations; many students of the Orthodox Rabbi Isaac Elchanan Theological Seminary and members of the OU's Rabbinical Council of America, or RCA, attended it. In 1926, RIETS and the JTS even negotiated a possible merger, though it was never materialized. Upon Schechter's death in 1915, the first generation of his disciples kept his non-sectarian legacy of striving for a united, traditional American Judaism. He was replaced by Cyrus Adler. The USoA grew rapidly as the Eastern European immigrant population slowly integrated. In 1923 it already had 150 affiliated communities, and 229 before 1930. Synagogues offered a more modernized ritual: English sermons, choir singing, late Friday evening services which tacitly acknowledging that most had to work until after the Sabbath began, and often mixed-gender seating. Men and women sat separately with no partition, and some houses of prayer already introduced family pews. Motivated by popular pressure and frowned upon by both RA and seminary faculty—in its own synagogue, the institute maintained a partition until 1983—this was becoming common among the OU as well. As both social conditions and apathy turned American Jews away from tradition (barely 20 per cent were attending prayers weekly), a young professor named Mordecai Kaplan promoted the idea of transforming the synagogue into a community center, a "Shul with a Pool", a policy which indeed stymied the tide somewhat. In 1927, the RA also established its own Committee of Jewish Law, entrusted with determining "halakhic" issues. Consisting of seven members, it was chaired by the traditionalist Rabbi Louis Ginzberg, who already distinguished himself in 1922, drafting a responsa that allowed to use grape juice rather than fermented wine for "Kiddush" on the background of Prohibition. Kaplan himself, who rose to become an influential and popular figure within the JTS, concluded that his fellow rabbis' ambiguity in matters of belief and the contradiction between full observance and critical study were untenable and hypocritical. He formulated his own approach of Judaism as a Civilization, rejecting the concept of Revelation and any supernatural belief in favour of a cultural-ethnic perception. While valuing received mores, he eventually suggested giving the past "a vote, not a veto". Though popular among students, Kaplan's nascent Reconstructionism was opposed by the new traditionalist Chancellor Louis Finkelstein, appointed in 1940, and a large majority among the faculty. Tensions within the JTS and RA grew. The Committee of Jewish Law consisted mainly of scholars who had little field experience, almost solely from the seminary's Talmudic department. They were greatly concerned with "halakhic" licitness and indifferent to the pressures exerted on the pulpit rabbis, who had to contend with an Americanized public which cared little for such considerations or for tradition in general. In 1935, the RA almost adopted a groundbreaking motion: Rabbi Louis Epstein offered a solution to the agunah predicament, a clause that would have had husbands appoint wives as their proxies to issue divorce. It was repealed under pressure from the Orthodox Union. As late as 1947, CJL Chair Rabbi Boaz Cohen, himself a historicist who argued that the Law evolved much through time, rebuked pulpit clergy who requested lenient or radical rulings, stating he and his peers were content to "progress in inches... Free setting up of new premises and the introduction of novel categories of ritual upon the basis of pure reason and thinking would be perilous, if not fatal, to the principles and continuity of Jewish Law." The boundaries between Orthodox and Conservative Judaism in America were institutionalized only in the aftermath of World War II. The 1940s saw the younger generation of JTS graduates less patient with the prudence of the CJL and Talmud faculty in face of popular demand. Kaplan's Reconstructionism, while its fully committed partisans were few, had much influence. The majority among recent alumni eschewed the epithet "Orthodox" and tended to employ "Conservative" exclusively. Succeeding Schechter's direct disciples who headed the RA, JTS and United Synagogue in the interwar period, a new strata of activist leaders was rising. Rabbi Robert Gordis, RA president in 1944–1946, represented the junior members in advocating more flexibility; Rabbi Jacob Agus, a RIETS graduate who joined the body only in 1945, clamored that "we need a law making body, not a law interpreting committee." Agus argued that the breach between the Jewish public and tradition was too wide to be bridged conventionally, and that the RA would always remain inferior to the Orthodox as long as it retained its policy of merely adopting lenient precedents in rabbinic literature. He offered to extensively apply the tool of "takkanah", rabbinic ordinance. In 1946, a committee chaired by Gordis issued the "Sabbath and Festival Prayerbook", the first clearly Conservative liturgy: references to the sacrificial cult were in the past tense instead of a petition for restoration, and it rephrased blessings such as "who hast made me according to thy will" for women to "who hast made me a woman". During the movement's national conference in Chicago, held 13–17 May 1948, the pulpit rabbis in the RA gained the upper hand. Spurned by Gordis, Agus and fellow leaders, They voted to reorganize the CJL into a Committee of Jewish Law and Standards, enfranchised to issue "takkanot" by a majority. Membership was conditioned on having experience as a congregational rabbi, and unseasoned JTS faculty were thus denied entrance. While the RA was asserting a Conservative distinctive identity, the seminary remained more cautious. Finkelstein opposed sectarianism and preferred the neutral epithet "traditional", later commenting that "Conservative Judaism is a gimmick to get Jews back to real Judaism". He and the very right-wing Talmud professor Saul Lieberman, who maintained ties with the Orthodox while also viewing them as obstructionist and ossified, dominated the JTS, providing a counterweight to the liberals in the Assembly. Kaplan, meanwhile, spent more time on consolidating his Society for Advancement of Judaism. Abraham Joshua Heschel, who espoused a mysticist understanding of Jewish religion, also became an important figure among the faculty. The CJLS now proceeded to demonstrate its independence. Sabbath was widely desecrated by a large majority of Jews, and the board believed arrival at synagogues should be encouraged. They therefore enacted an ordinance that allowed driving on the Sabbath (for worship alone) and the use of electricity. The driving responsum was later severely criticized by Conservative rabbis, and was charged with imparting the movement was overly keen to condone the laxity of congregants. It also signified the final break with the Orthodox, who were themselves being bolstered by more strictly observant immigrants from Europe. In 1954, the RCA reverted its 1948 ruling that allowed the use of microphones on Sabbath and festivals and declared that praying without a partition between sexes was banned. Though enforced slowly—in 1997, there were still seven OU congregations with no physical barrier, and so-called "Conservadox" remain extant—these two attributes became a demarcation line between Orthodox and Conservative synagogues. RA converts were denied ablution in Orthodox ritual baths, and rabbis from one movement would gradually cease serving in the other's communities. Rather than a force within American Judaism, the JTS-centered movement emerged as a third movement. The historicist and critical approach to "halakha", as well as other features, were emphasized by leaders eager to demonstrate their uniqueness. In their efforts to solidify a coherent identity, Conservative thinkers like Mordecai Waxman in his 1957 "Tradition and Change", ventured beyond Schechter's deem conceptions to Rabbi Zecharias Frankel and Breslau, presenting themselves as its direct inheritors via Alexander Kohut and others. The CJLS continued to issue groundbreaking ordinances and rulings. The postwar decades were a time of immense growth for the Conservative movement. Most of the 500,000 decommissioned Jewish GIs left the densely populated immigrant neighbourhoods of the East Coast, moving to suburbia. They were Americanized but still retained traditional sentiments, and Reform Judaism was too radical for most. The United Synagogue of America offered Jewish education for children and a familiar religious environment which was also comfortable and not strict. It expanded from 350 communities by 1945 to 832 by 1971, becoming the largest denomination, with some 350,000 dues-paying member households (1.5 million people) at synagogues and over 40 per cent of American Jewry identifying with it in polls, adding an estimated million more non-registered supporters. Already in a 1955 study, Marshall Sklare defined Conservative Judaism as the quintessential American Jewish movement, but stressed the gap between laity and clergy, noting "rabbis now recognize that they are not making decisions or writing responsa, but merely taking a poll of their membership." Most congregants, commented Edward S. Shapiro, were "Conservative Jews because their rabbi kept kosher and the Sabbath... Not because of their religious behavior." The movement established its presence outside the U.S. and Canada: In 1962, the young Rabbi Marshall Meyer founded the Seminario Rabinico Latinoamericano in Buenos Aires, which would serve as the basis for Conservative expansion in South America. In 1979, four communities formed the Israel Masorti Movement. Rabbi Louis Jacobs, dismissed in 1964 from the British Orthodox rabbinate on the charge of heresy after espousing a non-literal understanding of the Torah, joined with the Conservatives and founded his country's first Masorti community. The new branches were all united within the World Council of Synagogues, later to be named Masorti Olami. The movement peaked in numbers in the 1970s. During that decade, the tensions between the various elements within it intensified. The right wing, conservative in "halakhic" matters and often adhering to a verbal understanding of revelation, was dismayed by the failure to bolster observance among the laity and the resurgence of Orthodoxy. The left was influenced by the Reconstructionists, who formed their own seminary in 1968 and were slowly coalescing, as well as the growing appeal of Reform, which turned more traditional and threatened to sway congregants. While the rightists opposed further modifications, their left-wing peers demanded them. The Chavurah movement, consisting of nonaligned prayer quorums of young (and frequently, Conservative-raised) worshipers who sought a more intense religious experience, also weakened congregations. In 1972, the liberal wing gained an influential position with the appointment of Gerson D. Cohen as JTS Chancellor. During the same year, after Reform began to ordain female rabbis, a strong lobby rose to advocate the same. The CJLS rapidly enacted an ordinance which allowed women to be tallied for a "minyan", and by 1976 the percentage of synagogues allowing them to bless during the reading of the Torah grew from 7 per cent to 50 per cent. In 1979, ignoring the denominational leadership, Beth Israel Congregation of Chester County accepted the RRC-ordained Rabbi Linda Joy Holtzman. Pressures to allow women to assume rabbinical positions was mounting from the congregational level, though the RA agreed to delay any action until the JTS scholars would concur. Female ordination was a matter of great friction until 1983, when Rabbi Joel Roth devised a solution that entailed women voluntarily accepting the obligation to pray regularly. The leadership passed it not by scholarly consensus but via a popular vote of all JTS faculty, including non-specialists. Two years later, the first JTS-ordained female rabbi, Amy Eilberg, was admitted into the RA. David Weiss Halivni, professor of the Talmud faculty, claimed that Roth's method must have required waiting until a considerable number of women did prove sufficient commitment. He and his sympathizers regarded the vote as belying any claim to "halakhic" integrity. They formed the Union for Traditional Conservative Judaism in 1985, a right-wing lobby which numbered some 10,000 supporters from the Conservative observant elite. The UTJC withdrew from the movement and erased the word "Conservative" in 1990, attempting to merge with moderate Orthodox organizations. In the very same year, the Reconstructionist also seceded fully, joining the World Union for Progressive Judaism under observer status. The double defection narrowed the movement's spectrum of opinions, at a time when large swaths of congregants were abandoning in favour of Reform, which was more tolerant of intermarriage. RA leaders were engaged in introspection through the later 1980s, resulting in the 1988 "Emet ve-Emunah" platform, while Reform slowly bypassed them and became the largest American Jewish movement. After the issue of egalitarianism for women subsided, LGBT acceptance replaced it as the main source of contention between the declining right-wing wing and the liberal majority. A first attempt was rebuffed in 1992 by a harsh responsum written by Roth. The retirement of Chancellor Ismar Schorsch, a staunch opponent, allowed the CJLS to endorse a motion which still banned anal intercourse but not any other physical contact, and allowed the ordination of openly LGBT rabbis, in 2006. Roth and three other supporters resigned from the panel in protest, claiming the responsum was not valid; Masorti affiliates in South America, Israel and Hungary objected severely. The Seminario is yet to accept the resolution, while several Canadian congregations seceded from the United Synagogue in 2008 to form an independent union in protest of the slide to the left. Since the 2013 Pew survey, which assessed that only 18 per cent of American Jews identify with it, Conservative leadership is engaged in attempting to solve Conservative Judaism's demographic crisis.
https://en.wikipedia.org/wiki?curid=6623
Common Desktop Environment The Common Desktop Environment (CDE) is a desktop environment for Unix and OpenVMS, based on the Motif widget toolkit. It was part of the UNIX 98 Workstation Product Standard, and was for a long time the "classic" Unix desktop associated with commercial Unix workstations. After a long history as proprietary software, CDE was released as free software on August 6, 2012, under the GNU Lesser General Public License, version 2 or later. Since its release as free software, CDE has been ported to Linux and BSD derivatives. Hewlett-Packard, IBM, SunSoft, and USL announced CDE in June 1993 as a joint development within the Common Open Software Environment (COSE) initiative. Each development group contributed its own technology to CDE: After its release, HP endorsed CDE as the new standard desktop for Unix, and provided documentation and software for migrating HP VUE customizations to CDE. In March 1994 CDE became the responsibility of the "new OSF", a merger of the Open Software Foundation and Unix International; in September 1995, the merger of Motif and CDE into a single project, CDE/Motif, was announced. OSF became part of the newly formed Open Group in 1996. In February 1997, the Open Group released their last major version of CDE, version 2.1. Red Hat Linux was the only Linux distribution that proprietary CDE was ported to. In 1997, Red Hat began offering a version of CDE licensed from TriTeal Corporation. In 1998, Xi Graphics, a company specializing in the X Windowing System, offered a version of CDE bundled with Red Hat Linux, called "Xi Graphics maXimum cde/OS". These were phased out, and Red Hat moved to the GNOME desktop. Until about 2000, users of Unix desktops regarded CDE as the "de facto" standard, but at that time, other desktop environments such as GNOME and K Desktop Environment 2 were quickly becoming mature, and became widespread on Linux systems. In 2001, Sun Microsystems announced that they would phase out CDE as the standard desktop environment in Solaris in favor of GNOME. Solaris 10, released in early 2005, includes both CDE and the GNOME-based Java Desktop System. The OpenSolaris project, begun around the same time, did not include CDE, and had no intent to make Solaris CDE available as open-source. The original release of Solaris 11 in November 2011 only contained GNOME as standard desktop, though some CDE libraries, such as Motif and ToolTalk, remained for binary compatibility but Oracle Solaris 11.4, released in August 2018, removed support for the CDE runtime environment and background services. From its launch until 2012, CDE was proprietary software. Motif, the toolkit on which CDE is built, was released by The Open Group in 2000 as "Open Motif," under a "revenue sharing" license. That license did not meet either the open source or free software definitions. The Open Group had wished to make Motif open source, but did not succeed doing so at that time. In 2006, a petition was created asking The Open Group to release the source code for CDE and Motif under a free license. On August 6, 2012, CDE was open-sourced under the LGPL free software license. The CDE source code was then released to SourceForge. The free software project OpenCDE had been started in 2010 to reproduce the look and feel, organization, and feature set of CDE.
https://en.wikipedia.org/wiki?curid=6627
Children of Dune Children of Dune is a 1976 science fiction novel by Frank Herbert, the third in his "Dune" series of six novels. It was originally serialized in "Analog Science Fiction and Fact" in 1976, and was the last "Dune" novel to be serialized before book publication. At the end of "Dune Messiah", Paul Atreides walks into the desert, a blind man, leaving his twin children Leto and Ghanima in the care of the Fremen, while his sister Alia rules the universe as regent. Awakened in the womb by the spice, the children are the heirs to Paul's prescient vision of the fate of the universe, a role that Alia desperately craves. House Corrino schemes to return to the throne, while the Bene Gesserit make common cause with the Tleilaxu and Spacing Guild to gain control of the spice and the children of Paul Atreides. Initially selling over 75,000 copies, it became the first hardcover best-seller ever in the science fiction field. The novel was critically well-received for its gripping plot, action, and atmosphere, and was nominated for the Hugo Award for Best Novel in 1977. "Dune Messiah" (1969) and "Children of Dune" were collectively adapted by the Sci-Fi Channel in 2003 into a miniseries entitled "Frank Herbert's Children of Dune". Nine years after Emperor Paul Muad'Dib walked into the desert, blind, the ecological transformation of Dune has reached the point where some Fremen are living without stillsuits in the less arid climate and have started to move out of the sietches and into the villages and cities. As the old ways erode, more and more pilgrims arrive to experience the planet of Muad'Dib. The Imperial high council has lost its political might and is powerless to control the Jihad. Paul's young twin children, Leto II and Ghanima, have concluded that their guardian Alia has succumbed to Abomination—possession by her grandfather Baron Vladimir Harkonnen—and fear that a similar fate awaits them. They (and Alia) also realize that the terraforming of Dune will kill all the sandworms, thus destroying the source of the spice, but Harkonnen desires this outcome. Leto also fears that, like his father, he will become trapped by his prescience. Meanwhile, a new religious figure called "The Preacher" has risen in the desert, railing against the religious government's injustices and the changes among the Fremen. Some Fremen believe he is Paul Atreides. Princess Wensicia of the fallen House Corrino on Salusa Secundus plots to assassinate the twins and regain power for her House. Lady Jessica, Alia and Paul's mother, returns to Arrakis and recognizes that her daughter is possessed, but finds no signs of Abomination in the twins. Leto arranges for Fremen leader Stilgar to protect his sister if there is an attempt on their lives. The Preacher journeys to Salusa Secundus to meet Wensicia's son Farad'n, and in return pledges Duncan Idaho as an agent of House Corrino. Alia attempts to assassinate Jessica, who escapes into the desert with Duncan's help, precipitating a rebellion among the Fremen. The twins anticipate and survive the Corrino assassination plot. Leto leaves to seek out the Preacher, while Ghanima, changing her memory with self-hypnosis, reports (and believes) that her brother has been murdered. Duncan and Jessica flee to Salusa Secundus, where Jessica begins to mentor Farad'n. He seizes power from his regent mother Wensicia and allies with the Bene Gesserit, who promise to marry him to Ghanima and support his bid to become Emperor. A band of Fremen outlaws capture Leto and force him to undergo the spice trance at the suggestion of Gurney Halleck, who has infiltrated the group on Jessica's orders. Leto's spice-induced visions show him a myriad of possible futures where humanity becomes extinct and only one where it survives. He names this future "The Golden Path" and resolves to bring it to fruition—something that his father, who had already glimpsed this future, refused to do. He escapes his captors and sacrifices his humanity in pursuit of the Golden Path by physically fusing with a school of sandtrout, gaining superhuman strength and near-invulnerability. He travels across the desert and confronts the Preacher, who is indeed Paul. Duncan returns to Arrakis and provokes Stilgar into killing him so that Stilgar is forced to take Ghanima and go into hiding. Alia recaptures Ghanima and arranges her marriage to Farad'n, planning to exploit the expected chaos when Ghanima kills him to avenge her brother's murder. The Preacher and Leto return to the capital to confront Alia. Upon arriving, Paul is murdered, to Alia’s horror. Leto reveals himself in a display of superhuman strength and triggers the return of Ghanima's genuine memories. He confronts Alia and offers to help her overcome her possession, but Harkonnen resists. Alia manages to commit suicide by throwing herself off a high balcony. Leto declares himself Emperor and asserts control over the Fremen. Farad'n enlists in his service and delivers control of the Corrino armies. Leto marries his sister Ghanima to further his goals, but Farad'n is her true consort so the Atreides line can continue. Parts of "Dune Messiah" and "Children of Dune" were written before "Dune" was completed. "Children of Dune" was originally serialized in "Analog Science Fiction and Fact" in 1976, and was the last "Dune" novel to be serialized before book publication. "Dune Messiah" and "Children of Dune" were published in one volume by the Science Fiction Book Club in 2002. Herbert likened the initial trilogy of novels ("Dune", "Dune Messiah", and "Children of Dune") to a fugue, and while "Dune" was a heroic melody, "Dune Messiah" was its inversion. Paul rises to power in "Dune" by seizing control of the single critical resource in the universe, melange. His enemies are dead or overthrown, and he is set to take the reins of power and bring a hard but enlightened peace to the universe. Herbert chose in the books that followed to undermine Paul's triumph with a string of failures and philosophical paradoxes. Initially selling over 75,000 copies, "Children of Dune" became the first hardcover best-seller ever in the science fiction field. The novel was critically well-received for its gripping plot, action, and atmosphere, and was nominated for the Hugo Award for Best Novel in 1977. The "Los Angeles Times" called "Children of Dune" "a major event", and "Challenging Destiny" noted that "Herbert adds enough new twists and turns to the ongoing saga that familiarity with the recurring elements brings pleasure." "Publishers Weekly" wrote, "Ranging from palace intrigue and desert chases to religious speculation and confrontations with the supreme intelligence of the universe, there is something here for all science fiction fans." In a 1976 review, Spider Robinson found "Children of Dune" unsatisfying, faulting the ending as unconvincing and thematically overfamiliar. The novel is referenced in "A Thousand Plateaus" (1980) by Gilles Deleuze and Félix Guattari. David Pringle gave the novel a rating of two stars out of four and described the novel as "dark and convoluted stuff." "Dune Messiah" (1969) and "Children of Dune" were collectively adapted by the Sci-Fi Channel in 2003 into a miniseries entitled "Frank Herbert's Children of Dune". The three-part, six-hour miniseries covers the bulk of the plot of "Dune Messiah" in the first installment, and adapts "Children of Dune" in the second and third parts.
https://en.wikipedia.org/wiki?curid=6628
Candide ' ( , ) is a French satire first published in 1759 by Voltaire, a philosopher of the Age of Enlightenment. The novella has been widely translated, with English versions titled Candide: or, All for the Best (1759); Candide: or, The Optimist (1762); and Candide: Optimism""' (1947). It begins with a young man, Candide, who is living a sheltered life in an Edenic paradise and being indoctrinated with Leibnizian optimism by his mentor, Professor Pangloss. The work describes the abrupt cessation of this lifestyle, followed by Candide's slow and painful disillusionment as he witnesses and experiences great hardships in the world. Voltaire concludes Candide with, if not rejecting Leibnizian optimism outright, advocating a deeply practical precept, "we must cultivate our garden", in lieu of the Leibnizian mantra of Pangloss, "all is for the best" in the "best of all possible worlds". "Candide" is characterized by its tone as well as by its erratic, fantastical, and fast-moving plot. A picaresque novel with a story similar to that of a more serious coming-of-age narrative ("Bildungsroman"), it parodies many adventure and romance clichés, the struggles of which are caricatured in a tone that is bitter and matter-of-fact. Still, the events discussed are often based on historical happenings, such as the Seven Years' War and the 1755 Lisbon earthquake. As philosophers of Voltaire's day contended with the problem of evil, so does Candide in this short theological novel, albeit more directly and humorously. Voltaire ridicules religion, theologians, governments, armies, philosophies, and philosophers. Through "Candide", he assaults Leibniz and his optimism. "Candide" has enjoyed both great success and great scandal. Immediately after its secretive publication, the book was widely banned to the public because it contained religious blasphemy, political sedition, and intellectual hostility hidden under a thin veil of naïveté. However, with its sharp wit and insightful portrayal of the human condition, the novel has since inspired many later authors and artists to mimic and adapt it. Today, "Candide" is recognized as Voltaire's "magnum opus" and is often listed as part of the Western canon. It is among the most frequently taught works of French literature. The British poet and literary critic Martin Seymour-Smith listed "Candide" as one of the 100 most influential books ever written. A number of historical events inspired Voltaire to write "Candide", most notably the publication of Leibniz's "Monadology", a short metaphysical treatise, the Seven Years' War, and the 1755 Lisbon earthquake. Both of the latter catastrophes are frequently referred to in "Candide" and are cited by scholars as reasons for its composition. The 1755 Lisbon earthquake, tsunami, and resulting fires of All Saints' Day, had a strong influence on theologians of the day and on Voltaire, who was himself disillusioned by them. The earthquake had an especially large effect on the contemporary doctrine of optimism, a philosophical system which implies that such events should not occur. Optimism is founded on the theodicy of Gottfried Wilhelm Leibniz and says all is for the best because God is a benevolent deity. This concept is often put into the form, "all is for the best in the best of all possible worlds" (). Philosophers had trouble fitting the horrors of this earthquake into their optimistic world view. Voltaire actively rejected Leibnizian optimism after the natural disaster, convinced that if this were the best possible world, it should surely be better than it is. In both "Candide" and ("Poem on the Lisbon Disaster"), Voltaire attacks this optimist belief. He makes use of the Lisbon earthquake in both "Candide" and his to argue this point, sarcastically describing the catastrophe as one of the most horrible disasters "in the best of all possible worlds". Immediately after the earthquake, unreliable rumours circulated around Europe, sometimes overestimating the severity of the event. Ira Wade, a noted expert on Voltaire and "Candide", has analyzed which sources Voltaire might have referenced in learning of the event. Wade speculates that Voltaire's primary source for information on the Lisbon earthquake was the 1755 work by Ange Goudar. Apart from such events, contemporaneous stereotypes of the German personality may have been a source of inspiration for the text, as they were for , a 1669 satirical picaresque novel written by Hans Jakob Christoffel von Grimmelshausen and inspired by the Thirty Years' War. The protagonist of this novel, who was supposed to embody stereotypically German characteristics, is quite similar to the protagonist of "Candide". These stereotypes, according to Voltaire biographer Alfred Owen Aldridge, include "extreme credulousness or sentimental simplicity", two of Candide's and Simplicius's defining qualities. Aldridge writes, "Since Voltaire admitted familiarity with fifteenth-century German authors who used a bold and buffoonish style, it is quite possible that he knew as well." A satirical and parodic precursor of "Candide", Jonathan Swift's "Gulliver's Travels" (1726) is one of "Candide"s closest literary relatives. This satire tells the story of "a gullible ingenue", Gulliver, who (like Candide) travels to several "remote nations" and is hardened by the many misfortunes which befall him. As evidenced by similarities between the two books, Voltaire probably drew upon "Gulliver's Travels" for inspiration while writing "Candide". Other probable sources of inspiration for "Candide" are (1699) by François Fénelon and (1753) by Louis-Charles Fougeret de Monbron. "Candide"s parody of the is probably based on , which includes the prototypical parody of the tutor on whom Pangloss may have been partly based. Likewise, Monbron's protagonist undergoes a disillusioning series of travels similar to those of Candide. Born François-Marie Arouet, Voltaire (1694–1778), by the time of the Lisbon earthquake, was already a well-established author, known for his satirical wit. He had been made a member of the Académie Française in 1746. He was a deist, a strong proponent of religious freedom, and a critic of tyrannical governments. "Candide" became part of his large, diverse body of philosophical, political and artistic works expressing these views. More specifically, it was a model for the eighteenth- and early nineteenth-century novels called the "contes philosophiques". This genre, of which Voltaire was one of the founders, included previous works of his such as "Zadig" and "Micromegas". It is unknown exactly when Voltaire wrote "Candide", but scholars estimate that it was primarily composed in late 1758 and begun as early as 1757. Voltaire is believed to have written a portion of it while living at Les Délices near Geneva and also while visiting Charles Théodore, the Elector-Palatinate at Schwetzingen, for three weeks in the summer of 1758. Despite solid evidence for these claims, a popular legend persists that Voltaire wrote "Candide" in three days. This idea is probably based on a misreading of the 1885 work by Lucien Perey (real name: Clara Adèle Luce Herpin) and Gaston Maugras. The evidence indicates strongly that Voltaire did not rush or improvise "Candide", but worked on it over a significant period of time, possibly even a whole year. "Candide" is mature and carefully developed, not impromptu, as the intentionally choppy plot and the aforementioned myth might suggest. There is only one extant manuscript of "Candide" that was written before the work's 1759 publication; it was discovered in 1956 by Wade and since named the "La Vallière Manuscript". It is believed to have been sent, chapter by chapter, by Voltaire to the Duke and Duchess La Vallière in the autumn of 1758. The manuscript was sold to the Bibliothèque de l'Arsenal in the late eighteenth century, where it remained undiscovered for almost two hundred years. The "La Vallière Manuscript", the most original and authentic of all surviving copies of "Candide", was probably dictated by Voltaire to his secretary, Jean-Louis Wagnière, then edited directly. In addition to this manuscript, there is believed to have been another, one copied by Wagnière for the Elector Charles-Théodore, who hosted Voltaire during the summer of 1758. The existence of this copy was first postulated by Norman L. Torrey in 1929. If it exists, it remains undiscovered. Voltaire published "Candide" simultaneously in five countries no later than 15 January 1759, although the exact date is uncertain. Seventeen versions of "Candide" from 1759, in the original French, are known today, and there has been great controversy over which is the earliest. More versions were published in other languages: "Candide" was translated once into Italian and thrice into English that same year. The complicated science of calculating the relative publication dates of all of the versions of "Candide" is described at length in Wade's article "The First Edition of "Candide": A Problem of Identification". The publication process was extremely secretive, probably the "most clandestine work of the century", because of the book's obviously illicit and irreverent content. The greatest number of copies of "Candide" were published concurrently in Geneva by Cramer, in Amsterdam by Marc-Michel Rey, in London by Jean Nourse, and in Paris by Lambert. "Candide" underwent one major revision after its initial publication, in addition to some minor ones. In 1761, a version of "Candide" was published that included, along with several minor changes, a major addition by Voltaire to the twenty-second chapter, a section that had been thought weak by the Duke of Vallière. The English title of this edition was "Candide, or Optimism, Translated from the German of Dr. Ralph. With the additions found in the Doctor's pocket when he died at Minden, in the Year of Grace 1759." The last edition of "Candide" authorised by Voltaire was the one included in Cramer's 1775 edition of his complete works, known as , in reference to the border or frame around each page. Voltaire strongly opposed the inclusion of illustrations in his works, as he stated in a 1778 letter to the writer and publisher Charles Joseph Panckoucke: Despite this protest, two sets of illustrations for "Candide" were produced by the French artist Jean-Michel Moreau le Jeune. The first version was done, at Moreau's own expense, in 1787 and included in Kehl's publication of that year, "Oeuvres Complètes de Voltaire". Four images were drawn by Moreau for this edition and were engraved by Pierre-Charles Baquoy. The second version, in 1803, consisted of seven drawings by Moreau which were transposed by multiple engravers. The twentieth-century modern artist Paul Klee stated that it was while reading "Candide" that he discovered his own artistic style. Klee illustrated the work, and his drawings were published in a 1920 version edited by Kurt Wolff. "Candide" contains thirty episodic chapters, which may be grouped into two main schemes: one consists of two divisions, separated by the protagonist's hiatus in El Dorado; the other consists of three parts, each defined by its geographical setting. By the former scheme, the first half of "Candide" constitutes the rising action and the last part the resolution. This view is supported by the strong theme of travel and quest, reminiscent of adventure and picaresque novels, which tend to employ such a dramatic structure. By the latter scheme, the thirty chapters may be grouped into three parts each comprising ten chapters and defined by locale: I–X are set in Europe, XI–XX are set in the Americas, and XXI–XXX are set in Europe and the Ottoman Empire. The plot summary that follows uses this second format and includes Voltaire's additions of 1761. The tale of "Candide" begins in the castle of the Baron Thunder-ten-Tronckh in Westphalia, home to the Baron's daughter, Lady Cunégonde; his bastard nephew, Candide; a tutor, Pangloss; a chambermaid, Paquette; and the rest of the Baron's family. The protagonist, Candide, is romantically attracted to Cunégonde. He is a young man of "the most unaffected simplicity" (), whose face is "the true index of his mind" (). Dr. Pangloss, professor of "" (English: "metaphysico-theologo-cosmolonigology") and self-proclaimed optimist, teaches his pupils that they live in the "best of all possible worlds" and that "all is for the best". All is well in the castle until Cunégonde sees Pangloss sexually engaged with Paquette in some bushes. Encouraged by this show of affection, Cunégonde drops her handkerchief next to Candide, enticing him to kiss her. For this infraction, Candide is evicted from the castle, at which point he is captured by Bulgar (Prussian) recruiters and coerced into military service, where he is flogged, nearly executed, and forced to participate in a major battle between the Bulgars and the Avars (an allegory representing the Prussians and the French). Candide eventually escapes the army and makes his way to Holland where he is given aid by Jacques, an Anabaptist, who strengthens Candide's optimism. Soon after, Candide finds his master Pangloss, now a beggar with syphilis. Pangloss reveals he was infected with this disease by Paquette and shocks Candide by relating how Castle Thunder-ten-Tronckh was destroyed by Bulgars, that Cunégonde and her whole family were killed, and that Cunégonde was raped before her death. Pangloss is cured of his illness by Jacques, losing one eye and one ear in the process, and the three set sail to Lisbon. In Lisbon's harbor, they are overtaken by a vicious storm which destroys the boat. Jacques attempts to save a sailor, and in the process is thrown overboard. The sailor makes no move to help the drowning Jacques, and Candide is in a state of despair until Pangloss explains to him that Lisbon harbor was created in order for Jacques to drown. Only Pangloss, Candide, and the "brutish sailor" who let Jacques drown survive the wreck and reach Lisbon, which is promptly hit by an earthquake, tsunami and fire that kill tens of thousands. The sailor leaves in order to loot the rubble while Candide, injured and begging for help, is lectured on the optimistic view of the situation by Pangloss. The next day, Pangloss discusses his optimistic philosophy with a member of the Portuguese Inquisition, and he and Candide are arrested for heresy, set to be tortured and killed in an "" set up to appease God and prevent another disaster. Candide is flogged and sees Pangloss hanged, but another earthquake intervenes and he escapes. He is approached by an old woman, who leads him to a house where Lady Cunégonde waits, alive. Candide is surprised: Pangloss had told him that Cunégonde had been raped and disemboweled. She had been, but Cunégonde points out that people survive such things. However, her rescuer sold her to a Jewish merchant, Don Issachar, who was then threatened by a corrupt Grand Inquisitor into sharing her (Don Issachar gets Cunégonde on Mondays, Wednesdays, and the sabbath day). Her owners arrive, find her with another man, and Candide kills them both. Candide and the two women flee the city, heading to the Americas. Along the way, Cunégonde falls into self-pity, complaining of all the misfortunes that have befallen her. The old woman reciprocates by revealing her own tragic life: born the daughter of Pope Urban X and the Princess of Palestrina, she was raped and enslaved by African pirates, witnessed violent civil wars in Morocco under the bloodthirsty King Moulay Ismaïl (during which her mother was drawn and quartered), suffered further slavery and famine, nearly died from a plague in Algiers, and had a buttock cut off to feed starving Janissaries during the Russian siege of Azov. After traversing all the Russian Empire, she eventually became a servant of Don Issachar and met Cunégonde. The trio arrives in Buenos Aires, where Governor Don Fernando d'Ibarra y Figueroa y Mascarenes y Lampourdos y Souza asks to marry Cunégonde. Just then, an alcalde (a Spanish fortress commander) arrives, pursuing Candide for killing the Grand Inquisitor. Leaving the women behind, Candide flees to Paraguay with his practical and heretofore unmentioned manservant, Cacambo. At a border post on the way to Paraguay, Cacambo and Candide speak to the commandant, who turns out to be Cunégonde's unnamed brother. He explains that after his family was slaughtered, the Jesuits' preparation for his burial revived him, and he has since joined the order. When Candide proclaims he intends to marry Cunégonde, her brother attacks him, and Candide runs him through with his rapier. After lamenting all the people (mainly priests) he has killed, he and Cacambo flee. In their flight, Candide and Cacambo come across two naked women being chased and bitten by a pair of monkeys. Candide, seeking to protect the women, shoots and kills the monkeys, but is informed by Cacambo that the monkeys and women were probably lovers. Cacambo and Candide are captured by Oreillons, or Orejones; members of the Inca nobility who widened the lobes of their ears, and are depicted here as the fictional inhabitants of the area. Mistaking Candide for a Jesuit by his robes, the Oreillons prepare to cook Candide and Cacambo; however, Cacambo convinces the Oreillons that Candide killed a Jesuit to procure the robe. Cacambo and Candide are released and travel for a month on foot and then down a river by canoe, living on fruits and berries. After a few more adventures, Candide and Cacambo wander into El Dorado, a geographically isolated utopia where the streets are covered with precious stones, there exist no priests, and all of the king's jokes are funny. Candide and Cacambo stay a month in El Dorado, but Candide is still in pain without Cunégonde, and expresses to the king his wish to leave. The king points out that this is a foolish idea, but generously helps them do so. The pair continue their journey, now accompanied by one hundred red pack sheep carrying provisions and incredible sums of money, which they slowly lose or have stolen over the next few adventures. Candide and Cacambo eventually reach Suriname, where they split up: Cacambo travels to Buenos Aires to retrieve Lady Cunégonde, while Candide prepares to travel to Europe to await the two. Candide's remaining sheep are stolen, and Candide is fined heavily by a Dutch magistrate for petulance over the theft. Before leaving Suriname, Candide feels in need of companionship, so he interviews a number of local men who have been through various ill-fortunes and settles on a man named Martin. This companion, Martin, is a Manichaean scholar based on the real-life pessimist Pierre Bayle, who was a chief opponent of Leibniz. For the remainder of the voyage, Martin and Candide argue about philosophy, Martin painting the entire world as occupied by fools. Candide, however, remains an optimist at heart, since it is all he knows. After a detour to Bordeaux and Paris, they arrive in England and see an admiral (based on Admiral Byng) being shot for not killing enough of the enemy. Martin explains that Britain finds it necessary to shoot an admiral from time to time ""pour l'encouragement des autres"" (to encourage the others). Candide, horrified, arranges for them to leave Britain immediately. Upon their arrival in Venice, Candide and Martin meet Paquette, the chambermaid who infected Pangloss with his syphilis, in Venice. She is now a prostitute, and is spending her time with a Theatine monk, Brother Giroflée. Although both appear happy on the surface, they reveal their despair: Paquette has led a miserable existence as a sexual object, and the monk detests the religious order in which he was indoctrinated. Candide gives two thousand piastres to Paquette and one thousand to Brother Giroflée. Candide and Martin visit the Lord Pococurante, a noble Venetian. That evening, Cacambo—now a slave—arrives and informs Candide that Cunégonde is in Constantinople. Prior to their departure, Candide and Martin dine with six strangers who had come for Carnival of Venice. These strangers are revealed to be dethroned kings: the Ottoman Sultan Ahmed III, Emperor Ivan VI of Russia, Charles Edward Stuart (an unsuccessful pretender to the English throne), Augustus III of Poland, Stanisław Leszczyński, and Theodore of Corsica. On the way to Constantinople, Cacambo reveals that Cunégonde—now horribly ugly—currently washes dishes on the banks of the Propontis as a slave for a Transylvanian prince by the name of Rákóczi. After arriving at the Bosphorus, they board a galley where, to Candide's surprise, he finds Pangloss and Cunégonde's brother among the rowers. Candide buys their freedom and further passage at steep prices. The baron and Pangloss relate how they survived, but despite the horrors he has been through, Pangloss's optimism remains unshaken: "I still hold to my original opinions, because, after all, I'm a philosopher, and it wouldn't be proper for me to recant, since Leibniz cannot be wrong, and since pre-established harmony is the most beautiful thing in the world, along with the plenum and subtle matter." Candide, the baron, Pangloss, Martin, and Cacambo arrive at the banks of the Propontis, where they rejoin Cunégonde and the old woman. Cunégonde has indeed become hideously ugly, but Candide nevertheless buys their freedom and marries Cunégonde to spite her brother, who forbids Cunégonde from marrying anyone but a baron of the Empire (he is secretly sold back into slavery). Paquette and Brother Giroflée—having squandered their three thousand piastres—are reconciled with Candide on a small farm () which he just bought with the last of his finances. One day, the protagonists seek out a dervish known as a great philosopher of the land. Candide asks him why Man is made to suffer so, and what they all ought to do. The dervish responds by asking rhetorically why Candide is concerned about the existence of evil and good. The dervish describes human beings as mice on a ship sent by a king to Egypt; their comfort does not matter to the king. The dervish then slams his door on the group. Returning to their farm, Candide, Pangloss, and Martin meet a Turk whose philosophy is to devote his life only to simple work and not concern himself with external affairs. He and his four children cultivate a small area of land, and the work keeps them "free of three great evils: boredom, vice, and poverty." Candide, Pangloss, Martin, Cunégonde, Paquette, Cacambo, the old woman, and Brother Giroflée all set to work on this "commendable plan" () on their farm, each exercising his or her own talents. Candide ignores Pangloss's insistence that all turned out for the best by necessity, instead telling him "we must cultivate our garden" (). As Voltaire himself described it, the purpose of "Candide" was to "bring amusement to a small number of men of wit". The author achieves this goal by combining his sharp wit with a fun parody of the classic adventure-romance plot. Candide is confronted with horrible events described in painstaking detail so often that it becomes humorous. Literary theorist Frances K. Barasch described Voltaire's matter-of-fact narrative as treating topics such as mass death "as coolly as a weather report". The fast-paced and improbable plot—in which characters narrowly escape death repeatedly, for instance—allows for compounding tragedies to befall the same characters over and over again. In the end, "Candide" is primarily, as described by Voltaire's biographer Ian Davidson, "short, light, rapid and humorous". Behind the playful façade of "Candide" which has amused so many, there lies very harsh criticism of contemporary European civilization which angered many others. European governments such as France, Prussia, Portugal and England are each attacked ruthlessly by the author: the French and Prussians for the Seven Years' War, the Portuguese for their Inquisition, and the British for the execution of John Byng. Organised religion, too, is harshly treated in "Candide". For example, Voltaire mocks the Jesuit order of the Roman Catholic Church. Aldridge provides a characteristic example of such anti-clerical passages for which the work was banned: while in Paraguay, Cacambo remarks, "[The Jesuits] are masters of everything, and the people have no money at all …". Here, Voltaire suggests the Christian mission in Paraguay is taking advantage of the local population. Voltaire depicts the Jesuits holding the indigenous peoples as slaves while they claim to be helping them. The main method of "Candide"s satire is to contrast ironically great tragedy and comedy. The story does not invent or exaggerate evils of the world—it displays real ones starkly, allowing Voltaire to simplify subtle philosophies and cultural traditions, highlighting their flaws. Thus "Candide" derides optimism, for instance, with a deluge of horrible, historical (or at least plausible) events with no apparent redeeming qualities. A simple example of the satire of "Candide" is seen in the treatment of the historic event witnessed by Candide and Martin in Portsmouth harbour. There, the duo spy an anonymous admiral, supposed to represent John Byng, being executed for failing to properly engage a French fleet. The admiral is blindfolded and shot on the deck of his own ship, merely "to encourage the others" (, an expression Voltaire is credited with originating). This depiction of military punishment trivializes Byng's death. The dry, pithy explanation "to encourage the others" thus satirises a serious historical event in characteristically Voltairian fashion. For its classic wit, this phrase has become one of the more often quoted from "Candide". Voltaire depicts the worst of the world and his pathetic hero's desperate effort to fit it into an optimistic outlook. Almost all of "Candide" is a discussion of various forms of evil: its characters rarely find even temporary respite. There is at least one notable exception: the episode of El Dorado, a fantastic village in which the inhabitants are simply rational, and their society is just and reasonable. The positivity of El Dorado may be contrasted with the pessimistic attitude of most of the book. Even in this case, the bliss of El Dorado is fleeting: Candide soon leaves the village to seek Cunégonde, whom he eventually marries only out of a sense of obligation. Another element of the satire focuses on what William F. Bottiglia, author of many published works on "Candide", calls the "sentimental foibles of the age" and Voltaire's attack on them. Flaws in European culture are highlighted as "Candide" parodies adventure and romance clichés, mimicking the style of a picaresque novel. A number of archetypal characters thus have recognisable manifestations in Voltaire's work: Candide is supposed to be the drifting rogue of low social class, Cunégonde the sex interest, Pangloss the knowledgeable mentor and Cacambo the skilful valet. As the plot unfolds, readers find that Candide is no rogue, Cunégonde becomes ugly and Pangloss is a stubborn fool. The characters of "Candide" are unrealistic, two-dimensional, mechanical, and even marionette-like; they are simplistic and stereotypical. As the initially naïve protagonist eventually comes to a mature conclusion—however noncommittal—the novella is a "bildungsroman", if not a very serious one. Gardens are thought by many critics to play a critical symbolic role in "Candide". The first location commonly identified as a garden is the castle of the Baron, from which Candide and Cunégonde are evicted much in the same fashion as Adam and Eve are evicted from the Garden of Eden in the Book of Genesis. Cyclically, the main characters of "Candide" conclude the novel in a garden of their own making, one which might represent celestial paradise. The third most prominent "garden" is El Dorado, which may be a false Eden. Other possibly symbolic gardens include the Jesuit pavilion, the garden of Pococurante, Cacambo's garden, and the Turk's garden. These gardens are probably references to the Garden of Eden, but it has also been proposed, by Bottiglia, for example, that the gardens refer also to the "Encyclopédie", and that Candide's conclusion to cultivate "his garden" symbolises Voltaire's great support for this endeavour. Candide and his companions, as they find themselves at the end of the novella, are in a very similar position to Voltaire's tightly knit philosophical circle which supported the : the main characters of "Candide" live in seclusion to "cultivate [their] garden", just as Voltaire suggested his colleagues leave society to write. In addition, there is evidence in the epistolary correspondence of Voltaire that he had elsewhere used the metaphor of gardening to describe writing the . Another interpretative possibility is that Candide cultivating "his garden" suggests his engaging in only necessary occupations, such as feeding oneself and fighting boredom. This is analogous to Voltaire's own view on gardening: he was himself a gardener at his estates in Les Délices and Ferney, and he often wrote in his correspondence that gardening was an important pastime of his own, it being an extraordinarily effective way to keep busy. "Candide" satirises various philosophical and religious theories that Voltaire had previously criticised. Primary among these is Leibnizian optimism (sometimes called "Panglossianism" after its fictional proponent), which Voltaire ridicules with descriptions of seemingly endless calamity. Voltaire demonstrates a variety of irredeemable evils in the world, leading many critics to contend that Voltaire's treatment of evil—specifically the theological problem of its existence—is the focus of the work. Heavily referenced in the text are the Lisbon earthquake, disease, and the sinking of ships in storms. Also, war, thievery, and murder—evils of human design—are explored as extensively in "Candide" as are environmental ills. Bottiglia notes Voltaire is "comprehensive" in his enumeration of the world's evils. He is unrelenting in attacking Leibnizian optimism. Fundamental to Voltaire's attack is Candide's tutor Pangloss, a self-proclaimed follower of Leibniz and a teacher of his doctrine. Ridicule of Pangloss's theories thus ridicules Leibniz himself, and Pangloss's reasoning is silly at best. For example, Pangloss's first teachings of the narrative absurdly mix up cause and effect: Following such flawed reasoning even more doggedly than Candide, Pangloss defends optimism. Whatever their horrendous fortune, Pangloss reiterates "all is for the best" ("") and proceeds to "justify" the evil event's occurrence. A characteristic example of such theodicy is found in Pangloss's explanation of why it is good that syphilis exists: Candide, the impressionable and incompetent student of Pangloss, often tries to justify evil, fails, invokes his mentor and eventually despairs. It is by these failures that Candide is painfully cured (as Voltaire would see it) of his optimism. This critique of Voltaire's seems to be directed almost exclusively at Leibnizian optimism. "Candide" does not ridicule Voltaire's contemporary Alexander Pope, a later optimist of slightly different convictions. "Candide" does not discuss Pope's optimistic principle that "all is right", but Leibniz's that states, "this is the best of all possible worlds". However subtle the difference between the two, "Candide" is unambiguous as to which is its subject. Some critics conjecture that Voltaire meant to spare Pope this ridicule out of respect, although Voltaire's "Poème" may have been written as a more direct response to Pope's theories. This work is similar to "Candide" in subject matter, but very different from it in style: the "Poème" embodies a more serious philosophical argument than "Candide". The conclusion of the novel, in which Candide finally dismisses his tutor's optimism, leaves unresolved what philosophy the protagonist is to accept in its stead. This element of "Candide" has been written about voluminously, perhaps above all others. The conclusion is enigmatic and its analysis is contentious. Voltaire develops no formal, systematic philosophy for the characters to adopt. The conclusion of the novel may be thought of not as a philosophical alternative to optimism, but as a prescribed practical outlook (though it prescribes is in dispute). Many critics have concluded that one minor character or another is portrayed as having the right philosophy. For instance, a number believe that Martin is treated sympathetically, and that his character holds Voltaire's ideal philosophy—pessimism. Others disagree, citing Voltaire's negative descriptions of Martin's principles and the conclusion of the work in which Martin plays little part. Within debates attempting to decipher the conclusion of "Candide" lies another primary "Candide" debate. This one concerns the degree to which Voltaire was advocating a pessimistic philosophy, by which Candide and his companions give up hope for a better world. Critics argue that the group's reclusion on the farm signifies Candide and his companions' loss of hope for the rest of the human race. This view is to be compared to a reading that presents Voltaire as advocating a melioristic philosophy and a precept committing the travellers to improving the world through metaphorical gardening. This debate, and others, focuses on the question of whether or not Voltaire was prescribing passive retreat from society, or active industrious contribution to it. Separate from the debate about the text's conclusion is the "inside/outside" controversy. This argument centers on the matter of whether or not Voltaire was actually prescribing anything. Roy Wolper, professor emeritus of English, argues in a revolutionary 1969 paper that "Candide" does not necessarily speak for its author; that the work should be viewed as a narrative independent of Voltaire's history; and that its message is entirely (or mostly) it. This point of view, the "inside", specifically rejects attempts to find Voltaire's "voice" in the many characters of "Candide" and his other works. Indeed, writers have seen Voltaire as speaking through at least Candide, Martin, and the Turk. Wolper argues that "Candide" should be read with a minimum of speculation as to its meaning in Voltaire's personal life. His article ushered in a new era of Voltaire studies, causing many scholars to look at the novel differently. Critics such as Lester Crocker, Henry Stavan, and Vivienne Mylne find too many similarities between "Candide"s point of view and that of Voltaire to accept the "inside" view; they support the "outside" interpretation. They believe that Candide's final decision is the same as Voltaire's, and see a strong connection between the development of the protagonist and his author. Some scholars who support the "outside" view also believe that the isolationist philosophy of the Old Turk closely mirrors that of Voltaire. Others see a strong parallel between Candide's gardening at the conclusion and the gardening of the author. Martine Darmon Meyer argues that the "inside" view fails to see the satirical work in context, and that denying that "Candide" is primarily a mockery of optimism (a matter of historical context) is a "very basic betrayal of the text". Though Voltaire did not openly admit to having written the controversial "Candide" until 1768 (until then he signed with a pseudonym: "Monsieur le docteur Ralph", or "Doctor Ralph"), his authorship of the work was hardly disputed. Immediately after publication, the work and its author were denounced by both secular and religious authorities, because the book openly derides government and church alike. It was because of such polemics that Omer-Louis-François Joly de Fleury, who was Advocate General to the Parisian parliament when "Candide" was published, found parts of "Candide" to be "contrary to religion and morals". Despite much official indictment, soon after its publication, "Candide"s irreverent prose was being quoted. "Let us eat a Jesuit", for instance, became a popular phrase for its reference to a humorous passage in "Candide". By the end of February 1759, the Grand Council of Geneva and the administrators of Paris had banned "Candide". "Candide" nevertheless succeeded in selling twenty thousand to thirty thousand copies by the end of the year in over twenty editions, making it a best seller. The Duke de La Vallière speculated near the end of January 1759 that "Candide" might have been the fastest-selling book ever. In 1762, "Candide" was listed in the "Index Librorum Prohibitorum", the Roman Catholic Church's list of prohibited books. Bannings of "Candide" lasted into the twentieth century in the United States, where it has long been considered a seminal work of Western literature. At least once, "Candide" was temporarily barred from entering America: in February 1929, a US customs official in Boston prevented a number of copies of the book, deemed "obscene", from reaching a Harvard University French class. "Candide" was admitted in August of the same year; however by that time the class was over. In an interview soon after "Candide"s detention, the official who confiscated the book explained the office's decision to ban it, "But about 'Candide,' I'll tell you. For years we've been letting that book get by. There were so many different editions, all sizes and kinds, some illustrated and some plain, that we figured the book must be all right. Then one of us happened to read it. It's a filthy book". "Candide" is the most widely read of Voltaire's many works, and it is considered one of the great achievements of Western literature. However, "Candide" is not necessarily considered a true "classic". According to Bottiglia, "The physical size of "Candide", as well as Voltaire's attitude toward his fiction, precludes the achievement of artistic dimension through plenitude, autonomous '3D' vitality, emotional resonance, or poetic exaltation. "Candide", then, cannot in quantity or quality, measure up to the supreme classics." Bottiglia instead calls it a miniature classic, though others are more forgiving of its size. As the only work of Voltaire which has remained popular up to the present day, "Candide" is listed in Harold Bloom's "". It is included in the Encyclopædia Britannica collection "Great Books of the Western World". "Candide" has influenced modern writers of black humour such as Céline, Joseph Heller, John Barth, Thomas Pynchon, Kurt Vonnegut, and Terry Southern. Its parody and picaresque methods have become favourites of black humorists. Charles Brockden Brown, an early American novelist, may have been directly affected by Voltaire, whose work he knew well. Mark Kamrath, professor of English, describes the strength of the connection between "Candide" and "Edgar Huntly; or, Memoirs of a Sleep-Walker" (1799): "An unusually large number of parallels...crop up in the two novels, particularly in terms of characters and plot." For instance, the protagonists of both novels are romantically involved with a recently orphaned young woman. Furthermore, in both works the brothers of the female lovers are Jesuits, and each is murdered (although under different circumstances). Some twentieth-century novels that may have been influenced by "Candide" are dystopian science-fiction works. Armand Mattelart, a French critic, sees "Candide" in Aldous Huxley's "Brave New World", George Orwell's "Nineteen Eighty-Four" and Yevgeny Zamyatin's "We", three canonical works of the genre. Specifically, Mattelart writes that in each of these works, there exist references to "Candide"s popularisation of the phrase "the best of all possible worlds". He cites as evidence, for example, that the French version of "Brave New World" was entitled (). Readers of "Candide" often compare it with certain works of the modern genre the Theatre of the Absurd. Haydn Mason, a Voltaire scholar, sees in "Candide" a few similarities to this brand of literature. For instance, he notes commonalities of "Candide" and "Waiting for Godot" (1952). In both of these works, and in a similar manner, friendship provides emotional support for characters when they are confronted with harshness of their existences. However, Mason qualifies, "the must not be seen as a forerunner of the 'absurd' in modern fiction. Candide's world has many ridiculous and meaningless elements, but human beings are not totally deprived of the ability to make sense out of it." John Pilling, biographer of Beckett, does state that "Candide" was an early and powerful influence on Beckett's thinking. Rosa Luxemburg, in the aftermath of the First World War, remarked upon re-reading "Candide": "Before the war, I would have thought this wicked compilation of all human misery a caricature. Now it strikes me as altogether realistic." The American alternative rock band Bloodhound Gang refer to "Candide" in their song "Take the Long Way Home", from the American edition of their 1999 album "Hooray for Boobies". In 1760, one year after Voltaire published "Candide", a sequel was published with the name . This work is attributed both to Thorel de Campigneulles, a writer unknown today, and Henri Joseph Du Laurens, who is suspected of having habitually plagiarised Voltaire. The story continues in this sequel with Candide having new adventures in the Ottoman Empire, Persia, and Denmark. "Part II" has potential use in studies of the popular and literary receptions of "Candide", but is almost certainly apocryphal. In total, by the year 1803, at least ten imitations of "Candide" or continuations of its story were published by authors other than Voltaire. "Candide" was adapted for the radio anthology program "On Stage" in 1953. Richard Chandlee wrote the script; Elliott Lewis, Cathy Lewis, Edgar Barrier, Byron Kane, Jack Kruschen, Howard McNear, Larry Thor, Martha Wentworth, and Ben Wright performed. The operetta "Candide" was originally conceived by playwright Lillian Hellman, as a play with incidental music. Leonard Bernstein, the American composer and conductor who wrote the music, was so excited about the project that he convinced Hellman to do it as a "comic operetta". Many lyricists worked on the show, including James Agee, Dorothy Parker, John Latouche, Richard Wilbur, Leonard and Felicia Bernstein, and Hellman. Hershy Kay orchestrated all the pieces except for the overture, which Bernstein did himself. "Candide" first opened on Broadway as a musical on 1 December 1956. The premier production was directed by Tyrone Guthrie and conducted by Samuel Krachmalnick. While this production was a box office flop, the music was highly praised, and an original cast album was made. The album gradually became a cult hit, but Hellman's libretto was criticised as being too serious an adaptation of Voltaire's novel. "Candide" has been revised and reworked several times. The first New York revival, directed by Hal Prince, featured an entirely new libretto by Hugh Wheeler and additional lyrics by Stephen Sondheim. Bernstein revised the work again in 1987 with the collaboration of John Mauceri and John Wells. After Bernstein's death, further revised productions of the musical were performed in versions prepared by Trevor Nunn and John Caird in 1999, and Mary Zimmerman in 2010. "" (1977) or simply is a book by Leonardo Sciascia. It was at least partly based on Voltaire's "Candide", although the actual influence of "Candide" on is a hotly debated topic. A number of theories on the matter have been proposed. Proponents of one say that is very similar to "Candide", only with a happy ending; supporters of another claim that Voltaire provided Sciascia with only a starting point from which to work, that the two books are quite distinct. The BBC produced a television adaptation in 1973, with Ian Ogilvy as Candide, Emrys James as Dr. Pangloss, and Frank Finlay as Voltaire himself, acting as the narrator. Nedim Gürsel wrote his 2001 novel "Le voyage de Candide à Istanbul" about a minor passage in "Candide" during which its protagonist meets Ahmed III, the deposed Turkish sultan. This chance meeting on a ship from Venice to Istanbul is the setting of Gürsel's book. Terry Southern, in writing his popular novel "Candy" with Mason Hoffenberg adapted "Candide" for a modern audience and changed the protagonist from male to female. "Candy" deals with the rejection of a sort of optimism which the author sees in women's magazines of the modern era; "Candy" also parodies pornography and popular psychology. This adaptation of "Candide" was itself adapted for the cinema by director Christian Marquand in 1968. In addition to the above, "Candide" was made into a number of minor films and theatrical adaptations throughout the twentieth century. For a list of these, see (1989) with preface and commentaries by Pierre Malandain. In May 2009, a play called "Optimism", based on "Candide" opened at the CUB Malthouse Theatre in Melbourne. It followed the basic storyline of "Candide", incorporating anachronisms, music and stand up comedy from comedian Frank Woodley. It toured Australia and played at the Edinburgh International Festival. In 2010, the Icelandic writer Óttar M. Norðfjörð published a rewriting and modernisation of "Candide", entitled . In May 2019, a new theatrical adaptation of "Candide" by John Taflan and Krista D'Agostino entitled, """Candide, or, Optimism, or #blessed, or, a journey into pain and suffering and the demise of privileged innocence and--wait, how are middle schoolers going to do this?, or, Candide!," opened at Mudlark Theater in Evanston, IL. Using the book as a departure point, the production featured middle school-aged actors playing Candide and his friends alongside adult actors playing Pangloss, Martin, and the Old Woman. (plain text and HTML)
https://en.wikipedia.org/wiki?curid=6629
Chapterhouse: Dune Chapterhouse: Dune is a 1985 science fiction novel by Frank Herbert, the last in his "Dune" series of six novels. It rose to No. 2 on "The New York Times" Best Seller list. A direct follow-up to "Heretics of Dune", the novel chronicles the continued struggles of the Bene Gesserit Sisterhood against the violent Honored Matres, who are succeeding in their bid to seize control of the universe and destroy the factions and planets that oppose them. "Chapterhouse: Dune" ends with a cliffhanger, and Herbert's subsequent death in 1986 left some overarching plotlines of the series unresolved. Two decades later, Herbert's son Brian Herbert, along with Kevin J. Anderson, published two sequels – "Hunters of Dune" (2006) and "Sandworms of Dune" (2007) – based in part on notes left behind by Frank Herbert for what he referred to as "Dune 7", his own planned seventh novel in the "Dune" series. The situation is desperate for the Bene Gesserit as they find themselves the targets of the Honored Matres, whose conquest of the Old Empire is almost complete. The Matres are seeking to assimilate the technology and developed methods of the Bene Gesserit and exterminate the Sisterhood itself. Now in command of the Bene Gesserit, Mother Superior Darwi Odrade continues to develop her drastic, secret plan to overcome the Honored Matres. The Bene Gesserit are also terraforming the planet Chapterhouse to accommodate the all-important sandworms, whose native planet Dune had been destroyed by the Matres. Sheeana, in charge of the project, expects sandworms to appear soon. The Honored Matres have also destroyed the entire Bene Tleilax civilization, with Tleilaxu Master Scytale the only one of his kind left alive. In Bene Gesserit captivity, Scytale possesses the Tleilaxu secret of ghola production, which he has reluctantly traded for the Sisterhood's protection. The first ghola produced is that of their recently deceased military genius, Miles Teg. The Bene Gesserit have two other prisoners on Chapterhouse: the latest Duncan Idaho ghola, and former Honored Matre Murbella, whom they have accepted as a novice despite their suspicion that she intends to escape back to the Honored Matres. Lampadas, a center for Bene Gesserit education, has been destroyed by the Honored Matres. The planet's Chancellor, Reverend Mother Lucilla, manages to escape carrying the shared-minds of millions of Reverend Mothers. Lucilla is forced to land on Gammu where she seeks refuge with an underground group of Jews. The Rabbi gives Lucilla sanctuary, but to save his organization he must deliver her to the Matres. Before doing so, he reveals Rebecca, a "wild" Reverend Mother who has gained her Other Memory without Bene Gesserit training. Lucilla shares minds with Rebecca, who promises to take the memories of Lampadas safely back to the Sisterhood. Lucilla is then "betrayed", and taken before the Great Honored Matre Dama, who tries to persuade her to join the Honored Matres, preserving her life in exchange for Bene Gesserit secrets. Lucilla refuses, and Dama ultimately kills her. Back on Chapterhouse, Odrade confronts Duncan and forces him to admit that he is a Mentat, proving that he retains the memories of his many ghola lives. He does not reveal his mysterious visions of two people. Meanwhile, Murbella collapses under the pressure of Bene Gesserit training, giving in to "word weapons" that the Bene Gesserit had planted to undermine her earlier Honored Matre identity. Murbella realizes that she wants to be Bene Gesserit. Odrade believes that the Bene Gesserit made a mistake in fearing emotion, and that in order to evolve, the Bene Gesserit must learn to accept emotions. Odrade permits Duncan to watch Murbella undergo the spice agony, making him the first man ever to do so. Murbella survives the ordeal and becomes a Reverend Mother. Odrade then confronts Sheeana, discovering that Duncan and Sheeana have been allied together for some time. Sheeana does not reveal that they have been considering the option of reawakening Teg's memory through Imprinting, nor does Odrade discover that Sheeana has the keys to Duncan's no-ship prison. Odrade continues molding Scytale, with Sheeana showing him a baby sandworm, the Bene Gesserit's own long term supply of spice, and destroying Scytale's main bargaining card. Finally, Teg is awakened by Sheeana using imprinting techniques. Odrade appoints him again as Bashar of the military forces of the Sisterhood for the assault on the Honored Matres. Odrade next calls a meeting of all the Bene Gesserit, announcing her plan to attack the Honored Matres. She tells them that this attack will be led by Teg. She also announces candidates to succeed her as Mother Superior; she will share her memories with Murbella and Sheeana before she leaves. Odrade then goes to meet the Great Honored Matre. Under cover of Odrade's diplomacy, the Bene Gesserit forces under Teg attack Gammu with tremendous force. Teg uses his secret ability to see no-ships to secure control of the system. Survivors of the attack flee to Junction, and Teg follows them there and carries all with him. Victory for the Bene Gesserit seems inevitable. In the midst of this battle, the Jews (including Rebecca with her precious memories) take refuge with the Bene Gesserit fleet. Logno — chief advisor to Dama — assassinates Dama with poison and assumes control of the Honored Matres. Her first act surprises Odrade greatly. Too late, Odrade and Teg realize they have fallen into a trap, and the Honored Matres use a mysterious weapon to turn defeat into victory, as well as capturing Odrade. Murbella saves as much of the Bene Gesserit force as she can and they begin to withdraw to Chapterhouse. Odrade, however, had planned for the possible failure of the Bene Gesserit attack and left Murbella instructions for a last desperate gamble. Murbella pilots a small craft down to the surface, announcing herself as an Honored Matre who, in the confusion, has managed to escape the Bene Gesserit with all their secrets. She arrives on the planet and is taken to the Great Honored Matre. Unable to control her anger, Logno attacks but is killed by Murbella. Awed by her physical prowess, the remaining Honored Matres are forced to accept her as their new leader. Odrade is also killed in the melee and Murbella shares with Odrade to absorb her newest memories, as they had already shared prior to the battle. Murbella's ascension to leadership is not accepted as victory by all the Bene Gesserit. Some flee Chapterhouse, notably Sheeana, who has a vision of her own, and arranges to have some of the new worms that have emerged in the Chapterhouse desert brought aboard the no-ship. Sheeana is joined by Duncan. The two escape in the giant no-ship, with Scytale, Teg and the Jews. Murbella recognizes their plan at the last minute, but is powerless to stop them. Watching this escape with interest are Daniel and Marty, the observers of whom Duncan had been having visions. The story ends on a cliffhanger with several questions left unanswered regarding the merging of the Honored Matres and Bene Gesserit, the fates of those on the escaped no-ship (including the role of Scytale, the development of Idaho and Teg, and the role of the Jews), the identity of the god-like characters in the book's final chapter and the ultimate mystery of what chased the Honored Matres back into the Old Empire. "Chapterhouse: Dune" debuted at No. 5 and rose to No. 2 on "The New York Times" Best Seller list. Gerald Jonas of "The New York Times" noted that "Against all odds, the universe of "Dune" keeps getting richer in texture, more challenging in its moral dilemmas." Two decades after Frank Herbert's death, his son Brian Herbert, along with Kevin J. Anderson, published two sequels – "Hunters of Dune" (2006) and "Sandworms of Dune" (2007) – based on notes left behind by Frank Herbert for what he referred to as "Dune 7", his own planned seventh novel in the "Dune" series.
https://en.wikipedia.org/wiki?curid=6630
Bus (computing) In computer architecture, a bus (a contraction of the Latin "omnibus", and historically also called "data highway") is a communication system that transfers data between components inside a computer, or between computers. This expression covers all related hardware components (wire, optical fiber, etc.) and software, including communication protocols. Early computer buses were parallel electrical wires with multiple hardware connections, but the term is now used for any physical arrangement that provides the same logical function as a parallel electrical bus. Modern computer buses can use both parallel and bit serial connections, and can be wired in either a multidrop (electrical parallel) or daisy chain topology, or connected by switched hubs, as in the case of USB. Computer systems generally consist of three main parts: the central processing unit (CPU) that processes data, memory that holds the programs and data to be processed, and I/O (input/output) devices as peripherals that communicate with the outside world. An early computer might contain a hand-wired CPU of vacuum tubes, a magnetic drum for main memory, and a punch tape and printer for reading and writing data respectively. A modern system might have a multi-core CPU, DDR4 SDRAM for memory, a solid-state drive for secondary storage, a graphics card and LCD as a display system, a mouse and keyboard for interaction, and a Wi-Fi connection for networking. In both examples, computer buses of one form or another move data between all of these devices. In most traditional computer architectures, the CPU and main memory tend to be tightly coupled. A microprocessor conventionally is a single chip which has a number of electrical connections on its pins that can be used to select an "address" in the main memory and another set of pins to read and write the data stored at that location. In most cases, the CPU and memory share signalling characteristics and operate in synchrony. The bus connecting the CPU and memory is one of the defining characteristics of the system, and often referred to simply as the system bus. It is possible to allow peripherals to communicate with memory in the same fashion, attaching adaptors in the form of expansion cards directly to the system bus. This is commonly accomplished through some sort of standardized electrical connector, several of these forming the expansion bus or local bus. However, as the performance differences between the CPU and peripherals varies widely, some solution is generally needed to ensure that peripherals do not slow overall system performance. Many CPUs feature a second set of pins similar to those for communicating with memory, but able to operate at very different speeds and using different protocols. Others use smart controllers to place the data directly in memory, a concept known as direct memory access. Most modern systems combine both solutions, where appropriate. As the number of potential peripherals grew, using an expansion card for every peripheral became increasingly untenable. This has led to the introduction of bus systems designed specifically to support multiple peripherals. Common examples are the SATA ports in modern computers, which allow a number of hard drives to be connected without the need for a card. However, these high-performance systems are generally too expensive to implement in low-end devices, like a mouse. This has led to the parallel development of a number of low-performance bus systems for these solutions, the most common example being the standardized Universal Serial Bus (USB). All such examples may be referred to as peripheral buses, although this terminology is not universal. In modern systems the performance difference between the CPU and main memory has grown so great that increasing amounts of high-speed memory is built directly into the CPU, known as a cache. In such systems, CPUs communicate using high-performance buses that operate at speeds much greater than memory, and communicate with memory using protocols similar to those used solely for peripherals in the past. These system buses are also used to communicate with most (or all) other peripherals, through adaptors, which in turn talk to other peripherals and controllers. Such systems are architecturally more similar to multicomputers, communicating over a bus rather than a network. In these cases, expansion buses are entirely separate and no longer share any architecture with their host CPU (and may in fact support many different CPUs, as is the case with PCI). What would have formerly been a system bus is now often known as a front-side bus. Given these changes, the classical terms "system", "expansion" and "peripheral" no longer have the same connotations. Other common categorization systems are based on the bus's primary role, connecting devices internally or externally, PCI vs. SCSI for instance. However, many common modern bus systems can be used for both; SATA and the associated eSATA are one example of a system that would formerly be described as internal, while certain automotive applications use the primarily external IEEE 1394 in a fashion more similar to a system bus. Other examples, like InfiniBand and I²C were designed from the start to be used both internally and externally. The internal bus, also known as internal data bus, memory bus, system bus or Front-Side-Bus, connects all the internal components of a computer, such as CPU and memory, to the motherboard. Internal data buses are also referred to as local buses, because they are intended to connect to local devices. This bus is typically rather quick and is independent of the rest of the computer operations. The external bus, or expansion bus, is made up of the electronic pathways that connect the different external devices, such as printer etc., to the computer. An address bus is a bus that is used to specify a physical address. When a processor or DMA-enabled device needs to read or write to a memory location, it specifies that memory location on the address bus (the value to be read or written is sent on the data bus). The width of the address bus determines the amount of memory a system can address. For example, a system with a 32-bit address bus can address 232 (4,294,967,296) memory locations. If each memory location holds one byte, the addressable memory space is 4 GB. Early processors used a wire for each bit of the address width. For example, a 16-bit address bus had 16 physical wires making up the bus. As the buses became wider and lengthier, this approach became expensive in terms of the number of chip pins and board traces. Beginning with the Mostek 4096 DRAM, address multiplexing implemented with multiplexers became common. In a multiplexed address scheme, the address is sent in two equal parts on alternate bus cycles. This halves the number of address bus signals required to connect to the memory. For example, a 32-bit address bus can be implemented by using 16 lines and sending the first half of the memory address, immediately followed by the second half memory address Typically 2 additional pins in the control bus -- a row-address strobe (RAS) and the column-address strobe (CAS) -- are used to tell the DRAM whether the address bus is currently sending the first half of the memory address or the second half. Accessing an individual byte frequently requires reading or writing the full bus width (a word) at once. In these instances the least significant bits of the address bus may not even be implemented - it is instead the responsibility of the controlling device to isolate the individual byte required from the complete word transmitted. This is the case, for instance, with the VESA Local Bus which lacks the two least significant bits, limiting this bus to aligned 32-bit transfers. Historically, there were also some examples of computers which were only able to address words -- word machines. Buses can be parallel buses, which carry data words in parallel on multiple wires, or serial buses, which carry data in bit-serial form. The addition of extra power and control connections, differential drivers, and data connections in each direction usually means that most serial buses have more conductors than the minimum of one used in 1-Wire and UNI/O. As data rates increase, the problems of timing skew, power consumption, electromagnetic interference and crosstalk across parallel buses become more and more difficult to circumvent. One partial solution to this problem has been to double pump the bus. Often, a serial bus can be operated at higher overall data rates than a parallel bus, despite having fewer electrical connections, because a serial bus inherently has no timing skew or crosstalk. USB, FireWire, and Serial ATA are examples of this. Multidrop connections do not work well for fast serial buses, so most modern serial buses use daisy-chain or hub designs. Network connections such as Ethernet are not generally regarded as buses, although the difference is largely conceptual rather than practical. An attribute generally used to characterize a bus is that power is provided by the bus for the connected hardware. This emphasizes the busbar origins of bus architecture as supplying switched or distributed power. This excludes, as buses, schemes such as serial RS-232, parallel Centronics, IEEE 1284 interfaces and Ethernet, since these devices also needed separate power supplies. Universal Serial Bus devices may use the bus supplied power, but often use a separate power source. This distinction is exemplified by a telephone system with a connected modem, where the RJ11 connection and associated modulated signalling scheme is not considered a bus, and is analogous to an Ethernet connection. A phone line connection scheme is not considered to be a bus with respect to signals, but the Central Office uses buses with cross-bar switches for connections between phones. However, this distinctionthat power is provided by the busis not the case in many avionic systems, where data connections such as ARINC 429, ARINC 629, MIL-STD-1553B (STANAG 3838), and EFABus (STANAG 3910) are commonly referred to as “data buses” or, sometimes, "databuses". Such avionic data buses are usually characterized by having several equipments or Line Replaceable Items/Units (LRI/LRUs) connected to a common, shared media. They may, as with ARINC 429, be simplex, i.e. have a single source LRI/LRU or, as with ARINC 629, MIL-STD-1553B, and STANAG 3910, be duplex, allow all the connected LRI/LRUs to act, at different times (half duplex), as transmitters and receivers of data. Some processors use a dedicated wire for each bit of the address bus, data bus, and the control bus. For example, the 64-pin STEbus is composed of 8 physical wires dedicated to the 8-bit data bus, 20 physical wires dedicated to the 20-bit address bus, 21 physical wires dedicated to the control bus, and 15 physical wires dedicated to various power buses. Bus multiplexing requires fewer wires, which reduces costs in many early microprocessors and DRAM chips. One common multiplexing scheme, address multiplexing, has already been mentioned. Another multiplexing scheme re-uses the address bus pins as the data bus pins, an approach used by conventional PCI. The various "serial buses" can be seen as the ultimate limit of multiplexing, sending each of the address bits and each of the data bits, one at a time, through a single pin (or a single differential pair). Over time, several groups of people worked on various computer bus standards, including the IEEE Bus Architecture Standards Committee (BASC), the IEEE "Superbus" study group, the open microprocessor initiative (OMI), the open microsystems initiative (OMI), the "Gang of Nine" that developed EISA, etc. Early computer buses were bundles of wire that attached computer memory and peripherals. Anecdotally termed the ""digit trunk"", they were named after electrical power buses, or busbars. Almost always, there was one bus for memory, and one or more separate buses for peripherals. These were accessed by separate instructions, with completely different timings and protocols. One of the first complications was the use of interrupts. Early computer programs performed I/O by waiting in a loop for the peripheral to become ready. This was a waste of time for programs that had other tasks to do. Also, if the program attempted to perform those other tasks, it might take too long for the program to check again, resulting in loss of data. Engineers thus arranged for the peripherals to interrupt the CPU. The interrupts had to be prioritized, because the CPU can only execute code for one peripheral at a time, and some devices are more time-critical than others. High-end systems introduced the idea of channel controllers, which were essentially small computers dedicated to handling the input and output of a given bus. IBM introduced these on the IBM 709 in 1958, and they became a common feature of their platforms. Other high-performance vendors like Control Data Corporation implemented similar designs. Generally, the channel controllers would do their best to run all of the bus operations internally, moving data when the CPU was known to be busy elsewhere if possible, and only using interrupts when necessary. This greatly reduced CPU load, and provided better overall system performance. To provide modularity, memory and I/O buses can be combined into a unified system bus. In this case, a single mechanical and electrical system can be used to connect together many of the system components, or in some cases, all of them. Later computer programs began to share memory common to several CPUs. Access to this memory bus had to be prioritized, as well. The simple way to prioritize interrupts or bus access was with a daisy chain. In this case signals will naturally flow through the bus in physical or logical order, eliminating the need for complex scheduling. Digital Equipment Corporation (DEC) further reduced cost for mass-produced minicomputers, and mapped peripherals into the memory bus, so that the input and output devices appeared to be memory locations. This was implemented in the Unibus of the PDP-11 around 1969. Early microcomputer bus systems were essentially a passive backplane connected directly or through buffer amplifiers to the pins of the CPU. Memory and other devices would be added to the bus using the same address and data pins as the CPU itself used, connected in parallel. Communication was controlled by the CPU, which read and wrote data from the devices as if they are blocks of memory, using the same instructions, all timed by a central clock controlling the speed of the CPU. Still, devices interrupted the CPU by signaling on separate CPU pins. For instance, a disk drive controller would signal the CPU that new data was ready to be read, at which point the CPU would move the data by reading the "memory location" that corresponded to the disk drive. Almost all early microcomputers were built in this fashion, starting with the S-100 bus in the Altair 8800 computer system. In some instances, most notably in the IBM PC, although similar physical architecture can be employed, instructions to access peripherals (codice_1 and codice_2) and memory (codice_3 and others) have not been made uniform at all, and still generate distinct CPU signals, that could be used to implement a separate I/O bus. These simple bus systems had a serious drawback when used for general-purpose computers. All the equipment on the bus had to talk at the same speed, as it shared a single clock. Increasing the speed of the CPU becomes harder, because the speed of all the devices must increase as well. When it is not practical or economical to have all devices as fast as the CPU, the CPU must either enter a wait state, or work at a slower clock frequency temporarily, to talk to other devices in the computer. While acceptable in embedded systems, this problem was not tolerated for long in general-purpose, user-expandable computers. Such bus systems are also difficult to configure when constructed from common off-the-shelf equipment. Typically each added expansion card requires many jumpers in order to set memory addresses, I/O addresses, interrupt priorities, and interrupt numbers. "Second generation" bus systems like NuBus addressed some of these problems. They typically separated the computer into two "worlds", the CPU and memory on one side, and the various devices on the other. A "bus controller" accepted data from the CPU side to be moved to the peripherals side, thus shifting the communications protocol burden from the CPU itself. This allowed the CPU and memory side to evolve separately from the device bus, or just "bus". Devices on the bus could talk to each other with no CPU intervention. This led to much better "real world" performance, but also required the cards to be much more complex. These buses also often addressed speed issues by being "bigger" in terms of the size of the data path, moving from 8-bit parallel buses in the first generation, to 16 or 32-bit in the second, as well as adding software setup (now standardised as Plug-n-play) to supplant or replace the jumpers. However, these newer systems shared one quality with their earlier cousins, in that everyone on the bus had to talk at the same speed. While the CPU was now isolated and could increase speed, CPUs and memory continued to increase in speed much faster than the buses they talked to. The result was that the bus speeds were now very much slower than what a modern system needed, and the machines were left starved for data. A particularly common example of this problem was that video cards quickly outran even the newer bus systems like PCI, and computers began to include AGP just to drive the video card. By 2004 AGP was outgrown again by high-end video cards and other peripherals and has been replaced by the new PCI Express bus. An increasing number of external devices started employing their own bus systems as well. When disk drives were first introduced, they would be added to the machine with a card plugged into the bus, which is why computers have so many slots on the bus. But through the 1980s and 1990s, new systems like SCSI and IDE were introduced to serve this need, leaving most slots in modern systems empty. Today there are likely to be about five different buses in the typical machine, supporting various devices. "Third generation" buses have been emerging into the market since about 2001, including HyperTransport and InfiniBand. They also tend to be very flexible in terms of their physical connections, allowing them to be used both as internal buses, as well as connecting different machines together. This can lead to complex problems when trying to service different requests, so much of the work on these systems concerns software design, as opposed to the hardware itself. In general, these third generation buses tend to look more like a network than the original concept of a bus, with a higher protocol overhead needed than early systems, while also allowing multiple devices to use the bus at once. Buses such as Wishbone have been developed by the open source hardware movement in an attempt to further remove legal and patent constraints from computer design. The Compute Express Link (CXL) is an open standard interconnect for high-speed CPU-to-device and CPU-to-memory, designed to accelerate next-generation data center performance.
https://en.wikipedia.org/wiki?curid=6631
Chinese checkers Sternhalma, commonly known as Chinese Checkers (US and Canadian spelling) or Chinese Chequers (UK spelling), is a strategy board game of German origin which can be played by two, three, four, or six people, playing individually or with partners. The game is a modern and simplified variation of the game Halma. The objective is to be first to race all of one's pieces across the hexagram-shaped board into "home"—the corner of the star opposite one's starting corner—using single-step moves or moves that over other pieces. The remaining players continue the game to establish second-, third-, fourth-, fifth-, and last-place finishers. The rules are simple, so even young children can play. Despite its name, the game is not a variation of checkers, nor did it originate in China or any part of Asia. The game was invented in Germany in 1892 under the name "Stern-Halma" as a variation of the older American game Halma. The "Stern" (German for "star") refers to the board's star shape (in contrast to the square board used in Halma). The name "Chinese Checkers" originated in the United States as a marketing scheme by Bill and Jack Pressman in 1928. The Pressman company's game was originally called "Hop Ching Checkers". In Japan, the game is known as "Diamond Game" (ダイヤモンドゲーム). The game was introduced to Chinese-speaking regions mostly by the Japanese, where it is known as "Tiaoqi" (, "jump chess"). The aim is to race all one's pieces into the star corner on the opposite side of the board before the opponents do the same. The destination corner is called "home". Each player has 10 pieces, except in games between two players when 15 pieces are used. (On bigger star boards, 15 or 21 pieces are used.) In "hop across", the most popular variation, each player starts with their colored pieces on one of the six points or corners of the star and attempts to race them all home into the opposite corner. Players take turns moving a single piece, either by moving one step in any direction to an adjacent empty space, or by jumping in one or any number of available consecutive hops over other single pieces. A player may not combine hopping with a single-step move – a move consists of one or the other. There is no capturing in Sternhalma, so pieces that are hopped over remain active and in play. Turns proceed clockwise around the board. In the diagram, Green might move the topmost piece one space diagonally forward as shown. A "hop" consists of jumping over a single adjacent piece, either one's own or an opponent's, to the empty space directly beyond it in the same line of direction. Red might advance the indicated piece by a chain of three hops in a single move. It is not mandatory to make the most number of hops possible. (In some instances a player may choose to stop the jumping sequence part way in order to impede the opponent's progress, or to align pieces for planned future moves.) Can be played "all versus all", or three teams of two. When playing teams, teammates usually sit at opposite corners of the star, with each team member controlling their own colored set of pieces. The first team to advance both sets to their home destination corners is the winner. The remaining players usually continue play to determine second- and third-place finishers, etc. The four-player game is the same as the game for six players, except that two opposite corners will be unused. In a three-player game, all players control either one or two sets of pieces each. If one set is used, pieces race across the board into empty, opposite corners. If two sets are used, each player controls two differently colored sets of pieces at opposite corners of the star. In a two-player game, each player plays one, two, or three sets of pieces. If one set is played, the pieces usually go into the opponent's starting corner, and the number of pieces per side is increased to 15 (instead of the usual 10). If two sets are played, the pieces can either go into the opponent's starting corners, or one of the players' two sets can go into an opposite empty corner. If three sets are played, the pieces usually go into the opponent's starting corners. A basic strategy is to create or find the longest hopping path that leads closest to home, or immediately into it. (Multiple-jump moves are obviously faster to advance pieces than step-by-step moves.) Since either player can make use of any hopping 'ladder' or 'chain' created, a more advanced strategy involves hindering an opposing player in addition to helping oneself make jumps across the board. Of equal importance are the players' strategies for emptying and filling their starting and home corners. Games between top players are rarely decided by more than a couple of moves. Differing numbers of players result in different starting layouts, in turn imposing different best-game strategies. For example, if a player's home destination corner starts empty (i.e. is not an opponent's starting corner), the player can freely build a 'ladder' or 'bridge' with their pieces between the two opposite ends. But if a player's opponent occupies the home corner, the player may need to wait for opponent pieces to clear before filling the home vacancies. Variants While the standard rules allow hopping over only a single adjacent occupied position at a time (as in checkers), this version of the game allows pieces to catapult over multiple adjacent occupied positions in a line when hopping. In the "fast-paced" or "Super Chinese Checkers" variant popular in France, a piece may hop over a "non-adjacent" piece. A hop consists of jumping over a distant piece (friendly or enemy) to a symmetrical position on the opposite side, in the same line of direction. (For example, if there are two empty positions between the jumping piece and the piece being jumped, the jumping piece lands leaving exactly two empty positions immediately beyond the jumped piece.) As in the standard rules, a jumping move may consist of any number of a chain of hops. (When making a chain of hops, a piece is usually allowed to enter an empty corner, as long as it hops out again before the move is completed.) Jumping over two or more pieces in a hop is not allowed. Therefore, in this variant even more than in the standard version, it is sometimes strategically important to keep one's pieces bunched in order to prevent a long opposing hop. An alternative variant allows hops over "any" symmetrical arrangement, including pairs of pieces, pieces separated by empty positions, and so on. In the "capture" variant, all sixty game pieces start out in the hexagonal field in the center of the gameboard. The center position is left unoccupied, so pieces form a symmetric hexagonal pattern. Color is irrelevant in this variant, so players take turns hopping any game piece over any other eligible game piece(s) on the board. The hopped-over pieces are captured (retired from the game, as in English draughts) and collected in the capturing player's bin. Only jumping moves are allowed; the game ends when no further jumps are possible. The player with the most captured pieces is the winner. The board is tightly packed at the start of the game; as more pieces are captured, the board frees up, often allowing multiple captures to take place in a single move. Two or more players can compete in this variant, but if there are more than six players, not everyone will get a fair turn. This variant resembles the game Leap Frog. The main difference being that in Leap Frog the board is a square board. Diamond game is a variant of Sternhalma played in South Korea and Japan. It uses the same jump rule as in Sternhalma. The aim of the game is to enter all one's pieces into the star corner on the opposite side of the board, before opponents do the same. Each player has ten or fifteen pieces. Ten-piece diamond uses a smaller gameboard than Sternhalma, with 73 spaces. Fifteen-piece diamond uses the same board as in Sternhalma, with 121 spaces. To play diamond each player selects one color and places their 10 or 15 pieces on a triangle. Two or three players can compete.
https://en.wikipedia.org/wiki?curid=6635
Cantor Fitzgerald Cantor Fitzgerald is an American financial services firm that was founded in 1945. It specializes in institutional equity, fixed income sales and trading, and serving the middle market with investment banking services, prime brokerage, and commercial real estate financing. It is also active in new businesses, including advisory and asset management services, gaming technology, and e-commerce. It has more than 5,000 institutional clients. Cantor Fitzgerald is one of 22 primary dealers that are authorized to trade US government securities with the Federal Reserve Bank of New York. Cantor Fitzgerald's 1,600 employees work in more than 30 locations, including financial centers in the Americas, Europe, Asia-Pacific, and the Middle East. Together with its affiliates, Cantor Fitzgerald operates in more than 60 offices in 20 countries and has more than 8,500 employees. In 2011, Cantor's affiliate, BGC Partners, expanded into commercial real estate services by its purchase of Newmark Knight Frank and the assets of Grubb & Ellis, to form Newmark Grubb Knight Frank. Cantor Fitzgerald was formed in 1945 by Bernard Gerald Cantor and John Fitzgerald as an investment bank and brokerage business. It later became known for its computer-based bond brokerage, for the quality of its institutional distribution business model, and for being the market's premier dealer of government securities. In 1965, Cantor Fitzgerald began "large block" sales/trading of equities for institutional customers. It became the world's first electronic marketplace for US government securities in 1972 and in 1983, it was the first to offer worldwide screen brokerage services in US government securities. In 1991, Howard Lutnick was named president and CEO of Cantor Fitzgerald; he became chairman of Cantor Fitzgerald, L.P., in 1996. Cantor Fitzgerald's corporate headquarters and New York City office, on the 101st to the 105th floors of One World Trade Center in Lower Manhattan (2 to 6 floors above the impact zone of a hijacked airliner), were destroyed during the September 11, 2001 attacks. At 8:46:46 a.m., six seconds after the tower was struck by the plane, a Goldman Sachs server issued an alert saying that its trading system had gone offline because it was unable to connect with the server. Every employee that reported for work that morning was killed in the attacks; 658 of its 960 New York employees, 68.5% of its workforce, which was considerably more than any of the other World Trade Center tenants or the New York City Police Department, the Port Authority of New York and New Jersey Police Department, the New York City Fire Department, or the United States Department of Defense since all stairwells leading past the impact zone were destroyed by the initial crash or blocked with smoke, fire, or debris. Lutnick himself was not present because he was taking his son to his first day of kindergarten, but his younger brother, Gary, was among those killed. Lutnick vowed to keep the company alive, and the company was able to bring its trading markets back online within a week. On September 19, Cantor Fitzgerald made a pledge to distribute 25% of the firm's profits for the next five years, and it committed to paying for ten years of health care for the benefit of the families of its 658 former Cantor Fitzgerald, eSpeed, and TradeSpark employees (profits that would otherwise have been distributed to the Cantor Fitzgerald partners). In 2006, the company had completed its promise, having paid a total of $180 million (and an additional $17 million from a relief fund run by Lutnick's sister, Edie). Until the attacks, Cantor had handled about a quarter of the daily transactions in the multitrillion-dollar treasury security market. Cantor Fitzgerald has since rebuilt its infrastructure, partly by the efforts of its London office, and it now has its headquarters in Midtown Manhattan. The company's effort to regain its footing was the subject of Tom Barbash's 2003 book "On Top of the World: Cantor Fitzgerald, Howard Lutnick, and 9/11: A Story of Loss and Renewal" as well as a 2012 documentary, "Out of the Clear Blue Sky". On September 2, 2004, Cantor and other organizations filed a civil lawsuit against Saudi Arabia for allegedly providing money to the hijackers and Al Qaeda. It was later joined in the suit by the Port Authority of New York. Most of the claims against Saudi Arabia were dismissed on January 18, 2005. In December 2013, Cantor Fitzgerald settled its lawsuit against American Airlines for $135 million. Cantor Fitzgerald had been suing for loss of property and interruption of business by alleging the airline to have been negligent by allowing hijackers to board Flight 11. In 2003, the firm launched its fixed income sales and trading group. In 2006, the Federal Reserve added Cantor Fitzgerald & Co. to its list of primary dealers. In 2009, the firm launched Cantor Prime Services, a provider of multi-asset, perimeter brokerage prime brokerage platforms to exploit its clearing, financing, and execution capabilities. Cantor Fitzgerald began building its real estate business with the launch of CCRE in 2010. On December 5, 2014, two Cantor Fitzgerald analysts were said to be in the top 25 analysts on TipRanks. Cantor Fitzgerald has a prolific Special-purpose acquisition company underwriting practice, having led all banks in SPAC underwriting activity in both 2018 and 2019. Edie wrote "An Unbroken Bond: The Untold Story of How the 658 Cantor Fitzgerald Families Faced the Tragedy of 9/11 and Beyond". All proceeds from the sale of the book benefit the Cantor Fitzgerald Relief Fund and the charities that it assists. The Cantor Fitzgerald Relief Fund provided $10 million to families affected by Hurricane Sandy. Howard Lutnick and the Relief Fund "adopted" 19 elementary schools in impacted areas by distributing $1,000 prepaid debit cards to each family from the schools. A total of $10 million in funds was given to families affected by the storm. Two days after the 2013 Moore tornado struck Moore, Oklahoma, killing 24 people and injuring hundreds, Lutnick pledged to donate $2 million to families affected by the tornado. The donation was given out in the form of $1,000 debit cards given out to families. Each year, on September 11, Cantor Fitzgerald and its affiliate, BGC Partners, donate 100% of their revenue to charitable causes on their annual Charity Day, which was originally established to raise money to assist the families of the Cantor employees who died in the World Trade Center attacks. Since its inception, Charity Day has raised $110 million for charities globally. The firm has many subsidiaries and affiliates such as the following: An employee, Venetia Thompson, published, in February 2008, an article in "The Spectator" about her and her colleagues' behavior, with an emphasis on the drinking culture. She was subsequently fired for gross misconduct. She responded in 2010 by publishing a book, "Gross Misconduct: My Year of Excess in the City", about her experiences.
https://en.wikipedia.org/wiki?curid=6639
Cane toad The cane toad ("Rhinella marina"), also known as the giant neotropical toad or marine toad, is a large, terrestrial true toad native to South and mainland Central America, but which has been introduced to various islands throughout Oceania and the Caribbean, as well as Northern Australia. It is the world's largest toad. It is a member of the genus "Rhinella", which includes many true toad species found throughout Central and South America, but it was formerly assigned to the genus "Bufo". The cane toad is an old species. A fossil toad (specimen UCMP 41159) from the La Venta fauna of the late Miocene of Colombia is indistinguishable from modern cane toads from northern South America. It was discovered in a floodplain deposit, which suggests the "R. marina" habitat preferences have long been for open areas. The cane toad is a prolific breeder; females lay single-clump spawns with thousands of eggs. Its reproductive success is partly because of opportunistic feeding: it has a diet, unusual among anurans, of both dead and living matter. Adults average in length; the largest recorded specimen had a snout-vent length of . The cane toad has poison glands, and the tadpoles are highly toxic to most animals if ingested. Its toxic skin can kill many animals, both wild and domesticated, and cane toads are particularly dangerous to dogs. Because of its voracious appetite, the cane toad has been introduced to many regions of the Pacific and the Caribbean islands as a method of agricultural pest control. The common name of the species is derived from its use against the cane beetle ("Dermolepida albohirtum"), which damages sugar cane. The cane toad is now considered a pest and an invasive species in many of its introduced regions. The 1988 film "" documented the trials and tribulations of the introduction of cane toads in Australia. Historically, the cane toads were used to eradicate pests from sugarcane, giving rise to their common name. The cane toad has many other common names, including "giant toad" and "marine toad"; the former refers to its size, and the latter to the binomial name, "R. marina". It was one of many species described by Linnaeus in his 18th-century work "Systema Naturae" (1758). Linnaeus based the specific epithet "marina" on an illustration by Dutch zoologist Albertus Seba, who mistakenly believed the cane toad to inhabit both terrestrial and marine environments. Other common names include "giant neotropical toad", "Dominican toad", "giant marine toad", and "South American cane toad". In Trinidadian English, they are commonly called "crapaud", the French word for toad. The genus "Rhinella" is considered to constitute a distinct genus of its own, thus changing the scientific name of the cane toad. In this case, the specific name "marinus" (masculine) changes to "marina" (feminine) to conform with the rules of gender agreement as set out by the International Code of Zoological Nomenclature, changing the binomial name from "Bufo marinus" to "Rhinella marina"; the binomial "Rhinella marinus" was subsequently introduced as a synonym through misspelling by Pramuk, Robertson, Sites, and Noonan (2008). Though controversial (with many traditional herpetologists still using "Bufo marinus") the binomial "Rhinella marina" is gaining in acceptance with such bodies as the IUCN, Encyclopaedia of Life, Amphibian Species of the World and increasing numbers of scientific publications adopting its usage. Since 2016, cane toad populations native to Mesoamerica and northwestern South America are sometimes considered to be a separate species, "Rhinella horribilis". In Australia, the adults may be confused with large native frogs from the genera "Limnodynastes", "Cyclorana", and "Mixophyes". These species can be distinguished from the cane toad by the absence of large parotoid glands behind their eyes and the lack of a ridge between the nostril and the eye. Cane toads have been confused with the giant burrowing frog ("Heleioporus australiacus"), because both are large and warty in appearance; however, the latter can be readily distinguished from the former by its vertical pupils and its silver-grey (as opposed to gold) irises. Juvenile cane toads may be confused with species of the genus "Uperoleia", but their adult colleagues can be distinguished by the lack of bright colouring on the groin and thighs. In the United States, the cane toad closely resembles many bufonid species. In particular, it could be confused with the southern toad ("Bufo terrestris"), which can be distinguished by the presence of two bulbs in front of the parotoid glands. The cane toad genome has been sequenced and certain Australian academics believe this will help in understanding how the toad can quickly evolve to adapt to new environments, the workings of its infamous toxin, and hopefully provide new options for halting this species’ march across Australia and other places it has spread as an invasive pest. The cane toad is very large; the females are significantly longer than males, reaching a typical length of , with a maximum of . Larger toads tend to be found in areas of lower population density. They have a life expectancy of 10 to 15 years in the wild, and can live considerably longer in captivity, with one specimen reportedly surviving for 35 years. The skin of the cane toad is dry and warty. It has distinct ridges above the eyes, which run down the snout. Individual cane toads can be grey, yellowish, red-brown, or olive-brown, with varying patterns. A large parotoid gland lies behind each eye. The ventral surface is cream-coloured and may have blotches in shades of black or brown. The pupils are horizontal and the irises golden. The toes have a fleshy webbing at their base, and the fingers are free of webbing. Typically, juvenile cane toads have smooth, dark skin, although some specimens have a red wash. Juveniles lack the adults' large parotoid glands, so they are usually less poisonous. The tadpoles are small and uniformly black, and are bottom-dwellers, tending to form schools. Tadpoles range from in length. The common name "marine toad" and the scientific name "Rhinella marina" suggest a link to marine life, but cane toads do not live in the sea. However, laboratory experiments suggest that tadpoles can tolerate salt concentrations equivalent to 15% of seawater (~5.4‰), and recent field observations found living tadpoles and toadlets at salinities of 27.5‰ on Coiba Island, Panama. The cane toad inhabits open grassland and woodland, and has displayed a "distinct preference" for areas modified by humans, such as gardens and drainage ditches. In their native habitats, the toads can be found in subtropical forests, although dense foliage tends to limit their dispersal. The cane toad begins life as an egg, which is laid as part of long strings of jelly in water. A female lays 8,000–25,000 eggs at once and the strings can stretch up to in length. The black eggs are covered by a membrane and their diameter is about . The rate at which an egg grows into a tadpole increases with temperature. Tadpoles typically hatch within 48 hours, but the period can vary from 14 hours to almost a week. This process usually involves thousands of tadpoles—which are small, black, and have short tails—forming into groups. Between 12 and 60 days are needed for the tadpoles to develop into juveniles, with four weeks being typical. Similarly to their adult counterparts, eggs and tadpoles are toxic to many animals. When they emerge, toadlets typically are about in length, and grow rapidly. While the rate of growth varies by region, time of year, and gender, an average initial growth rate of per day is seen, followed by an average rate of per day. Growth typically slows once the toads reach sexual maturity. This rapid growth is important for their survival; in the period between metamorphosis and subadulthood, the young toads lose the toxicity that protected them as eggs and tadpoles, but have yet to fully develop the parotoid glands that produce bufotoxin. Because they lack this key defence, only an estimated 0.5% of cane toads reach adulthood. As with rates of growth, the point at which the toads become sexually mature varies across different regions. In New Guinea, sexual maturity is reached by female toads with a snout–vent length between , while toads in Panama achieve maturity when they are between in length. In tropical regions, such as their native habitats, breeding occurs throughout the year, but in subtropical areas, breeding occurs only during warmer periods that coincide with the onset of the wet season. The cane toad is estimated to have a critical thermal maximum of and a minimum of around . The ranges can change due to adaptation to the local environment. The cane toad has a high tolerance to water loss; some can withstand a 52.6% loss of body water, allowing them to survive outside tropical environments. Most frogs identify prey by movement, and vision appears to be the primary method by which the cane toad detects prey; however, it can also locate food using its sense of smell. They eat a wide range of material; in addition to the normal prey of small rodents, reptiles, other amphibians, birds, and even bats and a range of invertebrates, they also eat plants, dog food, and household refuse. The skin of the adult cane toad is toxic, as well as the enlarged parotoid glands behind the eyes, and other glands across its back. When the toad is threatened, its glands secrete a milky-white fluid known as bufotoxin. Components of bufotoxin are toxic to many animals; even human deaths have been recorded due to the consumption of cane toads. Dogs are especially prone to be poisoned by licking or biting toads. Pets showing excessive drooling, extremely red gums, head-shaking, crying, loss of coordination, and/or convulsions require immediate veterinary attention. Bufotenin, one of the chemicals excreted by the cane toad, is classified as a schedule 9 drug under Australian law, alongside heroin and LSD. The effects of bufotenin are thought to be similar to those of mild poisoning; the stimulation, which includes mild hallucinations, lasts less than an hour. As the cane toad excretes bufotenin in small amounts, and other toxins in relatively large quantities, toad licking could result in serious illness or death. In addition to releasing toxin, the cane toad is capable of inflating its lungs, puffing up, and lifting its body off the ground to appear taller and larger to a potential predator. Since 2011, experimenters in the Kimberley region of Western Australia have used poisonous sausages containing toad meat to try to protect native animals from cane toads' deadly impact. The Western Australian Department of Environment and Conservation, along with the University of Sydney, developed baits to train native animals not to eat the toads. By blending bits of toad with a nausea-inducing chemical, the baits train the animals to stay away from the amphibians. Many species prey on the cane toad and its tadpoles in its native habitat, including the broad-snouted caiman ("Caiman latirostris"), the banded cat-eyed snake ("Leptodeira annulata"), eels (family Anguillidae), various species of killifish, the rock flagtail ("Kuhlia rupestris"), some species of catfish (order Siluriformes), some species of ibis (subfamily Threskiornithinae), and "Paraponera clavata" (bullet ants). Predators outside the cane toad's native range include the whistling kite ("Haliastur sphenurus"), the rakali ("Hydromys chrysogaster"), the black rat ("Rattus rattus") and the water monitor ("Varanus salvator"). The tawny frogmouth ("Podargus strigoides") and the Papuan frogmouth ("Podargus papuensis") have been reported as feeding on cane toads; some Australian crows ("Corvus" spp.) have also learned strategies allowing them to feed on cane toads, such as using their beak to flip toads onto their backs. Rakalis have been observed eating the hearts and livers of the toads, where the toads have moved into their territory. Opossums of the genus "Didelphis" likely can eat cane toads with impunity. Meat ants are unaffected by the cane toads' toxins, so are able to kill them. The cane toad's normal response to attack is to stand still and let its toxin kill the attacker, which allows the ants to attack and eat the toad. The cane toad is native to the Americas, and its range stretches from the Rio Grande Valley in South Texas to the central Amazon and southeastern Peru, and some of the continental islands near Venezuela (such as Trinidad and Tobago). This area encompasses both tropical and semiarid environments. The density of the cane toad is significantly lower within its native distribution than in places where it has been introduced. In South America, the density was recorded to be 20 adults per 100 m (109 yd) of shoreline, 1 to 2% of the density in Australia. The cane toad has been introduced to many regions of the world—particularly the Pacific—for the biological control of agricultural pests. These introductions have generally been well documented, and the cane toad may be one of the most studied of any introduced species. Before the early 1840s, the cane toad had been introduced into Martinique and Barbados, from French Guiana and Guyana. An introduction to Jamaica was made in 1844 in an attempt to reduce the rat population. Despite its failure to control the rodents, the cane toad was introduced to Puerto Rico in the early 20th century in the hope that it would counter a beetle infestation ravaging the sugarcane plantations. The Puerto Rican scheme was successful and halted the economic damage caused by the beetles, prompting scientists in the 1930s to promote it as an ideal solution to agricultural pests. As a result, many countries in the Pacific region emulated the lead of Puerto Rico and introduced the toad in the 1930s. Introduced populations are in Australia, Florida, Papua New Guinea, the Philippines, the Ogasawara, Ishigaki Island and the Daitō Islands of Japan, most Caribbean islands, Fiji and many other Pacific islands, including Hawaii. Since then, the cane toad has become a pest in many host countries, and poses a serious threat to native animals. Following the apparent success of the cane toad in eating the beetles threatening the sugarcane plantations of Puerto Rico, and the fruitful introductions into Hawaii and the Philippines, a strong push was made for the cane toad to be released in Australia to negate the pests ravaging the Queensland cane fields. As a result, 102 toads were collected from Hawaii and brought to Australia. Queensland's sugar scientists released the toad into cane fields in August 1935. After this initial release, the Commonwealth Department of Health decided to ban future introductions until a study was conducted into the feeding habits of the toad. The study was completed in 1936 and the ban lifted, when large-scale releases were undertaken; by March 1937, 62,000 toadlets had been released into the wild. The toads became firmly established in Queensland, increasing exponentially in number and extending their range into the Northern Territory and New South Wales. In 2010, one was found on the far western coast in Broome, Western Australia. However, the toad was generally unsuccessful in reducing the targeted grey-backed cane beetles ("Dermolepida albohirtum"), in part because the cane fields provided insufficient shelter for the predators during the day, and in part because the beetles live at the tops of sugar cane – and cane toads are not good climbers. Since its original introduction, the cane toad has had a particularly marked effect on Australian biodiversity. The population of a number of native predatory reptiles has declined, such as the varanid lizards "Varanus mertensi", "V. mitchelli", and "V. panoptes", the land snakes "Pseudechis australis" and "Acanthophis antarcticus", and the crocodile species "Crocodylus johnstoni"; in contrast, the population of the agamid lizard "Amphibolurus gilberti"—known to be a prey item of "V. panoptes"—has increased. The cane toad was introduced to various Caribbean islands to counter a number of pests infesting local crops. While it was able to establish itself on some islands, such as Barbados, Jamaica, and Puerto Rico, other introductions, such as in Cuba before 1900 and in 1946, and on the islands of Dominica and Grand Cayman, were unsuccessful. The earliest recorded introductions were to Barbados and Martinique. The Barbados introductions were focused on the biological control of pests damaging the sugarcane crops, and while the toads became abundant, they have done even less to control the pests than in Australia. The toad was introduced to Martinique from French Guiana before 1944 and became established. Today, they reduce the mosquito and mole cricket populations. A third introduction to the region occurred in 1884, when toads appeared in Jamaica, reportedly imported from Barbados to help control the rodent population. While they had no significant effect on the rats, they nevertheless became well established. Other introductions include the release on Antigua—possibly before 1916, although this initial population may have died out by 1934 and been reintroduced at a later date— and Montserrat, which had an introduction before 1879 that led to the establishment of a solid population, which was apparently sufficient to survive the Soufrière Hills volcano eruption in 1995. In 1920, the cane toad was introduced into Puerto Rico to control the populations of white grub ("Phyllophaga" spp.), a sugarcane pest. Before this, the pests were manually collected by humans, so the introduction of the toad eliminated labor costs. A second group of toads was imported in 1923, and by 1932, the cane toad was well established. The population of white grubs dramatically decreased, and this was attributed to the cane toad at the annual meeting of the International Sugar Cane Technologists in Puerto Rico. However, there may have been other factors. The six-year period after 1931—when the cane toad was most prolific, and the white grub had a dramatic decline—had the highest-ever rainfall for Puerto Rico. Nevertheless, the cane toad was assumed to have controlled the white grub; this view was reinforced by a "Nature" article titled "Toads save sugar crop", and this led to large-scale introductions throughout many parts of the Pacific. The cane toad has been spotted in Carriacou and Dominica, the latter appearance occurring in spite of the failure of the earlier introductions. On September 8, 2013, the cane toad was also discovered on the island of New Providence in the Bahamas. The cane toad was first introduced deliberately into the Philippines in 1930 as a biological control agent of pests in sugarcane plantations, after the success of the experimental introductions into Puerto Rico. It subsequently became the most ubiquitous amphibian in the islands. It still retains the common name of "bakî" or "kamprag" in the Visayan languages, a corruption of 'American frog', referring to its origins. It is also commonly known as "bullfrog" in Philippine English. The cane toad was introduced into Fiji to combat insects that infested sugarcane plantations. The introduction of the cane toad to the region was first suggested in 1933, following the successes in Puerto Rico and Hawaii. After considering the possible side effects, the national government of Fiji decided to release the toad in 1953, and 67 specimens were subsequently imported from Hawaii. Once the toads were established, a 1963 study concluded, as the toad's diet included both harmful and beneficial invertebrates, it was considered "economically neutral". Today, the cane toad can be found on all major islands in Fiji, although they tend to be smaller than their counterparts in other regions. The cane toad was introduced into New Guinea to control the hawk moth larvae eating sweet potato crops. The first release occurred in 1937 using toads imported from Hawaii, with a second release the same year using specimens from the Australian mainland. Evidence suggests a third release in 1938, consisting of toads being used for human pregnancy tests—many species of toad were found to be effective for this task, and were employed for about 20 years after the discovery was announced in 1948. Initial reports argued the toads were effective in reducing the levels of cutworms and sweet potato yields were thought to be improving. As a result, these first releases were followed by further distributions across much of the region, although their effectiveness on other crops, such as cabbages, has been questioned; when the toads were released at Wau, the cabbages provided insufficient shelter and the toads rapidly left the immediate area for the superior shelter offered by the forest. A similar situation had previously arisen in the Australian cane fields, but this experience was either unknown or ignored in New Guinea. The cane toad has since become abundant in rural and urban areas. The cane toad naturally exists in South Texas, but attempts (both deliberate and accidental) have been made to introduce the species to other parts of the country. These include introductions to Florida and to the islands of Hawaii, as well as largely unsuccessful introductions to Louisiana. Initial releases into Florida failed. Attempted introductions before 1936 and 1944, intended to control sugarcane pests, were unsuccessful as the toads failed to proliferate. Later attempts failed in the same way. However, the toad gained a foothold in the state after an accidental release by an importer at Miami International Airport in 1957, and deliberate releases by animal dealers in 1963 and 1964 established the toad in other parts of Florida. Today, the cane toad is well established in the state, from the Keys to north of Tampa, and they are gradually extending further northward. In Florida, the toad is a regarded as a threat to native species and pets; so much so, the Florida Fish and Wildlife Conservation Commission recommends residents to kill them. Around 150 cane toads were introduced to Oahu in Hawaii in 1932, and the population swelled to 105,517 after 17 months. The toads were sent to the other islands, and more than 100,000 toads were distributed by July 1934; eventually over 600,000 were transported. Other than the use as a biological control for pests, the cane toad has been employed in a number of commercial and noncommercial applications. Traditionally, within the toad's natural range in South America, the Embera-Wounaan would "milk" the toads for their toxin, which was then employed as an arrow poison. The toxins may have been used as an entheogen by the Olmec people. The toad has been hunted as a food source in parts of Peru, and eaten after the careful removal of the skin and parotoid glands. When properly prepared, the meat of the toad is considered healthy and as a source of omega-3 fatty acids. More recently, the toad's toxins have been used in a number of new ways: bufotenin has been used in Japan as an aphrodisiac and a hair restorer, and in cardiac surgery in China to lower the heart rates of patients. New research has suggested that the cane toad's poison may have some applications in treating prostate cancer. Other modern applications of the cane toad include pregnancy testing, as pets, laboratory research, and the production of leather goods. Pregnancy testing was conducted in the mid-20th century by injecting urine from a woman into a male toad's lymph sacs, and if spermatozoa appeared in the toad's urine, the patient was deemed to be pregnant. The tests using toads were faster than those employing mammals; the toads were easier to raise, and, although the initial 1948 discovery employed "Bufo arenarum" for the tests, it soon became clear that a variety of anuran species were suitable, including the cane toad. As a result, toads were employed in this task for around 20 years. As a laboratory animal, the cane toad is regarded as ideal; they are plentiful, and easy and inexpensive to maintain and handle. The use of the cane toad in experiments started in the 1950s, and by the end of the 1960s, large numbers were being collected and exported to high schools and universities. Since then, a number of Australian states have introduced or tightened importation regulations. Even dead toads have value. Cane toad skin has been made into leather and novelty items; stuffed cane toads, posed and accessorised, have found a home in the tourist market, and attempts have been made to produce fertiliser from their bodies. Cane toads pose a serious threat to native species when introduced to a new ecosystem. Classified as an invasive species in over 20 countries, multiple reports exist of the cane toad moving into a new area to be followed by a decline in the biodiversity in that region. The most documented region of the cane toad's invasion and subsequent effect on native species is Australia, where multiple surveys and observations of the toad's conquest have been completed. The best way to illustrate this effect is through the plight of the northern quoll, as well as Mertens' water monitor, a large lizard native to South and Southeast Asia. Two sites were chosen to study the effects of cane toads on the northern quoll, one of which was at Mary River ranger station, which is located in the southern region of Kakadu National Park. The other site was located at the north end of the park. In addition to these two sites, a third site was located at the East Alligator ranger station, and this site was used as a control site, where the cane toads would not interact with the northern quoll population. Monitoring of the quoll population began at the Mary River ranger station using radio tracking in 2002, months before the first cane toads arrived at the site. After the arrival of the cane toads, the population of northern quolls in the Mary River site plummeted between October and December 2002, and by March 2003, the northern quoll appeared to be extinct in this section of the park, as no northern quolls were caught in the trapping trips in the following two months. In contrast, the population of northern quolls in the control site at the East Alligator ranger station remained relatively constant, not showing any symptoms of declining. The evidence from the Kakadu National Park is compelling not only because of the timing of the population of northern quolls plummeting just months after the arrival of the cane toad, but also because in the Mary River region 31% of mortalities within the quoll population were attributed to lethal toxic ingestion, as no signs of disease, parasite infestation, or any other obvious changes at the site were found that could have caused such a rapid decline. The most obvious evidence that supports the hypothesis that the invasion of the cane toads caused the local extinction of the northern quoll is that the closely monitored population of the control group, in the absence of cane toads, showed no signs of decline. In the case of Mertens' water monitor, only one region was monitored, but over the course of 18 months. This region is located 70 km south of Darwin, at the Manton Dam Recreation Area. Within the Manton Dam Recreation Area, 14 sites were set up to survey the population of water monitors, measuring abundance and site occupancy at each one. Seven surveys were conducted, each of which ran for 4 weeks and included 16 site visits, where each site was sampled twice per day for 2 consecutive days throughout the 4 weeks. Each site visit occurred between 7:30 and 10:30 am, and 4:00–7:00 pm, when "Varanus mertensi" can be viewed sunbathing on the shore or wrapped around a tree branch close to shore. The whole project lasted from December 2004 to May 2006, and had a total of 194 sightings of "Varanus mertensi" in 1568 site visits. Of the seven surveys, abundance was highest during the second survey, which took place in February 2005, 2 months into the project. Following this measurement, the abundance declined in the next four surveys, before declining sharply after the second to last survey in February 2006. In the final survey taken in May 2006, only two "V. mertensi" lizards were observed. Cane toads were first recorded in the region of study during the second survey during February 2005, also when the water monitor abundance was at its highest over the course of the study. Numbers of the cane toad population stayed low for the next year after introduction, and then skyrocketed to its peak in the last survey during May 2006. When compared, the two populations side by side clearly show that the onset of the cane toads had an immediate negative impact on the monitors, as their population began to drop in February 2005, which was when the first cane toads entered the Manton Dam Recreation Area. At the end of the study, some scattered populations of water monitors remained in the upper sites of the Manton Dam, which suggests that local extinctions occurred at certain shoreline sites within Manton Dam, but a complete extinction of the population did not occur. Notes Bibliography
https://en.wikipedia.org/wiki?curid=6641
Croquet Croquet (; (UK) or (US)) is a sport that involves hitting wooden or plastic balls with a mallet through hoops (often called "wickets" in the United States) embedded in a grass playing court. The oldest document to bear the word "croquet" with a description of the modern game is the set of rules registered by Isaac Spratt in November 1856 with the Stationers' Company in London. This record is now in the Public Record Office. In 1868, the first croquet all-comers meet was held at Moreton-in-Marsh, Gloucestershire and in the same year the All England Croquet Club was formed at Wimbledon, London. Regardless when and by what route it reached England and the British colonies in its recognizable form, croquet is, like golf, pall-mall, trucco, and kolven, among the later forms of ground billiards, which as a class have been popular in Western Europe back to at least the Late Middle Ages, with roots in classical antiquity, including sometimes the use of arches and pegs along with balls and mallets or other striking sticks (some more akin to modern field hockey sticks). By the 12th century, a team ball game called ' or ', akin to a chaotic version of hockey or football (depending on whether sticks were used), was regularly played in France and southern Britain between villages or parishes; it was attested in Cornwall as early as 1283. In the book "Queen of Games: The History of Croquet", Nicky Smith presents two theories of the origin of the modern game of croquet, which took England by storm in the 1860s and then spread overseas. The first explanation is that the ancestral game was introduced to Britain from France during the 1660–1685 reign of Charles II of England, and was played under the name of ' (among other spellings, today usually "pall-mall"), derived ultimately from Latin words for 'ball and mallet' (the latter also found in the name of the earlier French game, '). This was the explanation given in the ninth edition of "Encyclopædia Britannica", dated 1877. In his 1810 book "The Sports and Pastimes of the People of England", Joseph Strutt described the way pall-mall was played in England at the time:"Pale-maille is a game wherein a round box[wood] ball is struck with a mallet through a high arch of iron, which he that can do at the fewest blows, or at the number agreed upon, wins. It is to be observed, that there are two of these arches, that is one at either end of the alley. The game of mall was a fashionable amusement in the reign of Charles the Second, and the walk in Saint James's Park, now called the Mall, received its name from having been appropriated to the purpose of playing at mall, where Charles himself and his courtiers frequently exercised themselves in the practice of this pastime." While the name "pall-mall" and various games bearing this name also appeared elsewhere (France and Italy), the description above suggests that the croquet-like games in particular were popular in England by the early 17th century. Some other early modern sources refer to pall-mall being played over a large distance (as in golf); however, an image in Strutt's 1801 book shows a croquet-like ground billiards game (balls on ground, hoop, bats, and peg) being played over a , garden-sized distance. The image's caption describes the game as "a curious ancient pastime", confirming that croquet games were not new in early-19th-century England. In Samuel Johnson's 1755 dictionary, his definition of "pall-mall" clearly describes a game with similarities to modern croquet: "A play in which the ball is struck with a mallet through an iron ring". However, there is no evidence that pall-mall involved the croquet stroke which is the distinguishing characteristic of the modern game. The second theory is that the rules of the modern game of croquet arrived from Ireland during the 1850s, perhaps after being brought there from Brittany, where a similar game was played on the beaches. Regular contact between Ireland and France had continued since the Norman invasion of Ireland in 1169. By no later than the early 15th century, the game " (itself ancestral to pall-mall and perhaps to indoor billiards) was popular in France, including in the courts of Henry II in the 16th century and Louis XIV of the 17th. At least one version of it, " ('wheel') was a multi-ball lawn game. Records show a game called "crookey", similar to croquet, being played at Castlebellingham in County Louth, Ireland, in 1834, which was introduced to Galway in 1835 and played on the bishop's palace garden, and in the same year to the genteel Dublin suburb of Kingstown (today Dún Laoghaire) where it was first spelt as "croquet". There is, however, no pre-1858 Irish document that describes the way game was played, in particular there is no reference to the distinctive croquet stroke, which is described below under "Variations: Association". The noted croquet historian Dr Prior, in his book of 1872, makes the categoric statement "One thing only is certain: it is from Ireland that croquet came to England and it was on the lawn of the late Lord Lonsdale that it was first played in this country." This was about 1851. John Jaques apparently claimed in a letter to Arthur Lillie in 1873 that he had himself seen the game played in Ireland, writing "I made the implements and published directions (such as they were) before Mr. Spratt [mentioned above] introduced the subject to me." Whatever the truth of the matter, Jaques certainly played an important role in popularising the game, producing editions of the rules in 1857, 1860, and 1864. Croquet became highly popular as a social pastime in England during the 1860s. It was enthusiastically adopted and promoted by the Earl of Essex who held lavish croquet parties at Cassiobury House, his stately home in Watford, Hertfordshire, and the Earl even launched his own "Cassiobury" brand croquet set. By 1867, Jaques had printed 65,000 copies of his "Laws and Regulations" of the game. It quickly spread to other Anglophone countries, including Australia, Canada, New Zealand, South Africa, and the United States. No doubt one of the attractions was that the game could be played by both sexes; this also ensured a certain amount of adverse comment. By the late 1870s, however, croquet had been eclipsed by another fashionable game, lawn tennis, and many of the newly created croquet clubs, including the All England Club at Wimbledon, converted some or all of their lawns into tennis courts. There was a revival in the 1890s, but from then onwards, croquet was always a minority sport, with national individual participation amounting to a few thousand players. The All England Lawn Tennis and Croquet Club still has a croquet lawn, but has not hosted any significant tournaments. The English headquarters for the game is now in Cheltenham. The earliest known reference to croquet in Scotland is the booklet "The Game of Croquet, its Laws and Regulations" which was published in the mid-1860s for the proprietor of Eglinton Castle, the Earl of Eglinton. On the page facing the title page is a picture of Eglinton Castle with a game of "croquet" in full swing. The croquet lawn existed on the northern terrace, between Eglinton Castle and the Lugton Water. The 13th Earl developed a variation on croquet named Captain Moreton's Eglinton Castle croquet, which had small bells on the eight hoops "to ring the changes", two pegs, a double hoop with a bell and two tunnels for the ball to pass through. In 1865 the 'Rules of the Eglinton Castle and Cassiobury Croquet' was published by Edmund Routledge. Several incomplete sets of this form of croquet are known to exist, and one complete set is still used for demonstration games in the West of Scotland. There are several variations of croquet currently played, differing in the scoring systems, order of shots, and layout (particularly in social games where play must be adapted to smaller-than-standard playing courts). Two forms of the game, association croquet and golf croquet, have rules that are agreed internationally and are played in many countries around the world. The United States has its own set of rules for domestic games. Gateball, a sport originated in Japan under the influence of croquet, is played mainly in East and Southeast Asia and the Americas, and can also be regarded as a croquet variant. As well as club-level games, there are regular world championships and international matches between croquet-playing countries. The sport has particularly strong followings in the UK, US, New Zealand and Australia; every four years, these countries play the MacRobertson Shield tournament. Many other countries also play. The current world rankings show England in top place for association croquet, followed by Australia and New Zealand sharing second place, with the United States in fourth position; the same four countries appear in the top six of the golf croquet league table, below Egypt in top position, and with South Africa at number five. Croquet is popularly believed to be viciously competitive. This may derive from the fact that (unlike in golf) players will often attempt to move their opponents' balls to unfavourable positions. However, purely negative play is rarely a winning strategy: successful players (in all versions other than golf croquet) will use all four balls to set up a break for themselves, rather than simply making the game as difficult as possible for their opponents. At championship-standard association croquet, players can often make all 26 points (13 for each ball) in two turns. Croquet was an event at the 1900 Summer Olympics. Roque, an American variation on croquet, was an event at the 1904 Summer Olympics. Association croquet is the name of an advanced game of croquet, played at all levels up to international level. It involves four balls teamed in pairs, with both balls going through every hoop for one pair to win. The game's distinguishing feature is the "croquet" shot: when certain balls hit other balls, extra shots are allowed. The six hoops are arranged three at each end of the court, with a centre peg. One side takes the black and blue balls, the other takes red and yellow. At each turn, players can choose to play with either of their balls for that turn. At the start of a turn, the player plays a stroke. If the player either hits the ball through the correct hoop ("runs" the hoop), or hits another ball (a "roquet"), the turn continues. Following a roquet, the player picks up his or her own ball and puts it down next to the ball that it hit. The next shot is played with the two balls touching: this is the "croquet stroke" from which the game takes its name. By varying the speed and angle at which the mallet hits the striker's ball, a good player can control the final position of both balls: the horizontal angle determines how far the balls diverge in direction, while the vertical angle and the amount of follow-through determine the relative distance that the two balls travel. After the croquet stroke, the player plays a "continuation" stroke, during which the player may again attempt to make a roquet or run a hoop. Each of the other three balls may be roqueted once in a turn before a hoop is run, after which they become available to be roqueted again. The winner of the game is the team who completes the set circuit of six hoops (and then back again the other way), with both balls, and then strikes the centre peg (making a total of 13 points per ball = 26). Good players may make "s" or "s" of several hoops in a single turn. The best players may take a ball round a full circuit in one turn. "Advanced play" (a variant of association play for expert players) gives penalties to a player who runs certain hoops in a turn, to allow the opponent a chance of getting back into the game; feats of skill such as triple peels or better, in which the partner ball (or occasionally an opponent ball) is caused to run a number of hoops in a turn by the striker's ball, help avoid these penalties. A handicap system ("bisques") provides less experienced players a chance of winning against more formidable opponents. Players of all ages and both sexes compete on level terms. The World Championships are organised by the World Croquet Federation (WCF) and usually take place every two or three years. The 2018 championships took place in Wellington, New Zealand; the winner was Paddy Chapman of New Zealand. The current Women's Association Croquet World Champion (2015) is Miranda Chapman of England. Paddy and Miranda are married. The Australian team won the last MacRobertson International Croquet Shield tournament, which is the major international test tour trophy in association croquet. It is contested every three to four years between Australia, Britain, the United States and New Zealand. Historically the British have been the dominant force, winning 14 out of the 22 times that the event has been held. In individual competition, the UK is often divided by subnational country (England, Scotland and Wales), while Northern Ireland joins with the republic in an All Ireland association (as it does in several other sports). The world's top 10 association croquet players as of February 2018 were Robert Fletcher (Australia), Reg Bamford (South Africa), Robert Fulford (England), Paddy Chapman (New Zealand), Ben Rothman (USA), Malcolm Fletcher (Australia), Jamie Burch (England), Jose Riva (Spain), Stephen Mulliner (England), Greg Bryant (New Zealand). Unlike most sports, men and women compete and are ranked together. Three women have won the British Open Championship: Lily Gower in 1905, Dorothy Steel in 1925, 1933, 1935 and 1936, and Hope Rotherham in 1960. While male players are in the majority at club level in the UK, the opposite is the case in Australia and New Zealand. The governing body in England is The Croquet Association, which has been the driving force of the development of the game. The rules and tournament regulations are now maintained by the International Laws Committee, established by the croquet associations of England and Wales (CA), Australia (ACA), New Zealand (CNZ) and the United States (USCA). In golf croquet, a hoop is won by the first ball to go through each hoop. Unlike association croquet, there are no additional turns for hitting other balls. Each player takes a stroke in turn, each trying to hit a ball through the same hoop. The sequence of play is blue, red, black, yellow. Blue and black balls play against red and yellow. When a hoop is won, the sequence of play continues as before. The winner of the game is the player/team who wins the most hoops. Golf croquet is the fastest-growing version of the game, owing largely to its simplicity and competitiveness. There is an especially large interest with competitive success by players in Egypt. Golf croquet is easier to learn and play, but requires strategic skills and accurate play. In comparison with association croquet, play is faster and balls are more likely to be lifted off the ground. In April 2013, Reg Bamford of South Africa beat Ahmed Nasr of Egypt in the final of the Golf Croquet World Championship in Cairo, becoming the first person to simultaneously hold the title in both association croquet and golf croquet. As of 2017, the Golf Croquet World Champion was Reg Bamford (South Africa) and the Women's Golf Croquet World Champion was Judith Hanekom (South Africa). In 2018, two international championships open to both sexes were won by women: in May, Rachel Gee of England beat Pierre Beaudry to win the European Golf Croquet championship, and in October, Hanan Rashad of Egypt beat Yasser Fathy (also from Egypt) to win the World over-50s Golf Croquet championship. Garden croquet is widely played in the UK. The rules are easy to learn and the game can be played on lawns of almost any size but usually around by . The rules are similar to those described above for Association Croquet with three major differences: This version of the game is easy for beginners to learn. The main Garden Croquet Club in the UK is the Bygrave Croquet Club which is a private club with five lawns. Other clubs also use garden croquet as an introduction to the game, notably the Hampstead Heath Croquet Club and the Watford Croquet Club. The American-rules version of croquet, another six-hoop game, is the dominant version of the game in the United States and is also widely played in Canada. It is governed by the United States Croquet Association. Its genesis is mostly in association croquet, but it differs in a number of important ways that reflect the home-grown traditions of American "backyard" croquet. Two of the most notable differences are that the balls are always played in the same sequence (blue, red, black, yellow) throughout the game, and that a ball's "deadness" on other balls is carried over from turn to turn until the ball has been "cleared" by scoring its next hoop. A Deadness Board is used to keep track of deadness on all four balls. Tactics are simplified on the one hand by the strict sequence of play, and complicated on the other hand by the continuation of deadness. A further difference is the more restrictive boundary-line rules of American croquet. In the American game, roqueting a ball out of bounds or running a hoop out of bounds causes the turn to end, and balls that go out of bounds are replaced only from the boundary rather than as in association croquet. "Attacking" balls on the boundary line to bring them into play is thus far more challenging. Nine-wicket croquet, sometimes called "backyard croquet", is played mainly in Canada and the United States, and is the game most recreational players in those countries call simply "croquet". In this version of croquet there are nine wickets, two stakes, and up to six balls. The course is arranged in a double-diamond pattern, with one stake at each end of the course. Players start at one stake, navigate one side of the double diamond, hit the turning stake, then navigate the opposite side of the double diamond and hit the starting stake to end. If playing individually ("Cutthroat"), the first player to stake out is the winner. In partnership play, all members of a team must stake out, and a player might choose to avoid staking out (becoming a "Rover") in order to help a lagging teammate. Each time a ball is roqueted, the striker gets two bonus shots. For the first bonus shot, the player has four options: The second bonus shot ("continuation shot") is an ordinary shot played from where the striker ball came to rest. An alternate endgame is "poison": in this variant, a player who has scored the last wicket but not hit the starting stake becomes a "poison ball", which may eliminate other balls from the game by roqueting them. A non-poison ball that roquets a poison ball has the normal options. A poison ball that hits a stake or passes through any wicket (possibly by the action of a non-poison player) is eliminated. The last person remaining is the winner. This version of the game was invented by John Riches of Adelaide, Australia with help from Tom Armstrong in the 1980s. The game can be played by up to six people and is very easy to learn. For this reason it is often used as a stepping stone to association croquet. Ricochet has similar rules to association and garden croquet, except that when a ball is roqueted, the striker's ball remains live and two free shots are earned. This enables strikers to play their ball near to another opponent ball and ricochet that too thus earning two more free shots. Running a hoop earns one free shot. One-ball croquet has become popular in recent years as a way of bringing AC (association) and GC (golf) players together. The rules are essentially those of association croquet, except that each player/team has only one ball rather than two. This makes it very hard to create a break, which leads to more interactive play. The way croquet is depicted in paintings and books says much about popular perceptions of the game, though little about the reality of modern play. About 200 croquet clubs across the United States are members of the United States Croquet Association. Many colleges have croquet clubs as well, such as The University of Virginia, The University of Chicago, Pennsylvania State University, Bates College, SUNY New Paltz, Harvard University, and Dartmouth College. Notably, St. John's College and the US Naval Academy engage in a yearly match in Annapolis, Maryland. Both schools also compete at the collegiate level and the rivalry continues to be an Annapolis tradition, attracting thousands of spectators each April. In England and Wales, there are around 170 clubs affiliated with the Croquet Association. The All England Lawn Tennis and Croquet Club at Wimbledon is famous for its lawn tennis tournament, but retains an active croquet section. There are also clubs in many universities and colleges, with an annual Varsity match being played between Oxford and Cambridge. With over 1800 participants, the 2011 Oxford University "Cuppers" (inter-college) tournament claimed to be not only the largest croquet tournament ever, but the largest sporting event in the university's history.
https://en.wikipedia.org/wiki?curid=6643
Curling Curling is a sport in which players slide stones on a sheet of ice toward a target area which is segmented into four concentric circles. It is related to bowls, boules and shuffleboard. Two teams, each with four players, take turns sliding heavy, polished granite stones, also called "rocks", across the ice "curling sheet" toward the "house", a circular target marked on the ice. Each team has eight stones, with each player throwing two. The purpose is to accumulate the highest score for a "game"; points are scored for the stones resting closest to the centre of the house at the conclusion of each "end", which is completed when both teams have thrown all of their stones. A game usually consists of eight or ten ends. The player can induce a curved path, described as "curl", by causing the stone to slowly turn as it slides. The path of the rock may be further influenced by two sweepers with brooms or brushes, who accompany it as it slides down the sheet and sweep the ice in front of the stone. "Sweeping a rock" decreases the friction, which makes the stone travel a straighter path (with less "curl") and a longer distance. A great deal of strategy and teamwork go into choosing the ideal path and placement of a stone for each situation, and the skills of the curlers determine the degree to which the stone will achieve the desired result. This gives curling its nickname of "chess on ice". Evidence that curling existed in Scotland in the early 16th century includes a curling stone inscribed with the date 1511 found (along with another bearing the date 1551) when an old pond was drained at Dunblane, Scotland. The world's oldest curling stone and the world's oldest football are now kept in the same museum (the Stirling Smith Art Gallery and Museum) in Stirling. The first written reference to a contest using stones on ice coming from the records of Paisley Abbey, Renfrewshire, in February 1541. Two paintings, "" and "The Hunters in the Snow" (both dated 1565) by Pieter Bruegel the Elder depict Flemish peasants curling, albeit without brooms; Scotland and the Low Countries had strong trading and cultural links during this period, which is also evident in the history of golf. The word "curling" first appears in print in 1620 in Perth, Scotland, in the preface and the verses of a poem by Henry Adamson. The sport was (and still is, in Scotland and Scottish-settled regions like southern New Zealand) also known as "the roaring game" because of the sound the stones make while traveling over the "pebble" (droplets of water applied to the playing surface). The verbal noun "curling" is formed from the Scots (and English) verb "curl", which describes the motion of the stone. Kilsyth Curling Club claims to be the first club in the world, having been formally constituted in 1716; it is still in existence today. Kilsyth also claims the oldest purpose-built curling pond in the world at Colzium, in the form of a low dam creating a shallow pool some in size. The International Olympic Committee recognises the Royal Caledonian Curling Club (founded as the Grand Caledonian Curling Club in 1838) as developing the first official rules for the sport. In the early history of curling, the playing stones were simply flat-bottomed stones from rivers or fields, which lacked a handle and were of inconsistent size, shape and smoothness. Some early stones had holes for a finger and the thumb, akin to ten-pin bowling balls. Unlike today, the thrower had little control over the 'curl' or velocity and relied more on luck than on precision, skill and strategy. The sport was often played on frozen rivers although purpose-built ponds were later created in many Scottish towns. For example, the Scottish poet David Gray describes whisky-drinking curlers on the Luggie Water at Kirkintilloch. In Darvel, East Ayrshire, the weavers relaxed by playing curling matches using the heavy stone weights from the looms' "warp beams", fitted with a detachable handle for the purpose. Many a wife would keep her husband's brass curling stone handle on the mantelpiece, brightly polished until the next time it was needed. Central Canadian curlers often used 'irons' rather than stones until the early 1900s; Canada is the only country known to have done so, while others experimented with wood or ice-filled tins. Outdoor curling was very popular in Scotland between the 16th and 19th centuries because the climate provided good ice conditions every winter. Scotland is home to the international governing body for curling, the World Curling Federation in Perth, which originated as a committee of the Royal Caledonian Curling Club, the mother club of curling. Today, the sport is most firmly established in Canada, having been taken there by Scottish emigrants. The Royal Montreal Curling Club, the oldest established sports club still active in North America, was established in 1807. The first curling club in the United States was established in 1830, and the sport was introduced to Switzerland and Sweden before the end of the 19th century, also by Scots. Today, curling is played all over Europe and has spread to Brazil, Japan, Australia, New Zealand, China, and Korea. The first world championship for curling was limited to men and was known as the "Scotch Cup", held in Falkirk and Edinburgh, Scotland, in 1959. The first world title was won by the Canadian team from Regina, Saskatchewan, skipped by Ernie Richardson. (The "skip" is the team member who calls the shots; see below.) Curling was one of the first sports that was popular with women and girls. Curling has been a medal sport in the Winter Olympic Games since the 1998 Winter Olympics. It currently includes men's, women's and mixed doubles tournaments (the mixed doubles event was held for the first time in 2018). In February 2002, the International Olympic Committee retroactively decided that the curling competition from the 1924 Winter Olympics (originally called "Semaine des Sports d'Hiver", or International Winter Sports Week) would be considered official Olympic events and no longer be considered demonstration events. Thus, the first Olympic medals in curling, which at the time was played outdoors, were awarded for the 1924 Winter Games, with the gold medal won by Great Britain, two silver medals by Sweden, and the bronze by France. A demonstration tournament was also held during the 1932 Winter Olympic Games between four teams from Canada and four teams from the United States, with Canada winning 12 games to 4. Since the sport's official addition in the 1998 Olympics, Canada has dominated the sport with their men's teams winning gold in 2006, 2010, and 2014, and silver in 1998 and 2002. The women's team won gold in 1998 and 2014, a silver in 2010, and a bronze in 2002 and 2006. The mixed doubles team won gold in 2018. The playing surface or "curling sheet" is defined by the World Curling Federation Rules of Curling. It is a rectangular area of ice, carefully prepared to be as flat and level as possible, in length by in width. The shorter borders of the sheet are called the backboards. Because of the elongated shape, several sheets may be laid out side by side in the same arena, allowing multiple games to be played simultaneously. A target, the "house", is centred on the intersection of the "centre line", drawn lengthwise down the centre of the sheet and the "tee line", drawn from, and parallel to, the backboard. These lines divide the house into quarters. The house consists of a centre circle (the "button") and three concentric rings, of diameters 4, 8 and 12 feet, formed by painting or laying coloured vinyl sheet under the ice and are usually distinguished by colour. A stone must at least touch the outer ring in order to score (see Scoring below); otherwise the rings are merely a visual aid for aiming and judging which stone is closer to the button. Two "hog lines" are drawn from, and parallel to, the backboard. The "hacks", which give the thrower something to push against when making the throw, are fixed behind each button. On indoor rinks, there are usually two fixed hacks, rubber-lined holes, one on each side of the centre line, with the inside edge no more than from the centre line and the front edge on the hack line. A single moveable hack may also be used. The ice may be natural but is usually frozen by a refrigeration plant pumping a brine solution through numerous pipes fixed lengthwise at the bottom of a shallow pan of water. Most curling clubs have an ice maker whose main job is to care for the ice. At the major curling championships, ice maintenance is extremely important. Large events, such as national/international championships, are typically held in an arena that presents a challenge to the ice maker, who must constantly monitor and adjust the ice and air temperatures as well as air humidity levels to ensure a consistent playing surface. It is common for each sheet of ice to have multiple sensors embedded in order to monitor surface temperature, as well as probes set up in the seating area (to monitor humidity) and in the compressor room (to monitor brine supply and return temperatures). The surface of the ice is maintained at a temperature of around . A key part of the preparation of the playing surface is the spraying of water droplets onto the ice, which form "pebble" on freezing. The pebbled ice surface resembles an orange peel, and the stone moves on top of the pebbled ice. The pebble, along with the concave bottom of the stone, decreases the friction between the stone and the ice, allowing the stone to travel farther. As the stone moves over the pebble, any rotation of the stone causes it to "curl", or travel along a curved path. The amount of curl (commonly referred to as the "feet of curl") can change during a game as the pebble wears; the ice maker must monitor this and be prepared to scrape and re-pebble the surface prior to each game. The curling stone (also sometimes called a "rock" in North America) is made of granite and is specified by the World Curling Federation, which requires a weight between , a maximum circumference of and a minimum height of . The only part of the stone in contact with the ice is the "running surface", a narrow, flat annulus or ring, wide and about in diameter; the sides of the stone bulge convex down to the ring and the inside of the ring is hollowed concave to clear the ice. This concave bottom was first proposed by J. S. Russell of Toronto, Ontario, Canada sometime after 1870, and was subsequently adopted by Scottish stone manufacturer Andrew Kay. The granite for the stones comes from two sources: Ailsa Craig, an island off the Ayrshire coast of Scotland, and the Trefor Granite Quarry in Wales. Ailsa Craig is the traditional source and produces two types of granite, "Blue Hone" and "Ailsa Craig Common Green". "Blue Hone" has very low water absorption, which prevents the action of repeatedly freezing water from eroding the stone. "Ailsa Craig Common Green" is a lesser quality granite than "Blue Hone". In the past, most curling stones were made from "Blue Hone" but the island is now a wildlife reserve and the quarry is restricted by environmental conditions that exclude blasting. Kays of Scotland has been making curling stones in Mauchline, Ayrshire, since 1851 and has the exclusive rights to the Ailsa Craig granite, granted by the Marquess of Ailsa, whose family has owned the island since 1560. According to the 1881 Census, Andrew Kay employed 30 people in his curling stone factory in Mauchline. The last harvest of Ailsa Craig granite by Kays took place in 2013, after a hiatus of 11 years; 2,000 tons were harvested, sufficient to fill anticipated orders through at least 2020. Kays have been involved in providing curling stones for the Winter Olympics since Chamonix in 1924 and has been the exclusive manufacturer of curling stones for the Olympics since the 2006 Winter Olympics. "Trefor" granite comes from the Yr Eifl or Trefor Granite Quarry in the village of Trefor on the north coast of the Llŷn Peninsula in Gwynedd, Wales and has produced granite since 1850. "Trefor" granite comes in shades of pink, blue and grey. The quarry supplies curling stone granite exclusively to the Canada Curling Stone Company, which has been producing stones since 1992 and supplied the stones for the 2002 Winter Olympics. A handle is attached by a bolt running vertically through a hole in the centre of the stone. The handle allows the stone to be gripped and rotated upon release; on properly prepared ice the rotation will bend ("curl") the path of the stone in the direction in which the front edge of the stone is turning, especially as the stone slows. Handles are coloured to identify each team, two popular colours in major tournaments being red and yellow. In competition, an electronic handle known as the "eye on the hog" may be fitted to detect hog line violations. This electronically detects whether the thrower's hand is in contact with the handle as it passes the hog line and indicates a violation by lights at the base of the handle (see "delivery" below). The "eye on the hog" eliminates human error and the need for hog line officials. It is mandatory in high-level national and international competition, but its cost, around US$650 each, currently puts it beyond the reach of most curling clubs. The "curling broom", or "brush", is used to sweep the ice surface in the path of the stone (see "sweeping") and is also often used as a balancing aid during delivery of the stone. Prior to the 1950s, most curling brooms were made of corn strands and were similar to household brooms of the day. In 1958, Fern Marchessault of Montreal inverted the corn straw in the centre of the broom. This style of corn broom was referred to as "the Blackjack". Artificial brooms made from man-made fabrics rather than corn, such as the "Rink Rat", also became common later during this time period. Prior to the late sixties, "Scottish" curling brushes were used primarily by some of the Scots, as well as by recreational and elderly curlers, as a substitute for corn brooms, since the technique was easier to learn. In the late sixties, competitive curlers from Calgary, Alberta, such as John Mayer, Bruce Stewart, and, later, the world junior championship teams skipped by Paul Gowsell, proved that the curling brush could be just as (or more) effective without all the blisters common to corn broom use. During that time period, there was much debate in competitive curling circles as to which sweeping device was more effective: brush or broom. Eventually, the brush won out with the majority of curlers making the switch to the less costly and more efficient brush. Today, brushes have replaced traditional corn brooms at every level of curling; it is rare now to see a curler using a corn broom on a regular basis. Curling brushes may have fabric, hog hair, or horsehair heads. Modern curling brush handles are usually hollow tubes made of fibreglass or carbon fibre instead of a solid length of wooden dowel. These hollow tube handles are lighter and stronger than wooden handles, allowing faster sweeping and also enabling more downward force to be applied to the broom head with reduced shaft flex. New, "directional fabric" brooms, which players are worried will alter the fundamentals of the sport by reducing the level of skill required, have been accused of giving players an unfair advantage. The new brooms were temporarily banned by the World Curling Federation and Curling Canada for the 2015–2016 season. The new brooms give sweepers unprecedented control over the direction the stone goes. Curling shoes are similar to ordinary athletic shoes except for special soles; the "slider shoe" (usually known as a "slider") is designed for the sliding foot and the "gripper shoe" (usually known as a "gripper") for the foot that kicks off from the hack. The "slider" is designed to slide and typically has a Teflon sole. It is worn by the thrower during delivery from the hack and by sweepers or the skip to glide down the ice when sweeping or otherwise traveling down the sheet quickly. Stainless steel and "red brick" sliders with lateral blocks of PVC on the sole are also available as alternatives to Teflon. Most shoes have a full-sole sliding surface, but some shoes have a sliding surface covering only the outline of the shoe and other enhancements with the full-sole slider. Some shoes have small disc sliders covering the front and heel portions or only the front portion of the foot, which allow more flexibility in the sliding foot for curlers playing with tuck deliveries. When a player is not throwing, the player's slider shoe can be temporarily rendered non-slippery by using a slip-on gripper. Ordinary athletic shoes may be converted to sliders by using a step-on or slip-on Teflon slider or by applying electrical or gaffer tape directly to the sole or over a piece of cardboard. This arrangement often suits casual or beginning players. The "gripper" is worn by the thrower on the foot that kicks off from the hack during delivery and is designed to grip the ice. It may have a normal athletic shoe sole or a special layer of rubbery material applied to the sole of a thickness to match the sliding shoe. The toe of the hack foot shoe may also have a rubberised coating on the top surface or a flap that hangs over the toe to reduce wear on the top of the shoe as it drags on the ice behind the thrower. Other types of equipment include: The purpose of a game is to score points by getting stones closer to the house centre, or the "button", than the other team's stones. Players from either team alternate in taking shots from the far side of the sheet. An end is complete when all eight rocks from each team have been delivered, a total of sixteen stones. If the teams are tied at the end of regulation, often extra ends are played to break the tie. The winner is the team with the highest score after all ends have been completed (see Scoring below). A game may be conceded if winning the game is infeasible. International competitive games are generally ten ends, so most of the national championships that send a representative to the World Championships or Olympics also play ten ends. However, there is a movement on the World Curling Tour to make the games only eight ends. Most tournaments on that tour are eight ends, as are the vast majority of recreational games. In international competition, each side is given 73 minutes to complete all of its throws. Each team is also allowed two minute-long timeouts per 10-end game. If extra ends are required, each team is allowed 10 minutes of playing time to complete its throws and one added 60-second timeout for each extra end. However, the "thinking time" system, in which the delivering team's game timer stops as soon as the shooter's rock crosses the t-line during the delivery, is becoming more popular, especially in Canada. This system allows each team 38 minutes per 10 ends, or 30 minutes per 8 ends, to make strategic and tactical decisions, with 4 minutes and 30 seconds an end for extra ends. The "thinking time" system was implemented after it was recognized that using shots which take more time for the stones to come to rest was being penalized in terms of the time the teams had available compared to teams which primarily use hits which require far less time per shot. The process of sliding a stone down the sheet is known as the "delivery" or "throw". The players, with the exception of the skip, take turns throwing and sweeping; when one player (e.g., the lead) throws, the players not delivering (the second and third) sweep (see Sweeping, below). When the skip throws the vice-skip takes his or her role. The "skip", or the captain of the team, determines the desired stone placement and the required "weight", "turn", and "line" that will allow the stone to stop there. The placement will be influenced by the tactics at this point in the game, which may involve taking out, blocking or tapping another stone. The skip may communicate the "weight", "turn", "line," and other tactics by calling or tapping a broom on the ice. In the case of a takeout, guard, or a tap, the skip will indicate the stones involved. Before delivery, the running surface of the stone is wiped clean and the path across the ice swept with the broom if necessary, because any dirt on the bottom of a stone or in its path can alter the trajectory and ruin the shot. Intrusion by a foreign object is called a "pick-up" or "pick". The thrower starts from the "hack". The thrower's "gripper" shoe (with the non-slippery sole) is positioned against one of the hacks; for a right-handed curler the right foot is placed against the left hack and vice versa for a left-hander. The thrower, now "in the hack", lines the body up with shoulders square to the skip's broom at the far end for "line". The stone is placed in front of the foot now in the hack. Rising slightly from the hack, the thrower pulls the stone back (some older curlers may actually raise the stone in this backward movement) then lunges smoothly out from the hack pushing the stone ahead while the slider foot is moved in front of the gripper foot, which trails behind. The thrust from this lunge determines the "weight" and hence the distance the stone will travel. Balance may be assisted by a broom held in the free hand with the back of the broom down so that it slides. One older writer suggests the player keep "a basilisk glance" at the mark. There are two common types of delivery currently, the typical flat-foot delivery and the Manitoba tuck delivery where the curler slides on the front ball of his foot. When the player releases the stone a rotation (called the "turn)" is imparted by a slight clockwise or counter-clockwise twist of the handle from around the two or ten o'clock position to the twelve o'clock on release. A typical rate of turn is about rotations before coming to a rest. The stone must be released before its front edge crosses the near hog line, and it must clear the far hog line or else be removed from play ("hogged"); an exception is made if a stone fails to come to rest beyond the far hog line after rebounding from a stone in play just past the hog line. In major tournaments the "eye on the hog" sensor is commonly used to enforce this rule. The sensor is in the handle of the stone and will indicate whether the stone was released before the near hog line. The lights on the stone handle will either light up green, indicating that the stone has been legally thrown, or red, in which case the illegally thrown stone will be immediately pulled from play instead of waiting for the stone to come to rest. After the stone is delivered, its trajectory is influenced by the two sweepers under instruction from the skip. Sweeping is done for several reasons: to make the stone travel farther, to decrease the amount of curl, and to clean debris from the stone's path. Sweeping is able to make the stone travel farther and straighter by slightly melting the ice under the brooms, thus decreasing the friction as the stone travels across that part of the ice. The stones curl more as they slow down, so sweeping early in travel tends to increase distance as well as straighten the path, and sweeping after sideways motion is established can increase the sideways distance. One of the basic technical aspects of curling is knowing when to sweep. When the ice in front of the stone is swept a stone will usually travel both farther and straighter and in some situations one of those is not desirable. For example, a stone may be traveling too fast (said to have too much weight) but require sweeping to prevent curling into another stone. The team must decide which is better: getting by the other stone but traveling too far or hitting the stone. Much of the yelling that goes on during a curling game are the skip and sweepers exchanging information about the stone's "line" and "weight" and deciding whether to sweep. The skip evaluates the path of the stone and calls to the sweepers to sweep as necessary to maintain the intended track. The sweepers themselves are responsible for judging the weight of the stone, ensuring the length of travel is correct and communicating the weight of the stone back to the skip. Many teams use a "number system" to communicate in which of 10 zones the sweepers estimate the stone will stop. Some sweepers use stopwatches to time the stone from the back line or tee line to the nearest hog line to aid in estimating how far the stone will travel. Usually, the two sweepers will be on opposite sides of the stone's path, although depending on which side the sweepers' strengths lie this may not always be the case. Speed and pressure are vital to sweeping. In gripping the broom, one hand should be one third of the way from the top (non-brush end) of the handle while the other hand should be one third of the way from the head of the broom. The angle of the broom to the ice should be so that the most force possible can be exerted on the ice. The precise amount of pressure may vary from relatively light brushing ("just cleaning" - to ensure debris will not alter the stone's path) to maximum-pressure scrubbing. Sweeping is allowed anywhere on the ice up to the "tee line", once the leading edge of a stone crosses the tee line only one player may sweep it. Additionally, if a stone is behind the tee line one player from the opposing team is allowed to sweep it. This is the only case that a stone may be swept by an opposing team member. In international rules, this player must be the skip; or if the skip is throwing, then the sweeping player must be the third. Occasionally, players may accidentally touch a stone with their broom or a body part. This is often referred to as "burning" a stone. Players touching a stone in such a manner are expected to call their own infraction as a matter of good sportsmanship. Touching a stationary stone when no stones are in motion (there is no delivery in progress) is not an infraction as long as the stone is struck in such a manner that its position is not altered, and is a common way for the skip to indicate a stone that is to be taken out. When a stone is touched when stones are in play, the remedies vary between leaving the stones as they end up after the touch, replacing the stones as they would have been if no stone were touched, or removal of the touched stone from play. In non-officiated league play, the skip of the non-offending team has the final say on where the stones are placed after the infraction. Many different types of shots are used to carefully place stones for strategic or tactical reasons; they fall into three fundamental categories as follows: Guards are thrown in front of the house in the "free guard zone", usually to protect a stone or to make the opposing team's shot difficult. Guard shots include the "centre-guard", on the centreline and the "corner-guards" to the left or right sides of the centre line. See "Free Guard Zone" below. Draws are thrown only to reach the house. Draw shots include "raise", "come-around", and "freeze" shots. Takeouts are intended to remove stones from play and include the "peel", "hit-and-roll" and "double" shots. For a more complete listing, see Glossary of curling terms. The "free guard zone" is the area of the curling sheet between the hog line and tee line, excluding the house. Until five stones have been played (three from the side without hammer, and two from the side with hammer), stones in the free guard zone may not be removed by an opponent's stone, although they can be moved within the playing area. If a stone in the free guard zone is knocked out of play, it is placed back in the position it was in before the shot was thrown and the opponent's stone is removed from play. This rule is known as the "five-rock rule" or the "free guard zone rule" (previous versions of the free guard zone rule only limited removing guards from play in the first three or four rocks). This rule, a relatively recent addition to curling, was added in response to a strategy by teams of gaining a lead in the game and then "peeling" all of the opponents' stones (knocking them out of play at an angle that caused the shooter's stone to also roll out of play, leaving no stones on the ice). By knocking all stones out the opponents could at best score one point, if they had the last stone of the end (called the hammer). If the team peeling the rocks had the hammer they could peel rock after rock which would "blank the end" (leave the end scoreless), keeping the last rock advantage for another end. This strategy had developed (mostly in Canada) as ice-makers had become skilled at creating a predictable ice surface and newer brushes allowed greater control over the rock. While a sound strategy, this made for an unexciting game. Observers at the time noted that if two teams equally skilled in the peel game faced each other on good ice, the outcome of the game would be predictable from who won the coin flip to have last rock (or had earned it in the schedule) at the beginning of the game. The 1990 Brier (Canadian men's championship) was considered by many curling fans as boring to watch because of the amount of peeling and the quick adoption of the free guard zone rule the following year reflected how disliked this aspect of the game had become. The free guard zone rule was originally called the Modified Moncton Rule and was developed from a suggestion made by Russ Howard for the Moncton 100 cashspiel in Moncton, New Brunswick, in January 1990. "Howard's Rule" (later known as the Moncton Rule), used for the tournament and based on a practice drill his team used, had the first four rocks in play unable to be removed no matter where they were at any time during the end. This method of play was altered by restricting the area in which a stone was protected to the free guard zone only for the first four rocks thrown and adopted as a four-rock free guard zone rule for international competition shortly after. Canada kept to the traditional rules until a three-rock free guard zone rule was adopted for the 1993–94 season. After several years of having the three-rock rule used for the Canadian championships and the winners then having to adjust to the four-rock rule in the World Championships, the Canadian Curling Association adopted the four-rock free guard zone in the 2002–2003 season. One strategy that has been developed by curlers in response to the free guard zone (Kevin Martin from Alberta is one of the best examples) is the "tick" game, where a shot is made attempting to knock (tick) the guard to the side, far enough that it is difficult or impossible to use but still remaining in play while the shot itself goes out of play. The effect is functionally identical to peeling the guard but significantly harder, as a shot that hits the guard too hard (knocking it out of play) results in its being replaced, while not hitting it hard enough can result in it still being tactically useful for the opposition. There is also a greater chance that the shot will miss the guard entirely because of the greater accuracy required to make the shot. Because of the difficulty of making this type of shot, only the best teams will normally attempt it, and it does not dominate the game the way the peel formerly did. Steve Gould from Manitoba popularized ticks played across the face of the guard stone. These are easier to make because they impart less speed on the object stone, therefore increasing the chance that it remains in play even if a bigger chunk of it is hit. With the tick shot reducing the effectiveness of the four-rock rule, the Grand Slam of Curling series of bonspiels adopted a five-rock rule in 2014. In 2017, the five-rock rule was adopted by the World Curling Federation and member organizations for official play, beginning in the 2018–19 season. The last rock in an end is called the "hammer" and throwing the hammer gives a team a tactical advantage. Before the game, teams typically decide who gets the hammer in the first end either by chance (such as a coin toss), by a "draw-to-the-button" contest, where a representative of each team shoots to see who gets closer to the centre of the rings, or, particularly in tournament settings like the Winter Olympics, by a comparison of each team's win-loss record. In all subsequent ends the team that did not score in the preceding end gets to throw second, thus having the hammer. In the event that neither team scores, called a "blanked end", the hammer remains with the same team. Naturally, it is easier to score points with the hammer than without; the team with the hammer generally tries to score two or more points. If only one point is possible, the skip may try to avoid scoring at all in order to retain the hammer the next end, giving the team another chance to use the hammer advantage to try to score two points. Scoring without the hammer is commonly referred to as "stealing", or "a steal", and is much more difficult. Curling is a game of strategy, tactics and skill. The strategy depends on the team's skill, the opponent's skill, the conditions of the ice, the score of the game, how many ends remain and whether the team has last-stone advantage (the "hammer"). A team may play an end aggressively or defensively. Aggressive playing will put a lot of stones in play by throwing mostly draws; this makes for an exciting game and is very risky but the reward can be very great. Defensive playing will throw a lot of hits preventing a lot of stones in play; this tends to be less exciting and less risky. A good drawing team will usually opt to play aggressively, while a good hitting team will opt to play defensively. If a team does not have the hammer in an end, it will opt to try to clog up the four-foot zone in the house to deny the opposing team access to the button. This can be done by throwing "centre line" guards in front of the house on the centre line, which can be tapped into the house later or drawn around. If a team has the hammer, they will try to keep this four-foot zone free so that they have access to the button area at all times. A team with the hammer may throw a "corner guard" as their first stone of an end placed in front of the house but outside the four-foot zone to utilize the free guard zone. Corner guards are key for a team to score two points in an end, because they can either draw around it later or hit and roll behind it, making the opposing team's shot to remove it more difficult. Ideally, the strategy in an end for a team with the hammer is to score two points or more. Scoring one point is often a wasted opportunity, as they will then lose last-rock advantage for the next end. If a team cannot score two points, they will often attempt to "blank an end" by removing any leftover opposition rocks and rolling out; or, if there are no opposition rocks, just throwing the rock through the house so that no team scores any points, and the team with the hammer can try again the next end to score two or more with it. Generally, a team without the hammer would want to either force the team with the hammer to only one point (so that they can get the hammer back) or "steal" the end by scoring one or more points of their own. Generally, the larger the lead a team will have in a game, the more defensively they should play. By hitting all of the opponent's stones, it removes opportunities for their getting multiple points, therefore defending the lead. If the leading team is quite comfortable, leaving their own stones in play can also be dangerous. Guards can be drawn around by the other team, and stones in the house can be tapped back (if they are in front of the tee line) or frozen onto (if they are behind the tee line). A frozen stone is difficult to remove, because it is "frozen" (in front of and touching) to the opponents stone. At this point, a team will opt for "peels", meaning that the stones they throw will be to not only hit their opposition stones, but to roll out of play as well. Peels are hits that are thrown with the most amount of power. It is not uncommon at any level for a losing team to terminate the match before all ends are completed if it believes it no longer has a realistic chance of winning. Competitive games end once the losing team has "run out of rocks"—that is, once it has fewer stones in play and available for play than the number of points needed to tie the game. Most decisions about rules are left to the skips, although in official tournaments, decisions may be left to the officials. However, all scoring disputes are handled by the vice skip. No players other than the vice skip from each team should be in the house while score is being determined. In tournament play, the most frequent circumstance in which a decision has to be made by someone other than the vice skip is the failure of the vice skips to agree on which stone is closest to the button. An independent official (supervisor at Canadian and World championships) then measures the distances using a specially designed device that pivots at the centre of the button. When no independent officials are available, the vice skips measure the distances. The winner is the team having the highest number of accumulated points at the completion of ten ends. Points are scored at the conclusion of each of these ends as follows: when each team has thrown its eight stones, the team with the stone closest to the button wins that end; the winning team is then awarded one point for each of its own stones lying closer to the button than the opponent's closest stone. Only stones that are "in the house" are considered in the scoring. A stone is in the house if it lies within the zone or any portion of its edge lies over the edge of the ring. Since the bottom of the stone is rounded, a stone just barely in the house will not have any actual contact with the ring, which will pass under the rounded edge of the stone, but it still counts. This type of stone is known as a "biter". It may not be obvious to the eye which of two rocks is closer to the button (centre) or if a rock is actually biting or not. There are specialized devices to make these determinations, but these cannot be brought out until after an end is completed. Therefore, a team may make strategic decisions during an end based on assumptions of rock position that turn out to be incorrect. The score is marked on a scoreboard, of which there are two types; the baseball type and the club scoreboard. The baseball-style scoreboard was created for televised games for audiences not familiar with the club scoreboard. The "ends" are marked by columns 1 through 10 (or 11 for the possibility of an extra end to break ties) plus an additional column for the total. Below this are two rows, one for each team, containing the team's score for that end and their total score in the right hand column. The club scoreboard is traditional and used in most curling clubs. Scoring on this board only requires the use of (up to) 11 digit cards, whereas with baseball-type scoring an unknown number of multiples of the digits (especially low digits like "1") may be needed. The numbered centre row represents all possible accumulated scores, and the numbers placed in the team rows represent the end in which that team achieved that cumulative score. If the red team scores three points in the first end (called a "three-ender"), then a 1 (indicating the first end) is placed beside the number 3 in the red row. If they score two more in the second end, then a 2 will be placed beside the 5 in the red row, indicating that the red team has five points in total (3+2). This scoreboard works because only one team can get points in an end. However, some confusion may arise if neither team scores points in an end, this is called a "blank end". The blank end numbers are usually listed in the farthest column on the right in the row of the team that has the "hammer" (last rock advantage), or on a special spot for blank ends. The following example illustrates the difference between the two types. The example illustrates the men's final at the 2006 Winter Olympics. Eight points – all the rocks thrown by one team counting – is the highest score possible in an end, and is known as an "eight-ender" or "snowman". Scoring an eight-ender against a relatively competent team is very difficult; in curling, it is considered the equivalent of pitching a perfect game in baseball. Probably the best-known snowman came at the 2006 Players' Championships. Future (2007) World Champion Kelly Scott scored eight points in one of her games against 1998 World bronze medalist Cathy King. Competition teams are normally named after the skip, for example, Team Martin after skip Kevin Martin. Amateur league players can (and do) creatively name their teams, but when in competition (a bonspiel) the official team will have a standard name. Top curling championships are typically played by all-male or all-female teams. It is known as mixed curling when a team consists of two men and two women. For many years, in the absence of world championship or Olympic mixed curling events, national championships (of which the Canadian Mixed Curling Championship was the most prominent) were the highest-level mixed curling competitions. However, a European Mixed Curling Championship was inaugurated in 2005, a World Mixed Doubles Curling Championship was established in 2008, and the European Mixed Championship was replaced with the World Mixed Curling Championship in 2015. A mixed tournament was held at the Olympic level for the first time in 2018, although it was a doubles tournament, not a four-person. Curling tournaments may use the Schenkel system for determining the participants in matches. Curling is played in many countries, including Canada, the United Kingdom (especially Scotland), the United States, Norway, Sweden, Switzerland, Denmark, Finland and Japan, all of which compete in the world championships. Curling has been depicted by many artists including: George Harvey, John Levack, The Dutch School, Charles Martin Hardie, John Elliot Maguire, John McGhie, and John George Brown. Curling is particularly popular in Canada. Improvements in ice making and changes in the rules to increase scoring and promote complex strategy have increased the already high popularity of the sport in Canada, and large television audiences watch annual curling telecasts, especially the Scotties Tournament of Hearts (the national championship for women), the Tim Hortons Brier (the national championship for men), and the women's and men's world championships. Despite the Canadian province of Manitoba's small population (ranked 5th of 10 Canadian provinces), Manitoban teams have won the Brier more times than teams from any other province. The Tournament of Hearts and the Brier are contested by provincial and territorial champions, and the world championships by national champions. Curling is the provincial sport of Saskatchewan. From there Ernie Richardson and his family team dominated Canadian and international curling during the late 1950s and early 1960s and have been considered to be the best male curlers of all time. Sandra Schmirler led her team to the first ever gold medal in women's curling in the 1998 Winter Olympics. When she died two years later from cancer, over 15,000 people attended her funeral, and it was broadcast on national television. More so than in many other team sports, good sportsmanship, often referred to as the "Spirit of Curling", is an integral part of curling. The Spirit of Curling also leads teams to congratulate their opponents for making a good shot, strong sweeping or spectacular form. Perhaps most importantly, the Spirit of Curling dictates that one never cheers mistakes, misses or gaffes by one's opponent (unlike most team sports) and one should not celebrate one's own good shots during the game beyond modest acknowledgement of the shot such as a head nod, fist bump or thumbs-up gesture. Modest congratulation, however, may be exchanged between winning team members after the match. On-the-ice celebration is usually reserved for the winners of a major tournament after winning the final game of the championship. It is completely unacceptable to attempt to throw opposing players off their game by way of negative comment, distraction or heckling. A match traditionally begins with players shaking hands with and saying "good curling" or "have a pleasant game" to each member of the opposing team. It is also traditional in some areas for the winning team to buy the losing team a drink after the game. Even at the highest levels of play, players are expected to call their own fouls. It is not uncommon for a team to concede a curling match after it believes it no longer has any hope of winning. Concession is an honourable act and does not carry the stigma associated with quitting, and also allows for more socializing. To concede a match, members of the losing team offer congratulatory handshakes to the winning team. Thanks, wishes of future good luck and hugs are usually exchanged between the teams. To continue playing when a team has no realistic chance of winning can be seen as a breach of etiquette. Curling has been adapted for wheelchair users and people otherwise unable to throw the stone from the hack. These curlers may use a device known as a "delivery stick". The cue holds on to the handle of the stone and is then pushed along by the curler. At the end of delivery, the curler pulls back on the cue, which releases it from the stone. The Canadian Curling Association "Rules of Curling" allows the use of a delivery stick in club play but does not permit it in championships. The delivery stick was specifically invented for elderly curlers in Canada in 1999. In early 2016 an international initiative started to allow use of the delivery sticks by players over 60 years of age in World Curling Federation Senior Championships, as well as in any projected Masters (60+) Championship that develops in the future. Terms used to describe the game include: The ice in the game may be "fast (keen)" or "slow". If the ice is keen, a rock will travel farther with a given amount of weight (throwing force) on it. The speed of the ice is measured in seconds. One such measure, known as "hog-to-hog" time, is the speed of the stone and is the time in seconds the rock takes from the moment it crosses the near hog line until it crosses the far hog line. If this number is lower, the rock is moving faster, so again low numbers mean more speed. The ice in a match will be somewhat consistent and thus this measure of speed can also be used to measure how far down the ice the rock will travel. Once it is determined that a rock taking (for example) 13 seconds to go from hog line to hog line will stop on the tee line, the curler can know that if the hog-to-hog time is matched by a future stone, that stone will likely stop at approximately the same location. As an example, on keen ice, common times might be 16 seconds for guards, 14 seconds for draws, and 8 seconds for peel weight. "The back line to hog line speed" is used principally by sweepers to get an initial sense of the weight of a stone. As an example, on keen ice, common times might be 4.0 seconds for guards, 3.8 seconds for draws, 3.2 for normal hit weight, and 2.9 seconds for peel weight. Especially at the club level, this metric can be misleading, due to amateurs sometimes pushing stones on release, causing the stone to travel faster than the back-to-hog speed. In the 19th century several private railway stations in the United Kingdom were built to serve curlers attending bonspiels, such as those at Aboyne, Carsbreck and Drummuir. The Beatles participate in a game of curling during one scene of their 1965 film "Help!". The villains booby-trap one of the curling stones with a bomb; George sees the "fiendish thingy" and tells everyone to run. The bomb eventually goes off after a delay, creating a big hole in the ice. Curling is featured prominently in "Boy Meets Curl", the twelfth episode of the comedy series "The Simpsons"' twenty-first season. The episode aired on the Fox network in the United States on 14 February 2010. "Men with Brooms" is a 2002 Canadian film that takes a satirical look at curling. A TV adaptation, also titled "Men with Brooms", debuted in 2010 on CBC Television. The "Corner Gas" episode "Hurry Hard" involves the townspeople of Dog River competing in a local curling bonspiel for the fictitious "Clavet Cup". The episode also features cameos by Canadian curlers Randy Ferbey and Dave Nedohin.
https://en.wikipedia.org/wiki?curid=6644
Craven Cottage Craven Cottage is a football stadium located in Fulham, London. It has been the home ground of Fulham F.C. since 1896. The ground's capacity was 25,700, all-seated, until the closure of the Riverside Stand in the close season of 2019, though the record attendance is 49,335, for a game against Millwall, 8 October 1938. Located next to Bishop's Park on the banks of the River Thames, 'Craven Cottage' was originally a royal hunting lodge and has history dating back over 300 years. As well as by Fulham, the stadium has been also been used by the United States men's national football team, Australia national football team, the Republic of Ireland national football team (for a friendly match in 2012), and Canada men's national football team, and was formerly the home ground for rugby league team . The original 'Cottage' was built in 1780, by William Craven, the sixth Baron Craven and was located on the centre circle of the pitch. At the time, the surrounding areas were woods which made up part of Anne Boleyn's hunting grounds. The Cottage was lived in by Edward Bulwer-Lytton (who wrote "The Last Days of Pompeii") and other somewhat notable (and moneyed) persons until it was destroyed by fire in May 1888. Many rumours persist among Fulham fans of past tenants of Craven Cottage. Sir Arthur Conan Doyle, Jeremy Bentham, Florence Nightingale and even Queen Victoria are reputed to have stayed there, although there is no real evidence for this. Following the fire, the site was abandoned. Fulham had had 8 previous grounds before settling in at Craven Cottage for good. Therefore, The Cottagers have had 12 grounds overall (including a temporary stay at Loftus Road), meaning that only their former 'landlords' and rivals QPR have had more home grounds (14) in British football. Of particular note, was Ranelagh House, Fulham's palatial home from 1886–1888. When representatives of Fulham first came across the land, in 1894, it was so overgrown that it took two years to be made suitable for football to be played on it. A deal was struck for the owners of the ground to carry out the work, in return for which they would receive a proportion of the gate receipts. The first football match at which there were any gate receipts was when Fulham played against Minerva in the Middlesex Senior Cup, on 10 October 1896. The ground's first stand was built shortly after. Described as looking like an "orange box", it consisted of four wooden structures each holding some 250 seats, and later was affectionately nicknamed the "rabbit hutch". In 1904 London County Council became concerned with the level of safety at the ground, and tried to get it closed. A court case followed in January 1905, as a result of which Archibald Leitch, a Scottish architect who had risen to prominence after his building of the Ibrox Stadium, a few years earlier, was hired to work on the stadium. In a scheme costing £15,000 (a record for the time), he built a pavilion (the present-day 'Cottage' itself) and the Stevenage Road Stand, in his characteristic red brick style. The stand on Stevenage Road celebrated its centenary in the 2005–2006 season and, following the death of Fulham FC's favourite son, former England captain Johnny Haynes, in a car accident in October 2005 the Stevenage Road Stand was renamed the Johnny Haynes Stand after the club sought the opinions of Fulham supporters. Both the Johnny Haynes Stand and Cottage remain among the finest examples of Archibald Leitch football architecture to remain in existence and both have been designated as Grade II listed buildings. An England v Wales match was played at the ground in 1907, followed by a rugby league international between England and Australia in 1911. One of the club's directors Henry Norris, and his friend William Hall, took over Arsenal in the early 1910s, the plan being to merge Fulham with Arsenal, to form a "London superclub" at Craven Cottage. This move was largely motivated by Fulham's failure thus far to gain promotion to the top division of English football. There were also plans for Henry Norris to build a larger stadium on the other side of Stevenage Road but there was little need after the merger idea failed. During this era, the Cottage was used for choir singing and marching bands along with other performances, and Mass. In 1933 there were plans to demolish the ground and start again from scratch with a new 80,000 capacity stadium. These plans never materialised mainly due to the Great Depression. On 8 October 1938, 49,335 spectators watched Fulham play Millwall. It was the largest attendance ever at Craven Cottage and the record remains today, unlikely to be bettered as it is now an all-seater stadium with currently no room for more than 25,700. The ground hosted several football games for the 1948 Summer Olympics, and is one of the last extant that did. It was not until after Fulham first reached the top division, in 1949, that further improvements were made to the stadium. In 1962 Fulham became the final side in the first division to erect floodlights. The floodlights were said to be the most expensive in Europe at the time as they were so modern. The lights were like large pylons towering 50 metres over the ground and were similar in appearance to those at the WACA. An electronic scoreboard was installed on the Riverside Terrace at the same time as the floodlights were installed and flagpoles flying the flags of all of the other first division teams were flown from them. Following the sale of Alan Mullery to Tottenham Hotspur in 1964 (for £72,500) the Hammersmith End had a roof put over it at a cost of approximately £42,500. Although Fulham was relegated, the development of Craven Cottage continued. The Riverside terracing, infamous for the fact that fans occupying it would turn their heads annually to watch The Boat Race pass, was replaced by what was officially named the 'Eric Miller Stand', Eric Miller being a director of the club at the time. The stand, which cost £334,000 and held 4,200 seats, was opened with a friendly game against Benfica in February 1972, (which included Eusébio). Pelé was also to appear on the ground, with a friendly played against his team Santos F.C. The Miller stand bought the seated capacity up to 11,000 out of a total 40,000. Eric Miller committed suicide five years later after a political and financial scandal, and had shady dealing with trying to move Fulham away from the Cottage. The stand is now better known as the Riverside Stand. On Boxing Day 1963, Craven Cottage was the venue of the fastest hat-trick in the history of the English football league, which was completed in less than three minutes, by Graham Leggat. This helped his Fulham team to beat Ipswich 10–1 (a club record). The international record is held by Jimmy O'Connor, an Irish player who notched up his hat trick in 2 minutes 14 seconds in 1967. Between 1980 and 1984, Fulham rugby league played their home games at the Cottage. They have since evolved into the London Crusaders, the London Broncos and Harlequins Rugby League before reverting to London Broncos ahead of the 2012 season. Craven Cottage held the team's largest ever crowd at any ground with 15,013, at a game against Wakefield Trinity on 15 February 1981. When the Hillsborough disaster occurred in 1989, Fulham were in the second bottom rung of The Football League, but following the Taylor report Fulham's ambitious chairman Jimmy Hill tabled plans in 1996 for an all-seater stadium. These plans never came to fruition, partly due to local residents' pressure groups, and by the time Fulham reached the Premiership, they still had standing areas in the ground, something virtually unheard of at the time. A year remained to do something about this (teams reaching the second tier for the first time are allowed a three-year period to reach the required standards for the top two divisions), but by the time the last league game was played there, against Leicester City on 27 April 2002, no building plans had been made. Two more Intertoto Cup games were played there later that year (against FC Haka of Finland and Egaleo FC of Greece), and the eventual solution was to decamp to Loftus Road, home of local rivals QPR. During this time, many Fulham fans only went to away games in protest of moving from Craven Cottage. 'Back to the Cottage', later to become the 'Fulham Supporters Trust', was set up as a fans pressure group to encourage the chairman and his advisers that Craven Cottage was the only viable option for Fulham Football Club. After one and a half seasons at Loftus Road, no work had been done on the Cottage. In December 2003, plans were unveiled for £8million worth of major refurbishment work to bring it in line with Premier League requirements. With planning permission granted, work began in January 2004 in order to meet the deadline of the new season. The work proceeded as scheduled and the club were able to return to their home for the start of the 2004–05 season. Their first game in the new-look 22,000 all-seater stadium was a pre-season friendly against Watford on 10 July 2004. Fenway Sports Group originally partnered with Fulham in 2009, due to the perceived heritage and quirks shared between the Cottage and Fenway Park, saying no English club identifies with its stadium as much as Fulham. Boston Red Sox were later in bed with Liverpool FC, after the takeover by Fenway Sports Group. The current stadium was one of the Premiership's smallest grounds at the time of Fulham's relegation at the end of the 2013–14 season (it was third-smallest, after the KC Stadium and the Liberty Stadium). Much admired for its fine architecture, the stadium has recently hosted a few international games, mostly including Australia. This venue is suitable for Australia because most of the country's top players are based in Europe, and West London has a significant community of expatriate Australians. Also, Greece vs. South Korea was hosted on 6 February 2007. In 2011 Brazil played Ghana, in an international friendly, and the Women's Champions League Final was hosted. Craven Cottage often hosts many other events such as 5-a-side football tournaments and weddings. Also, many have Sunday Lunch at the Riverside restaurant or the 'Cottage Cafe' on non-match days. Craven Cottage hosted the Oxbridge Varsity Football match annually between 1991 and 2000 and again in 2003, 2006 (the same day as the famous 'Boat Race'), 2008, 2009, and 2014 as well as having a Soccer Aid warm-up match in 2006. The half-time entertainment often includes the SW6ers (previously called The Cravenettes) which are a group of female cheerleaders. However, other events have included brass bands, Michael Jackson (albeit just walking on the pitch once, as opposed to performing), Travis playing, Arabic dancing, keepie uppie professionals and presentational awards. Most games also feature the 'Fulham flutter', a half-time draw; and a shoot-out competition of some kind, usually involving scoring through a 'hoop' or 'beat the goalie'. On the first home game of the season, there is a carnival where every Fulham fan is expected to turn up in black-and-white colours. There is usually live rock bands, player signings, clowns, stilt walkers, a steel (calypso) band, food stalls and a free training session for children in Bishops Park. The Fulham Ladies (before their demise) and Reserve teams occasionally play home matches at the Cottage. Other than this, they generally play at the club's training ground at Motspur Park or at Kingstonian and AFC Wimbledon's stadium, Kingsmeadow. Craven Cottage is known by several affectionate nicknames from fans, including: The (River) Cottage, The Fortress (or Fortress Fulham), Thameside, The Friendy Confines, SW6, Lord of the Banks, The House of Hope, The Pavilion of Perfection, The 'True' Fulham Palace and The Palatial Home. The Thames at the banks of the Cottage is often referred to as 'Old Father' or The River of Dreams. The most accessible route to the ground is to walk through Bishops Park from Putney Bridge (the nearest Underground station), often known as 'The Green Mile' by Fulham fans (as it is roughly a mile walk through pleasant greenery). The Telegraph ranked the Cottage 9th out of 54 grounds to hold Premiership football. On 27 July 2012, Fulham FC were granted permission to redevelop the Riverside Stand, increasing the capacity of Craven Cottage to 30,000 seats. Beforehand various rumours arose including plans to return to ground-sharing with QPR in a new 40,000 seater White City stadium, although these now appear firmly on hold with the construction of the Westfield shopping centre on the proposed site. The board seem to have moved away from their ambition to make Fulham the "Manchester United of the south" as it became clear how expensive such a plan would be. With large spaces of land at a premium in south-west London, Fulham appear to be committed to a gradual increase of the ground's capacity often during the summer between seasons. The capacity of Craven Cottage has been increased during summers for instance in 2008 with a small increase in the capacity of the Hammersmith End. Fulham previously announced in 2007 that they are planning to increase the capacity of Craven Cottage by 4,000 seats, but this is yet to be implemented. There was also proposals for a bridge to span the Thames, for a redeveloped Riverside stand and a museum. More substantial plans arose in October 2011 with the 'Fulham Forever' campaign. With Al-Fayed selling Harrods department store for £1.5 billion in May 2010 a detailed plan emerged in the Riverside Stand as the only viable area for expansion. The scheme involved the demolition of the back of the Riverside Stand with a new tier of seating added on top of the current one and a row of corporate boxes; bringing Craven Cottage up to 30,000 capacity. Taking into account local residents, the proposal would: reopen the riverside walk; light pollution would be reduced with the removal of floodlight masts; new access points would make match-day crowds more manageable; and the new stand would be respectful in design to its position on the River Thames. Buckingham Group Contracting were chosen in March 2013 as the construction company for the project. In May 2019, the club confirmed that work on the new Riverside Stand would commence in the summer of 2019. During construction, set to occur through the 2020-21 season, the ground's capacity will be temporarily reduced to 19,000. The Hammersmith End (or Hammy) is the northernmost stand in the ground, the closest to Hammersmith. The roofing was financed through the sale of Alan Mullery to Tottenham Hotspur F.C. It is traditionally the "home" end where the more vocal Fulham fans sit, and many stand during games at the back rows of the stand. If Fulham win the toss, they usually choose to play towards the Hammersmith End in the second half. The hardcore fans tend to sit (or rather stand) in the back half of H6 and H7 zones (known as ‘H Block’ to the faithful). The stand had terracing until the reopening of the ground in 2004, when it was replaced with seating in order to comply with league rules following the Taylor Report. The Putney End is the southernmost stand in the ground, nearest to Putney and backing onto Bishops Park. This stand currently a mixture of home and away fans, separated by a "wall" of stewards, with away fans usually allocated blocks P5 and P6. When the ground became redeveloped, with the standing terraces replaced in 2003–04 the club applied for a licence to have a designated neutral area, in the rows closest to the Cottage, but this currently doesn’t exist due to the reduced capacity. Due to Fulham's history of having no segregation in the Putney End and no history of crowd trouble, the FA gave the club special dispensation to allow for this, making Fulham the only club currently in the UK to have such an area. Flags of every nationality in the Fulham squad were hung from the roofing, although they were removed after the 2006–07 season commenced and there is now an electronic scoreboard in place. There is a plane tree in the corner by the river. The Riverside was originally terracing that backed onto the Thames. It also featured large advertising hoardings above the fans. In 1971-72, an all-seater stand was built, originally known as the Riverside Stand (the name was confirmed in the Fulham v Carlisle United programme on 4 December 1971). Its hard lines and metallic and concrete finish are in stark contrast to the Johnny Haynes stand opposite. The stand was opened for a prestigious friendly against S.L. Benfica, who included Eusébio in the team. In the Fulham v Burnley programme on 4 October 1977, it was revealed that the stand would be re-named the Eric Miller Stand, following the recent death of the former vice-chairman. It is sometimes incorrectly stated that, contrary to the above, the name of the stand was changed from the Eric Miller Stand to The Riverside Stand after the discovery of Miller's suicide. He had been under investigation for fraud and embezzlement. The name of the stand actually reverted to "Riverside Stand" in the 1990s. The Riverside Stand backs onto the River Thames and is elevated above pitch level, unlike the other three stands. It contains the corporate hospitality seating alongside Fulham fans. Jimmy Hill once referred to the Riverside being "a bit like the London Palladium" as Blocks V & W (the middle section) are often filled with the rich and famous (including often Al-Fayed). There were then several Harrods advertising hoardings. Above the advertising hoardings is the gantry, for the press and cameras. Tickets in this area are often the easiest to buy, not surprisingly they are also some of the more expensive. It has the Hammersmith End to its left, the Putney End to its right and is opposite the Johnny Haynes Stand. During the 1970s, Craven Cottage flooded, with water gushing in from the Riverside. The stand houses the George Cohen restaurant, while on non-match days there is the Cottage Cafe, located near to the Cottage itself. (The River Café is also located nearby). Under Tommy Trinder's chairmanship in the 60s, flags of all other teams in the Division 1 were proudly flown along the Thames. However, when Fulham were relegated in 1968, Trinder decided not to change the flags as "Fulham won't be in this division next season". True to Trinder's prophecy, Fulham were relegated again. The roof of the stand has been used by sponsors, with VisitFlorida currently advertising in this way, and Pipex.com, FxPro, Lee Cooper Jeans and LG having previously done so. The end of the Riverside Stand towards the 'Hammy End' indicates the end of the 'Fulham Wall', a landmark in The Boat Race. From the 2019-20 season, this stand is being demolished and rebuilt, which will expand the capacity of the ground to around 29,600. It is due for completion for the beginning of the 2021-22 season. On 26 November 2019, the Chairman Shahid Khan provided an update in which it was announced that the new development will be known as Fulham Pier, a destination venue outside of match-day use. Originally called the Stevenage Road Stand, after the road it backs onto, the Johnny Haynes stand is the oldest remaining football stand in the Football League and professional football, originally constructed in 1905, and is a Grade II listed building. Designed by Archibald Leitch, the stand contains the ticket office and club shop and features original 'Bennet' wooden seating. Following his death in 2005, the stand was renamed after former player Johnny Haynes. The exterior facing Stevenage Road has a brick façade and features the club's old emblem in the artwork. Decorative pillars show the club's foundation date as 1880 though this is thought to be incorrect. Also, a special stone to commemorate the supporters' fund-raising group Fulham 2000, and The Cottagers' return to Craven Cottage, was engraved on the façade. The family enclosures are located in the two corners of the stand, one nearest to the Hammersmith End and one nearest to the Putney End. The front of the stand now contains plastic seating, but originally was a standing area. Children were often placed at the front of this enclosure and the area had a distinctive white picket fence to keep fans off the pitch up until the 1970s. The Cottage Pavilion dates back to 1905 along with the Johnny Haynes Stand, built by renowned football architect Archibald Leitch. Besides being the changing rooms, the Cottage (also called The Clubhouse) is traditionally used by the players' families and friends who sit on the balcony to watch the game. In the past, board meetings used to be held in The Cottage itself as well. There is a large tapestry draped from the Cottage which says "Still Believe". It encapsulates the now-famous moment, when fans facing defeat against Hamburg SV in the Europa League semi-final roused the players with the chant of "Stand up if you still believe". In the three other corners of the ground there are what have been described as large 'filing cabinets', which are corporate boxes on three levels, but currently the box on the other side of the Putney End has been removed due to the redevelopment of the Riverside Craven Cottage hosted the Northern Ireland versus Cyprus 1974 World Cup Qualifier on 8 May 1973, a match moved from Belfast due to The Troubles. Northern Ireland won 3-0, Sammy Morgan and a Trevor Anderson brace concluded the scoring in the first half. On 22 February 2000, it hosted England’s under 21s international under 21 friendly against Argentina’s under 21s. The hosts won 1–0 with Lee Hendrie’s sixty seventh-minute goal with 15,747 in attendance. In recent years, Craven Cottage has hosted several International Friendly matches, including the Ireland national team who played Colombia and Nigeria there in May 2008 and May 2009 respectively and Oman in 2012. The South Korea national football team have also used the ground thrice in recent years for international friendlies, first against Greece in February 2007 second against Serbia in November 2009, and then against Croatia in February 2013. On 17 November 2007 Australia beat Nigeria 1–0 in an international friendly at Craven Cottage. On 26 May 2011, Craven Cottage hosted the game of 2011 UEFA Women's Champions League Final between Lyon and Potsdam. In September 2011, a friendly between Ghana and Brazil was also held at Craven Cottage. On 15 October 2013, Australia beat Canada 3–0 at Craven Cottage. On 28 May 2014 Scotland played out a 2–2 draw with a Nigerian team who had qualified for the 2014 World Cup Finals. On 27 March 2018, Australia played host to Colombia in the international friendlies, the match ended 0-0, both teams having qualified for the 2018 World Cup Finals in Russia.
https://en.wikipedia.org/wiki?curid=6645
Cedar Falls, Iowa Cedar Falls is a city in Black Hawk County, Iowa, United States. As of the 2010 census, the city population was 39,260. It is home to the University of Northern Iowa, a public university. Cedar Falls was founded in 1845 by William Sturgis. It was originally named Sturgis Falls, for the first family who settled the site and who continued to live in the city for years. The city was called Sturgis Falls until it was merged with Cedar City (another city on the other side of the Cedar River), creating Cedar Falls. The city's founders are honored each year with a week long community-wide celebration named in their honor – the Sturgis Falls Celebration. Because of the availability of water power, Cedar Falls developed as a milling and industrial center prior to the Civil War. The establishment of the Civil War Soldiers' Orphans Home in Cedar Falls changed the direction in which the city developed when, following the war, it became the first building on the campus of the Iowa State Normal School (now the University of Northern Iowa). Cedar Falls is located at (42.523520, −92.446402). According to the United States Census Bureau, the city has a total area of , of which is land and is water. Natural forest, prairie and wetland areas are found within the city limits at the Hartman Reserve Nature Center. Cedar Falls is part of the Waterloo-Cedar Falls metropolitan area. As of the census of 2010, there were 39,260 people, 14,608 households, and 8,091 families living in the city. The population density was . There were 15,477 housing units at an average density of . The racial makeup of the city was 93.4% White, 2.1% African American, 0.2% Native American, 2.3% Asian, 0.5% from other races, and 1.7% from two or more races. Hispanic or Latino of any race were 2.0% of the population. There were 14,608 households of which 24.8% had children under the age of 18 living with them, 45.5% were married couples living together, 7.2% had a female householder with no husband present, 2.7% had a male householder with no wife present, and 44.6% were non-families. 28.0% of all households were made up of individuals and 10.4% had someone living alone who was 65 years of age or older. The average household size was 2.37 and the average family size was 2.88. The median age in the city was 26.8 years. 17.3% of residents were under the age of 18; 29.7% were between the ages of 18 and 24; 20.5% were from 25 to 44; 20.1% were from 45 to 64; and 12.4% were 65 years of age or older. The gender makeup of the city was 48.1% male and 51.9% female. As of the census of 2000, there were 36,145 people, 12,833 households, and 7,558 families living in the city. The population density was 1,277.2 people per square mile (493.1 per km2). There were 13,271 housing units at an average density of 468.9 per square mile (181.1 per km2). The racial makeup of the city was 95.14% White, 1.57% Black or African American, 0.15% Native American, 1.61% Asian, 0.02% Pacific Islander, 0.41% from other races, and 1.09% from two or more races. 1.08% of the population were Hispanic or Latino of any race. There were 12,833 households out of which 26.9% had children under the age of 18 living with them, 48.9% were married couples living together, 7.5% had a female householder with no husband present, and 41.1% were non-families. 25.5% of all households were made up of individuals and 9.4% had someone living alone who was 65 years of age or older. The average household size was 2.45 and the average family size was 2.91. Age spread: 18.0% under the age of 18, 30.6% from 18 to 24, 20.5% from 25 to 44, 19.0% from 45 to 64, and 11.9% who were 65 years of age or older. The median age was 26 years. For every 100 females, there were 88.5 males. For every 100 females age 18 and over, there were 85.7 males. The median income for a household in the city was $70,226, and the median income for a family was $85,158. Males had a median income of $60,235 versus $50,312 for females. The per capita income for the city was $27,140. About 5.6% of families and 4.7% of the population were below the poverty line, including 8.5% of those under age 18, and 6.1% of those age 65 or over. In 1986, the City of Cedar Falls established the Cedar Falls Art and Culture Board, which oversees the operation of the City's Cultural Division and the James & Meryl Hearst Center for the Arts. The Cedar Falls Public Library is housed in the Adele Whitenach Davis building located at 524 Main Street. The 47,000 square foot (4,400 m2) structure, designed by Struxture Architects, replaced the Carniege-Dayton building in early 2004. As of the 2016 fiscal year, the library's holdings included approximately 8,000 audio materials, 12,000 video materials, and 104,000 books and periodicals for a grand total of approximately 124,000 items. Patrons made 245,000 visits which took advantage of circulation services, adult, teen, and youth programming. Circulation of library materials for fiscal year 2016 was 543,134. The library also provides public access to more than 30 public computers which provide internet access, office software suites, high resolution color printing, wi-fi, and various games. The mission of the Cedar Falls Public Library is to promote literacy and provide open access to resources which facilitate lifelong learning. The library is a member of the Cedar Valley Library Consortium. Cedar Falls Public Library shares an Integrated Library System (SirsiDynix Symphony) with the Waterloo Public Library. Library management is provided by Kelly Stern, Director of the Cedar Falls Public Library. The Cedar Falls Historical Society has its offices in the Victorian Home and Carriage House Museum. It preserves Cedar Falls' history through its five museums, collection, archives, and public programs. Besides the Victorian House, the Society operates the Cedar Falls Ice House, Little Red Schoolhouse, and Behrens-Rapp Station. It hosts one of three public universities in Iowa, University of Northern Iowa (UNI). Cedar Falls Community Schools, which covers most of the city limits, includes Cedar Falls High School, two junior high schools, seven elementary schools. Waterloo Community School District covers a small section of Cedar Falls. There is a private Christian school, Valley Lutheran High School. Additionally there is a private Catholic elementary school at St. Patrick Catholic Church, under the Roman Catholic Archdiocese of Dubuque. A significant renovation occurred beginning in May 2014. The Malcolm Price Lab School/Northern University High School, was a state-funded K-12 school run by the university. It closed in 2012 following cuts at UNI. The city owns its power, gas and water, and cable TV service. Because of this, Cedar Falls Utilities provides gigabit speeds to residents, this became available on January 14, 2015. Cedar Falls has the power to do so because, unlike 19 other states, Iowa does not prohibit municipal broadband from competing with the private cable TV monopoly. In 2020, Cedar Falls Utilities was recognized by PC Magazine as having the nation's fastest internet, by a factor of three. The underground music scene in the Cedar Falls area from 1977 to present-day is well documented. The Wartburg College Art Gallery in Waverly, Iowa hosted a collaborative history of the bands, record labels, and music venues involved in the Cedar Falls music scene which ran from March 17 to April 14, 2007. This effort has been continued as a wiki-style website called "The Secret History of the Cedar Valley".
https://en.wikipedia.org/wiki?curid=6651
Cleveland Indians The Cleveland Indians are an American professional baseball team based in Cleveland, Ohio. They compete in Major League Baseball (MLB) as a member club of the American League (AL) Central division. Since , they have played at Progressive Field. The team's spring training facility is at Goodyear Ballpark in Goodyear, Arizona. Since their establishment as a major league franchise in 1901, the team has won 10 Central Division titles, six American League pennants, and two World Series championships, (in 1920 and 1948). The team's current 71-year World Series championship drought is the longest active among all 30 current Major League teams. The name "Indians" originated from a request by club owner Charles Somers to baseball writers to choose a new name to replace "Cleveland Naps" following the departure of Nap Lajoie after the 1914 season. It was a revival of the nickname that fans gave to the Cleveland Spiders while Louis Sockalexis, a Native American, was playing for the team. Common nicknames for the Indians include the "Tribe" and the "Wahoos", the latter referencing their former logo, Chief Wahoo. The team's mascot is named "Slider." The franchise originated in 1894 as the Grand Rapids Rustlers, a minor league team in the Western League. The team relocated to Cleveland in 1900 and was renamed the Cleveland Lake Shores. The Western League itself was renamed the American League while continuing its minor league status. One of the American League's eight charter franchises, the major league incarnation of the club was founded in Cleveland in 1901. Originally called the Cleveland Bluebirds, the team played in League Park until moving permanently to Cleveland Stadium in 1946. From August 24 to September 14, 2017, the Indians won 22 consecutive games, the longest winning streak in American League history. For 1901–2019, the Indians overall record is (). ""In 1857 baseball games were a daily spectacle in Cleveland's Public Squares. City authorities tried to find an ordinance forbidding it, to the joy of the crowd, they were unsuccessful. – Harold Seymour"" From 1865 to 1868 Forest Citys was an amateur ball club. During the 1869 season, Cleveland was among several cities that established professional baseball teams following the success of the 1869 Cincinnati Red Stockings, the first fully professional team. In the newspapers before and after 1870, the team was often called the Forest Citys, in the same generic way that the team from Chicago was sometimes called The Chicagos. In 1871 the Forest Citys joined the new National Association of Professional Base Ball Players (NA), the first professional league. Ultimately, two of the league's western clubs went out of business during the first season and the Chicago Fire left that city's White Stockings impoverished, unable to field a team again until 1874. Cleveland was thus the NA's westernmost outpost in 1872, the year the club folded. Cleveland played its full schedule to July 19 followed by two games versus Boston in mid-August and disbanded at the end of the season. In 1876, the National League (NL) supplanted the NA as the major professional league. Cleveland was not among its charter members, but by 1879 the league was looking for new entries and the city gained an NL team. The Cleveland Forest Citys were recreated, but rebranded in 1882 as the Cleveland Blues, because the National League required distinct colors for that season. The Blues had mediocre records for six seasons and were ruined by a trade war with the Union Association (UA) in 1884, when its three best players (Fred Dunlap, Jack Glasscock, and Jim McCormick) jumped to the UA after being offered higher salaries. The Cleveland Blues merged with the St. Louis Maroons UA team in 1885. Cleveland went without major league baseball for two seasons until gaining a team in the American Association (AA) in 1887. After the AA's Allegheny club jumped to the NL, Cleveland followed suit in 1889, as the AA began to crumble. The Cleveland ball club, named the Spiders (supposedly inspired by their "skinny and spindly" players) slowly became a power in the league. The next year the Spiders moved into League Park, which would serve as the home of Cleveland professional baseball for the next 55 years. Led by native Ohioan Cy Young, the Spiders became a contender in the mid-1890s, playing in the Temple Cup Series (that era's World Series) twice and winning it in 1895. The team began to fade after this success, and was dealt a severe blow under the ownership of the Robison brothers. Prior to the season, Frank Robison, the Spiders' owner, bought the St. Louis Browns, thus owning two clubs at the same time. The Browns were renamed the "Perfectos", and restocked with Cleveland talent. Just weeks before the season opener, most of the better Spiders were transferred to St. Louis, including three future Hall of Famers: Cy Young, Jesse Burkett and Bobby Wallace. The roster maneuvers failed to create a powerhouse Perfectos team, as St. Louis finished fifth in both 1899 and . The Spiders were left with essentially a minor league lineup, and began to lose games at a record pace. Drawing almost no fans at home, they ended up playing most of their season on the road, and became known as "The Wanderers." The team ended the season in 12th place, 84 games out of first place, with an all-time worst record of 20-134 (.130 winning percentage). Following the 1899 season, the National League disbanded four teams, including the Spiders franchise. The disastrous 1899 season would actually be a step toward a new future for Cleveland fans the next year. The Cleveland Infants competed in the Players' League, which was well-attended in some cities, but club owners lacked the confidence to continue beyond the one season. The Cleveland Infants finished with 55 wins and 75 losses, playing their home games at Brotherhood Park. The Grand Rapids Rustlers were founded in Michigan in 1894 and were part of the Western League. In 1900 the team moved to Cleveland and was named the Cleveland Lake Shores. Around the same time Ban Johnson changed the name of his minor league (Western League) to the American League. In 1900 the American League was still considered a minor league. In 1901 team was renamed the Cleveland Bluebirds when the American League broke with the National Agreement and declared itself a competing Major League. The Cleveland franchise was among its eight charter members, and is one of four teams that remain in its original city, along with Boston, Chicago, and Detroit. The new team was owned by coal magnate Charles Somers and tailor Jack Kilfoyl. Somers, a wealthy industrialist and also co-owner of the Boston Americans, lent money to other team owners, including Connie Mack's Philadelphia Athletics, to keep them and the new league afloat. Players didn't think the name "Bluebirds" was suitable for a baseball team. Writers frequently shortened it to Cleveland Blues due to the players' all-blue uniforms, but the players didn't like this unofficial name either. The players themselves tried to change the name to Cleveland Broncos in , but this unofficial name never really caught on. The Bluebirds suffered from financial problems in their first two seasons. This led Somers to seriously consider moving to either Pittsburgh or Cincinnati. Relief came in 1902 as a result of the conflict between the National and American Leagues. In 1901, Napoleon "Nap" Lajoie, the Philadelphia Phillies' star second baseman, jumped to the A's after his contract was capped at $2,400 per year—one of the highest-profile players to jump to the upstart AL. The Phillies subsequently filed an injunction to force Lajoie's return, which was granted by the Pennsylvania Supreme Court. The injunction appeared to doom any hopes of an early settlement between the warring leagues. However, a lawyer discovered that the injunction was only enforceable in the state of Pennsylvania. Mack, partly to thank Somers for his past financial support, agreed to trade Lajoie to the then-moribund Blues, who offered $25,000 salary over three years. Due to the injunction, however, Lajoie had to sit out any games played against the A's in Philadelphia. Lajoie arrived in Cleveland on June 4 and was an immediate hit, drawing 10,000 fans to League Park. Soon afterward, he was named team captain, and in 1903 the team was renamed the Cleveland Napoleons (soon shortened to Naps) after a newspaper conducted a write-in contest. Lajoie was named manager in , and the team's fortunes improved somewhat. They finished half a game short of the pennant in 1908. However, the success did not last and Lajoie resigned during the 1909 season as manager but remained on as a player. After that, the team began to unravel, leading Kilfoyl to sell his share of the team to Somers. Cy Young, who returned to Cleveland in 1909, was ineffective for most of his three remaining years and Addie Joss died from tubercular meningitis prior to the 1911 season. Despite a strong lineup anchored by the potent Lajoie and Shoeless Joe Jackson, poor pitching kept the team below third place for most of the next decade. One reporter referred to the team as the Napkins, "because they fold up so easily". The team hit bottom in 1914 and 1915, finishing in the cellar both years. 1915 brought significant changes to the team. Lajoie, nearly 40 years old, was no longer a top hitter in the league, batting only .258 in 1914. With Lajoie engaged in a feud with manager Joe Birmingham, the team sold Lajoie back to the A's. With Lajoie gone, the club needed a new name. Somers asked the local baseball writers to come up with a new name, and based on their input, the team was renamed the Cleveland Indians. The name referenced the nickname "Indians" that was applied to the Cleveland Spiders baseball club during the time when Louis Sockalexis, a Native American, played in Cleveland (1897–99). At the same time, Somers' business ventures began to fail, leaving him deeply in debt. With the Indians playing poorly, attendance and revenue suffered. Somers decided to trade Jackson midway through the 1915 season for two players and $31,500, one of the largest sums paid for a player at the time. By 1916, Somers was at the end of his tether, and sold the team to a syndicate headed by Chicago railroad contractor James C. "Jack" Dunn. Manager Lee Fohl, who had taken over in early 1915, acquired two minor league pitchers, Stan Coveleski and Jim Bagby and traded for center fielder Tris Speaker, who was engaged in a salary dispute with the Red Sox. All three would ultimately become key players in bringing a championship to Cleveland. Speaker took over the reins as player-manager in , and led the team to a championship in 1920. On August 16, 1920, the Indians were playing the Yankees at the Polo Grounds in New York. Shortstop Ray Chapman, who often crowded the plate, was batting against Carl Mays, who had an unusual underhand delivery. It was also late in the afternoon and the infield was completely shaded with the center field area (the batters' background) bathed in sunlight. As well, at the time, "part of every pitcher's job was to dirty up a new ball the moment it was thrown onto the field. By turns, they smeared it with dirt, licorice, tobacco juice; it was deliberately scuffed, sandpapered, scarred, cut, even spiked. The result was a misshapen, earth-colored ball that traveled through the air erratically, tended to soften in the later innings, and as it came over the plate, was very hard to see." In any case, Chapman did not move reflexively when Mays' pitch came his way. The pitch hit Chapman in the head, fracturing his skull. Chapman died the next day, becoming the only player to sustain a fatal injury from a pitched ball. The Indians, who at the time were locked in a tight three-way pennant race with the Yankees and White Sox, were not slowed down by the death of their teammate. Rookie Joe Sewell hit .329 after replacing Chapman in the lineup. In September 1920, the Black Sox Scandal came to a boil. With just a few games left in the season, and Cleveland and Chicago neck-and-neck for first place at 94–54 and 95–56 respectively, the Chicago owner suspended eight players. The White Sox lost two of three in their final series, while Cleveland won four and lost two in their final two series. Cleveland finished two games ahead of Chicago and three games ahead of the Yankees to win its first pennant, led by Speaker's .388 hitting, Jim Bagby's 30 victories and solid performances from Steve O'Neill and Stan Coveleski. Cleveland went on to defeat the Brooklyn Robins 5–2 in the World Series for their first title, winning four games in a row after the Robins took a 2–1 Series lead. The Series included three memorable "firsts", all of them in Game 5 at Cleveland, and all by the home team. In the first inning, right fielder Elmer Smith hit the first Series grand slam. In the fourth inning, Jim Bagby hit the first Series home run by a pitcher. And in the top of the fifth inning, second baseman Bill Wambsganss executed the first (and only, so far) unassisted triple play in World Series history, in fact the only Series triple play of any kind. The team would not reach the heights of 1920 again for 28 years. Speaker and Coveleski were aging and the Yankees were rising with a new weapon: Babe Ruth and the home run. They managed two second-place finishes but spent much of the decade in the cellar. In 1927 Dunn's widow, Mrs. George Pross (Dunn had died in 1922), sold the team to a syndicate headed by Alva Bradley. The Indians were a middling team by the 1930s, finishing third or fourth most years. brought Cleveland a new superstar in 17-year-old pitcher Bob Feller, who came from Iowa with a dominating fastball. That season, Feller set a record with 17 strikeouts in a single game and went on to lead the league in strikeouts from 1938–1941. On August 20, 1938, Indians catchers Hank Helf and Frank Pytlak set the "all-time altitude mark" by catching baseballs dropped from the Terminal Tower. By , Feller, along with Ken Keltner, Mel Harder and Lou Boudreau, led the Indians to within one game of the pennant. However, the team was wracked with dissension, with some players (including Feller and Mel Harder) going so far as to request that Bradley fire manager Ossie Vitt. Reporters lampooned them as the Cleveland Crybabies. Feller, who had pitched a no-hitter to open the season and won 27 games, lost the final game of the season to unknown pitcher Floyd Giebell of the Detroit Tigers. The Tigers won the pennant and Giebell never won another major league game. Cleveland entered 1941 with a young team and a new manager; Roger Peckinpaugh had replaced the despised Vitt; but the team regressed, finishing in fourth. Cleveland would soon be depleted of two stars. Hal Trosky retired in 1941 due to migraine headaches and Bob Feller enlisted in the Navy two days after the Attack on Pearl Harbor. Starting third baseman Ken Keltner and outfielder Ray Mack were both drafted in 1945 taking two more starters out of the lineup. In , Bill Veeck formed an investment group that purchased the Cleveland Indians from Bradley's group for a reported $1.6 million. Among the investors was Bob Hope, who had grown up in Cleveland, and former Tigers slugger, Hank Greenberg. A former owner of a minor league franchise in Milwaukee, Veeck brought to Cleveland a gift for promotion. At one point, Veeck hired rubber-faced Max Patkin, the "Clown Prince of Baseball" as a coach. Patkin's appearance in the coaching box was the sort of promotional stunt that delighted fans but infuriated the American League front office. Recognizing that he had acquired a solid team, Veeck soon abandoned the aging, small and lightless League Park to take up full-time residence in massive Cleveland Municipal Stadium. The Indians had briefly moved from League Park to Municipal Stadium in mid-1932, but moved back to League Park due to complaints about the cavernous environment. From 1937 onward, however, the Indians began playing an increasing number of games at Municipal, until by 1940 they played most of their home slate there. League Park was mostly demolished in 1951, but has since been rebuilt as a recreational park. Making the most of the cavernous stadium, Veeck had a portable center field fence installed, which he could move in or out depending on how the distance favored the Indians against their opponents in a given series. The fence moved as much as between series opponents. Following the 1947 season, the American League countered with a rule change that fixed the distance of an outfield wall for the duration of a season. The massive stadium did, however, permit the Indians to set the then record for the largest crowd to see a Major League baseball game. On October 10, 1948, Game 5 of the World Series against the Boston Braves drew over 84,000. The record stood until the Los Angeles Dodgers drew a crowd in excess of 92,500 to watch Game 5 of the 1959 World Series at the Los Angeles Memorial Coliseum against the Chicago White Sox. Under Veeck's leadership, one of Cleveland's most significant achievements was breaking the color barrier in the American League by signing Larry Doby, formerly a player for the Negro League's Newark Eagles in , 11 weeks after Jackie Robinson signed with the Dodgers. Similar to Robinson, Doby battled racism on and off the field but posted a .301 batting average in 1948, his first full season. A power-hitting center fielder, Doby led the American League twice in homers. In 1948, needing pitching for the stretch run of the pennant race, Veeck turned to the Negro League again and signed pitching great Satchel Paige amid much controversy. Barred from Major League Baseball during his prime, Veeck's signing of the aging star in 1948 was viewed by many as another publicity stunt. At an official age of 42, Paige became the oldest rookie in Major League baseball history, and the first black pitcher. Paige ended the year with a 6–1 record with a 2.48 ERA, 45 strikeouts and two shutouts. In , veterans Boudreau, Keltner, and Joe Gordon had career offensive seasons, while newcomers Doby and Gene Bearden also had standout seasons. The team went down to the wire with the Boston Red Sox, winning a one-game playoff, the first in American League history, to go to the World Series. In the series, the Indians defeated the Boston Braves four games to two for their first championship in 28 years. Boudreau won the American League MVP Award. The Indians appeared in a film the following year titled "The Kid From Cleveland", in which Veeck had an interest. The film portrayed the team helping out a "troubled teenaged fan" and featured many members of the Indians organization. However, filming during the season cost the players valuable rest days leading to fatigue towards the end of the season. That season, Cleveland again contended before falling to third place. On September 23, 1949, Bill Veeck and the Indians buried their 1948 pennant in center field the day after they were mathematically eliminated from the pennant race. Later in 1949, Veeck's first wife (who had a half-stake in Veeck's share of the team) divorced him. With most of his money tied up in the Indians, Veeck was forced to sell the team to a syndicate headed by insurance magnate Ellis Ryan. In , Al Rosen was an All Star for the second year in a row, was named "The Sporting News" Major League Player of the Year, and won the American League Most Valuable Player Award in a unanimous vote playing for the Indians after leading the AL in runs, home runs, RBIs (for the second year in a row), and slugging percentage, and coming in second by one point in batting average. Ryan was forced out in 1953 in favor of Myron Wilson, who in turn gave way to William Daley in . Despite this turnover in the ownership, a powerhouse team composed of Feller, Doby, Minnie Miñoso, Luke Easter, Bobby Ávila, Al Rosen, Early Wynn, Bob Lemon, and Mike Garcia continued to contend through the early 1950s. However, Cleveland only won a single pennant in the decade, in 1954, finishing second to the New York Yankees five times. The winningest season in franchise history came in 1954, when the Indians finished the season with a record of 111–43 (.721). That mark set an American League record for wins that stood for 44 years until the Yankees won 114 games in 1998 (a 162-game regular season). The Indians 1954 winning percentage of .721 is still an American League record. The Indians returned to the World Series to face the New York Giants. The team could not bring home the title, however, ultimately being upset by the Giants in a sweep. The series was notable for Willie Mays' over-the-shoulder catch off the bat of Vic Wertz in Game 1. Cleveland remained a talented team throughout the remainder of the decade, finishing in second place in 1959, George Strickland's last full year in the majors. From 1960 to 1993, the Indians managed one third-place finish (in 1968) and six fourth-place finishes (in 1960, 1974 to 1976, 1990, and 1992) but spent the rest of the time at or near the bottom of the standings. The Indians hired general manager Frank Lane, known as "Trader" Lane, away from the St. Louis Cardinals in 1957. Lane over the years had gained a reputation as a GM who loved to make deals. With the White Sox, Lane had made over 100 trades involving over 400 players in seven years. In a short stint in St. Louis, he traded away Red Schoendienst and Harvey Haddix. Lane summed up his philosophy when he said that the only deals he regretted were the ones that he didn't make. One of Lane's early trades in Cleveland was to send Roger Maris to the Kansas City Athletics in the middle of 1958. Indians executive Hank Greenberg was not happy about the trade and neither was Maris, who said that he could not stand Lane. After Maris broke Babe Ruth's home run record, Lane defended himself by saying he still would have done the deal because Maris was unknown and he received good ballplayers in exchange. After the Maris trade, Lane acquired 25-year-old Norm Cash from the White Sox for Minnie Miñoso and then traded him to Detroit before he ever played a game for the Indians; Cash went on to hit over 350 home runs for the Tigers. The Indians received Steve Demeter in the deal, who had only five at-bats for Cleveland. In 1960, Lane made the trade that would define his tenure in Cleveland when he dealt slugging right fielder and fan favorite Rocky Colavito to the Detroit Tigers for Harvey Kuenn just before Opening Day in . It was a blockbuster trade that swapped the AL home run co-champion (Colavito) for the AL batting champion (Kuenn). After the trade, however, Colavito hit over 30 home runs four times and made three All-Star teams for Detroit and Kansas City before returning to Cleveland in . Kuenn, on the other hand, played only one season for the Indians before departing for San Francisco in a trade for an aging Johnny Antonelli and Willie Kirkland. "Akron Beacon Journal" columnist Terry Pluto documented the decades of woe that followed the trade in his book "The Curse of Rocky Colavito". Despite being attached to the curse, Colavito said that he never placed a curse on the Indians but that the trade was prompted by a salary dispute with Lane. Lane also engineered a unique trade of managers in mid-season 1960, sending Joe Gordon to the Tigers in exchange for Jimmy Dykes. Lane left the team in 1961, but ill-advised trades continued. In 1965, the Indians traded pitcher Tommy John, who would go on to win 288 games in his career, and 1966 Rookie of the Year Tommy Agee to the White Sox to get Colavito back. Indians' pitchers also set numerous strikeout records. They led the league in K's every year from 1963 to 1968, and narrowly missed in 1969. The 1964 staff was the first to amass 1,100 strikeouts, and in 1968, they were the first to collect more strikeouts than hits allowed. The 1970s were not much better, with the Indians trading away several future stars, including Graig Nettles, Dennis Eckersley, Buddy Bell and 1971 Rookie of the Year Chris Chambliss, for a number of players who made no impact. Constant ownership changes did not help the Indians. In 1963, Daley's syndicate sold the team to a group headed by general manager Gabe Paul. Three years later, Paul sold the Indians to Vernon Stouffer, of the Stouffer's frozen-food empire. Prior to Stouffer's purchase, the team was rumored to be relocated due to poor attendance. Despite the potential for a financially strong owner, Stouffer had some non-baseball related financial setbacks and, consequently, the team was cash-poor. In order to solve some financial problems, Stouffer had made an agreement to play a minimum of 30 home games in New Orleans with a view to a possible move there. After rejecting an offer from George Steinbrenner and former Indian Al Rosen, Stouffer sold the team in 1972 to a group led by Cleveland Cavaliers and Cleveland Barons owner Nick Mileti. Steinbrenner went on to buy the New York Yankees in 1973. Only five years later, Mileti's group sold the team for $11 million to a syndicate headed by trucking magnate Steve O'Neill and including former general manager and owner Gabe Paul. O'Neill's death in 1983 led to the team going on the market once more. O'Neill's nephew Patrick O'Neill did not find a buyer until real estate magnates Richard and David Jacobs purchased the team in 1986. The team was unable to move out of the cellar, with losing seasons between 1969 and 1975. One highlight was the acquisition of Gaylord Perry in . The Indians traded fireballer "Sudden Sam" McDowell for Perry, who became the first Indian pitcher to win the Cy Young Award. In , Cleveland broke another color barrier with the hiring of Frank Robinson as Major League Baseball's first African American manager. Robinson served as player-manager and provided a franchise highlight when he hit a pinch hit home run on Opening Day. But the high-profile signing of Wayne Garland, a 20-game winner in Baltimore, proved to be a disaster after Garland suffered from shoulder problems and went 28–48 over five years. The team failed to improve with Robinson as manager and he was fired in . In 1977, pitcher Dennis Eckersley threw a no-hitter against the California Angels. The next season, he was be dealt to the Boston Red Sox where he won 20 games in 1978 and another 17 in 1979. The 1970s also featured the infamous Ten Cent Beer Night at Cleveland Municipal Stadium. The ill-conceived promotion at a 1974 game against the Texas Rangers ended in a riot by fans and a forfeit by the Indians. There were more bright spots in the 1980s. In May 1981, Len Barker threw a perfect game against the Toronto Blue Jays, joining Addie Joss as the only other Indian pitcher to do so. "Super Joe" Charbonneau won the American League Rookie of the Year award. Unfortunately, Charboneau was out of baseball by 1983 after falling victim to back injuries and Barker, who was also hampered by injuries, never became a consistently dominant starting pitcher. Eventually, the Indians traded Barker to the Atlanta Braves for Brett Butler and Brook Jacoby, who became mainstays of the team for the remainder of the decade. Butler and Jacoby were joined by Joe Carter, Mel Hall, Julio Franco and Cory Snyder, bringing new hope to fans in the late 1980s. Cleveland's struggles over the 30-year span were highlighted in the 1989 film "Major League", which comically depicted a hapless Cleveland ball club going from worst to first by the end of the film. Throughout the 1980s, the Indians' owners had pushed for a new stadium. Cleveland Stadium had been a symbol of the Indians' glory years in the 1940s and 1950s. However, during the lean years even crowds of 40,000 were swallowed up by the cavernous environment. The old stadium was not aging gracefully; chunks of concrete were falling off in sections and the old wooden pilings now petrified. In 1984, a proposal for a $150 million domed stadium was defeated in a referendum 2–1. Finally, in May 1990, Cuyahoga County voters passed an excise tax on sales of alcohol and cigarettes in the county. The tax proceeds were to be used for financing the construction of the Gateway Sports and Entertainment Complex, which would include Jacobs Field for the Indians and Gund Arena for the Cleveland Cavaliers basketball team. The team's fortunes started to turn in , ironically with a very unpopular trade. The team sent power-hitting outfielder Joe Carter to the San Diego Padres for two unproven players, Sandy Alomar, Jr. and Carlos Baerga. Alomar made an immediate impact, not only being elected to the All-Star team but also winning Cleveland's fourth Rookie of the Year award and a Gold Glove. Baerga became a three-time All-Star with consistent offensive production. Indians general manager John Hart made a number of moves that finally brought success to the team. In , he hired former Indian Mike Hargrove to manage and traded catcher Eddie Taubensee to the Houston Astros who, with a surplus of outfielders, were willing to part with Kenny Lofton. Lofton finished second in AL Rookie of the Year balloting with a .285 average and 66 stolen bases. The Indians were named "Organization of the Year" by "Baseball America" in 1992, in response to the appearance of offensive bright spots and an improving farm system. The team suffered a tragedy during spring training of , when a boat carrying pitchers Steve Olin, Tim Crews, and Bob Ojeda crashed into a pier. Olin and Crews were killed, and Ojeda was seriously injured. (Ojeda missed most of the season, and retired the following year). By the end of the 1993 season, the team was in transition, leaving Cleveland Stadium and fielding a talented nucleus of young players. Many of those players came from the Indians' new AAA farm team, the Charlotte Knights, who won the International League title that year. Indians General Manager John Hart and team owner Richard Jacobs managed to turn the team's fortunes around. The Indians opened Jacobs Field in 1994 with the aim of improving on the prior season's sixth-place finish. The Indians were only one game behind the division-leading Chicago White Sox on August 12 when a players strike wiped out the rest of the season. Having contended for the division in the aborted 1994 season, Cleveland sprinted to a 100–44 record (the season was shortened by 18 games due to player/owner negotiations) in 1995, winning its first ever divisional title. Veterans Dennis Martínez, Orel Hershiser and Eddie Murray combined with a young core of players including Omar Vizquel, Albert Belle, Jim Thome, Manny Ramírez, Kenny Lofton and Charles Nagy to lead the league in team batting average as well as team ERA. After defeating the Boston Red Sox in the Division Series and the Seattle Mariners in the ALCS, Cleveland clinched the American League pennant and a World Series berth, for the first time since 1954. The World Series ended in disappointment with the Indians falling in six games to the Atlanta Braves. Tickets for every Indians home game sold out several months before opening day in 1996. The Indians repeated as AL Central champions but lost to the Baltimore Orioles in the Division Series. In 1997 Cleveland started slow but finished with an 86–75 record. Taking their third consecutive AL Central title, the Indians defeated the New York Yankees in the Division Series, 3–2. After defeating the Baltimore Orioles in the ALCS, Cleveland went on to face the Florida Marlins in the World Series that featured the coldest game in World Series history. With the series tied after Game 6, the Indians went into the ninth inning of Game Seven with a 2–1 lead, but closer José Mesa allowed the Marlins to tie the game. In the eleventh inning, Édgar Rentería drove in the winning run giving the Marlins their first championship. Cleveland became the first team to lose the World Series after carrying the lead into the ninth inning of the seventh game. In 1998, the Indians made the postseason for the fourth straight year. After defeating the wild-card Boston Red Sox 3–1 in the Division Series, Cleveland lost the 1998 ALCS in six games to the New York Yankees, who had come into the postseason with a then-AL record 114 wins in the regular season. For the 1999 season, Cleveland added relief pitcher Ricardo Rincón and second baseman Roberto Alomar, brother of catcher Sandy Alomar, Jr, and won the Central Division title for the fifth consecutive year. The team scored 1,009 runs, becoming the first (and to date only) team since the 1950 Boston Red Sox to score more than 1,000 runs in a season. This time, Cleveland did not make it past the first round, losing the Division Series to the Red Sox, despite taking a 2–0 lead in the series. In game three, Indians starter Dave Burba went down with an injury in the 4th inning. Four pitchers, including presumed game four starter Jaret Wright, surrendered nine runs in relief. Without a long reliever or emergency starter on the playoff roster, Hargrove started both Bartolo Colón and Charles Nagy in games four and five on only three days rest. The Indians lost game four 23–7 and game five 12–8. Four days later, Hargrove was dismissed as manager. In 2000, the Indians had a 44–42 start, but caught fire after the All Star break and went 46–30 the rest of the way to finish 90–72. The team had one of the league's best offenses that year and a defense that yielded three gold gloves. However, they ended up five games behind the Chicago White Sox in the Central division and missed the wild card by one game to the Seattle Mariners. Mid-season trades brought Bob Wickman and Jake Westbrook to Cleveland. After the season, free agent outfielder Manny Ramírez departed for the Boston Red Sox. In 2000, Larry Dolan bought the Indians for $320 million from Richard Jacobs, who, along with his late brother David, had paid $45 million for the club in 1986. The sale set a record at the time for the sale of a baseball franchise. 2001 saw a return to the postseason. After the departures of Ramírez and Sandy Alomar, Jr., the Indians signed Ellis Burks and former MVP Juan González, who helped the team win the Central division with a 91–71 record. One of the highlights came on August 5, when the Indians completed the biggest comeback in MLB History. Cleveland rallied to close a 14–2 deficit in the seventh inning to defeat the Seattle Mariners 15–14 in 11 innings. The Mariners, who won an MLB record-tying 116 games that season, had a strong bullpen, and Indians manager Charlie Manuel had already pulled many of his starters with the game seemingly out of reach. Seattle and Cleveland met in the first round of the postseason, however the Mariners won the series 3–2. In the 2001–02 offseason, GM John Hart resigned and his assistant, Mark Shapiro, took the reins. Shapiro moved to rebuild by dealing aging veterans for younger talent. He traded Roberto Alomar to the New York Mets for a package that included outfielder Matt Lawton and prospects Alex Escobar and Billy Traber. When the team fell out of contention in mid-, Shapiro fired manager Charlie Manuel and traded pitching ace Bartolo Colón for prospects Brandon Phillips, Cliff Lee, and Grady Sizemore; acquired Travis Hafner from the Rangers for Ryan Drese and Einar Díaz; and picked up Coco Crisp from the St. Louis Cardinals for aging starter Chuck Finley. Jim Thome left after the season, going to the Phillies for a larger contract. Young Indians teams finished far out of contention in 2002 and under new manager Eric Wedge. They posted strong offensive numbers in , but continued to struggle with a bullpen that blew more than 20 saves. A highlight of the season was a 22–0 victory over the New York Yankees on August 31, one of the worst defeats suffered by the Yankees in team history. In early , the offense got off to a poor start. After a brief July slump, the Indians caught fire in August, and cut a 15.5 game deficit in the Central Division down to 1.5 games. However, the season came to an end as the Indians went on to lose six of their last seven games, five of them by one run, missing the playoffs by only two games. Shapiro was named Executive of the Year in 2005. The next season, the club made several roster changes, while retaining its nucleus of young players. The off-season was highlighted by the acquisition of top prospect Andy Marte from the Boston Red Sox. The Indians had a solid offensive season, led by career years from Travis Hafner and Grady Sizemore. Hafner, despite missing the last month of the season, tied the single season grand slam record of six, which was set in by Don Mattingly. Despite the solid offensive performance, the bullpen struggled with 23 blown saves (a Major League worst), and the Indians finished a disappointing fourth. In , Shapiro signed veteran help for the bullpen and outfield in the offseason. Veterans Aaron Fultz and Joe Borowski joined Rafael Betancourt in the Indians bullpen. The Indians improved significantly over the prior year and went into the All-Star break in second place. The team brought back Kenny Lofton for his third stint with the team in late July. The Indians finished with a 96–66 record tied with the Red Sox for best in baseball, their seventh Central Division title in 13 years and their first postseason trip since 2001. The Indians began their playoff run by defeating the Yankees in the ALDS three games to one. This series will be most remembered for the swarm of bugs that overtook the field in the later innings of Game Two. They also jumped out to a three-games-to-one lead over the Red Sox in the ALCS. The season ended in disappointment when Boston swept the final three games to advance to the 2007 World Series. Despite the loss, Cleveland players took home a number of awards. Grady Sizemore, who had a .995 fielding percentage and only two errors in 405 chances, won the Gold Glove award, Cleveland's first since 2001. Indians Pitcher CC Sabathia won the second Cy Young Award in team history with a 19–7 record, a 3.21 ERA and an MLB-leading 241 innings pitched. Eric Wedge was awarded the first Manager of the Year Award in team history. Shapiro was named to his second Executive of the Year in 2007. The Indians struggled during the 2008 season. Injuries to sluggers Travis Hafner and Victor Martinez, as well as starting pitchers Jake Westbrook and Fausto Carmona led to a poor start. The Indians, falling to last place for a short time in June and July, traded CC Sabathia to the Milwaukee Brewers for prospects Matt LaPorta, Rob Bryson, and Michael Brantley. and traded starting third baseman Casey Blake for catching prospect Carlos Santana. Pitcher Cliff Lee went 22–3 with an ERA of 2.54 and earned the AL Cy Young Award. Grady Sizemore had a career year, winning a Gold Glove Award and a Silver Slugger Award, and the Indians finished with a record of 81–81. Prospects for the 2009 season dimmed early when the Indians ended May with a record of 22–30. Shapiro made multiple trades: Cliff Lee and Ben Francisco to the Philadelphia Phillies for prospects Jason Knapp, Carlos Carrasco, Jason Donald and Lou Marson; Victor Martinez to the Boston Red Sox for prospects Bryan Price, Nick Hagadone and Justin Masterson; Ryan Garko to the Texas Rangers for Scott Barnes; and Kelly Shoppach to the Tampa Bay Rays for Mitch Talbot. The Indians finished the season tied for fourth in their division, with a record of 65–97. The team announced on September 30, 2009, that Eric Wedge and all of the team's coaching staff were released at the end of the 2009 season. Manny Acta was hired as the team's 40th manager on October 25, 2009. On February 18, 2010, it was announced that Shapiro (following the end of the 2010 season) would be promoted to team President, with current President Paul Dolan becoming the new Chairman/CEO, and longtime Shapiro assistant Chris Antonetti filling the GM role. On January 18, 2011, longtime popular former first baseman and manager Mike Hargrove was brought in as a special adviser. The Indians started the 2011 season strong – going 30–15 in their first 45 games and seven games ahead of the Detroit Tigers for first place. Injuries led to a slump where the Indians fell out of first place. Many minor leaguers such as Jason Kipnis and Lonnie Chisenhall got opportunities to fill in for the injuries. The biggest news of the season came on July 30 when the Indians traded four prospects for Colorado Rockies star pitcher, Ubaldo Jiménez. The Indians sent their top two pitchers in the minors, Alex White and Drew Pomeranz along with Joe Gardner and Matt McBride. On August 25, the Indians signed the team leader in home runs, Jim Thome off of waivers. He made his first appearance in an Indians uniform since he left Cleveland after the 2002 season. To honor Thome, the Indians placed him at his original position, third base, for one pitch against the Minnesota Twins on September 25. It was his first appearance at third base since 1996, and his last for Cleveland. The Indians finished the season in 2nd place, 15 games behind the division champion Tigers. The Indians broke Progressive Field's Opening Day attendance record with 43,190 against the Toronto Blue Jays on April 5, 2012. The game went 16 innings, setting the MLB Opening Day record, and lasted 5 hours and 14 minutes. On September 27, 2012, with six games left in the Indians' 2012 season, Manny Acta was fired; Sandy Alomar, Jr. was named interim manager for the remainder of the season. On October 6, the Indians announced that Terry Francona, who managed the Boston Red Sox to five playoff appearances and two World Series between 2004 and 2011, would take over as manager for 2013. The Indians entered the 2013 season following an active offseason of dramatic roster turnover. Key acquisitions included free agent 1B/OF Nick Swisher and CF Michael Bourn. The team added prized right handed pitching prospect Trevor Bauer, OF Drew Stubbs, and relief pitchers Bryan Shaw and Matt Albers in a three-way trade with the Arizona Diamondbacks and Cincinnati Reds that sent RF Shin-Soo Choo to the Reds, and Tony Sipp to the Arizona Diamondbacks Other notable additions included utility man Mike Avilés, catcher Yan Gomes, designated hitter Jason Giambi, and starting pitcher Scott Kazmir. The 2013 Indians increased their win total by 24 over 2012 (from 68 to 92), finishing in second place, one game behind Detroit in the Central Division, but securing the number one seed in the American League Wild Card Standings. In their first postseason appearance since 2007, Cleveland lost the 2013 American League Wild Card Game 4–0 at home to Tampa Bay. Francona was recognized for the turnaround with the 2013 American League Manager of the Year Award. With an 85–77 record, the 2014 Indians had consecutive winning seasons for the first time since 1999–2001, but they were eliminated from playoff contention during the last week of the season and finished third in the AL Central. In 2015, after struggling through the first half of the season, the Indians finished 81–80 for their third consecutive winning season, which the team had not done since 1999–2001. For the second straight year, the Tribe finished third in the Central and was eliminated from the Wild Card race during the last week of the season. Following the departure of longtime team executive Mark Shapiro on October 6, the Indians promoted GM Chris Antonetti to President of Baseball Operations, assistant general manager Mike Chernoff to GM, and named Derek Falvey as assistant GM. Falvey was later hired by the Minnesota Twins in 2016, becoming their President of Baseball Operations. The Indians set what was then a franchise record for longest winning streak when they won their 14th consecutive game, a 2–1 win over the Toronto Blue Jays in 19 innings on July 1, 2016 at Rogers Centre. The team clinched the Central Division pennant on September 26, their eighth division title overall and first since 2007, as well as returning to the playoffs for the first time since 2013. They finished the regular season at 94–67, marking their fourth straight winning season, a feat not accomplished since the 1990s and early 2000s. The Indians began the 2016 postseason by sweeping the Boston Red Sox in the best-of-five American League Division Series, then defeated the Blue Jays in five games in the 2016 American League Championship Series to claim their sixth American League pennant and advance to the World Series against the Chicago Cubs. It marked the first appearance for the Indians in the World Series since 1997 and first for the Cubs since 1945. The Indians took a 3–1 series lead following a victory in Game 4 at Wrigley Field, but the Cubs rallied to take the final three games and won the series 4 games to 3. The Indians' 2016 success led to Francona winning his second AL Manager of the Year Award with the club. From August 24 through September 15 during the 2017 season, the Indians set a new American League record by winning 22 games in a row. On September 28, the Indians won their 100th game of the season, marking only the third time in history the team has reached that milestone. They finished the regular season with 102 wins, second most in team history (behind 1954's 111 win team). The Indians earned the AL Central Title for the second consecutive year, along with home-field advantage throughout the American League playoffs, but they lost the 2017 ALDS to the Yankees 3–2 after being up 2–0. The rivalry with fellow Ohio team the Cincinnati Reds is known as the Battle of Ohio or Buckeye Series and features the Ohio Cup trophy for the winner. Prior to 1997, the winner of the cup was determined by an annual pre-season baseball game, played each year at minor-league Cooper Stadium in the state capital of Columbus, and staged just days before the start of each new Major League Baseball season. A total of eight Ohio Cup games were played, with the Indians winning six of them. It ended with the start of interleague play in 1997. The winner of the game each year was awarded the Ohio Cup in postgame ceremonies. The Ohio Cup was a favorite among baseball fans in Columbus, with attendances regularly topping 15,000. Since 1997, the two teams have played each other as part of the regular season, with the exception of 2002. The Ohio Cup was reintroduced in 2008 and is presented to the team who wins the most games in the series that season. Initially, the teams played one three-game series per season, meeting in Cleveland in 1997 and Cincinnati the following year. Since 1999, the teams have played two series per season against each other, one at each ballpark. Since 2013, each series is two games, except in years when the AL & NL Central divisions meet interleague play, it is extended to three. Through the 2019 meetings, the Indians lead the series 63–50. As the Indians play 19 games every year with each of their AL Central competitors, several rivalries have developed. The Indians have a geographic rivalry with the Detroit Tigers, highlighted in recent years by intense battles for the AL Central title. The match up has some carryover elements from the Ohio State-Michigan rivalry, as well as the general historic rivalry between Michigan and Ohio dating back to the Toledo War. The Chicago White Sox are another rival, dating back to the 1959 season, when the Sox slipped past the Tribe to win the AL pennant. The rivalry intensified when both clubs were moved to the new AL Central in 1994. During that season, the two teams challenged for the division title, with the Indians one game back of Chicago when the strike began in August. During a game in Chicago, the White Sox confiscated Albert Belle's corked bat, followed by an attempt by Indians pitcher Jason Grimsley to crawl through the Comiskey Park clubhouse ceiling to retrieve it. Belle later signed with the White Sox in 1997, adding additional intensity to the rivalry. For the first time in over 70 years, the Indians featured uniforms without the Chief Wahoo logo in 2019, with the official team colors being red, navy blue, and white. The primary home uniform is white with navy piping around each sleeve. Across the front of the jersey in script font is the word "Indians" in red with a navy blue outline, with navy blue undershirts, belts, and socks. The alternate home jersey is red with a navy blue script "Indians" trimmed in white on the front, and navy blue piping on the sleeves, with navy blue undershirts, belts, and socks. The home cap is navy blue with a red bill and features a red block "C" on the front. The primary road uniform is gray, with "Cleveland" in navy blue block letters trimmed in red across the front of the jersey, navy blue piping around the sleeves, and navy blue undershirts, belts, and socks. The alternate road jersey is navy blue with "Cleveland" in red block letters trimmed in white on the front of the jersey, and navy blue undershirts, belts, and socks. The road cap is similar to the home cap, with the only difference being the bill is navy blue. For all games, the team uses a navy blue batting helmet with a red block "C" on the front. During the 2019 season, all jerseys featured a patch on the left sleeve commemorating the 2019 Major League Baseball All-Star Game, which was played at Progressive Field. John Adams, known by baseball fans as "The Drummer", has played a bass drum at nearly every home game since 1973. He is the only fan for whom the team has dedicated a bobble head day. Adams originally paid for his tickets (one for himself, and one for his drum), but recently the Indians have paid for his seats in honor of the contributions he has made to the ballpark atmosphere. He has been featured and interviewed on national TV shows and newspaper articles. Between June 12, 1995 and April 4, 2001, the Indians sold out 455 consecutive home games, drawing a total of 19,324,248 fans to Jacobs Field. The demand for tickets was so great that all 81 home games were sold out before Opening Day on at least three separate occasions. The sellout streak set a Major League Baseball record; this was broken by the Boston Red Sox on September 8, 2008. The Indians play "Hang On Sloopy" by The McCoys during the middle of the 8th inning, bringing with it the Ohio State Buckeyes' tradition of the fans chanting "O-H-I-O" during the chorus. They also play "Cleveland Rocks" at the end of the game for Indians wins. The club nickname and its cartoon logo have been criticized by some for perpetuating Native American stereotypes. In 1997 and 1998, protesters were arrested after effigies were burned. Charges were dismissed in the 1997 case, and were not filed in the 1998 case. Protesters arrested in the 1998 incident subsequently fought and lost a lawsuit alleging that their First Amendment rights had been violated. Bud Selig (then-Commissioner of Baseball) said in 2014 that he had never received a complaint about the logo. He has heard that there are some protesting mascots, but individual teams such as the Indians and Atlanta Braves, whose nickname was also criticized for similar reasons, should make their own decisions. An organized group consisting of Native Americans protested Chief Wahoo on Opening Day 2015, as they have for many years, making note that this is the 100th anniversary of the team becoming the Indians. Owner Paul Dolan, while stating he is respectful of critics, says he mainly hears from fans who want to keep Chief Wahoo, and has no plans to change. On January 29, 2018, Major League Baseball announced that Chief Wahoo would be removed from the Indians' uniforms as of the 2019 season, stating that the logo was no longer appropriate for on-field use. The team will consider a new logo in the future, but the block "C" will be promoted to the primary logo; additionally, there are no plans to change the team's name. Cleveland radio stations WTAM (1100 AM) and WMMS (100.7 FM) serve as flagship stations for the Cleveland Indians Radio Network. Tom Hamilton and Jim Rosenhaus serve as play-by-play announcers. The television rights are held by SportsTime Ohio (STO), Matt Underwood and former Indians Gold Glove winning CF Rick Manning form the announcing team, with veteran Cleveland sportscaster Andre Knott as field reporter, and Al Pawlowski and former Indians pitcher Jensen Lewis as pregame/postgame hosts. Select games are shown on free TV, airing on NBC affiliate WKYC channel 3 via simulcast. Notable former Indians broadcasters include Tom Manning, Jack Graney (the first ex-baseball player to become a play-by-play announcer), Ken Coleman, Joe Castiglione, Van Patrick, Nev Chandler, Bruce Drennan, Jim "Mudcat" Grant, Rocky Colavito, Dan Coughlin, and Jim Donovan. Previous broadcasters who have had lengthy tenures with the team include Joe Tait (15 seasons between TV and radio), Jack Corrigan (18 seasons on TV), Ford C. Frick Award winner Jimmy Dudley (19 seasons on radio), Mike Hegan (23 seasons between TV and radio), and Herb Score (34 seasons between TV and radio). Over the years, the Indians have been featured in several movies and television shows. Examples include: Numerous Indians players have had statues made in their honor: (*) – Inducted into the Baseball Hall of Fame as an Indian/Nap. The Cleveland Indians farm system consists of nine minor league affiliates.
https://en.wikipedia.org/wiki?curid=6652
Cape Town Cape Town (Afrikaans: Kaapstad ; Xhosa: "iKapa;") is the second most populous city in South Africa after Johannesburg and also the legislative capital of South Africa. Colloquially named the Mother City, it is the largest city of the Western Cape province and forms part of the City of Cape Town metropolitan municipality. The Parliament of South Africa sits in Cape Town. The other two capitals are located in Pretoria (the executive capital where the Presidency is based) and Bloemfontein (the judicial capital where the Supreme Court of Appeal is located). The city is known for its harbour, for its natural setting in the Cape Floristic Region, and for landmarks such as Table Mountain and Cape Point. Cape Town is home to 64% of the Western Cape's population. The city was named the World Design Capital for 2014 by the International Council of Societies of Industrial Design. In 2014, Cape Town was named the best place in the world to visit by both "The New York Times" and "The Daily Telegraph". Cape Town was one of the host cities of the tournaments of 1995 Rugby World Cup and 2010 FIFA World Cup. Located on the shore of Table Bay, Cape Town, as the oldest urban area in South Africa, was developed by the United East India Company (VOC) as a supply station for Dutch ships sailing to East Africa, India, and the Far East. Jan van Riebeeck's arrival on 6 April 1652 established the VOC Cape Colony, the first permanent European settlement in South Africa. Cape Town outgrew its original purpose as the first European outpost at the Castle of Good Hope, becoming the economic and cultural hub of the Cape Colony. Until the Witwatersrand Gold Rush and the development of Johannesburg, Cape Town was the largest city in South Africa. The earliest known remnants in the region were found at Peers Cave in Fish Hoek and date to between 15,000 and 12,000 years ago. Little is known of the history of the region's first residents, since there is no written history from the area before it was first mentioned by Portuguese explorer Bartolomeu Dias in 1488 who was the first European to reach the area and named it "Cape of Storms" (Cabo das Tormentas). It was later renamed by John II of Portugal as "Cape of Good Hope" (Cabo da Boa Esperança) because of the great optimism engendered by the opening of a sea route to India and the East. Vasco da Gama recorded a sighting of the Cape of Good Hope in 1497. In 1510, at the Battle of Salt River, Francisco de Almeida and fifty of his men were killed and his party were defeated by ox-mounted !Uriǁ’aekua ("Goringhaiqua" in Dutch approximate spelling), which was one of the so-called Khoekhoe clans of the area that also included the !Uriǁ’aeǀ’ona ("Goringhaicona" in Dutch approximate spelling, also known as "Strandlopers"), said to be the ancestors of the !Ora nation of today. In the late 16th century, Portuguese, French, Danish, Dutch and English but mainly Portuguese ships regularly continued to stop over in Table Bay en route to the Indies. They traded tobacco, copper and iron with the Khoekhoe-speaking clans of the region, in exchange for fresh meat. In 1652, Jan van Riebeeck and other employees of the United East India Company (, VOC) were sent to the Cape to establish a way-station for ships travelling to the Dutch East Indies, and the Fort de Goede Hoop (later replaced by the Castle of Good Hope). The settlement grew slowly during this period, as it was hard to find adequate labour. This labour shortage prompted the authorities to import slaves from Indonesia and Madagascar. Many of these became ancestors of the first Cape Coloured communities. Under Van Riebeeck and his successors as VOC commanders and later governors at the Cape, an impressive range of useful plants were introduced to the Cape – in the process changing the natural environment forever. Some of these, including grapes, cereals, ground nuts, potatoes, apples and citrus, had an important and lasting influence on the societies and economies of the region. The Dutch Republic being transformed in Revolutionary France's vassal Batavian Republic, Great Britain moved to take control of its colonies. Britain captured Cape Town in 1795, but the Cape was returned to the Dutch by treaty in 1803. British forces occupied the Cape again in 1806 following the Battle of Blaauwberg. In the Anglo-Dutch Treaty of 1814, Cape Town was permanently ceded to the United Kingdom. It became the capital of the newly formed Cape Colony, whose territory expanded very substantially through the 1800s. With expansion came calls for greater independence from the UK, with the Cape attaining its own parliament (1854) and a locally accountable Prime Minister (1872). Suffrage was established according to the non-racial, but sexist Cape Qualified Franchise. During the 1850s and 1860s additional plant species were introduced from Australia by the British authorities. Notably rooikrans to stabilise the sand of the Cape Flats to allow for a road connecting the peninsula with the rest of the African continent and eucalyptus to drain marshes so as to help to eliminate malaria. The discovery of diamonds in Griqualand West in 1867, and the Witwatersrand Gold Rush in 1886, prompted a flood of immigrants to South Africa. Conflicts between the Boer republics in the interior and the British colonial government resulted in the Second Boer War of 1899–1902, which Britain won. In 1910, Britain established the Union of South Africa, which unified the Cape Colony with the two defeated Boer Republics and the British colony of Natal. Cape Town became the legislative capital of the Union, and later of the Republic of South Africa. Prior to the mid-twentieth century the Cape Town was arguably the most racially integrated city in the South Africa. In the 1948 national elections, the National Party won on a platform of "apartheid" (racial segregation) under the slogan of "swart gevaar". This led to the erosion and eventual abolition of the Cape's multiracial franchise, as well as to the Group Areas Act, which classified all areas according to race. Formerly multi-racial suburbs of Cape Town were either purged of residents deemed unlawful by apartheid legislation or demolished. The most infamous example of this in Cape Town was District Six. After it was declared a whites-only region in 1965, all housing there was demolished and over 60,000 residents were forcibly removed. Many of these residents were relocated to the Cape Flats. Under apartheid, the Cape was considered a "Coloured labour preference area", to the exclusion of "Bantus", i.e. Africans. The implementation of this policy was widely opposed by trade unions, civil society and opposition parties. It is notable that this policy was not advocated for by any coloured person, and its implementation was a unilateral decision by the apartheid government. School students from Langa, Gugulethu and Nyanga in Cape Town reacted to the news of protests against Bantu Education in Soweto in June 1976 and organised gatherings and marches, which were met with resistance from the police. A number of school buildings were burnt down. Cape Town was home to many leaders of the anti-apartheid movement. On Robben Island, a former penitentiary island from the city, many famous political prisoners were held for years. In one of the most famous moments marking the end of apartheid, Nelson Mandela made his first public speech since his imprisonment, from the balcony of Cape Town City Hall hours after being released on 11 February 1990. His speech heralded the beginning of a new era for the country, and the first democratic election, was held four years later, on 27 April 1994. Nobel Square in the Victoria & Alfred Waterfront features statues of South Africa's four Nobel Peace Prize winners: Albert Luthuli, Desmond Tutu, F. W. de Klerk and Nelson Mandela. Cape Town is located at latitude 33.55° S (approx. the same as Sydney and Buenos Aires and equivalent to Casablanca and Los Angeles in the northern hemisphere) and longitude 18.25° E. Table Mountain, with its near vertical cliffs and flat-topped summit over high, and with Devil's Peak and Lion's Head on either side, together form a dramatic mountainous backdrop enclosing the central area of Cape Town, the so-called City Bowl. A thin strip of cloud, known colloquially as the "tablecloth", sometimes forms on top of the mountain. To the immediate south, the Cape Peninsula is a scenic mountainous spine jutting southwards into the Atlantic Ocean and terminating at Cape Point. There are over 70 peaks above within Cape Town's official city limits. Many of the city's suburbs lie on the large plain called the Cape Flats, which extends over to the east and joins the peninsula to the mainland. The Cape Town region is characterised by an extensive coastline, rugged mountain ranges, coastal plains, inland valleys and semi-desert fringes. UNESCO declared Robben Island in the Western Cape a World Heritage Site in 1999. Robben Island is located in Table Bay, some west of Bloubergstrand in Cape Town, and stands some 30m above sea level. Robben Island has been used as a prison where people were isolated, banished and exiled to for nearly 400 years. It was also used as a leper colony, a post office, a grazing ground, a mental hospital, and an outpost. Currently visitors can only access the island via the Robben Island Museum boat service, which runs three times daily until the beginning of the peak season (1 September). The ferries depart from the Nelson Mandela Gateway at the V&A Waterfront. Cape Town has a warm Mediterranean climate (Köppen "Csb"), with mild, moderately wet winters and dry, warm summers. Winter, which lasts from the beginning of June to the end of August, may see large cold fronts entering for limited periods from the Atlantic Ocean with significant precipitation and strong north-westerly winds. Winter months in the city average a maximum of and minimum of Total annual rainfall in the city averages although in the southern suburbs, close to the mountains, rainfall is significantly higher and averages closer to 1000 millimetres (40 in). Summer, which lasts from December to March, is warm and dry with an average maximum of and minimum of . The region can get uncomfortably hot when the Berg Wind, meaning "mountain wind", blows from the Karoo interior for a couple of weeks in February or March. Spring and summer generally feature a strong wind from the south-east, known locally as the south- or the Cape Doctor, so called because it blows air pollution away. This wind is caused by a high-pressure system which sits in the South Atlantic to the west of Cape Town, known as the South Atlantic High. Cape Town receives 3,100 hours of sunshine per year. Water temperatures range greatly, between on the Atlantic Seaboard, to over in False Bay. Average annual Ocean temperatures are between on the Atlantic Seaboard (similar to Californian waters, such as San Francisco or Big Sur), and in False Bay (similar to Northern Mediterranean temperatures, such as Nice or Monte Carlo). Located in a CI Biodiversity hotspot as well as the unique Cape Floristic Region, the city of Cape Town has one of the highest levels of biodiversity of any equivalent area in the world. These protected areas are a World Heritage Site, and an estimated 2,200 species of plants are confined to Table Mountain – more than exist in the whole of the United Kingdom which has 1200 plant species and 67 endemic plant species. Many of these species, including a great many types of proteas, are endemic to the mountain and can be found nowhere else. It is home to a total of 19 different vegetation types, of which several are endemic to the city and occur nowhere else in the world. It is also the only habitat of hundreds of endemic species, and hundreds of others which are severely restricted or threatened. This enormous species diversity is mainly because the city is uniquely located at the convergence point of several different soil types and micro-climates. Table Mountain has an unusually rich biodiversity. Its vegetation consists predominantly of several different types of the unique and rich Cape Fynbos. The main vegetation type is endangered Peninsula Sandstone Fynbos, but critically endangered Peninsula Granite Fynbos, Peninsula Shale Renosterveld and Afromontane forest occur in smaller portions on the mountain. Unfortunately, rapid population growth and urban sprawl has covered much of these ecosystems with development. Consequently, Cape Town now has over 300 threatened plant species and 13 which are now extinct. The Cape Peninsula, which lies entirely within the city of Cape Town, has the highest concentration of threatened species of any continental area of equivalent size in the world. Tiny remnant populations of critically endangered or near extinct plants sometimes survive on road sides, pavements and sports fields. The remaining ecosystems are partially protected through a system of over 30 nature reserves – including the massive Table Mountain National Park. Cape Town reached first place in the 2019 iNaturalist City Nature Challenge in two out of the three categories: Most Observations, and Most Species. This was the first entry by Capetonians in this annual competition to observe and record the local biodiversity over a four-day long weekend during what is considered the worst time of the year for local observations. However, a worldwide survey showed that the extinction rate of endemic plants from the City of Cape Town is one of the highest in the world, at roughly three per year since 1900 - partly a consequence of the very small and localised habitats and high endemicity. Cape Town's urban geography is influenced by the contours of Table Mountain, its surrounding peaks, the Durbanville Hills, and the expansive lowland region known as the Cape Flats. These geographic features in part divide the city into several commonly known groupings of suburbs (equivalent to districts outside South Africa), many of which developed historically together and share common attributes of language and culture. The City Bowl is a natural amphitheatre-shaped area bordered by Table Bay and defined by the mountains of Signal Hill, Lion's Head, Table Mountain and Devil's Peak. The area includes the central business district of Cape Town, the harbour, the Company's Garden, and the residential suburbs of De Waterkant, Devil's Peak, District Six, Zonnebloem, Gardens, Bo-Kaap, Higgovale, Oranjezicht, Schotsche Kloof, Tamboerskloof, University Estate, Vredehoek, Walmer Estate and Woodstock. The Foreshore Freeway Bridge has stood in its unfinished state since construction officially ended in 1977. It was intended to be the Eastern Boulevard Highway in the city bowl, but is unfinished due to budget constraints. The Atlantic Seaboard lies west of the City Bowl and Table Mountain, and is characterised by its beaches, cliffs, promenade and hillside communities. The area includes, from north to south, the neighbourhoods of Green Point, Mouille Point, Three Anchor Bay, Sea Point, Fresnaye, Bantry Bay, Clifton, Camps Bay, Llandudno, and Hout Bay. The Atlantic Seaboard has some of the most expensive real estate in South Africa particularly on Nettleton and Clifton Roads in Clifton, Ocean View Drive and St Leon Avenue in Bantry Bay, Theresa Avenue in Bakoven and Fishermans Bend in Llandudno. Camps Bay is home to the highest concentration of multimillionaires in Cape Town and has the highest number of high-priced mansions in South Africa with more than 155 residential units exceeding R20 million (or $US1.8 million). The West Coast suburbs lie along the beach to the north of the Cape Town city centre, and include Bloubergstrand, Milnerton, Tableview, West Beach, Big Bay, Sunset Beach, Sunningdale, Parklands and Parklands North, as well as the exurbs of Atlantis and Melkbosstrand. The Koeberg Nuclear Power Station is located within this area and maximum housing density regulations are enforced in much of the area surrounding the nuclear plant. The Northern Suburbs are predominantly Afrikaans-speaking, and include Belhar, Bellville, Blue Downs, Bothasig, Burgundy Estate, Durbanville, Edgemead, Brackenfell, Elsie's River, Eerste River, Kraaifontein, Goodwood, Kensington, Maitland, Monte Vista, Panorama, Parow, Richwood, Kraaifontein and Kuils River.. The Northern Suburbs are home to Tygerberg Hospital, the largest hospital in the Western Cape and second largest in South Africa The Southern Suburbs lie along the eastern slopes of Table Mountain, southeast of the city centre. This area is predominantly English-speaking, and includes, from north to south, Observatory, Mowbray, Pinelands, Rosebank, Rondebosch, Rondebosch East, Newlands, Claremont, Lansdowne, Kenilworth, Bishopscourt, Constantia, Wynberg, Plumstead, Ottery, Bergvliet and Diep River. West of Wynberg lies Constantia which, in addition to being a wealthy neighbourhood, is a notable wine-growing region within the City of Cape Town, and attracts tourists for its well-known wine farms and Cape Dutch architecture. The South Peninsula is generally regarded as the area South of Muizenberg on False Bay and Noordhoek on the Atlantic Ocean, all the way to Cape Point. Until recently, this predominantly English-speaking area was quite rural, however the population of the area is growing quickly as new coastal developments proliferate and larger plots are subdivided to provide more compact housing. It includes Capri Village, Clovelly, Fish Hoek, Glencairn, Kalk Bay, Kommetjie, Masiphumelele, Muizenberg, Noordhoek, Ocean View, Scarborough, Simon's Town, St James, Sunnydale and Sun Valley. South Africa's largest naval base is located at Simon's Town harbour, and close by is Boulders Beach, the site of a large colony of African penguins. The Cape Flats is an expansive, low-lying, flat area situated to the southeast of the city centre. Due to the region having a Mediterranean climate, the wettest months on the Cape Flats are from April to September, with 82% most of its rainfall occurring between these months. The rainfall patterns on the Cape Flats vary with longitude, such that the eastern parts get a minimum of 214mm per year and the central and western parts get 800mm per year. A significant portion of this water ends up in the Cape Flats Aquifer, which lie beneath the central and southern parts of the Cape Flats. Most of the land of the Cape Flats is used for residential areas, the majority of which are formal, but with several informal settlements present. Light industrial areas are also found in the area. A part of the land in the south-east is used for cultivation and contains many smallholdings. The Helderberg consists of Somerset West, Strand, Gordons Bay and a few other suburbs which were previously towns in the Helderberg district. The district takes its name from the imposing Helderberg Mountain, which reaches a height of Cape Town is governed by a 231-member city council elected in a system of mixed-member proportional representation. The city is divided into 116 wards, each of which elects a councillor by first-past-the-post voting. The remaining 115 councillors are elected from party lists so that the total number of councillors for each party is proportional to the number of votes received by that party. In the local government elections of 3 August 2016, the Democratic Alliance (DA) won an outright majority, taking 154 of the 231 council seats. The African National Congress, the national ruling party, received 57 seats. As a result of this victory, Patricia de Lille of the Democratic Alliance was re-elected to a second term as Executive Mayor. However, De Lille resigned as Mayor on 31 October 2018. The Democratic Alliance designated Dan Plato as their candidate to replace her. According to the South African National Census of 2011, the population of the City of Cape Town metropolitan municipalityan area that includes suburbs and exurbs not always considered as part of Cape Townis 3,740,026 people. This represents an annual growth rate of 2.6% compared to the results of the previous census in 2001 which found a population of 2,892,243 people. In 1944, 47% of the city's population was White, 46% was Coloured, less than 6% was Black African and 1% was Asian. Of those residents who were asked about their first language, 35.7% spoke Afrikaans, 29.8% spoke Xhosa and 28.4% spoke English. 24.8% of the population is under the age of 15, while 5.5% is 65 or older. Of those residents aged 20 or older, 1.8% have no schooling, 8.1% have some schooling but did not finish primary school, 4.6% finished primary school but have no secondary schooling, 38.9% have some secondary schooling but did not finish Grade 12, 29.9% finished Grade 12 but have no higher education, and 16.7% have higher education. Overall, 46.6% have at least a Grade 12 education. Of those aged between 5 and 25, 67.8% are attending an educational institution. Amongst those aged between 15 and 65 the unemployment rate is 23.7%. The average annual household income is R161,762. There are 1,068,573 households in the municipality, giving an average household size of 3.3 people. Of those households, 78.4% are in formal structures (houses or flats), while 20.5% are in informal structures (shacks). 94.0% of households use electricity for lighting. 87.3% of households have piped water to the dwelling, while 12.0% have piped water through a communal tap. 94.9% of households have regular refuse collection service. 91.4% of households have a flush toilet or chemical toilet, while 4.5% still use a bucket toilet. 82.1% of households have a refrigerator, 87.3% have a television and 70.1% have a radio. Only 34.0% have a landline telephone, but 91.3% have a cellphone. 37.9% have a computer, and 49.3% have access to the Internet (either through a computer or a cellphone). Cape Town is the economic hub of the Western Cape province, South Africa's second main economic centre and Africa's third main economic hub city. It serves as the regional manufacturing centre in the Western Cape. In 2011 the city's GDP was US$56.8 billion with a GDP per capita of US$15,721. In the five years preceding 2014 Cape Town GDP grew at an average of 3.7% a year. As a proportion of GDP, the agriculture and manufacturing sectors have declined whilst finance, business services, transport and logistics have grown reflecting the growth in specialised services sectors of the local economy. Fishing, clothing and textiles, wood product manufacturing, electronics, furniture, hospitality, finance and business services are industries in which Cape Town's economy has the largest comparative advantage. Between 2001 and 2010 the city's Gini coefficient, a measure of inequality, improved by dropping from 0.59 in 2007 to 0.57 in 2010 only to increase to 0.58 by 2017. The city has the lowest rate of inequality in South Africa. Cape Town has recently enjoyed a booming real estate and construction market, because of the 2010 FIFA World Cup as well as many people buying summer homes in the city or relocating there permanently. Cape Town hosted nine World Cup matches: Six first-round matches, one second-round match, one quarter final and one semifinal. The central business district is under an extensive urban renewal programme, with numerous new buildings and renovations taking place under the guidance of the Cape Town Partnership. Cape Town has four major commercial nodes, with Cape Town Central Business District containing the majority of job opportunities and office space. Century City, the Bellville/Tygervalley strip and Claremont commercial nodes are well established and contain many offices and corporate headquarters as well. Most companies headquartered in the city are insurance companies, retail groups, publishers, design houses, fashion designers, shipping companies, petrochemical companies, architects and advertising agencies. The most notable companies headquartered in the city are food and fashion retailer Woolworths, supermarket chain Pick n Pay Stores and Shoprite, New Clicks Holdings Limited, fashion retailer Foschini Group, internet service provider MWEB, Mediclinic International, eTV, multinational mass media giant Naspers, and financial services giant Sanlam. Other notable companies include Belron (vehicle glass repair and replacement group operating worldwide), CapeRay (develops, manufactures and supplies medical imaging equipment for the diagnosis of breast cancer), Ceres Fruit Juices (produces fruit juice and other fruit based products), Coronation Fund Managers (third-party fund management company), ICS (was one of the largest meat processing and distribution companies in the world), Vida e Caffè (chain of coffee retailers), Capitec Bank (commercial bank in the Republic of South Africa). The city is a manufacturing base for several multinational companies including, Johnson & Johnson, GlaxoSmithKline, Levi Strauss & Co., Adidas, Bokomo Foods, Yoco and Nampak. Most goods are handled through the Port of Cape Town or Cape Town International Airport. Most major shipbuilding companies have offices and manufacturing locations in Cape Town. The province is also a centre of energy development for the country, with the existing Koeberg nuclear power station providing energy for the Western Cape's needs. The Western Cape is an important tourist region in South Africa; the tourism industry accounts for 9.8% of the GDP of the province and employs 9.6% of the province's workforce. In 2010, over 1.5 million international tourists visited the area. With the highest number of successful Information Technology companies in Africa, Cape Town is an important centre for the industry on the continent. Growing at an annual rate of 8.5% and an estimated worth of R77 billion in 2010, nationwide the IT industry in Cape Town is becoming increasingly important to the city's economy. The city was recently named as the most entrepreneurial city in South Africa, with the percentage of Capetonians pursuing business opportunities almost three times higher than the national average. Those aged between 18 and 64 were 190% more likely to pursue new business, whilst in Johannesburg, the same demographic group was only 60% more likely than the national average to pursue a new business. Cape Town has become the Silicon Valley of South Africa, hosting innovative tech startups such as Jumo, Yoco, Aerobotics, Luno and The Sun Exchange. Cape Town is not only a popular international tourist destination in South Africa, but Africa as a whole. This is due to its mild climate, natural setting, and well-developed infrastructure. The city has several well-known natural features that attract tourists, most notably Table Mountain, which forms a large part of the Table Mountain National Park and is the back end of the City Bowl. Reaching the top of the mountain can be achieved either by hiking up, or by taking the Table Mountain Cableway. Cape Point is recognised as the dramatic headland at the end of the Cape Peninsula. Many tourists also drive along Chapman's Peak Drive, a narrow road that links Noordhoek with Hout Bay, for the views of the Atlantic Ocean and nearby mountains. It is possible to either drive or hike up Signal Hill for closer views of the City Bowl and Table Mountain. Many tourists also visit Cape Town's beaches, which are popular with local residents. Due to the city's unique geography, it is possible to visit several different beaches in the same day, each with a different setting and atmosphere. Though the Cape's water ranges from cold to mild, the difference between the two sides of the city is dramatic. While the Atlantic Seaboard averages annual water temperatures barely above that of coastal California around , the False Bay coast is much warmer, averaging between annually. This is similar to water temperatures in much of the Northern Mediterranean (for example Nice). In summer, False Bay water averages slightly over , with a common high. Beaches located on the Atlantic Coast tend to have very cold water due to the Benguela current which originates from the Southern Ocean, whilst the water at False Bay beaches may be warmer by up to at the same moment due to the influence of the warm Agulhas current. It is a common misconception that False Bay is part of the Indian Ocean, with Cape Point being both the meeting point of the Indian and Atlantic Oceans, and the southernmost tip of Africa. The oceans in fact meet at the actual southernmost tip, Cape Agulhas, which lies approximately to the south east. The misconception is fuelled by the relative warmth of the False Bay water to the Atlantic Seaboard water, and the many confusing instances of "Two Oceans" in names synonymous with Cape Town, such as the Two Oceans Marathon, the Two Oceans Aquarium, and places such as Two Oceans wine farm. Both coasts are equally popular, although the beaches in affluent Clifton and elsewhere on the Atlantic Coast are better developed with restaurants and cafés, with a strip of restaurants and bars accessible to the beach at Camps Bay. The Atlantic seaboard, known as Cape Town's Riviera, is regarded as one of the most scenic routes in South Africa, along the slopes of the Twelve Apostles to the boulders and white sand beaches of Llandudno, which the route ending in Hout Bay, a diverse bustling suburb with a harbour and a seal island. This fishing village is flanked by the Constantia valley and the picturesque Chapman's Peak drive. Boulders Beach near Simon's Town is known for its colony of African penguins. Surfing is popular and the city hosts the Red Bull Big Wave Africa surfing competition every year. The city has several notable cultural attractions. The Victoria & Alfred Waterfront, built on top of part of the docks of the Port of Cape Town, is the city's most visited tourist attraction. It is also one of the city's most popular shopping venues, with several hundred shops and the Two Oceans Aquarium. The V&A also hosts the Nelson Mandela Gateway, through which ferries depart for Robben Island. It is possible to take a ferry from the V&A to Hout Bay, Simon's Town and the Cape fur seal colonies on Seal and Duiker Islands. Several companies offer tours of the Cape Flats, a mostly Coloured township, and Khayelitsha, a mostly black township. The most popular areas for visitors to stay include Camps Bay, Sea Point, the V&A Waterfront, the City Bowl, Hout Bay, Constantia, Rondebosch, Newlands, and Somerset West. In November 2013, Cape Town was voted the best global city in "The Daily Telegraph"'s annual Travel Awards. Cape Town offers tourists a range of air, land and sea based adventure activities, including paragliding and skydiving. The City of Cape Town works closely with Cape Town Tourism to promote the city both locally and internationally. The primary focus of Cape Town Tourism is to represent Cape Town as a tourist destination. Cape Town Tourism receives a portion of its funding from the City of Cape Town while the remainder is made up of membership fees and own-generated funds. The Tristan da Cunha government owns and operates a lodging facility in Cape Town which charges discounted rates to Tristan da Cunha residents and non-resident natives. Cape Town is noted for its architectural heritage, with the highest density of Cape Dutch style buildings in the world. Cape Dutch style, which combines the architectural traditions of the Netherlands, Germany, France and Indonesia, is most visible in Constantia, the old government buildings in the Central Business District, and along Long Street. The annual Cape Town Minstrel Carnival, also known by its Afrikaans name of "Kaapse Klopse", is a large minstrel festival held annually on 2 January or ""Tweede Nuwe Jaar"" (Afrikaans: Second New Year). Competing teams of minstrels parade in brightly coloured costumes, performing Cape Jazz, either carrying colourful umbrellas or playing an array of musical instruments. The Artscape Theatre Centre is the largest performing arts venue in Cape Town. The city also encloses the 36 hectare Kirstenbosch National Botanical Garden that contains protected natural forest and fynbos along with a variety of animals and birds. There are over 7,000 species in cultivation at Kirstenbosch, including many rare and threatened species of the Cape Floristic Region. In 2004 this Region, including Kirstenbosch, was declared a UNESCO World Heritage Site. Cape Town's transport system links it to the rest of South Africa; it serves as the gateway to other destinations within the province. The Cape Winelands and in particular the towns of Stellenbosch, Paarl and Franschhoek are popular day trips from the city for sightseeing and wine tasting. Whale watching is popular amongst tourists: southern right whales and humpback whales are seen off the coast during the breeding season (August to November) and Bryde's whales and killer whale can be seen any time of the year. The nearby town of Hermanus is known for its Whale Festival, but whales can also be seen in False Bay. Heaviside's dolphins are endemic to the area and can be seen from the coast north of Cape Town; dusky dolphins live along the same coast and can occasionally be seen from the ferry to Robben Island. The only complete windmill in South Africa is Mostert's Mill, Mowbray. It was built in 1796 and restored in 1935 and again in 1995. In recent years, the city has struggled with problems such as drugs, a surge in violent drug-related crime and more recently gang violence. In the Cape Flats alone, there are approximately 100,000 people in over 130 different gangs in 2018. While there are some alliances, this multitude and division is also cause for conflict between groups. At the same time, the economy has grown due to the boom in the tourism and the real estate industries. With a Gini coefficient of 0.58, Cape Town had the lowest inequality rate in South Africa in 2012. Between 2015 and 2018 Cape Town experienced a severe water shortage that resulted in temporary but widespread water rationing. Since July 2019 widespread violent crime in poorer gang dominated areas of greater Cape Town has resulted in an ongoing military presence in these neighbourhoods. Cape Town had the highest murder rate among large South African cities at 77 murders per 100,000 people in the period April 2018 to March 2019, with 3157 murders mostly occurring in poor townships created under the apartheid regime. Most places of worship in the city are Christian churches and cathedrals: Zion Christian Church, Apostolic Faith Mission of South Africa, Assemblies of God, Baptist Union of Southern Africa (Baptist World Alliance), Methodist Church of Southern Africa (World Methodist Council), Anglican Church of Southern Africa (Anglican Communion), Presbyterian Church of Africa (World Communion of Reformed Churches), Roman Catholic Archdiocese of Cape Town (Catholic Church). Islam is the city's second largest religion with a long history in Cape Town resulting in a number of mosques and other Muslim religious sites spread across the city such as the Auwal Mosque South Africa's first mosque. Cape Town's significant Jewish population supports a number of synagogues most notably the historic Gardens Shul. The Cape Town Progressive Jewish Congregation (CTPJC) also has three temples in the city. Other religious sites in the city include Hindu, Buddhist and Baha'i temples. Several newspapers, magazines and printing facilities have their offices in the city. Independent News and Media publishes the major English language papers in the city, the "Cape Argus" and the "Cape Times". Naspers, the largest media conglomerate in South Africa, publishes "Die Burger", the major Afrikaans language paper. Cape Town has many local community newspapers. Some of the largest community newspapers in English are the "Athlone News" from Athlone, the "Atlantic Sun", the "Constantiaberg Bulletin" from Constantiaberg, the "City Vision" from Bellville, the "False Bay Echo" from False Bay, the "Helderberg Sun" from Helderberg, the "Plainsman" from Michell's Plain, the "Sentinel News" from Hout Bay, the "Southern Mail" from the Southern Peninsula, the "Southern Suburbs Tatler" from the Southern Suburbs, "Table Talk" from Table View and "Tygertalk" from Tygervalley/Durbanville. Afrikaans language community newspapers include the "Landbou-Burger" and the "Tygerburger". "Vukani", based in the Cape Flats, is published in Xhosa. Cape Town is a centre for major broadcast media with several radio stations that only broadcast within the city. 94.5 Kfm (94.5 MHz FM) and Good Hope FM (94–97 MHz FM) mostly play pop music. Heart FM (104.9 MHz FM), the former P4 Radio, plays jazz and R&B, while Fine Music Radio (101.3 FM) plays classical music and jazz. Bush Radio is a community radio station (89.5 MHz FM). The Voice of the Cape (95.8 MHz FM) and Cape Talk (567 kHz MW) are the major talk radio stations in the city. Bokradio (98.9 MHz FM) is an Afrikaans music station. The University of Cape Town also runs its own radio station, UCT Radio (104.5 MHz FM). The SABC has a small presence in the city, with satellite studios located at Sea Point. e.tv has a greater presence, with a large complex located at Longkloof Studios in Gardens. M-Net is not well represented with infrastructure within the city. Cape Town TV is a local TV station, supported by numerous organisation and focusing mostly on documentaries. Numerous productions companies and their support industries are located in the city, mostly supporting the production of overseas commercials, model shoots, TV-series and movies. The local media infrastructure remains primarily in Johannesburg. Cape Town's most popular sports by participation are cricket, association football, swimming, and rugby union. In rugby union, Cape Town is the home of the Western Province side, who play at Newlands Stadium and compete in the Currie Cup. In addition, Western Province players (along with some from Wellington's Boland Cavaliers) comprise the Stormers in the Southern Hemisphere's Super Rugby competition. Cape Town also regularly hosts the national team, the Springboks, and hosted matches during the 1995 Rugby World Cup, including the opening ceremony and game, as well as the semi-final between New Zealand and England that saw Jonah Lomu run in four tries. Association football, which is also known as "soccer" in South Africa, is also popular. Two clubs from Cape Town play in the Premier Soccer League (PSL), South Africa's premier league. These teams are Ajax Cape Town, which formed as a result of the 1999 amalgamation of the Seven Stars and the Cape Town Spurs and resurrected Cape Town City F.C. Cape Town was also the location of several of the matches of the FIFA 2010 World Cup including a semi-final, held in South Africa. The Mother City built a new 70,000 seat stadium (Cape Town Stadium) in the Green Point area. In cricket, the Cape Cobras represent Cape Town at the Newlands Cricket Ground. The team is the result of an amalgamation of the Western Province Cricket and Boland Cricket teams. They take part in the Supersport and Standard Bank Cup Series. The Newlands Cricket Ground regularly hosts international matches. Cape Town has had Olympic aspirations. For example, in 1996, Cape Town was one of the five candidate cities shortlisted by the IOC to launch official candidatures to host the 2004 Summer Olympics. Although the Games ultimately went to Athens, Cape Town came in third place. There has been some speculation that Cape Town was seeking the South African Olympic Committee's nomination to be South Africa's bid city for the 2020 Summer Olympic Games. That however was quashed when the International Olympic Committee awarded the 2020 Games to Tokyo. The city of Cape Town has vast experience in hosting major national and international sports events. The Cape Town Cycle Tour is the world's largest individually timed cycle raceand the first event outside Europe to be included in the International Cycling Union's Golden Bike Series. It sees over 35,000 cyclists tackling a route around Cape Town. The Absa Cape Epic is the largest full-service mountain bike stage race in the world. Some notable events hosted by Cape Town have included the 1995 Rugby World Cup, 2003 ICC Cricket World Cup, and World Championships in various sports such as athletics, fencing, weightlifting, hockey, cycling, canoeing, gymnastics and others. Cape Town was also a host city to the 2010 FIFA World Cup from 11 June to 11 July 2010, further enhancing its profile as a major events city. It was also one of the host cities of the 2009 Indian Premier League cricket tournament. Public primary and secondary schools in Cape Town are run by the Western Cape Education Department. This provincial department is divided into seven districts; four of these are "Metropole" districtsMetropole Central, North, South, and Eastwhich cover various areas of the city. There are also many private schools, both religious and secular, in Cape Town. Cape Town has a well-developed higher system of public universities. Cape Town is served by three public universities: the University of Cape Town (UCT), the University of the Western Cape (UWC) and the Cape Peninsula University of Technology (CPUT). Stellenbosch University, while not in the city itself, is 50 kilometres from the City Bowl and has additional campuses, such as the Tygerberg Faculty of Medicine and Health Sciences and the Bellville Business Park closer to the City. Both the University of Cape Town and Stellenbosch University are leading universities in South Africa. This is due in large part to substantial financial contributions made to these institutions by both the public and private sector. UCT is an English-speaking institution. It has over 21,000 students and has an MBA programme that was ranked 51st by the Financial Times in 2006. It is also the top-ranked university in Africa, being the only African university to make the world's Top 200 university list at number 146. Since the African National Congress has become the country's ruling party, some restructuring of Western Cape universities has taken place and as such, traditionally non-white universities have seen increased financing, which has evidently benefitted the University of the Western Cape. The Cape Peninsula University of Technology was formed on 1 January 2005, when two separate institutions – Cape Technikon and Peninsula Technikon – were merged. The new university offers education primarily in English, although one may take courses in any of South Africa's official languages. The institution generally awards the National Diploma. Cape Town has also become a popular study abroad destination for many international college students. Many study abroad providers offer semester, summer, short-term, and internship programs in partnership with Cape Town universities as a chance for international students to gain intercultural understanding. Cape Town International Airport serves both domestic and international flights. It is the second-largest airport in South Africa and serves as a major gateway for travelers to the Cape region. Cape Town has regularly scheduled services to Southern Africa, East Africa, Mauritius, Middle East, Far East, Europe and the United States as well as eleven domestic destinations. Cape Town International Airport recently opened a brand new central terminal building that was developed to handle an expected increase in air traffic as tourism numbers increased in the lead-up to the 2010 FIFA World Cup. Other renovations include several large new parking garages, a revamped domestic departure terminal, a new Bus Rapid Transit system station and a new double-decker road system. The airport's cargo facilities are also being expanded and several large empty lots are being developed into office space and hotels. The Cape Town International Airport was among the winners of the World Travel Awards for being Africa's leading airport. Cape Town International Airport is located 18 km from the Central Business District Cape Town has a long tradition as a port city. The Port of Cape Town, the city's main port, is in Table Bay directly to the north of the CBD. The port is a hub for ships in the southern Atlantic: it is located along one of the busiest shipping corridors in the world. It is also a busy container port, second in South Africa only to Durban. In 2004, it handled 3,161 ships and 9.2 million tonnes of cargo. Simon's Town Harbour on the False Bay coast of the Cape Peninsula is the main operational base of the South African Navy. Until the 1970s the city was served by the Union Castle Line with service to the United Kingdom and St Helena. The RMS "St Helena" provided passenger and cargo service between Cape Town and St Helena until the opening of St Helena Airport. Currently the cargo vessel M/V "Helena", under AW Shipping Management, takes a limited number of passengers, between Cape Town and St Helena and Ascension Island on its voyages. Multiple vessels also take passengers to and from Tristan da Cunha, inaccessible by aircraft, to and from Cape Town. In addition takes passengers on its cargo service to the Canary Islands and Hamburg, Germany. The Shosholoza Meyl is the passenger rail operations of Spoornet and operates two long-distance passenger rail services from Cape Town: a daily service to and from Johannesburg via Kimberley and a weekly service to and from Durban via Kimberley, Bloemfontein and Pietermaritzburg. These trains terminate at Cape Town railway station and make a brief stop at Bellville. Cape Town is also one terminus of the luxury tourist-oriented Blue Train as well as the five-star Rovos Rail. Metrorail operates a commuter rail service in Cape Town and the surrounding area. The Metrorail network consists of 96 stations throughout the suburbs and outskirts of Cape Town. Cape Town is the origin of three national roads. The N1 and N2 begin in the foreshore area near the City Center and the N7, which runs North toward Namibia. The N1 runs East-North-East through Edgemead, Parow, Bellville, and Brackenfell. It connects Cape Town to major cities further inland, namely Bloemfontein, Johannesburg, and Pretoria An older at-grade road, the R101, runs parallel to the N1 from Bellville. The N2 runs East-South-East through Rondebosch, Guguletu, Khayelitsha, Macassar to Somerset West. It becomes a multiple-carriageway, at-grade road from the intersection with the R44 onwards. The N2 continues east along the coast, linking Cape Town to the coastal cities of Port Elizabeth, East London and Durban. An older at-grade road, the R101, runs parallel to the N1 initially, before veering south at Bellville, to join the N2 at Somerset West via the suburbs of Kuils River and Eerste River. The N7 originates from the N1 at Wingfield Interchange near Edgemead. It begins, initially as a highway, but becoming an at-grade road from the intersection with the M5 onwards. There are also a number of regional routes linking Cape Town with surrounding areas. The R27 originates from the N1 near the Foreshore and runs north parallel to the N7, but nearer to the coast. It passes through the suburbs of Milnerton, Table View and Bloubergstrand and links the City to the West Coast, ending at the town of Velddrif. The R44 enters the east of the metro from the north, from Stellenbosch. It connects Stellenbosch to Somerset West, then crosses the N2 to Strand and Gordon's Bay. It exits the metro heading south hugging the coast, leading to the towns of Betty's Bay and Kleinmond. Of the three-digit routes, the R300, is an expressway linking the N1 at Brackenfell to the N2 near Mitchells Plain and the Cape Town International Airport. The R302 runs from the R102 in Bellville, heading north across the N1 through Durbanville leaving the metro to Malmesbury. The R304 enters the northern limits of the metro from Stellenbosch, running NNW before veering west to cross the N7 at Philadelphia to end at Atlantis at a junction with the\sR307. This R307 starts north of Koeberg from the R27 and, after meeting the R304, continues north to Darling. The R310 originates from Muizenberg and runs along the coast, to the south of Mitchell's Plain and Khayelitsha, before veering north-east, crossing the N2 west of Macassar, and exiting the metro heading to Stellenbosch. Cape Town, like most South African cities, uses Metropolitan or "M" routes for important intra-city routes, a layer below National (N) roads and Regional (R) routes. Each city's M roads are independently numbered. Most are at-grade roads. However, the M3 splits from the N2 and runs to the south along the eastern slopes of Table Mountain, connecting the City Bowl with Muizenberg. Except for a section between Rondebosch and Newlands that has at-grade intersections, this route is a highway. The M5 splits from the N1 further east than the M3, and links the Cape Flats to the CBD. It is a highway as far as the interchange with the M68 at Ottery, before continuing as an at-grade road. Cape Town suffers from the worst traffic congestion in South Africa. Golden Arrow Bus Services operates scheduled bus services in the Cape Town metropolitan area. Several companies run long-distance bus services from Cape Town to the other cities in South Africa. Cape Town has a public transport system in about 10% of the city, running north to south along the west coastline of the city, comprising Phase 1 of the IRT system. This is known as the MyCiTi service. MyCiTi Phase 1 includes services linking the Airport to the Cape Town inner city, as well as the following areas: Blouberg / Table View, Dunoon, Atlantis and Melkbosstrand, Milnerton, Paarden Eiland, Century City, Salt River and Walmer Estate, and all suburbs of the City Bowl and Atlantic Seaboard all the way to Llandudno and Hout Bay. The MyCiTi N2 Express service consists of two routes each linking the Cape Town inner city and Khayelitsha and Mitchells Plain on the Cape Flats. The service use high floor articulated and standard size buses in dedicated busways, low floor articulated and standard size buses on the N2 Express service, and smaller Optare buses in suburban and inner city areas. It offers universal access through level boarding and numerous other measures, and requires cashless fare payment using the EMV compliant smart card system, called myconnect. Headway of services (i.e. the time between buses on the same route) range from 3 mins to 20 mins in peak times to 60 minutes during quiet off-peak periods. Cape Town has two kinds of taxis: metered taxis and minibus taxis. Unlike many cities, metered taxis are not allowed to drive around the city to solicit fares and instead must be called to a specific location. Cape Town metered taxi cabs mostly operate in the city bowl, suburbs and Cape Town International Airport areas. Large companies that operate fleets of cabs can be reached by phone and are cheaper than the single operators that apply for hire from taxi ranks and Victoria and Alfred Waterfront. There are about one thousand meter taxis in Cape Town. Their rates vary from R8 per kilometre to about R15 per kilometre. The larger taxi companies in Cape Town are Excite Taxis, Cabnet and Intercab and single operators are reachable by cellular phone. The seven seated Toyota Avanza are the most popular with larger Taxi companies. Meter cabs are mostly used by tourists and are safer to use than minibus taxis. Minibus taxis are the standard form of transport for the majority of the population who cannot afford private vehicles. Although essential, these taxis are often poorly maintained and are frequently not road-worthy. These taxis make frequent unscheduled stops to pick up passengers, which can cause accidents. With the high demand for transport by the working class of South Africa, minibus taxis are often filled over their legal passenger allowance. Minibuses are generally owned and operated in fleets. Cape Town currently has seventeen active sister city agreements
https://en.wikipedia.org/wiki?curid=6653
Chicago Cubs The Chicago Cubs are an American professional baseball team based in Chicago, Illinois. The Cubs compete in Major League Baseball (MLB) as a member club of the National League (NL) Central division. The team plays its home games at Wrigley Field, located on the city's North Side. The Cubs are one of two major league teams in Chicago; the other, the Chicago White Sox, is a member of the American League (AL) Central division. The Cubs, first known as the White Stockings, were a founding member of the NL in 1876, becoming the Chicago Cubs in 1903. The Cubs have appeared in a total of eleven World Series. The 1906 Cubs won 116 games, finishing 116–36 and posting a modern-era record winning percentage of , before losing the World Series to the Chicago White Sox ("The Hitless Wonders") by four games to two. The Cubs won back-to-back World Series championships in 1907 and 1908, becoming the first major league team to play in three consecutive World Series, and the first to win it twice. Most recently, the Cubs won the 2016 National League Championship Series and 2016 World Series, which ended a 71-year National League pennant drought and a 108-year World Series championship drought, both of which are record droughts in Major League Baseball. The 108-year drought was also the longest such occurrence in all major North American sports. Since the start of divisional play in 1969, the Cubs have appeared in the postseason ten times through the 2019 season. The Cubs are known as "the North Siders", a reference to the location of Wrigley Field within the city of Chicago, and in contrast to the White Sox, whose home field (Guaranteed Rate Field) is located on the South Side. The Cubs began playing in 1870 as the Chicago White Stockings, joining the National League (NL) in 1876 as a charter member. Owner William Hulbert signed multiple star players, such as pitcher Albert Spalding and infielders Ross Barnes, Deacon White, and Adrian "Cap" Anson, to join the team prior to the N.L.'s first season. The White Stockings played their home games at West Side Grounds and quickly established themselves as one of the new league's top teams. Spalding won forty-seven games and Barnes led the league in hitting at .429 as Chicago won the first ever National League pennant, which at the time was the game's top prize. After back-to-back pennants in 1880 and 1881, Hulbert died, and Spalding, who had retired to start Spalding sporting goods, assumed ownership of the club. The White Stockings, with Anson acting as player-manager, captured their third consecutive pennant in 1882, and Anson established himself as the game's first true superstar. In 1885 and 1886, after winning N.L. pennants, the White Stockings met the champions of the short-lived American Association in that era's version of a World Series. Both seasons resulted in matchups with the St. Louis Brown Stockings, with the clubs tying in 1885 and with St. Louis winning in 1886. This was the genesis of what would eventually become one of the greatest rivalries in sports. In all, the Anson-led Chicago Base Ball Club won six National League pennants between 1876 and 1886. As a result, Chicago's club nickname transitioned, and by 1890 they had become known as the Chicago Colts, or sometimes "Anson's Colts", referring to Cap's influence within the club. Anson was the first player in history credited with collecting 3,000 career hits. After a disappointing record of 59–73 and a ninth-place finish in 1897, Anson was released by the Cubs as both a player and manager. Due to Anson's absence from the club after 22 years, local newspaper reporters started to refer to the Colts as the "Orphans". After the 1900 season, the American Base-Ball League formed as a rival professional league, and incidentally the club's old White Stockings nickname (eventually shortened to White Sox) would be adopted by a new American League neighbor to the south. In 1902, Spalding, who by this time had revamped the roster to boast what would soon be one of the best teams of the early century, sold the club to Jim Hart. The franchise was nicknamed the Cubs by the "Chicago Daily News" in 1902, although not officially becoming the Chicago Cubs until the 1907 season. During this period, which has become known as baseball's dead-ball era, Cub infielders Joe Tinker, Johnny Evers, and Frank Chance were made famous as a double-play combination by Franklin P. Adams' poem "Baseball's Sad Lexicon". The poem first appeared in the July 18, 1910 edition of the "New York Evening Mail". Mordecai "Three-Finger" Brown, Jack Taylor, Ed Reulbach, Jack Pfiester, and Orval Overall were several key pitchers for the Cubs during this time period. With Chance acting as player-manager from 1905 to 1912, the Cubs won four pennants and two World Series titles over a five-year span. Although they fell to the "Hitless Wonders" White Sox in the 1906 World Series, the Cubs recorded a record 116 victories and the best winning percentage (.763) in Major League history. With mostly the same roster, Chicago won back-to-back World Series championships in 1907 and 1908, becoming the first Major League club to play three times in the Fall Classic and the first to win it twice. However, the Cubs would not win another World Series until 2016; this remains the longest championship drought in North American professional sports. The next season, veteran catcher Johnny Kling left the team to become a professional pocket billiards player. Some historians think Kling's absence was significant enough to prevent the Cubs from also winning a third straight title in 1909, as they finished 6 games out of first place. When Kling returned the next year, the Cubs won the pennant again, but lost to the Philadelphia Athletics in the 1910 World Series. In 1914, advertising executive Albert Lasker obtained a large block of the club's shares and before the 1916 season assumed majority ownership of the franchise. Lasker brought in a wealthy partner, Charles Weeghman, the proprietor of a popular chain of lunch counters who had previously owned the Chicago Whales of the short-lived Federal League. As principal owners, the pair moved the club from the West Side Grounds to the much newer Weeghman Park, which had been constructed for the Whales only two years earlier, where they remain to this day. The Cubs responded by winning a pennant in the war-shortened season of 1918, where they played a part in another team's curse: the Boston Red Sox defeated Grover Cleveland Alexander's Cubs four games to two in the 1918 World Series, Boston's last Series championship until 2004. Beginning in 1916, Bill Wrigley of chewing-gum fame acquired an increasing quantity of stock in the Cubs. By 1921 he was the majority owner, maintaining that status into the 1930s. Meanwhile, the year 1919 saw the start of the tenure of Bill Veeck, Sr. as team president. Veeck would hold that post throughout the 1920s and into the 30s. The management team of Wrigley and Veeck came to be known as the "double-Bills." Near the end of the first decade of the double-Bills' guidance, the Cubs won the NL Pennant in 1929 and then achieved the unusual feat of winning a pennant every three years, following up the 1929 flag with league titles in 1932, 1935, and 1938. Unfortunately, their success did not extend to the Fall Classic, as they fell to their AL rivals each time. The '32 series against the Yankees featured Babe Ruth's "called shot" at Wrigley Field in game three. There were some historic moments for the Cubs as well; In 1930, Hack Wilson, one of the top home run hitters in the game, had one of the most impressive seasons in MLB history, hitting 56 home runs and establishing the current runs-batted-in record of 191. That 1930 club, which boasted six eventual hall of fame members (Wilson, Gabby Hartnett, Rogers Hornsby, George "High Pockets" Kelly, Kiki Cuyler and manager Joe McCarthy) established the current team batting average record of .309. In 1935 the Cubs claimed the pennant in thrilling fashion, winning a record 21 games in a row in September. The '38 club saw Dizzy Dean lead the team's pitching staff and provided a historic moment when they won a crucial late-season game at Wrigley Field over the Pittsburgh Pirates with a walk-off home run by Gabby Hartnett, which became known in baseball lore as "The Homer in the Gloamin'". After the "Double-Bills" (Wrigley and Veeck) died in 1932 and 1933 respectively, P.K. Wrigley, son of Bill Wrigley, took over as majority owner. He was unable to extend his father's baseball success beyond 1938, and the Cubs slipped into years of mediocrity, although the Wrigley family would retain control of the team until 1981. The Cubs enjoyed one more pennant at the close of World War II, finishing 98–56. Due to the wartime travel restrictions, the first three games of the 1945 World Series were played in Detroit, where the Cubs won two games, including a one-hitter by Claude Passeau, and the final four were played at Wrigley. The Cubs lost the series, and did not return until the 2016 World Series. After losing the 1945 World Series to the Detroit Tigers, the Cubs finished with a respectable 82–71 record in the following year, but this was only good enough for third place. In the following two decades, the Cubs played mostly forgettable baseball, finishing among the worst teams in the National League on an almost annual basis. From 1947 to 1966, they only notched one winning season. Longtime infielder-manager Phil Cavarretta, who had been a key player during the 1945 season, was fired during spring training in 1954 after admitting the team was unlikely to finish above fifth place. Although shortstop Ernie Banks would become one of the star players in the league during the next decade, finding help for him proved a difficult task, as quality players such as Hank Sauer were few and far between. This, combined with poor ownership decisions such as the College of Coaches, and the ill-fated trade of future Hall of Fame member Lou Brock to the Cardinals for pitcher Ernie Broglio (who won only seven games over the next three seasons), hampered on-field performance. The late-1960s brought hope of a renaissance, with third baseman Ron Santo, pitcher Ferguson Jenkins, and outfielder Billy Williams joining Banks. After losing a dismal 103 games in 1966, the Cubs brought home consecutive winning records in '67 and '68, marking the first time a Cub team had accomplished that feat in over two decades. In the Cubs, managed by Leo Durocher, built a substantial lead in the newly created National League Eastern Division by mid-August. Ken Holtzman pitched a no-hitter on August 19, and the division lead grew to 8 games over the St. Louis Cardinals and by 9 games over the New York Mets. After the game of September 2, the Cubs record was 84–52 with the Mets in second place at 77–55. But then a losing streak began just as a Mets winning streak was beginning. The Cubs lost the final game of a series at Cincinnati, then came home to play the resurgent Pittsburgh Pirates (who would finish in third place). After losing the first two games by scores of 9–2 and 13–4, the Cubs led going into the ninth inning. A win would be a positive springboard since the Cubs were to play a crucial series with the Mets the next day. But Willie Stargell drilled a two-out, two-strike pitch from the Cubs' ace reliever, Phil Regan, onto Sheffield Avenue to tie the score in the top of the ninth. The Cubs would lose 7–5 in extra innings.[6] Burdened by a four-game losing streak, the Cubs traveled to Shea Stadium for a short two-game set. The Mets won both games, and the Cubs left New York with a record of 84–58 just 1⁄2 game in front. More of the same followed in Philadelphia, as a 99 loss Phillies team nonetheless defeated the Cubs twice, to extend Chicago's losing streak to eight games. In a key play in the second game, on September 11, Cubs starter Dick Selma threw a surprise pickoff attempt to third baseman Ron Santo, who was nowhere near the bag or the ball. Selma's throwing error opened the gates to a Phillies rally. After that second Philly loss, the Cubs were 84–60 and the Mets had pulled ahead at 85–57. The Mets would not look back. The Cubs' eight-game losing streak finally ended the next day in St. Louis, but the Mets were in the midst of a ten-game winning streak, and the Cubs, wilting from team fatigue, generally deteriorated in all phases of the game.[1] The Mets (who had lost a record 120 games 7 years earlier), would go on to win the World Series. The Cubs, despite a respectable 92–70 record, would be remembered for having lost a remarkable 17½ games in the standings to the Mets in the last quarter of the season. Following the 1969 season, the club posted winning records for the next few seasons, but no playoff action. After the core players of those teams started to move on, the 70s got worse for the team, and they became known as "the Loveable Losers." In , the team found some life, but ultimately experienced one of its biggest collapses. The Cubs hit a high-water mark on June 28 at 47–22, boasting an game NL East lead, as they were led by Bobby Murcer (27 HR/89 RBI), and Rick Reuschel (20–10). However, the Philadelphia Phillies cut the lead to two by the All-star break, as the Cubs sat 19 games over .500, but they swooned late in the season, going 20–40 after July 31. The Cubs finished in fourth place at 81–81, while Philadelphia surged, finishing with 101 wins. The following two seasons also saw the Cubs get off to a fast start, as the team rallied to over 10 games above .500 well into both seasons, only to again wear down and play poorly later on, and ultimately settling back to mediocrity. This trait became known as the "June Swoon". Again, the Cubs' unusually high number of day games is often pointed to as one reason for the team's inconsistent late season play. Wrigley died in 1977. The Wrigley family sold the team to the Chicago Tribune in 1981, ending a 65-year family relationship with the Cubs. After over a dozen more subpar seasons, in 1981 the Cubs hired GM Dallas Green from Philadelphia to turn around the franchise. Green had managed the 1980 Phillies to the World Series title. One of his early GM moves brought in a young Phillies minor-league 3rd baseman named Ryne Sandberg, along with Larry Bowa for Iván DeJesús. The 1983 Cubs had finished 71–91 under Lee Elia, who was fired before the season ended by Green. Green continued the culture of change and overhauled the Cubs roster, front-office and coaching staff prior to 1984. Jim Frey was hired to manage the 1984 Cubs, with Don Zimmer coaching 3rd base and Billy Connors serving as pitching coach. Green shored up the 1984 roster with a series of transactions. In December, 1983 Scott Sanderson was acquired from Montreal in a three-team deal with San Diego for Carmelo Martínez. Pinch hitter Richie Hebner (.333 BA in 1984) was signed as a free-agent. In spring training, moves continued: LF Gary Matthews and CF Bobby Dernier came from Philadelphia on March 26, for Bill Campbell and a minor leaguer. Reliever Tim Stoddard (10–6 3.82, 7 saves) was acquired the same day for a minor leaguer; veteran pitcher Ferguson Jenkins was released. The team's commitment to contend was complete when Green made a midseason deal on June 15 to shore up the starting rotation due to injuries to Rick Reuschel (5–5) and Sanderson. The deal brought 1979 NL Rookie of the Year pitcher Rick Sutcliffe from the Cleveland Indians. Joe Carter (who was with the Triple-A Iowa Cubs at the time) and right fielder Mel Hall were sent to Cleveland for Sutcliffe and back-up catcher Ron Hassey (.333 with Cubs in 1984). Sutcliffe (5–5 with the Indians) immediately joined Sanderson (8–5 3.14), Eckersley (10–8 3.03), Steve Trout (13–7 3.41) and Dick Ruthven (6–10 5.04) in the starting rotation. Sutcliffe proceeded to go 16–1 for Cubs and capture the Cy Young Award. The Cubs 1984 starting lineup was very strong. It consisted of LF Matthews (.291 14–82 101 runs 17 SB), C Jody Davis (.256 19–94), RF Keith Moreland (.279 16–80), SS Larry Bowa (.223 10 SB), 1B Leon "Bull" Durham (.279 23–96 16SB), CF Dernier (.278 45 SB), 3B Ron Cey (.240 25–97), Closer Lee Smith(9–7 3.65 33 saves) and 1984 NL MVP Ryne Sandberg (.314 19–84 114 runs, 19 triples,32 SB). Reserve players Hebner, Thad Bosley, Henry Cotto, Hassey and Dave Owen produced exciting moments. The bullpen depth of Rich Bordi, George Frazier, Warren Brusstar and Dickie Noles did their job in getting the game to Smith or Stoddard. At the top of the order, Dernier and Sandberg were exciting, aptly coined "the Daily Double" by Harry Caray. With strong defense – Dernier CF and Sandberg 2B, won the NL Gold Glove- solid pitching and clutch hitting, the Cubs were a well balanced team. Following the "Daily Double", Matthews, Durham, Cey, Moreland and Davis gave the Cubs an order with no gaps to pitch around. Sutcliffe anchored a strong top-to-bottom rotation, and Smith was one of the top closers in the game. The shift in the Cubs' fortunes was characterized June 23 on the "NBC Saturday Game of the Week" contest against the St. Louis Cardinals; it has since been dubbed simply "The Sandberg Game." With the nation watching and Wrigley Field packed, Sandberg emerged as a superstar with not one, but two game-tying home runs against Cardinals closer Bruce Sutter. With his shots in the 9th and 10th innings Wrigley Field erupted and Sandberg set the stage for a comeback win that cemented the Cubs as the team to beat in the East. No one would catch them. In early August the Cubs swept the Mets in a 4-game home series that further distanced them from the pack. An infamous Keith Moreland-Ed Lynch fight erupted after Lynch hit Moreland with a pitch, perhaps forgetting Moreland was once a linebacker at the University of Texas. It was the second game of a double header and the Cubs had won the first game in part due to a three run home run by Moreland. After the bench-clearing fight the Cubs won the second game, and the sweep put the Cubs at 68–45. In 1984, each league had two divisions, East and West. The divisional winners met in a best-of-5 series to advance to the World Series, in a "2–3" format, first two games were played at the home of the team who did not have home field advantage. Then the last three games were played at the home of the team, with home field advantage. Thus the first two games were played at Wrigley Field and the next three at the home of their opponents, San Diego. A common and unfounded myth is that since Wrigley Field did not have lights at that time the National League decided to give the home field advantage to the winner of the NL West. In fact, home field advantage had rotated between the winners of the East and West since 1969 when the league expanded. In even numbered years, the NL West had home field advantage. In odd numbered years, the NL East had home field advantage. Since the NL East winners had had home field advantage in 1983, the NL West winners were entitled to it. The confusion may stem from the fact that Major League Baseball did decide that, should the Cubs make it to the World Series, the American League winner would have home field advantage unless the Cubs hosted home games at an alternate site since the Cubs home field of Wrigley Field did not yet have lights. Rumor was the Cubs could hold home games across town at Comiskey Park, home of the American League's Chicago White Sox. Rather than hold any games in the cross town rival Sox Park, the Cubs made arrangements with the August A. Busch, owner of the St. Louis Cardinals, to use Busch Stadium in St. Louis as the Cubs "home field" for the World Series. This was approved by Major League Baseball and would have enabled the Cubs to host games 1 and 2, along with games 6 and 7 if necessary. At the time home field advantage was rotated between each league. Odd numbered years the AL had home field advantage. Even numbered years the NL had home field advantage. In the 1982 World Series the St. Louis Cardinals of the NL had home field advantage. In the 1983 World Series the Baltimore Orioles of the AL had home field advantage. In the NLCS, the Cubs easily won the first two games at Wrigley Field against the San Diego Padres. The Padres were the winners of the Western Division with Steve Garvey, Tony Gwynn, Eric Show, Goose Gossage and Alan Wiggins. With wins of 13–0 and 4–2, the Cubs needed to win only one game of the next three in San Diego to make it to the World Series. After being beaten in Game 3 7–1, the Cubs lost Game 4 when Smith, with the game tied 5–5, allowed a game-winning home run to Garvey in the bottom of the ninth inning. In Game 5 the Cubs took a 3–0 lead into the 6th inning, and a 3–2 lead into the seventh with Sutcliffe (who won the Cy Young Award that year) still on the mound. Then, Leon Durham had a sharp grounder go under his glove. This critical error helped the Padres win the game 6–3, with a 4-run 7th inning and keep Chicago out of the 1984 World Series against the Detroit Tigers. The loss ended a spectacular season for the Cubs, one that brought alive a slumbering franchise and made the Cubs relevant for a whole new generation of Cubs fans. The Padres would be defeated in 5 games by Sparky Anderson's Tigers in the World Series. The 1985 season brought high hopes. The club started out well, going 35–19 through mid-June, but injuries to Sutcliffe and others in the pitching staff contributed to a 13-game losing streak that pushed the Cubs out of contention. In 1989, the first full season with night baseball at Wrigley Field, Don Zimmer's Cubs were led by a core group of veterans in Ryne Sandberg, Rick Sutcliffe and Andre Dawson, who were boosted by a crop of youngsters such as Mark Grace, Shawon Dunston, Greg Maddux, Rookie of the Year Jerome Walton, and Rookie of the Year Runner-Up Dwight Smith. The Cubs won the NL East once again that season winning 93 games. This time the Cubs met the San Francisco Giants in the NLCS. After splitting the first two games at home, the Cubs headed to the Bay Area, where despite holding a lead at some point in each of the next three games, bullpen meltdowns and managerial blunders ultimately led to three straight losses. The Cubs couldn't overcome the efforts of Will Clark, whose home run off Maddux, just after a managerial visit to the mound, led Maddux to think Clark knew what pitch was coming. Afterward, Maddux would speak into his glove during any mound conversation, beginning what is a norm today. Mark Grace was 11–17 in the series with 8 RBI. Eventually, the Giants lost to the "Bash Brothers" and the Oakland A's in the famous ""Earthquake Series"." The 1998 season would begin on a somber note with the death of legendary broadcaster Harry Caray. After the retirement of Sandberg and the trade of Dunston, the Cubs had holes to fill, and the signing of Henry Rodríguez to bat cleanup provided protection for Sammy Sosa in the lineup, as Rodriguez slugged 31 round-trippers in his first season in Chicago. Kevin Tapani led the club with a career high 19 wins while Rod Beck anchored a strong bullpen and Mark Grace turned in one of his best seasons. The Cubs were swamped by media attention in 1998, and the team's two biggest headliners were Sosa and rookie flamethrower Kerry Wood. Wood's signature performance was one-hitting the Houston Astros, a game in which he tied the major league record of 20 strikeouts in nine innings. His torrid strikeout numbers earned Wood the nickname ""Kid K,"" and ultimately earned him the 1998 NL Rookie of the Year award. Sosa caught fire in June, hitting a major league record 20 home runs in the month, and his home run race with Cardinals slugger Mark McGwire transformed the pair into international superstars in a matter of weeks. McGwire finished the season with a new major league record of 70 home runs, but Sosa's .308 average and 66 homers earned him the National League MVP Award. After a down-to-the-wire Wild Card chase with the San Francisco Giants, Chicago and San Francisco ended the regular season tied, and thus squared off in a one-game playoff at Wrigley Field. Third baseman Gary Gaetti hit the eventual game winning homer in the playoff game. The win propelled the Cubs into the postseason for the first time since 1989 with a 90–73 regular season record. Unfortunately, the bats went cold in October, as manager Jim Riggleman's club batted .183 and scored only four runs en route to being swept by Atlanta in the National League Division Series. The home run chase between Sosa, McGwire and Ken Griffey, Jr. helped professional baseball to bring in a new crop of fans as well as bringing back some fans who had been disillusioned by the 1994 strike. The Cubs retained many players who experienced career years in 1998, but, after a fast start in 1999, they collapsed again (starting with being swept at the hands of the cross-town White Sox in mid-June) and finished in the bottom of the division for the next two seasons. Despite losing fan favorite Grace to free agency and the lack of production from newcomer Todd Hundley, skipper Don Baylor's Cubs put together a good season in 2001. The season started with Mack Newton being brought in to preach "positive thinking." One of the biggest stories of the season transpired as the club made a midseason deal for Fred McGriff, which was drawn out for nearly a month as McGriff debated waiving his no-trade clause. The Cubs led the wild card race by 2.5 games in early September, but crumbled when Preston Wilson hit a three run walk off homer off of closer Tom "Flash" Gordon, which halted the team's momentum. The team was unable to make another serious charge, and finished at 88–74, five games behind both Houston and St. Louis, who tied for first. Sosa had perhaps his finest season and Jon Lieber led the staff with a 20-win season. The Cubs had high expectations in 2002, but the squad played poorly. On July 5, 2002, the Cubs promoted assistant general manager and player personnel director Jim Hendry to the General Manager position. The club responded by hiring Dusty Baker and by making some major moves in 2003. Most notably, they traded with the Pittsburgh Pirates for outfielder Kenny Lofton and third baseman Aramis Ramírez, and rode dominant pitching, led by Kerry Wood and Mark Prior, as the Cubs led the division down the stretch. Chicago halted St. Louis' run to the playoffs by taking four of five games from the Cardinals at Wrigley Field in early September, after which they won their first division title in 14 years. They then went on to defeat the Atlanta Braves in a dramatic five-game Division Series, the franchise's first postseason series win since beating the Detroit Tigers in the 1908 World Series. After losing an extra-inning game in Game 1, the Cubs rallied and took a three-games-to-one lead over the Wild Card Florida Marlins in the National League Championship Series. Florida shut the Cubs out in Game 5, but the Cubs returned home to Wrigley Field with young pitcher Mark Prior to lead the Cubs in Game 6 as they took a 3–0 lead into the 8th inning. It was at this point when a now-infamous incident took place. Several spectators attempted to catch a foul ball off the bat of Luis Castillo. A Chicago Cubs fan by the name of Steve Bartman, of Northbrook, Illinois, reached for the ball and deflected it away from the glove of Moisés Alou for the second out of the eighth inning. Alou reacted angrily toward the stands and after the game stated that he would have caught the ball. Alou at one point recanted, saying he would not have been able to make the play, but later said this was just an attempt to make Bartman feel better and believing the whole incident should be forgotten. Interference was not called on the play, as the ball was ruled to be on the spectator side of the wall. Castillo was eventually walked by Prior. Two batters later, and to the chagrin of the packed stadium, Cubs shortstop Alex Gonzalez misplayed an inning-ending double play, loading the bases. The error would lead to eight Florida runs and a Marlin victory. Despite sending Kerry Wood to the mound and holding a lead twice, the Cubs ultimately dropped Game 7, and failed to reach the World Series. The "Steve Bartman incident" was seen as the "first domino" in the turning point of the era, and the Cubs did not win a playoff game for the next eleven seasons. In 2004, the Cubs were a consensus pick by most media outlets to win the World Series. The offseason acquisition of Derek Lee (who was acquired in a trade with Florida for Hee-seop Choi) and the return of Greg Maddux only bolstered these expectations. Despite a mid-season deal for Nomar Garciaparra, misfortune struck the Cubs again. They led the Wild Card by 1.5 games over San Francisco and Houston on September 25. On that day, both teams lost, giving the Cubs a chance at increasing the lead to 2.5 games with only eight games remaining in the season, but reliever LaTroy Hawkins blew a save to the Mets, and the Cubs lost the game in extra innings. The defeat seemingly deflated the team, as they proceeded to drop six of their last eight games as the Astros won the Wild Card. Despite the fact that the Cubs had won 89 games, this fallout was decidedly unlovable, as the Cubs traded superstar Sammy Sosa after he had left the season's final game early and then lied about it publicly. Already a controversial figure in the clubhouse after his corked-bat incident, Sammy's actions alienated much of his once strong fan base as well as the few teammates still on good terms with him, (many teammates grew tired of Sosa playing loud salsa music in the locker room) and possibly tarnished his place in Cubs' lore for years to come. The disappointing season also saw fans start to become frustrated with the constant injuries to ace pitchers Mark Prior and Kerry Wood. Additionally, the 2004 season led to the departure of popular commentator Steve Stone, who had become increasingly critical of management during broadcasts and was verbally attacked by reliever Kent Mercker. Things were no better in 2005, despite a career year from first baseman Derrek Lee and the emergence of closer Ryan Dempster. The club struggled and suffered more key injuries, only managing to win 79 games after being picked by many to be a serious contender for the N.L. pennant. In 2006, bottom fell out as the Cubs finished 66–96, last in the NL Central. After finishing last in the NL Central with 66 wins in 2006, the Cubs re-tooled and went from "worst to first" in 2007. In the offseason they signed Alfonso Soriano to a contract at eight years for $136 million, and replaced manager Dusty Baker with fiery veteran manager Lou Piniella. After a rough start, which included a brawl between Michael Barrett and Carlos Zambrano, the Cubs overcame the Milwaukee Brewers, who had led the division for most of the season. The Cubs traded Barrett to the Padres, and later acquired catcher Jason Kendall from Oakland. Kendall was highly successful with his management of the pitching rotation and helped at the plate as well. By September, Geovany Soto became the full-time starter behind the plate, replacing the veteran Kendall. Winning streaks in June and July, coupled with a pair of dramatic, late-inning wins against the Reds, led to the Cubs ultimately clinching the NL Central with a record of 85–77. They met Arizona in the NLDS, but controversy followed as Piniella, in a move that has since come under scrutiny, pulled Carlos Zambrano after the sixth inning of a pitcher's duel with D-Backs ace Brandon Webb, to "...save Zambrano for (a potential) Game 4." The Cubs, however, were unable to come through, losing the first game and eventually stranding over 30 baserunners in a three-game Arizona sweep. The Tribune company, in financial distress, was acquired by real-estate mogul Sam Zell in December 2007. This acquisition included the Cubs. However, Zell did not take an active part in running the baseball franchise, instead concentrating on putting together a deal to sell it. The Cubs successfully defended their National League Central title in 2008, going to the postseason in consecutive years for the first time since 1906–08. The offseason was dominated by three months of unsuccessful trade talks with the Orioles involving 2B Brian Roberts, as well as the signing of Chunichi Dragons star Kosuke Fukudome. The team recorded their 10,000th win in April, while establishing an early division lead. Reed Johnson and Jim Edmonds were added early on and Rich Harden was acquired from the Oakland Athletics in early July. The Cubs headed into the All-Star break with the N.L.'s best record, and tied the league record with eight representatives to the All-Star game, including catcher Geovany Soto, who was named Rookie of the Year. The Cubs took control of the division by sweeping a four-game series in Milwaukee. On September 14, in a game moved to Miller Park due to Hurricane Ike, Zambrano pitched a no-hitter against the Astros, and six days later the team clinched by beating St. Louis at Wrigley. The club ended the season with a 97–64 record and met Los Angeles in the NLDS. The heavily favored Cubs took an early lead in Game 1, but James Loney's grand slam off Ryan Dempster changed the series' momentum. Chicago committed numerous critical errors and were outscored 20–6 in a Dodger sweep, which provided yet another sudden ending. The Ricketts family acquired a majority interest in the Cubs in 2009, ending the Tribune years. Apparently handcuffed by the Tribune's bankruptcy and the sale of the club to the Ricketts siblings, led by chairman Thomas S. Ricketts, the Cubs' quest for a NL Central three-peat started with notice that there would be less invested into contracts than in previous years. Chicago engaged St. Louis in a see-saw battle for first place into August 2009, but the Cardinals played to a torrid 20–6 pace that month, designating their rivals to battle in the Wild Card race, from which they were eliminated in the season's final week. The Cubs were plagued by injuries in 2009, and were only able to field their Opening Day starting lineup three times the entire season. Third baseman Aramis Ramírez injured his throwing shoulder in an early May game against the Milwaukee Brewers, sidelining him until early July and forcing journeyman players like Mike Fontenot and Aaron Miles into more prominent roles. Additionally, key players like Derrek Lee (who still managed to hit .306 with 35 HR and 111 RBI that season), Alfonso Soriano, and Geovany Soto also nursed nagging injuries. The Cubs posted a winning record (83–78) for the third consecutive season, the first time the club had done so since 1972, and a new era of ownership under the Ricketts family was approved by MLB owners in early October. Rookie Starlin Castro debuted in early May (2010) as the starting shortstop. However, the club played poorly in the early season, finding themselves 10 games under .500 at the end of June. In addition, long-time ace Carlos Zambrano was pulled from a game against the White Sox on June 25 after a tirade and shoving match with Derrek Lee, and was suspended indefinitely by Jim Hendry, who called the conduct "unacceptable." On August 22, Lou Piniella, who had already announced his retirement at the end of the season, announced that he would leave the Cubs prematurely to take care of his sick mother. Mike Quade took over as the interim manager for the final 37 games of the year. Despite being well out of playoff contention the Cubs went 24–13 under Quade, the best record in baseball during that 37 game stretch, earning Quade the manager position going forward on October 19. On December 3, 2010, Cubs broadcaster and former third baseman, Ron Santo, died due to complications from bladder cancer and diabetes. He spent 13 seasons as a player with the Cubs, and at the time of his death was regarded as one of the greatest players not in the Hall of Fame. He was posthumously elected to the Major League Baseball Hall of Fame in 2012. Despite trading for pitcher Matt Garza and signing free-agent slugger Carlos Peña, the Cubs finished the 2011 season 20 games under .500 with a record of 71–91. Weeks after the season came to an end, the club was rejuvenated in the form of a new philosophy, as new owner Tom Ricketts signed Theo Epstein away from the Boston Red Sox, naming him club President and giving him a five-year contract worth over $18 million, and subsequently discharged manager Mike Quade. Epstein, a proponent of sabremetrics and one of the architects of the 2004 and 2007 World Series championships in Boston, brought along Jed Hoyer from the Padres to fill the role of GM and hired Dale Sveum as manager. Although the team had a dismal 2012 season, losing 101 games (the worst record since 1966), it was largely expected. The youth movement ushered in by Epstein and Hoyer began as longtime fan favorite Kerry Wood retired in May, followed by Ryan Dempster and Geovany Soto being traded to Texas at the All-Star break for a group of minor league prospects headlined by Christian Villanueva, but also included little thought of Kyle Hendricks. The development of Castro, Anthony Rizzo, Darwin Barney, Brett Jackson and pitcher Jeff Samardzija, as well as the replenishing of the minor-league system with prospects such as Javier Baez, Albert Almora, and Jorge Soler became the primary focus of the season, a philosophy which the new management said would carry over at least through the 2013 season. The 2013 season resulted in much as the same the year before. Shortly before the trade deadline, the Cubs traded Matt Garza to the Texas Rangers for Mike Olt, Carl Edwards Jr, Neil Ramirez, and Justin Grimm. Three days later, the Cubs sent Alfonso Soriano to the New York Yankees for minor leaguer Corey Black. The mid season fire sale led to another last place finish in the NL Central, finishing with a record of 66–96. Although there was a five-game improvement in the record from the year before, Anthony Rizzo and Starlin Castro seemed to take steps backward in their development. On September 30, 2013, Theo Epstein made the decision to fire manager Dale Sveum after just two seasons at the helm of the Cubs. The regression of several young players was thought to be the main focus point, as the front office said Sveum would not be judged based on wins and losses. In two seasons as skipper, Sveum finished with a record of 127–197. The 2013 season was also notable as the Cubs drafted future Rookie of the Year and MVP Kris Bryant with the second overall selection. On November 7, 2013, the Cubs hired San Diego Padres bench coach Rick Renteria to be the 53rd manager in team history. The Cubs finished the 2014 season in last place with a 73–89 record in Rentería's first and only season as manager. Despite the poor record, the Cubs improved in many areas during 2014, including rebound years by Anthony Rizzo and Starlin Castro, ending the season with a winning record at home for the first time since 2009, and compiling a 33–34 record after the All-Star Break. However, following unexpected availability of Joe Maddon when he exercised a clause that triggered on October 14 with the departure of General Manager Andrew Friedman to the Los Angeles Dodgers, the Cubs relieved Rentería of his managerial duties on October 31, 2014. During the season, the Cubs drafted Kyle Schwarber with the fourth overall selection. Hall of Famer Ernie Banks died of a heart attack on January 23, 2015, shortly before his 84th birthday. The 2015 uniform carried a commemorative #14 patch on both its home and away jerseys in his honor. On November 2, 2014, the Cubs announced that Joe Maddon had signed a five-year contract to be the 54th manager in team history. On December 10, 2014, Maddon announced that the team had signed free agent Jon Lester to a six-year, $155 million contract. Many other trades and acquisitions occurred during the off season. The opening day lineup for the Cubs contained five new players including center fielder Dexter Fowler. Rookies Kris Bryant and Addison Russell were in the starting lineup by mid-April, and rookie Kyle Schwarber was added in mid-June. On August 30, Jake Arrieta threw a no hitter against the Los Angeles Dodgers. The Cubs finished the 2015 season in third place in the NL Central, with a record of 97–65, the third best record in the majors and earned a wild card berth. On October 7, in the 2015 National League Wild Card Game, Arrieta pitched a complete game shutout and the Cubs defeated the Pittsburgh Pirates 4–0. The Cubs defeated the Cardinals in the NLDS three-games-to-one, qualifying for a return to the NLCS for the first time in 12 years, where they faced the New York Mets. This was the first time in franchise history that the Cubs had clinched a playoff series at Wrigley Field. However, they were swept in four games by the Mets and were unable to make it to their first World Series since 1945. Before the season, in an effort to shore up their lineup, free agents Ben Zobrist, Jason Heyward and John Lackey were signed. To make room for the Zobrist signing, Starlin Castro was traded to the Yankees for Adam Warren and Brendan Ryan, the latter of whom was released a week later. Also during the middle of the season, the Cubs traded their top prospect Gleyber Torres for Aroldis Chapman. In a season that included a no-hitter on April 21 by Jake Arrieta, the Cubs finished with the best record in Major League Baseball and won their first National League Central title since the 2008 season, winning by 17.5 games. The team also reached the 100-win mark for the first time since 1935 and won 103 total games, the most wins for the franchise since 1910. The Cubs defeated the San Francisco Giants in the National League Division Series and returned to the National League Championship Series for the second year in a row, where they defeated the Los Angeles Dodgers in six games. This was their first NLCS win since the series was created in 1969. The win earned the Cubs their first World Series appearance since 1945 and a chance for their first World Series win since 1908. Coming back from a three-games-to-one deficit, the Cubs defeated the Cleveland Indians in seven games in the 2016 World Series, They were the first team to come back from a three-games-to-one deficit since the Kansas City Royals in 1985. On November 4, the city of Chicago held a victory parade and rally for the Cubs that began at Wrigley Field, headed down Lake Shore Drive, and ended in Grant Park. The city estimated that over five million people attended the parade and rally, which made it one of the largest recorded gatherings in history. In an attempt to be the first team to repeat as World Series champions since the Yankees in 1998, 1999, and 2000, the Cubs struggled for most of the first half of the 2017 season, never moving more than four games over .500 and finishing the first half two games under .500. On July 15, the Cubs fell to a season-high 5.5 games out of first in the NL Central. The Cubs struggled mainly due to their pitching as Jake Arrieta and Jon Lester struggled and no starting pitcher managed to win more than 14 games (four pitchers won 15 games or more for the Cubs in 2016). The Cub offense also struggled as Kyle Schwarber batted near .200 for most of the first half and was even sent to the minors. However, the Cubs recovered in the second half of the season to finish 22 games over .500 and win the NL Central by six games over the Milwaukee Brewers. The Cubs pulled out a five-game NLDS series win over the Washington Nationals to advance to the NLCS for the third consecutive year. For the second consecutive year, they faced the Dodgers. This time, however, the Dodgers defeated the Cubs in five games. In May 2017, the Cubs and the Rickets family formed Marquee Sports & Entertainment as a central sales and marketing company for the various Rickets family sports and entertainment assets: the Cubs, Wrigley Rooftops and Hickory Street Capital. Prior to the 2018 season, the Cubs made several key free agent signings to bolster their pitching staff. The team signed starting pitcher Yu Darvish to a six-year, $126 million contract and veteran closer Brandon Morrow to two-year, $21-million contract, in addition to Tyler Chatwood and Steve Cishek. However, the Cubs struggled to stay healthy throughout the season. Anthony Rizzo missed much of April due to a back injury, and Bryant missed almost a month due to shoulder injury. However, Darvish, who only started eight games in 2018, was lost for the season due to elbow and triceps injuries. Morrow also faced two injuries before the team ruled him out for the season in September. The team maintained first place in their division for much of the season. The injury-depleted team only went 16–11 during September, which allowed the Milwaukee Brewers, to finish with the same record. The Brewers defeated the Cubs in a tie-breaker game to win the Central Division and secure the top-seed in the National League. The Cubs subsequently lost to the Colorado Rockies in the 2018 National League Wild Card Game for their earliest playoff exit in three seasons. The Cubs' roster remained largely intact going into the 2019 season. The team led the Central Division by a half-game over the Brewers at the All-Star Break. However, the team’s control over the division once again dissipated going into final months of the season. The Cubs lost several key players to injuries, including Javier Báez, Anthony Rizzo, and Kris Bryant during this stretch. The team's postseason chances were compromised after suffering a nine-game losing streak in late September. The Cubs were eliminated from playoff contention on September 25, marking the first time the team had failed to qualify for the playoffs since 2014. The Cubs announced they would not renew manager Joe Maddon's contract at the end of the season. On October 24, 2019, the Cubs hired David Ross as their new manager. The Cubs have played their home games at Wrigley Field, also known as ""The Friendly Confines"" since 1916. It was built in 1914 as Weeghman Park for the Chicago Whales, a Federal League baseball team. The Cubs also shared the park with the Chicago Bears of the NFL for 50 years. The ballpark includes a manual scoreboard, ivy-covered brick walls, and relatively small dimensions. Located in Chicago's Lake View neighborhood, Wrigley Field sits on an irregular block bounded by Clark and Addison Streets and Waveland and Sheffield Avenues. The area surrounding the ballpark is typically referred to as Wrigleyville. There is a dense collection of sports bars and restaurants in the area, most with baseball inspired themes, including Sluggers, Murphy's Bleachers and The Cubby Bear. Many of the apartment buildings surrounding Wrigley Field on Waveland and Sheffield Avenues have built bleachers on their rooftops for fans to view games and other sell space for advertisement. One building on Sheffield Avenue has a sign atop its roof which says "Eamus Catuli!" which is Latin for "Let's Go Cubs!" and another chronicles the time since the last Division title, pennant, and World Series championship. The 00 denotes the 2016 NL Central title, NL pennant, and the World Series championship. On game days, many residents rent out their yards and driveways to people looking for parking spots. The uniqueness of the neighborhood itself has ingrained itself into the culture of the Chicago Cubs as well as the Wrigleyville neighborhood, and has led to being used for concerts and other sporting events, such as the 2010 NHL Winter Classic between the Chicago Blackhawks and Detroit Red Wings, as well as a 2010 NCAA men's football game between the Northwestern Wildcats and Illinois Fighting Illini. In 2013, Tom Ricketts and team president Crane Kenney unveiled plans for a five-year, $575 million privately funded renovation of Wrigley Field. Called the 1060 Project, the proposed plans included vast improvements to the stadium's facade, infrastructure, restrooms, concourses, suites, press box, bullpens, and clubhouses, as well as a jumbotron to be added in the left field bleachers, batting tunnels, a video board in right field, and, eventually, an adjacent hotel, plaza, and office-retail complex. In previous years mostly all efforts to conduct any large-scale renovations to the field had been opposed by the city, former mayor Richard M. Daley (a staunch White Sox fan), and especially the rooftop owners. Months of negotiations between the team, a group of rooftop properties investors, local Alderman Tom Tunney, and Chicago Mayor Rahm Emanuel followed with the eventual endorsements of the city's Landmarks Commission, the Plan Commission and final approval by the Chicago City Council in July 2013. The project began at the conclusion of the 2014 season. The "Bleacher Bums" is a name given to fans, many of whom spend much of the day heckling, who sit in the bleacher section at Wrigley Field. Initially, the group was called "bums" because they attended most of the games, and as Wrigley did not yet have lights, these were all day games, so it was jokingly presumed these fans were jobless. A Broadway play, starring Joe Mantegna, Dennis Farina, Dennis Franz, and James Belushi ran for years and was based on a group of Cub fans who frequented the club's games. The group was started in 1967 by dedicated fans Ron Grousl, Tom Nall and "mad bugler" Mike Murphy, who was a sports radio host during mid days on Chicago-based WSCR AM 670 "The Score". Murphy alleges that Grousl started the Wrigley tradition of throwing back opposing teams' home run balls. Beginning in the days of P.K. Wrigley and the 1937 bleacher/scoreboard reconstruction, and prior to modern media saturation, a flag with either a "W" or an "L" has flown from atop the scoreboard masthead, indicating the day's result(s) when baseball was played at Wrigley. In case of a split doubleheader, both the "W" and "L" flags are flown. Past Cubs media guides show that originally the flags were blue with a white "W" and white with a blue "L". In 1978, consistent with the dominant colors of the flags, blue and white lights were mounted atop the scoreboard, denoting "win" and "loss" respectively for the benefit of nighttime passers-by. The flags were replaced by 1990, the first year in which the Cubs media guide reports the switch to the now familiar colors of the flags: White with blue "W" and blue with white "L". In addition to needing to replace the worn-out flags, by then the retired numbers of Banks and Williams were flying on the foul poles, as white with blue numbers; so the "good" flag was switched to match that scheme. This long-established tradition has evolved to fans carrying the white-with-blue-W flags to both home and away games, and displaying them after a Cub win. The flags are known as the Cubs Win Flag. The flags have become more and more popular each season since 1998, and are now even sold as T-shirts with the same layout. In 2009, the tradition spilled over to the NHL as Chicago Blackhawks fans adopted a red and black "W" flag of their own. During the early and mid-2000s, Chip Caray usually declared that a Cubs win at home meant it was ""White flag time at Wrigley!"" More recently, the Cubs have promoted the phrase "Fly the W!" among fans and on social media. The official Cubs team mascot is a young bear cub, named Clark, described by the team's press release as a young and friendly Cub. Clark made his debut at Advocate Health Care on January 13, 2014, the same day as the press release announcing his installation as the club's first ever official physical mascot. The bear cub itself was used in the clubs since the early 1900s and was the inspiration of the Chicago Staleys changing their team's name to the Chicago Bears, because the Cubs allowed the bigger football players—like bears to cubs—to play at Wrigley Field in the 1930s. The Cubs had no official physical mascot prior to Clark, though a man in a 'polar bear' looking outfit, called "The Bear-man" (or Beeman), which was mildly popular with the fans, paraded the stands briefly in the early 1990s. There is no record of whether or not he was just a fan in a costume or employed by the club. Through the 2013 season, there were "Cubbie-bear" mascots outside of Wrigley on game day, but none were employed by the team. They pose for pictures with fans for tips. The most notable of these was "Billy Cub" who worked outside of the stadium for over six years until July 2013, when the club asked him to stop. Billy Cub, who is played by fan John Paul Weier, had unsuccessfully petitioned the team to become the official mascot. Another unofficial but much more well-known mascot is Ronnie "Woo Woo" Wickers who is a longtime fan and local celebrity in the Chicago area. He is known to Wrigley Field visitors for his idiosyncratic cheers at baseball games, generally punctuated with an exclamatory "Woo!" (e.g., "Cubs, woo! Cubs, woo! Big-Z, woo! Zambrano, woo! Cubs, woo!") Longtime Cubs announcer Harry Caray dubbed Wickers "Leather Lungs" for his ability to shout for hours at a time. He is not employed by the team, although the club has on two separate occasions allowed him into the broadcast booth and allow him some degree of freedom once he purchases or is given a ticket by fans to get into the games. He is largely allowed to roam the park and interact with fans by Wrigley Field security. During the summer of 1969, a Chicago studio group produced a single record called "Hey Hey! Holy Mackerel! (The Cubs Song)" whose title and lyrics incorporated the catch-phrases of the respective TV and radio announcers for the Cubs, Jack Brickhouse and Vince Lloyd. Several members of the Cubs recorded an album called "Cub Power" which contained a cover of the song. The song received a good deal of local airplay that summer, associating it very strongly with that bittersweet season. It was played much less frequently thereafter, although it remained an unofficial Cubs theme song for some years after. For many years, Cubs radio broadcasts started with "It's a Beautiful Day for a Ball Game" by the Harry Simeone Chorale. In 1979, Roger Bain released a 45 rpm record of his song "Thanks Mr. Banks", to honor "Mr. Cub" Ernie Banks. The song "Go, Cubs, Go!" by Steve Goodman was recorded early in the 1984 season, and was heard frequently during that season. Goodman died in September of that year, four days before the Cubs clinched the National League Eastern Division title, their first title in 39 years. Since 1984, the song started being played from time to time at Wrigley Field; since 2007, the song has been played over the loudspeakers following each Cubs home victory. The Mountain Goats recorded a song entitled "Cubs in Five" on its 1995 EP Nine Black Poppies which refers to the seeming impossibility of the Cubs winning a World Series in both its title and Chorus. In 2007, Pearl Jam frontman Eddie Vedder composed a song dedicated to the team called "All the Way". Vedder, a Chicago native, and lifelong Cubs fan, composed the song at the request of Ernie Banks. Pearl Jam has played this song live multiple times several of which occurring at Wrigley Field. Eddie Vedder has played this song live twice, at his solo shows at the Chicago Auditorium on August 21 and 22, 2008. An album entitled "Take Me Out to a Cubs Game" was released in 2008. It is a collection of 17 songs and other recordings related to the team, including Harry Caray's final performance of "Take Me Out to the Ball Game" on September 21, 1997, the Steve Goodman song mentioned above, and a newly recorded rendition of "Talkin' Baseball" (subtitled "Baseball and the Cubs") by Terry Cashman. The album was produced in celebration of the 100th anniversary of the Cubs' 1908 World Series victory and contains sounds and songs of the Cubs and Wrigley Field. The 1986 film "Ferris Bueller's Day Off" showed a game played by the Cubs when Ferris' principal goes to a bar looking for him. The 1989 film "Back to the Future Part II" depicts the Chicago Cubs defeating a baseball team from Miami in the 2015 World Series, ending the longest championship drought in all four of the major North American professional sports leagues. In 2015, the Miami Marlins failed to make the playoffs but the Cubs were able to make it to the 2015 National League Wild Card round and move on to the 2015 National League Championship Series by October 21, 2015, the date where protagonist Marty McFly traveled to the future in the film. However, it was on October 21 that the Cubs were swept by the New York Mets in the NLCS. The 1993 film "Rookie of the Year", directed by Daniel Stern, centers on the Cubs as a team going nowhere into August when the team chances upon 12-year-old Cubs fan Henry Rowengartner (Thomas Ian Nicholas), whose right (throwing) arm tendons have healed tightly after a broken arm and granted him the ability to regularly pitch at speeds in excess of . Following the Cubs' win over the Cleveland Indians in Game 7 of the 2016 World Series, Nicholas, in celebration, tweeted the final shot from the movie: Henry holding his fist up to the camera to show a Cubs World Series ring. ""Baseball's Sad Lexicon,"" also known as ""Tinker to Evers to Chance"" after its refrain, is a 1910 baseball poem by Franklin Pierce Adams. The poem is presented as a single, rueful stanza from the point of view of a New York Giants fan seeing the talented Chicago Cubs infield of shortstop Joe Tinker, second baseman Johnny Evers, and first baseman Frank Chance complete a double play. The trio began playing together with the Cubs in 1902, and formed a double play combination that lasted through April 1912. The Cubs won the pennant four times between 1906 and 1910, often defeating the Giants en route to the World Series. The poem was first published in the "New York Evening Mail" on July 12, 1912. Popular among sportswriters, numerous additional verses were written. The poem gave Tinker, Evers, and Chance increased popularity and has been credited with their elections to the National Baseball Hall of Fame in 1946. Throughout the history of the Chicago Cubs' franchise, 15 different Cubs pitchers have pitched no-hitters; however, no Cubs pitcher has thrown a perfect game. As of 2018, the Chicago Cubs are ranked as the 16th most valuable sports team in the world, 12th in the United States, third in MLB behind the New York Yankees and the Los Angeles Dodgers, and first in the city of Chicago. The Chicago Cubs retired numbers are commemorated on pinstriped flags flying from the foul poles at Wrigley Field, with the exception of Jackie Robinson, the Brooklyn Dodgers player whose number 42 was retired for all clubs. The first retired number flag, Ernie Banks' number 14, was raised on the left field pole, and they have alternated since then. 14, 10 and 31 (Jenkins) fly on the left field pole; and 26, 23 and 31 (Maddux) fly on the right field pole. * Robinson's number was retired by all MLB clubs. The Chicago Cubs farm system consists of nine minor league affiliates. Before signing a developmental agreement with the Kane County Cougars in 2012, the Cubs had a Class A minor league affiliation on two occasions with the Peoria Chiefs (1985–1995 and 2004–2012). Ryne Sandberg managed the Chiefs from 2006 to 2010. In the period between those associations with the Chiefs, the club had affiliations with the Dayton Dragons and Lansing Lugnuts. The Lugnuts were often affectionately referred to by Chip Caray as "Steve Stone's favorite team." The 2007 developmental contract with the Tennessee Smokies was preceded by Double-A affiliations with the Orlando Cubs and West Tenn Diamond Jaxx. On September 16, 2014, the Cubs announced a move of their top Class A affiliate from Daytona in the Florida State League to Myrtle Beach in the Carolina League for the 2015 season. Two days later, the Cubs signed a four-year player development contract with the South Bend Silver Hawks of the Midwest League, ending their brief relationship with the Kane County Cougars and shortly thereafter renaming the Silver Hawks the South Bend Cubs. The Chicago White Stockings, (today's Chicago Cubs), began spring training in Hot Springs, Arkansas in 1886. President Albert Spalding (founder of Spalding Sporting Goods) and player/manager Cap Anson brought their players to Hot Springs and played at the Hot Springs Baseball Grounds. The concept was for the players to have training and fitness before the start of the regular season, utilizing the bath houses of Hot Springs after practices. After the White Stockings had a successful season in 1886, winning the National League Pennant, other teams began bringing their players to Hot Springs for "spring training". The Chicago Cubs, St. Louis Browns, New York Yankees, St. Louis Cardinals, Cleveland Spiders, Detroit Tigers, Pittsburgh Pirates, Cincinnati Reds, New York Highlanders, Brooklyn Dodgers and Boston Red Sox were among the early squads to arrive. Whittington Park (1894) and later Majestic Park (1909) and Fogel Field (1912) were all built in Hot Springs specifically to host Major League teams. The Cubs' current spring training facility is located in Sloan Park in Mesa, Arizona, where they play in the Cactus League. The park seats 15,000, making it Major League baseball's largest spring training facility by capacity. The Cubs annually sell out most of their games both at home and on the road. Before Sloan Park opened in 2014, the team played games at HoHoKam Park – Dwight Patterson Field from 1979. "HoHoKam" is literally translated from Native American as "those who vanished." The North Siders have called Mesa their spring home for most seasons since 1952. In addition to Mesa, the club has held spring training in Hot Springs, Arkansas (1886, 1896–1900), (1909–1910) New Orleans (1870, 1907, 1911–1912); Champaign, Illinois (1901–02, 1906); Los Angeles (1903–04, 1948–1949), Santa Monica, California (1905); French Lick, Indiana (1908, 1943–1945); Tampa, Florida (1913–1916); Pasadena, California (1917–1921); Santa Catalina Island, California (1922–1942, 1946–1947, 1950–1951); Rendezvous Park in Mesa (1952–1965); Blair Field in Long Beach, California (1966); and Scottsdale, Arizona (1967–1978). The curious location on Catalina Island stemmed from Cubs owner William Wrigley Jr.'s then-majority interest in the island in 1919. Wrigley constructed a ballpark on the island to house the Cubs in spring training: it was built to the same dimensions as Wrigley Field. The ballpark was called Wrigley Field of Avalon. (The ballpark is long gone, but a clubhouse built by Wrigley to house the Cubs exists as the Catalina County Club.) However, by 1951 the team chose to leave Catalina Island and spring training was shifted to Mesa, Arizona. The Cubs' 30-year association with Catalina is chronicled in the book, "The Cubs on Catalina", by Jim Vitti, which was named International 'Book of the Year' by "The Sporting News". The Cubs left Catalina after some bad weather in 1951, choosing to move to Mesa, a city where the Wrigleys also had interests. Today, there is an exhibit at the Catalina Museum dedicated to the Cubs' spring training on the island. The former location in Mesa is actually the second HoHoKam Park; the first was built in 1976 as the spring-training home of the Oakland Athletics who left the park in 1979. Apart from HoHoKam Park and Sloan Park the Cubs also have another Mesa training facility called Fitch Park, this complex provides of team facilities, including major league clubhouse, four practice fields, one practice infield, enclosed batting tunnels, batting cages, a maintenance facility, and administrative offices for the Cubs. Cubs radio rights are held by Entercom; its acquisition of the radio rights effective 2015 (under CBS Radio) ended the team's 90-year association with 720 WGN. During the first season of the contract, Cubs games aired on WBBM, taking over as flagship of the Chicago Cubs Radio Network. On November 11, 2015, CBS announced that the Cubs would move to WBBM's all-sports sister station, WSCR, beginning in the 2016 season. The move was enabled by WSCR's end of their rights agreement for the White Sox, who moved to WLS. The play-by-play voice of the Cubs is Pat Hughes, who has held the position since 1996, joined by Ron Coomer. Former Cubs third baseman and fan favorite Ron Santo had been Hughes' long-time partner until his death in 2010. Keith Moreland replaced Hall of Fame inductee Santo for three seasons, followed by Coomer for the 2014 season. The club also produces its own print media; the Cubs' official magazine "Vineline", which has 12 annual issues, is in its third decade, and spotlights players and events involving the club. The club also publishes a traditional media guide. As of the 2020 season, all Cubs games not aired on broadcast television will air on Marquee Sports Network, a joint venture between the team and Sinclair Broadcast Group. The venture was officially announced in February 2019. WGN-TV had a long-term association with the team, having aired Cubs games via its WGN Sports department from its establishment in 1948, through the 2019 season. For a period, WGN's Cubs games aired nationally on WGN America (formerly Superstation WGN); however, prior to the 2015 season, the Cubs, as well as all other Chicago sports programming, was dropped from the channel as part of its re-positioning as a general entertainment cable channel. To compensate, all games carried by over-the-air channels were syndicated to a network of other television stations within the Cubs' market, which includes Illinois and parts of Indiana and Iowa. Due to limits on program pre-emptions imposed by WGN's former affiliations with The WB and its successor The CW, WGN occasionally sub-licensed some of its sports broadcasts to another station in the market, particularly independent station WCIU-TV (and later MyNetworkTV station WPWR-TV). In November 2013, the Cubs exercised an option to terminate its existing broadcast rights with WGN-TV after the 2014 season, requesting a higher-valued contract lasting through the 2019 season (which would be aligned with the end of its contract with CSN Chicago). The team would split its over-the-air package with a second partner, ABC owned-and-operated station WLS-TV, who would acquire rights to 25 games per season from 2015 through 2019. On January 7, 2015, WGN announced that it would air 45 games per-season through 2019. From 1999, regional sports network FSN Chicago served as a cable rightsholder for games not on WGN or MLB's national television outlets. In 2003, the owners of the Cubs, White Sox, Blackhawks, and Bulls all broke away from FSN Chicago, and partnered with Comcast to form Comcast SportsNet Chicago (CSN Chicago, now NBC Sports Chicago) in 2004, assuming cable rights to all four teams. Len Kasper has been the Cubs' television play-by-play announcer since 2005 and was joined by Jim Deshaies in 2013. Bob Brenly (analyst, 2005–12), Chip Caray (play-by-play, 1998–2004), Steve Stone (analyst, 1983–2000, 2003–04), Joe Carter (analyst for WGN-TV games, 2001–02) and Dave Otto (analyst for FSN Chicago games, 2001–02) also have spent time broadcasting from the Cubs booth since the death of Harry Caray in 1998.
https://en.wikipedia.org/wiki?curid=6654
Coldcut Coldcut are an English electronic music duo composed of Matt Black and Jonathan More. Credited as pioneers for pop sampling in the 1980s, Coldcut are also considered the first stars of UK electronic dance music due to their innovative style, which featured cut-up samples of hip-hop, soul, funk, spoken word and various other types of music, as well as video and multimedia. According to "Spin", "in '87 Coldcut pioneered the British fad for 'DJ records'". Coldcut's records first introduced the public to pop artists Yazz and Lisa Stansfield, through which these artists achieved pop chart success. In addition, Coldcut has remixed and created productions on tracks by the likes of Eric B & Rakim, James Brown, Queen Latifah, Eurythmics, INXS, Steve Reich, Blondie, The Fall, Pierre Henry, Nina Simone, Fog, Red Snapper, and BBC Radiophonic Workshop. Beyond their work as a production duo, Coldcut are the founders of Ninja Tune, an independent record label in London, England (with a satellite office in Montreal) with an overall emphasis on encouraging interactive technology and finding innovative uses of software. The label's first releases (the first four volumes of DJ Food - "Jazz Brakes") were produced by Coldcut in the early 90s, and composed of instrumental hip-hop cuts that led the duo to help pioneer the trip hop genre, with artists such as Funki Porcini, The Herbaliser and DJ Vadim. In 1986, computer programmer Matt Black and ex-art teacher Jonathan More were part-time DJs on the rare groove scene. More also DJed on pirate radio, hosting the "Meltdown Show" on Kiss FM and worked at the Reckless Records store on Berwick Street, London where Black visited as a customer. The first collaboration between the two artists was "Say Kids What Time Is It?" on a white label in January 1987, which mixed Jungle Book's "King of the Swingers" with the break from James Brown's "Funky Drummer". The innovation of "Say Kids..." caused More and Black to be heralded by SPIN as "the first Brit artists to really get hip-hop's class-cutup aesthetic". It is regarded as the UK's first breaks record, the first UK record to be built entirely of samples and "the final link in the chain connecting European collage-experiment with the dance-remix-scratch edit". This was later sampled in "Pump Up the Volume" by MARRS, a single that reached #1 in the UK in October 1987. Though Black had joined Kiss FM with his own mix-based show, the pair eventually joined forces on its own show later in 1987 called "Solid Steel". The eclectic show became a unifying force in underground experimental electronic music and is still running, celebrating 25 years in 2013. The duo adopted the name "Coldcut" and set up a record label called Ahead Of Our Time to release the single "Beats + Pieces" (one of the formats also included "That Greedy Beat") in 1987. All of these tracks were assembled using cassette pause button edits and later spliced tape edits that would sometimes run "all over the room". The duo used sampling from Led Zeppelin to James Brown. Electronic act The Chemical Brothers have described "Beats + Pieces" as the "first bigbeat record", a style which appeared in the mid-1990s. Coldcut's first mainstream success came when Julian Palmer from Island Records asked them to remix Eric B. & Rakim's "Paid in Full". Released in October 1987, the landmark remix is said to have "laid the groundwork for hip hop's entry into the UK mainstream", becoming a breakthrough hit for Eric B & Rakim outside the U.S., reaching No. 15 in the UK, and the top 20 in a number of European countries. It featured a prominent Ofra Haza sample and many other vocal cut ups as well as a looped rhythm which later, when speeded up, proved popular in the Breakbeat genre. Off the back of its success in clubs, the Coldcut "Seven Minutes of Madness" remix ended up being promoted as the single in the UK. In 1988, More and Black formed Hex, a self-titled "multimedia pop group", with Mile Visman and Rob Pepperell. While working on videos for artists such as Kevin Saunderson, Queen Latifah and Spiritualized, Hex's collaborative work went on to incorporate 3D modelling, punk video art, and algorithmic visuals on desktop machines. The video for Coldcut's 'Christmas Break' in 1989 is arguably one of the first pop promos produced entirely on microcomputers. In 1988, Coldcut released "Out To Lunch With Ahead Of Our Time", a double LP of Coldcut productions and re-cuts, and the various aliases under which the duo had recorded. This continued the duo's tradition of releasing limited available vinyl. The next Coldcut single, released in February 1988, moved towards a more house-influenced style. "Doctorin' the House", which debuted singer Yazz, became a top ten hit, and peaked at No. 6. In the same year, under the guise Yazz and the Plastic Population, they produced "The Only Way Is Up", a cover of a Northern soul song. The record reached No. 1 in the UK in August, and remained there for five weeks, becoming 1988's second biggest selling single. Producer Youth of Killing Joke also helped Coldcut with this record. The duo had another top hit in September with "Stop This Crazy Thing", which featured reggae vocalist Junior Reid and reached number 21 in the UK. The single "People Hold On" became another UK Top 20 hit. Released in March 1989, it helped launch the career of the then relatively unknown singer Lisa Stansfield. Coldcut and Mark Saunders produced her debut solo single "This Is the Right Time", which became another UK Top 20 hit in August as well as reaching No. 21 on the U.S. "Billboard" Hot 100 the following year. As the duo started to enjoy critical and commercial success, their debut album "What's That Noise?" was released in April 1989 on Ahead of Our Time and distributed by Big Life Records. The album gave "breaks the full length treatment", and showcased "their heady blend of hip-hop production aesthetics and proto-acid house grooves". It also rounded up a heap of unconventional guest features, quoted by SPIN as having "somehow found room at the same table for Queen Latifah and Mark E. Smith". The album's track "I'm in Deep" (featuring Smith) prefigured the indie-dance guitar-breaks crossover of such bands as the Stone Roses and Happy Mondays, utilizing Smith's freestyle raucous vocals over an acid house backing, and also including psych guitar samples from British rock band Deep Purple. "What's That Noise?" reached the Top 20 in the UK and was certified Silver. Coldcut's second album, "Some Like It Cold", released in 1990 on Ahead Of Our Time, featured a collaboration with Queen Latifah on the single "Find a Way". Though "Find a Way" was a minor hit in the UK, no more singles were released from the album. The duo was given the BPI "Producer of the Year Award" in 1990. Hex - alongside some other London visual experimenters such as iE - produced a series of videos for a longform VHS version of the album. This continued Coldcut and Hex's pioneering of the use of microcomputers to synthesize electronic music visuals. After their success with Lisa Stansfield, Coldcut signed with her label, Arista. Conflicts arose with the major label, as Coldcut's "vision extended beyond the formulae of house and techno" and mainstream pop culture (CITATION: The Virgin Encyclopedia Of Nineties Music, 2000). Eventually, the duo's album Philosophy emerged in 1993. Singles "Dreamer" and "Autumn Leaves" (1994) sung by vocalist Janis Alexander were both minor hits but the album did not chart. "Autumn Leaves" had strings recorded at Abbey Road, with a 30-piece string section and an arrangement by film composer Ed Shearmur. The leader of the string section was Simon Jeffes of Penguin Cafe Orchestra. Coldcut's insistence on their friend Mixmaster Morris to remix "Autumn Leaves" led to one of Morris' most celebrated remixes, which became a minor legend in ambient music. It has appeared on numerous compilations. In 1990, whilst on their first tour in Japan (which also featured Norman Cook, who later became Fatboy Slim), Matt and Jon formed their second record label, Ninja Tune, as a self-titled "technocoloured escape pod", and a way to escape the creative control of major labels. The label enabled them to release music under different aliases (e.g. Bogus Order, DJ Food), which also helped them to avoid pigeonholing as producers. Ninja Tune's first release was Bogus Order's "Zen Brakes". The name Coldcut stayed with Arista so there were no official Coldcut releases for the next three years. During this time, Coldcut still produced for artists on their new label, releasing a flood of material under different names and continuing to work with young groups. They additionally kept on with "Solid Steel" on Kiss FM and running the night club Stealth (Club of the Year in the NME, "The Face", and "Mixmag" in 1996). In 1991, Hex released their first video game, "Top Banana", which was included on a Hex release for the Commodore CDTV machine in 1992, arguably the first complete purpose-designed multimedia system. "Top Banana" was innovative in that it used sampled graphics, contained an ecological theme and a female lead character (dubbed "KT"), and its music changed through random processes. Coldcut and Hex presented this multimedia project as an example of the forthcoming convergence of pop music and computer-game characters. In 1992, Hex's first single - "Global Chaos" / "Digital Love Opus 1" - combined rave visuals with techno and ambient interactive visuals. In November of that year, Hex released Global Chaos CDTV, which took advantage of the possibilities of the new CD-ROM medium. The Global Chaos CDTV disk (which contained the "Top Banana" game, interactive visuals and audio), was a forerunner of the "CD+" concept, uniting music, graphics, and video games into one. This multi-dimensional entertainment product received wide coverage in the national media, including features on Dance Energy, Kaleidoscope on BBC Radio 4, "What's Up Doc?" on ITV and "Reportage" on BBC Two. "i-D Magazine" was quoted as saying, "It's like your TV tripping". Coldcut videos were made for most songs, often by Hexstatic, and used a lot of stock and sampled footage. Their "Timber" video, which created an AV collage piece using analogous techniques to audio sample collage, was put on heavy rotation on MTV. Stuart Warren Hill of Hexstatic referred to this technique as: "What you see is what you hear". "Timber" (which appears on both "Let Us Play", Coldcut's fourth album, and "Let Us Replay", their fifth) won awards for its innovative use of repetitive video clips synced to the music, including being shortlisted at the Edinburgh Television and Film Festival in their top five music videos of the year in 1998. Coldcut began integrating video sampling into their live DJ gigs at the time, and incorporated multimedia content that caused press to credit the act as segueing "into the computer age". Throughout the 90s, Hex created visuals for Coldcut's live performances, and developed the CD-ROM portion of Coldcut's "Let Us Play" and "Let Us Replay", in addition to software developed specifically for the album's world tour. Hex's inclusion of music videos and "playtools" (playful art/music software programs) on Coldcut's CD-Roms was completely ahead of the curve at that time, offering viewers/listeners a high level of interactivity. Playtools such as My Little Funkit and Playtime were the prototypes for Ninja Jamm, the app Coldcut designed and launched 16 years later. Playtime followed on from Coldcut and Hex's Synopticon installation, developing the auto-cutup algorhythm, and using other random processes to generate surprising combinations. Coldcut and Hex performed live using Playtime at the 1st Sonar Festival in 1994. Playtime was also used to generate the backing track for Coldcut's collaboration with Jello Biafra, "Every Home a Prison". In 1994 Coldcut and Hex contributed an installation to the Glasgow Gallery of Modern Art. The piece, called "Generator" was installed in the Fire Gallery. "Generator" was an interactive installation which allowed users to mix sound, video, text and graphics and make their own audio-visual mix, modelled on the techniques and technology used by Coldcut in clubs and live performance events. It consisted of two consoles: the left controlling how the sounds are played, the right controlling how the images are played. As part of the JAM exhibition of "Style, Music and Media" at the Barbican Art Gallery in 1996, Coldcut and Hex were commissioned to produce an interactive audiovisual piece called Synopticon. Conceived and designed by Robert Pepperell and Matt Black, the digital culture synthesiser allows users to "remix" sounds, images, text and music in a partially random, partially controlled way. The year 1996 also brought the Coldcut name back to More and Black, and the pair celebrated with "70 Minutes of Madness", a mix CD that became part of the Journeys by DJ series. The release was credited with "bringing to wider attention the sort of freestyle mixing the pair were always known for through their radio show on KISS FM, Solid Steel, and their steady club dates". It was voted "Best Compilation of All Time" by Jockey Slut in 1998. In February 1997, they released a double pack single "Atomic Moog 2000" / "Boot the System", the first Coldcut release on Ninja Tune. This was not eligible for the UK chart because time and format restrictions prevented the inclusion of the "Natural Rhythm" video on the CD. In August 1997, a reworking of the early track "More Beats + Pieces" gave them their first UK Top 40 hit since 1989. The album "Let Us Play!" followed in September and also made the Top 40. The fourth album by Coldcut, "Let Us Play!" paid homage to the greats that inspired them. Their first album to be released on Ninja Tune, it featured guest appearances by Grandmaster Flash, Steinski, Jello Biafra, Jimpster, The Herbaliser, Talvin Singh, Daniel Pemberton and Selena Saliva. Coldcut's cut 'n' paste method on the album was compared to that of Dadaism and William Burroughs. Hex collaborated with Coldcut to produce the multimedia CD-Rom for the album. Hex later evolved the software into the engine that was used on the Let Us Play! world tour. In 1997, Matt Black - alongside Cambridge based developers Camart - created real-time video manipulation software VJAMM. It allowed users to be a "digital video jockey", remixing and collaging sound and images and trigger audio and visual samples simultaneously, subsequently bringing futuristic technology to the audio-visual field. VJAMM rivalled some of the features of high-end and high cost tech at the time. The VJAMM technology, praised as being proof of how far computers changed the face of live music, became seminal in both Coldcut's live sets (which were called a "revelaton" by Melody Maker and DJ sets. Their CCTV live show was featured at major festivals including Glastonbury, Roskilde, Sónar, the Montreux Jazz Festival, and John Peel's Meltdown. The "beautifully simple and devastatingly effective" software was deemed revolutionary, and became recognized as a major factor in the evolution of clubs. It eventually earned a place in the American Museum of the Moving Image's permanent collection. As quoted by The Independent, Coldcut's rallying cry was "Don't hate the media, be the media'". NME was quoted as saying: "Veteran duo Coldcut are so cool they invented the remix - now they are doing the same for television." Also working with Camart, Black designed DJamm software in 1998, which Coldcut used on laptops for their live shows, providing the audio bed alongside VJAMM's audiovisual samples. Matt Black explained they designed DJamm so they "could perform electronic music in a different way – i.e., not just taking a session band out to reproduce what you put together in the studio using samples. It had a relationship to DJing, but was more interactive and more effective." Excitingly at that time, DJamm was pioneering in its ability to shuffle sliced loops into intricate sequences, enabling users to split loops into any number of parts. In 1999, "Let Us Replay!" was released, a double-disc remix album where Coldcut's classic tunes were remixed by the likes of Cornelius (which was heralded as a highlight of the album, Irresistible Force, Shut Up And Dance, Carl Craig and J Swinscoe. Let Us Replay! pieces together "short sharp shocks that put the mental in 'experimental' and still bring the breaks till the breakadawn". It also includes a few live tracks from the duo's innovative world tour. The CD-Rom of the album, which also contained a free demo disc of the VJamm software, was one of the earliest audiovisual CD- ROMs on the market, and Muzik claimed deserved to "have them canonized...it's like buying an entire mini studio for under $15". In 2000, the "Solid Steel" show moved to BBC London. Coldcut continued to forge interesting collaborations, including 2001's "Re:volution" as an EP in which Coldcut created their own political party (The Guilty Party). Featuring scratches and samples of Tony Blair and William Hague speeches, the 3-track EP included Nautilus' "Space Journey", which won an Intermusic contest in 2000. The video was widely played on MTV. With "Space Journey", Coldcut were arguably the first group to give fans access to the multitrack parts, or "stems" of their songs, building on the idea of interactivity and sharing from "Let Us Play". In 2001, Coldcut produced tracks for the Sega music video game "Rez". "Rez" replaced typical video-game sound effects with electronic music; the player created sounds and melodies, intended to simulate a form of synesthesia. The soundtrack also featured Adam Freeland and Oval. In 2002, while utilizing VJamm and Detraktor, Coldcut and Juxta remixed Herbie Hancock's classic "Rockit", creating both an audio and video remix. Working with Marcus Clements in 2002, Coldcut released the sample manipulation algorhythm from their DJamm software as a standalone VST plugin that could be used in other software, naming it the "Coldcutter". Also in 2002, Coldcut with UK VJs Headspace (now mainly performing as the VJamm Allstars developed Gridio, an interactive, immersive audio-visual installation for the Pompidou Centre as part of the ‘'Sonic Process" exhibition. The "Sonic Process" exhibition was launched at the MACBA in Barcelona in conjunction with Sónar, featuring Gridio as its centerpiece. In 2003, a commission for Graz led to a specially built version of Gridio, in a cave inside the castle mountain in Austria. Gridio was later commissioned by O2 for two simultaneous customised installations at the O2 Wireless Festivals in Leeds and London in 2007. That same year, Gridio was featured as part of Optronica at the opening week of the new BFI Southbank development in London. In 2003, Black worked with Penny Rimbaud (ex Crass) on Crass Agenda's Savage Utopia project. Black performed the piece with Rimbaud, Eve Libertine and other players at London's Vortex Jazz Club. In 2004, Coldcut collaborated with American video mashup artist TV Sheriff to produce their cut-up entitled "Revolution USA". The tactical-media project (coordinated with Canadian art duo NomIg) followed on from the UK version and extended the premise "into an open access participatory project". Through the multimedia political art project, over 12 gigabytes of footage from the last 40 years of US politics were made accessible to download, allowing participants to create a cut-up over a Coldcut beat. Coldcut also collaborated with TV Sheriff and NomIg to produce two audiovisual pieces "World of Evil" (2004) and "Revolution '08" (2008), both composed of footage from the United States presidential elections of respective years. The music used was composed by Coldcut, with "Revolution '08" featuring a remix by the Qemists. Later that year, a collaboration with the British Antarctic Survey (BAS) led to the psychedelic art documentary "Wavejammer". Coldcut was given access to the BAS archive in order to create sounds and visuals for the short film. 2004 also saw Coldcut produce a radio play in conjunction with renowned young author Hari Kunzru for BBC Radio 3 (incidentally called "Sound Mirrors"). Coldcut returned with the single "Everything Is Under Control" at the end of 2005, featuring Jon Spencer (of Jon Spencer Blues Explosion) and Mike Ladd. It was followed in 2006 by their fifth studio album "Sound Mirrors", which was quoted as being "one of the most vital and imaginative records Jon Moore and Matt Black have ever made", and saw the duo "continue, impressively, to find new ways to present political statements through a gamut of pristine electronics and breakbeats" (CITATION: Future Music, 2007). The fascinating array of guest vocalists included Soweto Kinch, Annette Peacock, Ameri Baraka, and Saul Williams. The latter followed on from Coldcut's remix of Williams' "The Pledge" for a project with DJ Spooky. A 100-date audiovisual world tour commenced for "Sound Mirrors", which was considered "no small feat in terms of technology or human effort". Coldcut was accompanied by scratch DJ Raj and AV artist Juxta, in addition to guest vocalists from the album, including UK rapper Juice Aleem, Roots Manuva, Mpho Skeef, Jon Spencer and house legend Robert Owens. Three further singles were released from the album including the Top 75 hit "True Skool" with Roots Manuva. The same track appeared on the soundtrack of the video game FIFA Street 2. Sponsored by the British Council, in 2005 Coldcut introduced AV mixing to India with the Union project, alongside collaborators Howie B and Aki Nawaz of Fun-Da-Mental. Coldcut created an A/V remix of the Bollywood hit movie "Kal Ho Naa Ho". In 2006, Coldcut performed an A/V set based on "Music for 18 Musicians" as part of Steve Reich's 70th birthday gig at the Barbican Centre in London. This was originally written for the 1999 album "Reich Remixed". Coldcut remixed another classic song in 2007: Nina Simone's "Save Me". This was part of a remix album called "Nina Simone: Remixed & Re-imagined", featuring remixes from Tony Humphries, Francois K and Chris Coco. In February 2007, Coldcut and Mixmaster Morris created a psychedelic AV obituary/tribute Coldcut, Mixmaster Morris, Ken Campbell, Bill Drummond and Alan Moore (18 March 2007). Robert Anton Wilson tribute show. Queen Elizabeth Hall, London: Mixmaster Morris. (28 August 2009) to Robert Anton Wilson, the 60s author of Illuminatus! Trilogy. The tribute featured graphic novel writer Alan Moore and artist Bill Drummond and a performance by experimental theatre legend Ken Campbell. Coldcut and Morris' hour and a half performance resembled a documentary being remixed on the fly, cutting up nearly 15 hours' worth of Wilson's lectures. In 2008, an international group of party organisers, activists and artists including Coldcut received a grant from the Intelligent Energy Department of the European Union, to create a project that promoted intelligent energy and environmental awareness to the youth of Europe. The result was Energy Union, a piece of VJ cinema, political campaign, music tour, party, art exhibition and social media hub. Energy Union toured 12 EU countries throughout 2009 and 2010, completing 24 events in total. Coldcut created the Energy Union show for the tour, a one-hour Audio/Visual montage on the theme of Intelligent Energy. In presenting new ideas for climate, environmental and energy communication strategies, the Energy Union tour was well received, and reached a widespread audience in cities across the UK, Germany, Belgium, The Netherlands, Croatia, Slovenia, Austria, Hungary, Bulgaria, Spain and the Czech Republic. Also in 2008, Coldcut was asked to remix the theme song for British cult TV show "Doctor Who" for the program's 40th anniversary. In October 2008, Coldcut celebrated the legacy of the BBC Radiophonic Workshop (the place where the Doctor Who theme was created) with a live DJ mix at London's legendary Roundhouse. The live mix incorporated classic Radiophonic Workshop compositions with extended sampling of the original gear. Additionally in 2008, Coldcut remixed "Ourselves", a Japanese No. 1 hit from the single "&" by Ayumi Hamasaki. This mix was included on the album "". Starting in 2009, Matt Black, with musician/artist/coder Paul Miller (creator of the TX Modular Open Source synth), developed Granul8, a new type of visual fx/source Black termed a "granular video synthesiser". Granul8 allows the use of realtime VJ techniques including video feedback combined with VDMX VJ software. From 2009 onwards, Black has been collaborating with coder and psychedelic mathematician William Rood to create a forthcoming project called Liveloom, a social media AV mixer. In 2010, Coldcut celebrated 20 years of releasing music with its label, Ninja Tune. A book entitled "Ninja Tune: 20 Years of Beats and Pieces" was released on 12 August 2010, and an exhibition was held at Black Dog Publishing's Black Dog Space in London, showcasing artwork, design and photography from the label's 20-year history. A compilation album was released on 20 September in two formats: a regular version consisting of two 2-disc volumes, and a limited edition which contained six CDs, six 7" vinyl singles, a hardback copy of the book, a poster and additional items. Ninja Tune also incorporated a series of international parties. This repositioned Ninja as a continually compelling and influential label, being one of the "longest-running (and successful) UK indie labels to come out of the late-1980s/early-90s explosion in dance music and hip-hop" (Pitchfork, 28 September 2010). Pitchfork claimed it had a "right to show off a little". In July 2013, Coldcut produced a piece entitled "D'autre" based on the writings of French poet Arthur Rimbaud, for Forum Des Images in Paris. The following month, in August, Coldcut produced a new soundtrack for a section of André Sauvage's classic film "Études sur Paris", which was shown as part of Noise of Art at the BFI in London, which celebrated 100 years of electronic music and silent cinema. Coldcut put new music to films from the Russolo era, incorporating original recordings of Russolo's proto-synths. In 2014 Coldcut did 3 soundtracks as part of the project New City, a series of animated skylines of the near future developed by Tomorrow's Thought Today's Liam Young, with accompanying writing from sci-fi authors Jeff Noon, Pat Cadigan and Tim Maughan. Most recently, Coldcut released Ninja Jamm, a music making app, for Android and iOS, in collaboration with London-based arts and technology firm Seeper. Geared toward both casual listeners and more experienced DJs and music producers, the freemium app allows users to download, remix and make music with samplepacks and tunepacks that feature pro quality sample libraries and also original tracks and mixes by Coldcut, as well as other Ninja artists, creating something new altogether. With the "intuitive yet deep" app, users can turn instruments on and off, swap between clips, add glitches and effects, trigger and pitch-bend stabs and one-off samples, and change the tempo of the track instantly. Users can additionally record as they mix and instantly upload to SoundCloud or save the mixes locally. Tunepack releases for Ninja Jamm are increasingly synchronised with Ninja Tune releases on conventional formats. To date over 30 tunepacks have been released, including Amon Tobin, Bonobo, Coldcut, DJ Food, Martyn, Lapalux, Machinedrum, Raffertie, Irresistible Force, FaltyDL, Shuttle, Starkey. Ninja Jamm was featured by Apple in the New and Noteworthy section of the App Store in the week of release and it received over 100,000 downloads in the first week. Coldcut are developing Ninja Jamm further after the Android release garnered acclaim from the Guardian, Independent, Gizmodo and many more reviewers. In 2015, Coldcut are working on a new album, collaborating with producer Dave Taylor (a.k.a. Solid Groove a.k.a. Switch). This is planned for release in 2016. On 6 December 2017, BBC Radio 4 broadcast a play, "Billie Homeless Dies at the End" by Tom Kelly with electronic music by Coldcut.
https://en.wikipedia.org/wiki?curid=6655
Cuisine A cuisine is a style of cooking characterized by distinctive ingredients, techniques and dishes, and usually associated with a specific culture or geographic region. Regional food preparation traditions, customs and ingredients often combine to create dishes unique to a particular region. A cuisine is primarily influenced by the ingredients that are available locally or through trade, they can even be made into distinct ingredients themselves when they become popular within a region, take for example Japanese rice in Japanese cuisine and New Mexico chile in New Mexican cuisine. Religious food laws can also exercise a strong influence on cuisine, such as Hinduism in Indian cuisine, Sikhism in Punjabi cuisine, Buddhism in East Asian cuisine, Islam in Middle Eastern cuisine, and Judaism in Israeli cuisine. Some factors that have an influence on a region's cuisine include the area's climate, the trade among different countries, religiousness or sumptuary laws and culinary culture exchange. For example, a tropical diet may be based more on fruits and vegetables, while a polar diet might rely more on meat and fish. The area's climate, in large measure, determines the native foods that are available. In addition, climate influences food preservation. For example, foods preserved for winter consumption by smoking, curing, and pickling have remained significant in world cuisines for their altered gustatory properties. The trade among different countries also largely affects a region's cuisine. Dating back to the ancient spice trade, seasonings such as cinnamon, cassia, cardamom, ginger, and turmeric were important items of commerce in the earliest evolution of trade. Cinnamon and cassia found their way to the Middle East at least 4,000 years ago. Certain foods and food preparations are required or proscribed by the religiousness or sumptuary laws, such as Islamic dietary laws and Jewish dietary laws. Culinary culture exchange is also an important factor for cuisine in many regions: Japan's first substantial and direct exposure to the West came with the arrival of European missionaries in the second half of the 16th century. At that time, the combination of Spanish and Portuguese game frying techniques with an East Asian method for cooking vegetables in oil led to the development of "tempura", the "popular Japanese dish in which seafood and many different types of vegetables are coated with batter and deep fried". Cuisine dates back to the Antiquity. As food began to require more planning, there was an emergence of meals that situated around culture. Cuisines evolve continually, and new cuisines are created by innovation and cultural interaction. One recent example is fusion cuisine, which combines elements of various culinary traditions while not being categorized per any one cuisine style, and generally refers to the innovations in many contemporary restaurant cuisines since the 1970s. "Nouvelle cuisine" (New cuisine) is an approach to cooking and food presentation in French cuisine that was popularized in the 1960s by the food critics Henri Gault, who invented the phrase, and his colleagues André Gayot and Christian Millau in a new restaurant guide, the Gault-Millau, or "Le Nouveau Guide". Molecular cuisine, is a modern style of cooking which takes advantage of many technical innovations from the scientific disciplines (molecular cooking). The term was coined in 1999 by the French INRA chemist Hervé This because he wanted to distinguish it from the name Molecular cuisine that was previously introduced by him and the late Oxford physicist Nicholas Kurti. It is also named as multi sensory cooking, modernist cuisine, culinary physics, and experimental cuisine by some chefs. Besides, international trade brings new foodstuffs including ingredients to existing cuisines and leads to changes. The introduction of hot pepper to China from South America around the end of the 17th century, greatly influencing Sichuan cuisine, which combines the original taste (with use of Sichuan pepper) with the taste of newly introduced hot pepper and creates a unique mala () flavor that's mouth-numbingly spicy and pungent. A global cuisine is a cuisine that is practiced around the world, and can be categorized according to the common use of major foodstuffs, including grains, produce and cooking fats. Regional cuisines can vary based on availability and usage of specific ingredients, local cooking traditions and practices, as well as overall cultural differences. Such factors can be more-or-less uniform across wide swaths of territory, or vary intensely within individual regions. For example, in Central and South America, corn (maize), both fresh and dried, is a staple food, and is used in many different ways. In northern Europe, wheat, rye, and fats of animal origin predominate, while in southern Europe olive oil is ubiquitous and rice is more prevalent. In Italy, the cuisine of the north, featuring butter and rice, stands in contrast to that of the south, with its wheat pasta and olive oil. In some parts of China, rice is the staple, while in others this role is filled by noodles and bread. Throughout the Middle East and Mediterranean, common ingredients include lamb, olive oil, lemons, peppers, and rice. The vegetarianism practiced in much of India has made pulses (crops harvested solely for the dry seed) such as chickpeas and lentils as important as wheat or rice. From India to Indonesia, the extensive use of spices is characteristic; coconuts and seafood are also used throughout the region both as foodstuffs and as seasonings. African cuisines use a combination of locally available fruits, cereals and vegetables, as well as milk and meat products. In some parts of the continent, the traditional diet features a preponderance of milk, curd and whey products. In much of tropical Africa, however, cow's milk is rare and cannot be produced locally (owing to various diseases that affect livestock). The continent's diverse demographic makeup is reflected in the many different eating and drinking habits, dishes, and preparation techniques of its manifold populations. Asian cuisines are many and varied, and include East Asian cuisine, South Asian cuisine, Southeast Asian cuisine, Central Asian cuisine and West Asian cuisine. Ingredients common to East Asia and Southeast Asia (due to overseas Chinese influence) include rice, ginger, garlic, sesame seeds, chilies, dried onions, soy, and tofu, with stir frying, steaming, and deep frying being common cooking methods. While rice is common to most regional cuisines in Asia, different varieties are popular in the different regions: Basmati rice is popular in South Asia, Jasmine rice in Southeast Asia, and long-grain rice in China and short-grain rice in Japan and Korea. Curry is also a common ingredient found in South Asia, Southeast Asia, and East Asia (notably Japanese curry); however, they are not popular in West Asian and Central Asian cuisines. Those curry dishes with origins in South Asia usually have a yogurt base, with origins in Southeast Asia a coconut milk base, and in East Asia a stewed meat and vegetable base. South Asian cuisine and Southeast Asian cuisine are often characterized by their extensive use of spices and herbs native to the tropical regions of Asia. European cuisine (alternatively, "Western cuisine") include the cuisines of Europe and other Western countries. European cuisine includes that of Europe and to some extent Russia, as well as non-indigenous cuisines of North America, Australasia, Oceania, and Latin America. The term is used by East Asians to contrast with East Asian styles of cooking. When used in English, the term may refer more specifically to cuisine "in" (Continental) Europe; in this context, a synonym is Continental cuisine, especially in British English. Oceanian cuisines include Australian cuisine, New Zealand cuisine, and the cuisines from many other islands or island groups throughout Oceania. Australian cuisine consists of immigrant Anglo-Celtic derived cuisine, and Bushfood prepared and eaten by native Aboriginal Australian peoples, and various newer Asian influences. New Zealand cuisine also consists of European inspired dishes, such as Pavlova, and native Maori cuisine. Across Oceania, staples include the Kumura (Sweet potato) and Taro, which was/is a staple from Papua New Guinea to the South Pacific. On most islands in the south pacific, fish are widely consumed because of the proximity to the ocean. The cuisines of the Americas are found across North and South America, and are based on the cuisines of the countries from which the immigrant people came, primarily Europe. However, the traditional European cuisine has been adapted by the addition of many local and native ingredients, and many techniques have been added to traditional foods as well. Native American cuisine is prepared by indigenous populations across the continent, and its influences can be seen on multi-ethnic Latin American cuisine. Many staple foods eaten across the continent, such as corn (maize), beans, and potatoes have native origins. The regional cuisines are North American cuisine, Mexican cuisine, Central American cuisine, South American cuisine, and Caribbean cuisine.
https://en.wikipedia.org/wiki?curid=6656
Codec A codec is a device or computer program which encodes or decodes a digital data stream or signal. "Codec" is a portmanteau of "coder-decoder". A coder encodes a data stream or a signal for transmission or storage, possibly in encrypted form, and the decoder function reverses the encoding for playback or editing. Codecs are used in videoconferencing, streaming media, and video editing applications. In the mid-20th century, a codec was a device that coded analog signals into digital form using pulse-code modulation (PCM). Later, the name was also applied to software for converting between digital signal formats, including compander functions. An audio codec converts analog audio signals into digital signals for transmission or encodes them for storage. A receiving device converts the digital signals back to analog form using an audio decoder for playback. An example of this are the codecs used in the sound cards of personal computers. A video codec accomplishes the same task for video signals. In addition to encoding a signal, a codec may also compress the data to reduce transmission bandwidth or storage space. Compression codecs are classified primarily into lossy codecs and lossless codecs. Lossless codecs are often used for archiving data in a compressed form while retaining all information present in the original stream. If preserving the original quality of the stream is more important than eliminating the correspondingly larger data sizes, lossless codecs are preferred. This is especially true if the data is to undergo further processing (for example editing) in which case the repeated application of processing (encoding and decoding) on lossy codecs will degrade the quality of the resulting data such that it is no longer identifiable (visually, audibly or both). Using more than one codec or encoding scheme successively can also degrade quality significantly. The decreasing cost of storage capacity and network bandwidth has a tendency to reduce the need for lossy codecs for some media. Many popular codecs are lossy. They reduce quality in order to maximize compression. Often, this type of compression is virtually indistinguishable from the original uncompressed sound or images, depending on the codec and the settings used. The most widely used lossy data compression technique in digital media is based on the discrete cosine transform (DCT), used in compression standards such as JPEG images, H.26x and MPEG video, and MP3 and AAC audio. Smaller data sets ease the strain on relatively expensive storage sub-systems such as non-volatile memory and hard disk, as well as write-once-read-many formats such as CD-ROM, DVD and Blu-ray Disc. Lower data rates also reduce cost and improve performance when the data is transmitted. Two principal techniques are used in codecs, pulse-code modulation and delta modulation. Codecs are often designed to emphasize certain aspects of the media to be encoded. For example, a digital video (using a DV codec) of a sports event needs to encode motion well but not necessarily exact colors, while a video of an art exhibit needs to encode color and surface texture well. Audio codecs for cell phones need to have very low latency between source encoding and playback. In contrast, audio codecs for recording or broadcast can use high-latency audio compression techniques to achieve higher fidelity at a lower bit-rate. There are thousands of audio and video codecs, ranging in cost from free to hundreds of dollars or more. This variety of codecs can create compatibility and obsolescence issues. The impact is lessened for older formats, for which free or nearly-free codecs have existed for a long time. The older formats are often ill-suited to modern applications, however, such as playback in small portable devices. For example, raw uncompressed PCM audio (44.1 kHz, 16 bit stereo, as represented on an audio CD or in a .wav or .aiff file) has long been a standard across multiple platforms, but its transmission over networks is slow and expensive compared with more modern compressed formats, such as Opus and MP3. Many multimedia data streams contain both audio and video, and often some metadata that permits synchronization of audio and video. Each of these three streams may be handled by different programs, processes, or hardware; but for the multimedia data streams to be useful in stored or transmitted form, they must be encapsulated together in a container format. Lower bitrate codecs allow more users, but they also have more distortion. Beyond the initial increase in distortion, lower bit rate codecs also achieve their lower bit rates by using more complex algorithms that make certain assumptions, such as those about the media and the packet loss rate. Other codecs may not make those same assumptions. When a user with a low bitrate codec talks to a user with another codec, additional distortion is introduced by each transcoding. Audio Video Interleave (AVI) is sometimes erroneously described as a codec, but AVI is actually a container format, while a codec is a software or hardware tool that encodes or decodes audio or video into or from some audio or video format. Audio and video encoded with many codecs might be put into an AVI container, although AVI is not an ISO standard. There are also other well-known container formats, such as Ogg, ASF, QuickTime, RealMedia, Matroska, and DivX Media Format. MPEG transport stream, MPEG program stream, MP4, and ISO base media file format are examples of container formats that are ISO standardized.
https://en.wikipedia.org/wiki?curid=6660
Clyde Tombaugh Clyde William Tombaugh (February 4, 1906 January 17, 1997) was an American astronomer. He discovered Pluto in 1930, the first object to be discovered in what would later be identified as the Kuiper belt. At the time of discovery, Pluto was considered a planet but was later reclassified as a dwarf planet in 2006. Tombaugh also discovered many asteroids. He called for the serious scientific research of unidentified flying objects, or UFOs. Tombaugh was born in Streator, Illinois, son of Muron Dealvo Tombaugh, a farmer, and his wife Adella Pearl Chritton. After his family moved to Burdett, Kansas, in 1922, Tombaugh's plans for attending college were frustrated when a hailstorm ruined his family's farm crops. Starting in 1926, he built several telescopes with lenses and mirrors by himself. To better test his telescope mirrors, Tombaugh, with just a pick and shovel, dug a pit 24 feet long, 8 feet deep, and 7 feet wide. This provided a constant air temperature, free of air currents, and was also used by the family as a root cellar and emergency shelter. He sent drawings of Jupiter and Mars to the Lowell Observatory, at Flagstaff, Arizona, which offered him a job. Tombaugh worked there from 1929 to 1945. Following his discovery of Pluto, Tombaugh earned bachelor's and master's degrees in astronomy from the University of Kansas in 1936 and 1938. During World War II he taught naval personnel navigation at Northern Arizona University. He worked at White Sands Missile Range in the early 1950s, and taught astronomy at New Mexico State University from 1955 until his retirement in 1973. In 1980 he was inducted into the International Space Hall of Fame. The asteroid 1604 Tombaugh, discovered in 1931, is named after him. He discovered hundreds of asteroids, beginning with 2839 Annette in 1929, mostly as a by-product of his search for Pluto and his searches for other celestial objects. Tombaugh named some of them after his wife, children and grandchildren. The Royal Astronomical Society awarded him the Jackson-Gwilt Medal in 1931. Direct visual observation became rare in astronomy. By 1965 Robert S. Richardson called Tombaugh one of two great living experienced visual observers as talented as Percival Lowell or Giovanni Schiaparelli. In 1980, Tombaugh and Patrick Moore wrote a book "Out of the Darkness: The Planet Pluto". In August 1992, JPL scientist Robert Staehle called Tombaugh, requesting permission to visit his planet. "I told him he was welcome to it", Tombaugh later remembered, "though he's got to go one long, cold trip." The call eventually led to the launch of the "New Horizons" space probe to Pluto in 2006. Following the passage on July 14, 2015, of Pluto by the "New Horizons" spacecraft the "Cold Heart of Pluto" was named Tombaugh Regio. Tombaugh died on January 17, 1997, when he was in Las Cruces, New Mexico, at the age of 90. He was cremated. A small portion of his ashes was placed aboard the "New Horizons" spacecraft. The container includes the inscription: "Interred herein are remains of American Clyde W. Tombaugh, discoverer of Pluto and the Solar System's 'third zone'. Adelle and Muron's boy, Patricia's husband, Annette and Alden's father, astronomer, teacher, punster, and friend: Clyde W. Tombaugh (1906–1997)". Tombaugh was survived by his wife, Patricia (1912–2012), and their children, Annette and Alden. Tombaugh was an active Unitarian Universalist, and he and his wife helped found the Unitarian Universalist Church of Las Cruces, New Mexico. Clyde Tombaugh had five siblings. Through the daughter of his youngest brother Robert, he is the great-uncle of Los Angeles Dodgers pitcher Clayton Kershaw. While a young researcher working for the Lowell Observatory in Flagstaff, Arizona, Tombaugh was given the job to perform a systematic search for a trans-Neptunian planet (also called Planet X), which had been predicted by Percival Lowell based on calculations performed by his student mathematician Elizabeth Williams and William Pickering. Starting 6 April, 1929, Tombaugh used the observatory's 13-inch astrograph to take photographs of the same section of sky several nights apart. He then used a blink comparator to compare the different images. When he shifted between the two images, a moving object, such as a planet, would appear to jump from one position to another, while the more distant objects such as stars would appear stationary. Tombaugh noticed such a moving object in his search, near the place predicted by Lowell, and subsequent observations showed it to have an orbit beyond that of Neptune. This ruled out classification as an asteroid, and they decided this was the ninth planet that Lowell had predicted. The discovery was made on Tuesday, February 18, 1930, using images taken the previous month. Three classical mythological names were about equally popular among proposals for the new planet: Minerva, Cronus and Pluto. However, Minerva was already in use and the primary supporter of Cronus was widely disliked, leaving Pluto as the front-runner. Outside of Lowell staff, it was first proposed by an 11-year-old English schoolgirl, Venetia Burney. In its favor was that the Pluto of Roman mythology was able to render himself invisible, and that its first two letters formed Percival Lowell's initials. In order to avoid the name changes suffered by Neptune, the name was proposed to both the American Astronomical Society and the Royal Astronomical Society, both of which approved it unanimously. The name was officially adopted on May 1, 1930. Following the discovery, it was recognized that Pluto wasn't massive enough to be the expected ninth planet, and some astronomers began to consider it the first of a new class of object – and indeed Tombaugh searched for additional Trans-Neptunian objects for years, though due to the lack of any further discoveries he concluded that Pluto was indeed a planet. The idea that Pluto was not a true planet remained a minority position until the discovery of other Kuiper belt objects in the late 1990s, which showed that it did not orbit alone but was at best the largest of a number of icy bodies in its region of space. After it was shown that at least one such body was more massive than Pluto, the International Astronomical Union (IAU) reclassified Pluto on August 24, 2006, as a dwarf planet, leaving eight planets in the Solar System. Tombaugh's widow Patricia stated after the IAU's decision that while he might have been disappointed with the change since he had resisted attempts to remove Pluto's planetary status in his lifetime, he would have accepted the decision now if he were alive. She noted that he "was a scientist. He would understand they had a real problem when they start finding several of these things flying around the place." Hal Levison offered this perspective on Tombaugh's place in history: "Clyde Tombaugh discovered the Kuiper Belt. That's a helluva lot more interesting than the ninth planet." Tombaugh continued searching for over a decade after the discovery of Pluto, and the lack of further discoveries left him satisfied that no other object of a comparable apparent magnitude existed near the ecliptic. No more trans-Neptunian objects were discovered until 15760 Albion in 1992. However, more recently the relatively bright object has been discovered. It has a relatively high orbital inclination, but at the time of Tombaugh's discovery of Pluto, Makemake was only a few degrees from the ecliptic near the border of Taurus and Auriga at an apparent magnitude of 16. This position was also very near the galactic equator, making it almost impossible to find such an object within the dense concentration of background stars of the Milky Way. In the fourteen years of looking for planets, until he was drafted in July 1943, Tombaugh looked for motion in 90 million star images (two each of 45 million stars). Tombaugh is officially credited by the Minor Planet Center with discovering 15 asteroids, and he observed nearly 800 asteroids during his search for Pluto and years of follow-up searches looking for another candidate for the postulated Planet X. Tombaugh is also credited with the discovery of periodic comet 274P/Tombaugh–Tenagra. He also discovered hundreds of variable stars, as well as star clusters, galaxy clusters, and a galaxy supercluster. Tombaugh was probably the most eminent astronomer to have reported seeing unidentified flying objects. On August 20, 1949, Tombaugh saw several unidentified objects near Las Cruces, New Mexico. He described them as six to eight rectangular lights, stating: "I doubt that the phenomenon was any terrestrial reflection, because... nothing of the kind has ever appeared before or since... I was so unprepared for such a strange sight that I was really petrified with astonishment.". Tombaugh observed these rectangles of light for about 3 seconds and his wife saw them for about seconds. He never supported the interpretation as a spaceship that has often been attributed to him. He considered other possibilities, with a temperature inversion as the most likely cause.From my own studies of the solar system I cannot entertain any serious possibility for intelligent life on other planets, not even for Mars... The logistics of visitations from planets revolving around the nearer stars is staggering. In consideration of the hundreds of millions of years in the geologic time scale when such visits may have possibly occurred, the odds of a single visit in a given century or millennium are overwhelmingly against such an event. A much more likely source of explanation is some natural optical phenomenon in our own atmosphere. In my 1949 sightings the faintness of the object, together with the manner of fading in intensity as it traveled away from the zenith towards the southeastern horizon, is quite suggestive of a reflection from an optical boundary or surface of slight contrast in refractive index, as in an inversion layer. I have never seen anything like it before or since, and I have spent a lot of time where the night sky could be seen well. This suggests that the phenomenon involves a comparatively rare set of conditions or circumstances to produce it, but nothing like the odds of an interstellar visitation. Another sighting by Tombaugh a year or two later while at a White Sands observatory was of an object of −6 magnitude, four times brighter than Venus at its brightest, going from the zenith to the southern horizon in about 3 seconds. The object executed the same maneuvers as in Tombaugh's first sighting. Tombaugh later reported having seen three of the mysterious green fireballs, which suddenly appeared over New Mexico in late 1948 and continued at least through the early 1950s. A researcher on Project Twinkle reported that Tombaugh "... never observed an unexplainable aerial object despite his continuous and extensive observations of the sky." According to an entry in "UFO updates", Tombaugh said: "I have seen three objects in the last seven years which defied any explanation of known phenomenon, such as Venus, atmospheric optic, meteors or planes. I am a professional, highly skilled, professional astronomer. In addition I have seen three green fireballs which were unusual in behavior from normal green fireballs... I think that several reputable scientists are being unscientific in refusing to entertain the possibility of extraterrestrial origin and nature." Shortly after this, in January 1957, in an Associated Press article in the "Alamogordo Daily News" titled "Celestial Visitors May Be Invading Earth's Atmosphere", Tombaugh was again quoted on his sightings and opinion about them. "Although our own solar system is believed to support no other life than on Earth, other stars in the galaxy may have hundreds of thousands of habitable worlds. Races on these worlds may have been able to utilize the tremendous amounts of power required to bridge the space between the stars ...". Tombaugh stated that he had observed celestial phenomena which he could not explain, but had seen none personally since 1951 or 1952. "These things, which do appear to be directed, are unlike any other phenomena I ever observed. Their apparent lack of obedience to the ordinary laws of celestial motion gives credence." In 1949, Tombaugh had also told the Naval missile director at White Sands Missile Range, Commander Robert McLaughlin, that he had seen a bright flash on Mars on August 27, 1941, which he now attributed to an atomic blast. Tombaugh also noted that the first atomic bomb tested in New Mexico would have lit up the dark side of the Earth like a neon sign and that Mars was coincidentally quite close at the time, the implication apparently being that the atomic test would have been visible from Mars. In June 1952, Dr. J. Allen Hynek, an astronomer acting as a scientific consultant to the Air Force's Project Blue Book UFO study, secretly conducted a survey of fellow astronomers on UFO sightings and attitudes while attending an astronomy convention. Tombaugh and four other astronomers, including Dr. Lincoln LaPaz of the University of New Mexico, told Hynek about their sightings. Tombaugh also told Hynek that his telescopes were at the Air Force's disposal for taking photos of UFOs, if he was properly alerted. Tombaugh's offer may have led to his involvement in a search for near-Earth satellites, first announced in late 1953 and sponsored by the Army Office of Ordnance Research. Another public statement was made on the search in March 1954, emphasizing the rationale that such an orbiting object would serve as a natural space station. However, according to Donald Keyhoe, later director of the National Investigations Committee on Aerial Phenomena (NICAP), the real reason for the sudden search was because two near-Earth orbiting objects had been picked up on new long-range radar in the summer of 1953, according to his Pentagon source. By May 1954, Keyhoe was making public statements that his sources told him the search had indeed been successful, and either one or two objects had been found. However, the story did not break until August 23, 1954, when "Aviation Week" magazine stated that two satellites had been found only 400 and 600 miles out. They were termed "natural satellites" and implied that they had been recently captured, despite this being a virtual impossibility. The next day, the story was in many major newspapers. Dr. LaPaz was implicated in the discovery in addition to Tombaugh. LaPaz had earlier conducted secret investigations on behalf of the Air Force on the green fireballs and other unidentified aerial phenomena over New Mexico. "The New York Times" reported on August 29 that "a source close to the O. O. R. unit here described as 'quite accurate' the report in the magazine Aviation Week that two previously unobserved satellites had been spotted and identified by Dr. Lincoln LaPaz of the University of New Mexico as natural and not artificial objects. This source also said there was absolutely no connection between the reported satellites and flying saucer reports." However, in the October 10th issue, LaPaz said the magazine article was "false in every particular, in so far as reference to me is concerned." Both LaPaz and Tombaugh were to issue public denials that anything had been found. The October 1955 issue of "Popular Mechanics" magazine reported: "Professor Tombaugh is closemouthed about his results. He won't say whether or not any small natural satellites have been discovered. He does say, however, that newspaper reports of 18 months ago announcing the discovery of natural satellites at 400 and 600 miles out are not correct. He adds that there is no connection between the search program and the reports of so-called flying saucers." At a meteor conference in Los Angeles in 1957, Tombaugh reiterated that his four-year search for "natural satellites" had been unsuccessful. In 1959, Tombaugh was to issue a final report stating that nothing had been found in his search. His personal 16-inch telescope was reassembled and dedicated on September 17, 2009, at Rancho Hidalgo, New Mexico (near Animas, New Mexico), adjacent to "Astronomy" 's new observatory.
https://en.wikipedia.org/wiki?curid=6663
CPAN The Comprehensive Perl Archive Network (CPAN) is a repository of over 250,000 software modules and accompanying documentation for 39,000 distributions, written in the Perl programming language by over 12,000 contributors. "CPAN" can denote either the archive network or the Perl program that acts as an interface to the network and as an automated software installer (somewhat like a package manager). Most software on CPAN is free and open source software. CPAN was conceived in 1993 and has been active online since October 1995. It is based on the CTAN model and began as a place to unify the structure of scattered Perl archives. Like many programming languages, Perl has mechanisms to use external libraries of code, making one file contain common routines used by several programs. Perl calls these "modules". Perl modules are typically installed in one of several directories whose paths are placed in the Perl interpreter when it is first compiled; on Unix-like operating systems, common paths include "/usr/lib/perl5", "/usr/local/lib/perl5", and several of their subdirectories. Perl comes with a small set of "core modules". Some of these perform bootstrapping tasks, such as codice_1, which is used for building and installing other extension modules; others, like codice_2, are merely commonly used. CPAN's main purpose is to help programmers locate modules and programs not included in the Perl standard distribution. Its structure is decentralized. Authors maintain and improve their own modules. Forking, and creating competing modules for the same task or purpose, is common. There is a third-party bug tracking system that is automatically set up for any uploaded distribution, but authors may opt to use a different bug tracking system such as GitHub. Similarly, though GitHub is a popular location to store the source for distributions, it may be stored anywhere the author prefers, or may not be publicly accessible at all. Maintainers may grant permissions to others to maintain or take over their modules, and permissions may be granted by admins for those wishing to take over abandoned modules. Previous versions of updated distributions are retained on CPAN until deleted by the uploader, and a secondary mirror network called BackPAN retains distributions even if they are deleted from CPAN. Also, the complete history of the CPAN and all its modules is available as the GitPAN project, allowing to easily see the complete history for all the modules and for easy maintenance of forks. CPAN is also used to distribute new versions of Perl, as well as related projects, such as Parrot and Raku. Files on the CPAN are referred to as "distributions". A distribution may consist of one or more modules, documentation files, or programs packaged in a common archiving format, such as a gzipped tar archive or a ZIP file. Distributions will often contain installation scripts (usually called "Makefile.PL" or "Build.PL") and test scripts which can be run to verify the contents of the distribution are functioning properly. New distributions are uploaded to the Perl Authors Upload Server, or PAUSE (see the section Uploading distributions with PAUSE). In 2003, distributions started to include metadata files, called "META.yml", indicating the distribution's name, version, dependencies, and other useful information; however, not all distributions contain metadata. When metadata is not present in a distribution, the PAUSE's software will try to analyze the code in the distribution to look for the same information; this is not necessarily very reliable. In 2010, version 2 of this specification was created to be used via a new file called "META.json", with the YAML format file often also included for backward compatibility. With thousands of distributions, CPAN needs to be structured to be useful. Authors often place their modules in the natural hierarchy of Perl module names (such as codice_3 or codice_4) according to purpose or domain, though this is not enforced. CPAN module distributions usually have names in the form of "CGI-Application-3.1" (where the :: used in the module's name has been replaced with a dash, and the version number has been appended to the name), but this is only a convention; many prominent distributions break the convention, especially those that contain multiple modules. Security restrictions prevent a distribution from ever being replaced with an identical filename, so virtually all distribution names do include a version number. The distribution infrastructure of CPAN consists of its worldwide network of more than 250 mirrors in more than 60 countries. Each full mirror hosts around 20 gigabytes of data. Most mirrors update themselves hourly, daily or bidaily from the CPAN master site. Some sites are major FTP servers which mirror lots of other software, but others are simply servers owned by companies that use Perl heavily. There are at least two mirrors on every continent except Antarctica. Several search engines have been written to help Perl programmers sort through the CPAN. The official includes textual search, a browsable index of modules, and extracted copies of all distributions currently on the CPAN. On 16 May 2018, the Perl Foundation announced that search.cpan.org would be shut down on 29 June 2018 (after 19 years of operation), due to its aging codebase and maintenance burden. Users will be transitioned and redirected to the third-party alternative MetaCPAN. CPAN Testers are a group of volunteers, who will download and test distributions as they are uploaded to CPAN. This enables the authors to have their modules tested on many platforms and environments that they would otherwise not have access to, thus helping to promote portability, as well as a degree of quality. Smoke testers send reports, which are then collated and used for a variety of presentation websites, including the main reports site, statistics and dependencies. Authors can upload new distributions to the CPAN through the "Perl Authors Upload Server" (PAUSE). To do so, they must request a PAUSE account. Once registered, they may use a web interface at pause.perl.org, or an FTP interface to upload files to their directory and delete them. Modules in the upload will only be indexed as canonical if the module name has not been used before (granting "first-come" permission to the uploader), or if the uploader has permission for that name, and if the module is a higher version than any existing entry. This can be specified through PAUSE's web interface. There is also a Perl core module named CPAN; it is usually differentiated from the repository itself by using the name CPAN.pm. CPAN.pm is mainly an interactive shell which can be used to search for, download, and install distributions. An interactive shell called cpan is also provided in the Perl core, and is the usual way of running CPAN.pm. After a short configuration process and mirror selection, it uses tools available on the user's computer to automatically download, unpack, compile, test, and install modules. It is also capable of updating itself. An effort to replace CPAN.pm with something cleaner and more modern resulted in the CPANPLUS (or CPAN++) set of modules. CPANPLUS separates the back-end work of downloading, compiling, and installing modules from the interactive shell used to issue commands. It also supports several advanced features, such as cryptographic signature checking and test result reporting. Finally, CPANPLUS can uninstall a distribution. CPANPLUS was added to the Perl core in version 5.10.0, and removed from it in version 5.20.0. A smaller, leaner modern alternative to these CPAN installers was developed called cpanminus. cpanminus was designed to have a much smaller memory footprint as often required in limited memory environments, and to be usable as a standalone script such that it can even install itself, requiring only the expected set of core Perl modules to be available. It is also available from CPAN as the module App::cpanminus, which installs the cpanm script. It does not maintain or rely on a persistent configuration, but is configured only by the environment and command-line options. cpanminus does not have an interactive shell component. It recognizes the cpanfile format for specifying prerequisites, useful in ad-hoc Perl projects that may not be designed for CPAN installation. cpanminus also has the ability to uninstall distributions. Each of these modules can check a distribution's dependencies and recursively install any prerequisites, either automatically or with individual user approval. Each support FTP and HTTP and can work through firewalls and proxies. Experienced Perl programmers often comment that half of Perl's power is in the CPAN. It has been called Perl's killer app. It is roughly equivalent to the PECL and PEAR for PHP; the PyPI (Python Package Index) repository for Python; RubyGems for Ruby; CRAN for R; npm for Node.js; LuaRocks for Lua; Maven for Java; and Hackage for Haskell. CPAN's use of arbitrated name spaces, a testing regime and a well defined documentation style makes it unique. Given its importance to the Perl developer community, the CPAN both shapes and is shaped by Perl's culture. Its "self-appointed master librarian", Jarkko Hietaniemi, often takes part in the April Fools' Day jokes; on 1 April 2002 the site was temporarily named to "CJAN", where the "J" stood for "Java". In 2003, the codice_5 domain name was redirected to Matt's Script Archive, a site infamous in the Perl community for having badly written code. Some of the distributions on the CPAN are distributed as jokes. The codice_6 hierarchy is reserved for joke modules; for instance, codice_7 adds a codice_8 function that doesn't run the code given to it (to complement the codice_9 built-in, which does). Even outside the codice_6 hierarchy, some modules are still written largely for amusement; one example is codice_11, which can be used to write Perl programs in a subset of Latin. In 2005, a group of Perl developers who also had an interest in JavaScript got together to create JSAN, the JavaScript Archive Network. The JSAN is a near-direct port of the CPAN infrastructure for use with the JavaScript language, which for most of its lifespan did not have a cohesive "community". In 2008, after a chance meeting with CPAN admin Adam Kennedy at the Open Source Developers Conference, Linux kernel developer Rusty Russell created the CCAN, the Comprehensive C Archive Network. The CCAN is a direct port of the CPAN architecture for use with the C language. CRAN, the Comprehensive R Archive Network, is a set of mirrors hosting the R programming language distribution(s), documentation, and contributed extensions. Adam Kennedy is an Australian Perl programmer, and one of several CPAN administrators. Under his CPAN author id of ADAMK, he is the maintainer of over 200 module distributions on CPAN, which places him at the top of the CPAN contribution leaderboard. Kennedy is the first maintainer of more than 200 CPAN modules, many of which he has adopted from other authors and included in his Open Repository, which is available for use by any registered CPAN author. He is a frequent presenter at open source conferences such as OSDC, OSCON, and YAPC as well as the Perl QA hackathons. Kennedy has developed some significant modules for the Perl programming language, particularly in the area of tools to improve the development and build toolchain such as PPI (a Perl parser), CPAN::Metrics (generate metrics on the 16 m+ lines of code in CPAN), Portable Perl (a.k.a. "Perl on a Stick") and Padre (a Perl IDE). He has also been a strong advocate of platform equality for Perl on Windows, and started both the Win32 Perl Wiki and the Strawberry Perl distribution for Windows. The Perl PPI Parser has provided an essential building block for Perl 5 code analysis, documentation, and refactoring tools. Perl::Critic uses PPI to critique Perl source code against the criteria in Perl Best Practices; the Padre IDE uses PPI for code analysis and refactoring. The PPI documentation makes reference to the truism that "Only perl (the interpreter) can parse Perl (source code)" because it is a dynamic language; a post on PerlMonks posits a formal proof. Strawberry Perl is a binary distribution of Perl for the Windows operating system. Unlike most Windows-based Perl distributions, Strawberry Perl also bundles a C compiler, make tool, and some pre-configured modules to improve compatibility with the Windows environment. The aim of this distribution is to provide a practical environment to test and use the latest modules available from CPAN. Larry Wall, creator of Perl, has endorsed this as good port for Windows. Portable Perl is the first Perl distribution suitable for installation to portable flash memory devices, iPods, mobile phones, cameras, etc. The first release was distributed on USB thumbdrives at OSCON 2008. Initial work for this project was funded by a grant from The Perl Foundation. Adam Kennedy has spoken at a number of technical conferences worldwide, including at OSCON, OSDC, Linux.conf.au, and numerous YAPC events. These talks include: Adam moved to San Francisco, CA in the United States in 2012. He worked for Kaggle from 2012 to 2015 and began working as an engineer with Apple in 2016.. He currently holds the position of Data Engineering Manager, Siri Search at Apple
https://en.wikipedia.org/wiki?curid=6667
Colorado Rockies The Colorado Rockies are an American professional baseball team based in Denver, Colorado. The Rockies compete in Major League Baseball (MLB) as a member club of the National League (NL) West division. The team plays its home baseball games at Coors Field, which is located in the Lower Downtown area of Denver. It is owned by the Monfort brothers and managed by Bud Black. The Rockies began play as an expansion team for the 1993 season, and played their home games for their first two seasons at Mile High Stadium. Since 1995, they have played at Coors Field, which has earned a reputation as a hitter's park. The Rockies have qualified for the postseason five times, each time as a Wild Card winner. In 2007, the team earned its first (and only) NL pennant after winning 14 of their final 15 games in the regular season to secure a Wild Card position, capping the streak off with a 13 inning 9-8 victory against the San Diego Padres in the tiebreaker game affectionately known as "Game 163" by Rockies fans. The Rockies then proceeded to sweep the Philadelphia Phillies and Arizona Diamondbacks in the NLDS and NLCS respectively and entered the 2007 World Series as winners of 21 of their last 22 games. However, they were swept by the American League (AL) champions Boston Red Sox in four games. From 1993 to 2019, the Rockies have an overall record of 2,033–2,280 (a winning percentage). Denver had long been a hotbed of Denver Bears/Zephyrs minor league baseball and residents and businesses in the area desired a Major League team. Denver's Mile High Stadium was built originally as Denver Bears Stadium, a minor league baseball stadium that could be upgraded to major league standards. Following the Pittsburgh drug trials in 1985, an unsuccessful attempt was made to purchase the Pittsburgh Pirates and relocate them. However, in January 1990, Colorado's chances for a new team improved significantly when Coors Brewing Company became a limited partner with the AAA Denver Zephyrs. In 1991, as part of Major League Baseball's two-team expansion (along with the Florida (now Miami) Marlins), an ownership group representing Denver led by John Antonucci and Michael I. Monus was granted a franchise. They took the name "Rockies" due to Denver's proximity to the Rocky Mountains, which is reflected in their logo; the name was previously used by the city's first NHL team (who are now the New Jersey Devils). Monus and Antonucci were forced to drop out in 1992 after Monus' reputation was ruined by an accounting scandal. Trucking magnate Jerry McMorris stepped in at the 11th hour to save the franchise, allowing the team to begin play in 1993. The Rockies shared Mile High Stadium with the National Football League (NFL)'s Denver Broncos for their first two seasons while Coors Field was constructed. It was completed for the 1995 Major League Baseball season. In 1993, they started play in the West division of the National League. The Rockies were MLB's first team based in the Mountain Time Zone. They have reached the Major League Baseball postseason five times, each time as the National League wild card team. Twice (1995 and 2009) they were eliminated in the first round of the playoffs. In 2007, the Rockies advanced to the World Series, only to be swept by the Boston Red Sox. Like their expansion brethren, the Miami Marlins, they have never won a division title since their establishment; they are also one of two current MLB teams that have never won their division. The Rockies have played their home games at Coors Field since 1995. Their newest spring training home, Salt River Fields at Talking Stick in Scottsdale, Arizona, opened in March 2011 and is shared with the Arizona Diamondbacks. At the start of the 2012 season, the Rockies introduced "Purple Mondays" in which the team wears its purple uniform every Monday game day. In 2020, Larry Walker was the first Colorado Rockies player to be inducted to the Baseball Hall of Fame. Todd Helton is the first Colorado player to have his number (17) retired, which was done on Sunday, August 17, 2014. Newly-elected Hall of Fame member, Larry Walker, will have his number (33) retired on April 19, 2020 at Coors Field. Jackie Robinson's number, 42, was retired throughout all of baseball in 1997. Keli McGregor had worked with the Rockies since their inception in 1993, rising from senior director of operations to team president in 2002, until his death on April 20, 2010. He is honored at Coors Field alongside Helton and Robinson with his initials. The Colorado Rockies farm system consists of eight minor league affiliates. As of 2010, Rockies' flagship radio station is KOA 850AM, with some late-season games broadcast on KHOW 630 AM due to conflicts with Denver Broncos games. The Rockies Radio Network is composed of 38 affiliate stations in eight states. As of 2019, Jack Corrigifan is the radio announcer, serving as a backup TV announcer whenever Drew Goodman is not available. In January 2020, long-time KOA radio announcer Jerry Schemmel was let go from his role for budgetary reasons from KOA’s parent company. As of 2013, Spanish language radio broadcasts of the Rockies are heard on KNRV 1150 AM. As of 2013, all games are produced and televised by AT&T SportsNet Rocky Mountain. All 150 games produced by AT&T SportsNet Rocky Mountain are broadcast in HD. Jeff Huson and Drew Goodman are the usual the TV broadcast team, with Ryan Spilborghs and Taylor McGregor handling on-field coverage and clubhouse interviews. Jenny Cavnar, Jason Hirsh, and Cory Sullivan handle the pre-game and post-game shows. Schemmel, Corrigan, Spilborghs, Cavnar, and Sullivan also fill in as play-by-play or color commentator during absences of Huson or Goodman. The Rockies led MLB attendance records for the first seven years of their existence. The inaugural season is currently the MLB all-time record for home attendance. + = 57 home games in strike shortened season. ++ = 72 home games in strike shortened season.
https://en.wikipedia.org/wiki?curid=6669
Cement A cement is a binder, a substance used for construction that sets, hardens, and adheres to other materials to bind them together. Cement is seldom used on its own, but rather to bind sand and gravel (aggregate) together. Cement mixed with fine aggregate produces mortar for masonry, or with sand and gravel, produces concrete. Concrete is the most widely used material in existence and is only behind water as the planet's most-consumed resource. Cements used in construction are usually inorganic, often lime or calcium silicate based, which can be characterized as non-hydraulic or hydraulic respectively, depending on the ability of the cement to set in the presence of water (see hydraulic and non-hydraulic lime plaster). Non-hydraulic cement does not set in wet conditions or under water. Rather, it sets as it dries and reacts with carbon dioxide in the air. It is resistant to attack by chemicals after setting. Hydraulic cements (e.g., Portland cement) set and become adhesive due to a chemical reaction between the dry ingredients and water. The chemical reaction results in mineral hydrates that are not very water-soluble and so are quite durable in water and safe from chemical attack. This allows setting in wet conditions or under water and further protects the hardened material from chemical attack. The chemical process for hydraulic cement was found by ancient Romans who used volcanic ash (pozzolana) with added lime (calcium oxide). The word "cement" can be traced back to the Roman term "opus caementicium", used to describe masonry resembling modern concrete that was made from crushed rock with burnt lime as binder. The volcanic ash and pulverized brick supplements that were added to the burnt lime, to obtain a hydraulic binder, were later referred to as "cementum", "cimentum", "cäment", and "cement". In modern times, organic polymers are sometimes used as cements in concrete. If the cement industry were a country, it would be the third largest carbon dioxide emitter in the world with up to 2.8bn tonnes, surpassed only by China and the US. Cement materials can be classified into two distinct categories: non-hydraulic cements and hydraulic cements according to their respective setting and hardening mechanisms. Hydraulic cements setting and hardening involve hydration reactions and therefore require water, while non-hydraulic cements only react with a gas and can directly set under air. Non-hydraulic cement, such as slaked lime (calcium oxide mixed with water), hardens by carbonation in contact with carbon dioxide, which is present in the air (~ 412 vol. ppm ≃ 0.04 vol. %). First calcium oxide (lime) is produced from calcium carbonate (limestone or chalk) by calcination at temperatures above 825 °C (1,517 °F) for about 10 hours at atmospheric pressure: The calcium oxide is then "spent" (slaked) mixing it with water to make slaked lime (calcium hydroxide): Once the excess water is completely evaporated (this process is technically called "setting"), the carbonation starts: This reaction is slow, because the partial pressure of carbon dioxide in the air is low (~ 0.4 millibar). The carbonation reaction requires that the dry cement be exposed to air, so the slaked lime is a non-hydraulic cement and cannot be used under water. This process is called the "lime cycle". Conversely, hydraulic cement hardens by hydration of the clinker minerals when water is added. Hydraulic cements (such as Portland cement) are made of a mixture of silicates and oxides, the four main mineral phases of the clinker, abbreviated in the cement chemist notation, being: The silicates are responsible for the cement's mechanical properties — the tricalcium aluminate and brownmillerite are essential for the formation of the liquid phase during the sintering (firing) process of clinker at high temperature in the kiln. The chemistry of these reactions is not completely clear and is still the object of research. Perhaps the earliest known occurrence of cement is from twelve million years ago. A deposit of cement was formed after an occurrence of oil shale located adjacent to a bed of limestone burned due to natural causes. These ancient deposits were investigated in the 1960s and 1970s. Cement, chemically speaking, is a product that includes lime as the primary curing ingredient, but is far from the first material used for cementation. The Babylonians and Assyrians used bitumen to bind together burnt brick or alabaster slabs. In Egypt stone blocks were cemented together with a mortar made of sand and roughly burnt gypsum (CaSO4 · 2H2O), which often contained calcium carbonate (CaCO3). Lime (calcium oxide) was used on Crete and by the ancient Greeks. There is evidence that the Minoans of Crete used crushed potshards as an artificial pozzolan for hydraulic cement. Nobody knows who first discovered that a combination of hydrated non-hydraulic lime and a pozzolan produces a hydraulic mixture (see also: Pozzolanic reaction), but such concrete was used by the Ancient Macedonians, and three centuries later on a large scale by Roman engineers. The Greeks used volcanic tuff from the island of Thera as their pozzolan and the Romans used crushed volcanic ash (activated aluminium silicates) with lime. This mixture could set under water, increasing its resistance. The material was called "pozzolana" from the town of Pozzuoli, west of Naples where volcanic ash was extracted. In the absence of pozzolanic ash, the Romans used powdered brick or pottery as a substitute and they may have used crushed tiles for this purpose before discovering natural sources near Rome. The huge dome of the Pantheon in Rome and the massive Baths of Caracalla are examples of ancient structures made from these concretes, many of which still stand. The vast system of Roman aqueducts also made extensive use of hydraulic cement. Roman concrete was rarely used on the outside of buildings. The normal technique was to use brick facing material as the formwork for an infill of mortar mixed with an aggregate of broken pieces of stone, brick, potsherds, recycled chunks of concrete, or other building rubble. Any preservation of this knowledge in literature from the Middle Ages is unknown, but medieval masons and some military engineers actively used hydraulic cement in structures such as canals, fortresses, harbors, and shipbuilding facilities. A mixture of lime mortar and aggregate with brick or stone facing material was used in the Eastern Roman Empire as well as in the West into the Gothic period. The German Rhineland continued to use hydraulic mortar throughout the Middle Ages, having local pozzolana deposits called trass. Tabby is a building material made from oyster-shell lime, sand, and whole oyster shells to form a concrete. The Spanish introduced it to the Americas in the sixteenth century. The technical knowledge for making hydraulic cement was formalized by French and British engineers in the 18th century. John Smeaton made an important contribution to the development of cements while planning the construction of the third Eddystone Lighthouse (1755–59) in the English Channel now known as Smeaton's Tower. He needed a hydraulic mortar that would set and develop some strength in the twelve-hour period between successive high tides. He performed experiments with combinations of different limestones and additives including trass and pozzolanas and did exhaustive market research on the available hydraulic limes, visiting their production sites, and noted that the "hydraulicity" of the lime was directly related to the clay content of the limestone used to make it. Smeaton was a civil engineer by profession, and took the idea no further. In the South Atlantic seaboard of the United States, tabby relying on the oyster-shell middens of earlier Native American populations was used in house construction from the 1730s to the 1860s. In Britain particularly, good quality building stone became ever more expensive during a period of rapid growth, and it became a common practice to construct prestige buildings from the new industrial bricks, and to finish them with a stucco to imitate stone. Hydraulic limes were favored for this, but the need for a fast set time encouraged the development of new cements. Most famous was Parker's "Roman cement". This was developed by James Parker in the 1780s, and finally patented in 1796. It was, in fact, nothing like material used by the Romans, but was a "natural cement" made by burning septaria – nodules that are found in certain clay deposits, and that contain both clay minerals and calcium carbonate. The burnt nodules were ground to a fine powder. This product, made into a mortar with sand, set in 5–15 minutes. The success of "Roman cement" led other manufacturers to develop rival products by burning artificial hydraulic lime cements of clay and chalk. Roman cement quickly became popular but was largely replaced by Portland cement in the 1850s. Apparently unaware of Smeaton's work, the same principle was identified by Frenchman Louis Vicat in the first decade of the nineteenth century. Vicat went on to devise a method of combining chalk and clay into an intimate mixture, and, burning this, produced an "artificial cement" in 1817 considered the "principal forerunner" of Portland cement and "...Edgar Dobbs of Southwark patented a cement of this kind in 1811." In Russia, Egor Cheliev created a new binder by mixing lime and clay. His results were published in 1822 in his book "A Treatise on the Art to Prepare a Good Mortar" published in St. Petersburg. A few years later in 1825, he published another book, which described various methods of making cement and concrete, and the benefits of cement in the construction of buildings and embankments. Portland cement, the most common type of cement in general use around the world as a basic ingredient of concrete, mortar, stucco, and non-speciality grout, was developed in England in the mid 19th century, and usually originates from limestone. James Frost produced what he called "British cement" in a similar manner around the same time, but did not obtain a patent until 1822. In 1824, Joseph Aspdin patented a similar material, which he called "Portland cement", because the render made from it was in color similar to the prestigious Portland stone quarried on the Isle of Portland, Dorset, England. However, Aspdins' cement was nothing like modern Portland cement but was a first step in its development, called a "proto-Portland cement". Joseph Aspdins' son William Aspdin had left his father's company and in his cement manufacturing apparently accidentally produced calcium silicates in the 1840s, a middle step in the development of Portland cement. William Aspdin's innovation was counterintuitive for manufacturers of "artificial cements", because they required more lime in the mix (a problem for his father), a much higher kiln temperature (and therefore more fuel), and the resulting clinker was very hard and rapidly wore down the millstones, which were the only available grinding technology of the time. Manufacturing costs were therefore considerably higher, but the product set reasonably slowly and developed strength quickly, thus opening up a market for use in concrete. The use of concrete in construction grew rapidly from 1850 onward, and was soon the dominant use for cements. Thus Portland cement began its predominant role. Isaac Charles Johnson further refined the production of "meso-Portland cement" (middle stage of development) and claimed he was the real father of Portland cement. Setting time and "early strength" are important characteristics of cements. Hydraulic limes, "natural" cements, and "artificial" cements all rely on their belite content for strength development. Belite develops strength slowly. Because they were burned at temperatures below , they contained no alite, which is responsible for early strength in modern cements. The first cement to consistently contain alite was made by William Aspdin in the early 1840s: This was what we call today "modern" Portland cement. Because of the air of mystery with which William Aspdin surrounded his product, others ("e.g.," Vicat and Johnson) have claimed precedence in this invention, but recent analysis of both his concrete and raw cement have shown that William Aspdin's product made at Northfleet, Kent was a true alite-based cement. However, Aspdin's methods were "rule-of-thumb": Vicat is responsible for establishing the chemical basis of these cements, and Johnson established the importance of sintering the mix in the kiln. In the US the first large-scale use of cement was Rosendale cement, a natural cement mined from a massive deposit of a large dolomite deposit discovered in the early 19th century near Rosendale, New York. Rosendale cement was extremely popular for the foundation of buildings ("e.g.", Statue of Liberty, Capitol Building, Brooklyn Bridge) and lining water pipes. Sorel cement was patented in 1867 by Frenchman Stanislas Sorel. It was stronger than Portland cement but its poor water resistance and corrosive qualities limited its use in building construction. The next development in the manufacture of Portland cement was the introduction of the rotary kiln, which produced a stronger, more homogeneous mixture and facilitated a continuous manufacturing process. Calcium aluminate cements were patented in 1908 in France by Jules Bied for better resistance to sulfates. Also in 1908, Thomas Edison experimented with pre-cast concrete in houses in Union, N.J. In the US, after World War One, the long curing time of at least a month for Rosendale cement made it unpopular for constructing highways and bridges, and many states and construction firms turned to Portland cement. Because of the switch to Portland cement, by the end of the 1920s only one of the 15 Rosendale cement companies had survived. But in the early 1930s, builders discovered that, while Portland cement set faster, it was not as durable, especially for highways—to the point that some states stopped building highways and roads with cement. Bertrain H. Wait, an engineer whose company had helped construct the New York City's Catskill Aqueduct, was impressed with the durability of Rosendale cement, and came up with a blend of both Rosendale and Portland cements that had the good attributes of both. It was highly durable and had a much faster setting time. Wait convinced the New York Commissioner of Highways to construct an experimental section of highway near New Paltz, New York, using one sack of Rosendale to six sacks of Portland cement. It was a success, and for decades the Rosendale-Portland cement blend was used in highway and bridge construction. Cementitious materials have been used as a nuclear waste immobilizing matrix for more than a half-century. Technologies of waste cementation have been developed and deployed at industrial scale in many countries. Cementitious wasteforms require a careful selection and design process adapted to each specific type of waste to satisfy the strict waste acceptance criteria for long-term storage and disposal. Modern hydraulic development began with the start of the Industrial Revolution (around 1800), driven by three main needs: Modern cements are often Portland cement or Portland cement blends, but industry also uses other cements. Portland cement is by far the most common type of cement in general use around the world. This cement is made by heating limestone (calcium carbonate) with other materials (such as clay) to in a kiln, in a process known as calcination that liberates a molecule of carbon dioxide from the calcium carbonate to form calcium oxide, or quicklime, which then chemically combines with the other materials in the mix to form calcium silicates and other cementitious compounds. The resulting hard substance, called 'clinker', is then ground with a small amount of gypsum into a powder to make "ordinary Portland cement", the most commonly used type of cement (often referred to as OPC). Portland cement is a basic ingredient of concrete, mortar, and most non-specialty grout. The most common use for Portland cement is to make concrete. Concrete is a composite material made of aggregate (gravel and sand), cement, and water. As a construction material, concrete can be cast in almost any shape, and once it hardens, can be a structural (load bearing) element. Portland cement may be grey or white. Portland cement blends are often available as inter-ground mixtures from cement producers, but similar formulations are often also mixed from the ground components at the concrete mixing plant. Portland blast-furnace slag cement, or Blast furnace cement (ASTM C595 and EN 197-1 nomenclature respectively), contains up to 95% ground granulated blast furnace slag, with the rest Portland clinker and a little gypsum. All compositions produce high ultimate strength, but as slag content is increased, early strength is reduced, while sulfate resistance increases and heat evolution diminishes. Used as an economic alternative to Portland sulfate-resisting and low-heat cements. Portland-fly ash cement contains up to 40% fly ash under ASTM standards (ASTM C595), or 35% under EN standards (EN 197-1). The fly ash is pozzolanic, so that ultimate strength is maintained. Because fly ash addition allows a lower concrete water content, early strength can also be maintained. Where good quality cheap fly ash is available, this can be an economic alternative to ordinary Portland cement. Portland pozzolan cement includes fly ash cement, since fly ash is a pozzolan, but also includes cements made from other natural or artificial pozzolans. In countries where volcanic ashes are available (e.g., Italy, Chile, Mexico, the Philippines), these cements are often the most common form in use. The maximum replacement ratios are generally defined as for Portland-fly ash cement. Portland silica fume cement. Addition of silica fume can yield exceptionally high strengths, and cements containing 5–20% silica fume are occasionally produced, with 10% being the maximum allowed addition under EN 197-1. However, silica fume is more usually added to Portland cement at the concrete mixer. Masonry cements are used for preparing bricklaying mortars and stuccos, and must not be used in concrete. They are usually complex proprietary formulations containing Portland clinker and a number of other ingredients that may include limestone, hydrated lime, air entrainers, retarders, waterproofers and coloring agents. They are formulated to yield workable mortars that allow rapid and consistent masonry work. Subtle variations of Masonry cement in the US are Plastic Cements and Stucco Cements. These are designed to produce a controlled bond with masonry blocks. Expansive cements contain, in addition to Portland clinker, expansive clinkers (usually sulfoaluminate clinkers), and are designed to offset the effects of drying shrinkage normally encountered in hydraulic cements. This cement can make concrete for floor slabs (up to 60 m square) without contraction joints. White blended cements may be made using white clinker (containing little or no iron) and white supplementary materials such as high-purity metakaolin. Colored cements serve decorative purposes. Some standards allow the addition of pigments to produce "colored Portland cement". Other standards (e.g., ASTM) do not allow pigments in Portland cement, and colored cements are sold as "blended hydraulic cements". Very finely ground cements are cement mixed with sand or with slag or other pozzolan type minerals that are extremely finely ground together. Such cements can have the same physical characteristics as normal cement but with 50% less cement, particularly due to their increased surface area for the chemical reaction. Even with intensive grinding they can use up to 50% less energy (and thus less carbon emissions) to fabricate than ordinary Portland cements. "Pozzolan-lime cements" are mixtures of ground pozzolan and lime. These are the cements the Romans used, and are present in surviving Roman structures like the Pantheon in Rome. They develop strength slowly, but their ultimate strength can be very high. The hydration products that produce strength are essentially the same as those in Portland cement. "Slag-lime cements"—ground granulated blast-furnace slag is not hydraulic on its own, but is "activated" by addition of alkalis, most economically using lime. They are similar to pozzolan lime cements in their properties. Only granulated slag (i.e., water-quenched, glassy slag) is effective as a cement component. "Supersulfated cements" contain about 80% ground granulated blast furnace slag, 15% gypsum or anhydrite and a little Portland clinker or lime as an activator. They produce strength by formation of ettringite, with strength growth similar to a slow Portland cement. They exhibit good resistance to aggressive agents, including sulfate. Calcium aluminate cements are hydraulic cements made primarily from limestone and bauxite. The active ingredients are monocalcium aluminate CaAl2O4 (CaO · Al2O3 or CA in Cement chemist notation, CCN) and mayenite Ca12Al14O33 (12 CaO · 7 Al2O3, or C12A7 in CCN). Strength forms by hydration to calcium aluminate hydrates. They are well-adapted for use in refractory (high-temperature resistant) concretes, e.g., for furnace linings. "Calcium sulfoaluminate cements" are made from clinkers that include ye'elimite (Ca4(AlO2)6SO4 or C4A3 in Cement chemist's notation) as a primary phase. They are used in expansive cements, in ultra-high early strength cements, and in "low-energy" cements. Hydration produces ettringite, and specialized physical properties (such as expansion or rapid reaction) are obtained by adjustment of the availability of calcium and sulfate ions. Their use as a low-energy alternative to Portland cement has been pioneered in China, where several million tonnes per year are produced. Energy requirements are lower because of the lower kiln temperatures required for reaction, and the lower amount of limestone (which must be endothermically decarbonated) in the mix. In addition, the lower limestone content and lower fuel consumption leads to a CO2 emission around half that associated with Portland clinker. However, SO2 emissions are usually significantly higher. ""Natural" cements" corresponding to certain cements of the pre-Portland era, are produced by burning argillaceous limestones at moderate temperatures. The level of clay components in the limestone (around 30–35%) is such that large amounts of belite (the low-early strength, high-late strength mineral in Portland cement) are formed without the formation of excessive amounts of free lime. As with any natural material, such cements have highly variable properties. "Geopolymer cements" are made from mixtures of water-soluble alkali metal silicates, and aluminosilicate mineral powders such as fly ash and metakaolin. Polymer cements are made from organic chemicals that polymerise. Producers often use thermoset materials. While they are often significantly more expensive, they can give a water proof material that has useful tensile strength. Cement starts to set when mixed with water, which causes a series of hydration chemical reactions. The constituents slowly hydrate and the mineral hydrates solidify and harden. The interlocking of the hydrates gives cement its strength. Contrary to popular belief, hydraulic cement does not set by drying out — proper curing requires maintaining the appropriate moisture content necessary for the hydration reactions during the setting and the hardening processes. If hydraulic cements dry out during the curing phase, the resulting product can be insufficiently hydrated and significantly weakened. A minimum temperature of 5 °C is recommended, and no more than 30 °C. The concrete at young age must be protected against water evaporation due to direct insolation, elevated temperature, low relative humidity and wind. Bags of cement routinely have health and safety warnings printed on them because not only is cement highly alkaline, but the setting process is exothermic. As a result, wet cement is strongly caustic (pH = 13.5) and can easily cause severe skin burns if not promptly washed off with water. Similarly, dry cement powder in contact with mucous membranes can cause severe eye or respiratory irritation. Some trace elements, such as chromium, from impurities naturally present in the raw materials used to produce cement may cause allergic dermatitis. Reducing agents such as ferrous sulfate (FeSO4) are often added to cement to convert the carcinogenic hexavalent chromate (CrO42−) into trivalent chromium (Cr3+), a less toxic chemical species. Cement users need also to wear appropriate gloves and protective clothing. In 2010, the world production of hydraulic cement was . The top three producers were China with 1,800, India with 220, and USA with 63.5 million tonnes for a total of over half the world total by the world's three most populated states. For the world capacity to produce cement in 2010, the situation was similar with the top three states (China, India, and USA) accounting for just under half the world total capacity. Over 2011 and 2012, global consumption continued to climb, rising to 3585 Mt in 2011 and 3736 Mt in 2012, while annual growth rates eased to 8.3% and 4.2%, respectively. China, representing an increasing share of world cement consumption, remains the main engine of global growth. By 2012, Chinese demand was recorded at 2160 Mt, representing 58% of world consumption. Annual growth rates, which reached 16% in 2010, appear to have softened, slowing to 5–6% over 2011 and 2012, as China's economy targets a more sustainable growth rate. Outside of China, worldwide consumption climbed by 4.4% to 1462 Mt in 2010, 5% to 1535 Mt in 2011, and finally 2.7% to 1576 Mt in 2012. Iran is now the 3rd largest cement producer in the world and has increased its output by over 10% from 2008 to 2011. Due to climbing energy costs in Pakistan and other major cement-producing countries, Iran is in a unique position as a trading partner, utilizing its own surplus petroleum to power clinker plants. Now a top producer in the Middle-East, Iran is further increasing its dominant position in local markets and abroad. The performance in North America and Europe over the 2010–12 period contrasted strikingly with that of China, as the global financial crisis evolved into a sovereign debt crisis for many economies in this region and recession. Cement consumption levels for this region fell by 1.9% in 2010 to 445 Mt, recovered by 4.9% in 2011, then dipped again by 1.1% in 2012. The performance in the rest of the world, which includes many emerging economies in Asia, Africa and Latin America and representing some 1020 Mt cement demand in 2010, was positive and more than offset the declines in North America and Europe. Annual consumption growth was recorded at 7.4% in 2010, moderating to 5.1% and 4.3% in 2011 and 2012, respectively. As at year-end 2012, the global cement industry consisted of 5673 cement production facilities, including both integrated and grinding, of which 3900 were located in China and 1773 in the rest of the world. Total cement capacity worldwide was recorded at 5245 Mt in 2012, with 2950 Mt located in China and 2295 Mt in the rest of the world. "For the past 18 years, China consistently has produced more cement than any other country in the world. [...] (However,) China's cement export peaked in 1994 with 11 million tonnes shipped out and has been in steady decline ever since. Only 5.18 million tonnes were exported out of China in 2002. Offered at $34 a ton, Chinese cement is pricing itself out of the market as Thailand is asking as little as $20 for the same quality." In 2006, it was estimated that China manufactured 1.235 billion tonnes of cement, which was 44% of the world total cement production. "Demand for cement in China is expected to advance 5.4% annually and exceed 1 billion tonnes in 2008, driven by slowing but healthy growth in construction expenditures. Cement consumed in China will amount to 44% of global demand, and China will remain the world's largest national consumer of cement by a large margin." In 2010, 3.3 billion tonnes of cement was consumed globally. Of this, China accounted for 1.8 billion tonnes. Cement manufacture causes environmental impacts at all stages of the process. These include emissions of airborne pollution in the form of dust, gases, noise and vibration when operating machinery and during blasting in quarries, and damage to countryside from quarrying. Equipment to reduce dust emissions during quarrying and manufacture of cement is widely used, and equipment to trap and separate exhaust gases are coming into increased use. Environmental protection also includes the re-integration of quarries into the countryside after they have been closed down by returning them to nature or re-cultivating them. Carbon concentration in cement spans from ≈5% in cement structures to ≈8% in the case of roads in cement. Cement manufacturing releases in the atmosphere both directly when calcium carbonate is heated, producing lime and carbon dioxide, and also indirectly through the use of energy if its production involves the emission of CO2. The cement industry produces about 10% of global man-made CO2 emissions, of which 60% is from the chemical process, and 40% from burning fuel. A Chatham House study from 2018 estimates that the 4 billion tonnes of cement produced annually account for 8% of worldwide CO2 emissions. Nearly 900 kg of CO2 are emitted for every 1000 kg of Portland cement produced. In the European Union, the specific energy consumption for the production of cement clinker has been reduced by approximately 30% since the 1970s. This reduction in primary energy requirements is equivalent to approximately 11 million tonnes of coal per year with corresponding benefits in reduction of CO2 emissions. This accounts for approximately 5% of anthropogenic CO2. The majority of carbon dioxide emissions in the manufacture of Portland cement (approximately 60%) are produced from the chemical decomposition of limestone to lime, an ingredient in Portland cement clinker. These emissions may be reduced by lowering the clinker content of cement. They can also be reduced by alternative fabrication methods such as the intergrinding cement with sand or with slag or other pozzolan type minerals to a very fine powder. To reduce the transport of heavier raw materials and to minimize the associated costs, it is more economical to build cement plants closer to the limestone quarries rather than to the consumer centers. In certain applications, lime mortar reabsorbs some of the CO2 as was released in its manufacture, and has a lower energy requirement in production than mainstream cement. Newly developed cement types from Novacem and Eco-cement can absorb carbon dioxide from ambient air during hardening. In some circumstances, mainly depending on the origin and the composition of the raw materials used, the high-temperature calcination process of limestone and clay minerals can release in the atmosphere gases and dust rich in volatile heavy metals, e.g. thallium, cadmium and mercury are the most toxic. Heavy metals (Tl, Cd, Hg, ...) and also selenium are often found as trace elements in common metal sulfides (pyrite (FeS2), zinc blende (ZnS), galena (PbS), ...) present as secondary minerals in most of the raw materials. Environmental regulations exist in many countries to limit these emissions. As of 2011 in the United States, cement kilns are "legally allowed to pump more toxins into the air than are hazardous-waste incinerators." The presence of heavy metals in the clinker arises both from the natural raw materials and from the use of recycled by-products or alternative fuels. The high pH prevailing in the cement porewater (12.5 < pH < 13.5) limits the mobility of many heavy metals by decreasing their solubility and increasing their sorption onto the cement mineral phases. Nickel, zinc and lead are commonly found in cement in non-negligible concentrations. Chromium may also directly arise as natural impurity from the raw materials or as secondary contamination from the abrasion of hard chromium steel alloys used in the ball mills when the clinker is ground. As chromate (CrO42−) is toxic and may cause severe skin allergies at trace concentration, it is sometimes reduced into trivalent Cr(III) by addition of ferrous sulfate (FeSO4). A cement plant consumes 3 to 6 GJ of fuel per tonne of clinker produced, depending on the raw materials and the process used. Most cement kilns today use coal and petroleum coke as primary fuels, and to a lesser extent natural gas and fuel oil. Selected waste and by-products with recoverable calorific value can be used as fuels in a cement kiln (referred to as co-processing), replacing a portion of conventional fossil fuels, like coal, if they meet strict specifications. Selected waste and by-products containing useful minerals such as calcium, silica, alumina, and iron can be used as raw materials in the kiln, replacing raw materials such as clay, shale, and limestone. Because some materials have both useful mineral content and recoverable calorific value, the distinction between alternative fuels and raw materials is not always clear. For example, sewage sludge has a low but significant calorific value, and burns to give ash containing minerals useful in the clinker matrix. Scrap automobile and truck tires are useful in cement manufacturing as they have high calorific value and the iron embedded in tires is useful as a feed stock. Clinker is manufactured by heating raw materials inside the main burner of a kiln to a temperature of 1450 °C. The flame reaches temperatures of 1800 °C. The material remains at 1200 °C for 12–15 seconds at 1800 °C for 5–8 seconds (also referred to as residence time). These characteristics of a clinker kiln offer numerous benefits and they ensure a complete destruction of organic compounds, a total neutralization of acid gases, sulphur oxides and hydrogen chloride. Furthermore, heavy metal traces are embedded in the clinker structure and no by-products, such as ash of residues, are produced. The EU cement industry already uses more than 40% fuels derived from waste and biomass in supplying the thermal energy to the grey clinker making process. Although the choice for this so-called alternative fuels (AF) is typically cost driven, other factors are becoming more important. Use of alternative fuels provides benefits for both society and the company: CO2-emissions are lower than with fossil fuels, waste can be co-processed in an efficient and sustainable manner and the demand for certain virgin materials can be reduced. Yet there are large differences in the share of alternative fuels used between the European Union (EU) member states. The societal benefits could be improved if more member states increase their alternative fuels share. The Ecofys study assessed the barriers and opportunities for further uptake of alternative fuels in 14 EU member states. The Ecofys study found that local factors constrain the market potential to a much larger extent than the technical and economic feasibility of the cement industry itself. Ecological cement is a cementitious material that meets or exceeds the functional performance capabilities of ordinary Portland cement by incorporating and optimizing recycled materials, thereby reducing consumption of natural raw materials, water, and energy, resulting in a more sustainable construction material. One is Geopolymer cement. New manufacturing processes for producing ecological cement are being researched with the goal to reduce, or even eliminate, the production and release of damaging pollutants and greenhouse gasses, particularly CO2. Growing environmental concerns and the increasing cost of fuels of fossil origin have resulted in many countries in a sharp reduction of the resources needed to produce cement and effluents (dust and exhaust gases). A team at the University of Edinburgh has developed the 'DUPE' process based on the microbial activity of "Sporosarcina pasteurii", a bacterium precipitating calcium carbonate, which, when mixed with sand and urine, can produce mortar blocks with a compressive strength 70% of that of conventional construction materials.
https://en.wikipedia.org/wiki?curid=6670
Cincinnati Reds The Cincinnati Reds are an American professional baseball team based in Cincinnati, Ohio. They compete in Major League Baseball (MLB) as a member club of the National League (NL) Central division. They were a charter member of the American Association in 1882 and joined the NL in 1890. The Reds played in the NL West division from 1969 to 1993, before joining the Central division in 1994. They have won five World Series championships, nine NL pennants, one AA pennant, and 10 division titles. The team plays its home games at Great American Ball Park, which opened in 2003 replacing Riverfront Stadium. Bob Castellini has been chief executive officer since 2006. From 1882 to 2019, the Reds' overall win-loss record is 10,599–10,393 (a 0.505 winning percentage). The origins of the modern Cincinnati Reds can be traced to the expulsion of an earlier team bearing that name. In 1876, Cincinnati became one of the charter members of the new National League, but the club ran afoul of league organizer and long-time president William Hulbert for selling beer during games and renting out their ballpark on Sundays. Both were important activities to entice the city's large German population. While Hulbert made clear his distaste for both beer and Sunday baseball at the founding of the league, neither practice was actually against league rules in those early years. On October 6, 1880, however, seven of the eight team owners pledged at a special league meeting to formally ban both beer and Sunday baseball at the regular league meeting that December. Only Cincinnati president W. H. Kennett refused to sign the pledge, so the other owners formally expelled Cincinnati for violating a rule that would not actually go into effect for two more months. Cincinnati's expulsion from the National League incensed "Cincinnati Enquirer" sports editor O. P. Caylor, who made two attempts to form a new league on behalf of the receivers for the now bankrupt Reds franchise. When these attempts failed, he formed a new independent ballclub known as the Red Stockings in the Spring of 1881, and brought the team to St. Louis for a weekend exhibition. The Reds' first game was a 12–3 victory over the St. Louis club. After the 1881 series proved a success, Caylor and a former president of the old Reds named Justus Thorner received an invitation from Philadelphia businessman Horace Phillips to attend a meeting of several clubs in Pittsburgh with the intent of establishing a rival to the National League. Upon arriving in the city, however, Caylor and Thorner discovered that no other owners had decided to accept the invitation, with even Phillips not bothering to attend his own meeting. By chance, the duo met a former pitcher named Al Pratt, who hooked them up with former Pittsburgh Alleghenys president H. Denny McKnight. Together, the three men hatched a scheme to form a new league by sending a telegram to each of the other owners who were supposed to attend the meeting stating that he was the only person who did not attend and that everyone else was enthusiastic about the new venture and eager to attend a second meeting in Cincinnati. The ploy worked, and the American Association was officially formed at the Hotel Gibson in Cincinnati with the new Reds a charter member with Thorner as president. Led by the hitting of third baseman Hick Carpenter, the defense of future Hall of Fame second baseman Bid McPhee, and the pitching of 40-game-winner Will White, the Reds won the inaugural AA pennant in 1882. With the establishment of the Union Association Justus Thorner left the club to finance the Cincinnati Outlaw Reds and managed to acquire the lease on the Reds Bank Street Grounds playing field, forcing new president Aaron Stern to relocate three blocks away at the hastily built League Park. The club never placed higher than second or lower than fifth for the rest of its tenure in the American Association. The Cincinnati Red Stockings left the American Association on November 14, 1889 and joined the National League along with the Brooklyn Bridegrooms after a dispute with St. Louis Browns owner Chris Von Der Ahe over the selection of a new league president. The National League was happy to accept the teams in part due to the emergence of the new Player's League. This new league, an early failed attempt to break the reserve clause in baseball, threatened both existing leagues. Because the National League decided to expand while the American Association was weakening, the team accepted an invitation to join the National League. It was also at this time that the team first shortened their name from "Red Stockings" to "Reds". The Reds wandered through the 1890s signing local stars and aging veterans. During this time, the team never finished above third place (1897) and never closer than 10 games (1890). At the start of the 20th century, the Reds had hitting stars Sam Crawford and Cy Seymour. Seymour's .377 average in 1905 was the first individual batting crown won by a Red. In 1911, Bob Bescher stole 81 bases, which is still a team record. Like the previous decade, the 1900s (decade) were not kind to the Reds, as much of the decade was spent in the league's second division. In 1912, the club opened a new steel-and-concrete ballpark, Redland Field (later to be known as Crosley Field). The Reds had been playing baseball on that same site, the corner of Findlay and Western Avenues on the city's west side, for 28 years, in wooden structures that had been occasionally damaged by fires. By the late 1910s the Reds began to come out of the second division. The 1918 team finished fourth, and new manager Pat Moran led the Reds to an NL pennant in 1919, in what the club advertised as its "Golden Anniversary". The 1919 team had hitting stars Edd Roush and Heinie Groh while the pitching staff was led by Hod Eller and left-hander Harry "Slim" Sallee. The Reds finished ahead of John McGraw's New York Giants, and then won the world championship in eight games over the Chicago White Sox. By 1920, the "Black Sox" scandal had brought a taint to the Reds' first championship. After 1926, and well into the 1930s, the Reds were second division dwellers. Eppa Rixey, Dolf Luque and Pete Donohue were pitching stars, but the offense never lived up to the pitching. By 1931, the team was bankrupt, the Great Depression was in full swing and Redland Field was in a state of disrepair. Powel Crosley, Jr., an electronics magnate who, with his brother Lewis M. Crosley, produced radios, refrigerators, and other household items, bought the Reds out of bankruptcy in 1933, and hired Larry MacPhail to be the General Manager. Crosley had started WLW radio, the Reds flagship radio broadcaster, and the Crosley Broadcasting Corporation in Cincinnati, where he was also a prominent civic leader. MacPhail began to develop the Reds' minor league system and expanded the Reds' fan base. The Reds, throughout the 1930s, became a team of "firsts". The now-renamed Crosley Field became the host of the first night game in 1935, which was also the first baseball fireworks night, the fireworks at the game were shot by Joe Rozzi of Rozzi's Famous Fireworks. Johnny Vander Meer became the only pitcher in major league history to throw back-to-back no-hitters in 1938. Thanks to Vander Meer, Paul Derringer and second baseman/third baseman-turned-pitcher Bucky Walters, the Reds had a solid pitching staff. The offense came around in the late 1930s. By 1938 the Reds, now led by manager Bill McKechnie, were out of the second division finishing fourth. Ernie Lombardi was named the National League's Most Valuable Player in 1938. By 1939, they were National League champions, but in the World Series, they were swept by the New York Yankees. In 1940, they repeated as NL Champions, and for the first time in 21 years, the Reds captured a World championship, beating the Detroit Tigers 4 games to 3. Frank McCormick was the 1940 NL MVP. Other position players included Harry Craft, Lonny Frey, Ival Goodman, Lew Riggs and Bill Werber. World War II and age finally caught up with the Reds. Throughout the 1940s and early 1950s, Cincinnati finished mostly in the second division. In 1944, Joe Nuxhall (who was later to become part of the radio broadcasting team), at age 15, pitched for the Reds on loan from Wilson Junior High school in Hamilton, Ohio. He became the youngest player ever to appear in a major league game—a record that still stands today. Ewell "The Whip" Blackwell was the main pitching stalwart before arm problems cut short his career. Ted Kluszewski was the NL home run leader in 1954. The rest of the offense was a collection of over-the-hill players and not-ready-for-prime-time youngsters. In April 1953, the Reds announced a preference to be called the "Redlegs", saying that the name of the club had been "Red Stockings" and then "Redlegs". A newspaper speculated that it was due to the developing political connotation of the word 'red' to mean Communism. From 1956 to 1960, the club's logo was altered to remove the term "REDS" from the inside of the "wishbone "C" symbol. The "REDS" reappeared on the 1961 uniforms, but the point of the "C" was removed, leaving a smooth, non-wishbone curve. The traditional home-uniform logo was restored in 1967. In 1956, led by National League Rookie of the Year Frank Robinson, the Redlegs hit 221 HR to tie the NL record. By 1961, Robinson was joined by Vada Pinson, Wally Post, Gordy Coleman, and Gene Freese. Pitchers Joey Jay, Jim O'Toole, and Bob Purkey led the staff. The Reds captured the 1961 National League pennant, holding off the Los Angeles Dodgers and the San Francisco Giants, only to be defeated by the perennially powerful New York Yankees in the World Series. The Reds had winning teams during the rest of the 1960s, but did not produce any championships. They won 98 games in 1962, paced by Purkey's 23, but finished third. In 1964, they lost the pennant by one game to the Cardinals after having taken first place when the Phillies collapsed in September. Their beloved manager Fred Hutchinson died of cancer just weeks after the end of the 1964 season. The failure of the Reds to win the 1964 pennant led to owner Bill DeWitt's selling off key components of the team, in anticipation of relocating the franchise. In response to DeWitt's threatened move, the women of Cincinnati banded together to form the Rosie Reds to urge DeWitt to keep the franchise in Cincinnati. The Rosie Reds are still in existence, and are currently the oldest fan club in Major League Baseball. After the 1965 season he executed what may be the most lopsided trade in baseball history, sending former Most Valuable Player Frank Robinson to the Baltimore Orioles for pitchers Milt Pappas and Jack Baldschun, and outfielder Dick Simpson. Robinson went on to win the MVP and triple crown in the American league for 1966, and lead Baltimore to its first ever World Series title in a sweep of the Los Angeles Dodgers. The Reds did not recover from this trade until the rise of the "Big Red Machine" of the 1970s. Starting in the early 1960s, the Reds' farm system began producing a series of stars, including Jim Maloney (the Reds' pitching ace of the 1960s), Pete Rose, Tony Pérez, Johnny Bench, Lee May, Tommy Helms, Bernie Carbo, Hal McRae, Dave Concepción, and Gary Nolan. The tipping point came in 1967 with the appointment of Bob Howsam as general manager. That same year the Reds avoided a move to San Diego when the city of Cincinnati and Hamilton County agreed to build a state of the art, downtown stadium on the edge of the Ohio River. The Reds entered into a 30-year lease in exchange for the stadium commitment keeping the franchise in its original home city. In a series of strategic moves, Howsam brought in key personnel to complement the homegrown talent. The Reds' final game at Crosley Field, home to more than 4,500 baseball games, was played on June 24, 1970, a 5–4 victory over the San Francisco Giants. Under Howsam's administration starting in the late 1960s, the Reds instituted a strict rule barring the team's players from wearing facial hair and long hair. The clean cut look was meant to present the team as wholesome in an era of turmoil. All players coming to the Reds were required to shave and cut their hair for the next three decades. Over the years, the rule was controversial, but persisted well into the ownership of Marge Schott. On at least one occasion, in the early 1980s, enforcement of this rule lost them the services of star reliever and Ohio native Rollie Fingers, who would not shave his trademark handlebar mustache in order to join the team. The rule was not officially rescinded until 1999 when the Reds traded for slugger Greg Vaughn, who had a goatee. The New York Yankees continue to have a similar rule today, though unlike the Reds during this period, Yankees players are permitted to have mustaches. Much like when players leave the Yankees today, players who left the Reds took advantage with their new teams; Pete Rose, for instance, grew his hair out much longer than would be allowed by the Reds once he signed with the Philadelphia Phillies in 1979. The Reds' rules also included conservative uniforms. In Major League Baseball, a club generally provides most of the equipment and clothing needed for play. However, players are required to supply their gloves and shoes themselves. Many players enter into sponsorship arrangements with shoe manufacturers, but through the mid-1980s, the Reds had a strict rule that players were to wear only plain black shoes with no prominent logo. Reds players decried what they considered to be the boring color choice as well as the denial of the opportunity to earn more money through shoe contracts. A compromise was struck in 1985 in which players could paint red marks on their black shoes, then the following year, they were allowed to wear all-red shoes. In , little known George "Sparky" Anderson was hired as manager, and the Reds embarked upon a decade of excellence, with a team that came to be known as "The Big Red Machine". Playing at Crosley Field until June 30, 1970, when the Reds moved into brand-new Riverfront Stadium, a 52,000 seat multi-purpose venue on the shores of the Ohio River, the Reds began the 1970s with a bang by winning 70 of their first 100 games. Johnny Bench, Tony Pérez, Pete Rose, Lee May and Bobby Tolan were the early Red Machine offensive leaders; Gary Nolan, Jim Merritt, Wayne Simpson and Jim McGlothlin led a pitching staff which also contained veterans Tony Cloninger and Clay Carroll and youngsters Pedro Borbón and Don Gullett. The Reds breezed through the 1970 season, winning the NL West and captured the NL pennant by sweeping the Pittsburgh Pirates in three games. By the time the club got to the World Series, however, the Reds pitching staff had run out of gas and the veteran Baltimore Orioles, led by Hall of Fame third baseman and World Series MVP Brooks Robinson, beat the Reds in five games. After the disastrous season (the only season of the 1970s during which the Reds finished with a losing record) the Reds reloaded by trading veterans Jimmy Stewart, May, and Tommy Helms for Joe Morgan, César Gerónimo, Jack Billingham, Ed Armbrister, and Denis Menke. Meanwhile, Dave Concepción blossomed at shortstop. 1971 was also the year a key component of the future world championships was acquired in George Foster from the San Francisco Giants in a trade for shortstop Frank Duffy. The Reds won the NL West in baseball's first ever strike-shortened season and defeated the Pittsburgh Pirates in an exciting five-game playoff series. They then faced the Oakland Athletics in the World Series. Six of the seven games were won by one run. With powerful slugger Reggie Jackson sidelined by an injury incurred during Oakland's playoff series, Ohio native Gene Tenace got a chance to play in the series, delivering four home runs that tied the World Series record for homers, propelling Oakland to a dramatic seven-game series win. This was one of the few World Series in which no starting pitcher for either side pitched a complete game. The Reds won a third NL West crown in after a dramatic second half comeback, that saw them make up games on the Los Angeles Dodgers after the All-Star break. However they lost the NL pennant to the New York Mets in five games in the NLCS. In game one, Tom Seaver faced Jack Billingham in a classic pitching duel, with all three runs of the 2–1 margin being scored on home runs. John Milner provided New York's run off Billingham, while Pete Rose tied the game in the seventh inning off Seaver, setting the stage for a dramatic game ending home run by Johnny Bench in the bottom of the ninth. The New York series provided plenty of controversy with the riotous behavior of Shea Stadium fans towards Pete Rose when he and Bud Harrelson scuffled after a hard slide by Rose into Harrelson at second base during the fifth inning of Game 3. A full bench-clearing fight resulted after Harrelson responded to Rose's aggressive move to prevent him from completing a double play by calling him a name. This also led to two more incidents in which play was stopped. The Reds trailed 9–3 and New York's manager, Yogi Berra, and legendary outfielder Willie Mays, at the request of National League president Warren Giles, appealed to fans in left field to restrain themselves. The next day the series was extended to a fifth game when Rose homered in the 12th inning to tie the series at two games each. The Reds won 98 games in but they finished second to the 102-win Los Angeles Dodgers. The 1974 season started off with much excitement, as the Atlanta Braves were in town to open the season with the Reds. Hank Aaron entered opening day with 713 home runs, one shy of tying Babe Ruth's record of 714. The first pitch Aaron swung at in the 1974 season was the record tying home run off Jack Billingham. The next day the Braves benched Aaron, hoping to save him for his record-breaking home run on their season-opening homestand. The commissioner of baseball, Bowie Kuhn, ordered Braves management to play Aaron the next day, where he narrowly missed the historic home run in the fifth inning. Aaron went on to set the record in Atlanta two nights later. The 1974 season was also the debut of Hall of Fame radio announcer Marty Brennaman, who replaced Al Michaels, after Michaels left the Reds to broadcast for the San Francisco Giants. With 1975, the Big Red Machine lineup solidified with the "Great Eight" starting team of Johnny Bench (catcher), Tony Pérez (first base), Joe Morgan (second base), Dave Concepción (shortstop), Pete Rose (third base), Ken Griffey (right field), César Gerónimo (center field), and George Foster (left field). The starting pitchers included Don Gullett, Fred Norman, Gary Nolan, Jack Billingham, Pat Darcy, and Clay Kirby. The bullpen featured Rawly Eastwick and Will McEnaney combining for 37 saves, and veterans Pedro Borbón and Clay Carroll. On Opening Day, Rose still played in left field, Foster was not a starter, while John Vukovich, an off-season acquisition, was the starting third baseman. While Vuckovich was a superb fielder, he was a weak hitter. In May, with the team off to a slow start and trailing the Dodgers, Sparky Anderson made a bold move by moving Rose to third base, a position where he had very little experience, and inserting Foster in left field. This was the jolt that the Reds needed to propel them into first place, with Rose proving to be reliable on defense, while adding Foster to the outfield gave the offense some added punch. During the season, the Reds compiled two notable streaks: (1) by winning 41 out of 50 games in one stretch, and (2) by going a month without committing any errors on defense. In the 1975 season, Cincinnati clinched the NL West with 108 victories, then swept the Pittsburgh Pirates in three games to win the NL pennant. In the World Series, the Boston Red Sox were the opponents. After splitting the first four games, the Reds took Game 5. After a three-day rain delay, the two teams met in Game 6, one of the most memorable baseball games ever played and considered by many to be the best World Series game ever. The Reds were ahead 6–3 with 5 outs left, when the Red Sox tied the game on former Red Bernie Carbo's three-run home run. It was Carbo's second pinch-hit three-run homer in the series. After a few close-calls either way, Carlton Fisk hit a dramatic 12th inning home run off the foul pole in left field to give the Red Sox a 7–6 win and force a deciding Game 7. Cincinnati prevailed the next day when Morgan's RBI single won Game 7 and gave the Reds their first championship in 35 years. The Reds have not lost a World Series game since Carlton Fisk's home run, a span of 9 straight wins. 1976 saw a return of the same starting eight in the field. The starting rotation was again led by Nolan, Gullett, Billingham, and Norman, while the addition of rookies Pat Zachry and Santo Alcalá comprised an underrated staff in which four of the six had ERAs below 3.10. Eastwick, Borbon, and McEnaney shared closer duties, recording 26, 8, and 7 saves respectively. The Reds won the NL West by ten games. They went undefeated in the postseason, sweeping the Philadelphia Phillies (winning Game 3 in their final at-bat) to return to the World Series. They continued to dominate by sweeping the Yankees in the newly renovated Yankee Stadium, the first World Series games played in Yankee Stadium since 1964. This was only the second ever sweep of the Yankees in the World Series. In winning the Series, the Reds became the first NL team since the 1921–22 New York Giants to win consecutive World Series championships, and the Big Red Machine of 1975–76 is considered one of the best teams ever. So far in MLB history, the 1975 and '76 Reds were the last NL team to repeat as champions. Beginning with the 1970 National League pennant, the Reds beat either of the two Pennsylvania-based clubs, the Philadelphia Phillies or the Pittsburgh Pirates to win their pennants (Pirates in 1970, 1972, 1975, and 1990, Phillies in 1976), making The Big Red Machine part of the rivalry between the two Pennsylvania teams. In 1979, Pete Rose added further fuel in The Big Red Machine being part of the rivalry when he signed with the Phillies and helped them win their first World Series championship in . The later years of the 1970s brought turmoil and change. Popular Tony Pérez was sent to Montreal after the 1976 season, breaking up the Big Red Machine's starting lineup. Manager Sparky Anderson and General Manager Bob Howsam later considered this trade the biggest mistake of their careers. Starting pitcher Don Gullett left via free agency and signed with the New York Yankees. In an effort to fill that gap, a trade with the Oakland A's for starting ace Vida Blue was arranged during the 1976–77 off-season. However, Bowie Kuhn, the Commissioner of Baseball, vetoed the trade for the stated reason of maintaining competitive balance in baseball. Some have suggested that the actual reason had more to due with Kuhn's continued feud with Oakland A's owner Charlie Finley. On June 15, 1977, the Reds acquired Mets' franchise pitcher Tom Seaver for Pat Zachry, Doug Flynn, Steve Henderson, and Dan Norman. In other deals that proved to be less successful, the Reds traded Gary Nolan to the Angels for Craig Hendrickson, Rawly Eastwick to St. Louis for Doug Capilla and Mike Caldwell to Milwaukee for Rick O'Keeffe and Garry Pyka, and got Rick Auerbach from Texas. The end of the Big Red Machine era was heralded by the replacement of General Manager Bob Howsam with Dick Wagner. In Rose's last season as a Red, he gave baseball a thrill as he challenged Joe DiMaggio's 56-game hitting streak, tying for the second-longest streak ever at 44 games. The streak came to an end in Atlanta after striking out in his fifth at bat in the game against Gene Garber. Rose also earned his 3,000th hit that season, on his way to becoming baseball's all-time hits leader when he rejoined the Reds in the mid-1980s. The year also witnessed the only no-hitter of Hall of Fame pitcher Tom Seaver's career, coming against the St. Louis Cardinals on June 16, 1978. After the 1978 season and two straight second-place finishes, Wagner fired manager Anderson—an unpopular move. Pete Rose, who since 1963 had played almost every position for the team except pitcher, shortstop, and catcher, signed with Philadelphia as a free agent. By 1979, the starters were Bench (c), Dan Driessen (1b), Morgan (2b), Concepción (ss), Ray Knight (3b), with Griffey, Foster, and Geronimo again in the outfield. The pitching staff had experienced a complete turnover since 1976 except for Fred Norman. In addition to ace starter Tom Seaver; the remaining starters were Mike LaCoss, Bill Bonham, and Paul Moskau. In the bullpen, only Borbon had remained. Dave Tomlin and Mario Soto worked middle relief with Tom Hume and Doug Bair closing. The Reds won the 1979 NL West behind the pitching of Tom Seaver but were dispatched in the NL playoffs by Pittsburgh. Game 2 featured a controversial play in which a ball hit by Pittsburgh's Phil Garner was caught by Cincinnati outfielder Dave Collins but was ruled a trap, setting the Pirates up to take a 2–1 lead. The Pirates swept the series 3 games to 0 and went on to win the World Series against the Baltimore Orioles. The 1981 team fielded a strong lineup, but with only Concepción, Foster, and Griffey retaining their spots from the 1975–76 heyday. After Johnny Bench was able to play only a few games at catcher each year after 1980 due to ongoing injuries, Joe Nolan took over as starting catcher. Driessen and Bench shared 1st base, and Knight starred at third. Morgan and Geronimo had been replaced at second base and center field by Ron Oester and Dave Collins. Mario Soto posted a banner year starting on the mound, only surpassed by the outstanding performance of Seaver's Cy Young runner-up season. La Coss, Bruce Berenyi, and Frank Pastore rounded out the starting rotation. Hume again led the bullpen as closer, joined by Bair and Joe Price. In , Cincinnati had the best overall record in baseball, but they finished second in the division in both of the half-seasons that were created after a mid-season players' strike, and missed the playoffs. To commemorate this, a team photo was taken, accompanied by a banner that read "Baseball's Best Record 1981". By , the Reds were a shell of the original Red Machine; they lost 101 games that year. Johnny Bench, after an unsuccessful transition to 3rd base, retired a year later. After the heartbreak of 1981, General Manager Dick Wagner pursued the strategy of ridding the team of veterans including third-baseman Knight and the entire starting outfield of Griffey, Foster, and Collins. Bench, after being able to catch only seven games in 1981, was moved from platooning at first base to be the starting third baseman; Alex Treviño became the regular starting catcher. The outfield was staffed with Paul Householder, César Cedeño, and future Colorado Rockies & Pittsburgh Pirates manager Clint Hurdle on opening day. Hurdle was an immediate bust, and rookie Eddie Milner took his place in the starting outfield early in the year. The highly touted Householder struggled throughout the year despite extensive playing time. Cedeno, while providing steady veteran play, was a disappointment, and was unable to recapture his glory days with the Houston Astros. The starting rotation featured the emergence of a dominant Mario Soto, and featured strong years by Pastore and Bruce Berenyi, but Seaver was injured all year, and their efforts were wasted without a strong offensive lineup. Tom Hume still led the bullpen, along with Joe Price. But the colorful Brad "The Animal" Lesley was unable to consistently excel, and former all-star Jim Kern was a big disappointment. Kern was also publicly upset over having to shave off his prominent beard to join the Reds, and helped force the issue of getting traded during mid-season by growing it back. The season also saw the midseason firing of Manager John McNamara, who was replaced as skipper by Russ Nixon. The Reds fell to the bottom of the Western Division for the next few years. After the 1982 season, Seaver was traded back to the Mets. The year 1983 found Dann Bilardello behind the plate, Bench returning to part-time duty at first base, rookies Nick Esasky taking over at third base and Gary Redus taking over from Cedeno. Tom Hume's effectiveness as a closer had diminished, and no other consistent relievers emerged. Dave Concepción was the sole remaining starter from the Big Red Machine era. Wagner's tenure ended in 1983, when Howsam, the architect of the Big Red Machine, was brought back. The popular Howsam began his second term as Reds' General Manager by signing Cincinnati native Dave Parker as a free agent from Pittsburgh. In the Reds began to move up, depending on trades and some minor leaguers. In that season Dave Parker, Dave Concepción and Tony Pérez were in Cincinnati uniforms. In August 1984, Pete Rose was reacquired and hired to be the Reds player-manager. After raising the franchise from the grave, Howsam gave way to the administration of Bill Bergesch, who attempted to build the team around a core of highly regarded young players in addition to veterans like Parker. However, he was unable to capitalize on an excess of young and highly touted position players including Kurt Stillwell, Tracy Jones, and Kal Daniels by trading them for pitching. Despite the emergence of Tom Browning as rookie of the year in 1985 when he won 20 games, the rotation was devastated by the early demise of Mario Soto's career to arm injury. Under Bergesch, from –89 the Reds finished second four times. Among the highlights, Rose became the all-time hits leader, Tom Browning threw a perfect game, Eric Davis became the first player in baseball history to hit at least 35 home runs and steal 50 bases, and Chris Sabo was the 1988 National League Rookie of the Year. The Reds also had a bullpen star in John Franco, who was with the team from 1984 to 1989. Rose once had Concepción pitch late in a game at Dodger Stadium. Following the release of the Dowd Report which accused Rose for betting on baseball games, in Rose was banned from baseball by Commissioner Bart Giamatti, who declared Rose guilty of "conduct detrimental to baseball". Controversy also swirled around Reds owner Marge Schott, who was accused several times of ethnic and racial slurs. In , General Manager Bergesch was replaced by Murray Cook, who initiated a series of deals that would finally bring the Reds back to the championship, starting with acquisitions of Danny Jackson and José Rijo. An aging Dave Parker was let go after a revival of his career in Cincinnati following the Pittsburgh drug trials. Barry Larkin emerged as the starting shortstop over Kurt Stillwell, who along with reliever Power, was traded for Jackson. In , Cook was succeeded by Bob Quinn, who put the final pieces of the championship puzzle together, with the acquisitions of Hal Morris, Billy Hatcher and Randy Myers. In , the Reds under new manager Lou Piniella shocked baseball by leading the NL West from wire-to-wire. Winning their first nine games, they started off 33–12 and maintained their lead throughout the year. Led by Chris Sabo, Barry Larkin, Eric Davis, Paul O'Neill and Billy Hatcher in the field, and by José Rijo, Tom Browning and the "Nasty Boys" of Rob Dibble, Norm Charlton and Randy Myers on the mound, the Reds took out the Pirates in the NLCS. The Reds swept the heavily favored Oakland Athletics in four straight, and extended a Reds winning streak in the World Series to nine consecutive games. The World Series, however, saw Eric Davis severely bruise a kidney diving for a fly ball in Game 4, and his play was greatly limited the next year. In winning the World Series the Reds became the only National League team to go wire to wire. In , Quinn was replaced in the front office by Jim Bowden. On the field, manager Lou Piniella wanted outfielder Paul O'Neill to be a power-hitter to fill the void Eric Davis left when he was traded to the Los Angeles Dodgers in exchange for Tim Belcher. However, O'Neill only hit .246 and 14 homers. The Reds returned to winning after a losing season in 1991, but 90 wins was only enough for second place behind the division-winning Atlanta Braves. Before the season ended, Piniella got into an altercation with reliever Rob Dibble. In the off season, Paul O'Neill was traded to the New York Yankees for outfielder Roberto Kelly. Kelly was a disappointment for the Reds over the next couple of years, while O'Neill blossomed, leading a down-trodden Yankees franchise to a return to glory. Also, the Reds would replace their "Big Red Machine" era uniforms in favor of a pinstriped uniform with no sleeves. For the 1993 season Piniella was replaced by fan favorite Tony Pérez, but he lasted only 44 games at the helm, replaced by Davey Johnson. With Johnson steering the team, the Reds made steady progress. In 1994, the Reds were in the newly created National League Central Division with the Chicago Cubs, St. Louis Cardinals, as well as fellow rivals Pittsburgh Pirates and Houston Astros. By the time the strike hit, the Reds finished a half-game ahead of the Astros for first-place in the NL Central. By , the Reds won the division thanks to Most Valuable Player Barry Larkin. After defeating the NL West champion Dodgers in the first NLDS since 1981, they lost to the Atlanta Braves. Team owner Marge Schott announced mid-season that Johnson would be gone by the end of the year, regardless of outcome, to be replaced by former Reds third baseman Ray Knight. Johnson and Schott had never gotten along and she did not approve of Johnson living with his fiancée before they were married, In contrast, Knight, along with his wife, professional golfer Nancy Lopez, were friends of Schott. The team took a dive under Knight and he was unable to complete two full seasons as manager, subject to complaints in the press about his strict managerial style. In the Reds won 96 games, led by manager Jack McKeon, but lost to the New York Mets in a one game playoff. Earlier that year, Schott sold controlling interest in the Reds to Cincinnati businessman Carl Lindner. Despite an 85–77 finish in 2000, and being named 1999 NL manager of the year, McKeon was fired after the 2000 season. The Reds did not have another winning season until 2010. Riverfront Stadium, by then known as Cinergy Field, was demolished in . Great American Ball Park opened in with high expectations for a team led by local favorites, including outfielder Ken Griffey, Jr., shortstop Barry Larkin, and first baseman Sean Casey. Although attendance improved considerably with the new ballpark, the team continued to lose. Schott had not invested much in the farm system since the early 1990s, leaving the team relatively thin on talent. After years of promises that the club was rebuilding toward the opening of the new ballpark, General Manager Jim Bowden and manager Bob Boone were fired on July 28. This broke up the father-son combo of manager Bob Boone and third baseman Aaron Boone, and Aaron was soon traded to the New York Yankees. Tragedy struck in November when Dernell Stenson, a promising young outfielder for the Reds, was shot and killed during a carjack. Following the season Dan O'Brien was hired as the Reds' 16th General Manager. The and seasons continued the trend of big hitting, poor pitching, and poor records. Griffey, Jr. joined the 500 home run club in 2004, but was again hampered by injuries. Adam Dunn emerged as consistent home run hitter, including a home run against José Lima. He also broke the major league record for strikeouts in 2004. Although a number of free agents were signed before 2005, the Reds were quickly in last place and manager Dave Miley was forced out in the 2005 mid season and replaced by Jerry Narron. Like many other small market clubs, the Reds dispatched some of their veteran players and began entrusting their future to a young nucleus that included Adam Dunn and Austin Kearns. Late summer 2004 saw the opening of the Cincinnati Reds Hall of Fame (HOF). The Reds HOF had been in existence in name only since the 1950s, with player plaques, photos and other memorabilia scattered throughout their front offices. Ownership and management desired a stand-alone facility, where the public could walk through inter-active displays, see locker room recreations, watch videos of classic Reds moments and peruse historical items. The first floor houses a movie theater which resembles an older, ivy-covered brick wall ball yard. The hallways contain many vintage photographs. The rear of the building features a three-story wall containing a baseball for every hit Pete Rose had during his career. The third floor contains interactive exhibits including a pitcher's mound, radio booth, and children's area where the fundamentals of baseball are taught through videos featuring former Reds players. Robert Castellini took over as controlling owner from Lindner in 2006. Castellini promptly fired general manager Dan O'Brien and hired Wayne Krivsky. The Reds made a run at the playoffs but ultimately fell short. The 2007 season was again mired in mediocrity. Midway through the season Jerry Narron was fired as manager and replaced by Pete Mackanin. The Reds ended up posting a winning record under Mackanin, but finished the season in 5th place in the Central Division. Mackanin was manager in an interim capacity only, and the Reds, seeking a big name to fill the spot, ultimately brought in Dusty Baker. Early in the 2008 season, Krivsky was fired and replaced by Walt Jocketty. Though the Reds did not win under Krivsky, he is credited with revamping the farm system and signing young talent that could potentially lead the Reds to success in the future. The Reds failed to post winning records in both 2008 and 2009. In 2010, with NL MVP Joey Votto and Gold Glovers Brandon Phillips and Scott Rolen the Reds posted a 91-71 record and were NL Central champions. The following week, the Reds became only the second team in MLB history to be no-hit in a postseason game when Philadelphia's Roy Halladay shut down the National League's number one offense in game one of the NLDS. The Reds lost in a 3-game sweep of the NLDS to Philadelphia. After coming off their surprising 2010 NL Central Division Title, the Reds fell short of many expectations for the 2011 season. Multiple injuries and inconsistent starting pitching played a big role in their mid-season collapse, along with a less productive offense as compared to the previous year. The Reds ended the season at 79-83. The Reds won the 2012 NL Central Division Title. On September 28, Homer Bailey threw a 1-0 no-hitter against the Pittsburgh Pirates at PNC Park, this was the first Reds no-hitter since Tom Browning's perfect game in September of the 1988 season. Finishing with a 97–65 record, they earned the second seed in the Division Series and a match-up with the eventual World Series champion San Francisco Giants. After taking a 2–0 lead with road victories at AT&T Park, they headed home looking to win the series. However, they lost three straight at their home ballpark to become the first National League team since the Cubs in 1984 to lose a division series after leading 2–0. In the off-season, the team traded outfielder Drew Stubbs, as part of a three team deal with the Arizona Diamondbacks and Cleveland Indians, to the Indians, and in turn received right fielder Shin-Soo Choo. On July 2, 2013, Homer Bailey pitched a no-hitter against the San Francisco Giants for a 4-0 Reds victory, making Bailey the third pitcher in Reds history with two complete game no-hitters in their career. Following six consecutive losses to close out the 2013 season, including a loss to the Pittsburgh Pirates, at PNC Park, in the National League wild-card playoff game, the Reds decided to fire Dusty Baker. During his six years as manager, Baker led the Reds to the playoff three times; however, they never advanced beyond the first round. On October 22, 2013, the Reds hired pitching coach Bryan Price to replace Baker as manager. Under Bryan Price, the Reds were led by pitchers Johnny Cueto and the hard-throwing Cuban Aroldis Chapman. While the offense was led by all-star third baseman Todd Frazier, Joey Votto, and Brandon Phillips. Although with plenty of star power, the Reds never got off to a good start and ending the season in lowly fourth place in the division to go along with a 76-86 record. During the offseason, the Reds traded pitchers Alfredo Simón to the Tigers and Mat Latos to the Marlins. In return, they acquired young talents such as Eugenio Suárez and Anthony DeSclafani. They also acquired veteran slugger Marlon Byrd from the Phillies to play left field. The Reds' 2015 season wasn't much better, as they finished with the second worst record in the league with a record of 64-98, their worst finish since 1982. The Reds were forced to trade star pitchers Johnny Cueto (to the Kansas City Royals) and Mike Leake (to the San Francisco Giants), receiving minor league pitching prospects for both. Shortly after the season's end, the Reds traded home run derby champion Todd Frazier to the Chicago White Sox, and closing pitcher Aroldis Chapman to the New York Yankees. In 2016, the Reds broke the then record for home runs allowed during a single season, The Reds held this record until the 2019 season when it was broken by the Baltimore Orioles .The previous record holder was the 1996 Detroit Tigers with 241 longballs yielded to opposing teams. The Reds went 68-94, and again were one of the worst teams in the MLB. The Reds traded outfielder Jay Bruce to the Mets just before the July 31st non-waiver trade deadline in exchange for two prospects, infielder Dilson Herrera and pitcher Max Wotell. During the offseason, the Reds traded Brandon Phillips to the Atlanta Braves in exchange for two minor league pitchers. The Cincinnati Reds play their home games at Great American Ball Park, located at 100 Joe Nuxhall Way, in downtown Cincinnati. Great American Ball Park opened in 2003 at the cost of $290 million and has a capacity of 42,271. Along with serving as the home field for the Reds, the stadium also holds the Cincinnati Reds Hall of Fame. The Hall of Fame was added as a part of Reds tradition allowing fans to walk through the history of the franchise as well as participating in many interactive baseball features. Great American Ball Park is the seventh home of the Cincinnati Reds, built immediately to the north of the site on which Riverfront Stadium, later named Cinergy Field, once stood. The first ballpark the Reds occupied was Bank Street Grounds from 1882 to 1883 until they moved to League Park I in 1884, where they would remain until 1893. Through the late 1890s and early 1900s (decade), the Reds moved to two different parks where they stayed for less than ten years. League Park II was the third home field for the Reds from 1894 to 1901, and then moved to the Palace of the Fans which served as the home of the Reds in the 1910s. It was in 1912 that the Reds moved to Crosley Field which they called home for 58 years. Crosley served as the home field for the Reds for two World Series titles and five National League pennants. Beginning June 30, 1970, and during the dynasty of the Big Red Machine, the Reds played in Riverfront Stadium, appropriately named due to its location right by the Ohio River. Riverfront saw three World Series titles and five National League pennants. It was in the late 1990s that the city agreed to build two separate stadiums on the riverfront for the Reds and the Cincinnati Bengals. Thus, in 2003, the Reds began a new era with the opening of the current stadium. The Reds hold their spring training in Goodyear, Arizona at Goodyear Ballpark. The Reds moved into this stadium and the Cactus League in 2010 after staying in the Grapefruit League for most of their history. The Reds share Goodyear Park with their rivals in Ohio, the Cleveland Indians. Throughout the history of the Cincinnati Reds, many different variations of the classic wishbone "C" logo have been introduced. For most of the history of the Reds, especially during the early history, the Reds logo has been simply the wishbone "C" with the word "REDS" inside, the only colors used being red and white. However, during the 1950s, during the renaming and re-branding of the team as the Cincinnati Redlegs because of the connections to communism of the word 'Reds', the color blue was introduced as part of the Reds color combination. During the 1960s and 1970s the Reds saw a move towards the more traditional colors, abandoning the navy blue. A new logo also appeared with the new era of baseball in 1972, when the team went away from the script "REDS" inside of the "C", instead, putting their mascot Mr. Redlegs in its place as well as putting the name of the team inside of the wishbone "C". In the 1990s the more traditional, early logos of Reds came back with the current logo reflecting more of what the team's logo was when they were first founded. Along with the logo, the Reds' uniforms have been changed many different times throughout their history. Following their departure from being called the "Redlegs" in 1956 the Reds made a groundbreaking change to their uniforms with the use of sleeveless jerseys, seen only once before in the Major Leagues by the Chicago Cubs. At home and away, the cap was all-red with a white wishbone C insignia. The long-sleeved undershirts were red. The uniform was plain white with a red wishbone C logo on the left and the uniform number on the right. On the road the wishbone C was replaced by the mustachioed "Mr. Red" logo, the pillbox-hat-wearing man with a baseball for a head. The home stockings were red with six white stripes. The away stockings had only three white stripes. The Reds changed uniforms again in 1961, when they replaced the traditional wishbone C insignia with an oval C logo, but continued to use the sleeveless jerseys. At home, the Reds wore white caps with the red bill with the oval C in red, white sleeveless jerseys with red pinstripes, with the oval C-REDS logo in black with red lettering on the left breast and the number in red on the right. The gray away uniform included a gray cap with the red oval C and a red bill. Their gray away uniforms, which also included a sleeveless jersey, bore CINCINNATI in an arched block style across with the number below on the left. In 1964, players' last names were placed on the back of each set of uniforms, below the numbers. Those uniforms were scrapped after the 1966 season. However, the Cincinnati uniform design most familiar to baseball enthusiasts is the one whose basic form, with minor variations, held sway for the 26 seasons from 1967 to 1992. Most significantly, the point was restored to the C insignia, making it a wishbone again. During this era, the Reds wore all-red caps both at home and on the road. The caps bore the simple wishbone C insignia in white. The uniforms were standard short-sleeved jerseys and standard trousers—white at home and grey on the road. The home uniform featured the Wishbone C-REDS logo in red with white type on the left breast and the uniform number in red on the right. The away uniform bore CINCINNATI in an arched block style across the front with the uniform number below on the left. Red, long-sleeved undershirts and plain red stirrups over white sanitary stockings completed the basic design. The Reds wore pinstriped home uniforms in 1967 only, and the uniforms were flannel through 1971, changing to double-knits with pullover jerseys and beltless pants in 1972. Those uniforms lasted 21 seasons, and the 1992 Reds were the last MLB team to date whose primary uniforms featured pullover jerseys and beltless pants. The 1993 uniforms (which did away with the pullovers and brought back button-down jerseys) kept white and gray as the base colors for the home and away uniforms, but added red pinstripes. The home jerseys were sleeveless, showing more of the red undershirts. The color scheme of the C-REDS logo on the home uniform was reversed, now red lettering on a white background. A new home cap was created that had a red bill and a white crown with red pinstripes and a red wishbone C insignia. The away uniform kept the all-red cap, but moved the uniform number to the left, to more closely match the home uniform. The only additional change to these uniforms was the introduction of black as a primary color of the Reds in 1999, especially on their road uniforms. The Reds latest uniform change came in December 2006 which differed significantly from the uniforms worn during the previous eight seasons. The home caps returned to an all-red design with a white wishbone C, lightly outlined in black. Caps with red crowns and a black bill became the new road caps. Additionally, the sleeveless jersey was abandoned for a more traditional design. The numbers and lettering for the names on the backs of the jerseys were changed to an early-1900s style typeface, and a handlebar mustached "Mr. Redlegs" – reminiscent of the logo used by the Reds in the 1950s and 1960s – was placed on the left sleeve. The Cincinnati Reds have retired ten numbers in franchise history, as well as honoring Jackie Robinson, whose number is retired league-wide around Major League Baseball. All of the retired numbers are located at Great American Ball Park behind home-plate on the outside of the press box. Along with the retired player and manager number, the following broadcasters are honored with microphones by the broadcast booth: Marty Brennaman, Waite Hoyt, and Joe Nuxhall. On April 15, 1997, #42 was retired throughout Major League Baseball in honor of Jackie Robinson. The Reds have hosted the Major League Baseball All-Star Game five times: twice at Crosley Field (1938, 1953), twice at Riverfront Stadium (1970, 1988), and once at Great American Ball Park (2015). (Until 2019, the Reds shared this record with the Cleveland Indians and the Pittsburgh Pirates; Cleveland broke this tie in 2019, hosting the All-Star Game for the sixth time.) The Ohio Cup was an annual pre-season baseball game, which pitted the Ohio rivals Cleveland Indians and Cincinnati Reds. In its first series it was a single-game cup, played each year at minor-league Cooper Stadium in Columbus, was staged just days before the start of each new Major League Baseball season. A total of eight Ohio Cup games were played, in 1989 to 1996, with the Indians winning six of them. The winner of the game each year was awarded the Ohio Cup in postgame ceremonies. The Ohio Cup was a favorite among baseball fans in Columbus, with attendances regularly topping 15,000. The Ohio Cup games ended with the introduction of regular-season interleague play in 1997. Thereafter, the two teams competed annually in the regular-season Battle of Ohio or Buckeye Series. The Ohio Cup was revived in 2008 as a reward for the team with the better overall record in the Reds-Indians series each year. The Reds' flagship radio station has been WLW, 700AM since 1969. Prior to that, the Reds were heard over: WKRC, WCPO, WSAI and WCKY. WLW, a 50,000-watt station, is "clear channel" in more than one way, as iHeartMedia owns the "blowtorch" outlet which is also known as "The Nation's Station". Marty Brennaman has been the Reds' play-by-play voice since 1974 and has won the Ford C. Frick Award for his work, which includes his famous call of "... and this one belongs to the Reds!" after a win. Joining him for years on color was former Reds pitcher Joe Nuxhall, who worked in the radio booth from 1967 (the year after his retirement as an active player) until 2004, plus three more seasons doing select home games until his death, in 2007. In 2007, Thom Brennaman, a veteran announcer seen nationwide on Fox Sports, joined his father Marty in the radio booth. Retired relief pitcher Jeff Brantley, formerly of ESPN, also joined the network in 2007. , Brantley and Thom Brennaman's increased TV schedule (see below) has led to more appearances for Jim Kelch, who had filled in on the network since 2008. In 2019, the Reds hired then-Pensacola Blue Wahoos radio play-by-play announcer Tommy Thrall to provide in-game and post-game coverage as well as a fill in play-by-play announcer. Televised games are seen exclusively on Fox Sports Ohio and Fox Sports Indiana. In addition, Fox Sports South televises Fox Sports Ohio broadcasts of Reds games to Tennessee and western North Carolina. George Grande, who hosted the first "SportsCenter" on ESPN in 1979, was the play-by-play announcer, usually alongside Chris Welsh, from 1993 until his retirement during the final game of the 2009 season. Since 2009, Grande has worked part-time for the Reds as play-by-play announcer in September when Thom Brennaman is covering the NFL for Fox Sports. He has also made guest appearances throughout the season. Brennaman has been the head play-by-play commentator since 2010, with Welsh and Brantley sharing time as the color commentators. Paul Keels, who left in 2011 to become the play-by-play announcer for the Ohio State Buckeyes Radio Network, was the Reds' backup play-by-play television announcer during the 2010 season. Jim Kelch served as Keels' replacement. The Reds also added former Cincinnati First Baseman Sean Casey – known as "The Mayor" by Reds fans – to do color commentary for approximately 15 games in 2011. NBC affiliate WLWT carried Reds games from 1948 to 1995. Among those that have called games for WLWT include Waite Hoyt, Ray Lane, Steve Physioc, Johnny Bench, Joe Morgan, and Ken Wilson. Al Michaels, who established a long career with ABC and NBC, spent three years in Cincinnati early in his career. The last regularly-scheduled, over-the-air broadcasts of Reds games were on WSTR-TV from 1996 to 1998. Since 2010, WKRC-TV has simulcast Opening Day games with Fox Sports Ohio. The Cincinnati Reds farm system consists of eight minor league affiliates.
https://en.wikipedia.org/wiki?curid=6671
Caribbean cuisine Caribbean cuisine is a fusion of African, Creole, Cajun, Amerindian, European, Latin American, Indian/South Asian, Middle Eastern, and Chinese. These traditions were brought from many different countries when they came to the Caribbean. In addition, the population has created styles that are unique to the region. Ingredients that are common in most islands' dishes are rice, plantains, beans, cassava, culantro, bell peppers, chickpeas, tomatoes, sweet potatoes, coconut, and any of various meats that are locally available like beef, poultry, pork or fish. A characteristic seasoning for the region is a green herb-and-oil-based marinade which imparts a flavor profile which is quintessentially Caribbean in character. Ingredients may include garlic, onions, scotch bonnet peppers, celery, green onions, and herbs like culantro, Mexican mint, chives, marjoram, rosemary, tarragon and thyme. This green seasoning is used for a variety of dishes like curries, stews and roasted meats. Traditional dishes are so important to regional culture that, for example, the local version of Caribbean goat stew has been chosen as the official national dish of Montserrat and is also one of the signature dishes of St. Kitts and Nevis. Another popular dish in the Anglophone Caribbean is called "cook-up", or pelau. Ackee and saltfish is another popular dish that is unique to Jamaica. Callaloo is a dish containing leafy vegetables such as spinach and sometimes okra amongst others, widely distributed in the Caribbean, with a distinctively mixed African and indigenous character. The variety of dessert dishes in the area also reflects the mixed origins of the recipes. In some areas, black cake, a derivative of English Christmas pudding, may be served, especially on special occasions. Over time, food from the Caribbean has evolved into a narrative technique through which their culture has been accentuated and promoted. However, by studying Caribbean culture through a literary lens there then runs the risk of generalizing exoticist ideas about food practices from the tropical. Some food theorists argue that this depiction of Caribbean food in various forms of media contributes to the inaccurate conceptions revolving around their culinary practices, which are much more grounded in unpleasant historical events. Therefore, it can be argued that the connection between the idea of the Caribbean being the ultimate paradise and Caribbean food being exotic is based on inaccurate information.
https://en.wikipedia.org/wiki?curid=6672
Central Powers The Central Powers, also Central Empires (; ; / ; ), consisting of Germany, , the Ottoman Empire and Bulgaria – hence also known as the Quadruple Alliance ()—was one of the two main coalitions that fought World War I (1914–18). It faced and was defeated by the Allied Powers that had formed around the Triple Entente. The Powers' origin was the alliance of Germany and Austria-Hungary in 1879. Despite having nominally joined the Triple Alliance before, Italy did not take part in World War I on the side of the Central Powers; the Ottoman Empire and Bulgaria did not join until after World War I had begun, even though the Ottoman Empire had retained close relations with both Germany and Austria-Hungary since the beginning of the 20th century. The Central Powers consisted of the German Empire and the Austro-Hungarian Empire at the beginning of the war. The Ottoman Empire joined the Central Powers later in 1914. In 1915, the Kingdom of Bulgaria joined the alliance. The name "Central Powers" is derived from the location of these countries; all four (including the other groups that supported them except for Finland and Lithuania) were located between the Russian Empire in the east and France and the United Kingdom in the west. Finland, Azerbaijan, and Lithuania joined them in 1918 before the war ended and after the Russian Empire collapsed The Central Powers were composed of the following nations: In early July 1914, in the aftermath of the assassination of Austro-Hungarian Archduke Franz Ferdinand and the immediate likelihood of war between Austria-Hungary and Serbia, Kaiser Wilhelm II and the German government informed the Austro-Hungarian government that Germany would uphold its alliance with Austria-Hungary and defend it from possible Russian intervention if a war between Austria-Hungary and Serbia took place. When Russia enacted a general mobilization, Germany viewed the act as provocative. The Russian government promised Germany that its general mobilization did not mean preparation for war with Germany but was a reaction to the events between Austria-Hungary and Serbia. The German government regarded the Russian promise of no war with Germany to be nonsense in light of its general mobilization, and Germany, in turn, mobilized for war. On 1 August, Germany sent an ultimatum to Russia stating that since both Germany and Russia were in a state of military mobilization, an effective state of war existed between the two countries. Later that day, France, an ally of Russia, declared a state of general mobilization. In August 1914, Germany waged war on Russia, the German government justified military action against Russia as necessary because of Russian aggression as demonstrated by the mobilization of the Russian army that had resulted in Germany mobilizing in response. After Germany declared war on Russia, France with its alliance with Russia prepared a general mobilization in expectation of war. On 3 August 1914, Germany responded to this action by declaring war on France. Germany, facing a two-front war, enacted what was known as the Schlieffen Plan, that involved German armed forces needing to move through Belgium and swing south into France and towards the French capital of Paris. This plan was hoped to quickly gain victory against the French and allow German forces to concentrate on the Eastern Front. Belgium was a neutral country and would not accept German forces crossing its territory. Germany disregarded Belgian neutrality and invaded the country to launch an offensive towards Paris. This caused Great Britain to declare war against the German Empire, as the action violated the Treaty of London that both nations signed in 1839 guaranteeing Belgian neutrality and defense of the kingdom if a nation reneged. Subsequently, several states declared war on Germany in late August 1914, with Italy declaring war on Austria-Hungary in 1915 and Germany on 27 August 1916, the United States declaring war on Germany on 6 April 1917 and Greece declaring war on Germany in July 1917. Upon its founding in 1871, the German Empire controlled Alsace-Lorraine as an "imperial territory" incorporated from France after the Franco-Prussian War. It was held as part of Germany's sovereign territory. Germany held multiple African colonies at the time of World War I. All of Germany's African colonies were invaded and occupied by Allied forces during the war. Kamerun, German East Africa, and German Southwest Africa were German colonies in Africa. Togoland was a German protectorate in Africa. The Kiautschou Bay concession was a German dependency in East Asia leased from China in 1898. It was occupied by Japanese forces following the Siege of Tsingtao. German New Guinea was a German protectorate in the Pacific. It was occupied by Australian forces in 1914. German Samoa was a German protectorate following the Tripartite Convention. It was occupied by the New Zealand Expeditionary Force in 1914. Austria-Hungary regarded the assassination of Archduke Franz Ferdinand as being orchestrated with the assistance of Serbia. The country viewed the assassination as setting a dangerous precedent of encouraging the country's South Slav population to rebel and threaten to tear apart the multinational country. Austria-Hungary formally sent an ultimatum to Serbia demanding a full-scale investigation of Serbian government complicity in the assassination, and complete compliance by Serbia in agreeing to the terms demanded by Austria-Hungary. Serbia submitted to accept most of the demands, however Austria-Hungary viewed this as insufficient and used this lack of full compliance to justify military intervention. These demands have been viewed as a diplomatic cover for what was going to be an inevitable Austro-Hungarian declaration of war on Serbia. Austria-Hungary had been warned by Russia that the Russian government would not tolerate Austria-Hungary invading Serbia. However, with Germany supporting Austria-Hungary's actions, the Austro-Hungarian government hoped that Russia would not intervene and that the conflict with Serbia would be a regional conflict. Austria-Hungary's invasion of Serbia resulted in Russia declaring war on the country and Germany in turn declared war on Russia, setting off the beginning of the clash of alliances that resulted in the World War. Austria-Hungary was internally divided into two states with their own governments, joined in communion through the Habsburg throne. Austrian Cisleithania contained various duchies and principalities but also the Kingdom of Bohemia, the Kingdom of Dalmatia, the Kingdom of Galicia and Lodomeria. Hungarian Transleithania comprised the Kingdom of Hungary and the Kingdom of Croatia-Slavonia. In Bosnia and Herzegovina sovereign authority was shared by both Austria and Hungary. The Ottoman Empire joined the war on the side of the Central Powers in November 1914. The Ottoman Empire had gained strong economic connections with Germany through the Berlin-to-Baghdad railway project that was still incomplete at the time. The Ottoman Empire made a formal alliance with Germany signed on 2 August 1914. The alliance treaty expected that the Ottoman Empire would become involved in the conflict in a short amount of time. However, for the first several months of the war the Ottoman Empire maintained neutrality though it allowed a German naval squadron to enter and stay near the strait of Bosphorus. Ottoman officials informed the German government that the country needed time to prepare for conflict. Germany provided financial aid and weapons shipments to the Ottoman Empire. After pressure escalated from the German government demanding that the Ottoman Empire fulfill its treaty obligations, or else Germany would expel the country from the alliance and terminate economic and military assistance, the Ottoman government entered the war with the recently acquired cruisers from Germany, the "Yavuz Sultan Selim" (formerly "SMS Goeben") and the "Midilli" (formerly "SMS Breslau") launching a naval raid on the Russian port of Odessa, thus engaging in a military action in accordance with its alliance obligations with Germany. Russia and the Triple Entente declared war on the Ottoman Empire. Bulgaria was still resentful after its defeat in July 1913 at the hands of Serbia, Greece and Romania. It signed a treaty of defensive alliance with the Ottoman Empire on 19 August 1914. It was the last country to join the Central Powers, which Bulgaria did in October 1915 by declaring war on Serbia. It invaded Serbia in conjunction with German and Austro-Hungarian forces. Bulgaria held claims on the region of Vardar Macedonia then held by Serbia following the Balkan Wars of 1912–1913 and (from the Bulgarian point of view), the costly Treaty of Bucharest (1913). As a condition of entering WW1 on the side of the Central Powers, Bulgaria was granted the right to reclaim that territory. In opposition to offensive operations by Union of South Africa, which had joined the war, Boer army officers of what is now known as the Maritz Rebellion "refounded" the South African Republic in September 1914. Germany assisted the rebels, some rebels operating in and out of the German colony of German South-West Africa. The rebels were all defeated or captured by South African government forces by 4 February 1915. The Dervish movement was a rebel Somali movement which had existed since before World War I. It was led by Mohammed Abdullah Hassan, who was seeking the independence of Somali territories. The Dervish movement was supported by the Ottoman Empire and Germany, and also briefly by the Ethiopian Empire from 1915–1916. Dervish forces fought against Italian and British forces in Italian Somaliland and British Somaliland during the Somaliland Campaign. The Senussi Order was a Muslim political-religious tariqa (Sufi order) and clan in Libya, previously under Ottoman control, which had been lost to Italy in 1912. In 1915, they were courted by the Ottoman Empire and Germany and Grand Senussi Ahmed Sharif as-Senussi declared jihad and attacked the Italians in Libya and British controlled Egypt in the so called Senussi Campaign. In 1915 the Sultanate of Darfur renounced allegiance to the Sudan government and aligned with the Ottomans. The Anglo-Egyptian Darfur Expedition pre-emptively in March 1916 to prevent an attack on Sudan and took control of the Sultanate by November 1916. On 11 June 1916, while pursuing the Austro-Hungarian Army in Bukovina during the Brusilov Offensive, Russian forces inadvertently crossed into Romanian territory, where they overwhelmed the border guard at Mamornița and had a cavalry patrol disarmed and interned at Herța. Having no intention to force the hand of the Romanian Government, the Russians quickly left Romanian territory. The Romanian reaction to the Russian incursion left the Austro-Hungarian Minister to Romania, Count Ottokar Czernin, "fully satisfied". During 1917 and 1918, the Finns under Carl Gustaf Emil Mannerheim and Lithuanian nationalists fought Russia for a common cause. With the Bolshevik attack of late 1917, the General Secretariat of Ukraine sought military protection first from the Central Powers and later from the armed forces of the Entente. The Ottoman Empire also had its own allies in Azerbaijan and the Northern Caucasus. The three nations fought alongside each other under the Army of Islam in the Battle of Baku. The Belarusian People's Republic was a client state of Germany created in 1918. The Duchy of Courland and Semigallia was a client state of Germany created in 1918. The Crimean Regional Government was a client state of Germany created in 1918. The Don Republic proclaimed by Don Cossacks on May 18, 1918 fought against the Bolsheviks; it was closely associated with the German Empire Finland existed as an autonomous Grand Duchy of Russia since 1809 and collapse of Russia in 1917 gave it independence. Following the end of the Finnish Civil War, in which Germany supported the "White" against the Soviet-backed labour movement, in May 1918 there were moves to create a Kingdom of Finland. A German prince was elected but the Armistice intervened. The Kuban People's Republic was a client state of Germany created in 1918. The Kingdom of Lithuania was a client state of Germany created in 1918. The Mountainous Republic of the Northern Caucasus was associated with the Central Powers. The Democratic Republic of Georgia declared independence in 1918 which then led to border conflicts between newly formed republic and Ottoman Empire. Soon after Ottoman Empire invaded the republic and quickly reached Borjomi. This forced Georgia to ask for help from Germany which they were granted. Germany forced the Ottomans to withdraw from Georgian territories and recognize Georgian sovereignty. Germany, Georgia and the Ottomans signed a peace treaty, the Treaty of Batum which ended the conflict with the last two. In return Georgia become a German "ally". This time period of Georgian-German friendship was known as German Caucasus expedition. The Kingdom of Poland was a client state of Germany created in 1916. This government was recognized by the emperors of Germany and Austria-Hungary in November 1916, and it adopted a constitution in 1917. The decision to create a Polish State was taken by Germany in order to attempt to legitimize its military occupation amongst the Polish inhabitants, following upon German propaganda sent to Polish inhabitants in 1915 that German soldiers were arriving as liberators to free Poland from subjugation by Russia. The state was utilized by the German government alongside punitive threats to induce Polish landowners living in the German-occupied Baltic territories to move to the state and sell their Baltic property to Germans in exchange for moving to Poland, and efforts were made to induce similar emigration of Poles from Prussia to the state. The Ukrainian State was a client state of Germany led by Hetman Pavlo Skoropadskyi, who overthrew the government of the Ukrainian People's Republic. The United Baltic Duchy was a German protectorate proposed by Germans in the former Courland, Livonian, and Estonian governorates. It briefly existed in 1918. In 1918, the Azerbaijan Democratic Republic, facing Bolshevik revolution and opposition from the Muslim Musavat Party, was then occupied by the Ottoman Empire, which expelled the Bolsheviks while supporting the Musavat Party. The Ottoman Empire maintained a presence in Azerbaijan until the end of the war in November 1918. Jabal Shammar was an Arab state in the Middle East that was closely associated with the Ottoman Empire. States listed in this section were not officially members of the Central Powers, but at some point during the war engaged in cooperation with one or more Central Powers members on a level that makes their neutrality disputable. The Ethiopian Empire was officially neutral throughout World War I but widely suspected of sympathy for the Central Powers between 1915 and 1916. At the time, Ethiopia was one of the few independent states in Africa and a major power in the Horn of Africa. Its ruler, Lij Iyasu, was widely suspected of harbouring pro-Islamic sentiments and being sympathetic to the Ottoman Empire. The German Empire also attempted to reach out to Iyasu, dispatching several unsuccessful expeditions to the region to attempt to encourage it to collaborate in an Arab Revolt-style uprising in East Africa. One of the unsuccessful expeditions was led by Leo Frobenius, a celebrated ethnographer and personal friend of Kaiser Wilhelm II. Under Iyasu's directions, Ethiopia probably supplied weapons to the Muslim Dervish rebels during the Somaliland Campaign of 1915 to 1916, indirectly helping the Central Powers' cause. Fearing the rising influence of Iyasu and the Ottoman Empire, the Christian nobles of Ethiopia conspired against Iyasu over 1915. Iyasu was first excommunicated by the Ethiopian Orthodox Patriarch and eventually deposed in a coup d'état on 27 September 1916. A less pro-Ottoman regent, "Ras" Tafari Makonnen, was installed on the throne. Other movements supported the efforts of the Central Powers for their own reasons, such as the radical Irish Nationalists who launched the Easter Rising in Dublin in April 1916; they referred to their "gallant allies in Europe". However, the majority of Irish Nationalists supported the British and allied war effort up until 1916 when the Irish political landscape was changing. In 1914, Józef Piłsudski was permitted by Germany and Austria-Hungary to form independent Polish legions. Piłsudski wanted his legions to help the Central Powers defeat Russia and then side with France and the UK and win the war with them. Bulgaria signed an armistice with the Allies on 29 September 1918, following a successful Allied advance in Macedonia. The Ottoman Empire followed suit on 30 October 1918 in the face of British and Arab gains in Palestine and Syria. Austria and Hungary concluded ceasefires separately during the first week of November following the disintegration of the Habsburg Empire and the Italian offensive at Vittorio Veneto; Germany signed the armistice ending the war on the morning of 11 November 1918 after the Hundred Days Offensive, and a succession of advances by New Zealand, Australian, Canadian, Belgian, British, French and US forces in north-eastern France and Belgium. There was no unified treaty ending the war; the Central Powers were dealt with in separate treaties.
https://en.wikipedia.org/wiki?curid=6673
Conservatism Conservatism is a political and social philosophy promoting traditional social institutions in the context of culture and civilization. The central tenets of conservatism include tradition, organic society, hierarchy, authority, and property rights. Conservatives seek to preserve a range of institutions such as religion, parliamentary government, and property rights, with the aim of emphasizing social stability and continuity. The more traditional elements—reactionaries—oppose modernism and seek a return to "the way things were". The first established use of the term in a political context originated in 1818 with François-René de Chateaubriand during the period of Bourbon Restoration that sought to roll back the policies of the French Revolution. Historically associated with right-wing politics, the term has since been used to describe a wide range of views. There is no single set of policies regarded as conservative because the meaning of conservatism depends on what is considered traditional in a given place and time. Thus conservatives from different parts of the world—each upholding their respective traditions—may disagree on a wide range of issues. Edmund Burke, an 18th-century politician who opposed the French Revolution, but supported the American Revolution, is credited as one of the main theorists of conservatism in Great Britain in the 1790s. According to Quintin Hogg, the chairman of the British Conservative Party in 1959: "Conservatism is not so much a philosophy as an attitude, a constant force, performing a timeless function in the development of a free society, and corresponding to a deep and permanent requirement of human nature itself". Liberal conservatism incorporates the classical liberal view of minimal government intervention in the economy. Individuals should be free to participate in the market and generate wealth without government interference. However, individuals cannot be thoroughly depended on to act responsibly in other spheres of life, therefore liberal conservatives believe that a strong state is necessary to ensure law and order and social institutions are needed to nurture a sense of duty and responsibility to the nation. Liberal conservatism is a variant of conservatism that is strongly influenced by liberal stances. As these latter two terms have had different meanings over time and across countries, liberal conservatism also has a wide variety of meanings. Historically, the term often referred to the combination of economic liberalism, which champions "laissez-faire" markets, with the classical conservatism concern for established tradition, respect for authority and religious values. It contrasted itself with classical liberalism, which supported freedom for the individual in both the economic and social spheres. Over time, the general conservative ideology in many countries adopted economic liberal arguments and the term liberal conservatism was replaced with conservatism. This is also the case in countries where liberal economic ideas have been the tradition such as the United States and are thus considered conservative. In other countries where liberal conservative movements have entered the political mainstream, such as Italy and Spain, the terms liberal and conservative may be synonymous. The liberal conservative tradition in the United States combines the economic individualism of the classical liberals with a Burkean form of conservatism (which has also become part of the American conservative tradition, such as in the writings of Russell Kirk). A secondary meaning for the term liberal conservatism that has developed in Europe is a combination of more modern conservative (less traditionalist) views with those of social liberalism. This has developed as an opposition to the more collectivist views of socialism. Often this involves stressing what are now conservative views of free market economics and belief in individual responsibility, with social liberal views on defence of civil rights, environmentalism and support for a limited welfare state. In continental Europe, this is sometimes also translated into English as social conservatism. Conservative liberalism is a variant of liberalism that combines liberal values and policies with conservative stances. The roots of conservative liberalism are found at the beginning of the history of liberalism. Until the two World Wars, in most European countries the political class was formed by conservative liberals, from Germany to Italy. Events after World War I brought the more radical version of classical liberalism to a more conservative (i.e. more moderate) type of liberalism. Libertarian conservatism describes certain political ideologies most prominently within the United States which combine libertarian economic issues with aspects of conservatism. Its four main branches are constitutionalism, paleolibertarianism, small government conservatism and Christian libertarianism. They generally differ from paleoconservatives, in that they favor more personal and economic freedom. Agorists such as Samuel Edward Konkin III labeled libertarian conservatism right-libertarianism. In contrast to paleoconservatives, libertarian conservatives support strict "laissez-faire" policies such as free trade, opposition to any national bank and opposition to business regulations. They are vehemently opposed to environmental regulations, corporate welfare, subsidies and other areas of economic intervention. Many conservatives, especially in the United States, believe that the government should not play a major role in regulating business and managing the economy. They typically oppose efforts to charge high tax rates and to redistribute income to assist the poor. Such efforts, they argue, do not properly reward people who have earned their money through hard work. Fiscal conservatism is the economic philosophy of prudence in government spending and debt. In his "Reflections on the Revolution in France", Edmund Burke argued that a government does not have the right to run up large debts and then throw the burden on the taxpayer: [I]t is to the property of the citizen, and not to the demands of the creditor of the state, that the first and original faith of civil society is pledged. The claim of the citizen is prior in time, paramount in title, superior in equity. The fortunes of individuals, whether possessed by acquisition or by descent or in virtue of a participation in the goods of some community, were no part of the creditor's security, expressed or implied...[T]he public, whether represented by a monarch or by a senate, can pledge nothing but the public estate; and it can have no public estate except in what it derives from a just and proportioned imposition upon the citizens at large. National conservatism is a political term used primarily in Europe to describe a variant of conservatism which concentrates more on national interests than standard conservatism as well as upholding cultural and ethnic identity, while not being outspokenly nationalist or supporting a far-right approach. In Europe, national conservatives are usually eurosceptics. National conservatism is heavily oriented towards the traditional family and social stability as well as in favour of limiting immigration. As such, national conservatives can be distinguished from economic conservatives, for whom free market economic policies, deregulation and fiscal conservatism are the main priorities. Some commentators have identified a growing gap between national and economic conservatism: "[M]ost parties of the Right [today] are run by economic conservatives who, in varying degrees, have marginalized social, cultural, and national conservatives". National conservatism is also related to traditionalist conservatism. Traditionalist conservatism is a political philosophy emphasizing the need for the principles of natural law and transcendent moral order, tradition, hierarchy and organic unity, agrarianism, classicism and high culture as well as the intersecting spheres of loyalty. Some traditionalists have embraced the labels "reactionary" and "counterrevolutionary", defying the stigma that has attached to these terms since the Enlightenment. Having a hierarchical view of society, many traditionalist conservatives, including a few Americans, defend the monarchical political structure as the most natural and beneficial social arrangement. Cultural conservatives support the preservation of the heritage of one nation, or of a shared culture that is not defined by national boundaries. The shared culture may be as divergent as Western culture or Chinese culture. In the United States, the term "cultural conservative" may imply a conservative position in the culture war. Cultural conservatives hold fast to traditional ways of thinking even in the face of monumental change. They believe strongly in traditional values and traditional politics and often have an urgent sense of nationalism. Social conservatism is distinct from cultural conservatism, although there are some overlaps. Social conservatives may believe that society is built upon a fragile network of relationships which need to be upheld through duty, traditional values and established institutions; and that the government has a role in encouraging or enforcing traditional values or behaviours. A social conservative wants to preserve traditional morality and social mores, often by opposing what they consider radical policies or social engineering. Social change is generally regarded as suspect. A second meaning of the term social conservatism developed in the Nordic countries and continental Europe, where it refers to liberal conservatives supporting modern European welfare states. Social conservatives (in the first meaning of the phrase) in many countries generally favour the anti-abortion position in the abortion controversy and oppose human embryonic stem cell research (particularly if publicly funded); oppose both eugenics and human enhancement (transhumanism) while supporting bioconservatism; support a traditional definition of marriage as being one man and one woman; view the nuclear family model as society's foundational unit; oppose expansion of civil marriage and child adoption to couples in same-sex relationships; promote public morality and traditional family values; oppose atheism, especially militant atheism, secularism and the separation of church and state; support the prohibition of drugs, prostitution and euthanasia; and support the censorship of pornography and what they consider to be obscenity or indecency. Most conservatives in the United States support the death penalty. In contrast to the tradition-based definition of conservatism, some political theorists such as Corey Robin define conservatism primarily in terms of a general defense of social and economic inequality. From this perspective, conservatism is less an attempt to uphold traditional institutions and more "a meditation on—and theoretical rendition of—the felt experience of having power, seeing it threatened, and trying to win it back". Conversely, some conservatives may argue that they are seeking less to protect their own power than they are seeking to protect "inalienable rights" and promote norms and rules that they believe should stand timeless and eternal, applying to each citizen. Religious conservatism principally apply the teachings of particular religions to politics, sometimes by merely proclaiming the value of those teachings, at other times by having those teachings influence laws. In most democracies, political conservatism seeks to uphold traditional family structures and social values. Religious conservatives typically oppose abortion, LGBT behavior (or, in certain cases, identity), drug use, and sexual activity outside of marriage. In some cases, conservative values are grounded in religious beliefs, and conservatives seek to increase the role of religion in public life. Paternalistic conservatism is a strand in conservatism which reflects the belief that societies exist and develop organically and that members within them have obligations towards each other. There is particular emphasis on the paternalistic obligation of those who are privileged and wealthy to the poorer parts of society. Since it is consistent with principles such as organicism, hierarchy and duty, it can be seen as an outgrowth of traditional conservatism. Paternal conservatives support neither the individual nor the state in principle, but are instead prepared to support either or recommend a balance between the two depending on what is most practical. It stresses the importance of a social safety net to deal with poverty, support of limited redistribution of wealth along with government regulation of markets in the interests of both consumers and producers. Paternalistic conservatism first arose as a distinct ideology in the United Kingdom under Prime Minister Benjamin Disraeli's "One Nation" Toryism. There have been a variety of one nation conservative governments. In the United Kingdom, the Prime Ministers Disraeli, Stanley Baldwin, Neville Chamberlain, Winston Churchill, Harold Macmillan and Boris Johnson were or are one nation conservatives. In Germany, during the 19th-century German Chancellor Otto von Bismarck adopted policies of state-organized compulsory insurance for workers against sickness, accident, incapacity and old age. Chancellor Leo von Caprivi promoted a conservative agenda called the "New Course". In the United States, the administration of President William Howard Taft was a progressive conservative and he described himself as "a believer in progressive conservatism" and President Dwight D. Eisenhower declared himself an advocate of "progressive conservatism". In Canada, a variety of conservative governments have been part of the Red tory tradition, with Canada's former major conservative party being named the Progressive Conservative Party of Canada from 1942 to 2003. In Canada, the Prime Ministers Arthur Meighen, R. B. Bennett, John Diefenbaker, Joe Clark, Brian Mulroney, and Kim Campbell led Red tory federal governments. Authoritarian conservatism or reactionary conservatism refers to autocratic regimes that center their ideology around conservative nationalism, rather than ethnic nationalism, though certain racial components such as antisemitism may exist. Authoritarian conservative movements show strong devotion towards religion, tradition and culture while also expressing fervent nationalism akin to other far-right nationalist movements. Examples of authoritarian conservative leaders include António de Oliveira Salazar and Engelbert Dollfuss. Authoritarian conservative movements were prominent in the same era as fascism, with which it sometimes clashed. Although both ideologies shared core values such as nationalism and had common enemies such as communism and materialism, there was nonetheless a contrast between the traditionalist nature of authoritarian conservatism and the revolutionary, palingenetic and populist nature of fascism—thus it was common for authoritarian conservative regimes to suppress rising fascist and National Socialist movements. The hostility between the two ideologies is highlighted by the struggle for power for the National Socialists in Austria, which was marked by the assassination of Engelbert Dollfuss. Sociologist Seymour Martin Lipset has examined the class basis of right-wing extremist politics in the 1920–1960 era. He reports: Conservative or rightist extremist movements have arisen at different periods in modern history, ranging from the Horthyites in Hungary, the Christian Social Party of Dollfuss in Austria, "Der Stahlhelm" and other nationalists in pre-Hitler Germany, and Salazar in Portugal, to the pre-1966 Gaullist movements and the monarchists in contemporary France and Italy. The right extremists are conservative, not revolutionary. They seek to change political institutions in order to preserve or restore cultural and economic ones, while extremists of the centre and left seek to use political means for cultural and social revolution. The ideal of the right extremist is not a totalitarian ruler, but a monarch, or a traditionalist who acts like one. Many such movements in Spain, Austria, Hungary, Germany, and Italy-have been explicitly monarchist... The supporters of these movements differ from those of the centrists, tending to be wealthier, and more religious, which is more important in terms of a potential for mass support. In Great Britain, conservative ideas (though not yet called that) emerged in the Tory movement during the Restoration period (1660–1688). Toryism supported a hierarchical society with a monarch who ruled by divine right. Tories opposed the idea that sovereignty derived from the people and rejected the authority of parliament and freedom of religion. Robert Filmer's "Patriarcha: or the Natural Power of Kings" (published posthumously in 1680, but written before the English Civil War of 1642–1651) became accepted as the statement of their doctrine. However, the Glorious Revolution of 1688 destroyed this principle to some degree by establishing a constitutional government in England, leading to the hegemony of the Tory-opposed Whig ideology. Faced with defeat, the Tories reformed their movement, now holding that sovereignty was vested in the three estates of Crown, Lords and Commons rather than solely in the Crown. Toryism became marginalized during the long period of Whig ascendancy in the 18th century. Conservatives typically see Richard Hooker (1554–1600) as the founding father of conservatism, along with the Marquess of Halifax (1633–1695), David Hume (1711–1776) and Edmund Burke (1729–1797). Halifax promoted pragmatism in government whilst Hume argued against political rationalism and utopianism. Burke served as the private secretary to the Marquis of Rockingham and as official pamphleteer to the Rockingham branch of the Whig party. Together with the Tories, they were the conservatives in the late 18th century United Kingdom. Burke's views were a mixture of liberal and conservative. He supported the American Revolution of 1765–1783, but abhorred the violence of the French Revolution (1789–1799). He accepted the liberal ideals of private property and the economics of Adam Smith (1723–1790), but thought that economics should remain subordinate to the conservative social ethic, that capitalism should be subordinate to the medieval social tradition and that the business class should be subordinate to aristocracy. He insisted on standards of honor derived from the medieval aristocratic tradition and saw the aristocracy as the nation's natural leaders. That meant limits on the powers of the Crown, since he found the institutions of Parliament to be better informed than commissions appointed by the executive. He favored an established church, but allowed for a degree of religious toleration. Burke justified the social order on the basis of tradition: tradition represented the wisdom of the species and he valued community and social harmony over social reforms. Burke was a leading theorist in his day, finding extreme idealism (either Tory or Whig) an endangerment to broader liberties and (like Hume) rejecting abstract reason as an unsound guide for political theory. Despite their influence on future conservative thought, none of these early contributors were explicitly involved in Tory politics. Hooker lived in the 16th century, long before the advent of toryism, whilst Hume was an apolitical philosopher and Halifax similarly politically independent. Burke described himself as a Whig. Shortly after Burke's death in 1797, conservatism revived as a mainstream political force as the Whigs suffered a series of internal divisions. This new generation of conservatives derived their politics not from Burke, but from his predecessor, the Viscount Bolingbroke (1678–1751), who was a Jacobite and traditional Tory, lacking Burke's sympathies for Whiggish policies such as Catholic emancipation and American independence (famously attacked by Samuel Johnson in "Taxation No Tyranny"). In the first half of the 19th century, many newspapers, magazines, and journals promoted loyalist or right-wing attitudes in religion, politics and international affairs. Burke was seldom mentioned, but William Pitt the Younger (1759–1806) became a conspicuous hero. The most prominent journals included "The Quarterly Review", founded in 1809 as a counterweight to the Whigs' "Edinburgh Review" and the even more conservative "Blackwood's Edinburgh Magazine". Sack finds that the "Quarterly Review" promoted a balanced Canningite toryism as it was neutral on Catholic emancipation and only mildly critical of Nonconformist Dissent; it opposed slavery and supported the current poor laws; and it was "aggressively imperialist". The high-church clergy of the Church of England read the "Orthodox Churchman's Magazine" which was equally hostile to Jewish, Catholic, Jacobin, Methodist and Unitarian spokesmen. Anchoring the ultra Tories, "Blackwood's Edinburgh Magazine" stood firmly against Catholic emancipation and favoured slavery, cheap money, mercantilism, the Navigation Acts and the Holy Alliance. Conservatism evolved after 1820, embracing free trade in 1846 and a commitment to democracy, especially under Disraeli. The effect was to significantly strengthen conservatism as a grassroots political force. Conservatism no longer was the philosophical defense of the landed aristocracy, but had been refreshed into redefining its commitment to the ideals of order, both secular and religious, expanding imperialism, strengthened monarchy and a more generous vision of the welfare state as opposed to the punitive vision of the Whigs and liberals. As early as 1835, Disraeli attacked the Whigs and utilitarians as slavishly devoted to an industrial oligarchy, while he described his fellow Tories as the only "really democratic party of England" and devoted to the interests of the whole people. Nevertheless, inside the party there was a tension between the growing numbers of wealthy businessmen on the one side and the aristocracy and rural gentry on the other. The aristocracy gained strength as businessmen discovered they could use their wealth to buy a peerage and a country estate. Although conservatives opposed attempts to allow greater representation of the middle class in parliament, they conceded that electoral reform could not be reversed and promised to support further reforms so long as they did not erode the institutions of church and state. These new principles were presented in the Tamworth Manifesto of 1834, which historians regard as the basic statement of the beliefs of the new Conservative Party. Some conservatives lamented the passing of a pastoral world where the ethos of "noblesse oblige" had promoted respect from the lower classes. They saw the Anglican Church and the aristocracy as balances against commercial wealth. They worked toward legislation for improved working conditions and urban housing. This viewpoint would later be called Tory democracy. However, since Burke, there has always been tension between traditional aristocratic conservatism and the wealthy business class. In 1834, Tory Prime Minister Robert Peel issued the Tamworth Manifesto in which he pledged to endorse moderate political reform. This marked the beginning of the transformation of British conservatism from High Tory reactionism towards a more modern form based on "conservation". The party became known as the Conservative Party as a result, a name it has retained to this day. However, Peel would also be the root of a split in the party between the traditional Tories (led by the Earl of Derby and Benjamin Disraeli) and the "Peelites" (led first by Peel himself, then by the Earl of Aberdeen). The split occurred in 1846 over the issue of free trade, which Peel supported, versus protectionism, supported by Derby. The majority of the party sided with Derby whilst about a third split away, eventually merging with the Whigs and the radicals to form the Liberal Party. Despite the split, the mainstream Conservative Party accepted the doctrine of free trade in 1852. In the second half of the 19th century, the Liberal Party faced political schisms, especially over Irish Home Rule. Leader William Gladstone (himself a former Peelite) sought to give Ireland a degree of autonomy, a move that elements in both the left and right-wings of his party opposed. These split off to become the Liberal Unionists (led by Joseph Chamberlain), forming a coalition with the Conservatives before merging with them in 1912. The Liberal Unionist influence dragged the Conservative Party towards the left as Conservative governments passing a number of progressive reforms at the turn of the 20th century. By the late 19th century, the traditional business supporters of the Liberal Party had joined the Conservatives, making them the party of business and commerce. After a period of Liberal dominance before the First World War, the Conservatives gradually became more influential in government, regaining full control of the cabinet in 1922. In the inter-war period, conservatism was the major ideology in Britain as the Liberal Party vied with the Labour Party for control of the left. After the Second World War, the first Labour government (1945–1951) under Clement Attlee embarked on a program of nationalization of industry and the promotion of social welfare. The Conservatives generally accepted those policies until the 1980s. In the 1980s, the Conservative government of Margaret Thatcher, guided by neoliberal economics, reversed many of Labour's programmes. The Conservative Party also adopt soft eurosceptic politics, and oppose Federal Europe. Other conservative political parties, such as the United Kingdom Independence Party (UKIP, founded in 1993), Northern Ireland's Ulster Unionist Party (UUP) and the Democratic Unionist Party (DUP, founded in 1971), began to appear, although they have yet to make any significant impact at Westminster (, the DUP comprises the largest political party in the ruling coalition in the Northern Ireland Assembly), and from 2017 to 2019 the DUP provided support for the Conservative minority government. Another form of conservatism developed in France in parallel to conservatism in Britain. It was influenced by Counter-Enlightenment works by men such as Joseph de Maistre and Louis de Bonald. Many continental or traditionalist conservatives do not support separation of church and state, with most supporting state recognition of and cooperation with the Catholic Church, such as had existed in France before the Revolution. Eventually, conservatives added Gaullism, patriotism, and nationalism to the list of traditional values they support. Conservatives were the first to embrace nationalism, which was previously associated with liberalism and the Revolution in France. Conservatism developed alongside nationalism in Germany, culminating in Germany's victory over France in the Franco-Prussian War, the creation of the unified German Empire in 1871 and the simultaneous rise of Otto von Bismarck on the European political stage. Bismarck's "balance of power" model maintained peace in Europe for decades at the end of the 19th century. His "revolutionary conservatism" was a conservative state-building strategy designed to make ordinary Germans—not just the Junker elite—more loyal to state and emperor, he created the modern welfare state in Germany in the 1880s. According to Kees van Kersbergen and Barbara Vis, his strategy was: Bismarck also enacted universal male suffrage in the new German Empire in 1871. He became a great hero to German conservatives, who erected many monuments to his memory after he left office in 1890. With the rise of Nazism in 1933, agrarian movements faded and was supplanted by a more command-based economy and forced social integration. Though Adolf Hitler succeeded in garnering the support of many German industrialists, prominent traditionalists openly and secretly opposed his policies of euthanasia, genocide and attacks on organized religion, including Claus von Stauffenberg, Dietrich Bonhoeffer, Henning von Tresckow, Bishop Clemens August Graf von Galen and the monarchist Carl Friedrich Goerdeler. More recently, the work of conservative Christian Democratic Union leader and Chancellor Helmut Kohl helped bring about German reunification, along with the closer European integration in the form of the Maastricht Treaty. Today, German conservatism is often associated with politicians such as Chancellor Angela Merkel, whose tenure has been marked by attempts to save the common European currency (Euro) from demise. The German conservatives are divided under Merkel due to the refugee crisis in Germany and many conservatives in the CDU/CSU oppose the refugee and migrant policies developed under Merkel. American conservatism is a broad system of political beliefs in the United States that is characterized by respect for American traditions, support for Judeo-Christian values, economic liberalism, anti-communism and a defense of Western culture. Liberty within the bounds of conformity to conservatism is a core value, with a particular emphasis on strengthening the free market, limiting the size and scope of government and opposition to high taxes and government or labor union encroachment on the entrepreneur. The major conservative party in the United States is the Republican Party, also known as the GOP (Grand Old Party). American conservatives consider individual liberty, as long as it conforms to conservative values, small government, deregulation of the government, economic liberalism, and free trade, as the fundamental traits of democracy, which contrasts with modern American liberals, who generally place a greater value on social equality and social justice. Conservative political parties vary widely from country to country in the goals they wish to achieve. Both conservative and liberal parties tend to favor private ownership of property, in opposition to communist, socialist and green parties, which favor communal ownership or laws requiring social responsibility on the part of property owners. Where conservatives and liberals differ is primarily on social issues. Conservatives tend to reject behavior that does not conform to some social norm. Modern conservative parties often define themselves by their opposition to liberal or labor parties. The United States usage of the term "conservative" is unique to that country. According to Alan Ware, Belgium, Denmark, Finland, France, Greece, Iceland, Luxembourg, Netherlands, Norway, Sweden, Switzerland and the United Kingdom retained viable conservative parties into the 1980s. Ware said that Australia, Germany, Israel, Italy, Japan, Malta, New Zealand, Spain, and the United States had no conservative parties, although they had either Christian democrats or liberals as major right-wing parties. Canada, Ireland and Portugal had right-wing political parties that defied categorization: the Progressive Conservative Party of Canada; Fianna Fáil, Fine Gael and Progressive Democrats in Ireland; and the Social Democratic Party of Portugal. Since then, the Swiss People's Party has moved to the extreme right and is no longer considered to be conservative. Klaus von Beyme, who developed the method of party categorization, found that no modern Eastern European parties could be considered conservative, although the communist and communist-successor parties had strong similarities. In Italy, which was united by liberals and radicals ("Risorgimento"), liberals, not conservatives, emerged as the party of the right. In the Netherlands, conservatives merged into a new Christian democratic party in 1980. In Austria, Germany, Portugal and Spain, conservatism was transformed into and incorporated into fascism or the far-right. In 1940, all Japanese parties were merged into a single fascist party. Following the war, Japanese conservatives briefly returned to politics, but were largely purged from public office. Louis Hartz explained the absence of conservatism in Australia or the United States as a result of their settlement as radical or liberal fragments of Great Britain. Although he said English Canada had a negligible conservative influence, subsequent writers claimed that loyalists opposed to the American Revolution brought a Tory ideology into Canada. Hartz explained conservatism in Quebec and Latin America as a result of their settlement as feudal societies. The American conservative writer Russell Kirk provided the opinion that conservatism had been brought to the United States and interpreted the American Revolution as a "conservative revolution". Conservative elites have long dominated Latin American nations. Mostly, this has been achieved through control of and support for civil institutions, the church and the armed forces, rather than through party politics. Typically, the church was exempt from taxes and its employees immune from civil prosecution. Where national conservative parties were weak or non-existent, conservatives were more likely to rely on military dictatorship as a preferred form of government. However, in some nations where the elites were able to mobilize popular support for conservative parties, longer periods of political stability were achieved. Chile, Colombia and Venezuela are examples of nations that developed strong conservative parties. Argentina, Brazil, El Salvador and Peru are examples of nations where this did not occur. The Conservative Party of Venezuela disappeared following the Federal Wars of 1858–1863. Chile's conservative party, the National Party, disbanded in 1973 following a military coup and did not re-emerge as a political force following the subsequent return to democracy. Having its roots in the conservative Catholic Party, the Christian People's Party retained a conservative edge through the twentieth century, supporting the king in the Royal Question, supporting nuclear family as the cornerstone of society, defending Christian education, and opposing euthanasia. The Christian People's Party dominated politics in post-war Belgium. In 1999, the party's support collapsed, and it became the country's fifth-largest party. Currently, the N-VA (nieuw-vlaamse alliantie/New Flemish Alliance) is the largest party in Belgium. Canada's conservatives had their roots in the loyalists Tories who left America after the American Revolution. They developed in the socio-economic and political cleavages that existed during the first three decades of the 19th century and had the support of the business, professional and established Church (Anglican) elites in Ontario and to a lesser extent in Quebec. Holding a monopoly over administrative and judicial offices, they were called the "Family Compact" in Ontario and the "Chateau Clique" in Quebec. John A. Macdonald's successful leadership of the movement to confederate the provinces and his subsequent tenure as prime minister for most of the late 19th century rested on his ability to bring together the English-speaking Protestant oligarchy and the ultramontane Catholic hierarchy of Quebec and to keep them united in a conservative coalition. The conservatives combined pro-market liberalism and Toryism. They generally supported an activist government and state intervention in the marketplace and their policies were marked by "noblesse oblige", a paternalistic responsibility of the elites for the less well-off. From 1942, the party was known as the Progressive Conservatives until 2003, when the national party merged with the Canadian Alliance to form the Conservative Party of Canada. The conservative and autonomist Union Nationale, led my Maurice Duplessis, governed the province of Quebec in periods from 1936 to 1960 and in a close alliance with the Catholic Church, small rural elites, agricultors and business elites. This period, known by liberals as the Great Darkness, ended with the Quiet Revolution and the party went into terminal decline. By the end of the 1960s, the political debate in Quebec centered around the question of independence, opposing the social democratic and sovereignist Parti Québécois and the centrist and federalist Quebec Liberal Party, therefore marginalizing the conservative movement. Most French Canadian conservatives rallied either the Quebec Liberal Party or the Parti Québécois, while some of them still tried to offer an autonomist third-way with what was left of the Union Nationale or the more populists Ralliement créditiste du Québec and Parti national populaire, but by the 1981 provincial election politically organized conservatism had been obliterated in Quebec. It slowly started to revive at the 1994 provincial election with the Action démocratique du Québec, who served as Official opposition in the National Assembly from 2007 to 2008, before its merger with François Legault's Coalition Avenir Québec in 2012, that took power in 2018. The modern Conservative Party of Canada has rebranded conservatism and under the leadership of Stephen Harper, the Conservative Party added more conservative policies. The Colombian Conservative Party, founded in 1849, traces its origins to opponents of General Francisco de Paula Santander's 1833–1837 administration. While the term "liberal" had been used to describe all political forces in Colombia, the conservatives began describing themselves as "conservative liberals" and their opponents as "red liberals". From the 1860s until the present, the party has supported strong central government; supported the Catholic Church, especially its role as protector of the sanctity of the family; and opposed separation of church and state. Its policies include the legal equality of all men, the citizen's right to own property and opposition to dictatorship. It has usually been Colombia's second largest party, with the Colombian Liberal Party being the largest. Founded in 1915, the Conservative People's Party of Denmark. was the successor of "Højre" (literally "Right"). The conservative party led the government coalition from 1982 to 1993. The party was a junior partner in coalition with the Liberals from 2001 to 2011. The party is preceded by 11 years by the Young Conservatives (KU), today the youth movement of the party. The party suffered a major defeat in the parliamentary elections of September 2011 in which the party lost more than half of its seat and also lost governmental power. A liberal cultural policy dominated during the post-war period. However, by the 1990s, disagreements regarding immigrants from entirely different cultures ignited a conservative backlash. The conservative party in Finland is the National Coalition Party (in Finnish "Kansallinen Kokoomus", "Kok"). The party was founded in 1918 when several monarchist parties united. Although in the past the party was right-wing, today it is a moderate liberal conservative party. While the party advocates economic liberalism, it is committed to the social market economy. Conservatism in France focused on the rejection of the secularism of the French Revolution, support for the role of the Catholic Church and the restoration of the monarchy. The monarchist cause was on the verge of victory in the 1870s, but then collapsed because the proposed king refused to fly the tri-colored flag. Religious tensions heightened in the 1890–1910 era, but moderated after the spirit of unity in fighting the First World War. An extreme form of conservatism characterized the Vichy regime of 1940–1944 with heightened antisemitism, opposition to individualism, emphasis on family life and national direction of the economy. Following the Second World War, conservatives in France supported Gaullist groups and have been nationalistic and emphasized tradition, order and the regeneration of France. Gaullists held divergent views on social issues. The number of conservative groups, their lack of stability and their tendency to be identified with local issues defy simple categorization. Conservatism has been the major political force in France since the Second World War. Unusually, post-war French conservatism was formed around the personality of a leader, Charles de Gaulle; and did not draw on traditional French conservatism, but on the Bonapartism tradition. Gaullism in France continues under The Republicans (formerly Union for a Popular Movement), which was previously led by Nicolas Sarkozy, a conservative figure in France. The word "conservative" itself is a term of abuse in France. The main inter-war conservative party was called the People's Party (PP), which supported constitutional monarchy and opposed the republican Liberal Party. Both it and the Liberal party were suppressed by the authoritarian, arch-conservative and royalist 4th of August Regime of Ioannis Metaxas in 1936–1941. The PP was able to re-group after the Second World War as part of a United Nationalist Front which achieved power campaigning on a simple anticommunist, ultranationalist platform during the Greek Civil War (1946–1949). However, the vote received by the PP declined during the so-called "Centrist Interlude" in 1950–1952. In 1952, Marshal Alexandros Papagos created the Greek Rally as an umbrella for the right-wing forces. The Greek Rally came to power in 1952 and remained the leading party in Greece until 1963—after Papagos' death in 1955 reformed as the National Radical Union under Konstantinos Karamanlis. Right-wing governments backed by the palace and the army overthrew the Centre Union government in 1965 and governed the country until the establishment of the far-right Greek junta (1967–1974). After the regime's collapse in August 1974, Karamanlis returned from exile to lead the government and founded the New Democracy party. The new conservative party had four objectives: to confront Turkish expansionism in Cyprus, to reestablish and solidify democratic rule, to give the country a strong government and to make a powerful moderate party a force in Greek politics. The Independent Greeks, a newly formed political party in Greece, has also supported conservatism, particularly national and religious conservatism. The Founding Declaration of the Independent Greeks strongly emphasises in the preservation of the Greek state and its sovereignty, the Greek people and the Greek Orthodox Church. Founded in 1924 as the Conservative Party, Iceland's Independence Party adopted its current name in 1929 after the merger with the Liberal Party. From the beginning, they have been the largest vote-winning party, averaging around 40%. They combined liberalism and conservatism, supported nationalization of infrastructure and opposed class conflict. While mostly in opposition during the 1930s, they embraced economic liberalism, but accepted the welfare state after the war and participated in governments supportive of state intervention and protectionism. Unlike other Scandanivian conservative (and liberal) parties, it has always had a large working-class following. After the financial crisis in 2008, the party has sunk to a lower support level around 20–25%. Luxembourg's major Christian democratic conservative party, the Christian Social People's Party (CSV or PCS), was formed as the Party of the Right in 1914 and adopted its present name in 1945. It was consistently the largest political party in Luxembourg, and dominated politics throughout the 20th century. The Conservative Party of Norway (Norwegian: Høyre, literally "right") was formed by the old upper class of state officials and wealthy merchants to fight the populist democracy of the Liberal Party, but lost power in 1884, when parliamentarian government was first practised. It formed its first government under parliamentarism in 1889 and continued to alternate in power with the Liberals until the 1930s, when Labour became the dominant political party. It has elements both of paternalism, stressing the responsibilities of the state, and of economic liberalism. It first returned to power in the 1960s. During Kåre Willoch's premiership in the 1980s, much emphasis was laid on liberalizing the credit and housing market, and abolishing the NRK TV and radio monopoly, while supporting law and order in criminal justice and traditional norms in education Sweden's conservative party, the Moderate Party, was formed in 1904, two years after the founding of the Liberal Party. The party emphasizes tax reductions, deregulation of private enterprise and privatization of schools, hospitals, and kindergartens. There are a number of conservative parties in Switzerland's parliament, the Federal Assembly. These include the largest, the Swiss People's Party (SVP), the Christian Democratic People's Party (CVP) and the Conservative Democratic Party of Switzerland (BDP), which is a splinter of the SVP created in the aftermath to the election of Eveline Widmer-Schlumpf as Federal Council. The right-wing parties have a majority in the Federal Assembly. The Swiss People's Party (SVP or UDC) was formed from the 1971 merger of the Party of Farmers, Traders and Citizens, formed in 1917 and the smaller Swiss Democratic Party, formed in 1942. The SVP emphasized agricultural policy and was strong among farmers in German-speaking Protestant areas. As Switzerland considered closer relations with the European Union in the 1990s, the SVP adopted a more militant protectionist and isolationist stance. This stance has allowed it to expand into German-speaking Catholic mountainous areas. The Anti-Defamation League, a non-Swiss lobby group based in the United States has accused them of manipulating issues such as immigration, Swiss neutrality and welfare benefits, awakening antisemitism and racism. The Council of Europe has called the SVP "extreme right", although some scholars dispute this classification. For instance, Hans-Georg Betz describes it as "populist radical right". The SVP is the largest party since 2003. According to historian James Sack, English conservatives celebrate Edmund Burke as their intellectual father. Burke was affiliated with the Whig Party which eventually became the Liberal Party, but the modern Conservative Party is generally thought to derive from the Tory party and the MPs of the modern conservative party are still frequently referred to as Tories. While conservatism has been seen as an appeal to traditional, hierarchical society, some writers such as Samuel P. Huntington see it as situational. Under this definition, conservatives are seen as defending the established institutions of their time. The Liberal Party of Australia adheres to the principles of social conservatism and liberal conservatism. It is liberal in the sense of economics. Other conservative parties are the National Party of Australia, a sister party of the Liberals, Family First Party, Democratic Labor Party, Shooters, Fishers and Farmers Party, Australian Conservatives, and the Katter's Australian Party. The second largest party in the country is the Australian Labor Party and its dominant faction is Labor Right, a socially conservative element. Australia undertook significant economic reform under the Labor Party in the mid-1980s. Consequently, issues like protectionism, welfare reform, privatization and deregulation are no longer debated in the political space as they are in Europe or North America. Moser and Catley explain: "In America, 'liberal' means left-of-center, and it is a pejorative term when used by conservatives in adversarial political debate. In Australia, of course, the conservatives are in the Liberal Party". Jupp points out that, "[the] decline in English influences on Australian reformism and radicalism, and appropriation of the symbols of Empire by conservatives continued under the Liberal Party leadership of Sir Robert Menzies, which lasted until 1966". Conservatism in Brazil originates from the cultural and historical tradition of Brazil, whose cultural roots are Luso-Iberian and Roman Catholic. Brazilian conservatism from the 20th century on includes names such as Gerardo Melo Mourão and Otto Maria Carpeaux in literature; Oliveira Lima and Oliveira Torres in historiography; Sobral Pinto and Miguel Reale in law; Plinio Corrêa de Oliveira and Father Paulo Ricardo in the Catholic Church; Roberto Campos and Mario Henrique Simonsen in economics; Carlos Lacerda in the political arena; and Olavo de Carvalho in philosophy. Brazilian Labour Renewal Party, Patriota, Progressistas, Social Christian Party and Social Liberal Party are the conservative parties in Brazil. In India, the Bharatiya Janata Party (BJP), led by Narendra Modi, represent conservative politics. The BJP is the largest right-wing conservative party in the world. It promotes cultural nationalism, Hindu Nationalism, an aggressive foreign policy against Pakistan and a conservative social and fiscal policy. After World War II, in Italy the conservative parties were mainly represented by the Christian Democracy (DC) party, which government form the foundation of the Republic until the party's dissolution in 1994. Officially, DC refused the ideology of conservatism, but in many aspects, for example family values, it was a typical social conservative party. In 1994, the media tycoon and entrepreneur Silvio Berlusconi founded the liberal conservative party Forza Italia (FI). Berlusconi won three elections in 1994, 2001 and 2008, governing the country for almost ten years as Prime Minister. Forza Italia formed a coalition with right-wing regional party Lega Nord while in government. Besides FI, now the conservative ideas are mainly expressed by the New Centre-Right party led by Angelino Alfano, Berlusconi formed a new party, which is a rebirth of Forza Italia, thuds founding a new conservative movement. Alfano served as Minister of Foreign Affairs. After the 2018 election, Lega Nord and the Five Star Movement formed the current right-wing populist government. Under Vladimir Putin, the dominant leader since 1999, Russia has promoted explicitly conservative policies in social, cultural and political matters, both at home and abroad. Putin has attacked globalism and economic liberalism. Russian conservatism is unique in some respects as it supports Economic intervention with a mixed economy, with a strong nationalist sentiment and social conservatism with its views being largely populist. Russian conservatism as a result opposes libertarian ideals such as the aforementioned concept of economic liberalism found in other conservative movements around the world. Putin has as a result promoted new think tanks that bring together like-minded intellectuals and writers. For example, the Izborsky Club, founded in 2012 by Aleksandr Prokhanov, stresses Russian nationalism, the restoration of Russia's historical greatness and systematic opposition to liberal ideas and policies. Vladislav Surkov, a senior government official, has been one of the key ideologists during Putin's presidency. In cultural and social affairs, Putin has collaborated closely with the Russian Orthodox Church. Mark Woods provides specific examples of how the Church under Patriarch Kirill of Moscow has backed the expansion of Russian power into Crimea and eastern Ukraine. More broadly, "The New York Times" reports in September 2016 how that Church's policy prescriptions support the Kremlin's appeal to social conservatives: South Korea's major conservative party, the Liberty Korea Party, has changed its form throughout its history. First it was the Democratic-Republican Party (1963–1980) and its head was Park Chung-hee, who seized power in a 1961 military coup d'état and ruled as an unelected military strongman until his formal election as president in 1963. He was president for 16 years until his assassination on 26 October 1979. The Democratic Justice Party inherited the same ideology as the Democratic-Republican Party. Its head, Chun Doo-hwan, also gained power through a coup and his followers called themselves the Hanahae. The Democratic Justice Party changed its form and acted to suppress the opposition party and to follow the people's demand for direct elections. The party's Roh Tae-woo became the first President who was elected through direct election. The next form of the major conservative party was the Democratic-Liberal Party and again through election its second leader, Kim Young-sam, became the fourteenth President of Korea. When the conservative party was beaten by the opposition party in the general election, it changed its form again to follow the party members' demand for reforms. It became the New Korean Party, but it changed again one year later since the President Kim Young-sam was blamed by the citizen for the International Monetary Fund. It changed its name to Grand National Party (GNP). Since the late Kim Dae-jung assumed the presidency in 1998, GNP had been the opposition party until Lee Myung-bak won the presidential election of 2007. The meaning of "conservatism" in the United States has little in common with the way the word is used elsewhere. As Ribuffo (2011) notes, "what Americans now call conservatism much of the world calls liberalism or neoliberalism". Since the 1950s, conservatism in the United States has been chiefly associated with the Republican Party. However, during the era of segregation, many Southern Democrats were conservatives and they played a key role in the conservative coalition that largely controlled domestic policy in Congress from 1937 to 1963. The conservative Democrats continued to have influence in the US politics until 1994's Republican Revolution, when the American South shifted from solid Democrat to solid Republican, while maintaining its conservative values. Major priorities within American conservatism include support for the traditional family, law and order, the right to bear arms, Christian values, anti-communism and a defense of "Western civilization from the challenges of modernist culture and totalitarian governments". Economic conservatives and libertarians favor small government, low taxes, limited regulation and free enterprise. Some social conservatives see traditional social values threatened by secularism, so they support school prayer and oppose abortion and homosexuality. Neoconservatives want to expand American ideals throughout the world and show a strong support for Israel. Paleoconservatives, in opposition to multiculturalism, press for restrictions on immigration. Most US conservatives prefer Republicans over Democrats and most factions favor a strong foreign policy and a strong military. The conservative movement of the 1950s attempted to bring together these divergent strands, stressing the need for unity to prevent the spread of "godless communism", which Reagan later labeled an "evil empire". During the Reagan administration, conservatives also supported the so-called "Reagan Doctrine" under which the US as part of a Cold War strategy provided military and other support to guerrilla insurgencies that were fighting governments identified as socialist or communist. The Reagan administration also adopted neoliberalism and trickle-down economics, as well as Reaganomics, which made for economic growth in the 1980s, fueled by trillion-dollar deficits. Other modern conservative positions include opposition to big government and opposition to environmentalism. On average, American conservatives desire tougher foreign policies than liberals do. Economic liberalism, deregulation and social conservatism are main principles of the Republican Party. The Tea Party movement, founded in 2009, has proven a large outlet for populist American conservative ideas. Their stated goals include rigorous adherence to the US constitution, lower taxes, and opposition to a growing role for the federal government in health care. Electorally, it was considered a key force in Republicans reclaiming control of the US House of Representatives in 2010. This is a broad checklist of modern conservatism in seven countries. Following the Second World War, psychologists conducted research into the different motives and tendencies that account for ideological differences between left and right. The early studies focused on conservatives, beginning with Theodor W. Adorno's "The Authoritarian Personality" (1950) based on the F-scale personality test. This book has been heavily criticized on theoretical and methodological grounds, but some of its findings have been confirmed by further empirical research. In 1973, British psychologist Glenn Wilson published an influential book providing evidence that a general factor underlying conservative beliefs is "fear of uncertainty". A meta-analysis of research literature by Jost, Glaser, Kruglanski and Sulloway in 2003 found that many factors, such as intolerance of ambiguity and need for cognitive closure, contribute to the degree of one's political conservatism and its manifestations in decision-making. A study by Kathleen Maclay stated these traits "might be associated with such generally valued characteristics as personal commitment and unwavering loyalty". The research also suggested that while most people are resistant to change, liberals are more tolerant of it. According to psychologist Bob Altemeyer, individuals who are politically conservative tend to rank high in right-wing authoritarianism (RWA) on his RWA scale. This finding was echoed by Adorno. A study done on Israeli and Palestinian students in Israel found that RWA scores of right-wing party supporters were significantly higher than those of left-wing party supporters. However, a 2005 study by H. Michael Crowson and colleagues suggested a moderate gap between RWA and other conservative positions: "The results indicated that conservatism is not synonymous with RWA". Psychologist Felicia Pratto and her colleagues have found evidence to support the idea that a high social dominance orientation (SDO) is strongly correlated with conservative political views and opposition to social engineering to promote equality, though Pratto's findings have been highly controversial as Pratto and her colleagues found that high SDO scores were highly correlated with measures of prejudice. However, David J. Schneider argued for a more complex relationships between the three factors, writing "correlations between prejudice and political conservative are reduced virtually to zero when controls for SDO are instituted, suggesting that the conservatism–prejudice link is caused by SDO". Conservative political theorist Kenneth Minogue criticized Pratto's work, saying: "It is characteristic of the conservative temperament to value established identities, to praise habit and to respect prejudice, not because it is irrational, but because such things anchor the darting impulses of human beings in solidities of custom which we do not often begin to value until we are already losing them. Radicalism often generates youth movements, while conservatism is a condition found among the mature, who have discovered what it is in life they most value". A 1996 study on the relationship between racism and conservatism found that the correlation was stronger among more educated individuals, though "anti-Black affect had essentially no relationship with political conservatism at any level of educational or intellectual sophistication". They also found that the correlation between racism and conservatism could be entirely accounted for by their mutual relationship with social dominance orientation. A 2008 research report found that conservatives are happier than liberals and that—as income inequality increases—this difference in relative happiness increases because conservatives (more than liberals) possess an ideological buffer against the negative hedonic effects of economic inequality.
https://en.wikipedia.org/wiki?curid=6675
Classical liberalism Classical liberalism is a political ideology and a branch of liberalism which advocates civil liberties under the rule of law with an emphasis on economic freedom. Closely related to economic liberalism, it developed in the early 19th century, building on ideas from the previous century as a response to urbanisation and to the Industrial Revolution in Europe and North America. Notable liberal individuals whose ideas contributed to classical liberalism include John Locke, Jean-Baptiste Say, Thomas Robert Malthus and David Ricardo. It drew on classical economics, especially the economic ideas as espoused by Adam Smith in Book One of "The Wealth of Nations" and on a belief in natural law, progress and utilitarianism. As a term, "classical liberalism" has often been applied in retrospect to distinguish earlier 19th-century liberalism from social liberalism. Core beliefs of classical liberals included new ideas—which departed from both the older conservative idea of society as a family and from the later sociological concept of society as a complex set of social networks. Classical liberals believed that individuals are "egoistic, coldly calculating, essentially inert and atomistic" and that society is no more than the sum of its individual members. Classical liberals agreed with Thomas Hobbes that government had been created by individuals to protect themselves from each other and that the purpose of government should be to minimize conflict between individuals that would otherwise arise in a state of nature. These beliefs were complemented by a belief that laborers could be best motivated by financial incentive. This belief led to the passage of the Poor Law Amendment Act 1834, which limited the provision of social assistance, based on the idea that markets are the mechanism that most efficiently leads to wealth. Adopting Thomas Robert Malthus's population theory, they saw poor urban conditions as inevitable, believed population growth would outstrip food production and thus regarded that consequence desirable because starvation would help limit population growth. They opposed any income or wealth redistribution, believing it would be dissipated by the lowest orders. Drawing on ideas of Adam Smith, classical liberals believed that it is in the common interest that all individuals be able to secure their own economic self-interest. They were critical of what would come to be the idea of the welfare state as interfering in a free market. Despite Smith's resolute recognition of the importance and value of labor and of laborers, classical liberals selectively criticized labour's group rights being pursued at the expense of individual rights while accepting corporations' rights, which led to inequality of bargaining power. Classical liberals argued that individuals should be free to obtain work from the highest-paying employers while the profit motive would ensure that products that people desired were produced at prices they would pay. In a free market, both labor and capital would receive the greatest possible reward while production would be organized efficiently to meet consumer demand. Classical liberals argued for what they called a minimal state, limited to the following functions: Classical liberals asserted that rights are of a negative nature and therefore stipulate that other individuals and governments are to refrain from interfering with the free market, opposing social liberals who assert that individuals have positive rights, such as the right to vote, the right to an education, the right to health care and the right to a living wage. For society to guarantee positive rights, it requires taxation over and above the minimum needed to enforce negative rights. Core beliefs of classical liberals did not necessarily include democracy nor government by a majority vote by citizens because "there is nothing in the bare idea of majority rule to show that majorities will always respect the rights of property or maintain rule of law". For example, James Madison argued for a constitutional republic with protections for individual liberty over a pure democracy, reasoning that in a pure democracy a "common passion or interest will, in almost every case, be felt by a majority of the whole [...] and there is nothing to check the inducements to sacrifice the weaker party". In the late 19th century, classical liberalism developed into neo-classical liberalism, which argued for government to be as small as possible to allow the exercise of individual freedom. In its most extreme form, neo-classical liberalism advocated social Darwinism. Right-libertarianism is a modern form of neo-classical liberalism. Friedrich Hayek identified two different traditions within classical liberalism, namely the British tradition and the French tradition. Hayek saw the British philosophers Bernard Mandeville, David Hume, Adam Smith, Adam Ferguson, Josiah Tucker and William Paley as representative of a tradition that articulated beliefs in empiricism, the common law and in traditions and institutions which had spontaneously evolved but were imperfectly understood. The French tradition included Jean-Jacques Rousseau, Marquis de Condorcet, the Encyclopedists and the Physiocrats. This tradition believed in rationalism and sometimes showed hostility to tradition and religion. Hayek conceded that the national labels did not exactly correspond to those belonging to each tradition since he saw the Frenchmen Montesquieu, Benjamin Constant and Alexis de Tocqueville as belonging to the British tradition and the British Thomas Hobbes, Joseph Priestley, Richard Price and Thomas Paine as belonging to the French tradition. Hayek also rejected the label "laissez-faire" as originating from the French tradition and alien to the beliefs of Hume and Smith. Guido De Ruggiero also identified differences between "Montesquieu and Rousseau, the English and the democratic types of liberalism" and argued that there was a "profound contrast between the two Liberal systems". He claimed that the spirit of "authentic English Liberalism" had "built up its work piece by piece without ever destroying what had once been built, but basing upon it every new departure". This liberalism had "insensibly adapted ancient institutions to modern needs" and "instinctively recoiled from all abstract proclamations of principles and rights". Ruggiero claimed that this liberalism was challenged by what he called the "new Liberalism of France" that was characterised by egalitarianism and a "rationalistic consciousness". In 1848, Francis Lieber distinguished between what he called "Anglican and Gallican Liberty". Lieber asserted that "independence in the highest degree, compatible with safety and broad national guarantees of liberty, is the great aim of Anglican liberty, and self-reliance is the chief source from which it draws its strength". On the other hand, Gallican liberty "is sought in government [...]. [T]he French look for the highest degree of political civilization in organization, that is, in the highest degree of interference by public power". Classical liberalism in Britain traces its roots to the Whigs and radicals, and was heavily influenced by French physiocracy. Whiggery had become a dominant ideology following the Glorious Revolution of 1688 and was associated with supporting the British Parliament, upholding the rule of law, and defending landed property. The origins of rights were seen as being in an ancient constitution, which had existed from time immemorial. These rights, which some Whigs considered to include freedom of the press and freedom of speech, were justified by custom rather than as natural rights. These Whigs believed that the power of the executive had to be constrained. While they supported limited suffrage, they saw voting as a privilege rather than as a right. However, there was no consistency in Whig ideology and diverse writers including John Locke, David Hume, Adam Smith and Edmund Burke were all influential among Whigs, although none of them were universally accepted. From the 1790s to the 1820s, British radicals concentrated on parliamentary and electoral reform, emphasising natural rights and popular sovereignty. Richard Price and Joseph Priestley adapted the language of Locke to the ideology of radicalism. The radicals saw parliamentary reform as a first step toward dealing with their many grievances, including the treatment of Protestant Dissenters, the slave trade, high prices, and high taxes. There was greater unity among classical liberals than there had been among Whigs. Classical liberals were committed to individualism, liberty, and equal rights. They believed these goals required a free economy with minimal government interference. Some elements of Whiggery were uncomfortable with the commercial nature of classical liberalism. These elements became associated with conservatism. Classical liberalism was the dominant political theory in Britain from the early 19th century until the First World War. Its notable victories were the Catholic Emancipation Act of 1829, the Reform Act of 1832 and the repeal of the Corn Laws in 1846. The Anti-Corn Law League brought together a coalition of liberal and radical groups in support of free trade under the leadership of Richard Cobden and John Bright, who opposed aristocratic privilege, militarism, and public expenditure and believed that the backbone of Great Britain was the yeoman farmer. Their policies of low public expenditure and low taxation were adopted by William Ewart Gladstone when he became Chancellor of the Exchequer and later Prime Minister. Classical liberalism was often associated with religious dissent and nonconformism. Although classical liberals aspired to a minimum of state activity, they accepted the principle of government intervention in the economy from the early 19th century on, with passage of the Factory Acts. From around 1840 to 1860, "laissez-faire" advocates of the Manchester School and writers in "The Economist" were confident that their early victories would lead to a period of expanding economic and personal liberty and world peace, but would face reversals as government intervention and activity continued to expand from the 1850s. Jeremy Bentham and James Mill, although advocates of "laissez-faire", non-intervention in foreign affairs, and individual liberty, believed that social institutions could be rationally redesigned through the principles of utilitarianism. The Conservative Prime Minister Benjamin Disraeli rejected classical liberalism altogether and advocated Tory democracy. By the 1870s, Herbert Spencer and other classical liberals concluded that historical development was turning against them. By the First World War, the Liberal Party had largely abandoned classical liberal principles. The changing economic and social conditions of the 19th century led to a division between neo-classical and social (or welfare) liberals, who while agreeing on the importance of individual liberty differed on the role of the state. Neo-classical liberals, who called themselves "true liberals", saw Locke's "Second Treatise" as the best guide and emphasised "limited government" while social liberals supported government regulation and the welfare state. Herbert Spencer in Britain and William Graham Sumner were the leading neo-classical liberal theorists of the 19th century. Neo-classical liberalism has continued into the contemporary era, with writers such as John Rawls. The evolution from classical to social/welfare liberalism is for example reflected in Britain in the evolution of the thought of John Maynard Keynes. In the United States, liberalism took a strong root because it had little opposition to its ideals, whereas in Europe liberalism was opposed by many reactionary or feudal interests such as the nobility, the aristocracy, the landed gentry, the established church and the aristocratic army officers. Thomas Jefferson adopted many of the ideals of liberalism, but in the Declaration of Independence changed Locke's "life, liberty and property" to the more socially liberal "Life, Liberty and the pursuit of Happiness". As the United States grew, industry became a larger and larger part of American life; and during the term of its first populist President, Andrew Jackson, economic questions came to the forefront. The economic ideas of the Jacksonian era were almost universally the ideas of classical liberalism. Freedom, according to classical liberals, was maximised when the government took a "hands off" attitude toward the economy. Historian Kathleen G. Donohue argues: [A]t the center of classical liberal theory [in Europe] was the idea of "laissez-faire". To the vast majority of American classical liberals, however, "laissez-faire" did not mean no government intervention at all. On the contrary, they were more than willing to see government provide tariffs, railroad subsidies, and internal improvements, all of which benefited producers. What they condemned was intervention on behalf of consumers. Leading magazine "The Nation" espoused liberalism every week starting in 1865 under the influential editor Edwin Lawrence Godkin (1831–1902). The ideas of classical liberalism remained essentially unchallenged until a series of depressions, thought to be impossible according to the tenets of classical economics, led to economic hardship from which the voters demanded relief. In the words of William Jennings Bryan, "You shall not crucify the American farmer on a cross of gold". Classical liberalism remained the orthodox belief among American businessmen until the Great Depression. The Great Depression of the 1930s saw a sea change in liberalism, with priority shifting from the producers to consumers. Franklin D. Roosevelt's New Deal represented the dominance of modern liberalism in politics for decades. In the words of Arthur Schlesinger Jr.: Alan Wolfe summarizes the viewpoint that there is a continuous liberal understanding that includes both Adam Smith and John Maynard Keynes: The view that modern liberalism is a continuation of classical liberalism is not universally shared. James Kurth, Robert E. Lerner, John Micklethwait, Adrian Wooldridge and several other political scholars have argued that classical liberalism still exists today, but in the form of American conservatism. According to Deepak Lal, only in the United States does classical liberalism continue to be a significant political force—through American conservatism. Central to classical liberal ideology was their interpretation of John Locke's "Second Treatise of Government" and "A Letter Concerning Toleration", which had been written as a defence of the Glorious Revolution of 1688. Although these writings were considered too radical at the time for Britain's new rulers, they later came to be cited by Whigs, radicals and supporters of the American Revolution. However, much of later liberal thought was absent in Locke's writings or scarcely mentioned and his writings have been subject to various interpretations. For example, there is little mention of constitutionalism, the separation of powers and limited government. James L. Richardson identified five central themes in Locke's writing: individualism, consent, the concepts of the rule of law and government as trustee, the significance of property and religious toleration. Although Locke did not develop a theory of natural rights, he envisioned individuals in the state of nature as being free and equal. The individual, rather than the community or institutions, was the point of reference. Locke believed that individuals had given consent to government and therefore authority derived from the people rather than from above. This belief would influence later revolutionary movements. As a trustee, government was expected to serve the interests of the people, not the rulers; and rulers were expected to follow the laws enacted by legislatures. Locke also held that the main purpose of men uniting into commonwealths and governments was for the preservation of their property. Despite the ambiguity of Locke's definition of property, which limited property to "as much land as a man tills, plants, improves, cultivates, and can use the product of", this principle held great appeal to individuals possessed of great wealth. Locke held that the individual had the right to follow his own religious beliefs and that the state should not impose a religion against Dissenters, but there were limitations. No tolerance should be shown for atheists, who were seen as amoral, or to Catholics, who were seen as owing allegiance to the Pope over their own national government. Adam Smith's "The Wealth of Nations", published in 1776, was to provide most of the ideas of economics, at least until the publication of John Stuart Mill's "Principles of Political Economy" in 1848. Smith addressed the motivation for economic activity, the causes of prices and the distribution of wealth and the policies the state should follow to maximise wealth. Smith wrote that as long as supply, demand, prices and competition were left free of government regulation, the pursuit of material self-interest, rather than altruism, would maximise the wealth of a society through profit-driven production of goods and services. An "invisible hand" directed individuals and firms to work toward the public good as an unintended consequence of efforts to maximise their own gain. This provided a moral justification for the accumulation of wealth, which had previously been viewed by some as sinful. He assumed that workers could be paid wages as low as was necessary for their survival, which was later transformed by David Ricardo and Thomas Robert Malthus into the "iron law of wages". His main emphasis was on the benefit of free internal and international trade, which he thought could increase wealth through specialisation in production. He also opposed restrictive trade preferences, state grants of monopolies and employers' organisations and trade unions. Government should be limited to defence, public works and the administration of justice, financed by taxes based on income. Smith's economics was carried into practice in the nineteenth century with the lowering of tariffs in the 1820s, the repeal of the Poor Relief Act that had restricted the mobility of labour in 1834 and the end of the rule of the East India Company over India in 1858. In addition to Smith's legacy, Say's law, Thomas Robert Malthus' theories of population and David Ricardo's iron law of wages became central doctrines of classical economics. The pessimistic nature of these theories provided a basis for criticism of capitalism by its opponents and helped perpetuate the tradition of calling economics the "dismal science". Jean-Baptiste Say was a French economist who introduced Smith's economic theories into France and whose commentaries on Smith were read in both France and Britain. Say challenged Smith's labour theory of value, believing that prices were determined by utility and also emphasised the critical role of the entrepreneur in the economy. However, neither of those observations became accepted by British economists at the time. His most important contribution to economic thinking was Say's law, which was interpreted by classical economists that there could be no overproduction in a market and that there would always be a balance between supply and demand. This general belief influenced government policies until the 1930s. Following this law, since the economic cycle was seen as self-correcting, government did not intervene during periods of economic hardship because it was seen as futile. Malthus wrote two books, "An Essay on the Principle of Population" (published in 1798) and "Principles of Political Economy" (published in 1820). The second book which was a rebuttal of Say's law had little influence on contemporary economists. However, his first book became a major influence on classical liberalism. In that book, Malthus claimed that population growth would outstrip food production because population grew geometrically while food production grew arithmetically. As people were provided with food, they would reproduce until their growth outstripped the food supply. Nature would then provide a check to growth in the forms of vice and misery. No gains in income could prevent this and any welfare for the poor would be self-defeating. The poor were in fact responsible for their own problems which could have been avoided through self-restraint. Ricardo, who was an admirer of Smith, covered many of the same topics, but while Smith drew conclusions from broadly empirical observations he used deduction, drawing conclusions by reasoning from basic assumptions While Ricardo accepted Smith's labour theory of value, he acknowledged that utility could influence the price of some rare items. Rents on agricultural land were seen as the production that was surplus to the subsistence required by the tenants. Wages were seen as the amount required for workers' subsistence and to maintain current population levels. According to his iron law of wages, wages could never rise beyond subsistence levels. Ricardo explained profits as a return on capital, which itself was the product of labour, but a conclusion many drew from his theory was that profit was a surplus appropriated by capitalists to which they were not entitled. Utilitarianism provided the political justification for implementation of economic liberalism by British governments, which was to dominate economic policy from the 1830s. Although utilitarianism prompted legislative and administrative reform and John Stuart Mill's later writings on the subject foreshadowed the welfare state, it was mainly used as a justification for "laissez-faire". The central concept of utilitarianism, which was developed by Jeremy Bentham, was that public policy should seek to provide "the greatest happiness of the greatest number". While this could be interpreted as a justification for state action to reduce poverty, it was used by classical liberals to justify inaction with the argument that the net benefit to all individuals would be higher. Classical liberals saw utility as the foundation for public policies. This broke both with conservative "tradition" and Lockean "natural rights", which were seen as irrational. Utility, which emphasises the happiness of individuals, became the central ethical value of all liberalism. Although utilitarianism inspired wide-ranging reforms, it became primarily a justification for "laissez-faire" economics. However, classical liberals rejected Smith's belief that the "invisible hand" would lead to general benefits and embraced Malthus' view that population expansion would prevent any general benefit and Ricardo's view of the inevitability of class conflict. "Laissez-faire" was seen as the only possible economic approach and any government intervention was seen as useless and harmful. The Poor Law Amendment Act 1834 was defended on "scientific or economic principles" while the authors of the Elizabethan Poor Law of 1601 were seen as not having had the benefit of reading Malthus. However, commitment to "laissez-faire" was not uniform and some economists advocated state support of public works and education. Classical liberals were also divided on free trade as Ricardo expressed doubt that the removal of grain tariffs advocated by Richard Cobden and the Anti-Corn Law League would have any general benefits. Most classical liberals also supported legislation to regulate the number of hours that children were allowed to work and usually did not oppose factory reform legislation. Despite the pragmatism of classical economists, their views were expressed in dogmatic terms by such popular writers as Jane Marcet and Harriet Martineau. The strongest defender of "laissez-faire" was "The Economist" founded by James Wilson in 1843. "The Economist" criticised Ricardo for his lack of support for free trade and expressed hostility to welfare, believing that the lower orders were responsible for their economic circumstances. "The Economist" took the position that regulation of factory hours was harmful to workers and also strongly opposed state support for education, health, the provision of water and granting of patents and copyrights. "The Economist" also campaigned against the Corn Laws that protected landlords in the United Kingdom of Great Britain and Ireland against competition from less expensive foreign imports of cereal products. A rigid belief in "laissez-faire" guided the government response in 1846–1849 to the Great Famine in Ireland, during which an estimated 1.5 million people died. The minister responsible for economic and financial affairs, Charles Wood, expected that private enterprise and free trade, rather than government intervention, would alleviate the famine. The Corn Laws were finally repealed in 1846 by the removal of tariffs on grain which kept the price of bread artificially high, but it came too late to stop the Irish famine, partly because it was done in stages over three years. Several liberals, including Smith and Cobden, argued that the free exchange of goods between nations could lead to world peace. Erik Gartzke states: "Scholars like Montesquieu, Adam Smith, Richard Cobden, Norman Angell, and Richard Rosecrance have long speculated that free markets have the potential to free states from the looming prospect of recurrent warfare". American political scientists John R. Oneal and Bruce M. Russett, well known for their work on the democratic peace theory, state: In "The Wealth of Nations", Smith argued that as societies progressed from hunter gatherers to industrial societies the spoils of war would rise, but that the costs of war would rise further and thus making war difficult and costly for industrialised nations: Cobden believed that military expenditures worsened the welfare of the state and benefited a small, but concentrated elite minority, summing up British imperialism, which he believed was the result of the economic restrictions of mercantilist policies. To Cobden and many classical liberals, those who advocated peace must also advocate free markets. The belief that free trade would promote peace was widely shared by English liberals of the 19th and early 20th century, leading the economist John Maynard Keynes (1883–1946), who was a classical liberal in his early life, to say that this was a doctrine on which he was "brought up" and which he held unquestioned only until the 1920s. In his review of a book on Keynes, Michael S. Lawlor argues that it may be in large part due to Keynes' contributions in economics and politics, as in the implementation of the Marshall Plan and the way economies have been managed since his work, "that we have the luxury of not facing his unpalatable choice between free trade and full employment". A related manifestation of this idea was the argument of Norman Angell (1872–1967), most famously before World War I in "The Great Illusion" (1909), that the interdependence of the economies of the major powers was now so great that war between them was futile and irrational; and therefore unlikely.
https://en.wikipedia.org/wiki?curid=6677
Cat The cat ("Felis catus") is a domestic species of small carnivorous mammal. It is the only domesticated species in the family Felidae and is often referred to as the domestic cat to distinguish it from the wild members of the family. A cat can either be a house cat, a farm cat or a feral cat; the latter ranges freely and avoids human contact. Domestic cats are valued by humans for companionship and their ability to hunt rodents. About 60 cat breeds are recognized by various cat registries. The cat is similar in anatomy to the other felid species: it has a strong flexible body, quick reflexes, sharp teeth and retractable claws adapted to killing small prey. Its night vision and sense of smell are well developed. Cat communication includes vocalizations like meowing, purring, trilling, hissing, growling and grunting as well as cat-specific body language. It is a solitary hunter but a social species. It can hear sounds too faint or too high in frequency for human ears, such as those made by mice and other small mammals. It is a predator that is most active at dawn and dusk. It secretes and perceives pheromones. Female domestic cats can have kittens from spring to late autumn, with litter sizes often ranging from two to five kittens. Domestic cats are bred and shown at events as registered pedigreed cats, a hobby known as cat fancy. Failure to control breeding of pet cats by spaying and neutering, as well as abandonment of pets, resulted in large numbers of feral cats worldwide, contributing to the extinction of entire bird, mammal, and reptile species, and evoking population control. Cats were first domesticated in the Near East around 7500 BC. It was long thought that cat domestication was initiated in Ancient Egypt, as since around 3100 BC veneration was given to cats in ancient Egypt. , the domestic cat was the second-most popular pet in the United States by number of pets owned, after freshwater fish, with 95 million cats owned. In the United Kingdom, around 7.3 million cats lived in more than 4.8 million households . The origin of the English word 'cat', Old English , is thought to be the Late Latin word , which was first used at the beginning of the 6th century. It was suggested that the word 'cattus' is derived from an Egyptian precursor of Coptic , "tomcat", or its feminine form suffixed with . The Late Latin word is also thought to be derived from Afro-Asiatic languages. The Nubian word "wildcat" and Nobiin are possible sources or cognates. The Nubian word may be a loan from Arabic ~ . It is "equally likely that the forms might derive from an ancient Germanic word, imported into Latin and thence to Greek and to Syriac and Arabic". The word may be derived from Germanic and Northern European languages, and ultimately be borrowed from Uralic, cf. Northern Sami , "female stoat", and Hungarian , "stoat"; from Proto-Uralic "*käďwä", "female (of a furred animal)". The English "puss", extended as "pussy" and "pussycat", is attested from the 16th century and may have been introduced from Dutch or from Low German , related to Swedish , or Norwegian , . Similar forms exist in Lithuanian and Irish or . The etymology of this word is unknown, but it may have simply arisen from a sound used to attract a cat. A male cat is called a "tom" or "tomcat" (or a "gib", if neutered) An unspayed female is called a "queen", especially in a cat-breeding context. A juvenile cat is referred to as a "kitten". In Early Modern English, the word "kitten" was interchangeable with the now-obsolete word "catling". A group of cats can be referred to as a "clowder" or a "glaring". The scientific name "Felis catus" was proposed by Carl Linnaeus in 1758 for a domestic cat. "Felis catus domesticus" was proposed by Johann Christian Polycarp Erxleben in 1777. "Felis daemon" proposed by Konstantin Alekseevich Satunin in 1904 was a black cat from the Transcaucasus, later identified as a domestic cat. In 2003, the International Commission on Zoological Nomenclature ruled that the domestic cat is a distinct species, namely "Felis catus". In 2007, it was considered a subspecies of the European wildcat, "F. silvestris catus", following results of phylogenetic research. In 2017, the IUCN Cat Classification Taskforce followed the recommendation of the ICZN in regarding the domestic cat as a distinct species, "Felis catus". The domestic cat is a member of the Felidae, a family that had a common ancestor about 10–15 million years ago. The genus "Felis" diverged from the Felidae around 6–7 million years ago. Results of phylogenetic research confirm that the wild "Felis" species evolved through sympatric or parapatric speciation, whereas the domestic cat evolved through artificial selection. The domesticated cat and its closest wild ancestor are both diploid organisms that possess 38 chromosomes and roughly 20,000 genes. The leopard cat ("Prionailurus bengalensis") was tamed independently in China around 5500 BC. This line of partially domesticated cats leaves no trace in the domestic cat populations of today. The earliest known indication for the taming of an African wildcat ("F. lybica") was excavated close by a human Neolithic grave in Shillourokambos, southern Cyprus, dating to about 9,200 to 9,500 years before present. As there is no evidence of native mammalian fauna on Cyprus, the inhabitants of this Neolithic village most likely brought the cat and other wild mammals to the island from the Middle Eastern mainland. Scientists therefore assume that African wildcats were attracted to early human settlements in the Fertile Crescent by rodents, in particular the house mouse ("Mus musculus"), and were tamed by Neolithic farmers. This commensal relationship between early farmers and tamed cats lasted thousands of years. As agricultural practices spread, so did tame and domesticated cats. Wildcats of Egypt contributed to the maternal gene pool of the domestic cat at a later time. The earliest known evidence for the occurrence of the domestic cat in Greece dates to around 1200 BC. Greek, Phoenician, Carthaginian and Etruscan traders introduced domestic cats to southern Europe. During the Roman Empire they were introduced to Corsica and Sardinia before the beginning of the 1st millennium. By the 5th century BC, they were familiar animals around settlements in Magna Graecia and Etruria. By the end of the Roman Empire in the 5th century, the Egyptian domestic cat lineage had arrived in a Baltic Sea port in northern Germany. During domestication, cats have undergone only minor changes in anatomy and behavior, and they are still capable of surviving in the wild. Several natural behaviors and characteristics of wildcats may have preadapted them for domestication as pets. These traits include their small size, social nature, obvious body language, love of play and relatively high intelligence. Captive "Leopardus" cats may also display affectionate behavior toward humans, but were not domesticated. House cats often mate with feral cats, producing hybrids such as the Kellas cat in Scotland. Hybridisation between domestic and other Felinae species is also possible. Development of cat breeds started in the mid 19th century. An analysis of the domestic cat genome revealed that the ancestral wildcat genome was significantly altered in the process of domestication as specific mutations were selected to develop cat breeds. Most breeds are founded on random-bred domestic cats. Genetic diversity of these breeds varies between regions, and is lowest in purebred populations, which show more than 20 deleterious genetic disorders. The domestic cat has a smaller skull and shorter bones than the European wildcat. It averages about in head-to-body length and in height, with about long tails. Males are larger than females. Adult domestic cats typically weigh between . Cats have seven cervical vertebrae (as do most mammals); 13 thoracic vertebrae (humans have 12); seven lumbar vertebrae (humans have five); three sacral vertebrae (as do most mammals, but humans have five); and a variable number of caudal vertebrae in the tail (humans have only vestigial caudal vertebrae, fused into an internal coccyx). The extra lumbar and thoracic vertebrae account for the cat's spinal mobility and flexibility. Attached to the spine are 13 ribs, the shoulder, and the pelvis. Unlike human arms, cat forelimbs are attached to the shoulder by free-floating clavicle bones which allow them to pass their body through any space into which they can fit their head. The cat skull is unusual among mammals in having very large eye sockets and a powerful specialized jaw. Within the jaw, cats have teeth adapted for killing prey and tearing meat. When it overpowers its prey, a cat delivers a lethal neck bite with its two long canine teeth, inserting them between two of the prey's vertebrae and severing its spinal cord, causing irreversible paralysis and death. Compared to other felines, domestic cats have narrowly spaced canine teeth relative to the size of their jaw, which is an adaptation to their preferred prey of small rodents, which have small vertebrae. The premolar and first molar together compose the carnassial pair on each side of the mouth, which efficiently shears meat into small pieces, like a pair of scissors. These are vital in feeding, since cats' small molars cannot chew food effectively, and cats are largely incapable of mastication. Although cats tend to have better teeth than most humans, with decay generally less likely because of a thicker protective layer of enamel, a less damaging saliva, less retention of food particles between teeth, and a diet mostly devoid of sugar, they are nonetheless subject to occasional tooth loss and infection. The cat is digitigrade. It walks on the toes, with the bones of the feet making up the lower part of the visible leg. Unlike most mammals, it uses a "pacing" gait and moves both legs on one side of the body before the legs on the other side. It registers directly by placing each hind paw close to the track of the corresponding fore paw, minimizing noise and visible tracks. This also provides sure footing for hind paws when navigating rough terrain. As it speeds up walking to trotting, its gait changes to a "diagonal" gait: the diagonally opposite hind and fore legs move simultaneously. Cats have protractable and retractable claws. In their normal, relaxed position, the claws are sheathed with the skin and fur around the paw's toe pads. This keeps the claws sharp by preventing wear from contact with the ground and allows the silent stalking of prey. The claws on the fore feet are typically sharper than those on the hind feet. Cats can voluntarily extend their claws on one or more paws. They may extend their claws in hunting or self-defense, climbing, kneading, or for extra traction on soft surfaces. Cats shed the outside layer of their claw sheaths when scratching rough surfaces. Most cats have five claws on their front paws, and four on their rear paws. The dewclaw is proximal to the other claws. More proximally is a protrusion which appears to be a sixth "finger". This special feature of the front paws, on the inside of the wrists has no function in normal walking, but is thought to be an antiskidding device used while jumping. Some cat breeds are prone to having extra digits (polydactyly). Polydactylous cats occur along North America's northeast coast and in Great Britain. Cats have excellent night vision and can see at only one-sixth the light level required for human vision. This is partly the result of cat eyes having a tapetum lucidum, which reflects any light that passes through the retina back into the eye, thereby increasing the eye's sensitivity to dim light. Large pupils are an adaptation to dim light. The domestic cat has slit pupils, which allow it to focus bright light without chromatic aberration. At low light, a cat's pupils expand to cover most of the exposed surface of its eyes. However, the domestic cat has rather poor color vision and only two types of cone cells, optimized for sensitivity to blue and yellowish green; its ability to distinguish between red and green is limited. A response to middle wavelengths from a system other than the rod cells might be due to a third type of cone. However, this appears to be an adaptation to low light levels rather than representing true trichromatic vision. The domestic cat's hearing is most acute in the range of 500 Hz to 32 kHz. It can detect an extremely broad range of frequencies ranging from 55 Hz to 79,000 Hz. It can hear a range of 10.5 octaves, while humans and dogs can hear ranges of about 9 octaves. Its hearing sensitivity is enhanced by its large movable outer ears, the pinnae, which amplify sounds and help detect the location of a noise. It can detect ultrasound, which enables it to detect ultrasonic calls made by rodent prey. Cats have an acute sense of smell, due in part to their well-developed olfactory bulb and a large surface of olfactory mucosa, about in area, which is about twice that of humans. Cats and many other animals have a Jacobson's organ in their mouths that is used in the behavioral process of flehmening. It allows them to sense certain aromas in a way that humans cannot. Cats are sensitive to pheromones such as 3-mercapto-3-methylbutan-1-ol, which they use to communicate through urine spraying and marking with scent glands. Many cats also respond strongly to plants that contain nepetalactone, especially catnip, as they can detect that substance at less than one part per billion. About 70–80% of cats are affected by nepetalactone. This response is also produced by other plants, such as silver vine ("Actinidia polygama") and the herb valerian; it may be caused by the smell of these plants mimicking a pheromone and stimulating cats' social or sexual behaviors. Cats have relatively few taste buds compared to humans (470 or so versus more than 9,000 on the human tongue). Domestic and wild cats share a taste receptor gene mutation that keeps their sweet taste buds from binding to sugary molecules, leaving them with no ability to taste sweetness. Their taste buds instead respond to acids, amino acids like protein, and bitter tastes. Cats also have a distinct temperature preference for their food, preferring food with a temperature around which is similar to that of a fresh kill and routinely rejecting food presented cold or refrigerated (which would signal to the cat that the "prey" item is long dead and therefore possibly toxic or decomposing). To aid with navigation and sensation, cats have dozens of movable whiskers (vibrissae) over their body, especially their faces. These provide information on the width of gaps and on the location of objects in the dark, both by touching objects directly and by sensing air currents; they also trigger protective blink reflexes to protect the eyes from damage. Most breeds of cat have a noted fondness for sitting in high places, or perching. A higher place may serve as a concealed site from which to hunt; domestic cats strike prey by pouncing from a perch such as a tree branch. Another possible explanation is that height gives the cat a better observation point, allowing it to survey its territory. A cat falling from heights of up to 3 meters can right itself and land on its paws. During a fall from a high place, a cat reflexively twists its body and rights itself to land on its feet using its acute sense of balance and flexibility. This reflex is known as the cat righting reflex. An individual cat always rights itself in the same way during a fall, provided it has sufficient time to do so. The height required for this to occur is around . Cats without a tail also have this reflex. Several explanations have been proposed for this phenomenon since the late 19th century: Outdoor cats are active both day and night, although they tend to be slightly more active at night. Domestic cats spend the majority of their time in the vicinity of their homes, but can range many hundreds of meters from this central point. They establish territories that vary considerably in size, in one study ranging from . The timing of cats' activity is quite flexible and varied, which means house cats may be more active in the morning and evening, as a response to greater human activity at these times. Cats conserve energy by sleeping more than most animals, especially as they grow older. The daily duration of sleep varies, usually between 12 and 16 hours, with 13 and 14 being the average. Some cats can sleep as much as 20 hours. The term "cat nap" for a short rest refers to the cat's tendency to fall asleep (lightly) for a brief period. While asleep, cats experience short periods of rapid eye movement sleep often accompanied by muscle twitches, which suggests they are dreaming. The social behavior of the domestic cat ranges from widely dispersed individuals to feral cat colonies that gather around a food source, based on groups of co-operating females. Within such groups, one cat is usually dominant over the others. Each cat in a colony holds a distinct territory, with sexually active males having the largest territories, which are about 10 times larger than those of female cats and may overlap with several females' territories. These territories are marked by urine spraying, by rubbing objects at head height with secretions from facial glands, and by defecation. Between these territories are neutral areas where cats watch and greet one another without territorial conflicts. Outside these neutral areas, territory holders usually chase away stranger cats, at first by staring, hissing, and growling and, if that does not work, by short but noisy and violent attacks. Despite some cats cohabiting in colonies, they do not have a social survival strategy, or a pack mentality and always hunt alone. However, some pet cats are poorly socialized. In particular, older cats show aggressiveness towards newly arrived kittens, which include biting and scratching; this type of behavior is known as feline asocial aggression. Life in proximity to humans and other domestic animals has led to a symbiotic social adaptation in cats, and cats may express great affection toward humans or other animals. Ethologically, the human keeper of a cat functions as a sort of surrogate for the cat's mother. Adult cats live their lives in a kind of extended kittenhood, a form of behavioral neoteny. Their high-pitched sounds may mimic the cries of a hungry human infant, making them particularly difficult for humans to ignore. Domestic cats' scent rubbing behavior towards humans or other cats is thought to be a feline means for social bonding. Domestic cats use many vocalizations for communication, including purring, trilling, hissing, growling/snarling, grunting, and several different forms of meowing. Their body language, including position of ears and tail, relaxation of the whole body, and kneading of the paws, are all indicators of mood. The tail and ears are particularly important social signal mechanisms in cats. A raised tail indicates a friendly greeting, and flattened ears indicates hostility. Tail-raising also indicates the cat's position in the group's social hierarchy, with dominant individuals raising their tails less often than subordinate ones. Feral cats are generally silent. Nose-to-nose touching is also a common greeting and may be followed by social grooming, which is solicited by one of the cats raising and tilting its head. Purring may have developed as an evolutionary advantage as a signalling mechanism of reassurance between mother cats and nursing kittens. Post-nursing cats often purr as a sign of contentment: when being petted, becoming relaxed, or eating. The mechanism by which cats purr is elusive. The cat has no unique anatomical feature that is clearly responsible for the sound. Cats are known for spending considerable amounts of time licking their coats to keep them clean. The cat's tongue has backwards-facing spines about 500 μm long, which are called papillae. These contain keratin which makes them rigid so the papillae act like a hairbrush. Some cats, particularly longhaired cats, occasionally regurgitate hairballs of fur that have collected in their stomachs from grooming. These clumps of fur are usually sausage-shaped and about long. Hairballs can be prevented with remedies that ease elimination of the hair through the gut, as well as regular grooming of the coat with a comb or stiff brush. Among domestic cats, males are more likely to fight than females. Among feral cats, the most common reason for cat fighting is competition between two males to mate with a female. In such cases, most fights are won by the heavier male. Another common reason for fighting in domestic cats is the difficulty of establishing territories within a small home. Female cats also fight over territory or to defend their kittens. Neutering will decrease or eliminate this behavior in many cases, suggesting that the behavior is linked to sex hormones. When cats become aggressive, they try to make themselves appear larger and more threatening by raising their fur, arching their backs, turning sideways and hissing or spitting. Often, the ears are pointed down and back to avoid damage to the inner ear and potentially listen for any changes behind them while focused forward. They may also vocalize loudly and bare their teeth in an effort to further intimidate their opponent. Fights usually consist of grappling and delivering powerful slaps to the face and body with the forepaws as well as bites. Cats also throw themselves to the ground in a defensive posture to rake their opponent's belly with their powerful hind legs. Serious damage is rare, as the fights are usually short in duration, with the loser running away with little more than a few scratches to the face and ears. However, fights for mating rights are typically more severe and injuries may include deep puncture wounds and lacerations. Normally, serious injuries from fighting are limited to infections of scratches and bites, though these can occasionally kill cats if untreated. In addition, bites are probably the main route of transmission of feline immunodeficiency virus. Sexually active males are usually involved in many fights during their lives, and often have decidedly battered faces with obvious scars and cuts to their ears and nose. The shape and structure of cats' cheeks is insufficient to suck. They lap with the tongue to draw liquid upwards into their mouths. Lapping at a rate of four times a second, the cat touches the smooth tip of its tongue to the surface of the water, and quickly retracts it like a corkscrew, drawing water upwards. Free-fed feral cats and house cats consume several small meals in a day. The frequency and size of meals varies between individuals. They select food based on its temperature, smell and texture; they dislike chilled foods and respond most strongly to moist foods rich in amino acids, which are similar to meat. Cats reject novel flavors (a response termed neophobia) and learn quickly to avoid foods that have tasted unpleasant in the past. They also avoid sweet food and milk. Most adult cats are lactose intolerant; the sugar in milk is not easily digested and may cause soft stools or diarrhea. Some also develop odd eating habits and like to eat or chew on things like wool, plastic, cables, paper, string, aluminum foil, or even coal. This condition, pica, can threaten their health, depending on the amount and toxicity of the items eaten. Cats hunt small prey, primarily birds and rodents, and are often used as a form of pest control. Cats use two hunting strategies, either stalking prey actively, or waiting in ambush until an animal comes close enough to be captured. The strategy used depends on the prey species in the area, with cats waiting in ambush outside burrows, but tending to actively stalk birds. Domestic cats are a major predator of wildlife in the United States, killing an estimated 1.4 to 3.7 billion birds and 6.9 to 20.7 billion mammals annually. Certain species appear more susceptible than others; for example, 30% of house sparrow mortality is linked to the domestic cat. In the recovery of ringed robins ("Erithacus rubecula") and dunnocks ("Prunella modularis"), 31% of deaths were a result of cat predation. In parts of North America, the presence of larger carnivores such as coyotes which prey on cats and other small predators reduces the effect of predation by cats and other small predators such as opossums and raccoons on bird numbers and variety. Perhaps the best-known element of cats' hunting behavior, which is commonly misunderstood and often appalls cat owners because it looks like torture, is that cats often appear to "play" with prey by releasing it after capture. This cat and mouse behavior is due to an instinctive imperative to ensure that the prey is weak enough to be killed without endangering the cat. Another poorly understood element of cat hunting behavior is the presentation of prey to human guardians. One explanation is that cats adopt humans into their social group and share excess kill with others in the group according to the dominance hierarchy, in which humans are reacted to as if they are at, or near, the top. Another explanation is that they attempt to teach their guardians to hunt or to help their human as if feeding "an elderly cat, or an inept kitten". This hypothesis is inconsistent with the fact that male cats also bring home prey, despite males having negligible involvement in raising kittens. On islands, birds can contribute as much as 60% of a cat's diet. In nearly all cases, however, the cat cannot be identified as the sole cause for reducing the numbers of island birds, and in some instances, eradication of cats has caused a "mesopredator release" effect; where the suppression of top carnivores creates an abundance of smaller predators that cause a severe decline in their shared prey. Domestic cats are, however, known to be a contributing factor to the decline of many species, a factor that has ultimately led, in some cases, to extinction. The South Island piopio, Chatham rail, and the New Zealand merganser are a few from a long list, with the most extreme case being the flightless Lyall's wren, which was driven to extinction only a few years after its discovery. One feral cat in New Zealand killed 102 New Zealand lesser short-tailed bats in seven days. In the US, feral and free-ranging domestic cats kill an estimated 6.3 – 22.3 billion mammals annually. In Australia, the impact of cats on mammal populations is even greater than the impact of habitat loss. More than one million reptiles are killed by feral cats each day, representing 258 species. Cats have contributed to the extinction of the Navassa curly-tailed lizard and "Chioninia coctei." Domestic cats, especially young kittens, are known for their love of play. This behavior mimics hunting and is important in helping kittens learn to stalk, capture, and kill prey. Cats also engage in play fighting, with each other and with humans. This behavior may be a way for cats to practice the skills needed for real combat, and might also reduce any fear they associate with launching attacks on other animals. Cats also tend to play with toys more when they are hungry. Owing to the close similarity between play and hunting, cats prefer to play with objects that resemble prey, such as small furry toys that move rapidly, but rapidly lose interest. They become habituated to a toy they have played with before. String is often used as a toy, but if it is eaten, it can become caught at the base of the cat's tongue and then move into the intestines, a medical emergency which can cause serious illness, even death. Owing to the risks posed by cats eating string, it is sometimes replaced with a laser pointer's dot, which cats may chase. Female cats called queens are polyestrous with several estrus cycles during a year, lasting usually 21 days. They are usually ready to mate between early February and August. Several males, called tomcats, are attracted to a female in heat. They fight over her, and the victor wins the right to mate. At first, the female rejects the male, but eventually, the female allows the male to mate. The female utters a loud yowl as the male pulls out of her because a male cat's penis has a band of about 120–150 backward-pointing penile spines, which are about long; upon withdrawal of the penis, the spines rake the walls of the female's vagina, which acts to induce ovulation. This act also occurs to clear the vagina of other sperm in the context of a second (or more) mating, thus giving the later males a larger chance of conception. After mating, the female cleans her vulva thoroughly. If a male attempts to mate with her at this point, the female attacks him. After about 20 to 30 minutes, once the female is finished grooming, the cycle will repeat. Because ovulation is not always triggered by a single mating, females may not be impregnated by the first male with which they mate. Furthermore, cats are superfecund; that is, a female may mate with more than one male when she is in heat, with the result that different kittens in a litter may have different fathers. The morula forms 124 hours after conception. At 148 hours, early blastocysts form. At 10–12 days, implantation occurs. The gestation of queens lasts between 64 and 67 days, with an average of 65 days. Data on the reproductive capacity of more than 2,300 free-ranging queens were collected during a study between May 1998 and October 2000. They had one to six kittens per litter, with an average of three kittens. They produced a mean of 1.4 litters per year, but a maximum of three litters in a year. Of 169 kittens, 127 died before they were six months old due to a trauma caused in most cases by dog attacks and road accidents. The first litter is usually smaller than subsequent litters. Kittens are weaned between six and seven weeks of age. Queens normally reach sexual maturity at 5–10 months, and males at 5–7 months. This varies depending on breed. Kittens reach puberty at the age of 9–10 months. Cats are ready to go to new homes at about 12 weeks of age, when they are ready to leave their mother. They can be surgically sterilized (spayed or castrated) as early as seven weeks to limit unwanted reproduction. This surgery also prevents undesirable sex-related behavior, such as aggression, territory marking (spraying urine) in males and yowling (calling) in females. Traditionally, this surgery was performed at around six to nine months of age, but it is increasingly being performed before puberty, at about three to six months. In the United States, about 80% of household cats are neutered. The average lifespan of pet cats has risen in recent decades. In the early 1980s, it was about seven years, rising to 9.4 years in 1995 and 15.1 years in 2018. Some cats have been reported as surviving into their 30s, with the oldest known cat, Creme Puff, dying at a verified age of 38. Spaying or neutering increases life expectancy: one study found neutered male cats live twice as long as intact males, while spayed female cats live 62% longer than intact females. Having a cat neutered confers health benefits, because castrated males cannot develop testicular cancer, spayed females cannot develop uterine or ovarian cancer, and both have a reduced risk of mammary cancer. Despite widespread concern about the welfare of free-roaming cats, the lifespans of neutered feral cats in managed colonies compare favorably with those of pet cats. About two hundred fifty heritable genetic disorders have been identified in cats, many similar to human inborn errors of metabolism. The high level of similarity among the metabolism of mammals allows many of these feline diseases to be diagnosed using genetic tests that were originally developed for use in humans, as well as the use of cats as animal models in the study of the human diseases. Diseases affecting domestic cats include acute infections, parasitic infestations, injuries, and chronic diseases such as kidney disease, thyroid disease, and arthritis. Vaccinations are available for many infectious diseases, as are treatments to eliminate parasites such as worms and fleas. The domestic cat is a cosmopolitan species and occurs across much of the world. It is adaptable and now present on all continents except Antarctica, and on 118 of the 131 main groups of islands—even on isolated islands such as the Kerguelen Islands. Due to its ability to thrive in almost any terrestrial habitat, it is among the world's most invasive species. As it is little altered from the wildcat, it can readily interbreed with the wildcat. This hybridization poses a danger to the genetic distinctiveness of some wildcat populations, particularly in Scotland and Hungary and possibly also the Iberian Peninsula. It lives on small islands with no human inhabitants. Feral cats can live in forests, grasslands, tundra, coastal areas, agricultural land, scrublands, urban areas, and wetlands. Feral cats are domestic cats that were born in or have reverted to a wild state. They are unfamiliar with and wary of humans and roam freely in urban and rural areas. The numbers of feral cats is not known, but estimates of the United States feral population range from twenty-five to sixty million. Feral cats may live alone, but most are found in large colonies, which occupy a specific territory and are usually associated with a source of food. Famous feral cat colonies are found in Rome around the Colosseum and Forum Romanum, with cats at some of these sites being fed and given medical attention by volunteers. Public attitudes towards feral cats vary widely, ranging from seeing them as free-ranging pets, to regarding them as vermin. One common approach to reducing the feral cat population is termed "trap-neuter-return", where the cats are trapped, neutered, immunized against diseases such as rabies and the feline panleukopenia and leukemia viruses, and then released. Before releasing them back into their feral colonies, the attending veterinarian often nips the tip off one ear to mark it as neutered and inoculated, since these cats may be trapped again. Volunteers continue to feed and give care to these cats throughout their lives. Given this support, their lifespans are increased, and behavior and nuisance problems caused by competition for food are reduced. Some feral cats can be successfully socialised and 're-tamed' for adoption; young cats, especially kittens and cats that have had prior experience and contact with humans are the most receptive to these efforts. Cats are common pets throughout the world, and their worldwide population exceeds 500 million as of 2007. As well as being kept as pets, cats are also used in the international fur and leather industries for making coats, hats, blankets, and stuffed toys; and shoes, gloves, and musical instruments respectively (about 24 cats are needed to make a cat-fur coat). This use has been outlawed in the United States, Australia, and the European Union in 2007. Cat pelts have been used for superstitious purposes as part of the practise of witchcraft, and are still made into blankets in Switzerland as folk remedies believed to help rheumatism. In the Western intellectual tradition, the idea of cats as everyday objects have served to illustrate problems of quantum mechanics in the Schrödinger's cat thought experiment. A few attempts to build a cat census have been made over the years, both through associations or national and international organizations (such as the Canadian Federation of Humane Societies's one) and over the Internet, but such a task does not seem simple to achieve. General estimates for the global population of domestic cats range widely from anywhere between 200 million to 600 million. Walter Chandoha made his career photographing cats after his 1949 images of "Loco", an especially charming stray taken in, were published around the world. He is reported to have photographed 90,000 cats during his career and maintained an archive of 225,000 images that he drew from for publications during his lifetime. A cat show is a judged event in which the owners of cats compete to win titles in various cat-registering organizations by entering their cats to be judged after a breed standard. Both pedigreed and non-purebred companion ("moggy") cats are admissible, although the rules differ from organization to organization. Competing cats are compared to the applicable breed standard, and assessed for temperament and apparent health; the owners of those judged to be most ideal awarded a prize. Moggies are judged based on their temperament and healthy appearance. Some events also include activity judging, such as trained navigation of obstacle course. Often, at the end of the year, all of the points accrued at various shows are added up and more national and regional titles are awarded to champion cats. Cats can be infected or infested with viruses, bacteria, fungus, protozoans, arthropods or worms that can transmit diseases to humans. In some cases, the cat exhibits no symptoms of the disease, However, the same disease can then become evident in a human. The likelihood that a person will become diseased depends on the age and immune status of the person. Humans who have cats living in their home or in close association are more likely to become infected, however, those who do not keep cats as pets might also acquire infections from cat feces and parasites exiting the cat's body. Some of the infections of most concern include salmonella, cat-scratch disease and toxoplasmosis. In ancient Egypt, cats were worshipped, and the goddess Bastet often depicted in cat form, sometimes taking on the war-like aspect of a lioness. The Greek historian Herodotus reported that killing a cat was forbidden, and when a household cat died, the entire family mourned and shaved their eyebrows. Families took their dead cats to the sacred city of Bubastis, where they were embalmed and buried in sacred repositories. Herodotus expressed astonishment at the domestic cats in Egypt, because he had only ever seen wildcats. Ancient Greeks and Romans kept weasels as pets, which were seen as the ideal rodent-killers. The earliest unmistakable evidence of the Greeks having domestic cats comes from two coins from Magna Graecia dating to the mid-fifth century BC showing Iokastos and Phalanthos, the legendary founders of Rhegion and Taras respectively, playing with their pet cats. The usual ancient Greek word for 'cat' was , meaning 'thing with the waving tail'. Cats are rarely mentioned in ancient Greek literature. Aristotle remarked in his "History of Animals" that "female cats are naturally lecherous." The Greeks later syncretized their own goddess Artemis with the Egyptian goddess Bastet, adopting Bastet's associations with cats and ascribing them to Artemis. In Ovid's "Metamorphoses", when the deities flee to Egypt and take animal forms, the goddess Diana turns into a cat. Cats eventually displaced ferrets as the pest control of choice because they were more pleasant to have around the house and were more enthusiastic hunters of mice. During the Middle Ages, many of Artemis's associations with cats were grafted onto the Virgin Mary. Cats are often shown in icons of Annunciation and of the Holy Family and, according to Italian folklore, on the same night that Mary gave birth to Jesus, a cat in Bethlehem gave birth to a kitten. Domestic cats were spread throughout much of the rest of the world during the Age of Discovery, as ships' cats were carried on sailing ships to control shipboard rodents and as good-luck charms. Several ancient religions believed cats are exalted souls, companions or guides for humans, that are all-knowing but mute so they cannot influence decisions made by humans. In Japan, the cat is a symbol of good fortune. In Norse mythology, Freyja, the goddess of love, beauty, and fertility, is depicted as riding a chariot drawn by cats. In Jewish legend, the first cat was living in the house of the first man Adam as a pet that got rid of mice. The cat was once partnering with the first dog before the latter broke an oath they had made which resulted in enmity between the descendants of these two animals. It is also written that neither cats nor foxes are represented in the water, while every other animal has an incarnation species in the water. Although no species are sacred in Islam, cats are revered by Muslims. Some Western writers have stated Muhammad had a favorite cat, "Muezza". He is reported to have loved cats so much, "he would do without his cloak rather than disturb one that was sleeping on it". The story has no origin in early Muslim writers, and seems to confuse a story of a later Sufi saint, Ahmed ar-Rifa'i, centuries after Muhammad. One of the companions of Muhammad was known as Abu Hurayrah ("father of the kitten"), in reference to his documented affection to cats. Many cultures have negative superstitions about cats. An example would be the belief that a black cat "crossing one's path" leads to bad luck, or that cats are witches' familiars used to augment a witch's powers and skills. The killing of cats in Medieval Ypres, Belgium, is commemorated in the innocuous present-day Kattenstoet (cat parade). In medieval France, cats would be burnt alive as a form of entertainment. According to Norman Davies, the assembled people "shrieked with laughter as the animals, howling with pain, were singed, roasted, and finally carbonized". "It was the custom to burn a basket, barrel, or sack full of live cats, which was hung from a tall mast in the midst of the bonfire; sometimes a fox was burned. The people collected the embers and ashes of the fire and took them home, believing that they brought good luck. The French kings often witnessed these spectacles and even lit the bonfire with their own hands. In 1648 Louis XIV, crowned with a wreath of roses and carrying a bunch of roses in his hand, kindled the fire, danced at it and partook of the banquet afterwards in the town hall. But this was the last occasion when a monarch presided at the midsummer bonfire in Paris. At Metz midsummer fires were lighted with great pomp on the esplanade, and a dozen cats, enclosed in wicker cages, were burned alive in them, to the amusement of the people. Similarly at Gap, in the department of the Hautes-Alpes, cats used to be roasted over the midsummer bonfire." According to a myth in many cultures, cats have multiple lives. In many countries, they are believed to have nine lives, but in Italy, Germany, Greece, Brazil and some Spanish-speaking regions, they are said to have seven lives, while in Turkish and Arabic traditions, the number of lives is six. The myth is attributed to the natural suppleness and swiftness cats exhibit to escape life-threatening situations. Also lending credence to this myth is the fact that falling cats often land on their feet, using an instinctive righting reflex to twist their bodies around. Nonetheless, cats can still be injured or killed by a high fall.
https://en.wikipedia.org/wiki?curid=6678