text
stringlengths
10
951k
source
stringlengths
39
44
Organelle In cell biology, an organelle is a specialized subunit, usually within a cell, that has a specific function. Organelles are either separately enclosed within their own lipid bilayers (also called membrane-bound organelles) or are spatially distinct functional units without a surrounding lipid bilayer (non-membrane bound organelles). Although most organelles are functional units within cells, some functional units that extend outside of cells are often termed organelles, such as cilia, the flagellum and archaellum, and the trichocyst. The name "organelle" comes from the idea that these structures are parts of cells, as organs are to the body, hence "organelle," the suffix "-elle" being a diminutive. Organelles are identified by microscopy, and can also be purified by cell fractionation. There are many types of organelles, particularly in eukaryotic cells. While prokaryotes do not possess intracellular organelles "per se", some do contain protein-based bacterial microcompartments, which are thought to act as primitive prokaryotic organelles. Also, the prokaryotic flagellum which protrudes outside the cell, and its motor, as well as the largely extracellular pilus, are often spoken of as organelles. In biology "organs" are defined as confined functional units within an organism. The analogy of bodily organs to microscopic cellular substructures is obvious, as from even early works, authors of respective textbooks rarely elaborate on the distinction between the two. In the 1830s, Félix Dujardin refuted Ehrenberg theory which said that microorganisms have the same organs of multicellular animals, only minor. Credited as the first to use a diminutive of "organ" (i.e., little organ) for cellular structures was German zoologist Karl August Möbius (1884), who used the term "organula" (plural of "organulum", the diminutive of Latin "organum"). In a footnote, which was published as a correction in the next issue of the journal, he justified his suggestion to call organs of unicellular organisms "organella" since they are only differently formed parts of one cell, in contrast to multicellular organs of multicellular organisms. While most cell biologists consider the term "organelle" to be synonymous with cell compartment, a space often bound by one or two lipid bilayers, some cell biologists choose to limit the term to include only those cell compartments that contain deoxyribonucleic acid (DNA), having originated from formerly autonomous microscopic organisms acquired via endosymbiosis. Under this definition, there would only be two broad classes of organelles (i.e. those that contain their own DNA, and have originated from endosymbiotic bacteria): Other organelles are also suggested to have endosymbiotic origins, but do not contain their own DNA (notably the flagellum – see evolution of flagella). A second, less restrictive definition of organelles is that they are membrane-bound structures. However, even by using this definition, some parts of the cell that have been shown to be distinct functional units do not qualify as organelles. Therefore, the use of organelle to also refer to non-membrane bound structures such as ribosomes is common and accepted. This has led many texts to delineate between membrane-bound and non-membrane bound organelles. The non-membrane bound organelles, also called large biomolecular complexes, are large assemblies of macromolecules that carry out particular and specialized functions, but they lack membrane boundaries. Many of these are referred to as "proteinaceous organelles" as there many structure is made of proteins. Such cell structures include: The mechanisms by which such non-membrane bound organelles form and retain their spatial integrity have been likened to liquid-liquid phase separation. Eukaryotic cells are structurally complex, and by definition are organized, in part, by interior compartments that are themselves enclosed by lipid membranes that resemble the outermost cell membrane. The larger organelles, such as the nucleus and vacuoles, are easily visible with the light microscope. They were among the first biological discoveries made after the invention of the microscope. Not all eukaryotic cells have each of the organelles listed below. Exceptional organisms have cells that do not include some organelles that might otherwise be considered universal to eukaryotes (such as mitochondria). There are also occasional exceptions to the number of membranes surrounding organelles, listed in the tables below (e.g., some that are listed as double-membrane are sometimes found with single or triple membranes). In addition, the number of individual organelles of each type found in a given cell varies depending upon the function of that cell. Mitochondria and plastids, including chloroplasts, have double membranes and their own DNA. According to the endosymbiotic theory, they are believed to have originated from incompletely consumed or invading prokaryotic organisms. Other related structures: Prokaryotes are not as structurally complex as eukaryotes, and were once thought not to have any internal structures enclosed by lipid membranes. In the past, they were often viewed as having little internal organization, but slowly, details are emerging about prokaryotic internal structures. An early false turn was the idea developed in the 1970s that bacteria might contain membrane folds termed mesosomes, but these were later shown to be artifacts produced by the chemicals used to prepare the cells for electron microscopy. However, more recent research has revealed that at least some prokaryotes have microcompartments such as carboxysomes. These subcellular compartments are 100–200 nm in diameter and are enclosed by a shell of proteins. Even more striking is the description of membrane-bound magnetosomes in bacteria, reported in 2006, as well as the nucleus-like structures of the "Planctomycetes" that are surrounded by lipid membranes, reported in 2005.
https://en.wikipedia.org/wiki?curid=22393
Ouida Ouida (; 1 January 1839 – 25 January 1908) was the pseudonym of the English novelist Maria Louise Ramé (although she preferred to be known as Marie Louise de la Ramée). During her career, Ouida wrote more than 40 novels, as well as short stories, children's books and essays. Moderately successful, she lived a life of luxury, entertaining many of the literary figures of the day. "Under Two Flags", one of her most famous novels, described the British in Algeria. It expressed sympathy for the French colonists—with whom Ouida deeply identified—and, to some extent, the Arabs. The novel was adapted for the stage, and was filmed six times. Her novel "A Dog of Flanders" is considered a children's classic in much of Asia. The American author Jack London cited her novel "Signa", as one of the reasons for his literary success. Her lavish lifestyle eventually led her to penury, and her works were put up for auction to pay her debts. She died in Italy from pneumonia. Soon after her death, her friends organized a public subscription in Bury St Edmunds, where they had a fountain for horses and dogs installed in her name. Maria Louise Ramé was born at Bury St Edmunds, England. Her mother, Susan Sutton, was a wine merchant's daughter; her father was from France. She derived her pen name from her own childish pronunciation of her given name "Louise". Her opinion of her birthplace fluctuated; she wrote:— "That clean, quiet antiquated town, that always puts me in the mind of an old maid dressed for a party; that lowest and dreariest of Boroughs, where the streets are as full of grass as an acre of pasture land. Why, the inhabitants are driven to ringing their own doorbells lest they rust from lack of use." She moved into the Langham Hotel, London in 1867. There, according to the hotel promotional materials, she wrote in bed, by candlelight, with the curtains drawn to keep out daylight and surrounded by purple flowers. She ran up huge hotel and florists bills of up to 200 pounds per week and commanded soirees that included soldiers, politicians, literary lights (including Oscar Wilde, Algernon Swinburne, Robert Browning and Wilkie Collins), and artists (including John Millais). Many of her stories and characters were based upon people she invited to her salons at The Langham. Ouida was described by William Allingham in his diary of 1872 as of short stature, with a "sinister, clever face" and with a "voice like a carving knife." For many years Ouida lived in London, but about 1871 she moved to Italy. In 1874, she settled permanently with her mother in Florence, and there long pursued her work as a novelist. At first she rented an 'apartment' at the Palazzo Vagnonville. Later she removed to the Villa Farinola at Scandicci, south of Bellosguardo, three miles from Florence, where she lived in great style, entertained largely, collected objets d'art, dressed expensively but not tastefully, drove good horses, and kept many dogs, to which she was deeply attached. She lived in Bagni di Lucca for a period, where there is a commemorative plaque on the outside wall. She declared that she never received from her publishers more than £1600. for any one novel, but that she found America 'a mine of wealth.' In 'The Massarenes' (1897) she gave a lurid picture of the parvenu millionaire in smart London society. This book was greatly prized by Ouida, and was very successful in terms of sales. Thenceforth she chiefly wrote for the leading magazines essays on social questions or literary criticisms, which were not remunerative. As before, she used her locations as inspiration for the setting and characters in her novels. The British and American colony in Florence was satirised in her novel, "Friendship" (1878). Ouida considered herself a serious artist. She was inspired by Byron in particular, and was interested in other artists of all kinds. Sympathetic descriptions of tragic painters and singers occurred in her later novels. Her work often combines romanticism with social criticism. In her novel, "Puck", a talking dog narrates his views on society. "Views and Opinions" includes essays in her own voice on a variety of social topics. She was an animal lover and rescuer, and at times owned as many as thirty dogs. Although successful, she did not manage her money well. A civil list pension of £150 a year was offered to her by the prime minister, Sir Henry Campbell-Bannerman, on the application of Alfred Austin, George Wyndham, and Walburga, Lady Paget, which she reluctantly accepted after request by her friend, Lady Howard of Glossop, on 16 July 1906. She continued to live in Italy until her death on 25 January 1908, at 70 Via Zanardelli, Viareggio, of pneumonia. She is buried in the English Cemetery in Bagni di Lucca, Italy. During her career, Ouida wrote more than 40 novels, children's books and collections of short stories and essays. Her work had several phases. In 1863, when she was 24, she published her first novel, "Held in Bondage". (She later claimed to have written her well-received novel "Idalia" (1867) at the age of 16. It featured a rebellious ingenue heroine who was sympathetic to Italian independence.) In her early period, her novels were considered "racy" and "swashbuckling", a contrast to "the moralistic prose of early Victorian literature" (Tom Steele), and a hybrid of the sensationalism of the 1860s and the proto-adventure novels being published as part of the romanticisation of imperial expansion. Later her work was more typical of historical romance, though she never stopped commenting on contemporary society. She also wrote several stories for children. "Under Two Flags", one of her most famous novels, described the British in Algeria. It expressed sympathy for the French colonists (called "pieds noirs")—with whom Ouida deeply identified—and, to some extent, the Arabs. The novel was adapted for the stage, and was filmed six times. The American author Jack London cited her novel "Signa", which he read at age eight, as one of the eight reasons for his literary success. The British composer Frederic Hymen Cowen and his librettists Gilbert Arthur à Beckett, H.A. Rudall, and Frederic Edward Weatherly acquired the rights to Ouida's 1875 novel "Signa" to create an opera for Richard D'Oyly Carte's Royal English Opera House to succeed Arthur Sullivan's "Ivanhoe" in 1891. Between Cowen not being ready with his work and the collapse of Carte's venture, Cowen eventually took his finished "Signa" to Italy with an Italian translation of the original English text by G.A. Mazzucato. After many delays and production troubles, Cowen's "Signa" was first performed in a reduced three-act version at the Teatro Dal Verme, Milan on 12 November 1893. After further revision and much cutting, it was later given in a two-act version at Covent Garden, London on 30 June 1894, at which point Cowen wondered if there was any sense left in the opera at all. Ouida's impression of the work is unknown. Later, Pietro Mascagni bought the rights for her story "Two Little Wooden Shoes", intending to adapt it for an opera. His friend Giacomo Puccini became interested in the story and began a court action, claiming that because Ouida was in debt, the rights to her works should be put up for public auction to raise funds for creditors. He won the court challenge and persuaded his publisher Ricordi to bid for the story. After Ricordi won, Puccini lost interest and never composed the opera. Mascagni later composed one based on the story, under the title "Lodoletta". Soon after her death, her friends organized a public subscription in Bury St Edmunds, where they had a fountain for horses and dogs installed in her name. Its inscription was composed by Lord Curzon: Fellow author "Rita" Humphreys (Eliza Margaret Jane Humphreys, 1850-1938) wrote a eulogy to Ouida and sent it to the press soon after her death. It was read at the unveiling of Ouida's memorial. During Rita's youth, Ouida had been popular but the girl was forbidden to read her. She made up for it later by purchasing every book written by Ouida and keeping them in her library for the rest of her life. Notes References Sources
https://en.wikipedia.org/wiki?curid=22400
Natural law Natural law (, "lex naturalis") is law as seen as being independent of, and pre-existent to, the positive law of any given political order, society or nation-state. Such genesis is seen as determined by nature (whether that reflects creation, evolution, or random chance), and a notional law of nature treated as objective fact that is universally applicable; that is, it exists and is recognizable, without any dependence on human understanding, or on the positive law of any given state, political order, or legislature — and even of society at large. Historically, natural law refers to the use of reason to analyze human nature in deducing binding rules of moral behavior, via (dominant or insurgent) accounts of observed and/or posited aspects of reality and of “the human condition“. The concept of natural law was documented in ancient Greek philosophy, including Aristotle, and was referred to in Roman philosophy by Cicero. References to it are also to be found in the Old and New Testaments of the Bible, and were later expounded upon in the Middle Ages by Christian philosophers such as Albert the Great and Thomas Aquinas. The School of Salamanca made notable contributions during the Renaissance. Modern natural law theories were greatly developed in the Age of Enlightenment, combining inspiration from Roman law with philosophies like social contract theory. It was used in challenging theory of the divine right of kings, and became an alternative justification for the establishment of a social contract, positive law, and government—and thus legal rights—in the form of classical republicanism. Conversely, the concept of natural rights is used by others to challenge the legitimacy of all such establishments. In the early decades of the 21st century, the concept of natural law is closely related to the concept of natural rights. Indeed, many philosophers, jurists and scholars use natural law synonymously with natural rights (), or natural justice, though others distinguish between natural law and natural right. Because of the intersection between natural law and natural rights, natural law has been claimed or attributed as a key component in the "Declaration of Independence" (1776) of the United States, the "Declaration of the Rights of Man and of the Citizen" (1789) of France, the "Universal Declaration of Human Rights" (1948) of the United Nations, as well as the "European Convention on Human Rights" (1953) of the Council of Europe. Although Plato did not have an explicit theory of natural law (he rarely used the phrase 'natural law' except in "Gorgias" 484 and "Timaeus" 83e), his concept of nature, according to John Wild, contains some of the elements found in many natural law theories. According to Plato, we live in an orderly universe. The basis of this orderly universe or nature are the forms, most fundamentally the Form of the Good, which Plato describes as "the brightest region of Being". The Form of the Good is the cause of all things, and when it is seen it leads a person to act wisely. In the "Symposium", the Good is closely identified with the Beautiful. In the "Symposium", Plato describes how the experience of the Beautiful by Socrates enabled him to resist the temptations of wealth and sex. In the "Republic", the ideal community is "a city which would be established in accordance with nature". Greek philosophy emphasized the distinction between "nature" ("physis", "φúσις") on the one hand and "law", "custom", or "convention" ("nomos", "νóμος") on the other. What the law commanded would be expected to vary from place to place, but what was "by nature" should be the same everywhere. A "law of nature" would therefore have the flavor more of a paradox than something that obviously existed. Against the conventionalism that the distinction between nature and custom could engender, Socrates and his philosophic heirs, Plato and Aristotle, posited the existence of natural justice or natural right ("dikaion physikon", "δίκαιον φυσικόν", Latin "ius naturale"). Of these, Aristotle is often said to be the father of natural law. Aristotle's association with natural law may be due to the interpretation given to his works by Thomas Aquinas. But whether Aquinas correctly read Aristotle is in dispute. According to some, Aquinas conflates natural law and natural right, the latter of which Aristotle posits in Book V of the "Nicomachean Ethics" (Book IV of the "Eudemian Ethics"). According to this interpretation, Aquinas's influence was such as to affect a number of early translations of these passages in an unfortunate manner, though more recent translations render those more literally. Aristotle notes that natural justice is a species of political justice, specifically the scheme of distributive and corrective justice that would be established under the best political community; were this to take the form of law, this could be called a natural law, though Aristotle does not discuss this and suggests in the "Politics" that the best regime may not rule by law at all. The best evidence of Aristotle's having thought there was a natural law comes from the "Rhetoric", where Aristotle notes that, aside from the "particular" laws that each people has set up for itself, there is a "common" law that is according to nature. Specifically, he quotes Sophocles and Empedocles: Universal law is the law of Nature. For there really is, as every one to some extent divines, a natural justice and injustice that is binding on all men, even on those who have no association or covenant with each other. It is this that Sophocles' Antigone clearly means when she says that the burial of Polyneices was a just act in spite of the prohibition: she means that it was just by nature: And so Empedocles, when he bids us kill no living creature, he is saying that to do this is not just for some people, while unjust for others: Some critics believe that the context of this remark suggests only that Aristotle advised that it could be rhetorically advantageous to appeal to such a law, especially when the "particular" law of one's own city was averse to the case being made, not that there actually was such a law; Moreover, they claim that Aristotle considered two of the three candidates for a universally valid, natural law provided in this passage to be wrong. Aristotle's paternity of natural law tradition is consequently disputed. The development of this tradition of natural justice into one of natural law is usually attributed to the Stoics. The rise of natural law as a universal system coincided with the rise of large empires and kingdoms in the Greek world. Whereas the "higher" law that Aristotle suggested one could appeal to was emphatically natural, in contradistinction to being the result of divine positive legislation, the Stoic natural law was indifferent to either the natural or divine source of the law: the Stoics asserted the existence of a rational and purposeful order to the universe (a divine or eternal law), and the means by which a rational being lived in accordance with this order was the natural law, which inspired actions that accorded with virtue. As the English historian A. J. Carlyle (1861–1943) notes: There is no change in political theory so startling in its completeness as the change from the theory of Aristotle to the later philosophical view represented by Cicero and Seneca ... We think that this cannot be better exemplified than with regard to the theory of the equality of human nature." Charles H. McIlwain likewise observes that "the idea of the equality of men is the most profound contribution of the Stoics to political thought" and that "its greatest influence is in the changed conception of law that in part resulted from it. Natural law first appeared among the stoics who believed that God is everywhere and in everyone (see classical pantheism). According to this belief, within humans there is a "divine spark" which helps them to live in accordance with nature. The stoics felt that there was a way in which the universe had been designed, and that natural law helped us to harmonise with this. Cicero wrote in his De Legibus that both justice and law originate from what nature has given to humanity, from what the human mind embraces, from the function of humanity, and from what serves to unite humanity. For Cicero, natural law obliges us to contribute to the general good of the larger society. The purpose of positive laws is to provide for "the safety of citizens, the preservation of states, and the tranquility and happiness of human life." In this view, "wicked and unjust statutes" are "anything but 'laws,'" because "in the very definition of the term 'law' there inheres the idea and principle of choosing what is just and true." Law, for Cicero, "ought to be a reformer of vice and an incentive to virtue." Cicero expressed the view that "the virtues which we ought to cultivate, always tend to our own happiness, and that the best means of promoting them consists in living with men in that perfect union and charity which are cemented by mutual benefits." In De Re Publica, he writes: Cicero influenced the discussion of natural law for many centuries to come, up through the era of the American Revolution. The jurisprudence of the Roman Empire was rooted in Cicero, who held "an extraordinary grip ... upon the imagination of posterity" as "the medium for the propagation of those ideas which informed the law and institutions of the empire." Cicero's conception of natural law "found its way to later centuries notably through the writings of Saint Isidore of Seville and the Decretum of Gratian." Thomas Aquinas, in his summary of medieval natural law, quoted Cicero's statement that "nature" and "custom" were the sources of a society's laws. The Renaissance Italian historian Leonardo Bruni praised Cicero as the person "who carried philosophy from Greece to Italy, and nourished it with the golden river of his eloquence." The legal culture of Elizabethan England, exemplified by Sir Edward Coke, was "steeped in Ciceronian rhetoric." The Scottish moral philosopher Francis Hutcheson, as a student at Glasgow, "was attracted most by Cicero, for whom he always professed the greatest admiration." More generally in eighteenth-century Great Britain, Cicero's name was a household word among educated people. Likewise, "in the admiration of early Americans Cicero took pride of place as orator, political theorist, stylist, and moralist." The British polemicist Thomas Gordon "incorporated Cicero into the radical ideological tradition that travelled from the mother country to the colonies in the course of the eighteenth century and decisively shaped early American political culture." Cicero's description of the immutable, eternal, and universal natural law was quoted by Burlamaqui and later by the American revolutionary legal scholar James Wilson. Cicero became John Adams's "foremost model of public service, republican virtue, and forensic eloquence." Adams wrote of Cicero that "as all the ages of the world have not produced a greater statesman and philosopher united in the same character, his authority should have great weight." Thomas Jefferson "first encountered Cicero as a schoolboy while learning Latin, and continued to read his letters and discourses throughout his life. He admired him as a patriot, valued his opinions as a moral philosopher, and there is little doubt that he looked upon Cicero's life, with his love of study and aristocratic country life, as a model for his own." Jefferson described Cicero as "the father of eloquence and philosophy." The New Testament carries a further exposition on the Abrahamic dialogue and links to the later Greek exposition on the subject, when Paul's Epistle to the Romans states: "For when the Gentiles, which have not the law, do by nature the things contained in the law, these, having not the law, are a law unto themselves: Which shew the work of the law written in their hearts, their conscience also bearing witness, and their thoughts the meanwhile accusing or else excusing one another." The intellectual historian A. J. Carlyle has commented on this passage, "There can be little doubt that St Paul's words imply some conception analogous to the 'natural law' in Cicero, a law written in men's hearts, recognized by man's reason, a law distinct from the positive law of any State, or from what St Paul recognized as the revealed law of God. It is in this sense that St Paul's words are taken by the Fathers of the fourth and fifth centuries like St Hilary of Poitiers, St Ambrose, and St Augustine, and there seems no reason to doubt the correctness of their interpretation." Because of its origins in the Old Testament, early Church Fathers, especially those in the West, saw natural law as part of the natural foundation of Christianity. The most notable among these was Augustine of Hippo, who equated natural law with humanity's prelapsarian state; as such, a life according to unbroken human nature was no longer possible and persons needed instead to seek healing and salvation through the divine law and grace of Jesus Christ. The natural law was inherently teleological, however, it is most assuredly not deontological. For Christians, natural law is how human beings manifest the divine image in their life. This mimicry of God's own life is impossible to accomplish except by means of the power of grace. Thus, whereas deontological systems merely require certain duties be performed, Christianity explicitly states that no one can, in fact, perform any duties if grace is lacking. For Christians, natural law flows not from divine commands, but from the fact that humanity is made in God's image, humanity is empowered by God's grace. Living the natural law is how humanity displays the gifts of life and grace, the gifts of all that is good. Consequences are in God's hands, consequences are generally not within human control, thus in natural law, actions are judged by three things: (1) the person's intent, (2) the circumstances of the act and (3) the nature of the act. The apparent good or evil consequence resulting from the moral act is not relevant to the act itself. The specific content of the natural law is therefore determined by how each person's acts mirror God's internal life of love. Insofar as one lives the natural law, temporal satisfaction may or may not be attained, but salvation will be attained. The state, in being bound by the natural law, is conceived as an institution whose purpose is to assist in bringing its subjects to true happiness. True happiness derives from living in harmony with the mind of God as an image of the living God. After the Protestant Reformation, some Protestant denominations maintained parts of the Catholic concept of natural law. The English theologian Richard Hooker from the Church of England adapted Thomistic notions of natural law to Anglicanism five principles: to live, to learn, to reproduce, to worship God, and to live in an ordered society. In the twelfth century, Gratian equated the natural law with divine law. Albertus Magnus would address the subject a century later, and his pupil, St. Thomas Aquinas, in his "Summa Theologica" I-II qq. 90–106, restored Natural Law to its independent state, asserting natural law as the rational creature's participation in the eternal law. Yet, since human reason could not fully comprehend the Eternal law, it needed to be supplemented by revealed Divine law. (See also Biblical law in Christianity.) Meanwhile, Aquinas taught that all human or positive laws were to be judged by their conformity to the natural law. An unjust law is not a law, in the full sense of the word. It retains merely the 'appearance' of law insofar as it is duly constituted and enforced in the same way a just law is, but is itself a 'perversion of law.' At this point, the natural law was not only used to pass judgment on the moral worth of various laws, but also to determine what those laws meant in the first place. This principle laid the seed for possible societal tension with reference to tyrants. The Catholic Church holds the view of natural law introduced by Albertus Magnus and elaborated by Thomas Aquinas, particularly in his "Summa Theologiae", and often as filtered through the School of Salamanca. This view is also shared by some Protestants, and was delineated by Anglican writer C. S. Lewis in his works "Mere Christianity" and "The Abolition of Man. The Catholic Church understands human beings to consist of body and mind, the physical and the non-physical (or soul perhaps), and that the two are inextricably linked. Humans are capable of discerning the difference between good and evil because they have a conscience. There are many manifestations of the good that we can pursue. Some, like procreation, are common to other animals, while others, like the pursuit of truth, are inclinations peculiar to the capacities of human beings. To know what is right, one must use one's reason and apply it to Thomas Aquinas' precepts. This reason is believed to be embodied, in its most abstract form, in the concept of a primary precept: "Good is to be sought, evil avoided." St. Thomas explains that: there belongs to the natural law, first, certain most general precepts, that are known to all; However, while the primary and immediate precepts cannot be "blotted out", the secondary precepts can be. Therefore, for a deontological ethical theory they are open to a surprisingly large amount of interpretation and flexibility. Any rule that helps humanity to live up to the primary or subsidiary precepts can be a secondary precept, for example: Natural moral law is concerned with both exterior and interior acts, also known as action and motive. Simply doing the right thing is not enough; to be truly moral one's motive must be right as well. For example, helping an old lady across the road (good exterior act) to impress someone (bad interior act) is wrong. However, good intentions don't always lead to good actions. The motive must coincide with the cardinal or theological virtues. Cardinal virtues are acquired through reason applied to nature; they are: The theological virtues are: According to Aquinas, to lack any of these virtues is to lack the ability to make a moral choice. For example, consider a person who possesses the virtues of justice, prudence, and fortitude, yet lacks temperance. Due to their lack of self-control and desire for pleasure, despite their good intentions, they will find themself swaying from the moral path. In the 16th century, the School of Salamanca (Francisco Suárez, Francisco de Vitoria, etc.) further developed a philosophy of natural law. Abū Rayhān al-Bīrūnī, a medieval scholar, scientist, and polymath, understood "natural law" as the survival of the fittest. He argued that the antagonism between human beings can be overcome only through a divine law, which he believed to have been sent through prophets. This is also said to be the general position of the Ashari school, the largest school of Sunni theology, as well as Ibn Hazm. Conceptualized thus, all "laws" are viewed as originating from subjective attitudes actuated by cultural conceptions and individual preferences, and so the notion of "divine revelation" is justified as some kind of "divine intervention" that replaces human positive laws, which are criticized as being relative, with a single divine positive law. This, however, also entails that anything may be included in "the divine law" as it would in "human laws", but unlike the latter, "God's law" is seen as binding regardless of the nature of the commands by virtue of "God's might": since God is not subject to human laws and conventions, He may command what He wills just as He may do what He wills. The Maturidi school, the second-largest school of Sunni theology, as well as the Mu'tazilites, posits the existence of a form of natural, or "objective," law that humans can comprehend. Abu Mansur al-Maturidi stated that the human mind could know of the existence of God and the major forms of "good" and "evil" without the help of revelation. Al-Maturidi gives the example of stealing, which, he believes, is known to be evil by reason alone due to people's working hard for their property. Similarly, killing, fornication, and drunkenness are all "discernible evils" that the human mind could know of according to al-Maturidi. Likewise, Averroes (Ibn Rushd), in his treatise on "Justice and Jihad" and his commentary on Plato's "Republic", writes that the human mind can know of the unlawfulness of killing and stealing and thus of the five maqasid or higher intents of the Islamic sharia, or the protection of religion, life, property, offspring, and reason. His Aristotelian commentaries also influenced the subsequent Averroist movement and the writings of Thomas Aquinas. Ibn Qayyim Al-Jawziyya also posited that human reason could discern between "great sins" and "good deeds". Nonetheless, he, like Ibn Taymiyah, emphasized the authority of "divine revelation" and asserted that it must be followed even if it "seems" to contradict human reason, though he stressed that most, if not all, of "God's commands" are both sensible (that is, rationalizable) and advantageous to humans in both "this life" and "the hereafter". The concept of "Istislah" in Islamic law bears some similarities to the natural law tradition in the West, as exemplified by Thomas Aquinas. However, whereas natural law deems good what is self-evidently good, according as it tends towards the fulfillment of the person, "istislah" typically calls good whatever is related to one of five "basic goods". Many jurists, theologians, and philosophers attempted to abstract these "basic and fundamental goods" from legal precepts. Al-Ghazali, for instance, defined them as religion, life, reason, lineage, and property, while others add "honor" also. Early Irish law, An Senchus Mor (The Great Tradition) mentions in a number of places "recht aicned" or natural law. This is a concept predating European legal theory, and reflects a type of law that is universal and may be determined by reason and observation of natural action. Neil McLeod identifies concepts that law must accord with: fír (truth) and dliged (right or entitlement). These two terms occur frequently, though Irish law never strictly defines them. Similarly, the term córus (law in accordance with proper order) occurs in some places, and even in the titles of certain texts. These were two very real concepts to the jurists and the value of a given judgment with respect to them was apparently ascertainable. McLeod has also suggested that most of the specific laws mentioned have passed the test of time and thus their truth has been confirmed, while other provisions are justified in other ways because they are younger and have not been tested over time The laws were written in the oldest dialect of the Irish language, called Bérla Féini [Bairla-faina], which even at the time was so difficult that persons about to become brehons had to be specially instructed in it, the length of time from beginning to becoming a learned Brehon was usually 20 years. Although under the law any third person could fulfill the duty if both parties agreed, and both were sane. It has been included in an Ethno-Celtic breakaway subculture, as it has religious undertones and freedom of religious expression allows it to once again be used as a valid system in Western Europe. Heinrich A. Rommen remarked upon "the tenacity with which the spirit of the English common law retained the conceptions of natural law and equity which it had assimilated during the Catholic Middle Ages, thanks especially to the influence of Henry de Bracton (d. 1268) and Sir John Fortescue (d. cir. 1476)." Bracton's translator notes that Bracton "was a trained jurist with the principles and distinctions of Roman jurisprudence firmly in mind"; but Bracton adapted such principles to English purposes rather than copying slavishly. In particular, Bracton turned the imperial Roman maxim that "the will of the prince is law" on its head, insisting that the king is "under" the law. The legal historian Charles F. Mullett has noted Bracton's "ethical definition of law, his recognition of justice, and finally his devotion to natural rights." Bracton considered justice to be the "fountain-head" from which "all rights arise." For his definition of justice, Bracton quoted the twelfth-century Italian jurist Azo: "'Justice is the constant and unfailing will to give to each his right.'" Bracton's work was the second legal treatise studied by the young apprentice lawyer Thomas Jefferson. Fortescue stressed "the supreme importance of the law of God and of nature" in works that "profoundly influenced the course of legal development in the following centuries." The legal scholar Ellis Sandoz has noted that "the historically ancient and the ontologically higher law—eternal, divine, natural—are woven together to compose a single harmonious texture in Fortescue's account of English law." As the legal historian Norman Doe explains: "Fortescue follows the general pattern set by Aquinas. The objective of every legislator is to dispose people to virtue. It is by means of law that this is accomplished. Fortescue's definition of law (also found in Accursius and Bracton), after all, was 'a sacred sanction commanding what is virtuous ["honesta"] and forbidding the contrary.'" Fortescue cited the great Italian Leonardo Bruni for his statement that "virtue alone produces happiness." Christopher St. Germain's "Doctor and Student" was a classic of English jurisprudence, and it was thoroughly annotated by Thomas Jefferson. St. Germain informs his readers that English lawyers generally don't use the phrase "law of nature", but rather use "reason" as the preferred synonym. Norman Doe notes that St. Germain's view "is essentially Thomist," quoting Thomas Aquinas's definition of law as "an ordinance of reason made for the common good by him who has charge of the community, and promulgated". Sir Edward Coke was the preeminent jurist of his time. Coke's preeminence extended across the ocean: "For the American revolutionary leaders, 'law' meant Sir Edward Coke's custom and right reason." Coke defined law as "perfect reason, which commands those things that are proper and necessary and which prohibits contrary things". For Coke, human nature determined the purpose of law; and law was superior to any one person's reason or will. Coke's discussion of natural law appears in his report of "Calvin's Case" (1608): "The law of nature is that which God at the time of creation of the nature of man infused into his heart, for his preservation and direction." In this case the judges found that "the ligeance or faith of the subject is due unto the King by the law of nature: secondly, that the law of nature is part of the law of England: thirdly, that the law of nature was before any judicial or municipal law: fourthly, that the law of nature is immutable." To support these findings, the assembled judges (as reported by Coke, who was one of them) cited as authorities Aristotle, Cicero, and the Apostle Paul; as well as Bracton, Fortescue, and St. Germain. After Coke, the most famous common law jurist of the seventeenth century is Sir Matthew Hale. Hale wrote a treatise on natural law that circulated among English lawyers in the eighteenth century and survives in three manuscript copies. This natural-law treatise has been published as "Of the Law of Nature" (2015). Hale's definition of the natural law reads: "It is the Law of Almighty God given by him to Man with his Nature discovering the morall good and moral evill of Moral Actions, commanding the former, and forbidding the latter by the secret voice or dictate of his implanted nature, his reason, and his concience." He viewed natural law as antecedent, preparatory, and subsequent to civil government, and stated that human law "cannot forbid what the Law of Nature injoins, nor Command what the Law of Nature prohibits." He cited as authorities Plato, Aristotle, Cicero, Seneca, Epictetus, and the Apostle Paul. He was critical of Hobbes's reduction of natural law to self-preservation and Hobbes's account of the state of nature, but drew positively on Hugo Grotius's "De jure belli ac pacis", Francisco Suárez's "Tractatus de legibus ac deo legislatore", and John Selden's "De jure naturali et gentium juxta disciplinam Ebraeorum". As early as the thirteenth century, it was held that "the law of nature...is the ground of all laws" and by the Chancellor and Judges that "it is required by the law of nature that every person, before he can be punish'd, ought to be present; and if absent by contumacy, he ought to be summoned and make default". Further, in 1824, we find it held that "proceedings in our Courts are founded upon the law of England, and that law is again founded upon the law of nature and the revealed law of God. If the right sought to be enforced is inconsistent with either of these, the English municipal courts cannot recognize it." By the 17th century, the medieval teleological view came under intense criticism from some quarters. Thomas Hobbes instead founded a contractarian theory of legal positivism on what all men could agree upon: what they sought (happiness) was subject to contention, but a broad consensus could form around what they feared (violent death at the hands of another). The natural law was how a rational human being, seeking to survive and prosper, would act. Natural law, therefore, was discovered by considering humankind's natural rights, whereas previously it could be said that natural rights were discovered by considering the natural law. In Hobbes' opinion, the only way natural law could prevail was for men to submit to the commands of the sovereign. Because the ultimate source of law now comes from the sovereign, and the sovereign's decisions need not be grounded in morality, legal positivism is born. Jeremy Bentham's modifications on legal positivism further developed the theory. As used by Thomas Hobbes in his treatises "Leviathan" and "De Cive", natural law is "a precept, or general rule, found out by reason, by which a man is forbidden to do that which is destructive of his life, or takes away the means of preserving the same; and to omit that by which he thinks it may best be preserved." According to Hobbes, there are nineteen Laws. The first two are expounded in chapter XIV of Leviathan ("of the first and second natural laws; and of contracts"); the others in chapter XV ("of other laws of nature"). Hobbes's philosophy includes a frontal assault on the founding principles of the earlier natural legal tradition, disregarding the traditional association of virtue with happiness, and likewise re-defining "law" to remove any notion of the promotion of the common good. Hobbes has no use for Aristotle's association of nature with human perfection, inverting Aristotle's use of the word "nature." Hobbes posits a primitive, unconnected state of nature in which men, having a "natural proclivity...to hurt each other" also have "a Right to every thing, even to one anothers body"; and "nothing can be Unjust" in this "warre of every man against every man" in which human life is "solitary, poore, nasty, brutish, and short." Rejecting Cicero's view that people join in society primarily through "a certain social spirit which nature has implanted in man," Hobbes declares that men join in society simply for the purpose of "getting themselves out from that miserable condition of Warre, which is necessarily consequent...to the naturall Passions of men, when there is no visible Power to keep them in awe." As part of his campaign against the classical idea of natural human sociability, Hobbes inverts that fundamental natural legal maxim, the Golden Rule. Hobbes's version is ""Do not that to another, which thou wouldst not have done to thy selfe."" The English cleric Richard Cumberland wrote a lengthy and influential attack on Hobbes's depiction of individual self-interest as the essential feature of human motivation. Historian Knud Haakonssen has noted that in the eighteenth century, Cumberland was commonly placed alongside Alberico Gentili, Hugo Grotius and Samuel Pufendorf "in the triumvirate of seventeenth-century founders of the 'modern' school of natural law." The eighteenth-century philosophers Shaftesbury and Hutcheson "were obviously inspired in part by Cumberland." Historian Jon Parkin likewise describes Cumberland's work as "one of the most important works of ethical and political theory of the seventeenth century." Parkin observes that much of Cumberland's material "is derived from Roman Stoicism, particularly from the work of Cicero, as "Cumberland deliberately cast his engagement with Hobbes in the mould of Cicero's debate between the Stoics, who believed that nature could provide an objective morality, and Epicureans, who argued that morality was human, conventional and self-interested." In doing so, Cumberland de-emphasized the overlay of Christian dogma (in particular, the doctrine of "original sin" and the corresponding presumption that humans are incapable of "perfecting" themselves without divine intervention) that had accreted to natural law in the Middle Ages. By way of contrast to Hobbes's multiplicity of laws, Cumberland states in the very first sentence of his "Treatise of the Laws of Nature" that "all the Laws of Nature are reduc'd to that one, of Benevolence toward all Rationals." He later clarifies: "By the name "Rationals" I beg leave to understand, as well "God" as "Man"; and I do it upon the Authority of Cicero." Cumberland argues that the mature development ("perfection") of human nature involves the individual human willing and acting for the common good. For Cumberland, human interdependence precludes Hobbes's natural right of each individual to wage war against all the rest for personal survival. However, Haakonssen warns against reading Cumberland as a proponent of "enlightened self-interest." Rather, the "proper moral love of humanity" is "a disinterested love of God through love of humanity in ourselves as well as others." Cumberland concludes that actions "principally conducive to our Happiness" are those that promote "the Honour and Glory of God" and also "Charity and Justice towards men." Cumberland emphasizes that desiring the well-being of our fellow humans is essential to the "pursuit of our own Happiness." He cites "reason" as the authority for his conclusion that happiness consists in "the most extensive Benevolence," but he also mentions as "Essential Ingredients of Happiness" the "Benevolent Affections," meaning "Love and Benevolence towards others," as well as "that Joy, which arises from their Happiness." The U.S. Declaration of Independence states that it has become necessary for the people of the United States to assume "the separate and equal station to which the Laws of Nature and of Nature's God entitle them". Some early American lawyers and judges perceived natural law as too tenuous, amorphous, and evanescent a legal basis for grounding concrete rights and governmental limitations. Natural law did, however, serve as authority for legal claims and rights in some judicial decisions, legislative acts, and legal pronouncements. Robert Lowry Clinton argues that the U.S. Constitution rests on a common law foundation and the common law, in turn, rests on a classical natural law foundation. Liberal natural law grew out of the medieval Christian natural law theories and out of Hobbes' revision of natural law, sometimes in an uneasy balance of the two. Sir Alberico Gentili and Hugo Grotius based their philosophies of international law on natural law. In particular, his writings on freedom of the seas and just war theory directly appealed to natural law. About natural law itself, he wrote that "even the will of an omnipotent being cannot change or abrogate" natural law, which "would maintain its objective validity even if we should assume the impossible, that there is no God or that he does not care for human affairs." ("De iure belli ac pacis", Prolegomeni XI). This is the famous argument "etiamsi daremus" ("non esse Deum"), that made natural law no longer dependent on theology. However, German church-historians Ernst Wolf and M. Elze disagreed and claimed that Grotius' concept of natural law did have a theological basis. In Grotius' view, the Old Testament contained moral precepts (e.g. the Decalogue) which Christ confirmed and therefore were still valid. Moreover, they were useful in explaining the content of natural law. Both biblical revelation and natural law originated in God and could therefore not contradict each other. In a similar way, Samuel Pufendorf gave natural law a theological foundation and applied it to his concepts of government and international law. John Locke incorporated natural law into many of his theories and philosophy, especially in "Two Treatises of Government". There is considerable debate about whether his conception of natural law was more akin to that of Aquinas (filtered through Richard Hooker) or Hobbes' radical reinterpretation, though the effect of Locke's understanding is usually phrased in terms of a revision of Hobbes upon Hobbesian contractarian grounds. Locke turned Hobbes' prescription around, saying that if the ruler went against natural law and failed to protect "life, liberty, and property," people could justifiably overthrow the existing state and create a new one. While Locke spoke in the language of natural law, the content of this law was by and large protective of natural rights, and it was this language that later liberal thinkers preferred. Political philosopher Jeremy Waldron has pointed out that Locke's political thought was based on "a particular set of Protestant Christian assumptions." To Locke, the content of natural law was identical with biblical ethics as laid down especially in the Decalogue, Christ's teaching and exemplary life, and St. Paul's admonitions. Locke derived the concept of basic human equality, including the equality of the sexes ("Adam and Eve"), from Genesis 1, 26–28, the starting-point of the theological doctrine of Imago Dei. One of the consequences is that as all humans are created equally free, governments need the consent of the governed. Thomas Jefferson, arguably echoing Locke, appealed to unalienable rights in the "Declaration of Independence", "We hold these truths to be self-evident, that all men are "created" equal, that they are endowed by their "Creator" with certain unalienable Rights, that among these are Life, Liberty and the pursuit of Happiness." The Lockean idea that governments need the consent of the governed was also fundamental to the Declaration of Independence, as the American Revolutionaries used it as justification for their separation from the British crown. The Belgian philosopher of law Frank van Dun is one among those who are elaborating a secular conception of natural law in the liberal tradition. Libertarian theorist Murray Rothbard argues that "the very existence of a natural law discoverable by reason is a potentially powerful threat to the status quo and a standing reproach to the reign of blindly traditional custom or the arbitrary will of the State apparatus." Ludwig von Mises states that he relaid the general sociological and economic foundations of the liberal doctrine upon utilitarianism, rather than natural law, but R. A. Gonce argues that "the reality of the argument constituting his system overwhelms his denial." Murray Rothbard, however, says that Gonce makes a lot of errors and distortions in the analysis of Mises's works, including making confusions about the term which Mises uses to refer to scientific laws, "laws of nature", saying it characterizes Mises as a natural law philosopher. David Gordon notes, "When most people speak of natural law, what they have in mind is the contention that morality can be derived from human nature. If human beings are rational animals of such-and-such a sort, then the moral virtues are...(filling in the blanks is the difficult part)." Economist and philosopher F. A. Hayek said that, originally, "the term 'natural' was used to describe an orderliness or regularity that was not the product of deliberate human will. Together with 'organism' it was one of the two terms generally understood to refer to the spontaneously grown in contrast to the invented or designed. Its use in this sense had been inherited from the stoic philosophy, had been revived in the twelfth century, and it was finally under its flag that the late Spanish Schoolmen developed the foundations of the genesis and functioning of spontaneously formed social institutions." The idea that 'natural' was "the product of designing reason" is a product of a seventeenth century rationalist reinterpretation of the law of nature. Luis Molina, for example, when referred to the 'natural' price, explained that it is "so called because 'it results from the thing itself without regard to laws and decrees, but is dependent on many circumstances which alter it, such as the sentiments of men, their estimation of different uses, often even in consequence of whims and pleasures". And even John Locke, when talking about the foundations of natural law and explaining what he thought when citing "reason", said: "By reason, however, I do not think is meant here that faculty of the understanding which forms traint of thought and deduces proofs, but certain definite principles of action from which spring all virtues and whatever is necessary for the proper moulding of morals." This anti-rationalist approach to human affairs, for Hayek, was the same which guided Scottish enlightenment thinkers, such as Adam Smith, David Hume and Adam Ferguson, to make their case for liberty. For them, no one can have the knowledge necessary to plan society, and this "natural" or "spontaneous" order of society shows how it can efficiently "plan" bottom-up. Also, the idea that law is just a product of deliberate design, denied by natural law and linked to legal positivism, can easily generate totalitarianism: "If law is wholly the product of deliberate design, whatever the designer decrees to be law is just by definition and unjust law becomes a contradiction in terms. The will of the duly authorized legislator is then wholly unfettered and guided solely by his concrete interests". This idea is wrong because law cannot be just a product of "reason": "no system of articulated law can be applied except within a framework of generally recognized but often unarticulated rules of justice". However, a secular critique of the natural law doctrine was stated by Pierre Charron in his "De la sagesse" (1601): "The sign of a natural law must be the universal respect in which it is held, for if there was anything that nature had truly commanded us to do, we would undoubtedly obey it universally: not only would every nation respect it, but every individual. Instead there is nothing in the world that is not subject to contradiction and dispute, nothing that is not rejected, not just by one nation, but by many; equally, there is nothing that is strange and (in the opinion of many) unnatural that is not approved in many countries, and authorized by their customs." One modern articulation of the concept of natural laws was given by Belina and Dzudzek: "By constant repetition, those practices develop into structures in the form of discourses which can become so natural that we abstract from their societal origins, that the latter are forgotten and seem to be natural laws." In jurisprudence, "natural law" can refer to the several doctrines: Whereas legal positivism would say that a law can be unjust without it being any less a law, a natural law jurisprudence would say that there is something legally deficient about an unjust norm. Besides utilitarianism and Kantianism, natural law jurisprudence has in common with virtue ethics that it is a live option for a first principles ethics theory in analytic philosophy. The concept of natural law was very important in the development of the English common law. In the struggles between Parliament and the monarch, Parliament often made reference to the Fundamental Laws of England, which were at times said to embody natural law principles since time immemorial and set limits on the power of the monarchy. According to William Blackstone, however, natural law might be useful in determining the content of the common law and in deciding cases of equity, but was not itself identical with the laws of England. Nonetheless, the implication of natural law in the common law tradition has meant that the great opponents of natural law and advocates of legal positivism, like Jeremy Bentham, have also been staunch critics of the common law. Natural law jurisprudence is currently undergoing a period of reformulation (as is legal positivism). The most prominent contemporary natural law jurist, Australian John Finnis, is based in Oxford, but there are also Americans Germain Grisez, Robert P. George, and Canadian Joseph Boyle and Brazilian Emídio Brasileiro. All have tried to construct a new version of natural law. The 19th-century anarchist and legal theorist, Lysander Spooner, was also a figure in the expression of modern natural law. "New Natural Law" as it is sometimes called, originated with Grisez. It focuses on "basic human goods", such as human life, knowledge, and aesthetic experience, which are self-evidently and intrinsically worthwhile, and states that these goods reveal themselves as being incommensurable with one another. The tensions between the natural law and the positive law have played, and continue to play, a key role in the development of international law.
https://en.wikipedia.org/wiki?curid=22063
Nestorianism Nestorianism is a Christian theological doctrine that upholds several distinctive teachings in the fields of Christology and Mariology. It opposes the concept of hypostatic union and emphasizes that the two natures (human and divine) of Jesus Christ were joined by will rather than personhood. This Christological position is defined as "radical dyophysitism". Nestorianism was named after Christian theologian Nestorius (386–450), Patriarch of Constantinople from 428 to 431, who was influenced by Christological teachings of Theodore of Mopsuestia at the School of Antioch. Nestorius' teachings brought him into conflict with other prominent church leaders, most notably Cyril of Alexandria, who criticized especially his rejection of the title "Theotokos" ("Mother of God") for Mary, the mother of Jesus, and issued 12 anathemas against him at a council in Rome in 430. Nestorius and his teachings were eventually condemned as heretical at the Council of Ephesus in 431, and again at the Council of Chalcedon in 451, which led to the Nestorian Schism; churches supporting Nestorian teachings broke with the rest of the Christian Church. Following that, many of Nestorius's supporters relocated to the Sasanian Empire, where they affiliated with the local Christian community, known as the Church of the East. However, some such as Sebastian Brock consider referring to the Assyrian Church of the East as the "Nestorian Church" to be "inappropriate and misleading." Nestorianism is a form of dyophysitism, differing from orthodox dyophysitism on several points, mainly by opposition to the concept of hypostatic union. It can be seen as the antithesis to Eutchyan Monophysitism, which emerged in reaction to Nestorianism. Where Nestorianism holds that Christ had two loosely united natures, divine and human, Monophysitism holds that he had but a single nature, his human nature being absorbed into his divinity. A brief definition of Nestorian Christology can be given as: "Jesus Christ, who is not identical with the Son but personally united with the Son, who lives in him, is one hypostasis and one nature: human." This contrasts with Nestorius' own teaching that the Word, which is eternal, and the Flesh, which is not, came together in a hypostatic union, 'Jesus Christ', Jesus thus being both fully man and God, of two "ousia" () but of one "prosopon". Both Nestorianism and Monophysitism were condemned as heretical at the Council of Chalcedon. Nestorianism was condemned as heresy at the Council of Ephesus. The Armenian Church rejected the Council of Chalcedon (451) because they believed Chalcedonian Definition was too similar to Nestorianism. The Persian Nestorian Church, on the other hand, supported the spread of Nestorianism in Persarmenia. The Armenian Church and other eastern churches saw the rise of Nestorianism as a threat to the independence of their Church. Peter the Iberian, a Georgian prince, also strongly opposed the Chalcedonian Creed. Thus, in 491, Catholicos Babken I of Armenia, along with the Albanian and Iberian bishops met in Vagharshapat and issued a condemnation of the Chalcedonian Definition. Nestorians held that the Council of Chalcedon proved the orthodoxy of their faith and had started persecuting non-Chalcedonian or Monophysite Syrian Christians during the reign of Peroz I. In response to pleas for assistance from the Syrian Church, Armenian prelates issued a letter addressed to Persian Christians reaffirming their condemnation of the Nestorianism as heresy. Following the exodus to Persia, scholars expanded on the teachings of Nestorius and his mentors, particularly after the relocation of the School of Edessa to the (then) Persian city of Nisibis (modern-day Nusaybin in Turkey) in 489, where it became known as the School of Nisibis. Nestorian monasteries propagating the teachings of the Nisbis school flourished in 6th century Persarmenia. Despite this initial Eastern expansion, the Nestorians' missionary success was eventually deterred. David J. Bosch observes, "By the end of the fourteenth century, however, the Nestorian and other churches—which at one time had dotted the landscape of all of Central and even parts of East Asia—were all but wiped out. Isolated pockets of Christianity survived only in India. The religious victors on the vast Central Asian mission field of the Nestorians were Islam and Buddhism". Nestorius developed his Christological views as an attempt to understand and explain rationally the incarnation of the divine Logos, the Second Person of the Holy Trinity as the man Jesus. He had studied at the School of Antioch where his mentor had been Theodore of Mopsuestia; Theodore and other Antioch theologians had long taught a literalist interpretation of the Bible and stressed the distinctiveness of the human and divine natures of Jesus. Nestorius took his Antiochene leanings with him when he was appointed Patriarch of Constantinople by Byzantine emperor Theodosius II in 428. Nestorius's teachings became the root of controversy when he publicly challenged the long-used title "Theotokos" ("God-Bearer") for Mary. He suggested that the title denied Christ's full humanity, arguing instead that Jesus had two persons (dyoprosopism), the divine Logos and the human Jesus. As a result of this prosopic duality, he proposed "Christotokos" (Bringer forth of Christ) as a more suitable title for Mary. Nestorius' opponents found his teaching too close to the heresy of adoptionism – the idea that Christ had been born a man who had later been "adopted" as God's son. Nestorius was especially criticized by Cyril, Patriarch of Alexandria, who argued that Nestorius's teachings undermined the unity of Christ's divine and human natures at the Incarnation. Some of Nestorius's opponents argued that he put too much emphasis on the human nature of Christ, and others debated that the difference that Nestorius implied between the human nature and the divine nature created a fracture in the singularity of Christ, thus creating two Christ figures. Nestorius himself always insisted that his views were orthodox, though they were deemed heretical at the Council of Ephesus in 431, leading to the Nestorian Schism, when churches supportive of Nestorius and the rest of the Christian Church separated. A more elaborate Nestorian theology developed from there, which came to see Christ as having two natures united, or hypostases, the divine Logos and the human Christ. However, this formulation was never adopted by all churches termed "Nestorian". Indeed, the modern Assyrian Church of the East, which reveres Nestorius, does not fully subscribe to Nestorian doctrine, though it does not employ the title "Theotokos". Nestorianism became a distinct sect following the Nestorian Schism, beginning in the 430s. Nestorius had come under fire from Western theologians, most notably Cyril of Alexandria. Cyril had both theological and political reasons for attacking Nestorius; on top of feeling that Nestorianism was an error against true belief, he also wanted to denigrate the head of a competing patriarchate. Cyril and Nestorius asked Pope Celestine I to weigh in on the matter. Celestine found that the title "Theotokos" was orthodox, and authorized Cyril to ask Nestorius to recant. Cyril, however, used the opportunity to further attack Nestorius, who pleaded with Emperor Theodosius II to call a council so that all grievances could be aired. In 431 Theodosius called the Council of Ephesus. However, the council ultimately sided with Cyril, who held that the Christ contained two natures in one divine person ("hypostasis", unity of subsistence), and that the Virgin Mary, conceiving and bearing this divine person, is truly called the Mother of God ("Theotokos", meaning, God-bearer). The council accused Nestorius of heresy, and deposed him as patriarch. Upon returning to his monastery in 436, he was banished to Upper Egypt. Nestorianism was officially anathematized, a ruling reiterated at the Council of Chalcedon in 451. However, a number of churches, particularly those associated with the School of Edessa, supported Nestorius – though not necessarily his doctrine – and broke with the churches of the West. Many of Nestorius' supporters relocated to the Sasanian Empire of Iran, home to a vibrant but persecuted Christian minority. In Upper Egypt, Nestorius wrote his "Book of Heraclides", responding to the two councils at Ephesus (431, 449). Persia had long been home to Christian communities that had been persecuted by the Zoroastrian majority, which had accused local Christians of pro-Roman leanings. In 424, the Church in Persia declared itself independent of the Byzantine Church and all other churches, in order to ward off allegations of foreign allegiance. Following the Nestorian Schism, the Persian Church increasingly aligned itself with the Nestorians, a measure encouraged by the Zoroastrian ruling class. The Persian Church became increasingly Nestorian in doctrine over the next decades, furthering the divide between Chalcedonian Christianity and the Nestorians. In 486 the Metropolitan of Nisibis, Barsauma, publicly accepted Nestorius' mentor, Theodore of Mopsuestia, as a spiritual authority. In 489 when the School of Edessa in Mesopotamia was closed by Byzantine Emperor Zeno for its Nestorian teachings, the school relocated to its original home of Nisibis, becoming again the School of Nisibis, leading to a wave of Nestorian immigration into Persia. The Persian patriarch Babai (497–502) reiterated and expanded upon the church's esteem for Theodore, solidifying the church's adoption of Nestorianism. Now firmly established in Iran, with centers in Nisibis, Ctesiphon, and Gundeshapur, and several "metropoleis", the Nestorian Persian Church began to branch out beyond the Sasanian Empire. However, through the sixth century, the church was frequently beset with internal strife and persecution by Zoroastrians. The infighting led to a schism, which lasted from 521 until around 539 when the issues were resolved. However, immediately afterward Roman-Persian conflict led to the persecution of the church by the Sassanid emperor Khosrow I; this ended in 545. The church survived these trials under the guidance of Patriarch Aba I, who had converted to Christianity from Zoroastrianism. The church emerged stronger after this period of ordeal, and increased missionary efforts farther afield. Missionaries established dioceses in the Arabian Peninsula and India (the Saint Thomas Christians). They made some advances in Egypt, despite the strong miaphysite presence there. Missionaries entered Central Asia and had significant success converting local Turkic tribes. Nestorian missionaries were firmly established in China during the early part of the Tang dynasty (618–907); the Chinese source known as the Nestorian Stele records a mission under a Persian proselyte named Alopen as introducing Nestorian Christianity to China in 635. Following the Arab conquest of Persia, completed in 644, the Persian Church became a "dhimmi" community under the Rashidun Caliphate. The church and its communities abroad grew larger under the Caliphate. By the 10th century it had 15 metropolitan sees within the Caliphate's territories, and another five elsewhere, including in China and India. After that time, however, Nestorianism went into decline. In a 1996 article published in the Bulletin of the John Rylands Library, Fellow of the British Academy Sebastian Brock wrote: "...the term 'Nestorian Church' has become the standard designation for the ancient oriental church which in the past called itself 'The Church of the East', but which today prefers the fuller title 'The Assyrian Church of the East'. Such a designation ['Nestorian Church'] is not only discourteous to modern members of this venerable church, but also--as this paper aims to show--both inappropriate and misleading." The Common Christological Declaration between the Catholic Church and the Assyrian Church of the East signed by Pope John Paul II and Mar Dinkha IV underlines the Chalcedonian Christological formulation as the expression of the common faith of these Churches and recognizes the legitimacy of the title "Theotokos". In a 2017 paper, Mar Awa Royel, Bishop of the Assyrian Church, stated the position of that church: "After the Council of Ephesus (431), when the Nestorius the patriarch of Constantinople was condemned for his views on the unity of the Godhead and the humanity in Christ, the Church of the East was branded as 'Nestorian' on account of its refusal to anathematize the patriarch."
https://en.wikipedia.org/wiki?curid=22065
Naomi Klein Naomi Klein (born May 8, 1970) is a Canadian author, social activist, and filmmaker known for her political analyses and criticism of corporate globalization and of capitalism. On a three-year appointment from September 2018, she is the Gloria Steinem Chair in Media, Culture, and Feminist Studies at Rutgers University. Klein first became known internationally for her book "No Logo" (1999); "The Take" (2004), a documentary film about Argentina's occupied factories, written by her, and directed by her husband Avi Lewis; and significantly for "The Shock Doctrine" (2007), a critical analysis of the history of neoliberal economics that was adapted into a six-minute companion film by Alfonso and Jonás Cuarón, as well as a feature-length documentary by Michael Winterbottom. Klein's "This Changes Everything: Capitalism vs. the Climate" (2014) was a "New York Times" Best Seller list non-fiction bestseller and the winner of the Hilary Weston Writers' Trust Prize for Nonfiction in its year. In 2016, Klein was awarded the Sydney Peace Prize for her activism on climate justice. Klein frequently appears on global and national lists of top influential thinkers, including the 2014 Thought Leaders ranking compiled by the Gottlieb Duttweiler Institute, "Prospect" magazine's world thinkers 2014 poll, and Maclean's 2014 Power List. She is a member of the board of directors of the climate activist group 350.org. Naomi Klein was born in Montreal, Quebec, and brought up in a Jewish family with a history of peace activism. Her parents were self-described "hippies" who emigrated from the United States in 1967 as war resisters to the Vietnam War. Her mother, documentary film-maker Bonnie Sherr Klein, is best known for her anti-pornography film "". Her father, Michael Klein, is a physician and a member of Physicians for Social Responsibility. Her brother, Seth Klein, is director of the British Columbian office of the Canadian Centre for Policy Alternatives. Before World War II, her paternal grandparents were communists, but they began to turn against the Soviet Union after the Molotov–Ribbentrop Pact in 1939. In 1942, her grandfather, an animator at Disney, was fired after the 1941 strike, and had to switch to working in a shipyard instead. By 1956, they had abandoned communism. Klein's father grew up surrounded by ideas of social justice and racial equality, but found it "difficult and frightening to be the child of Communists", a so-called red diaper baby. Klein's husband, Avi Lewis, was born into a well-connected political and journalistic family; he works as a TV journalist and documentary filmmaker. The couple's only child, son Toma, was born on June 13, 2012. Klein spent much of her teenage years in shopping malls, obsessed with designer labels. As a child and teenager, she found it "very oppressive to have a very public feminist mother" and she rejected politics, instead embracing "full-on consumerism". She has attributed her change in worldview to two catalysts. One was when she was 17 and preparing for the University of Toronto, her mother had a stroke and became severely disabled. Naomi, her father, and her brother took care of Bonnie through the period in hospital and at home, making educational sacrifices to do so. That year off prevented her "from being such a brat". The next year, after beginning her studies at the University of Toronto, the second catalyst occurred: the 1989 École Polytechnique massacre of female engineering students, which proved to be a wake-up call to feminism. Klein's writing career began with contributions to "The Varsity", a student newspaper, where she served as editor-in-chief. After her third year at the University of Toronto, she dropped out of university to take a job at "The Globe and Mail", followed by an editorship at "This Magazine". In 1995, she returned to the University of Toronto with the intention of finishing her degree but left academia for a journalism internship before acquiring the final credits required to complete her degree. In 1999, Klein published the book "No Logo", which for many became a manifesto of the anti-globalization movement. In it, she attacks brand-oriented consumer culture and the operations of large corporations. She also accuses several such corporations of unethically exploiting workers in the world's poorest countries in pursuit of greater profits. In this book, Klein criticized Nike so severely that Nike published a point-by-point response. "No Logo" became an international bestseller, selling over one million copies in over 28 languages. Klein's "Fences and Windows" (2002) is a collection of her articles and speeches written on behalf of the anti-globalization movement (all proceeds from the book go to benefit activist organizations through The Fences and Windows Fund). "The Take" (2004), a documentary film collaboration by Klein and Lewis, concerns factory workers in Argentina who took over a closed plant and resumed production, operating as a collective. The first African screening was in the Kennedy Road shack settlement in the South African city of Durban, where the Abahlali baseMjondolo movement began. An article in "Z Communications" criticized "The Take" for its portrayal of the Argentine general and politician Juan Domingo Perón arguing that he was falsely portrayed as a social democrat. Klein's third book, "The Shock Doctrine: The Rise of Disaster Capitalism", was published on September 4, 2007. The book argues that the free market policies of Nobel Laureate Milton Friedman and the Chicago School of Economics have risen to prominence in countries such as Chile, under Pinochet, Poland, Russia, under Yeltsin. The book also argues that policy initiatives (for instance, the privatization of Iraq's economy under the Coalition Provisional Authority) were rushed through while the citizens of these countries were in shock from disasters, upheavals, or invasion. The book became an international and "New York Times" bestseller and was translated into 28 languages. Central to the book's thesis is the contention that those who wish to implement unpopular free market policies now routinely do so by taking advantage of certain features of the aftermath of major disasters, be they economic, political, military or natural. The suggestion is that when a society experiences a major 'shock' there is a widespread desire for a rapid and decisive response to correct the situation; this desire for bold and immediate action provides an opportunity for unscrupulous actors to implement policies which go far beyond a legitimate response to disaster. The book suggests that when the rush to act means the specifics of a response will go unscrutinized, that is the moment when unpopular and unrelated policies will intentionally be rushed into effect. The book appears to claim that these shocks are in some cases intentionally encouraged or even manufactured. Klein identifies the "shock doctrine", elaborating on Joseph Schumpeter, as the latest in capitalism's phases of "creative destruction". "The Shock Doctrine" was adapted into a short film of the same name, released onto YouTube. The original is no longer available on the site, however, a duplicate was published in 2008. The film was directed by Jonás Cuarón, produced and co-written by his father Alfonso Cuarón. The original video was viewed over one million times. The publication of "The Shock Doctrine" increased Klein's prominence, with "The New Yorker" judging her "the most visible and influential figure on the American left—what Howard Zinn and Noam Chomsky were thirty years ago." On February 24, 2009, the book was awarded the inaugural Warwick Prize for Writing from the University of Warwick in England. The prize carried a cash award of £50,000. Klein's fourth book, "This Changes Everything: Capitalism vs. the Climate" was published in September 2014. The book puts forth the argument that the hegemony of neoliberal market fundamentalism is blocking any serious reforms to halt climate change and protect the environment. Questioned about Klein's claim that capitalism and controlling climate change were incompatible, Benoit Blarel, manager of the Environment and Natural Resources global practice at the World Bank, said that the write-off of fossil fuels necessary to control climate change "will have a huge impact all over" and that the World Bank was "starting work on this". The book won the 2014 Hilary Weston Writers' Trust Prize for Nonfiction, and was a shortlisted nominee for the 2015 Shaughnessy Cohen Prize for Political Writing. Klein's fifth book, "No Is Not Enough: Resisting Trump's Shock Politics and Winning the World We Need" was published in June 2017. It has also been published Internationally with the alternative subtitle "Defeating the New Shock Politics". Released in June 2018 as paperback and e-book, "The Battle For Paradise: Puerto Rico Takes on the Disaster Capitalists" covers what San Juan Mayor Carmen Yulín Cruz refers to as the post-Hurricane Maria unmasked colonialism leading to inequality and "creating a fierce humanitarian crisis." In April 2019, Simon & Schuster announced they would be publishing Klein's seventh book, "On Fire: The (Burning) Case for a Green New Deal", which was published on September 17, 2019. "On Fire" is a collection of essays focusing on climate change and the urgent actions needed to preserve the world. Klein relates her meeting with Greta Thunberg in the opening essay in which she discusses the entrance of young people into those speaking out for climate awareness and change. She supports the Green New Deal throughout the book and in the final essay she discusses the 2020 U.S. election stating: "The stakes of the election are almost unbearably high. It’s why I wrote the book and decided to put it out now and why I’ll be doing whatever I can to help push people toward supporting a candidate with the most ambitious Green New Deal platform—so that they win the primaries and then the general." Klein has written about the Iraq War. In "Baghdad Year Zero" ("Harper's Magazine", September 2004), Klein argues that, contrary to popular belief, the George W. Bush administration "did" have a clear plan for post-invasion Iraq: to build a completely unconstrained free market economy. She describes plans to allow foreigners to extract wealth from Iraq and the methods used to achieve those goals. Her "Baghdad Year Zero" was one of the inspirations for the 2008 film "War, Inc." Klein's "Bring Najaf to New York" ("The Nation", August 2004) argued that Muqtada Al Sadr's Mahdi Army "represents the overwhelmingly mainstream sentiment in Iraq" and that, if he were elected, "Sadr would try to turn Iraq into a theocracy like Iran," although his immediate demands were for "direct elections and an end to foreign occupation". Marc Cooper, a former "Nation" columnist, attacked the assertion that Al Sadr represented mainstream Iraqi sentiment and that American forces had brought the war to the holy city of Najaf. "Klein should know better," he wrote. "All enemies of the U.S. occupation she opposes are not her friends. Or ours. Or those of the Iraqi people. I don’t think that Mullah Al Sadr, in any case, is much desirous of support issuing from secular Jewish feminist-socialists." Klein signed a 2004 petition entitled, “We would vote for Hugo Chávez.” In 2007, she described Venezuela under the Chávez government as a country where "citizens had renewed their faith in the power of democracy to improve their lives," and described Venezuela as a place sheltered by Chávez's policies from the economic shocks produced by capitalism. Rather, according to Klein, Chávez protected his country from financial crisis by building “a zone of relative economic calm and predictability.” According to reviewer Todd Gitlin, who described the overall argument of Klein's book "The Shock Doctrine" (2007) as "more right than wrong," Klein is "a romantic," who expected that the Chávez government would produce a bright future in which worker-controlled co-operatives would run the economy. "The Shock Doctrine" was consistent with her prior thinking about globalization, and in that book she describes Chávez' policies as an example of public control of some sectors of the economy as protecting poor people from harm caused by globalization. After the collapse of the Venezuelan economy and alleged erosion of its democratic institutions under Chávez' successor Nicolás Maduro, Klein and other people who had supported Chávez were criticized by writers such as James Kirchick and Mark Milke. In March 2008, Klein was the keynote speaker at the first national conference of the Alliance of Concerned Jewish Canadians. In January 2009, during the Gaza War, Klein supported the Boycott, Divestment and Sanctions (BDS) campaign against Israel, arguing that "the best strategy to end the increasingly bloody occupation is for Israel to become the target of the kind of global movement that put an end to apartheid in South Africa." In summer 2009, on the occasion of the publication of the Hebrew translation of her book "The Shock Doctrine", Klein visited Israel, the West Bank, and Gaza, combining the promotion of her book and the BDS campaign. In an interview to the Israeli newspaper "Haaretz" she emphasized that it is important to her "not to boycott Israelis but rather to boycott the normalization of Israel and the conflict." In a speech in Ramallah on June 27, she apologized to the Palestinians for not joining the BDS campaign earlier. Her remarks, particularly that "[Some Jews] even think we get one get-away-with-genocide-free card" were characterized by Noam Schimmel, an op-ed columnist in "The Jerusalem Post", as "violent" and "unethical", and as the "most perverse of aspersions on Jews, an age-old stereotype of Jews as intrinsically evil and malicious." Klein was also a spokesperson for the protest against the spotlight on Tel Aviv at the 2009 Toronto International Film Festival, a spotlight that Klein said was a very selective and misleading portrait of Israel. Since 2009, Klein's attention has turned to environmentalism, with particular focus on climate change, the subject of her book "This Changes Everything" (2014). According to her website, the book and its accompanying film (released in 2015) will be about "how the climate crisis can spur economic and political transformation." She sits on the board of directors of campaign group 350.org and took part in their "Do the Math" tour in 2013, encouraging a divestment movement. She has encouraged the Occupy movement to join forces with the environmental movement, saying the financial crisis and the climate crisis have the same root—unrestrained corporate greed. She gave a speech at Occupy Wall Street where she described the world as "upside down", where we act as if "there is no end to what is actually finite—fossil fuels and the atmospheric space to absorb their emissions," and as if there are "limits to what is actually bountiful—the financial resources to build the kind of society we need." She has been a particularly vocal critic of the Athabasca oil sands in Alberta, describing it in a TED talk as a form of "terrestrial skinning." On September 2, 2011, she attended the demonstration against the Keystone XL pipeline outside the White House and was arrested. Klein celebrated Obama's decision to postpone a decision on the Keystone pipeline until 2013 pending an environmental review as a victory for the environmental movement. She attended the Copenhagen Climate Summit of 2009. She put the blame for the failure of Copenhagen on President Barack Obama, and described her own country, Canada, as a "climate criminal." She presented the Angry Mermaid Award (a satirical award designed to recognise the corporations who have best sabotaged the climate negotiations) to Monsanto. Writing in the wake of Hurricane Sandy, she warned that the climate crisis constitutes a massive opportunity for disaster capitalists and corporations seeking to profit from crisis. But equally, the climate crisis "can be a historic moment to usher in the next great wave of progressive change," or a so-called "People's Shock." On November 9, 2016, following the election of Donald Trump as the 45th President of the United States, Klein called for an international campaign to impose economic sanctions on the United States if his administration refuses to abide by the terms of the Paris Agreement. Klein contributes to "The Nation", "In These Times", "The Globe and Mail", "This Magazine", "Harper's Magazine", and "The Guardian", and is a senior contributor for "The Intercept". She is a former Miliband Fellow and lectured at the London School of Economics on the anti-globalization movement. Her appointment as the inaugural Gloria Steinem Endowed Chair in Media, Culture and Feminist Studies at Rutgers University–New Brunswick began in October 2018 and runs for 3 years. The position is funded by foundations, endowments and individuals. Klein ranked 11th in an internet poll of the top global intellectuals of 2005, a list of the world's top 100 public intellectuals compiled by the "Prospect" magazine in conjunction with "Foreign Policy" magazine. She was involved in 2010 G-20 Toronto summit protests, condemning police force and brutality. She spoke to a rally seeking the release of protesters in front of police headquarters on June 28, 2010. On October 6, 2011, she visited Occupy Wall Street and gave a speech declaring the protest movement "the most important thing in the world". On November 10, 2011, she participated in a panel discussion about the future of Occupy Wall Street with four other panelists, including Michael Moore, William Greider, and Rinku Sen, in which she stressed the crucial nature of the evolving movement. Klein also made an appearance in the British radio show "Desert Island Discs" on BBC Radio 4 in 2017. In November 2017, the Democracy in Europe Movement 2025 announced that Klein had been appointed to their Advisory Panel. In November 2019, along with other public figures, Klein signed a letter supporting Labour Party leader Jeremy Corbyn describing him as "a beacon of hope in the struggle against emergent far-right nationalism, xenophobia and racism in much of the democratic world" and endorsed him in the 2019 UK general election.
https://en.wikipedia.org/wiki?curid=22068
Nonsteroidal anti-inflammatory drug Nonsteroidal anti-inflammatory drugs (NSAIDs) are members of a drug class that reduces pain, decreases fever, prevents blood clots, and in higher doses, decreases inflammation. Side effects depend on the specific drug but largely include an increased risk of gastrointestinal ulcers and bleeds, heart attack, and kidney disease. The term "nonsteroidal" distinguishes these drugs from steroids, which while having a similar eicosanoid-depressing, anti-inflammatory action, have a broad range of other effects. First used in 1960, the term served to distance these medications from steroids, which were particularly stigmatised at the time due to the connotations with anabolic steroid abuse. NSAIDs work by inhibiting the activity of cyclooxygenase enzymes (COX-1 or COX-2). In cells, these enzymes are involved in the synthesis of key biological mediators, namely prostaglandins, which are involved in inflammation, and thromboxanes, which are involved in blood clotting. There are two types of NSAIDs available: non-selective and COX-2 selective. Most NSAIDs are non-selective and inhibit the activity of both COX-1 and COX-2. These NSAIDs, while reducing inflammation, also inhibit platelet aggregation (especially aspirin) and increase the risk of gastrointestinal ulcers/bleeds. COX-2 selective inhibitors have less gastrointestinal side effects but promote thrombosis and substantially increase the risk of heart attack. As a result, COX-2 selective inhibitors are generally contraindicated due to the high risk of undiagnosed vascular disease. These differential effects are due to the different roles and tissue localisations of each COX isoenzyme. By inhibiting physiological COX activity, all NSAIDs increase the risk of kidney disease and through a related mechanism, heart attack. In addition, NSAIDS can blunt the production of erythropoietin resulting in anaemia, since haemoglobin needs this hormone to be produced. Prolonged use is dangerous and case studies have shown the health risk with celecoxib. The most prominent NSAIDs are aspirin, ibuprofen, and naproxen, all available over the counter (OTC) in most countries. Paracetamol (acetaminophen) is generally not considered an NSAID because it has only minor anti-inflammatory activity. It treats pain mainly by blocking COX-2 and inhibiting endocannabinoid reuptake almost exclusively within the brain, but not much in the rest of the body. NSAIDs are usually used for the treatment of acute or chronic conditions where pain and inflammation are present. NSAIDs are generally used for the symptomatic relief of the following conditions: Aspirin, the only NSAID able to irreversibly inhibit COX-1, is also indicated for antithrombosis through inhibition of platelet aggregation. This is useful for the management of arterial thrombosis and prevention of adverse cardiovascular events like heart attacks. Aspirin inhibits platelet aggregation by inhibiting the action of thromboxane A2. In a more specific application, the reduction in prostaglandins is used to close a patent ductus arteriosus in neonates if it has not done so physiologically after 24 hours. NSAIDs are useful in the management of post-operative dental pain following invasive dental procedures such as dental extraction. When not contra-indicated they are favoured over the use of paracetamol alone due to the anti-inflammatory effect they provide. When used in combination with paracetamol the analgesic effect has been proven to be improved. There is weak evidence suggesting that taking pre-operative analgesia can reduce the length of post operative pain associated with placing orthodontic spacers under local anaesthetic. Combination of NSAIDs with pregabalin as preemptive analgesia has shown promising results for decreasing post operative pain intensity. The effectiveness of NSAIDs for treating non-cancer chronic pain and cancer-related pain in children and adolescents is not clear. There have not been sufficient numbers of high-quality randomized controlled trials conducted. Differences in anti-inflammatory activity between NSAIDs are small, but there is considerable variation in individual response and tolerance to these drugs. About 60% of patients will respond to any NSAID; of the others, those who do not respond to one may well respond to another. Pain relief starts soon after taking the first dose and a full analgesic effect should normally be obtained within a week, whereas an anti-inflammatory effect may not be achieved (or may not be clinically assessable) for up to 3 weeks. If appropriate responses are not obtained within these times, another NSAID should be tried. NSAIDs may be used with caution by people with the following conditions: NSAIDs should usually be avoided by people with the following conditions: The widespread use of NSAIDs has meant that the adverse effects of these drugs have become increasingly common. Use of NSAIDs increases risk of a range of gastrointestinal (GI) problems, kidney disease and adverse cardiovascular events. As commonly used for post-operative pain, there is evidence of increased risk of kidney complications. Their use following gastrointestinal surgery remains controversial, given mixed evidence of increased risk of leakage from any bowel anastomosis created. An estimated 10–20% of NSAID patients experience indigestion. In the 1990s high doses of prescription NSAIDs were associated with serious upper gastrointestinal adverse events, including bleeding. Over the past decade, deaths associated with gastric bleeding have declined. NSAIDs, like all medications, may interact with other medications. For example, concurrent use of NSAIDs and quinolones may increase the risk of quinolones' adverse central nervous system effects, including seizure. There is an argument over the benefits and risks of NSAIDs for treating chronic musculoskeletal pain. Each drug has a benefit-risk profile and balancing the risk of no treatment with the competing potential risks of various therapies is the clinician's responsibility. If a COX-2 inhibitor is taken, a traditional NSAID (prescription or over-the-counter) should not be taken at the same time. In addition, people on daily aspirin therapy (e.g., for reducing cardiovascular risk) must be careful if they also use other NSAIDs, as these may inhibit the cardioprotective effects of aspirin. Rofecoxib (Vioxx) was shown to produce significantly fewer gastrointestinal adverse drug reactions (ADRs) compared with naproxen. The study, the VIGOR trial, raised the issue of the cardiovascular safety of the coxibs. A statistically significant increase in the incidence of myocardial infarctions was observed in patients on rofecoxib. Further data, from the APPROVe trial, showed a statistically significant relative risk of cardiovascular events of 1.97 versus placebo—which caused a worldwide withdrawal of rofecoxib in October 2004. Use of methotrexate together with NSAIDS in rheumatoid arthritis is safe, if adequate monitoring is done. NSAIDs, aside from aspirin, increase the risk of myocardial infarction and stroke. This occurs at least within a week of use. They are not recommended in those who have had a previous heart attack as they increase the risk of death or recurrent MI. Evidence indicates that naproxen may be the least harmful out of these. NSAIDs aside from (low-dose) aspirin are associated with a doubled risk of heart failure in people without a history of cardiac disease. In people with such a history, use of NSAIDs (aside from low-dose aspirin) was associated with a more than 10-fold increase in heart failure. If this link is proven causal, researchers estimate that NSAIDs would be responsible for up to 20 percent of hospital admissions for congestive heart failure. In people with heart failure, NSAIDs increase mortality risk (hazard ratio) by approximately 1.2–1.3 for naproxen and ibuprofen, 1.7 for rofecoxib and celecoxib, and 2.1 for diclofenac. On 9 July 2015, the Food and Drug Administration (FDA) toughened warnings of increased heart attack and stroke risk associated with nonsteroidal anti-inflammatory drugs (NSAIDs) "other than aspirin". A 2005 Finnish study linked long term (over 3 months) use of NSAIDs with an increased risk of erectile dysfunction. The study was correlational only, and depended solely on self-reports (questionnaires). A 2011 publication in "The Journal of Urology" received widespread publicity. According to the study, men who used NSAIDs regularly were at significantly increased risk of erectile dysfunction. A link between NSAID use and erectile dysfunction still existed after controlling for several conditions. However, the study was observational and not controlled, with low original participation rate, potential participation bias, and other uncontrolled factors. The authors warned against drawing any conclusion regarding cause. The main adverse drug reactions (ADRs) associated with NSAID use relate to direct and indirect irritation of the gastrointestinal (GI) tract. NSAIDs cause a dual assault on the GI tract: the acidic molecules directly irritate the gastric mucosa, and inhibition of COX-1 and COX-2 reduces the levels of protective prostaglandins. Inhibition of prostaglandin synthesis in the GI tract causes increased gastric acid secretion, diminished bicarbonate secretion, diminished mucus secretion and diminished trophic effects on the epithelial mucosa. Common gastrointestinal side effects include: Clinical NSAID ulcers are related to the systemic effects of NSAID administration. Such damage occurs irrespective of the route of administration of the NSAID (e.g., oral, rectal, or parenteral) and can occur even in people who have achlorhydria. Ulceration risk increases with therapy duration, and with higher doses. To minimize GI side effects, it is prudent to use the lowest effective dose for the shortest period of time—a practice that studies show is often not followed. Over 50% of patients who take NSAIDs have sustained some mucosal damage to their small intestine. The risk and rate of gastric adverse effects is different depending on the type of NSAID medication a person is taking. Indomethacin, ketoprofen, and piroxicam use appear to lead to the highest rate of gastric adverse effects, while ibuprofen (lower doses) and diclofenac appear to have lower rates. Certain NSAIDs, such as aspirin, have been marketed in enteric-coated formulations that manufacturers claim reduce the incidence of gastrointestinal ADRs. Similarly, some believe that rectal formulations may reduce gastrointestinal ADRs. However, consistent with the systemic mechanism of such ADRs, and in clinical practice, these formulations have not demonstrated a reduced risk of GI ulceration. Numerous "gastro-protective" drugs have been developed with the goal of preventing gastrointestinal toxicity in people who need to take NSAIDS on a regular basis. Gastric adverse effects may be reduced by taking medications that suppress acid production such as proton pump inhibitors (e.g.: omeprazole and esomeprazole), or by treatment with a drug that mimics prostaglandin in order to restore the lining of the GI tract (e.g.: a prostaglandin analog misoprostol). Diarrhea is a common side effect of misoprostol, however, higher doses of misoprostol have been shown to reduce the risk of a person having a complication related to a gastric ulcer while taking NSAIDS. While these techniques may be effective, they are expensive for maintenance therapy. Hydrogen sulfide NSAID hybrids prevent the gastric ulceration/bleeding associated with taking the NSAIDs alone. Hydrogen sulphide is known to have a protective effect on the cardiovascular and gastrointestinal system. NSAIDs should be used with caution in individuals with inflammatory bowel disease (e.g., Crohn's disease or ulcerative colitis) due to their tendency to cause gastric bleeding and form ulceration in the gastric lining. NSAIDs are also associated with a fairly high incidence of adverse drug reactions (ADRs) on the kidney and over time can lead to chronic kidney disease. The mechanism of these kidney ADRs is due to changes in kidney blood flow. Prostaglandins normally dilate the afferent arterioles of the glomeruli. This helps maintain normal glomerular perfusion and glomerular filtration rate (GFR), an indicator of kidney function. This is particularly important in kidney failure where the kidney is trying to maintain renal perfusion pressure by elevated angiotensin II levels. At these elevated levels, angiotensin II also constricts the afferent arteriole into the glomerulus in addition to the efferent arteriole it normally constricts. Since NSAIDs block this prostaglandin-mediated effect of afferent arteriole dilation, particularly in kidney failure, NSAIDs cause unopposed constriction of the afferent arteriole and decreased RPF (renal perfusion flow) and GFR. Common ADRs associated with altered kidney function include: These agents may also cause kidney impairment, especially in combination with other nephrotoxic agents. Kidney failure is especially a risk if the patient is also concomitantly taking an ACE inhibitor (which removes angiotensin II's vasoconstriction of the efferent arteriole) and a diuretic (which drops plasma volume, and thereby RPF)—the so-called "triple whammy" effect. In rarer instances NSAIDs may also cause more severe kidney conditions: NSAIDs in combination with excessive use of phenacetin or paracetamol (acetaminophen) may lead to analgesic nephropathy. Photosensitivity is a commonly overlooked adverse effect of many of the NSAIDs. The 2-arylpropionic acids are the most likely to produce photosensitivity reactions, but other NSAIDs have also been implicated including piroxicam, diclofenac, and benzydamine. Benoxaprofen, since withdrawn due to its liver toxicity, was the most photoactive NSAID observed. The mechanism of photosensitivity, responsible for the high photoactivity of the 2-arylpropionic acids, is the ready decarboxylation of the carboxylic acid moiety. The specific absorbance characteristics of the different chromophoric 2-aryl substituents, affects the decarboxylation mechanism. NSAIDs are not recommended during pregnancy, particularly during the third trimester. While NSAIDs as a class are not direct teratogens, they may cause premature closure of the fetal ductus arteriosus and kidney ADRs in the fetus. Additionally, they are linked with premature birth and miscarriage. Aspirin, however, is used together with heparin in pregnant women with antiphospholipid syndrome. Additionally, indomethacin is used in pregnancy to treat polyhydramnios by reducing fetal urine production via inhibiting fetal kidney blood flow. In contrast, paracetamol (acetaminophen) is regarded as being safe and well tolerated during pregnancy, but Leffers et al. released a study in 2010 indicating that there may be associated male infertility in the unborn. Doses should be taken as prescribed, due to risk of liver toxicity with overdoses. In France, the country's health agency contraindicates the use of NSAIDs, including aspirin, after the sixth month of pregnancy. A variety of allergic or allergic-like NSAID hypersensitivity reactions follow the ingestion of NSAIDs. These hypersensitivity reactions differ from the other adverse reactions listed here which are toxicity reactions, i.e. unwanted reactions that result from the pharmacological action of a drug, are dose-related, and can occur in any treated individual; hypersensitivity reactions are idiosyncratic reactions to a drug. Some NSAID hypersensitivity reactions are truly allergic in origin: 1) repetitive IgE-mediated urticarial skin eruptions, angioedema, and anaphylaxis following immediately to hours after ingesting one structural type of NSAID but not after ingesting structurally unrelated NSAIDs; and 2) Comparatively mild to moderately severe T cell-mediated delayed onset (usually more than 24 hour), skin reactions such as maculopapular rash, fixed drug eruptions, photosensitivity reactions, delayed urticaria, and contact dermatitis; or 3) far more severe and potentially life-threatening t-cell-mediated delayed systemic reactions such as the DRESS syndrome, acute generalized exanthematous pustulosis, the Stevens–Johnson syndrome, and toxic epidermal necrolysis. Other NSAID hypersensitivity reactions are allergy-like symptoms but do not involve true allergic mechanisms; rather, they appear due to the ability of NSAIDs to alter the metabolism of arachidonic acid in favor of forming metabolites that promote allergic symptoms. Afflicted individuals may be abnormally sensitive to these provocative metabolites or overproduce them and typically are susceptible to a wide range of structurally dissimilar NSAIDs, particularly those that inhibit COX1. Symptoms, which develop immediately to hours after ingesting any of various NSAIDs that inhibit COX-1, are: 1) exacerbations of asthmatic and rhinitis (see aspirin-induced asthma) symptoms in individuals with a history of asthma or rhinitis and 2) exacerbation or first-time development of wheals or angioedema in individuals with or without a history of chronic urticarial lesions or angioedema. It has been hypothesized that NSAIDs may delay healing from bone and soft-tissue injuries by inhibiting inflammation. On the other hand, it has also been hypothesized that NSAIDs might speed recovery from soft tissue injuries by preventing inflammatory processes from damaging adjacent, non-injured muscles. There is moderate evidence that they delay bone healing. Their overall effect on soft-tissue healing is unclear. The use of NSAIDS for analgesia following gastrointestinal surgery remains controversial, given mixed evidence of an increased risk of leakage from any bowel anastomosis created. This risk may vary according to the class of NSAID prescribed. Common adverse drug reactions (ADR), other than listed above, include: raised liver enzymes, headache, dizziness. Uncommon ADRs include an abnormally high level of potassium in the blood, confusion, spasm of the airways, and rash. Ibuprofen may also rarely cause irritable bowel syndrome symptoms. NSAIDs are also implicated in some cases of Stevens–Johnson syndrome. Most NSAIDs penetrate poorly into the central nervous system (CNS). However, the COX enzymes are expressed constitutively in some areas of the CNS, meaning that even limited penetration may cause adverse effects such as somnolence and dizziness. NSAIDs may increase the risk of bleeding in patients with Dengue fever For this reason, NSAIDs are only available with a prescription in India. In very rare cases, ibuprofen can cause aseptic meningitis. As with other drugs, allergies to NSAIDs might exist. While many allergies are specific to one NSAID, up to 1 in 5 people may have unpredictable cross-reactive allergic responses to other NSAIDs as well. NSAIDs reduce kidney blood flow and thereby decrease the efficacy of diuretics, and inhibit the elimination of lithium and methotrexate. NSAIDs cause decreased ability to form blood clots, which can increase the risk of bleeding when combined with other drugs that also decrease blood clotting, such as warfarin. NSAIDs may aggravate hypertension (high blood pressure) and thereby antagonize the effect of antihypertensives, such as ACE inhibitors. NSAIDs may interfere and reduce efficiency of SSRI antidepressants. Various widely used nonsteroidal anti-inflammatory drugs (NSAIDs) enhance endocannabinoid signaling by blocking the anandamide-degrading membrane enzyme fatty acid amide hydrolase (FAAH). Most NSAIDs act as nonselective inhibitors of the cyclooxygenase (COX) enzymes, inhibiting both the cyclooxygenase-1 (COX-1) and cyclooxygenase-2 (COX-2) isoenzymes. This inhibition is competitively reversible (albeit at varying degrees of reversibility), as opposed to the mechanism of aspirin, which is irreversible inhibition. COX catalyzes the formation of prostaglandins and thromboxane from arachidonic acid (itself derived from the cellular phospholipid bilayer by phospholipase A2). Prostaglandins act (among other things) as messenger molecules in the process of inflammation. This mechanism of action was elucidated in 1970 by John Vane (1927–2004), who received a Nobel Prize for his work (see Mechanism of action of aspirin). COX-1 is a constitutively expressed enzyme with a "house-keeping" role in regulating many normal physiological processes. One of these is in the stomach lining, where prostaglandins serve a protective role, preventing the stomach mucosa from being eroded by its own acid. COX-2 is an enzyme facultatively expressed in inflammation, and it is inhibition of COX-2 that produces the desirable effects of NSAIDs. When nonselective COX-1/COX-2 inhibitors (such as aspirin, ibuprofen, and naproxen) lower stomach prostaglandin levels, ulcers of the stomach or duodenum and internal bleeding can result. NSAIDs have been studied in various assays to understand how they affect each of these enzymes. While the assays reveal differences, unfortunately, different assays provide differing ratios. The discovery of COX-2 led to research to the development of selective COX-2 inhibiting drugs that do not cause gastric problems characteristic of older NSAIDs. Paracetamol (acetaminophen) is not considered an NSAID because it has little anti-inflammatory activity. It treats pain mainly by blocking COX-2 mostly in the central nervous system, but not much in the rest of the body. However, many aspects of the mechanism of action of NSAIDs remain unexplained, and for this reason, further COX pathways are hypothesized. The COX-3 pathway was believed to fill some of this gap but recent findings make it appear unlikely that it plays any significant role in humans and alternative explanation models are proposed. NSAIDs interact with the endocannabinoid system and its endocannabinoids, as COX2 have been shown to utilize endocannabinoids as substrates, and may have a key role in both the therapeutic effects and adverse effects of NSAIDs, as well as in NSAID-induced placebo responses. NSAIDs are also used in the acute pain caused by gout because they inhibit urate crystal phagocytosis besides inhibition of prostaglandin synthase. NSAIDS have antipyretic activity and can be used to treat fever. Fever is caused by elevated levels of prostaglandin E2, which alters the firing rate of neurons within the hypothalamus that control thermoregulation. Antipyretics work by inhibiting the enzyme COX, which causes the general inhibition of prostanoid biosynthesis (PGE2) within the hypothalamus. PGE2 signals to the hypothalamus to increase the body's thermal setpoint. Ibuprofen has been shown more effective as an antipyretic than paracetamol (acetaminophen). Arachidonic acid is the precursor substrate for cyclooxygenase leading to the production of prostaglandins F, D, and E. NSAIDs can be classified based on their chemical structure or mechanism of action. Older NSAIDs were known long before their mechanism of action was elucidated and were for this reason classified by chemical structure or origin. Newer substances are more often classified by mechanism of action. The following NSAIDs are derived from fenamic acid. which is a derivative of anthranilic acid, which in turn is a nitrogen isostere of salicylic acid, which is the active metabolite of aspirin. Most NSAIDs are chiral molecules; diclofenac is a notable exception. However, the majority are prepared as racemic mixtures. Typically, only a single enantiomer is pharmacologically active. For some drugs (typically profens), an isomerase enzyme "in vivo" converts the inactive enantiomer into the active form, although its activity varies widely in individuals. This phenomenon is likely responsible for the poor correlation between NSAID efficacy and plasma concentration observed in older studies when specific analysis of the active enantiomer was not performed. Ibuprofen and ketoprofen are now available in single-enantiomer preparations (dexibuprofen and dexketoprofen), which purport to offer quicker onset and an improved side-effect profile. Naproxen has always been marketed as the single active enantiomer. NSAIDs within a group tend to have similar characteristics and tolerability. There is little difference in clinical efficacy among the NSAIDs when used at equivalent doses. Rather, differences among compounds usually relate to dosing regimens (related to the compound's elimination half-life), route of administration, and tolerability profile. Regarding adverse effects, selective COX-2 inhibitors have lower risk of gastrointestinal bleeding. With the exception of naproxen, nonselective NSAIDs increase the risk of having a heart attack. Some data also supports that the partially selective nabumetone is less likely to cause gastrointestinal events. A consumer report noted that ibuprofen, naproxen, and salsalate are less expensive than other NSAIDs, and essentially as effective and safe when used appropriately to treat osteoarthritis and pain. Most nonsteroidal anti-inflammatory drugs are weak acids, with a pKa of 3–5. They are absorbed well from the stomach and intestinal mucosa. They are highly protein-bound in plasma (typically >95%), usually to albumin, so that their volume of distribution typically approximates to plasma volume. Most NSAIDs are metabolized in the liver by oxidation and conjugation to inactive metabolites that typically are excreted in the urine, though some drugs are partially excreted in bile. Metabolism may be abnormal in certain disease states, and accumulation may occur even with normal dosage. Ibuprofen and diclofenac have short half-lives (2–3 hours). Some NSAIDs (typically oxicams) have very long half-lives (e.g. 20–60 hours). From the era of Greek medicine to the mid-19th century, the discovery of medicinal agents was classed as an empirical art; folklore and mythological guidance were combined in deploying the vegetable and mineral products that made up the expansive pharmacopeia of the time. Myrtle leaves were in use by 1500 BCE. Hippocrates (460–377 BCE) first reported using willow bark and in 30 BCE Celsus described the signs of inflammation and also used willow bark to mitigate them. On 25 April 1763, Edward Stone wrote to the Royal Society describing his observations on the use of willow bark-based medicines in febrile patients. The active ingredient of willow bark, a glycoside called salicin, was first isolated by Johann Andreas Buchner in 1827. By 1829, French chemist Henri Leroux had improved the extraction process to obtain about 30g of purified salicin from 1.5kg of bark. By hydrolysis, salicin releases glucose and salicyl alcohol which can be converted into salicylic acid, both in vivo and through chemical methods. The acid is more effective than salicin and, in addition to its fever-reducing properties, is anti-inflammatory and analgesic. In 1869, Hermann Kolbe synthesised salicylate, although it was too acidic for the gastric mucosa. The reaction used to synthesise aromatic acid from a phenol in the presence of CO2 is known as the Kolbe-Schmitt reaction. By 1897 the German chemist Felix Hoffmann and the Bayer company prompted a new age of pharmacology by converting salicylic acid into acetylsalicylic acid—named aspirin by Heinrich Dreser. Other NSAIDs like ibuprofen were developed from the 1950s forward. In 2001, NSAIDs accounted for 70,000,000 prescriptions and 30billion over-the-counter doses sold annually in the United States. While studies have been conducted to see if various NSAIDs can improve behavior in transgenic mouse models of Alzheimer's disease and observational studies in humans have shown promise, there is no good evidence from randomized clinical trials that NSAIDs can treat or prevent Alzheimer's in humans; clinical trials of NSAIDs for treatment of Alzheimer's have found more harm than benefit. NSAIDs coordinate with metal ions affecting cellular function. Research supports the use of NSAIDs for the control of pain associated with veterinary procedures such as dehorning and castration of calves. The best effect is obtained by combining a short-term local anesthetic such as lidocaine with an NSAID acting as a longer term analgesic. However, as different species have varying reactions to different medications in the NSAID family, little of the existing research data can be extrapolated to animal species other than those specifically studied, and the relevant government agency in one area sometimes prohibits uses approved in other jurisdictions. For example, ketoprofen's effects have been studied in horses more than in ruminants but, due to controversy over its use in racehorses, veterinarians who treat livestock in the United States more commonly prescribe flunixin meglumine, which, while labeled for use in such animals, is not indicated for post-operative pain. In the United States, meloxicam is approved for use only in canines, whereas (due to concerns about liver damage) it carries warnings against its use in cats except for one-time use during surgery. In spite of these warnings, meloxicam is frequently prescribed "off-label" for non-canine animals including cats and livestock species. In other countries, for example The European Union (EU), there is a label claim for use in cats. Also Carprofen.
https://en.wikipedia.org/wiki?curid=22071
NC (complexity) In computational complexity theory, the class NC (for "Nick's Class") is the set of decision problems decidable in polylogarithmic time on a parallel computer with a polynomial number of processors. In other words, a problem is in NC if there exist constants "c" and "k" such that it can be solved in time using parallel processors. Stephen Cook coined the name "Nick's class" after Nick Pippenger, who had done extensive research on circuits with polylogarithmic depth and polynomial size. Just as the class P can be thought of as the tractable problems (Cobham's thesis), so NC can be thought of as the problems that can be efficiently solved on a parallel computer. NC is a subset of P because polylogarithmic parallel computations can be simulated by polynomial-time sequential ones. It is unknown whether NC = P, but most researchers suspect this to be false, meaning that there are probably some tractable problems that are "inherently sequential" and cannot significantly be sped up by using parallelism. Just as the class NP-complete can be thought of as "probably intractable", so the class P-complete, when using NC reductions, can be thought of as "probably not parallelizable" or "probably inherently sequential". The parallel computer in the definition can be assumed to be a "parallel, random-access machine" (PRAM). That is a parallel computer with a central pool of memory, and any processor can access any bit of memory in constant time. The definition of NC is not affected by the choice of how the PRAM handles simultaneous access to a single bit by more than one processor. It can be CRCW, CREW, or EREW. See PRAM for descriptions of those models. Equivalently, NC can be defined as those decision problems decidable by a uniform Boolean circuit (which can be calculated from the length of the input, for NC, we suppose we can compute the Boolean circuit of size "n" in logarithmic space in "n") with polylogarithmic depth and a polynomial number of gates. RNC is a class extending NC with access to randomness. As with P, by a slight abuse of language, one might classify function problems and search problems as being in NC. NC is known to include many problems, including Often algorithms for those problems had to be separately invented and could not be naïvely adapted from well-known algorithms – Gaussian elimination and Euclidean algorithm rely on operations performed in sequence. One might contrast ripple carry adder with a carry-lookahead adder. NC"i" is the class of decision problems decidable by uniform boolean circuits with a polynomial number of gates of at most two inputs and depth , or the class of decision problems solvable in time "O"(log"i" "n") on a parallel computer with a polynomial number of processors. Clearly, we have which forms the NC-hierarchy. We can relate the NC classes to the space classes L and NL and AC. The NC classes are related to the AC classes, which are defined similarly, but with gates having unbounded fan-in. For each "i", we have As an immediate consequence of this, we have that NC = AC. It is known that both inclusions are strict for "i" = 0. Similarly, we have that NC is equivalent to the problems solvable on an alternating Turing machine restricted to at most two options at each step with "O"(log "n") space and formula_4 alternations. One major open question in complexity theory is whether or not every containment in the NC hierarchy is proper. It was observed by Papadimitriou that, if NC"i" = NC"i"+1 for some "i", then NC"i" = NC"j" for all "j" ≥ "i", and as a result, NC"i" = NC. This observation is known as NC-hierarchy collapse because even a single equality in the chain of containments implies that the entire NC hierarchy "collapses" down to some level "i". Thus, there are 2 possibilities: It is widely believed that (1) is the case, although no proof as to the truth of either statement has yet been discovered. A branching program with "n" variables of width "k" and length "m" consists of a sequence of "m" instructions. Each of the instructions is a tuple ("i", "p", "q") where "i" is the index of variable to check (1 ≤ "i" ≤ "n"), and "p" and "q" are functions from {1, 2, ..., "k"} to {1, 2, ..., "k"}. Numbers 1, 2, ..., "k" are called states of the branching program. The program initially starts in state 1, and each instruction ("i", "p", "q") changes the state from "x" to "p"("x") or "q"("x"), depending on whether the "i"th variable is 0 or 1. A family of branching programs consists of a branching program with "n" variables for each "n". It is easy to show that every language "L" on {0,1} can be recognized by a family of branching programs of width 5 and exponential length, or by a family of exponential width and linear length. Every regular language on {0,1} can be recognized by a family of branching programs of constant width and linear number of instructions (since a DFA can be converted to a branching program). BWBP denotes the class of languages recognizable by a family of branching programs of bounded width and polynomial length. Barrington's theorem says that BWBP is exactly nonuniform NC1. The proof uses the nonsolvability of the symmetric group S5. The theorem is rather surprising. For instance, it implies that the majority function can be computed by a family of branching programs of constant width and polynomial size, while intuition might suggest that to achieve polynomial size, one needs a linear number of states. A branching program of constant width and polynomial size can be easily converted (via divide-and-conquer) to a circuit in NC1. Conversely, suppose a circuit in NC1 is given. Without loss of generality, assume it uses only AND and NOT gates. Lemma 1: If there exists a branching program that sometimes works as a permutation "P" and sometimes as a permutation "Q", by right-multiplying permutations in the first instruction by α, and in the last instruction left-multiplying by β, we can make a circuit of the same length that behaves as β"P"α or β"Q"α, respectively. Call a branching program α-computing a circuit "C" if it works as identity when C's output is 0, and as α when C's output is 1. As a consequence of Lemma 1 and the fact that all cycles of length 5 are conjugate, for any two 5-cycles α, β, if there exists a branching program α-computing a circuit "C", then there exists a branching program β-computing the circuit "C", of the same length. Lemma 2: There exist 5-cycles γ, δ such that their commutator ε=γδγ−1δ−1 is a 5-cycle. For example, γ = (1 2 3 4 5), δ = (1 3 5 4 2) giving ε = (1 3 2 5 4). We will now prove Barrington's theorem by induction: Suppose we have a circuit "C" which takes inputs "x"1...,"x"n and assume that for all subcircuits "D" of "C" and 5-cycles α, there exists a branching program α-computing "D". We will show that for all 5-cycles α, there exists a branching program α-computing "C". By assuming the subcircuits have branching programs so that they are α-computing for all 5-cycles α∈"S"5, we have shown "C" also has this property, as required. The size of the branching program is at most 4d, where "d" is the depth of the circuit. If the circuit has logarithmic depth, the branching program has polynomial length.
https://en.wikipedia.org/wiki?curid=22073
Nori The finished dried sheets are made by a shredding and rack-drying process that resembles papermaking. They are sold in packs in grocery stores for culinary purposes. Since "nori" sheets easily absorb water from the air and degrade, a desiccant is needed when storing "nori" for any significant time. Originally, the term "nori" was generic and referred to seaweeds, including "hijiki". One of the oldest descriptions of nori is dated to around the 8th century. In the Taihō Code enacted in 701, "nori" was already included in the form of taxation. Local people have been described as drying nori in Hitachi Province fudoki (721–721), and harvesting of nori was mentioned in Izumo Province fudoki (713–733), showing that nori was used as food from ancient times. In "Utsubo Monogatari", written around 987, "nori" was recognized as a common food. Nori had been consumed as paste form until the sheet form was invented in Asakusa, Edo (contemporary Tokyo), around 1750 in the Edo period through the method of Japanese paper-making. The word ""nori"" first appeared in an English-language publication in "C. P. Thunberg's Trav.", published in 1796. It was used in conjugation as ""Awa nori"", probably referring to what is now called aonori. The Japanese nori industry was in decline after WWII, when Japan was in need of all food that could be produced. The decline was due to a lack of understanding of nori's three-stage life cycle, such that local people did not understand why traditional cultivation methods were not effective. The industry was rescued by knowledge deriving from the work of British phycologist Kathleen Mary Drew-Baker, who had been researching the organism "Porphyria umbilicalis", which grew in the seas around Wales and was harvested for food (bara lafwr or bara lawr), as in Japan. Her work was discovered by Japanese scientists who applied it to artificial methods of seeding and growing the nori, rescuing the industry. Kathleen Baker was hailed as the "Mother of the Sea" in Japan and a statue erected in her memory; she is still revered as the savior of the Japanese nori industry. In the 21st century, the Japanese nori industry faces a new decline due to increased competition from seaweed producers in China and Korea and domestic sales tax hikes. The word "nori" started to be used widely in the United States, and the product (imported in dry form from Japan) became widely available at natural food stores and Asian-American grocery stores in the 1960s due to the macrobiotic movement and in the 1970s with the increase of sushi bars and Japanese restaurants. Production and processing of "nori" is an advanced form of agriculture. The biology of "Pyropia", although complicated, is well understood, and this knowledge is used to control the production process. Farming takes place in the sea where the "Pyropia" plants grow attached to nets suspended at the sea surface and where the farmers operate from boats. The plants grow rapidly, requiring about 45 days from "seeding" until the first harvest. Multiple harvests can be taken from a single seeding, typically at about ten-day intervals. Harvesting is accomplished using mechanical harvesters of a variety of configurations. Processing of raw product is mostly accomplished by highly automated machines that accurately duplicate traditional manual processing steps, but with much improved efficiency and consistency. The final product is a paper-thin, black, dried sheet of approximately and in weight. Several grades of "nori" are available in the United States. The most common, and least expensive, grades are imported from China, costing about six cents per sheet. At the high end, ranging up to 90 cents per sheet, are "delicate "shin-nori"" ("nori" from the first of the year's several harvests) cultivated in Ariake Sea, off the island of Kyushu in Japan. In Japan, over of coastal waters are given to producing of "nori", worth over a billion dollars. China produces about a third of this amount. "Nori" is commonly used as a wrap for sushi and "onigiri". It is also a garnish or flavoring in noodle preparations and soups. It is most typically toasted prior to consumption ("yaki-nori"). A common secondary product is toasted and flavored "nori" ("ajitsuke-nori"), in which a flavoring mixture (variable, but typically soy sauce, sugar, sake, mirin, and seasonings) is applied in combination with the toasting process. It is also eaten by making it into a soy sauce-flavored paste, "nori no tsukudani" (). "Nori" is sometimes also used as a form of food decoration. A related product, prepared from the unrelated green algae "Monostroma" and "Enteromorpha", is called "aonori" ( literally blue/green "nori") and is used like herbs on everyday meals, such as "okonomiyaki" and "yakisoba". Raw seaweed is 85% water, 6% protein, 5% carbohydrates, and has negligible fat (table). In a 100 gram reference amount, seaweed is a rich source (20% or more of the Daily Value, DV) of vitamin A, vitamin C, riboflavin, and folate (table). Seaweed is a moderate source (less than 20% DV) of niacin, iron, and zinc. Seaweed has a high content of iodine, providing a substantial amount in just one gram. A 2014 study reported that dried purple laver ("nori") contains vitamin B12 in sufficient quantities to meet the RDA requirement (Vitamin B12 content: 77.6 μg /100 g dry weight). By contrast, a 2017 review concluded that vitamin B12 may be destroyed during metabolism or is converted into inactive B12 analogs during drying and storage. The Academy of Nutrition and Dietetics established in 2016 that nori is not an adequate source of vitamin B12 for humans. Nori contains toxic metals (arsenic and cadmium), whose levels are highly variable among nori products. It also contains amphipod allergens that can cause serious allergic reactions, especially in highly sensitized crustacean-allergic people. Therefore, daily consumption of high amounts of dried nori is discouraged.
https://en.wikipedia.org/wiki?curid=22076
Normative ethics Normative ethics is the study of ethical behaviour, and is the branch of philosophical ethics that investigates the questions that arise in regards to how one ought to act, in a moral sense. Normative ethics is distinct from meta-ethics in that the former examines standards for the rightness and wrongness of actions, whereas the latter studies the meaning of moral language and the metaphysics of moral facts. Likewise, normative ethics is distinct from applied ethics in that the former is more concerned with 'who ought one be' rather than the ethics of a specific issue (e.g. if, or when, abortion is acceptable). Normative ethics is also distinct from descriptive ethics, as the latter is an empirical investigation of people’s moral beliefs. In this context normative ethics is sometimes called "prescriptive", as opposed to "descriptive" ethics. However, on certain versions of the meta-ethical view of moral realism, moral facts are both descriptive and prescriptive at the same time. Most traditional moral theories rest on principles that determine whether an action is right or wrong. Classical theories in this vein include utilitarianism, Kantianism, and some forms of contractarianism. These theories mainly offered the use of overarching moral principles to resolve difficult moral decisions. There are disagreements about what precisely gives an action, rule, or disposition its ethical force. There are three competing views on how moral questions should be answered, along with hybrid positions that combine some elements of each: The former focuses on the character of those who are acting. In contrast, both deontological ethics and consequentialism focus on the status of the action, rule, or disposition itself, and come in various forms. Virtue ethics, advocated by Aristotle with some aspects being supported by Saint Thomas Aquinas, focuses on the inherent character of a person rather than on specific actions. There has been a significant revival of virtue ethics in the past half-century, through the work of such philosophers as G. E. M. Anscombe, Philippa Foot, Alasdair Macintyre, Mortimer J. Adler, Jacques Maritain, Yves Simon, and Rosalind Hursthouse. Deontology argues that decisions should be made considering the factors of one's duties and one's rights. Some deontological theories include: Consequentialism, or teleology, argues that the morality of an action is contingent on the action's outcome or result. Consequentialist theories, varying in what they consider to be valuable (i.e., axiology), include: It can be unclear what it means to say that a person "ought to do X because it is moral, whether they like it or not." Morality is sometimes presumed to have some kind of special binding force on behaviour, though some philosophers believe that, used this way, the word "ought" seems to wrongly attribute magic powers to morality. For instance, G. E. M. Anscombe worries that "ought" has become "a word of mere mesmeric force." British ethicist Philippa Foot elaborates that morality does not seem to have any special binding force, and she clarifies that people only behave morally when motivated by other factors. Foot says "People talk, for instance, about the 'binding force' of morality, but it is not clear what this means if not that we feel ourselves unable to escape." The idea is that, faced with an opportunity to steal a book because we can get away with it, moral obligation itself has no power to stop us unless we "feel" an obligation. Morality may therefore have no binding force beyond regular human motivations, and people must be motivated to behave morally. The question then arises: what role does reason play in motivating moral behaviour? The categorical imperative perspective suggests that proper reason always leads to particular moral behaviour. As mentioned above, Foot instead believes that humans are actually motivated by desires. Proper reason, on this view, allows humans to discover actions that get them what they want (i.e., hypothetical imperatives)—not necessarily actions that are moral. Social structure and motivation can make morality binding in a sense, but only because it makes moral norms feel inescapable, according to Foot. John Stuart Mill adds that external pressures, to please others for instance, also influence this felt binding force, which he calls human "conscience." Mill says that humans must first reason about what is moral, then try to bring the feelings of our conscience in line with our reason. At the same time, Mill says that a good moral system (in his case, utilitarianism) ultimately appeals to aspects of human nature—which, must themselves be nurtured during upbringing. Mill explains: This firm foundation is that of the social feelings of mankind; the desire to be in unity with our fellow creatures, which is already a powerful principle in human nature, and happily one of those which tend to become stronger, even without express inculcation, from the influences of advancing civilisation. Mill thus believes that it is important to appreciate that it is feelings that drive moral behavior, but also that they may not be present in some people (e.g. psychopaths). Mill goes on to describe factors that help ensure people develop a conscience and behave morally. Popular texts such as Joseph Daleiden's "The Science of Morality: The Individual, Community, and Future Generations" (1998) describe how societies can use science to figure out how to make people more likely to be good.
https://en.wikipedia.org/wiki?curid=22081
Negotiation Negotiation is a dialogue between two or more people or parties intended to reach a beneficial outcome over one or more issues where a conflict exists with respect to at least one of these issues. Negotiation is an interaction and process between entities who compromise to agree on matters of mutual interest, while optimizing their individual utilities. This beneficial outcome can be for all of the parties involved, or just for one or some of them. Negotiators need to understand the negotiation process and other negotiators to increase their chances to close deals, avoid conflicts, establishing relationship with other parties and gain profit. It is aimed to resolve points of difference, to gain advantage for an individual or collective, or to craft outcomes to satisfy various interests. It is often conducted by putting forward a position and making concessions to achieve an agreement. The degree to which the negotiating parties trust each other to implement the negotiated solution is a major factor in determining whether negotiations are successful. People negotiate daily, often without considering it a negotiation. Negotiation occurs in organizations, including businesses, non-profits, and within and between governments as well as in sales and legal proceedings, and in personal situations such as marriage, divorce, parenting, etc. Professional negotiators are often specialized, such as union negotiators, leverage buyout negotiators, peace negotiator, or hostage negotiators. They may also work under other titles, such as diplomats, legislators, or brokers. There is also negotiation conducted by algorithms or machines known as autonomous negotiation. For automation, the negotiation participants and process have to be modeled correctly. Negotiation can take a wide variety of forms, from a multilateral conference of all United Nations members to establish a new international norm (such as the UN Convention on the Law of the Sea) to a meeting of parties to a conflict to end violence or resolve the underlying issue (such as constitutional negotiations in South Africa in 1990-1994 or in Colombia with the FARC in 2012-2015) to a business encounter to make a deal to a face-off between parents (or between parent and child) over the child's proper behavior. Mediation is a form of negotiation with a third-party catalyst who helps the conflicting parties negotiate when they cannot do so by themselves. Negotiation can be contrasted with arbitration, where the decision lies with the third party, which the conflicting parties are committed to accept. Negotiation theorists generally distinguish between two types of negotiation The difference in the usage of the two type depends on the mindset of the negotiator but also on the situation: one-off encounters where lasting relationships do not obtain are more likely to produce distributive negotiations whereas lasting relationships are more likely to require integrative negotiating Different theorists use different labels for the two general types and distinguish them in different ways. Distributive negotiation is also sometimes called positional or hard-bargaining negotiation and attempts to distribute a "fixed pie" of benefits. Distributive negotiation operates under zero-sum conditions and implies that any gain one party makes is at the expense of the other and vice versa. For this reason, distributive negotiation is also sometimes called "win-lose" because of the assumption that one person's gain is another person's loss. Distributive negotiation examples include haggling prices on an open market, including the negotiation of the price of a car or a home. In a distributive negotiation, each side often adopts an extreme or fixed position, knowing it will not be accepted—and then seeks to cede as little as possible before reaching a deal. Distributive bargainers conceive of negotiation as a process of distributing a fixed amount of value. A distributive negotiation often involves people who have never had a previous interactive relationship, nor are they likely to do so again in the near future, although all negotiations usually have a distributive element. In the distributive approach each negotiator fights for the largest possible piece of the pie, so parties tend to regard each other more as an adversary than a partner and to take a harder line. Since Prospect Theory indicates that people value losses more than gains and are more risk-averse about losses, concession-convergence bargaining is likely to be more acrimonious and less productive of an agreement Integrative negotiation is also called interest-based, merit-based, or principled negotiation. It is a set of techniques that attempts to improve the quality and likelihood of negotiated agreement by taking advantage of the fact that different parties often value various outcomes differently. While distributive negotiation assumes there is a fixed amount of value (a "fixed pie") to be divided between the parties, integrative negotiation attempts to create value in the course of the negotiation ("expand the pie") by either "compensating" loss of one item with gains from another ("trade-offs" or logrolling), or by constructing or reframing the issues of the conflict in such a way that both parties benefit ("win-win" negotiation). However, even integrative negotiation is likely to have some distributive elements, especially when the different parties both value different items to the same degree or when details are left to be allocated at the end of the negotiation. While concession is mandatory for negotiations, research shows that people who concede more quickly, are less likely to explore all integrative and mutually beneficial solutions. Therefore, early conceding reduces the chance of an integrative negotiation. Integrative negotiation often involves a higher degree of trust and the formation of a relationship. It can also involve creative problem-solving that aims to achieve mutual gains. It sees a good agreement as not one with maximum individual gain, but one that provides optimum gain for all parties. Gains in this scenario are not at the expense of the Other, but with it. Each seeks to accord the Other enough benefit that it will hold to the agreement that gives the first party an agreeable outcome, and vice versa. Productive negotiation focuses on the underlying interests of the parties rather than their starting positions, approaches negotiation as a shared problem-solving rather than a personalized battle, and insists upon adherence to objective, principled criteria as the basis for agreement. However, negotiators need not sacrifice effective negotiation in favor of a positive relationship between parties. Rather than conceding, each side can appreciate that the other has emotions and motivations of their own and use this to their advantage in discussing the issue. In fact, perspective-taking can help move parties toward a more integrative solution. Fisher et al. illustrate a few techniques that effectively improve perspective-taking in their book "Getting to Yes", and through the following, negotiators can separate people from the problem itself. Additionally, negotiators can use certain communication techniques to build a stronger relationship and develop more meaningful negotiation solution. "Integrated negotiation" is a strategic approach to influence that maximizes value in any single negotiation through the astute linking and sequencing of other negotiations and decisions related to one's operating activities. This approach in complex settings is best executed by mapping out all potentially relevant negotiations, conflicts and operating decisions in order to integrate helpful connections among them, while minimizing any potentially harmful connections (see examples below). "Integrated negotiation" is not to be confused with "integrative negotiation", a different concept (as outlined above) related to a non-zero-sum approach to creating value in negotiations. Integrated negotiation was first identified and labeled by international negotiator and author Peter Johnston in his book "Negotiating with Giants". One of the examples cited in Johnston's book is that of J. D. Rockefeller deciding where to build his first major oil refinery. Instead of taking the easier, cheaper route from the oil fields to refine his petroleum in Pittsburgh, Rockefeller chose to build his refinery in Cleveland. Why? Because rail companies would be transporting his refined oil to market. Pittsburgh had just one major railroad, meaning it could dictate prices in negotiations, while Cleveland had three railroads that Rockefeller knew would compete for his business, potentially reducing his costs significantly. The leverage gained in these rail negotiations more than offset the additional operating costs of sending his oil to Cleveland for refining, helping establish Rockefeller's empire, while undermining his competitors who failed to integrate their core operating decisions with their negotiation strategies. Other examples of integrated negotiation include the following: When a party pretends to negotiate, but secretly has no intention of compromising, the party is considered negotiating in bad faith. Bad faith is a concept in negotiation theory whereby parties pretend to reason to reach settlement, but have no intention to do so, for example, one political party may pretend to negotiate, with no intention to compromise, for political effect. Bad faith negotiations are often used in political science and political psychology to refer to negotiating strategies in which there is no real intention to reach compromise, or a model of information processing. The "inherent bad faith model" of information processing is a theory in political psychology that was first put forth by Ole Holsti to explain the relationship between John Foster Dulles' beliefs and his model of information processing. It is the most widely studied model of one's opponent. A state is presumed implacably hostile, and contra-indicators of this are ignored. They are dismissed as propaganda ploys or signs of weakness. Examples are John Foster Dulles' position regarding the Soviet Union, or Hamas's position on the state of Israel. The total of advantages and disadvantages to be distributed in a negotiation is illustrated with the term negotiation pie. The course of the negotiation can either lead to an increase, shrinking, or stagnation of these values. If the negotiation parties are able to expand the total pie a win-win situation is possible assuming that both parties profit from the expansion of the pie. In practice, however, this maximisation approach is oftentimes impeded by the so-called small pie bias, i.e. the psychological underestimation of the negotiation pie's size. Likewise, the possibility to increase the pie may be underestimated due to the so-called incompatibility bias. Contrary to enlarging the pie, the pie may also shrink during negotiations e.g. due to (excessive) negotiation costs. There are many different ways to categorize the essential elements of negotiation. One view of negotiation involves three basic elements: "process", "behavior" and "substance". The process refers to how the parties negotiate: the context of the negotiations, the parties to the negotiations, the tactics used by the parties, and the sequence and stages in which all of these play out. Behavior refers to the relationships among these parties, the communication between them and the styles they adopt. The substance refers to what the parties negotiate over: the agenda, the issues (positions and – more helpfully – interests), the options, and the agreement(s) reached at the end. Another view of negotiation comprises four elements: "strategy", "process", "tools", and "tactics". Strategy comprises the top level goals – typically including relationship and the final outcome. Processes and tools include the steps to follow and roles to take in preparing for and negotiating with the other parties. Tactics include more detailed statements and actions and responses to others' statements and actions. Some add to this "persuasion and influence", asserting that these have become integral to modern day negotiation success, and so should not be omitted. A skilled negotiator may serve as an advocate for one party to the negotiation. The advocate attempts to obtain the most favorable outcomes possible for that party. In this process the negotiator attempts to determine the minimum outcome(s) the other party is (or parties are) willing to accept, then adjusts their demands accordingly. A "successful" negotiation in the advocacy approach is when the negotiator is able to obtain all or most of the outcomes their party desires, but without driving the other party to permanently break off negotiations. Skilled negotiators may use a variety of tactics ranging from negotiation hypnosis, to a straightforward presentation of demands or setting of preconditions, to more deceptive approaches such as cherry picking. Intimidation and salami tactics may also play a part in swaying the outcome of negotiations. Another negotiation tactic is bad guy/good guy. Bad guy/good guy is when one negotiator acts as a bad guy by using anger and threats. The other negotiator acts as a good guy by being considerate and understanding. The good guy blames the bad guy for all the difficulties while trying to get concessions and agreement from the opponent. The best alternative to a negotiated agreement, or BATNA, is the most advantageous alternative course of action a negotiator can take should the current negotiation end without reaching agreement. The quality of a BATNA has the potential to improve a party's negotiation outcome. Understanding one's BATNA can empower an individual and allow him or her to set higher goals when moving forward. Alternatives need to be actual and actionable to be of value. Negotiators may also consider the other party's BATNA and how it compares to what they are offering during the negotiation. Kenneth W. Thomas identified five styles or responses to negotiation. These five strategies have been frequently described in the literature and are based on the dual-concern model. The dual concern model of conflict resolution is a perspective that assumes individuals' preferred method of dealing with conflict is based on two themes or dimensions: Based on this model, individuals balance the concern for personal needs and interests with the needs and interests of others. The following five styles can be used based on individuals' preferences depending on their pro-self or pro-social goals. These styles can change over time, and individuals can have strong dispositions towards numerous styles. Three basic kinds of negotiators have been identified by researchers involved in The Harvard Negotiation Project. These types of negotiators are: soft bargainers, hard bargainers, and principled bargainers. Researchers from The Harvard Negotiation Project recommend that negotiators explore a number of alternatives to the problems they face in order to reach the best solution, but this is often not the case (as when you may be dealing with an individual using soft or hard bargaining tactics) (Forsyth, 2010). Tactics are always an important part of the negotiating process. More often than not they are subtle, difficult to identify and used for multiple purposes. Tactics are more frequently used in distributive negotiations and when the focus in on taking as much value off the table as possible. Many negotiation tactics exist. Below are a few commonly used tactics. Auction: The bidding process is designed to create competition. When multiple parties want the same thing, pit them against one another. When people know that they may lose out on something, they want it even more. Not only do they want the thing that is being bid on, they also want to win, just to win. Taking advantage of someone's competitive nature can drive up the price. Brinksmanship: One party aggressively pursues a set of terms to the point where the other negotiating party must either agree or walk away. Brinkmanship is a type of "hard nut" approach to bargaining in which one party pushes the other party to the "brink" or edge of what that party is willing to accommodate. Successful brinksmanship convinces the other party they have no choice but to accept the offer and there is no acceptable alternative to the proposed agreement. Bogey: Negotiators use the bogey tactic to pretend that an issue of little or no importance is very important. Then, later in the negotiation, the issue can be traded for a major concession of actual importance. Calling a higher authority: To mitigate too far reaching concessions, deescalate, or overcome a deadlock situation, one party makes the further negotiation process dependent on the decision of a decision maker, not present at the negotiation table. Chicken: Negotiators propose extreme measures, often bluffs, to force the other party to chicken out and give them what they want. This tactic can be dangerous when parties are unwilling to back down and go through with the extreme measure. Defence in Depth: Several layers of decision-making authority is used to allow further concessions each time the agreement goes through a different level of authority. In other words, each time the offer goes to a decision maker, that decision maker asks to add another concession to close the deal. Deadlines: Give the other party a deadline forcing them to make a decision. This method uses time to apply pressure to the other party. Deadlines given can be actual or artificial. Flinch: Flinching is showing a strong negative physical reaction to a proposal. Common examples of flinching are gasping for air, or a visible expression of surprise or shock. The flinch can be done consciously or unconsciously. The flinch signals to the opposite party that you think the offer or proposal is absurd in hopes the other party will lower their aspirations. Seeing a physical reaction is more believable than hearing someone saying, "I'm shocked." Good Guy/Bad Guy: Within the tactic of good guy/bad guy (synonyms are good cop/bad cop or black hat/white hat) oftentimes positive and unpleasant tasks are divided between two negotiators on the same negotiation side or unpleasant tasks or decisions are allocated to an (real or fictitious) outsider. The good guy supports the conclusion of the contract and emphasises positive aspects of the negotiation (mutual interests). The bad guy criticises negative aspects (opposing interests). The division of the two roles allows more consistent behaviour and credibility of the individual negotiators. As the good guy promotes the contract, he/she can build trust with the other side. Highball/Lowball or Ambit claim: Depending on whether selling or buying, sellers or buyers use a ridiculously high, or ridiculously low opening offer that is not achievable. The theory is that the extreme offer makes the other party reevaluate their own opening offer and move close to the resistance point (as far as you are willing to go to reach an agreement). Another advantage is that the party giving the extreme demand appears more flexible when they make concessions toward a more reasonable outcome. A danger of this tactic is that the opposite party may think negotiating is a waste of time. The Nibble: Also known under salami tactic or quivering quill, nibbling is the demand of proportionally small concessions that haven't been discussed previously just before closing the deal. This method takes advantage of the other party's desire to close by adding "just one more thing." Snow Job: Negotiators overwhelm the other party with so much information that they have difficulty determining what information is important, and what is a diversion. Negotiators may also use technical language or jargon to mask a simple answer to a question asked by a non-expert. Mirroring: When people get on well, the outcome of a negotiation is likely to be more positive. To create trust and a rapport, a negotiator may mimic or mirror the opponent's behavior and repeat what they say. Mirroring refers to a person repeating the core content of what another person just said, or repeating a certain expression. It indicates attention to the subject of negotiation and acknowledges the other party's point or statement. Mirroring can help create trust and establish a relationship. Communication is a key element of negotiation. Effective negotiation requires that participants effectively convey and interpret information. Participants in a negotiation communicate information not only verbally but non-verbally through body language and gestures. By understanding how nonverbal communication works, a negotiator is better equipped to interpret the information other participants are leaking non-verbally while keeping secret those things that would inhibit his/her ability to negotiate. Non-verbal "anchoring" In a negotiation, a person can gain the advantage by verbally expressing a position first. By anchoring one's position, one establishes the position from which the negotiation proceeds. In a like manner, one can "anchor" and gain advantage with nonverbal (body language) cues. Reading non-verbal communication Being able to read the non-verbal communication of another person can significantly aid in the communication process. By being aware of inconsistencies between a person's verbal and non-verbal communication and reconciling them, negotiators can to come to better resolutions. Examples of incongruity in body language include: Conveying receptivity The way negotiation partners position their bodies relative to each other may influence how receptive each is to the other person's message and ideas. Receptive negotiators tend to appear relaxed with their hands open and palms visibly displayed. Emotions play an important part in the negotiation process, although it is only in recent years that their effect is being studied. Emotions have the potential to play either a positive or negative role in negotiation. During negotiations, the decision as to whether or not to settle rests in part on emotional factors. Negative emotions can cause intense and even irrational behavior, and can cause conflicts to escalate and negotiations to break down, but may be instrumental in attaining concessions. On the other hand, positive emotions often facilitate reaching an agreement and help to maximize joint gains, but can also be instrumental in attaining concessions. Positive and negative discrete emotions can be strategically displayed to influence task and relational outcomes and may play out differently across cultural boundaries. Dispositions for affects affect various stages of negotiation: which strategies to use, which strategies are actually chosen, the way the other party and their intentions are perceived, their willingness to reach an agreement and the final negotiated outcomes. Positive affectivity (PA) and negative affectivity (NA) of one or more of the negotiating sides can lead to very different outcomes. Even before the negotiation process starts, people in a positive mood have more confidence, and higher tendencies to plan to use a cooperative strategy. During the negotiation, negotiators who are in a positive mood tend to enjoy the interaction more, show less contentious behavior, use less aggressive tactics and more cooperative strategies. This in turn increases the likelihood that parties will reach their instrumental goals, and enhance the ability to find integrative gains. Indeed, compared with negotiators with negative or natural affectivity, negotiators with positive affectivity reached more agreements and tended to honor those agreements more. Those favorable outcomes are due to better decision making processes, such as flexible thinking, creative problem solving, respect for others' perspectives, willingness to take risks and higher confidence. Post-negotiation positive affect has beneficial consequences as well. It increases satisfaction with achieved outcome and influences one's desire for future interactions. The PA aroused by reaching an agreement facilitates the dyadic relationship, which brings commitment that sets the stage for subsequent interactions. PA also has its drawbacks: it distorts perception of self performance, such that performance is judged to be relatively better than it actually is. Thus, studies involving self reports on achieved outcomes might be biased. "Don't fear to negotiate don't negotiate out of fear" JFK. Negative affect has detrimental effects on various stages in the negotiation process. Although various negative emotions affect negotiation outcomes, by far the most researched is anger. Angry negotiators plan to use more competitive strategies and to cooperate less, even before the negotiation starts. These competitive strategies are related to reduced joint outcomes. During negotiations, anger disrupts the process by reducing the level of trust, clouding parties' judgment, narrowing parties' focus of attention and changing their central goal from reaching agreement to retaliating against the other side. Angry negotiators pay less attention to opponent's interests and are less accurate in judging their interests, thus achieve lower joint gains. Moreover, because anger makes negotiators more self-centered in their preferences, it increases the likelihood that they will reject profitable offers. Opponents who get really angry (or cry, or otherwise lose control) are more likely to make errors: make sure they are in your favor. Anger does not help achieve negotiation goals either: it reduces joint gains and does not boost personal gains, as angry negotiators do not succeed. Moreover, negative emotions lead to acceptance of settlements that are not in the positive utility function but rather have a negative utility. However, expression of negative emotions during negotiation can sometimes be beneficial: legitimately expressed anger can be an effective way to show one's commitment, sincerity, and needs. Moreover, although NA reduces gains in integrative tasks, it is a better strategy than PA in distributive tasks (such as zero-sum). In his work on negative affect arousal and white noise, Seidner found support for the existence of a negative affect arousal mechanism through observations regarding the devaluation of speakers from other ethnic origins. Negotiation may be negatively affected, in turn, by submerged hostility toward an ethnic or gender group. Research indicates that negotiator's emotions do not necessarily affect the negotiation process. Albarracın et al. (2003) suggested that there are two conditions for emotional affect, both related to the ability (presence of environmental or cognitive disturbances) and the motivation: According to this model, emotions affect negotiations only when one is high and the other is low. When both ability and motivation are low, the affect is identified, and when both are high the affect is identified but discounted as irrelevant to judgment. A possible implication of this model is, for example, that the positive effects PA has on negotiations (as described above) is seen only when either motivation or ability are low. Most studies on emotion in negotiations focus on the effect of the negotiator's own emotions on the process. However, what the other party feels might be just as important, as group emotions are known to affect processes both at the group and the personal levels. When it comes to negotiations, trust in the other party is a necessary condition for its emotion to affect, and visibility enhances the effect. Emotions contribute to negotiation processes by signaling what one feels and thinks and can thus prevent the other party from engaging in destructive behaviors and to indicate what steps should be taken next: PA signals to keep in the same way, while NA points that mental or behavioral adjustments are needed. Partner's emotions can have two basic effects on negotiator's emotions and behavior: mimetic/ reciprocal or complementary. For example, disappointment or sadness might lead to compassion and more cooperation. In a study by Butt et al. (2005) that simulated real multi-phase negotiation, most people reacted to the partner's emotions in reciprocal, rather than complementary, manner. Specific emotions were found to have different effects on the opponent's feelings and strategies chosen: Negotiation is a rather complex interaction. Capturing all its complexity is a very difficult task, let alone isolating and controlling only certain aspects of it. For this reason most negotiation studies are done under laboratory conditions, and focus only on some aspects. Although lab studies have their advantages, they do have major drawbacks when studying emotions: While negotiations involving more than two parties is less often researched, some results from two-party negotiations still apply with more than two parties. One such result is that in negotiations it is common to see language similarity arise between the two negotiating parties. In three-party negotiations, language similarity still arose, and results were particularly efficient when the party with the most to gain from the negotiation adopted language similarities from the other parties. Due to globalization and growing business trends, negotiation in the form of teams is becoming widely adopted. Teams can effectively collaborate to break down a complex negotiation. There is more knowledge and wisdom dispersed in a team than in a single mind. Writing, listening, and talking, are specific roles team members must satisfy. The capacity base of a team reduces the amount of blunder, and increases familiarity in a negotiation. However, unless a team can appropriately utilize the full capacity of its potential, effectiveness can suffer. One factor in the effectiveness of team negotiation is a problem that occurs through solidarity behavior. Solidarity behavior occurs when one team member reduces his or her own utility (benefit) in order to increase the benefits of other team members. This behavior is likely to occur when interest conflicts rise. When the utility/needs of the negotiation opponent does not align with every team member's interests, team members begin to make concessions and balance the benefits gained among the team. Intuitively, this may feel like a cooperative approach. However, though a team may aim to negotiate in a cooperative or collaborative nature, the outcome may be less successful than is possible, especially when integration is possible. Integrative potential is possible when different negotiation issues are of different importance to each team member. Integrative potential is often missed due to the lack of awareness of each member's interests and preferences. Ultimately, this leads to a poorer negotiation result. Thus, a team can perform more effectively if each member discloses his or her preferences prior to the negotiation. This step will allow the team to recognize and organize the team's joint priorities, which they can take into consideration when engaging with the opposing negotiation party. Because a team is more likely to discuss shared information and common interests, teams must make an active effort to foster and incorporate unique viewpoints from experts from different fields. Research by Daniel Thiemann, which largely focused on computer-supported collaborative tasks, found that the Preference Awareness method is an effective tool for fostering the knowledge about joint priorities and further helps the team judge which negotiation issues were of highest importance. Many of the strategies in negotiation vary across genders, and this leads to variations in outcomes for different genders, often with women experiencing less success in negotiations as a consequence. This is due to a number of factors, including that it has been shown that it is more difficult for women to be self-advocating when they are negotiating. Many of the implications of these findings have strong financial impacts in addition to the social backlash faced by self-advocating women in negotiations, as compared to other advocating women, self-advocating men, and other advocating men. Research in this area has been studied across platforms, in addition to more specific areas like women as physician assistants. The backlash associated with this type of behavior is attributed to the fact that to be self-advocated is considered masculine, whereas the alternative, being accommodating, is considered more feminine. Males, however, do not appear to face any type of backlash for not being self-advocating. This research has been supported by multiple studies, including one which evaluated candidates participating in a negotiation regarding compensation. This study showed that women who initiated negotiations were evaluated more poorly than men who initiated negotiations. In another variation of this particular setup, men and women evaluated videos of men and women either accepting a compensation package or initiating negotiations. Men evaluated women more poorly for initiating negotiations, while women evaluated both men and women more poorly for initiating negotiations. In this particular experiment, women were less likely to initiate a negotiation with a male, citing nervousness, but there was no variation with the negotiation was initiated with another female. Research also supports the notion that the way individuals respond in a negotiation varies depending on the gender of the opposite party. In all-male groups, the use of deception showed no variation upon the level of trust between negotiating parties, however in mixed-sex groups there was an increase in deceptive tactics when it was perceived that the opposite party was using an accommodating strategy. In all-female groups, there were many shifts in when individuals did and did not employ deception in their negotiation tactics. The academic world contains a unique management system, wherein faculty members, some of which have tenure, reside in academic units (e.g. departments) and are overseen by chairs, or heads. These chairs/heads are in turn supervised by deans of the college where their academic unit resides. Negotiation is an area where faculty, chairs/heads and their deans have little preparation; their doctoral degrees are typically in a highly specialized area according to their academic expertise. However, the academic environment frequently presents with situations where negotiation takes place. For example, many faculty are hired with an expectation that they will conduct research and publish scholarly works. For these faculty, where their research requires equipment, space, and/or funding, negotiation of a "start-up" package is critical for their success and future promotion. Also, department chairs often find themselves in situations, typically involving resource redistribution where they must negotiate with their dean, on behalf of their unit. And deans oversee colleges where they must optimize limited resources, such as research space or operating funds while at the same time creating an environment that fosters student success, research accomplishments and more. Integrative negotiation is the type predominately found in academic negotiation – where trust and long-term relationships between personnel are valued. Techniques found to be particularly useful in academic settings include: (1) doing your homework – grounding your request in facts; (2) knowing your value; (3) listening actively and acknowledging what is being said, (4) putting yourself in their shoes, (5) asking – negotiation begins with an ask, (6) not committing immediately, (7) managing emotion and (8) keeping in mind the principle of a "wise agreement", with its associated emphasis on meeting the interests of both parties to the extent possible as a key working point. The articles by Callahan, et al. and Amekudzi-Kennedy, et al. contain several case studies of academic negotiations. The word "negotiation" originated in the early 15th century from the Old French "negociacion" from Latin "negotiatio" from "neg"- "no" and "otium" "leisure". These terms mean "business, trade, traffic". By the late 1570s negotiation had the definition, "to communicate in search of mutual agreement." With this new introduction and this meaning, it showed a shift in "doing business" to "bargaining about" business.
https://en.wikipedia.org/wiki?curid=22083
Fertility awareness Fertility awareness (FA) refers to a set of practices used to determine the fertile and infertile phases of a woman's menstrual cycle. Fertility awareness methods may be used to avoid pregnancy, to achieve pregnancy, or as a way to monitor gynecological health. Methods of identifying infertile days have been known since antiquity, but scientific knowledge gained during the past century has increased the number, variety, and especially accuracy of methods. Systems of fertility awareness rely on observation of changes in one or more of the primary fertility signs (basal body temperature, cervical mucus, and cervical position), tracking menstrual cycle length and identifying the fertile window based on this information, or both. Other signs may also be observed: these include breast tenderness and mittelschmerz (ovulation pains), urine analysis strips known as ovulation predictor kits (OPKs), and microscopic examination of saliva or cervical fluid. Also available are computerized fertility monitors. Symptoms-based methods involve tracking one or more of the three primary fertility signs: basal body temperature, cervical mucus, and cervical position. Systems relying exclusively on cervical mucus include the Billings Ovulation Method, the Creighton Model, and the Two-Day Method. Symptothermal methods combine observations of basal body temperature (BBT), cervical mucus, and sometimes cervical position. Calendar-based methods rely on tracking a woman's cycle and identifying her fertile window based on the lengths of her cycles. The best known of these methods is the Standard Days Method. The Calendar-Rhythm method is also considered a calendar-based method, though it is not well defined and has many different meanings to different people. Systems of fertility awareness may be referred to as fertility awareness–based methods (FAB methods); the term Fertility Awareness Method (FAM) refers specifically to the system taught by Toni Weschler. The term natural family planning (NFP) is sometimes used to refer to any use of FA methods, the Lactational amenorrhea method and periodic abstinence during fertile times. A method of FA may be used by NFP users to identify these fertile times. Women who are breastfeeding a child and wish to avoid pregnancy may be able to practice the lactational amenorrhea method (LAM). LAM is distinct from fertility awareness, but because it also does not involve contraceptives, it is often presented alongside FA as a method of "natural" birth control. Within the Catholic Church and some Protestant denominations, the term Natural Family Planning is often used to refer to Fertility Awareness pointing out it is the only method of Family Planning approved by the Church. It is not known exactly when it was first discovered that women have predictable periods of fertility and infertility. It is already clearly stated in the Talmud tractate Niddah, that a woman only becomes pregnant in specific periods in the month, which seemingly refers to ovulation. St. Augustine wrote about periodic abstinence to avoid pregnancy in the year 388 (the Manichaeans attempted to use this method to remain childfree, and Augustine condemned their use of periodic abstinence). One book states that periodic abstinence was recommended ""by a few secular thinkers since the mid-nineteenth century,"" but the dominant force in the twentieth century popularization of fertility awareness-based methods was the Roman Catholic Church. In 1905 Theodoor Hendrik van de Velde, a Dutch gynecologist, showed that women only ovulate once per menstrual cycle. In the 1920s, Kyusaku Ogino, a Japanese gynecologist, and Hermann Knaus, from Austria, independently discovered that ovulation occurs about fourteen days before the next menstrual period. Ogino used his discovery to develop a formula for use in aiding infertile women to time intercourse to achieve pregnancy. In 1930, John Smulders, Roman Catholic physician from the Netherlands, used this discovery to create a method for "avoiding" pregnancy. Smulders published his work with the Dutch Roman Catholic medical association, and this was the first formalized system for periodic abstinence: the rhythm method. In the 1930s, Reverend Wilhelm Hillebrand, a Catholic priest in Germany, developed a system for avoiding pregnancy based on basal body temperature. This temperature method was found to be more effective at helping women avoid pregnancy than were calendar-based methods. Over the next few decades, both systems became widely used among Catholic women. Two speeches delivered by Pope Pius XII in 1951 gave the highest form of recognition to the Catholic Church's approval—for couples who needed to avoid pregnancy—of these systems. In the early 1950s, John Billings discovered the relationship between cervical mucus and fertility while working for the Melbourne Catholic Family Welfare Bureau. Billings and several other physicians, including his wife, Dr. Evelyn Billings, studied this sign for a number of years, and by the late 1960s had performed clinical trials and begun to set up teaching centers around the world. While Dr. Billings initially taught both the temperature and mucus signs, they encountered problems in teaching the temperature sign to largely illiterate populations in developing countries. In the 1970s they modified the method to rely on only mucus. The international organization founded by Dr. Billings is now known as the World Organization Ovulation Method Billings (WOOMB). The first organization to teach a symptothermal method was founded in 1971. John and Sheila Kippley, lay Catholics, joined with Dr. Konald Prem in teaching an observational method that relied on all three signs: temperature, mucus, and cervical position. Their organization is now called Couple to Couple League International. The next decade saw the founding of other now-large Catholic organizations, Family of the Americas (1977), teaching the Billings method, and the Pope Paul VI Institute (1985), teaching a new mucus-only system called the Creighton Model. Up until the 1980s, information about fertility awareness was only available from Catholic sources. The first secular teaching organization was the Fertility Awareness Center in New York, founded in 1981. Toni Weschler started teaching in 1982 and published the bestselling book "Taking Charge of Your Fertility" in 1995. Justisse was founded in 1987 in Edmonton, Canada. These secular organizations all teach symptothermal methods. Although the Catholic organizations are significantly larger than the secular fertility awareness movement, independent secular teachers have become increasingly common since the 1990s. Development of fertility awareness methods is ongoing. In the late 1990s, the Institute for Reproductive Health at Georgetown University introduced two new methods. The Two-Day Method, a mucus-only system, and CycleBeads and iCycleBeads (the digital version), based on the Standard Days Method, are designed to be both effective and simple to teach, learn, and use. In 2019, Urrutia et al. released a study as well as interactive graph over-viewing all studied fertility awareness based methods. Femtech companies such as Dot and Natural Cycles have also produced new studies and apps to help women avoid pregnancy. Natural Cycles is the first app of its kind to receive FDA approval. Most menstrual cycles have several days at the beginning that are infertile (pre-ovulatory infertility), a period of fertility, and then several days just before the next menstruation that are infertile (post-ovulatory infertility). The first day of red bleeding is considered day one of the menstrual cycle. Different systems of fertility awareness calculate the fertile period in slightly different ways, using primary fertility signs, cycle history, or both. The three primary signs of fertility are "basal body temperature" (BBT), "cervical mucus", and "cervical position". A woman practicing symptoms-based fertility awareness may choose to observe one sign, two signs, or all three. Many women experience secondary fertility signs that correlate with certain phases of the menstrual cycle, such as abdominal pain and heaviness, back pain, breast tenderness, and mittelschmerz (ovulation pains). This usually refers to a temperature reading collected when a person first wakes up in the morning (or after their longest sleep period of the day). The true BBT can only be obtained by continuous temperature monitoring through internally worn temperature sensors. In women, ovulation will trigger a rise in BBT between 0.2º and 0.5 °C. (0.5 and 1.°F) that lasts approximately until the next menstruation. This temperature shift may be used to determine the onset of post-ovulatory infertility. (See ref. 30) The appearance of cervical mucus and vulvar sensation are generally described together as two ways of observing the same sign. Cervical mucus is produced by the cervix, which connects the uterus to the vaginal canal. Fertile cervical mucus promotes sperm life by decreasing the acidity of the vagina, and also it helps guide sperm through the cervix and into the uterus. The production of fertile cervical mucus is caused by estrogen, the same hormone that prepares a woman's body for ovulation. By observing her cervical mucus and paying attention to the sensation as it passes the vulva, a woman can detect when her body is gearing up for ovulation, and also when ovulation has passed. When ovulation occurs, estrogen production drops slightly and progesterone starts to rise. The rise in progesterone causes a distinct change in the quantity and quality of mucus observed at the vulva. The cervix changes position in response to the same hormones that cause cervical mucus to be produced and to dry up. When a woman is in an infertile phase of her cycle, the cervix will be low in the vaginal canal; it will feel firm to the touch (like the tip of a person's nose); and the os—the opening in the cervix—will be relatively small, or "closed". As a woman becomes more fertile, the cervix will rise higher in the vaginal canal, it will become softer to the touch (more like a person's lips), and the os will become more open. After ovulation has occurred, the cervix will revert to its infertile position. Calendar-based systems determine both pre-ovulatory and post-ovulatory infertility based on cycle history. When used to avoid pregnancy, these systems have higher perfect-use failure rates than symptoms-based systems, but are still comparable with barrier methods, such as diaphragms and cervical caps. Mucus- and temperature-based methods used to determine post-ovulatory infertility, when used to avoid conception, result in very low perfect-use pregnancy rates. However, mucus and temperature systems have certain limitations in determining pre-ovulatory infertility. A temperature record alone provides no guide to fertility or infertility before ovulation occurs. Determination of pre-ovulatory infertility may be done by observing the absence of fertile cervical mucus; however, this results in a higher failure rate than that seen in the period of post-ovulatory infertility. Relying only on mucus observation also means that unprotected sexual intercourse is not allowed during menstruation, since any mucus would be obscured. Use of certain calendar rules to determine the length of the pre-ovulatory infertile phase allows unprotected intercourse during the first few days of the menstrual cycle while maintaining a very low risk of pregnancy. With mucus-only methods, there is a possibility of incorrectly identifying mid-cycle or anovulatory bleeding as menstruation. Keeping a BBT chart enables accurate identification of menstruation, when pre-ovulatory calendar rules may be reliably applied. In temperature-only systems, a calendar rule may be relied on alone to determine pre-ovulatory infertility. In symptothermal systems, the calendar rule is cross-checked by mucus records: observation of fertile cervical mucus overrides any calendar-determined infertility. Calendar rules may set a standard number of days, specifying that (depending on a woman's past cycle lengths) the first three to six days of each menstrual cycle are considered infertile. Or, a calendar rule may require calculation, for example holding that the length of the pre-ovulatory infertile phase is equal to the length of a woman's shortest cycle minus 21 days. Rather than being tied to cycle length, a calendar rule may be determined from the cycle day on which a woman observes a thermal shift. One system has the length of the pre-ovulatory infertile phase equal to a woman's earliest historical day of temperature rise minus seven days. Ovulation predictor kits (OPKs) can detect imminent ovulation from the concentration of lutenizing hormone (LH) in a woman's urine. A positive OPK is usually followed by ovulation within 12–36 hours. Saliva microscopes, when correctly used, can detect ferning structures in the saliva that precede ovulation. Ferning is usually detected beginning three days before ovulation, and continuing until ovulation has occurred. During this window, ferning structures occur in cervical mucus as well as saliva. Computerized fertility monitors, such as Lady-Comp, are available under various brand names. These monitors may use BBT-only systems, they may analyze urine test strips, they may use symptothermal observations, they may monitor the electrical resistance of saliva and vaginal fluids, or a combination of any of these factors. A symptohormonal method of FAM developed at Marquette University uses the ClearBlue Easy fertility monitor to determine the fertile window. The monitor measures estrogen and LH to determine the peak day. This method is also applicable during postpartum, breastfeeding, and perimenopause, and requires less abstinence than other FAM methods. Some couples prefer this method because the monitor reading is objective and is not affected by sleep quality as BBT can be. Fertility awareness has a number of unique characteristics: By restricting unprotected sexual intercourse to the infertile portion of the menstrual cycle, a woman and her partner can prevent pregnancy. During the fertile portion of the menstrual cycle, the couple may use barrier contraception or abstain from sexual intercourse. The effectiveness of fertility awareness, as of most forms of contraception, can be assessed two ways. "Perfect use" or "method" effectiveness rates only include people who follow all observational rules, correctly identify the fertile phase, and refrain from unprotected intercourse on days identified as fertile. "Actual use" or "typical use" effectiveness rates include all women relying on fertility awareness to avoid pregnancy, including those who fail to meet the "perfect use" criteria. Rates are generally presented for the first year of use. Most commonly, the Pearl Index is used to calculate effectiveness rates, but some studies use decrement tables. The failure rate of fertility awareness varies widely depending on the system used to identify fertile days, the instructional method, and the population being studied. Some studies have found actual failure rates of 25% per year or higher. At least one study has found a failure rate of less than 1% per year with continuous intensive coaching and monthly review, and several studies have found actual failure rates of 2%–3% per year. When used correctly and consistently (i.e., with perfect use) with ongoing coaching, under study conditions some studies have found some forms of FA to be 99% effective. From "Contraceptive Technology": Several factors account for typical-use effectiveness being lower than perfect-use effectiveness: The most common reason for the lower actual effectiveness is not mistakes on the part of instructors or users, but conscious user non-compliance—that is, the couple knowing that the woman is likely to be fertile at the time but engaging in sexual intercourse nonetheless. This is similar to failures of barrier methods, which are primarily caused by non-use of the method. A 2015 review found insufficient evidence to draw any conclusions about the effect of timing intercourse on the rate of live births or pregnancies, compared to regular intercourse. A study by Barrett and Marshall has shown that random acts of intercourse achieve a 24% pregnancy rate per cycle. That study also found that timed intercourse based on information from a BBT-only method of FA increased pregnancy rates to 31%–68%. Studies of cervical-mucus methods of fertility awareness have found pregnancy rates of 67–81% in the first cycle if intercourse occurred on the Peak Day of the mucus sign. Because of high rates of very early miscarriage (25% of pregnancies are lost within the first six weeks since the woman's last menstrual period, or LMP), the methods used to detect pregnancy may lead to bias in conception rates. Less-sensitive methods will detect lower conception rates, because they miss the conceptions that resulted in early pregnancy loss. A Chinese study of couples practicing random intercourse to achieve pregnancy used very sensitive pregnancy tests to detect pregnancy. It found a 40% conception rate per cycle over the 12-month study period. Pregnancy rates for sexual intercourse are also affected by several other factors. Regarding frequency, there are recommendations of sexual intercourse every 1 or 2 days, or every 2 or 3 days. Studies have shown no significant difference between different sex positions and pregnancy rate, as long as it results in ejaculation into the vagina. Regular menstrual cycles are sometimes taken as evidence that a woman is ovulating normally, and irregular cycles as evidence she is not. However, many women with irregular cycles do ovulate normally, and some with regular cycles are actually annovulatory or have a luteal phase defect. Records of basal body temperatures, especially, but also of cervical mucus and position, can be used to accurately determine if a woman is ovulating, and if the length of the post-ovulatory (luteal) phase of her menstrual cycle is sufficient to sustain a pregnancy. Fertile cervical mucus is important in creating an environment that allows sperm to pass through the cervix and into the fallopian tubes where they wait for ovulation. Fertility charts can help diagnose hostile cervical mucus, a common cause of infertility. If this condition is diagnosed, some sources suggest taking guaifenesin in the few days before ovulation to thin out the mucus. Pregnancy tests are not accurate until 1–2 weeks after ovulation. Knowing an estimated date of ovulation can prevent a woman from getting false negative results due to testing too early. Also, 18 consecutive days of elevated temperatures means a woman is almost certainly pregnant. Estimated ovulation dates from fertility charts are a more accurate method of estimating gestational age than the traditional pregnancy wheel or last menstrual period (LMP) method of tracking menstrual periods.
https://en.wikipedia.org/wiki?curid=22085
Nicaragua Canal The Nicaraguan Canal (), formally the Nicaraguan Canal and Development Project (also referred to as the Nicaragua Grand Canal, or the Grand Interoceanic Canal) was a proposed shipping route through Nicaragua to connect the Caribbean Sea (and therefore the Atlantic Ocean) with the Pacific Ocean. Scientists were concerned about the project's environmental impact, as Lake Nicaragua is Central America's key freshwater reservoir while the project's viability was questioned by shipping experts and engineers. Construction of a canal using the San Juan River as an access route to Lake Nicaragua was first proposed in the early colonial era. The United States abandoned plans to construct a waterway in Nicaragua in the early 20th century after it purchased the French interests in the Panama Canal. In June 2013, Nicaragua's National Assembly approved a bill to grant a 50-year concession to finance and manage the project to the private HK Nicaragua Canal Development Investment (HKND) headed by Wang Jing, a Chinese billionaire. The concession could have been extended for another 50 years once the waterway was operational. In 2015, media reports suggested the project would be delayed and possibly cancelled because Wang's personal wealth declined greatly as a result of the 2015–16 Chinese stock market crash. "Major works" such as dredging were to take place after the finishing of a Pacific Ocean wharf, whose construction was planned to start in late 2016. The Nicaraguan government failed to present reliable information about whether or not the project can be financed, thus casting doubt over whether it would be completed. The HKND Group stated that financing would come from debt and equity sales and a potential initial public offering (IPO). By May 2017, no concrete action had been reportedly taken constructing the canal and further doubts were expressed about its financing. In February 2018, analysts widely viewed the project as defunct, though the head of the project insisted work was on-going and HKND retained the legal rights to the concession for the canal as well as side projects. Despite HKND vanishing, the Nicaraguan government indicates that it will go ahead with the 908 km2 dry land expropriations anywhere within Nicaragua, under land expropriation Canal Law 840. The idea of constructing a man-made waterway through Central America has been thought about throughout history. The colonial administration of New Spain conducted preliminary surveys. The routes suggested usually ran across Nicaragua, Panama, or the Isthmus of Tehuantepec in Mexico. The history of attempts to build a Nicaragua canal connecting the Caribbean Sea and thus the Atlantic Ocean and the Pacific Ocean goes back at least to 1825 when the Federal Republic of Central America hired surveyors to study a route via Lake Nicaragua, above sea level. Many other proposals have followed. Despite the operation of the Panama Canal, which opened in 1914, interest in a Nicaragua canal has continued. With emergence of globalization, an increase in commerce and the cost of fuel, and the limitations of the Panama Canal, the concept of a second canal across the American land bridge became more attractive, and in 2006 the president of Nicaragua, Enrique Bolaños, announced an intention to proceed with such a project. Even with the Panama Canal expansion project, which began commercial operation to modern New Panamax vessels on 26 June 2016, some ships would be too big for the Panama Canal. On 26 September 2012, the Nicaraguan Government and the newly formed Hong Kong Nicaragua Canal Development Group (HKND) signed a memorandum of understanding that committed HKND to financing and building the "Nicaraguan Canal and Development Project". HKND Group is a private enterprise. The Nicaraguan government subsequently approved the "Master Concession Agreement" with HKND on 13 June 2013 thereby granting "the sole rights to the HKND Group to plan, design, construct and thereafter to operate and manage the Nicaragua Grand Canal and other related projects, including ports, a free trade zone, an international airport and other infrastructure development projects." The agreement would have lasted for 50 years and was renewable for another 50 years. HKND would have paid the Government of Nicaragua US$10M annually for 10 years, and thereafter a portion of the revenue starting at 1% and increasing later. Stratfor indicated that after 10 years, ownership shares would periodically be handed over to Nicaragua, so that after 50 years Nicaragua would be the majority shareholder. HKND Group performed a preliminary study phase of development to assess the technological and economic feasibility of constructing a canal in Nicaragua, as well as the potential environmental, social, and regional implications of various routes. The canal and other associated projects would be financed by investors throughout the world and would generate jobs for Nicaragua and other Central American countries. Initial findings of the commercial analysis conducted by HKND Group indicate that the combined effect of growth in east–west trade and in ship sizes could provide a compelling argument for the construction of a second canal, substantially larger than the expanded Panama Canal, across Central America. In the 2020s, growth in global maritime trade is expected to cause congestion and delays in transit through the Panama Canal without a complementary route through the isthmus, and by 2030, the volume of trade that a Nicaragua Canal could serve would have grown by 240%. On 10 June 2013, The Associated Press reported that the National Assembly's Infrastructure Committee voted nearly unanimously in favor of the project, with four members abstaining. On 13 June, Nicaragua's legislature passed the legislation granting the concession. On 15 June, Nicaraguan President Daniel Ortega and the billionaire chairman of HKND Group, Wang Jing, signed the concession agreement giving HKND Group the rights to construct and manage the canal and associated projects for 50 years. An HKND Group press release read, "HKND Group Successfully Obtains Exclusive Right to Develop and Manage Nicaragua Grand Canal for 100 Years." Under the exclusive contract, Wang can skip building the canal (and making any payments to Nicaragua) and instead simply operate lucrative tax-free side projects. Wang announced at a press briefing in June 2013 that he had successfully attracted global investors to the $40 billion project. In January 2014, Wang and President Ortega issued a statement that the project's construction would begin in December 2014, and that it would be completed in 2019. On 7 July 2014, a route for the Nicaragua Canal was approved. The route starts from the mouth of the Brito River on the Pacific side, passes through Lake Nicaragua, and ends in the Punta Gorda River on the Caribbean. The proposed canal would be between 230 meters and 520 meters (754.6 feet and 1,706 feet) wide and deep. The "Toronto Star" noted that Chinese engineer Dong Yung Song said the canal's design called for the creation of a artificial lake. The water to fill the canal's giant locks would come from the artificial lake, not from Lake Nicaragua. Daniel Ortega whose government approved the agreement within one week in June 2013, reportedly perceived the canal as the second phase of the Nicaraguan Revolution, predicting that it would pull Nicaragua out of poverty and lead to the creation of 250,000 jobs, but HKND said the project would create 50,000 jobs, though about half would come from abroad, mainly China. "The Moscow Times" reported in 2014 that Russia was willing to help build the Nicaragua Canal, viewing the project in part as an opportunity to pursue strategic interests in the region. Construction was to begin on 29 December 2014, and officially started a week earlier. However, owing to Nicaragua's volatile climate and seismic activity, feasibility concerns emerged over the project's future. In November 2015, HKND announced that there would be a delay in the construction of locks and excavations until late 2016 in order to fine-tune the design. The Nicaragua canal project saw business rivalry greatly intensify in late 2014. China Harbor Engineering Company, an experienced construction company, offered to design, construct, and finance a fourth set of locks in Panama, where it opened a regional headquarters. If built to the width of the proposed Nicaragua Canal, it would cut across far fewer kilometers, and still cost only $10 billion, according to the firm. Panama is in a much better financial situation than Nicaragua to afford taking on such debt and already has a stream of income from its existing canals. Alternative motives have been explored and other projects have taken priority against the water transportation venture. Bloomberg reported in 2015 that "conspiracy theories abound" including the project was a land grab by Ortega, an attempt by Ortega to "whip up" support in elections, and part of a Chinese plan to gain influence in the region. By November 2016, the president of the canal commission, Manual Coronel Kautz said "According to our schedule, we should initiate major works by the end of the year." However, Carlos Fernando Chamorro, editor of the Confidencial newspaper, said "If the People's Republic of China does not step forward, it won't happen. Wang Jing does not have the reputation to push this through. If it is just him, then the chances of this happening are zero. If the PRC steps in, then it is a big possibility." Following financial difficulties, HKND finally closed its Headquarters offices in China in April 2018, leaving no forwarding address or telephone numbers to be reached. Despite HKND abandoning its attempt to construct the canal, the Nicaragua government indicated that it will go ahead with the vast land expropriations (2,900 km2) under land expropriation Canal Law 840 enacted in 2013, which includes a concession for carrying out seven sub-projects, among them ports, oil pipelines, free-trade zones, and developing tourist areas that could be realized in any part of the national territory. In particular, this law denies any right to appeal against the expropriation decision and provides a derisory level of compensation. It also allows the investor (HKND) to buy and sell its rights over the various sub-projects "in parts", which is a highly profitable enterprise. This has been called a "land grab" and it has prompted protests, and some violent confrontations against security forces. Activists noted that the canal contract established that it must be dissolved in 72 months, if the investor has not obtained the money to start the project; that deadline was 14 June 2019, so they assert that the Law 840 must be repealed. Protests against the canal's construction occurred shortly after the official ceremony marking its beginning. Farmers feared it could cause their eviction and land expropriation. Opposition leader Eliseo Nuñez has called the deal "part of one of the biggest international scams in the world". Legal challenges that the deal violates constitutional rights were rejected by the Supreme Court of Nicaragua and a retrospective rewriting of the Constitution of Nicaragua placed HKND beyond legal challenge. HKND has been granted the right to expropriate land within on each side of the canal and pay only cadastral value, not market value, for property. Wang, however, promised to pay fair market value. The estimates of the number of people who would be displaced range from 29,000 to more than 100,000. There are indications of local opposition to intended expropriations. Thus, according to an activist leader, an unrest in Rivas in December 2014, in opposition to the canal, left two protesters dead, although no evidence was ever produced to justify his claim. The CIDH, Nicaragua's Human Rights Commission, has strongly criticized the government for not looking into the project's effect on citizens, amid claims that citizens were not involved in decision-making. The British firm ERM, who carried out the Environmental Impact Assessment, claims it held consultations with around 6,000 people in the communities along its planned route, and estimated that the property of about 30,000 people would be affected. National opinion polls show that support for the project is about 70%. Investor Wang had financial setbacks unrelated to the Nicaragua project, losing 80% of his net worth during the 2015–16 Chinese stock market turbulence. In March 2017, the "Havana Times" reported that the public relations agency handling Wang's interests in Nicaragua had been let go, in absence of any developments on the project to report, and Wang had not been in the country in more than two years. In May 2017, the "PanAm Post" indicated that "no concrete action has been taken to begin the project" and suggested that the project was either "paralyzed, or nonexistent." In September 2017 Agence France-Presse reported that the work had been "pushed back indefinitely," although the government renewed the project's environmental permit in April 2017. In February 2018, Manuel Coronel Kautz, head of Interoceanic Grand Canal Authority of Nicaragua, told Agence France Presse work on the canal was still ongoing, but by that point analysts and activists widely viewed the canal project as defunct, with China having shifted its investment focus to Panama, the main competitor to a Nicaraguan canal. Following financial difficulties, HKND finally closed its offices in April 2018, leaving no forwarding address or telephone numbers to be reached. Absent a 60% vote to revoke the legislation, HKND maintains the legal concessions established by the 2013 law, including for other infrastructures projects in Nicaragua, including ports, roads, railway and an airport. The construction company provided a project description for review on open source, dated December 2014. The canal as planned would have been and would have three sections. The West Canal runs from Brito on the Pacific Ocean up the Rio Brito valley, crosses the continental divide, and after passing through the Rio Las Lajas valley enters Lake Nicaragua; its length would be . The Nicaragua Lake section measures and runs from south of San Jorge to south of San Miguelito. The Eastern Canal would be the longest section at and would be built along the Rio Tule valley through the Caribbean highland to the Rio Punta Gorda valley to meet the Caribbean Sea. A channel would have to be dug in the lake bottom, as it is not deep enough for large vessels to transit the canal. Both the West Canal and the East Canal would each have one lock with 3 consecutive chambers to raise ships to the level of Lake Nicaragua that has an average water elevation of 31.3 m, range 30.2–33.0 m. The western Brito Lock would be inland from the Pacific, and the eastern Camilo Lock would be inland from the Caribbean Sea. The dimensions of each of the locks' chambers are long, wide, and threshold depth. As locks generally define the limit on the size of ships that can be handled, the Nicaragua Canal would have allowed passage for larger ships than those that pass through the Panama Canal. For comparison, the new third set of locks in the Panama expansion will only be long, wide, and deep. No water from Lake Nicaragua was planned to be used to flood the locks; water would have come from local rivers and recycling using water-saving basins. The Camilo lock would have been built adjacent to a new dam of the upper Punta Gorda River that creates a reservoir. This Atlanta Reservoir (or Lake Atlanta) would have a surface area of 395 km2. West of the Atlanta reservoir, the Rio Agua Zarca would have been dammed to create a second reservoir. This reservoir would have had a surface area of 48.5 km2 and hold 1,100 gigalitres. A hydropower facility would be built at the dam and would have generated over 10 megawatts of power to be used for Camilo Lock operations. Both locks would also be connected to the country's power grid and have back-up generator facilities. It was estimated that each lock would have used about 9 megawatts of power. At each oceanic canal entrance, breakwaters and port facilities would have been constructed. The Pacific port would be named Brito Port and the Caribbean one Aguila Port. Initially these two ports would have helped during construction and later become international ports. Their design capacity was 1.68 million TEU/year and 2.5 million TEU/year, respectively. Existing port facilities at Corinto and Bluefields would have been improved to allow for shipment of material to the entry ports under construction. Fuel storage sites would be placed at the two port sites. Four lighthouses would be constructed at the entrances to the East and West Canals. In addition, the channel entrance on sea would have been be marked on both sides with a large sailing buoy about offshore and 2 light buoys would mark the passage through Lake Nicaragua. A free trade zone with commercial facilities as well as tourist hotels and an international airport at Rivas were planned to be built when canal construction was advanced. Appropriate road improvements were planned. The Pan-American Highway would have crossed the canal via a bridge. Nicaragua Route 25 (Acoyapa-San Carlos) on the eastern side of Lake Nicaragua would have gotten a ferry service. Both ports would get public road connections. HKND plans to construct a private gravel maintenance road on both sides of the canal. The estimate for the workforce in 2020 was 3,700 people, and 12,700 in 2050 when traffic had increased. Transit time would have been about 30 hours. It was projected that by 2020 3,576 ships would have transited the canal annually. The transit rate was expected to increase to 4,138 by 2030, and to 5,097 by 2050. For comparison, the Panama Canal handled 12,855 transits in 2009. No significant construction took place. No "major works" such as dredging were planned to take place until after a Pacific Ocean wharf was built. The apparent lack of experience of Wang and his HKND Group in large-scale engineering was cited as a risk. On December 22, 2014, Wang announced construction started in Rivas, Nicaragua. Wang spoke during the starting ceremony of the first works of the Interoceanic Grand Canal in Brito town. Construction of the new waterway would have been by HKND Group—Hong Kong–based HK Nicaragua Canal Development Investment Co Ltd., which is controlled by Wang. According to HKND's announced plans in 2015, the project entailed the canal's development and building, and a supporting infrastructure. There would have been four main phases. The preconstruction phase included getting permits, acquiring land and machinery, and finalizing designs and plans. The early construction phase, started in December 2014, lasted through September 2015; it secured access to construction sites, but it did not provide the critical infrastructure nor mobilized the workforce. During the construction phase from September 2015 to March 2020, the canal would have been dug and the locks built along with accompanying infrastructure. The commissioning phase projected from April 2020 to June 2020 included lock testing and lock and tug boat operator training. HKND described the project as the largest civil earth-moving operation in history. Most of this would have consisted of dry excavation to form the canal with an estimate of 4,019 million cubic metres of rock and soil. There would have been 739 MCM of freshwater dredging (Lake Nicaragua) and 241 MCM of marine dredging. Marine dredging of the oceanic access canal would be required on the Pacific side for 1.7 km and on the Caribbean Sea for 14.4 km. Disposal of excavation material would have been done along the canal in designated disposal areas typically within 3 km of the canal. Two concrete plants and a steel plant were planned to support the project. While cement would have likely been imported, construction aggregate would have come from local quarries near the two locks. HKND estimates that about 50,000 people would be employed during the five-year construction, about half of them from Nicaragua, 25% from China, and the remainder from various other countries. 1,400 workers would be in office or administrative positions and the rest in the field. The management offices would be rented or purchased near Rivas. Workers would live in one of nine camps, which besides food and shelter would also provide health care and security. These are “closed” camps—that is, workers cannot leave the camp unless part of an organized activity. The work schedule calls for 12-hour shifts for seven days a week. Domestic workers work two weeks and get one week off, while foreign workers are six weeks on and get two weeks off (management) or 22 weeks on, four weeks off (blue collar workers). On 2 September 2015, Pang Wai Kwok (executive VP of HKND Group) was interviewed by Nicaraguan journalist Carlos Solis and said up to 3,000 people might be employed on the canal project within the year. However, the labor force depends on the contract bid's winner and Kwok said anyone in the world is eligible to work on the canal. Project costs were estimated in the region of $40 billion to $50 billion Beside private money provided by Wang at the start-up, further influx of financial support was expected from investors. An IPO was reported to be in preparation by the end of 2014. XCMG, a state-owned Chinese construction company would have provided machinery and take 1.5% to 3% of HKND shares in return. By the end of 2014, no major investors had been named. There had been speculation that the Chinese government would provide financial backing for the project, but China, as well as Wang, denied this. Wang lost nearly 85% of his wealth during the 2015 Chinese stock market crash, according to the Bloomberg Billionaires Index. In addition, Wang has had a string of setbacks for projects around the world since 2014. The economic development potential for the canal project is relatively measurable with Panama; however, the World Bank describes the country of Nicaragua as the second poorest in Latin America and the Caribbean. The World Bank has compiled a data list of projects that the impoverished nation has on record and the majority of the efforts are geared towards infrastructure and agricultural needs, but there is no explicit title project that would support the canal line of effort. Wang admitted that the project has financial, political, and engineering risks. With the high cost of the project that independently has been estimated to be about $100 billion, it was not fully funded. The project was expected in 2014 to be completed in 2020, but Stratfor, an analyst agency, stated then that was an "unrealistic goal." While the Nicaraguan National Bank reserves served as collateral on the part of Nicaragua, HKND had no such potential liability. Following financial difficulties, HKND finally closed its Headquarter offices in China in April 2018, leaving no forwarding address or telephone numbers to be reached. Some of the natural habitat of at least 22 endangered species would be destroyed in the construction. Another major environmental concern is the project's impact on Lake Nicaragua, the largest source of freshwater in Nicaragua. An oil spill would have serious and lasting consequences. Other problems include the possibility of dredging bringing up toxic sediments, the disruption of migration patterns of animal species, and the potential to introduce invasive species to the Lake. Environmental studies had not been released by HKND when the project officially started in December 2014. The Nicaraguan Academy of Sciences noted that hundreds of thousands of hectares of pristine forests and wetlands would be destroyed and pointed out that the environmental study performed for the canal was not independent. President Daniel Ortega stated that he is "not concerned about harming the lake because it is already contaminated." Protesters fear that the canal would bring massive environmental destruction to Lake Nicaragua and the Atlantic Autonomous Regions. 400,000 hectares of tropical rain forest and wetlands would be destroyed. It would also encroach upon the habitats of animals such as Baird's tapir, the spider monkey, and the jaguar. Richard Condit, from the Smithsonian Tropical Research Institute, believes that the project could be used as leverage for forest protection in a country that currently lacks "institutional capacity" to meet conservation needs. A Canadian pilot was the first fatality during the canal project. Atkinson was flying alone on the western side of Lake Nicaragua during an aerial survey. The survey site was on the same side as NicarAgua–Dulce, which is the only ecotourism group in Nicaragua that is affiliated with The International Ecotourism Society, and it is located north of the proposed canal site. Falling in line with ecotourism, Nicaragua's Ministry of Environment and Natural Resources has promoted formal workshops at each level of education (primary, secondary, and post-secondary); however, there is no curriculum relevant to the pending canal project. The American-led Foundation for Sustainable Development is another partner that provides training initiatives to Nicaraguans that cannot access formal education. One of FSD's support sites is located at Tola, which is within close proximity of the proposed Brito–Pacific canal opening. As the original Panama Canal still has capacity for Panamax-sized shipping and Panama has completed its Panama Canal expansion project, adding more capacity and allowing transit for even larger New Panamax size ships, projections for the Nicaragua canal's traffic may be optimistic. While the proposed Nicaragua Canal would be wide enough to accommodate Triple E class of mega container ships, which are too wide for the expanded Panama Canal, few ports are able to handle these ships at the present. Further, a coast-to-coast railway line may be built by China in Honduras and could affect utilization of the Nicaragua Canal. Also, North American land bridges in Mexico and the United States will compete in the traffic between Asia and the U.S. east coast. Thus, competition may undermine the Nicaragua Canal's economic viability if it was ever built. The canal would affect neighboring economies, like Honduras and El Salvador, as they are part of the commercial treaty known as North Triangle of Central America (Triángulo Norte de Centroamérica). The GDP of each nation would be influenced by expanded export/import operations and trade cooperation through agencies like the promotion authority in El Salvador. According to the official Environmental Impact Assessment carried out by ERM, the canal would require the relocation of around 30,000 people. However, according to human rights group Amnesty International it would "forcibly displace an estimated 120,000 people, including Rama and Creole communities from protected indigenous territories on the Caribbean coast." The report claims that communities established in the area of canals have been visited by foreigners, guarded by Nicaraguan authorities, measuring the lands of the inhabitants, and that laws passed by the Ortega government "authorizes HKND to expropriate whatever land it wants, while denying displaced families the right to appeal." Amnesty International also states that excessive forces and unjust arrests have been performed by Nicaraguan officials. However, the 2015 Nicaraguan government's report on the canal states: "The Nicaraguan government and HKND will guarantee that persons and families on the route of the canal's construction will have living conditions superior to those they currently have [without the canal]."
https://en.wikipedia.org/wiki?curid=22087
Nu metal Nu metal (sometimes stylized as nü-metal) is a subgenre of that combines elements of with elements of other music genres such as hip hop, alternative rock, funk, industrial, and grunge. bands have drawn elements and influences from a variety of musical styles, including multiple genres of heavy metal. rarely features guitar solos; the genre is heavily syncopated and based on guitar riffs. Many nu metal guitarists use that are down-tuned to play a heavier sound. DJs are occasionally featured in nu metal to provide instrumentation such as sampling, turntable scratching and electronic backgrounds. Vocal styles in include singing, rapping, screaming and growling. Nu metal is one of the key genres of the new wave of American heavy metal. Nu metal became popular in the late 1990s with bands and artists such as Korn, Limp Bizkit, and Kid Rock all releasing albums that sold millions of copies. Nu metal's popularity continued during the early 2000s, with bands such as Papa Roach, Staind, and P.O.D. all selling multi-platinum albums, and came to a peak with Linkin Park's diamond-selling album "Hybrid Theory". However, by the the oversaturation of bands combined with the under-performance of a number of releases led to nu metal's decline, leading to the rise of metalcore and many nu metal bands disbanding or abandoning their established sound in favor of other genres. During the 2010s, there was a nu metal revival; many bands that combine nu metal with other genres (for example, metalcore and deathcore) emerged and some nu metal bands from the 1990s and early 2000s returned to the nu metal sound. Bands like Staind, Linkin Park and Papa Roach went back to making nu metal songs after abandoning the genre. Bands like Of Mice & Men, Emmure, Issues and My Ticket Home would combine nu metal with metalcore or deathcore. In the 2010s, nu metal bands like From Ashes to New, Islander and Ded emerged. Nu metal received criticism from many fans of heavy metal and nu metal was often labelled with pejorative words like "mallcore". Some nu metal musicians did not view their own music as heavy metal while other nu metal musicians rejected the nu metal label. Nu metal is also known as nü-metal and aggro-metal. It is a subgenre of alternative metal. MTV states that the early nu metal group Korn "arrived in 1993 into the burgeoning alternative metal scene, which would morph into nü-metal the way college rock became alternative rock." "Stereogum" has similarly claimed that nu metal was a "weird outgrowth of the Lollapalooza-era alt-metal scene". Nu metal merges elements of heavy metal music with elements of other music genres such as grunge, hip hop, and alternative rock. Nu metal bands have been influenced by and have used elements of a variety of musical genres, including electronic music, funk, gothic rock, hardcore punk, punk rock, dance music, new wave, jazz, post-punk, symphonic rock and synth-pop. Nu metal bands also are influenced by and use elements of genres of heavy metal music such as death metal, rap metal, groove metal, funk metal, and thrash metal. Some nu metal bands, such as Static-X and Dope, made nu metal music with elements of industrial metal. In contrast with other heavy metal subgenres, nu metal tends to use the same structure of verses, choruses and bridges as those in pop music. Nu metal is heavily syncopated and is based mostly on guitar riffs. Mid-song bridges and a general lack of guitar solos contrasts it with other genres of heavy metal. Kory Grow of "Revolver" wrote, "... [i]n its efforts to tune down and simplify riffs, effectively drove a stake through the heart of the guitar solo". Another contrast with other heavy metal genres is nu metal's emphasis on rhythm, rather than on complexity or mood, often its rhythm sounds like that of groove metal. The wah pedal is occasionally featured in nu metal music. Nu metal guitar riffs occasionally are similar to those of death metal. Nu metal bassists and drummers are often influenced by funk and hip hop, respectively, adding to nu metal's rhythmic nature. Blast beats, which are common in heavy metal subgenres such as black metal and death metal, are extremely rare in nu metal. Nu metal's similarities with many heavy metal subgenres include its use of common time, distorted guitars, power chords and note structures primarily revolving around Dorian, Aeolian or Phrygian modes. While loud and heavily distorted electric guitars are a core feature of all metal genres, nu metal guitarists took the sounds of "violence and destruction" to new levels with their overdriven guitar tone, which music journalists Kitts and Tolinski compared to the "...sound [of] a Mack truck being crushed by a collapsing skyscraper." Some nu metal bands use seven-string guitars that are generally down-tuned, rather than traditional Likewise, some bass guitarists use five-string and six-string instruments. in nu metal often features an emphasis on funk elements. In nu metal music, DJs are sometimes featured to provide instrumentation such as sampling, turntable scratching and electronic backgrounds. tends to have hip hop grooves and rhythms. Vocal styles used in nu metal music include singing, rapping, screaming and growling. Vocals in nu metal are often rhythmic and influenced by hip hop. While some nu metal bands, such as Limp Bizkit and Linkin Park have rapping in their music, other nu metal bands, such as Godsmack and Staind, do not. Nu metal bands occasionally feature hip hop musicians as guests in their songs; Korn's song "Children of the Korn" features the rapper Ice Cube, who performed on the band's 1998 Family Values Tour. The hip hop musician Nas was featured on Korn's song "Play Me", which is on the band's album "Take a Look in the Mirror". Limp Bizkit has recorded with multiple hip hop musicians including Method Man, Lil Wayne, Xzibit, Redman, DMX and Snoop Dogg. Linkin Park collaborated with hip hop musician Jay Z on their 2004 extended play "Collision Course". Kid Rock has recorded with hip hop musicians Eminem and Snoop Dogg. Trevor Baker of "The Guardian" wrote, "Bands such as Linkin Park, Korn and even the much reviled Limp Bizkit ... did far more to break down the artificial barriers between 'urban music' and rock than any of their more critically acceptable counterparts." Lyrics in nu metal songs are often angry or nihilistic; many of the genre's lyrics focus on topics such as pain, angst, bullying, emotional issues, abandonment, betrayal, and personal alienation, in a way similar to those of grunge. Many nu metal lyrics that are about these topics tend to be in a very direct tone. However, some songs have lyrics that are about other topics. P.O.D. have used positive lyrics about promise and hope. The nu metal song "Bodies" by Drowning Pool is about moshing. Saliva guitarist Wayne Swinny said that the band's song "Badass" was "meant to be one of those 'sports anthem kind of songs'". "The Michigan Daily" wrote about Limp Bizkit's lyrics, writing that the band "used the nu-metal sound as a way to spin testosterone fueled fantasies into snarky white-boy rap. Oddly, audiences took frontman Fred Durst more seriously than he wanted, failing to see the intentional silliness in many of his songs". Limp Bizkit's lyrics also have been described as misogynistic. Dope's lyrics are usually about sex, drugs, parties, women, violence and relationships. According to Josh Chesler of the "Phoenix New Times", the lyrics of Deftones, who were once a nu metal band, "tend to have complex allusions and leave the songs open to many different interpretations." Nu metal clothing typically consists of baggy pants, shirts, and shorts, JNCO jeans, Adidas tracksuits, sports jerseys, baseball caps, baggy hoodies, cargo pants, and sweatpants. Nu metal hairstyles and facial hairstyles include dreadlocks, Braids, spiky hair, chin beards, bald heads, goatees, frosted tips, and bleached or dyed hair. Common accessories in nu metal fashion include wallet chains, tattoos, and piercings, especially facial piercings. Nu metal fashion has been compared to hip hop fashion. Some nu metal bands such as Motograter, Mushroomhead, Mudvayne, and Slipknot wear masks, jumpsuits, costumes, face paint, corpse paint or body paint. A few nu metal bands, such as Coal Chamber, Evanescence, and Kittie, are known for having gothic appearances. Many heavy metal, alternative metal, industrial, funk metal, alternative rock, rap metal, and industrial metal artists and bands of the 1980s and early 1990s have been credited with laying groundwork for the development of nu metal by combining heavy guitar riffs with pop music structures and drawing influences from subgenres of heavy metal and other music genres; Faith No More, Primus, Helmet, Boo-Yaa T.R.I.B.E., Tool, Fear Factory, 24-7 Spyz, Hot Dawgz, Fishbone, Biohazard, Suicidal Tendencies, Infectious Grooves, Godflesh, Red Hot Chili Peppers, Nine Inch Nails, White Zombie, Mr. Bungle, Prong, Rage Against the Machine, and Ministry all have been highlighted as examples of this. Groove metal and thrash metal bands of the same period such as Machine Head, Sepultura, Metallica, Pantera, Slayer, and Anthrax all have been cited as influential to nu metal as well. For example, Anthrax pioneered the rap metal genre by combining hip hop and rap with heavy metal on their 1987 EP "I'm the Man", which laid groundwork for development. Korn's lead vocalist Jonathan Davis said about Pantera guitarist Dimebag Darrell, "if there was no Dimebag Darrell, there would be no Korn". In the 1990s, bands described as "neo-metal" by the author Garry Sharpe-Young emerged; these bands include Pantera, Strapping Young Lad, Machine Head, Biohazard and Fear Factory. Sharpe-Young wrote that these bands "had chosen to strip metal down to its raw, primal element" and that "neo-metal paved the way for nu-metal". Nu metal is often influenced by hip hop. hip hop musicians Dr. Dre and Ice Cube have been a big influence on nu metal creators and pioneers Korn; guitarist Munky said the band were trying to emulate the samples of Dr. Dre's 1992 album "The Chronic". Munky and fellow Korn guitarist Head also said they tried to emulate samples by the hip hop group Cypress Hill. Both the Geto Boys and N.W.A. also have been a major influence on Korn. Fred Durst of Limp Bizkit has cited the hip hop group The Fat Boys as a major influence on him. The nu metal band Papa Roach cited rapper Nas and hip hop groups and Fugees as influences. Shifty Shellshock of the nu metal band Crazy Town cited Run–D.M.C. and Beastie Boys as influences. Josey Scott of the nu metal band Saliva cited LL Cool J, Beastie Boys, Public Enemy, N.W.A., Chuck D, Doug E. Fresh, and Whodini as influences. Sonny Sandoval of the nu metal band P.O.D. cited hip hop groups Boogie Down Productions and Run–D.M.C. as influences. Linkin Park member Mike Shinoda's hip hop influences include Boogie Down Productions, Public Enemy, N.W.A., and the Juice Crew. Chester Bennington, another member of Linkin Park, cited A Tribe Called Quest, KRS-One, Run–D.M.C., Public Enemy, N.W.A., Beastie Boys, and Rob Base as influences. Beastie Boys are a hip hop music group that influenced nu metal. Hip hop group Run–DMC was one of the first groups to combine rap with rock, paving the way for nu metal. Joel McIver acknowledged Korn as the band that created and pioneered the nu metal genre with its demo "Neidermayer's Mind", which was released in 1993. McIver also acknowledged Korn as the band that started the new wave of American heavy metal, which is a heavy metal music movement that started in the 1990s. The aggressive riffs of Korn, the rapping of Limp Bizkit, and the melodic ballads of Staind created the sonic template for nu metal. The origins of the term "nu metal" are often attributed to the work of producer Ross Robinson, who has been called "The Godfather of Nu Metal" between producers. Robinson has produced for nu metal bands such as Korn, Limp Bizkit and Slipknot. Many of the first nu metal bands, such as Korn and Deftones, came from California; however, the genre soon spread across the United States and many bands arose from various states, including Limp Bizkit from Florida, Staind from Massachusetts, and Slipknot from Iowa. In the book "Brave Nu World", Tommy Udo wrote about the nu metal band Coal Chamber, "There's some evidence to suggest that Coal Chamber were the first band to whom the tag 'nu metal' was actually applied, in a live review in Spin magazine." In 1994, Korn released their self-titled debut album, which is widely considered the first popular nu metal album. Korn had experienced underground popularity at this time; their debut album peaked at number 72 on the "Billboard" 200. However, earlier the same year, P.O.D.'s album "Snuff the Punk" was also released, which was later recognized as the first nu metal album. In 1995, the band Sugar Ray released its debut studio album "Lemonade and Brownies", an album described as nu metal. In 1995, Deftones released their debut album "Adrenaline". The album peaked at number 23 on the Heatseekers Albums chart on October 5, 1996. Deftones also were temporarily controversial in 1996 when their vocalist Chino Moreno was blamed by TV news reports for a riot that occurred at the 1996 U-Fest festival. Deftones' 1997 album "Around the Fur" peaked at number 29 on the "Billboard" 200 on November 15, 1997. Both "Adrenaline" and "Around the Fur" were certified gold by the Recording Industry Association of America (RIAA) in the summer of 1999."Adrenaline" and "Around the Fur" were certified platinum by the RIAA in September 2008 and June 2011, respectively. Sepultura's 1996 album "Roots" features nu metal elements that were considered influential to the genre, while "Roots" itself was influenced by Korn's self-titled debut album. Few bands were playing nu metal until 1997 when bands such as Coal Chamber, Limp Bizkit, and Papa Roach all released their debut albums. Attention through MTV and Ozzy Osbourne's 1995 introduction of Ozzfest was integral to the launching of the careers of many nu metal bands, including Limp Bizkit in 1998. Nu metal began to rise in popularity when Korn's 1996 album "Life Is Peachy" peaked at number 3 on the "Billboard" 200 and sold 106,000 copies in its first week of release. In 1997, Sugar Ray released its second studio album "Floored". The album achieved mainstream success very quickly and was certified 2x platinum by the RIAA on February 20, 1998. Although "Floored" is a nu metal album, the only song from the album that achieved chart success was the song "Fly", which is instead a reggae song. Although Sugar Ray continued to be extremely popular, the band abandoned the nu metal genre and became a pop rock band with its 1999 studio album "". In 1998, nu metal became one of the most mainstream genres of music when Korn's third album "Follow the Leader" peaked at number 1 on the "Billboard" 200, was certified 5x platinum by the RIAA, and paved the way for other nu metal bands. At this point, many nu metal bands were signed to major record labels, and were playing combinations of heavy metal, hip hop, industrial, grunge and hardcore punk styles. Hip hop artists Vanilla Ice and Cypress Hill, along with heavy metal bands Sepultura, Primus, Fear Factory, Machine Head, and Slayer released albums that draw from the nu metal genre. In 1999, Korn's fourth studio album "Issues" peaked at number 1 on the "Billboard" 200. The album was certified 3× platinum in one month. The album sold at least 573,000 copies in its first week of release and its first single "Falling Away From Me" peaked at number 8 on the Bubbling Under Hot 100 Singles chart. A little before the album was released, Korn appeared on an episode of "South Park" titled "Korn's Groovy Pirate Ghost Mystery", in which "Falling Away from Me" was premiered. During the late 1990s and early 2000s, multiple nu metal bands such as Korn, Limp Bizkit and P.O.D. appeared repeatedly on MTV's "Total Request Live". The Woodstock 1999 festival featured multiple nu metal artists and bands such as Korn, Kid Rock, Godsmack, Limp Bizkit and Sevendust. During and after Limp Bizkit's performance at the festival, violence occurred and people tore plywood from the walls during the performance of the band's song "Break Stuff". Several sexual assaults were reported to have happened during the festival; a rape that was reported during Limp Bizkit's performance, and gang rape was reported to have occurred during Korn's set at the festival. Despite the incidents at the festival, Limp Bizkit's popularity and the sales of their then-recent album "Significant Other" were not affected. The album peaked at number 1 on the "Billboard" 200, selling 643,874 copies in its first week of release, topping over one million sold in two weeks, and eventually being certified 7x platinum in 2001. "Significant Other" sold at least 7,237,123 copies in the United States. Orgy became popular in the late 1990s with their album "Candyass", which was certified platinum by the RIAA in July 1999. The band's cover of "Blue Monday" by New Order peaked at number 56 on the "Billboard" Hot 100. Godsmack's was released in 1998 and was certified 4× platinum in December 2001. In April 1999, Kid Rock's album "Devil Without a Cause" was certified by gold by the RIAA. The following month, "Devil Without a Cause", as Kid Rock predicted, went platinum. The album sold at least 9,300,000 copies in the United States and was certified 11x platinum. In 1999, Slipknot emerged with an extremely heavy nu metal sound, releasing their self-titled album, which was certified platinum in 2000 and 2x platinum in 2005. In a review of the band's self-titled album, Rick Anderson of AllMusic wrote about Slipknot, "You thought Limp Bizkit was hard? They're the Osmonds. These guys are something else entirely." In 1999, Staind's second album "Dysfunction" was released; the track "Mudshovel" peaked at number 10 on the Mainstream Rock chart. "Dysfunction" was certified 2x platinum by the RIAA. In 2000, Limp Bizkit's third studio album "Chocolate Starfish and the Hot Dog Flavored Water" set a record for highest week-one sales of a rock album, selling over 1,000,000 copies in the United States in its first week of release—400,000 of which sold on its first day of release, making it the fastest-selling rock album ever and breaking the world record held for seven years by Pearl Jam's "Vs." "Chocolate Starfish and the Hot Dog Flavored Water" by Limp Bizkit was certified 6x platinum and sold at least 8,000,000 copies in the United States. That same year, both Papa Roach's second studio album "Infest" and Disturbed's debut studio album "The Sickness" were released. The RIAA certified "The Sickness" 4× platinum and "Infest" 3× platinum. Disturbed's song "Down with the Sickness" was certified platinum by the RIAA. Papa Roach's song "Last Resort" peaked at number 57 on the "Billboard" Hot 100 and at number 1 on the Modern Rock Tracks chart. In 2000, P.O.D.'s album "The Fundamental Elements of Southtown" went platinum in the United States and was the 143rd best-selling album of 2000. The album's song "Rock the Party (Off the Hook)" went to number 1 on MTV's "Total Request Live". In 2000, the hip hop group Cypress Hill released their fifth studio album "Skull & Bones", which features a and rap metal style. The album went platinum in the United States in two months. During the early 2000s, the nu metal band Incubus was very popular and made the albums "Make Yourself" and "Morning View", which both were certified 2x platinum by the RIAA. Late in 2000, Linkin Park released their debut album "Hybrid Theory", which was the best-selling debut album by any artist of any genre in the 21st century. The album was also the best-selling album of 2001, selling more than albums such as "Celebrity" by NSYNC and "Hot Shot" by Shaggy. Linkin Park earned a Grammy Award for their second single "Crawling". Their fourth single, "In the End", was released late in 2001 and peaked at number 2 on the "Billboard" Hot 100 in March 2002. In 2001, Linkin Park's album "Hybrid Theory" sold 4,800,000 copies in the United States, making it the highest-selling album of the year. Linkin Park's album "Hybrid Theory" was certified diamond by the RIAA and sold at least 10,222,000 copies in the United States. In 2000, Godsmack released their second studio album "Awake", which was certified 2x platinum in March 2002. The album's title track peaked at number 1 on the Mainstream Rock chart. Both the album's title track and the song "Sick of Life" have been featured on the United States Navy's television commercials. Crazy Town's debut album "The Gift of Game" peaked at number 9 on the "Billboard" 200, went platinum in February 2001, and sold at least 1,500,000 copies in the United States. Worldwide, the album sold at least 2,500,000 copies. Staind's 2001 album "Break the Cycle" debuted at number 1 on the Billboard 200 with at least 716,000 copies sold in its first week of release, selling more than albums such as "Survivor" by Destiny's Child, "Lateralus" by Tool and "Miss E... So Addictive" by Missy Elliott. "Break the Cycle" by Staind was certified 5x platinum by the RIAA in 2003. In March 2001, Saliva released their second album "Every Six Seconds" and the album was certified platinum. The album's song "Click Click Boom" was used as the theme song for WWE's No Mercy event of 2001. "Click Click Boom" also has been played during football games. Saliva's song "Your Disease" peaked at number 7 on "Billboard" Modern Rock Tracks chart and peaked at number 3 on "Billboard" Mainstream Rock chart. In August 2001, Slipknot released their album "Iowa", which peaked at number 3 on the "Billboard" 200 and went platinum in October 2001. Critic John Mulvey called the album the "absolute triumph of nu metal". P.O.D.'s 2001 album "Satellite" went and peaked at number 6 on the "Billboard" 200. P.O.D.'s popularity continued in the year 2002. On June 5, 2001, Drowning Pool released a nu metal album titled "Sinner", which features the song "Bodies". The album went platinum on August 23, 2001 and its song "Bodies" became one of the most frequently played videos on MTV for new bands. "Bodies" went to number 6 on the Mainstream Rock chart and was used by Boston Red Sox closer Jonathan Papelbon as his theme song. Alien Ant Farm's album "Anthology", which was released in 2001, sold at least 1,900,000 copies in the United States and was certified platinum by the RIAA the same year. Alien Ant Farm's cover of Michael Jackson's song "Smooth Criminal" peaked at number 23 on the "Billboard" Hot 100. In 2001, System of a Down's album "Toxicity" peaked at number 1 on the "Billboard" 200. In November 2002, "Toxicity" was certified 3x platinum by the RIAA. In 2002, the soundtrack album for the film "The Scorpion King" was released and peaked at number 1 on the Top Soundtracks chart; it features multiple nu metal bands such as Drowning Pool, Coal Chamber, Lifer, Sevendust, Flaw and Godsmack. Godsmack's track "I Stand Alone" was the most played active rock song in 2002 for fourteen consecutive weeks. "I Stand Alone" also peaked at number 1 on the Mainstream Rock chart. In 2003, MTV wrote that nu metal's mainstream popularity was declining, citing that Korn's fifth album "Untouchables" and Papa Roach's third album "Lovehatetragedy" both sold less than the bands' previous releases. Korn's lead vocalist Jonathan Davis blamed music piracy for the amount of sales of "Untouchables" because the album had been leaked to the Internet more than four months before its official release date. MTV also wrote that nu metal bands were played less frequently on radio stations and MTV began focusing on other musical genres. MTV wrote that Papa Roach's third album "Lovehatetragedy" has less hip hop elements than the band's previous album "Infest" and also said that Saliva's 2002 album "Back into Your System" has less elements than the band's 2001 album "Every Six Seconds". MTV also wrote that Crazy Town's second album "Darkhorse" had no hit singles and sold less than the band's previous album "The Gift of Game". MTV wrote that although Kid Rock's album "Cocky" had characteristics of the musician's 1998 album "Devil Without a Cause", "Cocky" song "Forever", which featured the style of Kid Rock's song "Bawitdaba", was not as popular as "Cocky" country song "Picture". MTV also wrote, "Another cause for nü-metal and rap-rock's slip from the spotlight could be a diluted talent pool caused by so many similar-sounding bands. American Head Charge, Primer 55, Adema, Cold, the Union Underground, Dope, Apartment 26, Hed (Planet Earth) and Skrape—all of whom released albums between 2000 and 2001—left more of a collective impression than individual ones". Despite what MTV wrote, the RIAA certified Korn's album "Untouchables" platinum in July 2002, and one of the album's singles, "Here to Stay", peaked at number 72 on the "Billboard" Hot 100, won a Grammy, received much radio play, and peaked at number one on MTV's "Total Request Live" twice. "Untouchables" sold at least 434,000 copies in first week of release and peaked at number 2 on the Billboard 200. "Thoughtless", another single from Korn's album "Untouchables", also was successful; the single reached number 11 on the "Billboard" Modern Rock Tracks chart and number 6 on the "Billboard" Mainstream Rock chart. However, "Untouchables" still did not sell as many copies as Korn's most commercially successful album, "Follow the Leader". Papa Roach's song "She Loves Me Not", which is from the band's 2002 album "Lovehatetragedy", peaked at number 76 on the "Billboard" Hot 100. Despite the MTV report that nu metal was declining, nu metal remained extremely popular with bands such as Linkin Park, Godsmack, Trapt, and Evanescence. Linkin Park's remix album "Reanimation" was released in July 2002 and sold more than a million copies that year, which MTV described as "impressive for a remix album". Trapt's 2002 song "Headstrong" launched the band into the mainstream; the song peaked at number 16 on the "Billboard" Hot 100, number 4 on the Mainstream Top 40 chart and number 1 on the "Billboard" Mainstream Rock Tracks chart. Trapt's song "Still Frame" peaked at number 69 on the "Billboard" Hot 100. The band's self-titled album was certified platinum by the RIAA in 2003. Evanescence's debut album "Fallen" was released in March 2003. Johnny Loftus of AllMusic noted the nu metal sound of the album. "Fallen" Grammy Award-winning lead single "Bring Me to Life" peaked at number 5 on the "Billboard" Hot 100 chart and number 1 on the Mainstream Top 40 chart. In 2003, Linkin Park's album "Meteora" peaked at number 1 on the "Billboard" 200 and sold at least 810,000 copies in its first week of being released. "Meteora" by Linkin Park and "Fallen" by Evanescence ranked third and fourth respectively on the best-selling albums of 2003. Both Linkin Park and Evanescence released high-charting singles throughout 2003 to "Fallen" by Evanescence sold at least 7,600,000 copies in the United States and "Meteora" by Linkin Park sold at least 6,100,000 copies in the United States. In 2003, Korn released a song called "Did My Time", which peaked at number 38 on the "Billboard" Hot 100. That same year, Godsmack released their third studio album "Faceless", which peaked at number 1 on the "Billboard" 200 and was certified platinum by the RIAA in its first five weeks of being released. Most of nu metal's mainstream popularity sharply declined in 2003 and 2004. After a period of mainstream success with bands such as Godsmack, Trapt, Linkin Park and Evanescence, nu metal declined in popularity. Limp Bizkit's 2003 album "Results May Vary", which features alternative rock music and peaked at number 3 on the "Billboard" 200, with sales of at least 325,000 copies in its first week of being released. In three weeks of being released, the album had sold at least 500,000 copies. In 2004, Blabbermouth.net reported that, according to Nielsen SoundScan, "Results May Vary" sold 1,337,356 copies in the United States. However, the album garnered very poor critical reception and consequently performed much weaker than previous Limp Bizkit albums such as "Significant Other" and "Chocolate Starfish and the Hot Dog Flavored Water". Although Korn's album "Take a Look in the Mirror" song "Did My Time" peaked at number 38 on the "Billboard" Hot 100, the album sold less than previous Korn albums "Issues" and "Untouchables". In 2004, post-punk revival bands such as Jet and The Darkness were achieving mainstream success as the popularity of nu metal declined. During the the popularity of emo exceeded the declining popularity of nu metal. Also, during the metalcore, a fusion of extreme metal and hardcore punk, became one of the most popular genres in the new wave of American heavy metal. In the mid-to-late 2000s, many nu metal bands experimented with other genres and sounds. Linkin Park's third studio album "Minutes to Midnight", released in 2007, was noted for its complete departure from the band's nu metal sound. Nu metal bands such as Disturbed and Drowning Pool moved to a hard rock or standard heavy metal sound. Slipknot also departed from their nu metal sound and included elements of groove metal, death metal and thrash metal into their music. Staind and Papa Roach moved to lighter sounds. Staind's 2003 album "14 Shades Of Grey" does not express as much anger as the band's previous albums and shows the band's departure from heavy metal elements and a movement towards a lighter sound. Papa Roach abandoned the nu metal genre with their 2004 album "Getting Away with Murder", moving to a hard rock style. Soulfly moved away from the nu metal style and moved to styles such as death metal and thrash metal. Kittie abandoned the style and started making music with elements of genres such as black metal and death metal. Korn and Mudvayne continued being mainstream during the mid-2000s. Nonetheless, they did not completely abandon the nu metal style. Korn combined their earlier sound with influences from other genres, such as industrial. Korn's songs "Coming Undone" and "Twisted Transistor", which both are on their 2005 album "See You on the Other Side", both reached the "Billboard" Hot 100; Pop music producers The Matrix helped produce the album. Mudvayne's 2005 album "Lost and Found" was seen as gravitating towards a more accessible sound. The album's song "Happy?" peaked at number 89 on the "Billboard" Hot 100 and at number 91 on "Billboard" Pop 100 chart. In 2005, Limp Bizkit released a record called "The Unquestionable Truth (Part 1)" without promoting and advertising the record. The album was not very popular; its sales fell 67% during its second week of release. In 2006, Limp Bizkit went on hiatus. During the 2010s, there was a discussion within media of a possible nu metal revival because of bands fusing nu metal with other genres, the return of nu metal bands, formerly nu metal bands going back to the nu metal genre and nu metal bands forming. Despite the lack of radio play and popularity, some nu metal bands recaptured some of their former popularity as they released albums in a nu metal style. Korn's 2010 studio album "" sold 63,000 copies during its first week of release and peaked at number 2 on the "Billboard" 200. As of December 6, 2011, the album had sold at least 185,000 units in the United States. Korn's vocalist Jonathan Davis said with their new album the band "want to go back to that old-school vibe". He also said "It's gonna be very raw, it's gonna be old school like the first Korn records". In 2011, Limp Bizkit's sixth studio album "Gold Cobra" was released; it sold 27,000 copies during its first week in the United States and peaked at number 16 on the "Billboard" 200. That same year, Staind's self-titled album was released; it shows the band returning to their heavier nu metal style. The album debuted at number 5 on the "Billboard" 200, selling 47,000 copies in its first week of release, making it the band's fifth consecutive top-five album. In October 2011, Evanescence's self-titled album debuted at number 1 on the "Billboard" 200 and other United States charts and sold over 127,000 copies in the first week. In December that year, Korn released their album "The Path of Totality", which sold 55,000 copies in its first week. The album combines nu metal with dubstep. Both the "Phoenix New Times" and the "LA Weekly" cited "The Path of Totality" as a new direction for nu metal. The album won a Revolver Golden God award for "Album of the Year". In 2014, Linkin Park returned to their nu-metal roots with their sixth studio album "The Hunting Party". The album peaked at number three on the "Billboard" 200, with first-week sales of 110,000 copies in the United States. In 2014, Slipknot released its fifth studio album "". With ".5: The Gray Chapter", Slipknot returned to the nu metal genre. ".5: The Gray Chapter" peaked at number one on the "Billboard" 200. Many metalcore and deathcore groups such as My Ticket Home, Stray from the Path, Emmure, Of Mice & Men, Suicide Silence, and Issues, all gained moderate popularity in the 2010s and used elements from nu metal. This fusion has sometimes been referred to as "nu metalcore". Suicide Silence's 2011 album "The Black Crown", which features elements of nu metal and deathcore, peaked at number 28 on the "Billboard" 200. In 2014, Issues' self-titled debut album peaked at number 9 on the same chart. The album features elements of metalcore, nu metal, pop and R&B. Of Mice & Men's 2014 album "Restoring Force", which features elements of nu metal, peaked at number 4 on the "Billboard" 200. Bring Me the Horizon, previously known for a much heavier style of music, released their fifth album "That's the Spirit", which peaked at number 2 on the "Billboard" 200, in 2015. The album draws from multiple genres, including nu metal; however, the band completely abandoned their metalcore style. In the 2010s, nu metal bands like From Ashes to New, Ded, and Islander emerged. The idea of a revival of the genre has been explicitly mentioned by journalists. The likes of Islander, My Ticket Home, Sworn In, DangerKids, Stray from the Path, and Issues have been mentioned as examples of this musical movement. In the mid–late 2010s, genres like emo rap and trap metal emerged. Despite its popularity in the late 1990s and early 2000s, nu metal has often been criticized by many fans of heavy metal music, often being labelled with derogatory terms such as "mallcore" and "whinecore". Gregory Heaney of AllMusic called nu metal "one of metal's more unfortunate pushes into the mainstream". Lucy Jones of "NME" called nu metal "the worst genre of all time". In "Metal: The Definitive Guide : Heavy, NWOBH, Progressive, Thrash, Death ... ", Garry Sharpe-Young described as "a dumbed-down and—thankfully short[-]lived exercise". When Machine Head moved to the nu metal genre with their album "The Burning Red" and their vocalist Robb Flynn spiked his hair in the fashion of many nu metal musicians, the band were accused of "selling out" and many fans criticized their change of appearance and musical style. Machine Head's drummer Dave McClain said, "Pissing people off isn't a bad thing, you know? For people to be narrow-minded is bad ... [i]t doesn't bother us at all, we know we're going to piss people off with this record, but some people hopefully will actually sit down and listen to the whole record". Robb Flynn, Machine Head's vocalist, said Jonathan Davis, the vocalist of Korn, spoke about the criticism of nu metal from heavy metal fans, saying Lamb of God's vocalist Randy Blythe criticized the nu metal genre and spoke about its loss of popularity in 2004, saying, "Nu-metal sucks, so that's why that's dying off. And I think ... ... people are ready for angrier music. I think people are ready for something that's real, not, you know, 'I did it all for the nookie.'" Megadeth frontman Dave Mustaine said he would "rather have his eyelids pulled out" than listen to nu metal. Guitarist Gary Holt of Exodus and Slayer said that he "was so glad about" the decline of . Some musicians who influenced nu metal have tried to distance themselves from the subgenre and its bands. Mike Patton, the vocalist of Faith No More and Mr. Bungle, tried to distance himself from the subgenre and criticized it, even though he is featured on the song "Lookaway" on heavy metal band Sepultura's album "Roots", which also features Jonathan Davis. Patton said of his music's influence on nu metal, "I feel no responsibility for that, it's their mothers' fault, not mine". Helmet member Page Hamilton said, "It's frustrating that people write [us] off because we're affiliated with or credited with or discredited with creating and rap metal ... which we sound nothing like". However, Page Hamilton appeared on the song "All for Nothing" on Linkin Park's album "The Hunting Party". Although Trent Reznor of Nine Inch Nails has said he knows some Korn members and that he thinks they are "cool guys", he also criticized nu metal, saying: In response to reports that Fred Durst, lead singer of Limp Bizkit, is a big fan of Tool, the latter's vocalist Maynard James Keenan said, "If the lunch-lady in high school hits on you, you appreciate the compliment, but you're not really gonna start dating the lunch-lady, are ya?" While Durst has cited Rage Against the Machine as a major influence, Rage Against the Machine's bassist Tim Commerford is open about his hatred of Limp Bizkit, describing them as "one of the dumbest bands in the history of music". At the 2000 MTV Video Music Awards, Limp Bizkit won the Best Rock Video category for their song "Break Stuff", beating Rage Against the Machine's "Sleep Now in the Fire". When Limp Bizkit accepted their award, Commerford went on stage and climbed up a backdrop, rocking back and forth. After the incident, Commerford was arrested and spent a night in jail. Commerford said in 2015, "I do apologize for Limp Bizkit. I really do. I feel really bad that we inspired such bullshit ... They're gone, though. That's the beautiful thing." Some nu metal musicians have rejected the label nu metal and have tried to distance themselves from it. Slipknot prefer to distance themselves from other nu metal groups, describing their own music as "metal metal" and equate their link to nu metal as a coincidence of their time of emergence. Jonathan Davis has rejected the nu metal label, saying "We're not 'rap rock,' we're not 'nu-metal ... We might have invented a new genre of heavy music or rock, but I believe the term 'nu-metal' was made up for all the bands that followed us. Those guys to me are nu-metal. And we're just Korn." In 2014, Davis spoke about the nu metal label, saying: Staind's vocalist Aaron Lewis rejected the nu metal label, saying, "if we get called a 'nu metal' band one more time, I don't even know what I'm going to do!" Chino Moreno, vocalist of Deftones, rejected the nu metal label saying "We told motherfuckers not to lump us in with nu metal because when those bands go down we aren't going to be with them". As Deftones abandoned the nu metal sound of their early work, Moreno tried to distance himself from nu metal bands and began to criticize the bands and their albums, including Korn's 2002 album "Untouchables"; he said, "As Korn go on, it's the same things—bad childhoods and mean moms. It gets too old after a while. How old is Jonathan [Davis]? Thirty? How long has it been since he lived with his parents?" Davis responded saying, "Obviously, Chino hasn't listened to the words on the rest of my albums because they're nothing about my parents or my childhood." Moreno also said, "A big problem for me was opening for Limp Bizkit and Linkin Park, two bands that wouldn't exist if it weren't for me, straight up!". Mike Shinoda of Linkin Park spoke about the nu metal label in an interview with "NME", saying "We never held the flag for nu-metal—it was associated with frat rock. Arrogant, misogynistic, and full of testosterone; we were reacting against that." Wes Borland of Limp Bizkit said that he "never liked or condoned" the term "nu metal" in any way, and said he does not understand "how so many bands that sound nothing alike can be put into" the nu metal genre. Mike Wengren of Disturbed said that he doesn't think Disturbed "were ever a nu-metal band to begin with". Jody MacGregor of FasterLouder called nu metal "music's most hated genre" conversely, he also wrote that nu metal is "not as bad as people think", praising several examples of the genre. Despite the fact that multiple nu metal musicians rejected the nu metal label, Limp Bizkit's vocalist Fred Durst defended it, saying "Nu metal let people open up and it meant something to people. It really did." Coal Chamber's vocalist Dez Fafara also defended nu metal. He said he is proud to be associated with the subgenre and that nu metal bands "broke new musical ground" saying, "I think 'hair metal' was cheesy. [But] I think 'nu metal' was different. I think what's beautiful about 'nu metal' is it's different. And you've got so many different influences." Chester Bennington of Linkin Park said he accepted the nu metal label, saying: Jack Porter of "The Michigan Daily" defended , writing In addition to criticizing nu metal, many heavy metal musicians have rejected nu metal as a legitimate subgenre of heavy metal, saying it is not "true heavy metal". Some nu metal musicians have tried to distance themselves from being heavy metal at all. For example, Korn's Jonathan Davis rejected the "heavy metal" label. When talking with "Vice", Davis spoke about Korn being called a heavy metal band, saying, "I never thought of us to be metal to begin with. Yeah, we're heavy and downtuned, but metal, to me, is like Judas Priest and Iron Maiden. That's metal, man. I always thought of us as a funk band. That funky, groovy shit." Godsmack's vocalist Sully Erna also rejected the "heavy metal" label and said he views Godsmack as a hard rock band. Though he was originally more tolerant of the concept, Linkin Park's vocalist Chester Bennington later expressed some disagreement with his band being labeled a heavy metal group because he felt the term limited the scope of the band's actual style. He elaborated:
https://en.wikipedia.org/wiki?curid=22090
Ncurses ncurses (new curses) is a programming library providing an application programming interface (API) that allows the programmer to write text-based user interfaces in a terminal-independent manner. It is a toolkit for developing "GUI-like" application software that runs under a terminal emulator. It also optimizes screen changes, in order to reduce the latency experienced when using remote shells. There are bindings for ncurses in a variety of programming languages, including Ada, Python, Gambas, Ruby, PHP, JavaScript, and Perl. As the new version, ncurses is a free-software emulation of the System V Release 4.0 (SVr4) curses, which was itself an enhancement over the discontinued 4.4 BSD curses. The XSI Curses standard issued by X/Open is explicitly and closely modeled on System V. The first curses library was developed at the University of California at Berkeley, for a BSD operating system, around 1980 to support Rogue, a text-based adventure game. It originally used the termcap library, which was used in other programs, such as the vi editor. The success of the BSD curses library prompted Bell Labs to release an enhanced curses library in their System V Release 2 Unix systems. This library was more powerful and instead of using termcap, it used terminfo. However, due to AT&T policy regarding source-code distribution, this improved curses library did not have much acceptance in the BSD community. Around 1982, Pavel Curtis started work on a freeware clone of the Bell Labs curses, named pcurses, which was maintained by various people through 1986. The pcurses library was further improved when Zeyd Ben-Halim took over the development effort in late 1991. The new library was released as ncurses in November 1993, with version 1.8.1 as the first major release. Subsequent work, through version 1.8.8 (M1995), was driven by Eric S. Raymond, who added the form and menu libraries written by Juergen Pfeifer. Since 1996, it has been maintained by Thomas E. Dickey. Most ncurses calls can be easily ported to the old curses. System V curses implementations can support BSD curses programs with just a recompilation. However, a few areas are problematic, such as handling terminal resizing, since no counterpart exists in the old curses. Ncurses can use either terminfo (with extensible data) or termcap. Other implementations of curses generally use terminfo; a minority use termcap. Few (mytinfo was an older exception) use both. Ncurses is a part of the GNU Project. It is one of the few GNU files not distributed under the GNU GPL or LGPL; it is distributed under a permissive free software licence, similar to the MIT License. This is due to the agreement made with the Free Software Foundation at the time the developers assigned their copyright. When the agreement was made to pass on the rights to the FSF, there was a clause that stated: The Foundation promises that all distribution of the Package, or of any work "based on the Package", that takes place under the control of the Foundation or its agents or assignees, shall be on terms that explicitly and perpetually permit anyone possessing a copy of the work to which the terms apply, and possessing accurate notice of these terms, to redistribute copies of the work to anyone on the same terms. According to the maintainer Thomas E. Dickey, this precludes relicensing to the GPL in any version, since it would place restrictions on the programs that will be able to link to the libraries. There are hundreds of programs which use ncurses. Some, such as GNU Screen and w3m, use only the termcap interface and perform screen management themselves. Others, such as GNU Midnight Commander and YaST, use the curses programming interface.
https://en.wikipedia.org/wiki?curid=22091
National Basketball Association The National Basketball Association (NBA) is a men's professional basketball league in North America, composed of 30 teams (29 in the United States and 1 in Canada). It is one of the four major professional sports leagues in the United States and Canada, and is widely considered to be the premier men's professional basketball league in the world. The league was founded in New York City on June 6, 1946, as the Basketball Association of America (BAA). It changed its name to the National Basketball Association on August 3, 1949, after merging with the competing National Basketball League (NBL). The NBA's regular season runs from October to April, with each team playing 82 games. The league's playoff tournament extends into June. As of 2015, NBA players are the world's best paid athletes by average annual salary per player. The NBA is an active member of USA Basketball (USAB), which is recognized by FIBA (also known as the International Basketball Federation) as the national governing body for basketball in the United States. The league's several international as well as individual team offices are directed out of its head offices in Midtown Manhattan, while its NBA Entertainment and NBA TV studios are directed out of offices located in Secaucus, New Jersey. The Basketball Association of America was founded in 1946 by owners of the major ice hockey arenas in the Northeastern and Midwestern United States and Canada. On November 1, 1946, in Toronto, Ontario, Canada, the Toronto Huskies hosted the New York Knickerbockers at Maple Leaf Gardens, in a game the NBA now refers to as the first game played in NBA history. The first basket was made by Ossie Schectman of the Knickerbockers. Although there had been earlier attempts at professional basketball leagues, including the American Basketball League and the NBL, the BAA was the first league to attempt to play primarily in large arenas in major cities. During its early years, the quality of play in the BAA was not significantly better than in competing leagues or among leading independent clubs such as the Harlem Globetrotters. For instance, the 1948 ABL finalist Baltimore Bullets moved to the BAA and won that league's 1948 title, and the 1948 NBL champion Minneapolis Lakers won the 1949 BAA title. Prior to the 1948–49 season, however, NBL teams from Fort Wayne, Indianapolis, Minneapolis, and Rochester jumped to the BAA, which established the BAA as the league of choice for collegians looking to turn professional. On August 3, 1949, the remaining NBL teams–Syracuse, Anderson, Tri-Cities, Sheboygan, Denver, and Waterloo–merged into the BAA. In deference to the merger and to avoid possible legal complications, the league name was changed to the present National Basketball Association, even though the merged league retained the BAA's governing body, including Maurice Podoloff as president. To this day, the NBA claims the BAA's history as its own. It now reckons the arrival of the NBL teams as an expansion, not a merger, and does not recognize NBL records and statistics. The new league had seventeen franchises located in a mix of large and small cities, as well as large arenas and smaller gymnasiums and armories. In 1950, the NBA consolidated to eleven franchises, a process that continued until 1953–54, when the league reached its smallest size of eight franchises: the New York Knicks, Boston Celtics, Philadelphia Warriors, Minneapolis Lakers, Rochester Royals, Fort Wayne Pistons, Tri-Cities Blackhawks, and Syracuse Nationals, all of which remain in the league today. The process of contraction saw the league's smaller-city franchises move to larger cities. The Hawks shifted from the Tri-Cities to Milwaukee in 1951, and then to St. Louis in 1955. The Rochester Royals moved from Rochester, New York, to Cincinnati in 1957 and the Pistons moved from Fort Wayne, Indiana, to Detroit in 1957. Japanese-American Wataru Misaka broke the NBA color barrier in the 1947–48 season when he played for the New York Knicks. He remained the only non-white player in league history prior to the first African-American, Harold Hunter, signing with the Washington Capitols in 1950. Hunter was cut from the team during training camp, but several African-American players did play in the league later that year, including Chuck Cooper with the Celtics, Nathaniel "Sweetwater" Clifton with the Knicks, and Earl Lloyd with the Washington Capitols. During this period, the Minneapolis Lakers, led by center George Mikan, won five NBA Championships and established themselves as the league's first dynasty. To encourage shooting and discourage stalling, the league introduced the 24-second shot clock in 1954. If a team does not attempt to score a field goal (or the ball fails to make contact with the rim) within 24 seconds of obtaining the ball, play is stopped and the ball given to its opponent. In 1957, rookie center Bill Russell joined the Boston Celtics, which already featured guard Bob Cousy and coach Red Auerbach, and went on to lead the franchise to eleven NBA titles in thirteen seasons. Center Wilt Chamberlain entered the league with the Warriors in 1959 and became a dominant individual star of the 1960s, setting new single game records in scoring (100) and rebounding (55). Russell's rivalry with Chamberlain became one of the greatest rivalries in the history of American team sports. The 1960s were dominated by the Celtics. Led by Russell, Cousy, and Auerbach, Boston won eight straight championships in the NBA from 1959 to 1966. This championship streak is the longest in NBA history. They did not win the title in 1966–67, but regained it in the 1967–68 season and repeated in 1969. The domination totaled nine of the ten championship banners of the 1960s. Through this period, the NBA continued to evolve with the shift of the Minneapolis Lakers to Los Angeles, the Philadelphia Warriors to San Francisco, the Syracuse Nationals to Philadelphia to become the Philadelphia 76ers, and the St. Louis Hawks moving to Atlanta, as well as the addition of its first expansion franchises. The Chicago Packers (now Washington Wizards) became the ninth NBA team in 1961. From 1966 to 1968, the league expanded from 9 to 14 teams, introducing the Chicago Bulls, Seattle SuperSonics (now Oklahoma City Thunder), San Diego Rockets (who moved to Houston four years later), Milwaukee Bucks, and Phoenix Suns. In 1967, the league faced a new external threat with the formation of the American Basketball Association (ABA). The leagues engaged in a bidding war. The NBA landed the most important college star of the era, Kareem Abdul-Jabbar (then known as Lew Alcindor). However, the NBA's leading scorer, Rick Barry, jumped to the ABA, as did four veteran referees—Norm Drucker, Earl Strom, John Vanak, and Joe Gushue. In 1969, Alan Siegel, who oversaw the design of Jerry Dior's Major League Baseball logo a year prior, created the modern NBA logo inspired by the MLB's. It incorporates the silhouette of Jerry West, based on a photo by Wen Roberts, although NBA officials denied a particular player as being its influence because, according to Siegel, "They want to institutionalize it rather than individualize it. It's become such a ubiquitous, classic symbol and focal point of their identity and their licensing program that they don't necessarily want to identify it with one player." The iconic logo debuted in 1971 (with a small change to the typeface on the NBA wordmark in 2017) and would remain a fixture of the NBA brand. The ABA succeeded in signing a number of major stars in the 1970s, including Julius Erving of the Virginia Squires, in part because it allowed teams to sign college undergraduates. The NBA expanded rapidly during this period, one purpose being to tie up the most viable cities. From 1966 to 1974, the NBA grew from nine franchises to 18. In 1970, the Portland Trail Blazers, Cleveland Cavaliers, and Buffalo Braves (now the Los Angeles Clippers) all made their debuts expanding the league to 17. The New Orleans Jazz (now in Utah) came aboard in 1974 bringing the total to 18. Following the 1976 season, the leagues reached a settlement that provided for the addition of four ABA franchises to the NBA, raising the number of franchises in the league at that time to 22. The franchises added were the San Antonio Spurs, Denver Nuggets, Indiana Pacers, and New York Nets (now the Brooklyn Nets). Some of the biggest stars of this era were Abdul-Jabbar, Barry, Dave Cowens, Erving, Elvin Hayes, Walt Frazier, Moses Malone, Artis Gilmore, George Gervin, Dan Issel, and Pete Maravich. The end of the decade, however, saw declining TV ratings, low attendance and drug-related player issues – both perceived and real – that threatened to derail the league. The league added the ABA's three-point field goal beginning in 1979. That same year, rookies Larry Bird and Magic Johnson joined the Boston Celtics and Los Angeles Lakers respectively, initiating a period of significant growth in fan interest in the NBA. The two had faced each other in the 1979 NCAA Division I Basketball Championship Game, and they later played against each other in three NBA Finals (1984, 1985, and 1987). In the 10 seasons of the 1980s, Johnson led the Lakers to five titles while Bird led the Celtics to three titles. Also in the early 1980s, the NBA added one more expansion franchise, the Dallas Mavericks, bringing the total to 23 teams. Later on, Larry Bird won the first three three-point shooting contests. On February 1, 1984 David Stern became commissioner of the NBA. Stern has been recognized as playing a major role in the growth of the league during his career. Michael Jordan entered the league in 1984 with the Chicago Bulls, spurring more interest in the league. In 1988 and 1989, four cities got their wishes as the Charlotte Hornets, Miami Heat, Orlando Magic, and Minnesota Timberwolves made their NBA debuts, bringing the total to 27 teams. The Detroit Pistons won the back-to-back NBA Championships in 1989 and 1990, led by coach Chuck Daly and guard Isiah Thomas. Jordan and Scottie Pippen led the Bulls to two three-peats in eight years during the 1991–1998 seasons. Hakeem Olajuwon won back-to-back titles with the Houston Rockets in 1994 and 1995. The 1992 Olympic basketball Dream Team, the first to use current NBA stars, featured Michael Jordan as the anchor, along with Bird, Johnson, David Robinson, Patrick Ewing, Scottie Pippen, Clyde Drexler, Karl Malone, John Stockton, Chris Mullin, Charles Barkley, and star NCAA amateur Christian Laettner. The team was elected to the Naismith Memorial Basketball Hall of Fame, and 11 players and three coaches have been elected to the Hall of Fame as individuals. In 1995, the NBA expanded to Canada with the addition of the Vancouver Grizzlies and the Toronto Raptors. In 1996, the NBA created a women's league, the Women's National Basketball Association (WNBA). In 1998, the NBA owners began a lockout that suspended all league business until a new labor agreement could be reached. By the time the lockout was settled later that year, half the season was lost. In 1999, the San Antonio Spurs became the first former ABA team to win the NBA championship. It was the team's first championship win. Since the breakup of the Chicago Bulls championship roster in the summer of 1998, the Western Conference has dominated. The Los Angeles Lakers of coach Phil Jackson and the San Antonio Spurs of Gregg Popovich combined to make 13 Finals in 16 seasons, with 10 titles. Tim Duncan and David Robinson won the 1999 championship with the Spurs, and Shaquille O'Neal and Kobe Bryant started the 2000s with three consecutive championships for the Lakers. The Spurs reclaimed the title in 2003 against the Nets. In 2004, the Lakers returned to the Finals, only to lose in five games to the Detroit Pistons. The league's image was marred by a violent incident between players and fans in a November 2004 game between the Indiana Pacers and Detroit Pistons. In response, players were suspended for a total of 146 games with $11 million total lost in salary, and the league tightened security and limited the sale of alcohol. After the Spurs took home the Larry O'Brien Championship Trophy in 2005, the 2006 Finals featured two franchises making their inaugural Finals appearances. The Miami Heat, led by their star shooting guard, Dwyane Wade, and Shaquille O'Neal, who had been traded from the Lakers during summer 2004, won the series over the Dallas Mavericks in six after losing the first two games. The Lakers/Spurs dominance continued in 2007 with a four-game sweep by the Spurs over the Cleveland Cavaliers, who were led by LeBron James. The 2008 Finals saw a rematch of the league's highest profile rivalry, the Boston Celtics and Los Angeles Lakers, with the Celtics winning, for their 17th championship, thanks to their new big three of Paul Pierce, Ray Allen, and Kevin Garnett. In 2009, Kobe Bryant and the Lakers returned to the Finals, this time defeating the Dwight Howard-led Orlando Magic. Bryant won his first Bill Russell NBA Finals Most Valuable Player Award in his 13th season after leading the Lakers to their first NBA championship since the departure of Shaquille O'Neal. The 2010 NBA All-Star Game was held at Cowboys Stadium in front of the largest crowd ever, 108,713. At the end of that season, the Celtics and the Lakers renewed their rivalry from 2008 when they met again in the NBA Finals for a record 12th time. The Lakers won the title by winning Game 7, 83–79. Before the start of the 2010–11 season the NBA had an exciting summer with one of the most anticipated free agent classes of all time. Two free agents signed, and one resigned, with the Miami Heat, leading to a season that was heavily centered on their eventual success or failure at taking home the championship. The Heat, led by LeBron James, Dwyane Wade, and Chris Bosh, did in fact make the Finals against the Dallas Mavericks, in a rematch for the franchises of the 2006 Finals. The Mavericks, led by Dirk Nowitzki (the eventual NBA Finals MVP), took the series in six games. This was the Mavericks' first title. Veterans Shawn Marion, Jason Kidd, Jason Terry, and Peja Stojaković celebrated their first NBA championship. On July 1, 2011, at 12:01 am, the NBA announced another lockout. After the first few weeks of the season were canceled, the players and owners ratified a new collective bargaining agreement on December 8, 2011, setting up a shortened 66-game season. Following the shortened season, the Miami Heat made a return to the Finals with the trio of Dwyane Wade, LeBron James, and Chris Bosh against Oklahoma City Thunder's Kevin Durant, Russell Westbrook, and James Harden. The Heat went on to defeat the Thunder in five games, capturing their second NBA title in six years. Their success would continue into the following season, which concluded with their victory over the San Antonio Spurs in the 2013 NBA Finals. The two teams would meet for a rematch in the following year's Finals, where the Spurs defeated the Heat in five games. Off the court, commissioner David Stern retired on February 1, 2014, 30 years to the day after taking office. He was succeeded by his deputy, Adam Silver. After four seasons with the Miami Heat, LeBron James returned to Ohio upon the 2014–15 season to once again play for the Cleveland Cavaliers. He became the head of the team and led them to their second Finals appearance with the help of Kyrie Irving and Kevin Love. The Golden State Warriors defeated the Cavaliers in six games, led by Stephen Curry and Klay Thompson, also known as the Splash Brothers. The following year, the Warriors obtained the best season record in NBA history by win percentage, further establishing their dominance within the league. The 2016 NBA Finals became a rematch between the two teams. It showed the Cavaliers win their first NBA Championship as well as become the first team to return from a 3–1 series deficit in the Finals. In the 2016–17 season, the Warriors benefited from the recruitment of Kevin Durant, who had joined the team as a free agent in 2016. He became the Most Valuable Player of the 2017 NBA Finals and led the team to another Championship victory in five games over the Cavaliers, then again in 2018 when the Warriors won in a four-game sweep against the Cavaliers. During free agency in 2018, LeBron James made the decision to sign with the Los Angeles Lakers. The Cavaliers were unable to perform in the 2018–19 season, finishing second to last in their conference after four consecutive Finals appearances. However, the Warriors once again returned to the Finals for the fifth straight year where they faced the Toronto Raptors, a team that had never previously made it to the Finals. Led by Kawhi Leonard, the Raptors won their first NBA championship, defeating the Warriors in a total of six games. On March 11, 2020, due to the COVID-19 pandemic, the league held a conference call among the NBA owners regarding the formulation of a plan should the epidemic become more serious. The same day, the Oklahoma City Thunder postponed their game and the Warriors cancelled all home games in the foreseeable future. Eventually, the NBA suspended the remainder of its season "until further notice" after Utah Jazz center Rudy Gobert tested positive for the coronavirus. After the suspension of the season, the Dallas Mavericks' owner Mark Cuban started off the initiative of helping pay arena workers by setting up a payment procedure for the hourly workers, which was also joined by players and other NBA personnel. The NBA and the National Basketball Players Association have cut the players' salaries by 25 percent due to the economic fallout from the pandemic. Following pioneers like Vlade Divac (Serbia) and Dražen Petrović (Croatia) who joined the NBA in the late 1980s, an increasing number of international players have moved directly from playing elsewhere in the world to starring in the NBA. Since 2006, the NBA has faced EuroLeague teams in exhibition matches in the NBA Europe Live Tour, and since 2009, in the EuroLeague American Tour. The 2013–14 season opened with a record 92 international players on the opening night rosters, representing 39 countries and comprising over 20% of the league. The beginning of the 2017–18 season saw a record 108 international players representing 42 different countries marking 4 consecutive years of at least 100 international players and each team having at least one international player. In 2018, the Phoenix Suns hired Serbian coach Igor Kokoškov as their new head coach, replacing Canadian interim coach Jay Triano, making Kokoškov the first European coach to become a head coach for a team in the NBA. In 2001, an affiliated minor league, the National Basketball Development League, now called the NBA G League, was created. Two years after the Hornets' move to New Orleans, the NBA returned to North Carolina, as the Charlotte Bobcats were formed as an expansion team in 2004. The Hornets temporarily moved to Oklahoma City in 2005 for two seasons because of damage caused by Hurricane Katrina. The team returned to New Orleans in 2007. A new official game ball was introduced on June 28, 2006, for the 2006–07 season, marking the first change to the ball in over 35 years and only the second ball in 60 seasons. Manufactured by Spalding, the new ball featured a new design and new synthetic material that Spalding claimed offered a better grip, feel, and consistency than the original ball. However, many players were vocal in their disdain for the new ball, saying that it was too sticky when dry, and too slippery when wet. Commissioner Stern announced on December 11, 2006, that beginning January 1, 2007, the NBA would return to the traditional leather basketball in use prior to the 2006–07 season. The change was influenced by frequent player complaints and confirmed hand injuries (cuts) caused by the microfiber ball. The Players' Association had filed a suit on behalf of the players against the NBA over the new ball. As of the 2017–18 season, the NBA team jerseys are manufactured by Nike, replacing the previous supplier, Adidas. All teams will wear jerseys with the Nike logo except the Charlotte Hornets, whose jerseys will instead have the Jumpman logo associated with longtime Nike endorser Michael Jordan, who owns the Hornets. The Federal Bureau of Investigation (FBI) began an investigation on July 19, 2007, over allegations that veteran NBA referee Tim Donaghy bet on basketball games he officiated over the past two seasons and that he made calls affecting the point spread in those games. On August 15, 2007, Donaghy pleaded guilty to two federal charges related to the investigation. Donaghy claimed in 2008 that certain referees were friendly with players and "company men" for the NBA, and he alleged that referees influenced the outcome of certain playoff and finals games in 2002 and 2005. NBA commissioner David Stern denied the allegations and said Donaghy was a convicted felon and a "singing, cooperating witness". Donaghy served 15 months in prison and was released in November 2009. According to an independent study by Ronald Beech of Game 6 of the 2002 Western Conference Finals between the Los Angeles Lakers and Sacramento Kings, although the refs increased the Lakers' chances of winning through foul calls during the game, there was no collusion to fix the game. On alleged "star treatment" during Game 6 by the referees toward certain players, Beech claimed, "there does seem to be issues with different standards and allowances for different players." The NBA Board of Governors approved the request of the Seattle SuperSonics to move to Oklahoma City on April 18, 2008. The team, however, could not move until it had settled a lawsuit filed by the city of Seattle, which was intended to keep the SuperSonics in Seattle for the remaining two seasons of the team's lease at KeyArena. Following a court case, the city of Seattle settled with the ownership group of the SuperSonics on July 2, 2008, allowing the team to move to Oklahoma City immediately in exchange for terminating the final two seasons of the team's lease at KeyArena. The Oklahoma City Thunder began playing in the 2008–09 season. The first outdoor game in the modern era of the league was played at the Indian Wells Tennis Garden on October 11, 2008, between the Phoenix Suns and the Denver Nuggets. A referee lockout began on September 1, 2009, when the contract between the NBA and its referees expired. The first preseason games were played on October 1, 2009, and replacement referees from the WNBA and NBA Development League were used, the first time replacement referees had been used since the beginning of the 1995–96 season. The NBA and the regular referees reached a deal on October 23, 2009. The first official NBA league games on European ground took place in 2011. In two matchups, the New Jersey Nets faced the Toronto Raptors at the O2 Arena in London in front of over 20,000 fans. The 2011–12 NBA season, scheduled to begin November 1, 2011, with a matchup between the defending champion Dallas Mavericks and the Chicago Bulls, was postponed due to a labor dispute. The lockout officially ended on December 8, 2011, when players and owners ratified a new collective bargaining agreement, and the season began on Christmas Day. After the 2012–13 season, the New Orleans Hornets were renamed the Pelicans. During the 2013–14 season, Stern retired as commissioner after 30 years, and deputy commissioner Adam Silver ascended to the position of commissioner. During that season's playoffs, the Bobcats officially reclaimed the Hornets name, and by agreement with the league and the Pelicans, also received sole ownership of all history, records, and statistics from the Pelicans' time in Charlotte. As a result, the Hornets are now officially considered to have been founded in 1988, suspended operations in 2002, and resumed in 2004 as the Bobcats, while the Pelicans are officially treated as a 2002 expansion team. (This is somewhat similar to the relationship between the Cleveland Browns and Baltimore Ravens in the NFL.) Donald Sterling, who was then-owner of the Los Angeles Clippers, received a lifetime ban from the NBA on April 29, 2014 after racist remarks he made became public. Sterling was also fined US$2.5 million, the maximum allowed under the NBA Constitution. Becky Hammon was hired by the San Antonio Spurs on August 5, 2014, as an assistant coach, becoming the second female coach in NBA history but the first full-time coach. This also makes her the first full-time female coach in any of the four major professional sports in North America. The NBA announced on April 15, 2016, that it would allow all 30 of its teams to sell corporate sponsor advertisement patches on official game uniforms, beginning with the 2017–18 season. The sponsorship advertisement patches would appear on the left front of jerseys, opposite Nike's logo, marking the first time a manufacturer's logo would appear on NBA jerseys, and would measure approximately 2.5 by 2.5 inches. The NBA would become the first major North American professional sports league to allow corporate sponsorship logos on official team uniforms, and the last to have a uniform manufacturer logo appear on its team uniforms. The first team to announce a jersey sponsorship was the Philadelphia 76ers, who agreed to a deal with StubHub. On July 6, 2017, the NBA unveiled an updated rendition of its logo; it was largely identical to the previous design, except with revised typography and a "richer" color scheme. The league began to phase in the updated logo across its properties during the 2017 NBA Summer League, but it will not immediately be used on equipment or uniforms due to lead time. The NBA also officially released new Nike uniforms for all 30 teams beginning with the 2017–18 season. The league eliminated "home" and "away" uniform designations. Instead, each team would have four or six uniforms: the "Association" edition, which is the team's white uniform, the "Icon" edition, which is the team's color uniform, and the "Statement" and "City" uniforms, which most teams use as an alternate uniform. In 2018, the NBA also released the "Earned" uniform. The NBA originated in 1946 with 11 teams, and through a sequence of team expansions, reductions, and relocations, currently consists of 30 teams. The United States is home to 29 teams; another is in Canada. The current league organization divides 30 teams into two conferences of three divisions with five teams each. The current divisional alignment was introduced in the 2004–05 season. Reflecting the population distribution of the United States and Canada as a whole, most teams are in the eastern half of the country: 13 teams are in the Eastern Time Zone, nine in the Central, three in the Mountain, and five in the Pacific. Following the summer break, teams begin training camps in late September. Training camps allow the coaching staff to evaluate players (especially rookies), scout the team's strengths and weaknesses, prepare the players for the rigorous regular season, and determine the 12-man active roster (and a 3-man inactive list) with which they will begin the regular season. Teams have the ability to assign players with less than two years of experience to the NBA G League. After training camp, a series of preseason exhibition games are held. Preseason matches are sometimes held in non-NBA cities, both in the United States and overseas. The NBA regular season begins in the last week of October. During the regular season, each team plays 82 games, 41 each home and away. A team faces opponents in its own division four times a year (16 games). Each team plays six of the teams from the other two divisions in its conference four times (24 games), and the remaining four teams three times (12 games). Finally, each team plays all the teams in the other conference twice apiece (30 games). This asymmetrical structure means the strength of schedule will vary between teams (but not as significantly as the NFL or MLB). Over five seasons, each team will have played 80 games against their division (20 games against each opponent, 10 at home, 10 on the road), 180 games against the rest of their conference (18 games against each opponent, 9 at home, 9 on the road), and 150 games against the other conference (10 games against each team, 5 at home, 5 on the road). The NBA is one of only two of the four major professional sports leagues in the United States and Canada in which teams play every other team during the regular season (the other being the National Hockey League). Each team hosts and visits every other team at least once every season. From 2005 to 2008, the NBA had the distinction of being the only one of the four major leagues in which all teams play every other team. The NBA is also the only league that regularly schedules games on Christmas Day. The league has been playing games regularly on the holiday since 1947, though the first Christmas Day games were not televised until . Games played on this day have featured some of the best teams and players. Christmas is also notable for NBA on television, as the holiday is when the first NBA games air on network television each season. Games played on this day have been some of the highest-rated games during a particular season. In February, the regular season pauses to celebrate the annual NBA All-Star Game. Fans vote throughout the United States, Canada, and on the Internet, and the top vote-getters in each conference are named captains. Fan votes determine the rest of the allstar starters. Coaches vote to choose the remaining 14 All-Stars. Then, the top vote-getters in each conference draft their own team from a player pool of allstars. The top vote-getter in the league earns first pick and so forth. The player with the best performance during the game is rewarded with a Game MVP award. Other attractions of the All-Star break include the Rising Stars Challenge (originally Rookie Challenge), where the top rookies and second-year players in the NBA play in a 5-on-5 basketball game, with the current format pitting U.S. players against those from the rest of the world; the Skills Challenge, where players compete to finish an obstacle course consisting of shooting, passing, and dribbling in the fastest time; the Three Point Contest, where players compete to score the highest number of three-point field goals in a given time; and the NBA Slam Dunk Contest, where players compete to dunk the ball in the most entertaining way according to the judges. These other attractions have varying names which include the names of the various sponsors who have paid for naming rights. Shortly after the All-Star break is the trade deadline, which is set to fall on the 16th Thursday of the season (usually in February) at 3pm Eastern Time. After this date, teams are not allowed to exchange players with each other for the remainder of the season, although they may still sign and release players. Major trades are often completed right before the trading deadline, making that day a hectic time for general managers. Around the middle of April, the regular season ends. It is during this time that voting begins for individual awards, as well as the selection of the honorary, league-wide, post-season teams. The Sixth Man of the Year Award is given to the best player coming off the bench (must have more games coming off the bench than actual games started). The Rookie of the Year Award is awarded to the most outstanding first-year player. The Most Improved Player Award is awarded to the player who is deemed to have shown the most improvement from the previous season. The Defensive Player of the Year Award is awarded to the league's best defender. The Coach of the Year Award is awarded to the coach that has made the most positive difference to a team. The Most Valuable Player Award is given to the player deemed the most valuable for (his team) that season. Additionally, "Sporting News" awards an unofficial (but widely recognized) Executive of the Year Award to the general manager who is adjudged to have performed the best job for the benefit of his franchise. The post-season teams are the All-NBA Team, the All-Defensive Team, and the All-Rookie Team; each consists of five players. There are three All-NBA teams, consisting of the top players at each position, with first-team status being the most desirable. There are two All-Defensive teams, consisting of the top defenders at each position. There are also two All-Rookie teams, consisting of the top first-year players regardless of position. The NBA playoffs begin in April after the conclusion of the regular season with the top eight teams in each conference, regardless of divisional alignment, competing for the league's championship title, the Larry O'Brien Championship Trophy. Seeds are awarded in strict order of regular season record (with a tiebreaker system used as needed). Having a higher seed offers several advantages. Since the first seed begins the playoffs playing against the eighth seed, the second seed plays the seventh seed, the third seed plays the sixth seed, and the fourth seed plays the fifth seed, having a higher seed means a team faces a weaker team in the first round. The team in each series with the better record has home court advantage, including the First Round. Before the league changed its playoff determination format for the 2006–07 season, this meant that, for example, if the team that received the sixth seed had a better record than the team with the third seed (by virtue of a divisional championship), the sixth seed would have home court advantage, even though the other team had a higher seed. Therefore, the team with the best regular season record in the league is guaranteed home court advantage in every series it plays. For example, in 2006, the Denver Nuggets won 44 games and captured the Northwest Division and the third seed. Their opponent was the sixth-seeded Los Angeles Clippers, who won 47 games and finished second in the Pacific Division. Although Denver won its much weaker division, the Clippers had home-court advantage and won the series in 5. The playoffs follow a tournament format. Each team plays an opponent in a best-of-seven series, with the first team to win four games advancing into the next round, while the other team is eliminated from the playoffs. In the next round, the successful team plays against another advancing team of the same conference. All but one team in each conference are eliminated from the playoffs. Since the NBA does not re-seed teams, the playoff bracket in each conference uses a traditional design, with the winner of the series matching the first- and eighth-seeded teams playing the winner of the series matching the fourth- and fifth-seeded teams, and the winner of the series matching the second- and seventh-seeded teams playing the winner of the series matching the third- and sixth-seeded teams. In every round, the best-of-7 series follows a 2–2–1–1–1 home-court pattern, meaning that one team will have home court in games 1, 2, 5, and 7, while the other plays at home in games 3, 4, and 6. From 1985 to 2013, the NBA Finals followed a 2–3–2 pattern, meaning that one team had home court in games 1, 2, 6, and 7, while the other played at home in games 3, 4, and 5. The final playoff round, a best-of-seven series between the victors of both conferences, is known as the NBA Finals, and is held annually in June. The winner of the NBA Finals receives the Larry O'Brien Championship Trophy. Each player and major contributor—including coaches and the general manager—on the winning team receive a championship ring. In addition, the league awards the Bill Russell NBA Finals Most Valuable Player Award to the best performing player of the series. The league began using its current format, with the top eight teams in each conference advancing regardless of divisional alignment, in the 2015–16 season. Previously, the top three seeds went to the division winners. The Boston Celtics have won the most championships with 17 NBA Finals wins. The second most successful franchise is the Los Angeles Lakers, who have 16 overall championships (11 in Los Angeles, 5 in Minneapolis). Following the Lakers, are the Chicago Bulls and Golden State Warriors (2 in Philadelphia, 4 in Oakland) with six championships each. The San Antonio Spurs have five championships, all since 1999. Current teams that have no NBA Finals appearances: As one of the major sports leagues in North America, the NBA has a long history of partnerships with television networks in the United States. The NBA signed a contract with DuMont Television Network in its eighth season, the 1953–54 season, marking the first year the NBA had a national television broadcaster. Similar to the National Football League, the lack of television stations led to NBC taking over the rights from the 1954-55 season until April 7, 1962–NBC's first tenure with the NBA. Currently in the U.S., the NBA has a contract with ESPN and TNT through the 2024–25 season. Games that are not broadcast nationally are usually aired over regional sports networks specific to the area where the teams are located. The National Basketball Association has sporadically participated in international club competitions. From 1987 to 1999 the NBA champions played against the continental champions of the Fédération Internationale de Basketball (FIBA) in the McDonald's Championship. This tournament was won by the NBA invitee every year it was held. In 2012, a ticket cost from $10 to $3,000 apiece, depending on the location of the seat and the success of the teams that were playing. In 2020, ticket prices for the NBA All Star Game became more expensive than ever before, averaging around $2,600, and even more on the secondary market. According to Nielsen's survey, in 2013 the NBA had the youngest audience, with 45 percent of its viewers under 35, but the least likely, along with Major League Baseball, to be watched by women, who make up only 30% of the viewership. It also has the highest share of black viewers with 45 percent of its viewers being black and only about 40 percent of viewers being white, making it the only top North American sport that does not have a white majority audience. As of 2017, the NBA's popularity further declined among White Americans, who during the 2016–17 season, made up only 34% of the viewership. At the same time, the black viewership increased to 47 percent, while Hispanic (of any race) stood at 11% and Asian viewership stood at 8%. According to the same poll, the NBA was favored more strongly by Democrats than Republicans. The NBA has been involved in a number of controversies over the years and has received a significant amount of criticism. Following pioneers like Vlade Divac (Serbia) and Dražen Petrović (Croatia) who joined the NBA in the late 1980s, an increasing number of international players have moved directly from playing elsewhere in the world to starring in the NBA. Below is a short list of foreign players who have won NBA awards or have been otherwise recognized for their contributions to basketball, either currently or formerly active in the league: On some occasions, young players, most but not all from the English-speaking world, have attended U.S. colleges before playing in the NBA. Notable examples are: Since 2006, the NBA has faced EuroLeague teams in exhibition matches in the NBA Europe Live Tour, and since 2009 in the EuroLeague American Tour. The 2013–14 season opened with a record 92 international players on the opening night rosters, representing 39 countries and comprising over 20% of the league. The NBA defines "international" players as those born outside the 50 United States and Washington, D.C. This means that: The league has a global social responsibility program, NBA Cares, that is responsible for the league's stated mission of addressing important social issues worldwide.
https://en.wikipedia.org/wiki?curid=22093
Nutation Nutation (from Latin "nūtātiō", "nodding, swaying") is a rocking, swaying, or nodding motion in the axis of rotation of a largely axially symmetric object, such as a gyroscope, planet, or bullet in flight, or as an intended behaviour of a mechanism. In an appropriate reference frame it can be defined as a change in the second Euler angle. If it is not caused by forces external to the body, it is called "free nutation" or "Euler nutation". A "pure nutation" is a movement of a rotational axis such that the first Euler angle is constant. In spacecraft dynamics, precession (a change in the first Euler angle) is sometimes referred to as nutation. If a top is set at a tilt on a horizontal surface and spun rapidly, its rotational axis starts precessing about the vertical. After a short interval, the top settles into a motion in which each point on its rotation axis follows a circular path. The vertical force of gravity produces a horizontal torque about the point of contact with the surface; the top rotates in the direction of this torque with an angular velocity such that at any moment where is the instantaneous angular momentum of the top. Initially, however, there is no precession, and the top falls straight downward. This gives rise to an imbalance in torques that starts the precession. In falling, the top overshoots the level at which it would precess steadily and then oscillates about this level. This oscillation is called "nutation". If the motion is damped, the oscillations will die down until the motion is a steady precession. The physics of nutation in tops and gyroscopes can be explored using the model of a "heavy symmetrical top" with its tip fixed. (A symmetrical top is one with rotational symmetry, or more generally one in which two of the three principal moments of inertia are equal.) Initially, the effect of friction is ignored. The motion of the top can be described by three Euler angles: the tilt angle between the symmetry axis of the top and the vertical; the azimuth of the top about the vertical; and the rotation angle of the top about its own axis. Thus, precession is the change in and nutation is the change in . If the top has mass and its center of mass is at a distance from the pivot point, its gravitational potential relative to the plane of the support is In a coordinate system where the axis is the axis of symmetry, the top has angular velocities and moments of inertia about the , and axes. Since we are taking a symmetric top, we have =. The kinetic energy is In terms of the Euler angles, this is If the Euler–Lagrange equations are solved for this system, it is found that the motion depends on two constants and (each related to a constant of motion). The rate of precession is related to the tilt by The tilt is determined by a differential equation for of the form where is a cubic polynomial that depends on parameters and as well as constants that are related to the energy and the gravitational torque. The roots of are cosines of the angles at which the rate of change of is zero. One of these is not related to a physical angle; the other two determine the upper and lower bounds on the tilt angle, between which the gyroscope oscillates. The nutation of a planet occurs because the gravitational effects of other bodies cause the speed of its axial precession to vary over time, so that the speed is not constant. English astronomer James Bradley discovered the nutation of Earth's axis in 1728. Nutation subtly changes the axial tilt of Earth with respect to the ecliptic plane, shifting the major circles of latitude that are defined by the Earth's tilt (the tropical circles and the polar circles). In the case of Earth, the principal sources of tidal force are the Sun and Moon, which continuously change location relative to each other and thus cause nutation in Earth's axis. The largest component of Earth's nutation has a period of 18.6 years, the same as that of the precession of the Moon's orbital nodes. However, there are other significant periodic terms that must be accounted for depending upon the desired accuracy of the result. A mathematical description (set of equations) that represents nutation is called a "theory of nutation". In the theory, parameters are adjusted in a more or less "ad hoc" method to obtain the best fit to data. Simple rigid body dynamics do not give the best theory; one has to account for deformations of the Earth, including mantle inelasticity and changes in the core–mantle boundary. The principal term of nutation is due to the regression of the Moon's nodal line and has the same period of 6798 days (18.61 years). It reaches plus or minus 17″ in longitude and 9.2″ in obliquity. All other terms are much smaller; the next-largest, with a period of 183 days (0.5 year), has amplitudes 1.3″ and 0.6″ respectively. The periods of all terms larger than 0.0001″ (about as accurately as one can measure) lie between 5.5 and 6798 days; for some reason (as with ocean tidal periods) they seem to avoid the range from 34.8 to 91 days, so it is customary to split the nutation into long-period and short-period terms. The long-period terms are calculated and mentioned in the almanacs, while the additional correction due to the short-period terms is usually taken from a table. They can also be calculated from the Julian day according to IAU 2000B methodology. In the 1961 disaster film "The Day the Earth Caught Fire", the near-simultaneous detonation of two super-hydrogen bombs near the poles causes a change in Earth's nutation, as well as an 11° shift in the axial tilt and a change in Earth's orbit around the Sun.
https://en.wikipedia.org/wiki?curid=22094
Narcissus (mythology) In Greek mythology, Narcissus (; Ancient Greek: Νάρκισσος "Nárkissos") was a hunter from Thespiae in Boeotia who was known for his beauty. According to Tzetzes, he was a Laconian hunter who loved everything beautiful. Narcissus was proud, in that he disdained those who loved him, causing some to take their own lives to prove their devotion to his striking beauty. Narcissus is the origin of the term "narcissism", a fixation with oneself and one's physical appearance or public perception. The name is of uncertain etymology. According to R. S. P. Beekes, "[t]he suffix [-ισσος] clearly points to a Pre-Greek word." The word narcissus has come to be used for the daffodil, but there is no clarity on whether the flower is named for the myth, or the myth for the flower, or if there is any true connection at all. Pliny the Elder wrote that the plant was named for its fragrance ( , "I grow numb") not the youth. Narcissus was the son of the river god Cephissus and nymph Liriope, while Nonnus instead has him as the son of the lunar goddess Selene and her mortal lover Endymion. Several versions of the myth have survived from ancient sources. The classic version is by Ovid, found in book 3 of his "Metamorphoses" (completed 8 AD); this is the story of Echo and Narcissus. One day Narcissus was walking in the woods when Echo, an Oread (mountain nymph) saw him, fell deeply in love, and followed him. Narcissus sensed he was being followed and shouted "Who's there?". Echo repeated "Who's there?" She eventually revealed her identity and attempted to embrace him. He stepped away and told her to leave him alone. She was heartbroken and spent the rest of her life in lonely glens until nothing but an echo sound remained of her. Nemesis (as an aspect of Aphrodite), the goddess of revenge, noticed this behaviour after learning the story and decided to punish Narcissus. Once, during the summer, he was getting thirsty after hunting, and the goddess lured him to a pool where he leaned upon the water and saw himself in the bloom of youth. Narcissus did not realize it was merely his own reflection and fell deeply in love with it, as if it were somebody else. Unable to leave the allure of his image, he eventually realized that his love could not be reciprocated and he melted away from the fire of passion burning inside him, eventually turning into a gold and white flower. An earlier version ascribed to the poet Parthenius of Nicaea, composed around 50 BC, was discovered in 2004 by Dr Benjamin Henry among the Oxyrhynchus papyri at Oxford. Unlike Ovid's version, it ended with Narcissus who lost his will to live and committed suicide. A version by Conon, a contemporary of Ovid, also ends in suicide ("Narrations," 24). In it, a young man named Ameinias fell in love with Narcissus, who had already spurned his male suitors. Narcissus also spurned him and gave him a sword. Ameinias committed suicide at Narcissus's doorstep. He had prayed to the gods to give Narcissus a lesson for all the pain he provoked. Narcissus walked by a pool of water and decided to drink some. He saw his reflection, became entranced by it, and killed himself because he could not have his object of desire. A century later the travel writer Pausanias recorded a novel variant of the story, in which Narcissus falls in love with his twin sister rather than himself ("Guide to Greece", 9.31.7). In all versions, his body disappears and all that is left is a narcissus flower. In 1898 Havelock Ellis, an English sexologist, used the term "narcissus-like" in reference to excessive masturbation, whereby the person becomes his or her own sex object. In 1899, Paul Näcke was the first person to use the term "narcissism" in a study of sexual perversions. Otto Rank, in 1911, published the first psychoanalytical paper specifically concerned with narcissism, linking it to vanity and self-admiration. Sigmund Freud published a paper exclusively devoted to narcissism in 1914, called "On Narcissism: An Introduction". One of the personality disorders is called narcissistic personality disorder. The myth of Narcissus has inspired artists for at least two thousand years, even before the Roman poet Ovid featured a version in book III of his "Metamorphoses". This was followed in more recent centuries by other poets (e.g. Keats and Alfred Edward Housman) and painters (Caravaggio, Poussin, Turner, Dalí (see "Metamorphosis of Narcissus"), and Waterhouse). In Stendhal's novel "Le Rouge et le Noir" (1830), there is a classic narcissist in the character of Mathilde. Says Prince Korasoff to Julien Sorel, the protagonist, with respect to his beloved girl: She looks at herself instead of looking at you, and so doesn't know you. During the two or three little outbursts of passion she has allowed herself in your favor, she has, by a great effort of imagination, seen in you the hero of her dreams, and not yourself as you really are. (Page 401, 1953 Penguin Edition, trans. Margaret R.B. Shaw). The myth had a decided influence on English Victorian homoerotic culture, via André Gide's study of the myth, "Le Traité du Narcisse" ('The Treatise of the Narcissus', 1891), and the only novel by Oscar Wilde, "The Picture of Dorian Gray". Paulo Coelho's "The Alchemist" also starts with a story about Narcissus, found (we are told) by the alchemist in a book brought by someone in the caravan. The alchemist's (and Coelho's) source was very probably Hesketh Pearson's "The Life of Oscar Wilde" (1946) in which this story is recorded (Penguin edition, p. 217) as one of Wilde's inspired inventions. This version of the Narcissus story is based on Wilde's "The Disciple" from his "Poems in Prose (Wilde) ". Author and poet Rainer Maria Rilke visits the character and symbolism of Narcissus in several of his poems. Seamus Heaney references Narcissus in his poem "Personal Helicon" from his first collection "Death of a Naturalist":"To stare, big-eyed Narcissus, into some spring Is beneath all adult dignity." In Rick Riordan's "Heroes of Olympus series," Narcissus appears as a minor antagonist in the third book "The Mark of Athena". In the fantasy series "Harry Potter", Narcissa Malfoy, a minor antagonist, is named for Narcissus. William Faulkner's character "Narcissa" in "Sanctuary", sister of Horace Benbow, was also named after Narcissus. Throughout the novel, she allows the arrogant, pompous pressures of high-class society to overrule the unconditional love that she should have for her brother. Hermann Hesse's character "Narcissus" in "Narcissus and Goldmund" shares several of mythical Narcissus' traits, although his narcissism is based on his intellect rather than his physical beauty. A. E. Housman refers to the 'Greek Lad', Narcissus, in his poem "Look not in my Eyes" from "A Shropshire Lad" set to music by several English composers including George Butterworth. At the end of the poem stands a jonquil, a variety of daffodil, Narcissus Jonquilla, which like Narcissus looks sadly down into the water. Herman Melville references the myth of Narcissus in his novel Moby-Dick, in which Ishmael explains the myth as "the key to it all," referring to the greater theme of finding the essence of Truth through the physical world. On Sophia de Mello Breyner Andresen's A Fada Oriana, the eponymous protagonist is punished with mortality for abandoning her duties in order to stare at herself in the surface of a river. Joseph Conrad's novel The Nigger of the 'Narcissus' features a merchant ship named "Narcissus". An incident involving the ship, and the difficult decisions made by the crew, explore themes involving self-interest vs. altruism and humanitarianism. Naomi Iizuka's play "Polaroid Stories", a contemporary rewrite of the Orpheus and Eurydice myth, features Narcissus as a character. In the play he is portrayed as a self obsessed, and drug addicted young man who was raised on the streets. He is alluded to being a member of the LGBT+ community and mentions his sexual endeavors with older men, some ending with the death of these men due to drug overdoses. He is accompanied by the character Echo who he continuously spurns. In the TV series Boardwalk Empire, a Dr. Narcisse (Valentin Narcisse) is introduced as a condescending intellectual. Scottish-Canadian animator Norman McLaren finished his career with a short film named "Narcissus", re-telling the Greek legend through ballet. Narcissus appears in the Disney adaptation of "Hercules". In the film, he is portrayed as an Olympian god with purple skin. In the film Bab'Aziz, directed by Nacer Khemir, a Narcissus like character was portrayed by an ancient prince who sat by a pond for days after days and looked at the reflection of his own soul. He was referred to as 'The prince who contemplated his soul'. "Pink Narcissus" is an artistic film by James Bidgood about the fantasies of a hustler. The escape craft Ripley boards in the 1979 Ridley Scott film Alien is called the Narcissus. Narcissus is the name of Laurel and Hardy's goat in their 1940 film "Saps At Sea". The Neon Demon, a 2016 psychological horror film by Nicolas Winding Refn, is loosely based on the story of Narcissus. Narcissus is the name of the host club in the 2018 Japanese drama Todome no Kiss. The lead character, Otaro Dojima (Kento Yamazaki), works in the nightclub as a sought-after host under the stage name Eight and just like Narcissus, he is narcissistic and disregards the feelings of others; he uses women for money and power. National Medal Of Arts recipient Morten Lauridsen wrote a choral work entitled "Dirait-on" based on the poem by Rainer Maria Rilke. "Narcissus" is a popular melody from "Water Scenes" by American composer Ethelbert Nevin. "Supper's Ready" by Genesis (ca. 1972), a near-23-minute epic song laden with religious and mythological imagery, refers to the myth of Narcissus as follows: "A young figure sits still by the pool / He's been stamped "Human Bacon" by some butchery tool / (He is you) / Social Security took care of this lad. / We watch in reverence, as Narcissus is turned to a flower. / A flower?". The movement is titled "How Dare I Be So Beautiful?". Progressive metal band Threshold referenced the myth with an 11-minute epic titled "Narcissus", the closing track on their album "Hypothetical". Greek metal band Septic Flesh recorded a song about Narcissus (called "Narcissus") on their album "Communion". The cerebral rock band "Glass Wave" retells the Narcissus story from the perspective of the nymph Echo in their song "Echo," from their self-titled album "Glass Wave" (2010). "Narcissus in a Red Dress" by The Like was released on "The Like EP" and their album "Release Me". The Canadian band Hedley has written a song about Narcissus (called "Narcissist"). One line goes "He falls in love with his reflection in the glass / He can't resist who's staring back" Composer Nikolai Tcherepnin wrote his ballet "Narcisse et Echo, Op. 40" in 1911 for Sergei Diaghilev's Ballets Russes and was danced by Nijinski. Uruguayan band El Cuarteto de Nos wrote the song "Me Amo" (I Love Myself) in which the chorus sings "como Narciso soy" (I am like Narcissus). In 2010, Swedish electronic artist pacific! released "Narcissus" an album and ballet staged in Gothenburg. In 1994, composer Mark Applebaum composed Narcissus: Strata/Panacea for marimba solo. This work comprised one movement of the larger Janus Cycle, for mixed instrumentation. In 1987, Thea Musgrave was commissioned by a consortium of four flutists for a solo work. She composed Narcissus for flute and digital delay. German composer Matthias Pintscher composed his first cello concerto based on this mythology figure, titled "Reflections on Narcissus". In Marilyn Manson's song Deep Six, the first verse mentions Zeus in conversation with Narcissus. Narcissus has been a subject for many painters including: Caravaggio, Poussin, Turner, Dalí, Waterhouse, Carpioni, Lagrenée, and Roos. Sculptors such as Paul Dubois, John Gibson, Henri-Léon Gréber, Benvenuto Cellini and Hubert Netzer have sculpted Narcissus.
https://en.wikipedia.org/wiki?curid=22099
Naval mine A naval mine is a self-contained explosive device placed in water to damage or destroy surface ships or submarines. Unlike depth charges, mines are deposited and left to wait until they are triggered by the approach of, or contact with, any vessel. Naval mines can be used offensively, to hamper enemy shipping movements or lock vessels into a harbour; or defensively, to protect friendly vessels and create "safe" zones. Mines allow the minelaying force commander to concentrate warships or defensive assets in mine-free areas giving the adversary three choices: undertake an expensive and time-consuming minesweeping effort, accept the casualties of challenging the minefield, or use the unmined waters where the greatest concentration of enemy firepower will be encountered. Modern mines containing high explosives detonated by complex electronic fuze mechanisms are much more effective than early gunpowder mines requiring physical ignition. Mines may be placed by aircraft, ships, submarines, or individual swimmers and boatmen. Although international law requires signatory nations to declare mined areas, precise locations remain secret; and non-complying individuals may not disclose minelaying. While mines threaten only those who choose to traverse waters which may be mined, the possibility of activating a mine is a powerful disincentive to shipping. Mines can be laid in many ways: by purpose-built minelayers, refitted ships, submarines or aircraft—and even by dropping them into a harbour by hand. They can be inexpensive: some variants can cost as little as US$2000, though more sophisticated mines can cost millions of dollars, be equipped with several kinds of sensors and deliver a warhead by rocket or torpedo. Their flexibility and cost-effectiveness make mines attractive to the less powerful belligerent in asymmetric warfare. The cost of producing and laying a mine is usually between 0.5% and 10% of the cost of removing it, and it can take up to 200 times as long to clear a minefield as to lay it. Parts of some World War II naval minefields still exist because they are too extensive and expensive to clear. Some 1940s-era mines may remain dangerous for many years. Mines have been employed as offensive or defensive weapons in rivers, lakes, estuaries, seas and oceans, but they can also be used as tools of psychological warfare. Offensive mines are placed in enemy waters, outside harbours and across important shipping routes to sink both merchant and military vessels. Defensive minefields safeguard key stretches of coast from enemy ships and submarines, forcing them into more easily defended areas, or keeping them away from sensitive ones. Ship owners are reluctant to send their ships though known minefields. Port authorities may attempt to clear a mined area, but those without effective minesweeping equipment may cease using the area. Transit of a mined area will be attempted only when strategic interests outweigh potential losses. The decision-maker's perception of the minefield is a critical factor. Minefields designed for psychological effect are usually placed on trade routes to stop ships from reaching an enemy nation. They are often spread thinly, to create an impression of minefields existing across large areas. A single mine inserted strategically on a shipping route can stop maritime movements for days while the entire area is swept. A mine's capability to sink ships makes it a credible threat, but minefields work more on the mind than on ships. International law, specifically the Eighth Hague Convention of 1907, requires nations to declare when they mine an area, to make it easier for civil shipping to avoid the mines. The warnings do not have to be specific; for example, during World War II, Britain declared simply that it had mined the English Channel, North Sea and French coast. Precursors to naval mines were first invented by Chinese innovators of Imperial China and were described in thorough detail by the early Ming dynasty artillery officer Jiao Yu, in his 14th century military treatise known as the "Huolongjing". Chinese records tell of naval explosives in the 16th century, used to fight against Japanese pirates ("wokou"). This kind of naval mine was loaded in a wooden box, sealed with putty. General Qi Jiguang made several timed, drifting explosives, to harass Japanese pirate ships. The "Tiangong Kaiwu" ("The Exploitation of the Works of Nature") treatise, written by Song Yingxing in 1637, describes naval mines with a rip cord pulled by hidden ambushers located on the nearby shore who rotated a steel wheellock flint mechanism to produce sparks and ignite the fuse of the naval mine. Although this is the rotating steel wheellock's first use in naval mines, Jiao Yu described their use for land mines in the 14th century. The first plan for a sea mine in the West was by Ralph Rabbards, who presented his design to Queen Elizabeth I of England in 1574. The Dutch inventor Cornelius Drebbel was employed in the Office of Ordnance by King Charles I of England to make weapons, including the failed "floating petard". Weapons of this type were apparently tried by the English at the Siege of La Rochelle in 1627. American David Bushnell developed the first American naval mine, for use against the British in the American War of Independence. It was a watertight keg filled with gunpowder that was floated toward the enemy, detonated by a sparking mechanism if it struck a ship. It was used on the Delaware River as a drift mine. In 1812 Russian engineer Pavel Shilling exploded an underwater mine using an electrical circuit. In 1842 Samuel Colt used an electric detonator to destroy a moving vessel to demonstrate an underwater mine of his own design to the United States Navy and President John Tyler. However, opposition from former president John Quincy Adams, scuttled the project as "not fair and honest warfare". In 1854, during the unsuccessful attempt of the Anglo-French fleet to seize the Kronstadt fortress, British steamships HMS "Merlin" (9 June 1855, the first successful mining in history), HMS "Vulture" and HMS "Firefly" suffered damage due to the underwater explosions of Russian naval mines. Russian naval specialists set more than 1500 naval mines, or "infernal machines", designed by Moritz von Jacobi and by Immanuel Nobel, in the Gulf of Finland during the Crimean War of 1853–1856. The mining of "Vulcan" led to the world's first minesweeping operation. During the next 72 hours, 33 mines were swept. The Jacobi mine was designed by German-born, Russian engineer Jacobi, in 1853. The mine was tied to the sea bottom by an anchor. A cable connected it to a galvanic cell which powered it from the shore, the power of its explosive charge was equal to of black powder. In the summer of 1853, the production of the mine was approved by the Committee for Mines of the Ministry of War of the Russian Empire. In 1854, 60 Jacobi mines were laid in the vicinity of the Forts Pavel and Alexander (Kronstadt), to deter the British Baltic Fleet from attacking them. It gradually phased out its direct competitor the Nobel mine on the insistence of Admiral Fyodor Litke. The Nobel mines were bought from Swedish industrialist Immanuel Nobel who had entered into collusion with Russian head of navy Alexander Sergeyevich Menshikov. Despite their high cost (100 Russian rubles) the Nobel mines proved to be faulty, exploding while being laid, failing to explode or detaching from their wires and drifting uncontrollably, at least 70 of them were subsequently disarmed by the British. In 1855, 301 more Jacobi mines were laid around Krostadt and Lisy Nos. British ships did not dare to approach them. In the 19th century, mines were called torpedoes, a name probably conferred by Robert Fulton after the torpedo fish, which gives powerful electric shocks. A spar torpedo was a mine attached to a long pole and detonated when the ship carrying it rammed another one and withdrew a safe distance. The submarine used one to sink on 17 February 1864. A Harvey torpedo was a type of floating mine towed alongside a ship and was briefly in service in the Royal Navy in the 1870s. Other "torpedoes" were attached to ships or propelled themselves. One such weapon called the Whitehead torpedo after its inventor, caused the word "torpedo" to apply to self-propelled underwater missiles as well as to static devices. These mobile devices were also known as "fish torpedoes". The American Civil War of 1861–1865 also saw the successful use of mines. The first ship sunk by a mine, , foundered in 1862 in the Yazoo River. Rear Admiral David Farragut's famous/apocryphal command during the Battle of Mobile Bay in 1864, "Damn the torpedoes, full speed ahead!" refers to a minefield laid at Mobile, Alabama. After 1865 the United States adopted the mine as its primary weapon for coastal defense. In the decade following 1868, Major Henry Larcom Abbot carried out a lengthy set of experiments to design and test moored mines that could be exploded on contact or be detonated at will as enemy shipping passed near them. This initial development of mines in the United States took place under the purview of the U.S. Army Corps of Engineers, which trained officers and men in their use at the Engineer School of Application at Willets Point, New York (later named Fort Totten). In 1901 underwater minefields became the responsibility of the US Army's Artillery Corps, and in 1907 this was a founding responsibility of the United States Army Coast Artillery Corps. The Imperial Russian Navy, a pioneer in mine warfare, successfully deployed mines against the Ottoman Navy during both the Crimean War and the Russo-Turkish War (1877-1878). During the Battle of Tamsui (1884), in the Keelung Campaign of the Sino-French War, Chinese forces in Taiwan under Liu Mingchuan took measures to reinforce Tamsui against the French; they planted nine torpedo mines in the river and blocked the entrance. During the Boxer Rebellion, Imperial Chinese forces deployed a command-detonated mine field at the mouth of the Peiho river before the Dagu forts, to prevent the western Allied forces from sending ships to attack. The next major use of mines was during the Russo-Japanese War of 1904–1905. Two mines blew up when the struck them near Port Arthur, sending the holed vessel to the bottom and killing the fleet commander, Admiral Stepan Makarov, and most of his crew in the process. The toll inflicted by mines was not confined to the Russians, however. The Japanese Navy lost two battleships, four cruisers, two destroyers and a torpedo-boat to offensively laid mines during the war. Most famously, on 15 May 1904, the Russian minelayer "Amur" planted a 50-mine minefield off Port Arthur and succeeded in sinking the Japanese battleships and . Following the end of the Russo-Japanese War, several nations attempted to have mines banned as weapons of war at the Hague Peace Conference (1907). Many early mines were fragile and dangerous to handle, as they contained glass containers filled with nitroglycerin or mechanical devices that activated a blast upon tipping. Several mine-laying ships were destroyed when their cargo exploded. Beginning around the start of the 20th century, submarine mines played a major role in the defense of U.S. harbors against enemy attack as part of the Endicott and Taft Programs. The mines employed were controlled mines, anchored to the bottoms of the harbors and detonated under control from large mine casemates on shore. During World War I, mines were used extensively to defend coasts, coastal shipping, ports and naval bases around the globe. The Germans laid mines in shipping lanes to sink merchant and naval vessels serving Britain. The Allies targeted the German U-boats in the Strait of Dover and the Hebrides. In an attempt to seal up the northern exits of the North Sea, the Allies developed the North Sea Mine Barrage. During a period of five months from June 1918 almost 70,000 mines were laid spanning the North Sea's northern exits. The total number of mines laid in the North Sea, the British East Coast, Straits of Dover, and Heligoland Bight is estimated at 190,000 and the total number during the whole of WWI was 235,000 sea mines. Clearing the barrage after the war took 82 ships and five months, working around the clock. It was also during World War I, that the naval mine sunk its largest vessel ever, the British hospital ship, HMHS "Britannic", which was the sister ship of the RMS "Titanic." During World War II, the U-boat fleet, which dominated much of the battle of the Atlantic, was small at the beginning of the war and much of the early action by German forces involved mining convoy routes and ports around Britain. German submarines also operated in the Mediterranean Sea, in the Caribbean Sea, and along the U.S. coast. Initially, contact mines (requiring a ship to physically strike a mine to detonate it) were employed, usually tethered at the end of a cable just below the surface of the water. Contact mines usually blew a hole in ships' hulls. By the beginning of World War II, most nations had developed mines that could be dropped from aircraft, some of which floated on the surface, making it possible to lay them in enemy harbours. The use of dredging and nets was effective against this type of mine, but this consumed valuable time and resources, and required harbours to be closed. Later, some ships survived mine blasts, limping into port with buckled plates and broken backs. This appeared to be due to a new type of mine, detecting ships by their proximity to the mine (an influence mine) and detonating at a distance, causing damage with the shock wave of the explosion. Ships that had successfully run the gantlet of the Atlantic crossing were sometimes destroyed entering freshly cleared British harbours. More shipping was being lost than could be replaced, and Churchill ordered the intact recovery of one of these new mines to be of the highest priority. The British experienced a stroke of luck in November 1939, when a German mine was dropped from an aircraft onto the mud flats off Shoeburyness during low tide. Additionally, the land belonged to the army and a base with men and workshops was at hand. Experts were dispatched from HMS Vernon to investigate the mine. The Royal Navy knew that mines could use magnetic sensors, Britain having developed magnetic mines in World War I, so everyone removed all metal, including their buttons, and made tools of non-magnetic brass. They disarmed the mine and rushed it to the labs at HMS Vernon, where scientists discovered that the mine had a magnetic arming mechanism. A large ferrous object passing through the Earth's magnetic field will concentrate the field through it; the mine's detector was designed to trigger as a ship passed over, when the Earth's magnetic field was concentrated in the ship and away from the mine. The mine detected this loss of the magnetic field which caused it to detonate. The mechanism had an adjustable sensitivity, calibrated in milligauss. The U.S. began adding delay counters to their magnetic mines in June 1945. From this data, known methods were used to clear these mines. Early methods included the use of large electromagnets dragged behind ships or below low-flying aircraft (a number of older bombers like the Vickers Wellington were used for this). Both of these methods had the disadvantage of "sweeping" only a small strip. A better solution was found in the "Double-L Sweep" using electrical cables dragged behind ships that passed large pulses of current through the seawater. This created a large magnetic field and swept the entire area between the two ships. The older methods continued to be used in smaller areas. The Suez Canal continued to be swept by aircraft, for instance. Wartime Japanese sweep methods, by contrast, never advanced much past 1930s standards, and failed entirely to keep up with new American mines, clearing no more than 15% of all the mines laid in Japan's coastal waters. IJN's minesweeping force was insufficient, with 350 ships, and 20,000 men. While these methods were useful for clearing mines from local ports, they were of little or no use for enemy-controlled areas. These were typically visited by warships, and the majority of the fleet then underwent a massive degaussing process, where their hulls had a slight "south" bias induced into them which offset the concentration effect almost to zero. Initially, major warships and large troopships had a copper "degaussing coil" fitted around the perimeter of the hull, energized by the ship's electrical system whenever in suspected magnetic-mined waters. Some of the first to be so fitted were the carrier HMS "Ark Royal" and the liners and . It was a photo of one of these liners in New York harbor, showing the degaussing coil, which revealed to German Naval Intelligence the fact that the British were using degaussing methods to combat their magnetic mines. This was felt to be impractical for smaller warships and merchant vessels, mainly because the ships lacked the generating capacity to energise such a coil. It was found that "wiping" a current-carrying cable up and down a ship's hull temporarily cancelled the ships' magnetic signature sufficiently to nullify the threat. This started in late 1939, and by 1940 merchant vessels and the smaller British warships were largely immune for a few months at a time until they once again built up a field. The cruiser HMS Belfast is just one example of a ship which was struck by a magnetic mine during this time. On 21 November 1939, a mine broke her keel which damaged her engine and boiler rooms, as well as injuring 46 men with one man later dying from his injuries. She was towed to Rosyth for repairs. Incidents like this, resulted in many of the boats that sailed to Dunkirk being degaussed in a marathon four-day effort by degaussing stations. The Allies and Germany deployed acoustic mines in WW II, against which even wooden-hulled ships (in particular minesweepers) remained vulnerable. Japan developed sonic generators to sweep these; the gear was not ready by war's end. The primary method Japan used was small air-delivered bombs. This was profligate and ineffectual; used against acoustic mines at Penang, 200 bombs were needed to detonate just 13 mines. The Germans developed a pressure-activated mine and planned to deploy it as well, but they saved it for later use when it became clear the British had defeated the magnetic system. The U.S. also deployed these, adding "counters" which would allow a variable number of ships to pass unharmed before detonating. This made them a great deal harder to sweep. Japan's antiquated sweep methods, lifting mines in nets, accidentally proved useful against these mines; it remained too slow and hazardous to be truly effective, especially in light of the high numbers being laid. Mining campaigns could have devastating consequences. The U.S. effort against Japan, for instance, closed major ports, such as Hiroshima, for days, and by the end of the Pacific War had cut the amount of freight passing through Kobe–Yokohama by 90%. When the war ended, more than 25,000 U.S.-laid mines were still in place, and the Navy proved unable to sweep them all, limiting efforts to critical areas. After sweeping for almost a year, in May 1946, the Navy abandoned the effort with 13,000 mines still unswept. Over the next thirty years, more than 500 minesweepers (of a variety of types) were damaged or sunk clearing them. Since World War II, mines have damaged 14 United States Navy ships, whereas air and missile attacks have damaged four. During the Korean War, mines laid by North Korean forces caused 70% of the casualties suffered by U.S. naval vessels and caused 4 sinkings. During the Iran–Iraq War from 1980 to 1988, the belligerents mined several areas of the Persian Gulf and nearby waters. On 24 July 1987, the supertanker "SS" Bridgeton was mined by Iran near Farsi Island. On 14 April 1988, struck an Iranian M-08/39 mine in the central Persian Gulf shipping lane, wounding 10 sailors. In the summer of 1984, magnetic sea mines damaged at least 19 ships in the Red Sea. The U.S. concluded Libya was probably responsible for the minelaying. In response the U.S., Britain, France, and three other nations launched Operation Intense Look, a minesweeping operation in the Red Sea involving more than 46 ships. On the orders of the Reagan administration, the CIA mined Nicaragua's Sandino port in 1984 in support of the Contra guerrilla group. A Soviet tanker was among the ships damaged by these mines. In 1986, in the case of "Nicaragua v. United States", the International Court of Justice ruled that this mining was a violation of international law. During the Gulf War, Iraqi naval mines severely damaged and . When the war concluded, eight countries conducted clearance operations. Houthi forces in the Yemeni Civil War have made frequent use of naval mines, laying over 150 in the Red Sea throughout the conflict. Naval mines may be classified into three major groups; contact, remote and influence mines. The earliest mines were usually of this type. They are still used today, as they are extremely low cost compared to any other anti-ship weapon and are effective, both as a psychological weapon and as a method to sink enemy ships. Contact mines need to be touched by the target before they detonate, limiting the damage to the direct effects of the explosion and usually affecting only the vessel that triggers them. Early mines had mechanical mechanisms to detonate them, but these were superseded in the 1870s by the "Hertz horn" (or "chemical horn"), which was found to work reliably even after the mine had been in the sea for several years. The mine's upper half is studded with hollow lead protuberances, each containing a glass vial filled with sulfuric acid. When a ship's hull crushes the metal horn, it cracks the vial inside it, allowing the acid to run down a tube and into a lead–acid battery which until then contained no acid electrolyte. This energizes the battery, which detonates the explosive. Earlier forms of the detonator employed a vial of sulfuric acid surrounded by a mixture of potassium perchlorate and sugar. When the vial was crushed, the acid ignited the perchlorate-sugar mix, and the resulting flame ignited the gunpowder charge. During the initial period of World War I, the Royal Navy used contact mines in the English Channel and later in large areas of the North Sea to hinder patrols by German submarines. Later, the American antenna mine was widely used because submarines could be at any depth from the surface to the seabed. This type of mine had a copper wire attached to a buoy that floated above the explosive charge which was weighted to the seabed with a steel cable. If a submarine's steel hull touched the copper wire, the slight voltage change caused by contact between two dissimilar metals was amplified and detonated the explosives. Limpet mines are a special form of contact mine that are manually attached to the target by magnets and remain in place. They are named because of the similarity to the limpet, a mollusk. Generally, this mine type is set to float just below the surface of the water or as deep as five meters. A steel cable connecting the mine to an anchor on the seabed prevents it from drifting away. The explosive and detonating mechanism is contained in a buoyant metal or plastic shell. The depth below the surface at which the mine floats can be set so that only deep draft vessels such as aircraft carriers, battleships or large cargo ships are at risk, saving the mine from being used on a less valuable target. In littoral waters it is important to ensure that the mine does not become visible when the sea level falls at low tide, so the cable length is adjusted to take account of tides. During WWII there were mines that could be moored in 300m-deep water. Floating mines typically have a mass of around 200 kg, including 80 kg of explosives e.g. TNT, minol or amatol. A special form of moored contact mines are those equipped with a plummet. When the mine is launched (1), the mine with the anchor floats first and the lead plummet sinks from it (2). In doing so, the plummet unwinds a wire, the deep line, which is used to set the depth of the mine below the water surface before it is launched (3). When the deep line has been unwound to a set length, the anchor is flooded and the mine is released from the anchor (4). The anchor begins to sink and the mooring cable unwinds until the plummet reaches the sea floor (5). Due to the decreasing tension on the deep line, the mooring cable is clamped. The anchor sinks further down to the bottom of the sea pulling the mine as deep below the water surface as the deep line has been unwound (6). Thus, even without knowing the exact depth, an exact depth of the mine below the water surface can be set. Limited only by the maximum length of the mooring cable. Drifting mines were occasionally used during World War I and World War II. However, they were more feared than effective. Sometimes floating mines break from their moorings and become drifting mines; modern mines are designed to deactivate in this event. After several years at sea, the deactivation mechanism might not function as intended and the mines may remain live. Admiral Jellicoe's British fleet did not pursue and destroy the outnumbered German High Seas Fleet when it turned away at the Battle of Jutland because he thought they were leading him into a trap: he believed it possible that the Germans were either leaving floating mines in their wake, or were drawing him towards submarines, although neither of these was the case. After World War I the drifting contact mine was banned, but was occasionally used during World War II. The drifting mines were much harder to remove than tethered mines after the war, and they caused about the same damage to both sides. Churchill promoted "Operation Royal Marine" in 1940 and again in 1944 where floating mines were put into the Rhine in France to float down the river, becoming active after a time calculated to be long enough to reach German territory. Frequently used in combination with coastal artillery and hydrophones, controlled mines (or command detonation mines) can be in place in peacetime, which is a huge advantage in blocking important shipping routes. The mines can usually be turned into "normal" mines with a switch (which prevents the enemy from simply capturing the controlling station and deactivating the mines), detonated on a signal or be allowed to detonate on their own. The earliest ones were developed around 1812 by Robert Fulton. The first remotely controlled mines were moored mines used in the American Civil War, detonated electrically from shore. They were considered superior to contact mines because they did not put friendly shipping at risk. The extensive American fortifications program initiated by the Board of Fortifications in 1885 included remotely controlled mines, which were emplaced or in reserve from the 1890s through the end of World War II. Modern examples usually weigh , including of explosives (TNT or torpex). These mines are triggered by the influence of a ship or submarine, rather than direct contact. Such mines incorporate electronic sensors designed to detect the presence of a vessel and detonate when it comes within the blast range of the warhead. The fuses on such mines may incorporate one or more of the following sensors: magnetic, passive acoustic or water pressure displacement caused by the proximity of a vessel. First used during WWI, their use became more general in WWII. The sophistication of influence mine fuses has increased considerably over the years as first transistors and then microprocessors have been incorporated into designs. Simple magnetic sensors have been superseded by total-field magnetometers. Whereas early magnetic mine fuses would respond only to changes in a single component of a target vessel's magnetic field, a total field magnetometer responds to changes in the magnitude of the total background field (thus enabling it to better detect even degaussed ships). Similarly, the original broadband hydrophones of 1940s acoustic mines (which operate on the integrated volume of all frequencies) have been replaced by narrow-band sensors which are much more sensitive and selective. Mines can now be programmed to listen for highly specific acoustic signatures (e.g. a gas turbine powerplant or cavitation sounds from a particular design of propeller) and ignore all others. The sophistication of modern electronic mine fuzes incorporating these digital signal processing capabilities makes it much more difficult to detonate the mine with electronic countermeasures because several sensors working together (e.g. magnetic, passive acoustic and water pressure) allow it to ignore signals which are not recognised as being the unique signature of an intended target vessel. Modern influence mines such as the BAE Stonefish are computerised, with all the programmability this implies, such as the ability to quickly load new acoustic signatures into fuses, or program them to detect a single, highly distinctive target signature. In this way, a mine with a passive acoustic fuze can be programmed to ignore all friendly vessels and small enemy vessels, only detonating when a very large enemy target passes over it. Alternatively, the mine can be programmed specifically to ignore all surface vessels regardless of size and exclusively target submarines. Even as far back as WWII it was possible to incorporate a "ship counter" function in mine fuzes. This might set the mine to ignore the first two ships passing over it (which could be minesweepers deliberately trying to trigger mines) but detonate when the third ship passes overhead, which could be a high-value target such as an aircraft carrier or oil tanker. Even though modern mines are generally powered by a long life lithium battery, it is important to conserve power because they may need to remain active for months or even years. For this reason, most influence mines are designed to remain in a semi-dormant state until an unpowered (e.g. deflection of a mu-metal needle) or low-powered sensor detects the possible presence of a vessel, at which point the mine fuze powers up fully and the passive acoustic sensors will begin to operate for some minutes. It is possible to program computerised mines to delay activation for days or weeks after being laid. Similarly, they can be programmed to self-destruct or render themselves safe after a preset period of time. Generally, the more sophisticated the mine design, the more likely it is to have some form of anti-handling device to hinder clearance by divers or remotely piloted submersibles. The moored mine is the backbone of modern mine systems. They are deployed where water is too deep for bottom mines. They can use several kinds of instruments to detect an enemy, usually a combination of acoustic, magnetic and pressure sensors, or more sophisticated optical shadows or electro potential sensors. These cost many times more than contact mines. Moored mines are effective against most kinds of ships. As they are cheaper than other anti-ship weapons they can be deployed in large numbers, making them useful area denial or "channelizing" weapons. Moored mines usually have lifetimes of more than 10 years, and some almost unlimited. These mines usually weigh , including of explosives (RDX). In excess of of explosives the mine becomes inefficient, as it becomes too large to handle and the extra explosives add little to the mine's effectiveness. Bottom mines (sometimes called ground mines) are used when the water is no more than deep or when mining for submarines down to around . They are much harder to detect and sweep, and can carry a much larger warhead than a moored mine. Bottom mines commonly utilize multiple types of sensors, which are less sensitive to sweeping. These mines usually weigh between , including between of explosives. Several specialized mines have been developed for other purposes than the common minefield. The bouquet mine is a single anchor attached to several floating mines. It is designed so that when one mine is swept or detonated, another takes its place. It is a very sensitive construction and lacks reliability. The anti-sweep mine is a very small mine (40 kg warhead) with as small a floating device as possible. When the wire of a mine sweep hits the anchor wire of the mine, it drags the anchor wire along with it, pulling the mine down into contact with the sweeping wire. That detonates the mine and cuts the sweeping wire. They are very cheap and usually used in combination with other mines in a minefield to make sweeping more difficult. One type is the Mark 23 used by the United States during World War II. The mine is hydrostatically controlled to maintain a pre-set depth below the water's surface independently of the rise and fall of the tide. The ascending mine is a floating distance mine that may cut its mooring or in some other way float higher when it detects a target. It lets a single floating mine cover a much larger depth range. These are mines containing a moving weapon as a warhead, either a torpedo or a rocket. Rocket mine: a Russian invention, the rocket mine is a bottom distance mine that fires a homing high-speed rocket (not torpedo) upwards towards the target. It is intended to allow a bottom mine to attack surface ships as well as submarines from a greater depth. One type is the Te-1 rocket propelled mine. Torpedo mine: the torpedo mine is a self-propelled variety, able to lie in wait for a target and then pursue it e.g. the Mark 60 CAPTOR. Generally, torpedo mines incorporate computerised acoustic and magnetic fuzes. The U.S. Mark 24 "mine", code-named Fido, was actually an ASW homing torpedo. The mine designation was disinformation to conceal its function. The mine is propelled to its intended position by propulsion equipment such as a torpedo. After reaching its destination, it sinks to the seabed and operates like a standard mine. It differs from the homing mine in that its mobile stage is before it lies in wait, rather than as part of the attacking phase. One such design is the Mk 67 submarine launched mobile mine (which is based on a Mark 37 torpedo) are capable of travelling as far as 10 miles through or into a channel, harbor, shallow water area and other zones which would normally be inaccessible to craft laying the device. After reaching the target area they sink to the sea bed and act like conventionally laid influence mines. During the Cold War a test was conducted with naval mine fitted with tactical nuclear warheads for the "Baker" shot of Operation Crossroads. This weapon was experimental and never went into production. There have been some reports that North Korea may be developing a nuclear mine The Seabed Arms Control Treaty prohibits the placement of nuclear weapons on the seabed beyond a 12-mile coast zone. This comprises two moored, floating contact mines which are tethered together by a length of steel cable or chain. Typically, each mine is situated approximately away from its neighbour, and each floats a few metres below the surface of the ocean. When the target ship hits the steel cable, the mines on either side are drawn down the side of the ship's hull, exploding on contact. In this manner it is almost impossible for target ships to pass safely between two individually moored mines. Daisy-chained mines are a very simple concept which was used during World War II. Plastic drums filled with sand or concrete are periodically rolled off the side of ships as real mines are laid in large mine-fields. These inexpensive false targets (designed to be of a similar shape and size as genuine mines) are intended to slow down the process of mine clearance: a mine-hunter is forced to investigate each suspicious sonar contact on the sea bed, whether it is real or not. Often a maker of naval mines will provide both training and dummy versions of their mines. Historically several methods were used to lay mines. During WWI and WWII, the Germans used U-boats to lay mines around the UK. In WWII, aircraft came into favour for mine laying with one of the largest examples being the mining of the Japanese sea routes in Operation Starvation. Laying a minefield is a relatively fast process with specialized ships, which is today the most common method. These minelayers can carry several thousand mines and manoeuvre with high precision. The mines are dropped at predefined intervals into the water behind the ship. Each mine is recorded for later clearing, but it is not unusual for these records to be lost together with the ships. Therefore, many countries demand that all mining operations be planned on land and records kept so that the mines can later be recovered more easily. Other methods to lay minefields include: In some cases, mines are automatically activated upon contact with the water. In others, a safety lanyard is pulled (one end attached to the rail of a ship, aircraft or torpedo tube) which starts an automatic timer countdown before the arming process is complete. Typically, the automatic safety-arming process takes some minutes to complete. This allows the people laying the mines sufficient time to move out of its activation and blast zones. In the 1930s, Germany had experimented with the laying of mines by aircraft. It became a crucial element in their overall mining strategy. Aircraft had the advantage of speed, and they would never get caught in their own minefields. German mines held a large explosive charge. From April to June 1940, the Luftwaffe laid 1,000 mines in British waters. Soviet ports were mined, as was the Arctic convoy route to Murmansk. The Heinkel He 115 could carry two medium or one large mine while the Heinkel He 59, Dornier Do 18, Junkers Ju 88 and Heinkel He 111 could carry more. The USSR was relatively ineffective in its use of naval mines in WWII in comparison with its record in previous wars. Small mines were developed for use in rivers and lakes, and special mines for shallow water. A very large chemical mine was designed to sink through ice with the aid of a melting compound. Special aerial mine designs finally arrived in 1943–1944, the AMD-500 and AMD-1000. Various Soviet Naval Aviation torpedo bombers were pressed into the role of aerial mining in the Baltic Sea and the Black Sea, including Ilyushin DB-3s, Il-4s and Lend Lease Douglas Boston IIIs. In September 1939, the UK announced the placement of extensive defensive minefields in waters surrounding the Home Islands. Offensive aerial mining operations began in April 1940 when 38 mines were laid at each of these locations: the Elbe River, the port of Lübeck and the German naval base at Kiel. In the next 20 months, mines delivered by aircraft sank or damaged 164 Axis ships with the loss of 94 aircraft. By comparison, direct aerial attacks on Axis shipping had sunk or damaged 105 vessels at a cost of 373 aircraft lost. The advantage of aerial mining became clear, and the UK prepared for it. A total of 48,000 aerial mines were laid by the Royal Air Force (RAF) in the European Theatre during World War II. As early as 1942, American mining experts such as Naval Ordnance Laboratory scientist Dr. Ellis A. Johnson, CDR USNR, suggested massive aerial mining operations against Japan's "outer zone" (Korea and northern China) as well as the "inner zone", their home islands. First, aerial mines would have to be developed further and manufactured in large numbers. Second, laying the mines would require a sizable air group. The US Army Air Forces had the carrying capacity but considered mining to be the navy's job. The US Navy lacked suitable aircraft. Johnson set about convincing General Curtis LeMay of the efficacy of heavy bombers laying aerial mines. B-24 Liberators, PBY Catalinas and other bomber aircraft took part in localized mining operations in the Southwest Pacific and the China Burma India (CBI) theaters, beginning with a successful attack on the Yangon River in February 1943. Aerial minelaying operations involved a coalition of British, Australian and American aircrews, with the RAF and the Royal Australian Air Force (RAAF) carrying out 60% of the sorties and the USAAF and US Navy covering 40%. Both British and American mines were used. Japanese merchant shipping suffered tremendous losses, while Japanese mine sweeping forces were spread too thin attending to far-flung ports and extensive coastlines. Admiral Thomas C. Kinkaid, who directed nearly all RAAF mining operations in CBI, heartily endorsed aerial mining, writing in July 1944 that "aerial mining operations were of the order of 100 times as destructive to the enemy as an equal number of bombing missions against land targets." A single B-24 dropped three mines into Haiphong harbor in October 1943. One of those mines sank a Japanese freighter. Another B-24 dropped three more mines into the harbor in November, and a second freighter was sunk by a mine. The threat of the remaining mines prevented a convoy of ten ships from entering Haiphong; and six of those ships were sunk by attacks before they reached a safe harbor. The Japanese closed Haiphong to all steel-hulled ships for the remainder of the war after another small ship was sunk by one of the remaining mines, although they may not have realized no more than three mines remained. Using Grumman TBF Avenger torpedo bombers, the US Navy mounted a direct aerial mining attack on enemy shipping in Palau on 30 March 1944 in concert with simultaneous conventional bombing and strafing attacks. The dropping of 78 mines deterred 32 Japanese ships from escaping Koror harbor; and 23 of those immobilized ships were sunk in a subsequent bombing raid. The combined operation sank or damaged 36 ships. Two Avengers were lost, and their crews were recovered. The mines brought port usage to a halt for 20 days. Japanese mine sweeping was unsuccessful; and the Japanese abandoned Palau as a base when their first ship attempting to traverse the swept channel was damaged by a mine detonation. In March 1945, Operation Starvation began in earnest, using 160 of LeMay's B-29 Superfortress bombers to attack Japan's inner zone. Almost half of the mines were the US-built Mark 25 model, carrying 1250 lbs of explosives and weighing about 2,000 lbs. Other mines used included the smaller 1,000 lb Mark 26. Fifteen B-29s were lost while 293 Japanese merchant ships were sunk or damaged. Twelve thousand aerial mines were laid, a significant barrier to Japan's access to outside resources. Prince Fumimaro Konoe said after the war that the aerial mining by B-29s had been "equally as effective as the B-29 attacks on Japanese industry at the closing stages of the war when all food supplies and critical material were prevented from reaching the Japanese home islands." The United States Strategic Bombing Survey (Pacific War) concluded that it would have been more efficient to combine the United States's effective anti-shipping submarine effort with land- and carrier-based air power to strike harder against merchant shipping and begin a more extensive aerial mining campaign earlier in the war. Survey analysts projected that this would have starved Japan, forcing an earlier end to the war. After the war, Dr. Johnson looked at the Japan inner zone shipping results, comparing the total economic cost of submarine-delivered mines versus air-dropped mines and found that, though 1 in 12 submarine mines connected with the enemy as opposed to 1 in 21 for aircraft mines, the aerial mining operation was about ten times less expensive per enemy ton sunk. Between 600,000 and 1,000,000 naval mines of all types were laid in WWII. Advancing military forces worked to clear mines from newly-taken areas, but extensive minefields remained in place after the war. Air-dropped mines had an additional problem for mine sweeping operations: they were not meticulously charted. In Japan, much of the B-29 mine-laying work had been performed at high altitude, with the drifting on the wind of mines carried by parachute adding a randomizing factor to their placement. Generalized danger areas were identified, with only the quantity of mines given in detail. Mines used in Operation Starvation were supposed to be self-sterilizing, but the circuit did not always work. Clearing the mines from Japanese waters took so many years that the task was eventually given to the Japan Maritime Self-Defense Force. For the purpose of clearing all types of naval mines, the Royal Navy employed German crews and minesweepers from June 1945 to January 1948, organised in the German Mine Sweeping Administration (GMSA), which consisted of 27,000 members of the former "Kriegsmarine" and 300 vessels. Mine clearing was not always successful: a number of ships were damaged or sunk by mines after the war. Two such examples were the liberty ships "Pierre Gibault" which was scrapped after hitting a mine in a previously cleared area off the Greek island of Kythira in June 1945, and "Nathaniel Bacon" which hit a minefield off Civitavecchia, Italy in December 1945, caught fire, was beached, and broke in two. The damage that may be caused by a mine depends on the "shock factor value", a combination of the initial strength of the explosion and of the distance between the target and the detonation. When taken in reference to ship hull plating, the term "Hull Shock Factor" (HSF) is used, while keel damage is termed "Keel Shock Factor" (KSF). If the explosion is directly underneath the keel, then HSF is equal to KSF, but explosions that are not directly underneath the ship will have a lower value of KSF. Usually only created by contact mines, direct damage is a hole blown in the ship. Among the crew, fragmentation wounds are the most common form of damage. Flooding typically occurs in one or two main watertight compartments, which can sink smaller ships or disable larger ones. Contact mine damage often occurs at or close to the waterline near the bow, but depending on circumstances a ship could be hit anywhere on its outer hull surface (the mine attack being a good example of a contact mine detonating amidships and underneath the ship). The bubble jet effect occurs when a mine or torpedo detonates in the water a short distance away from the targeted ship. The explosion creates a bubble in the water, and due to the difference in pressure, the bubble will collapse from the bottom. The bubble is buoyant, and so it rises towards the surface. If the bubble reaches the surface as it collapses, it can create a pillar of water that can go over a hundred meters into the air (a "columnar plume"). If conditions are right and the bubble collapses onto the ship's hull, the damage to the ship can be extremely serious; the collapsing bubble forms a high-energy jet similar to shaped charge can break a metre-wide hole straight through the ship, flooding one or more compartments, and is capable of breaking smaller ships apart. The crew in the areas hit by the pillar are usually killed instantly. Other damage is usually limited. The Baengnyeong incident, in which the ROKS "Cheonan" broke in half and sank off the coast South Korea in 2010, was caused by the bubble jet effect, according to an international investigation. If the mine detonates at a distance from the ship, the change in water pressure causes the ship to resonate. This is frequently the most deadly type of explosion, if it is strong enough. The whole ship is dangerously shaken and everything on board is tossed around. Engines rip from their beds, cables from their holders, etc.. A badly shaken ship usually sinks quickly, with hundreds, or even thousands of small leaks all over the ship and no way to power the pumps. The crew fare no better, as the violent shaking tosses them around. This shaking is powerful enough to cause disabling injury to knees and other joints in the body, particularly if the affected person stands on surfaces connected directly to the hull (such as steel decks). The resulting gas cavitation and shock-front-differential over the width of the human body is sufficient to stun or kill divers. Weapons are frequently a few steps ahead of countermeasures, and mines are no exception. In this field the British, with their large seagoing navy, have had the bulk of world experience, and most anti-mine developments, such as degaussing and the double-L sweep, were British inventions. When on operational missions, such as the recent invasion of Iraq, the US still relies on British and Canadian minesweeping services. The US has worked on some innovative mine-hunting countermeasures, such as the use of military dolphins to detect and flag mines. However, they are of questionable effectiveness. Mines in nearshore environments remain a particular challenge. They are small and as technology has developed they can have anechoic coatings, be non-metallic, and oddly shaped to resist detection. Further, oceanic conditions and the sea bottoms of the area of operations can degrade sweeping and hunting efforts. Mining countermeasures are far more expensive and time-consuming than mining operations, and that gap is only growing with new technologies. Ships can be designed to be difficult for mines to detect, to avoid detonating them. This is especially true for minesweepers and mine hunters that work in minefields, where a minimal signature outweighs the need for armour and speed. These ships have hulls of glass fibre or wood instead of steel to avoid magnetic signatures. These ships may use special propulsion systems, with low magnetic electric motors, to reduce magnetic signature, and Voith-Schneider propellers, to limit the acoustic signature. They are built with hulls that produce a minimal pressure signature. These measures create other problems. They are expensive, slow, and vulnerable to enemy fire. Many modern ships have a mine-warning sonar—a simple sonar looking forward and warning the crew if it detects possible mines ahead. It is only effective when the ship is moving slowly. (See also SQQ-32 Mine-hunting sonar) A steel-hulled ship can be "degaussed" (more correctly, de-oerstedted or depermed) using a special degaussing station that contains many large coils and induces a magnetic field in the hull with alternating current to demagnetize the hull. This is a rather problematic solution, as magnetic compasses need recalibration and all metal objects must be kept in exactly the same place. Ships slowly regain their magnetic field as they travel through the Earth's magnetic field, so the process has to be repeated every six months. A simpler variation of this technique, called "wiping", was developed by Charles F. Goodeve which saved time and resources. Between 1941 and 1943 the US Naval Gun factory (a division of the Naval Ordnance Laboratory) in Washington, D.C., built physical models of all US naval ships. Three kinds of steel were used in shipbuilding: mild steel for bulkheads, a mixture of mild steel and high tensile steel for the hull, and special treatment steel for armor plate. The models were placed within coils which could simulate the Earth's magnetic field at any location. The magnetic signatures were measured with degaussing coils. The objective was to reduce the vertical component of the combination of the Earth's field and the ship's field at the usual depth of German mines. From the measurements, coils were placed and coil currents determined to minimize the chance of detonation for any ship at any heading at any latitude. Some ships are built with magnetic inductors, large coils placed along the ship to counter the ship's magnetic field. Using magnetic probes in strategic parts of the ship, the strength of the current in the coils can be adjusted to minimize the total magnetic field. This is a heavy and clumsy solution, suited only to small-to-medium-sized ships. Boats typically lack the generators and space for the solution, while the amount of power needed to overcome the magnetic field of a large ship is impractical. Active countermeasures are ways to clear a path through a minefield or remove it completely. This is one of the most important tasks of any mine warfare flotilla. A sweep is either a contact sweep, a wire dragged through the water by one or two ships to cut the mooring wire of floating mines, or a distance sweep that mimics a ship to detonate the mines. The sweeps are dragged by minesweepers, either purpose-built military ships or converted trawlers. Each run covers between one and two hundred meters, and the ships must move slowly in a straight line, making them vulnerable to enemy fire. This was exploited by the Turkish army in the Battle of Gallipoli in 1915, when mobile howitzer batteries prevented the British and French from clearing a way through minefields. If a contact sweep hits a mine, the wire of the sweep rubs against the mooring wire until it is cut. Sometimes "cutters", explosive devices to cut the mine's wire, are used to lessen the strain on the sweeping wire. Mines cut free are recorded and collected for research or shot with a deck gun. Minesweepers protect themselves with an oropesa or paravane instead of a second minesweeper. These are torpedo-shaped towed bodies, similar in shape to a Harvey Torpedo, that are streamed from the sweeping vessel thus keeping the sweep at a determined depth and position. Some large warships were routinely equipped with paravane sweeps near the bows in case they inadvertently sailed into minefields—the mine would be deflected towards the paravane by the wire instead of towards the ship by its wake. More recently, heavy-lift helicopters have dragged minesweeping sleds, as in the 1991 Persian Gulf War. The distance sweep mimics the sound and magnetism of a ship and is pulled behind the sweeper. It has floating coils and large underwater "drums". It is the only sweep effective against bottom mines. During WWII, RAF Coastal Command used Vickers Wellington bombers Wellington DW.Mk I fitted with degaussing coils to trigger magnetic mines. Modern influence mines are designed to discriminate against false inputs and are, therefore, much harder to sweep. They often contain inherent anti-sweeping mechanisms. For example, they may be programmed to respond to the unique noise of a particular ship-type, its associated magnetic signature and the typical pressure displacement of such a vessel. As a result, a mine-sweeper must accurately mimic the required target signature to trigger detonation. The task is complicated by the fact that an influence mine may have one or more of a hundred different potential target signatures programmed into it. Another anti-sweeping mechanism is a ship-counter in the mine fuze. When enabled, this allows detonation only after the mine fuze has been triggered a pre-set number of times. To further complicate matters, influence mines may be programmed to arm themselves (or disarm automatically—known as "self-sterilization") after a pre-set time. During the pre-set arming delay (which could last days or even weeks) the mine would remain dormant and ignore any target stimulus, whether genuine or false. When influence mines are laid in an ocean minefield, they may have various combinations of fuze settings configured. For example, some mines (with the acoustic sensor enabled) may become active within three hours of being laid, others (with the acoustic and magnetic sensors enabled) may become active after two weeks but have the ship-counter mechanism set to ignore the first two trigger events, and still others in the same minefield (with the magnetic and pressure sensors enabled) may not become armed until three weeks have passed. Groups of mines within this mine-field may have different target signatures which may or may not overlap. The fuzes on influence mines allow many different permutations, which complicates the clearance process. Mines with ship-counters, arming delays and highly specific target signatures in mine fuzes can falsely convince a belligerent that a particular area is clear of mines or has been swept effectively because a succession of vessels have already passed through safely. As naval mines have become more sophisticated, and able to discriminate between targets, so they have become more difficult to deal with by conventional sweeping. This has given rise to the practice of mine-hunting. Mine hunting is very different from sweeping, although some minehunters can do both tasks. Minehunting pays little attention to the nature of the mine itself. Nor does the method change much. At the current state of the art, Minehunting remains the best way to deal with influence mines proving to be both safer and more effective than sweeping. Specialized high-frequency sonars and high fidelity sidescaning sonar are used for mine location. Mines are hunted using sonar, then inspected and destroyed either by divers or ROVs (remote controlled unmanned mini-submarines). It is slow, but also the most reliable way to remove mines. Minehunting started during the Second World War, but it was only after the war that it became truly effective. Sea mammals (mainly the bottlenose dolphin) have been trained to hunt and mark mines, most famously by the U.S. Navy Marine Mammal Program. Mine-clearance dolphins were deployed in the Persian Gulf during the Iraq War in 2003. The US Navy claims that these dolphins were effective in helping to clear more than 100 antiship mines and underwater booby traps from Umm Qasr Port. French naval officer Jacques Yves Cousteau's Undersea Research Group was once involved in mine-hunting operations: They removed or detonated a variety of German mines, but one particularly defusion-resistant batch—equipped with acutely sensitive pressure, magnetic, and acoustic sensors and wired together so that one explosion would trigger the rest—was simply left undisturbed for years until corrosion would (hopefully) disable the mines. A more drastic method is simply to run a ship through the minefield, letting other ships safely follow the same path. An early example of this was Farragut's actions at Mobile Bay during the American Civil War. However, as mine warfare became more developed this method became uneconomical. This method was revived by the German "Kriegsmarine" during WWII. Left with a surfeit of idle ships due to the Allied blockade, the "Kriegsmarine" introduced a ship known as "Sperrbrecher" ("block breaker"). Typically an old cargo ship, loaded with cargo that made her less vulnerable to sinking (wood for example), the "Sperrbrecher" was run ahead of the ship to be protected, detonating any mines that might be in their path. The use of "Sperrbrecher" obviated the need to continuous and painstaking sweeping, but the cost was high. Over half the 100 or so ships used as "Sperrbrecher" were sunk during the war. Alternatively, a shallow draught vessel can be steamed through the minefield at high speed to generate a pressure wave sufficient to trigger mines, with the minesweeper moving fast enough to be sufficiently clear of the pressure wave so that triggered mines do not destroy the ship itself. These techniques are the only publicly known to be employed way to sweep pressure mines. The technique can be simply countered by use of a ship-counter, set to allow a certain number of passes before the mine is actually triggered. Modern doctrine calls for ground mines to be hunted rather than swept. A new system is being introduced for sweeping pressure mines, however counters are going to remain a problem. An updated form of this method is the use of small unmanned ROVs (such as the "Seehund" drone) that simulate the acoustic and magnetic signatures of larger ships and are built to survive exploding mines. Repeated sweeps would be required in case one or more of the mines had its "ship counter" facility enabled i.e. were programmed to ignore the first 2, 3, or even 6 target activations. The United States Navy MK56 ASW mine (the oldest still in use by the United States) was developed in 1966. More advanced mines include the MK60 CAPTOR (short for "encapsulated torpedo"), the MK62 and MK63 Quickstrike and the MK67 SLMM (Submarine Launched Mobile Mine). Today, most U.S. naval mines are delivered by aircraft. MK67 SLMM Submarine Launched Mobile Mine The SLMM was developed by the United States as a submarine deployed mine for use in areas inaccessible for other mine deployment techniques or for covert mining of hostile environments. The SLMM is a shallow-water mine and is basically a modified Mark 37 torpedo. General characteristics MK65 Quickstrike The Quickstrike is a family of shallow-water aircraft-laid mines used by the United States, primarily against surface craft. The MK65 is a 2,000-lb (900 kg) dedicated, purpose-built mine. However, other Quickstrike versions (MK62, MK63, and MK64) are converted general-purpose bombs. These latter three mines are actually a single type of electronic fuze fitted to Mk82, Mk83 and Mk84 air-dropped bombs. Because this latter type of Quickstrike fuze only takes up a small amount of storage space compared to a dedicated sea mine, the air-dropped bomb casings have dual purpose i.e. can be fitted with conventional contact fuzes and dropped on land targets, or have a Quickstrike fuze fitted which converts them into sea mines. General characteristics MK56 General characteristics According to a statement made to the UK Parliament in 2002: ...the Royal Navy does not have any mine stocks and has not had since 1992. Notwithstanding this, the United Kingdom retains the capability to lay mines and continues research into mine exploitation. Practice mines, used for exercises, continue to be laid in order to retain the necessary skills. However, a British company (BAE Systems) does manufacture the Stonefish influence mine for export to friendly countries such as Australia, which has both war stock and training versions of Stonefish, in addition to stocks of smaller Italian MN103 Manta mines. The computerised fuze on a Stonefish mine contains acoustic, magnetic and water pressure displacement target detection sensors. Stonefish can be deployed by fixed-wing aircraft, helicopters, surface vessels and submarines. An optional kit is available to allow Stonefish to be air-dropped, comprising an aerodynamic tail-fin section and parachute pack to retard the weapon's descent. The operating depth of Stonefish ranges between 30 and 200 metres. The mine weighs 990 kilograms and contains a 600 kilogram aluminised PBX explosive warhead. Mine warfare remains the most cost-effective of asymmetrical naval warfare. Mines are relatively cheap and being small allows them to be easily deployed. Indeed, with some kinds of mines, trucks and rafts will suffice. At present there are more than 300 different mines available. Some 50 countries currently have mining ability. The number of naval mine producing countries has increased by 75% since 1988. It is also noted that these mines are of an increasing sophistication while even the older type mines present a significant problem. It has been noted that mine warfare may become an issue with terrorist organizations. Mining busy shipping straits and mining shipping harbors remain some of the most serious threats.
https://en.wikipedia.org/wiki?curid=22102
North Melbourne Football Club The North Melbourne Football Club, nicknamed the Kangaroos or less formally the Roos, the Kangas or North, is an Australian rules football club that competes in the Australian Football League (AFL), the sport's premier competition. Founded in the suburb of North Melbourne in 1869, it is based at its traditional home ground, Arden Street Oval, and currently plays its home matches at both the nearby Docklands Stadium and Blundstone Arena in Hobart, Tasmania. The club's mascot is a grey kangaroo donning the club uniform, and its use dates from the mid-20th century. The club is also unofficially known as "The Shinboners", a term which dates back to its 19th-century abattoir-worker origins. The club's motto is "Victoria amat curam", Latin for "Victory Demands Dedication". Aside from their representation in the AFL, the Kangaroos also field teams in the following competitions; AFL Women's, VFL Women's and Victorian Football League. North Melbourne Football Club was founded in North Melbourne in 1869 by local cricketers desiring to keep fit over the winter months. One thought is that the club was connected to the St Mary's Church of England Cricket Club, now the St Mary's Anglican Church North Melbourne, whose colours – blue and white – are reflected in the club's colours today. The association between the St Mary's Church of England Cricket Club and the establishment of the North Melbourne Football Club is believed to have been an informal gathering to play some competitive sport. Information on the club's first ever match is limited, but it is known that it took place in Royal Park, which also served as the club's home ground until 1882. The ball used in the match was purchased by a local resident called Tom Jacks, who sold some roofing iron to pay for it. James Henry Gardiner is considered the founder of the club. He continued an active role with North Melbourne until his death in 1921. Regular premiership matches of Australian Football commenced in Victoria in 1870. Although North Melbourne was a part of this, it was classed as a "junior club". "The Australasian" noted them as being "one of the best of many junior clubs". The club continued to develop, graduating to senior ranks in 1874, finishing 4th. Along with the promotion, the club adopted its first uniform of blue and white horizontal stripes. In 1876, North Melbourne disbanded, and many of its player and members joined Albert-park, giving the club such a strong North Melbourne character that many described it as "Albert-park "cum" North Melbourne". In 1877, the club was re-established as a stand-alone club under the new name of "Hotham". Football took a giant step forward in 1877, with the formation of Victoria's first colonial football league, the VFA. Hotham were prime movers in establishing this league and were afforded a place in light of their previous contributions to Australian Football. The 1880s marked the emergence of the modern identity we now associate with North today. In 1882, the club amalgamated with the Hotham Cricket Club and moved into the North Melbourne Recreation Reserve (Arden St Oval), which remains the home of the club today. The joint venture was aimed at affecting improvements at the Hotham Cricket Ground, which was the name of the Reserve at the time. Four years later the club adopted the traditional uniform of blue and white vertical stripes at the insistence of the VFA, who wanted a visible contrast between Geelong's and Hotham's uniforms. The third significant development occurred in 1888 with the club returning to its original name of the North Melbourne Football Club. This followed the name of the local area reverting from Hotham to North Melbourne. The 1880s saw the club develop a penchant for inter-colonial travel with trips to Tasmania (1881/1887) and South Australia (1889). Hotham also found itself well represented at the first ever inter-colonial representative game in 1879 with four players from the club gaining selection for Victoria. The VFA grew to 13 senior clubs in the 1890s. Led by Geelong and Essendon, the largest clubs of the VFA formed their own break away league, the Victorian Football League (VFL), in 1896. Despite finishing 6th in 1896, North Melbourne was not invited to the breakaway competition. The main reasons for being excluded were: North continued on in the depleted VFA, emerging as a powerhouse, finishing 2nd in 1897, 1898 and 1899. In 1903, after 34 years of competing, the club won its first premiership, defeating Richmond in the final. The club became back to back premiers in 1904 after Richmond forfeited the grand final due to the appointment of an umpire whose performance when the two teams met earlier in the year was severely criticised by Richmond players and officials. North merged with fellow VFA football club West Melbourne in 1907, which at the time had lost its home ground. The joint venture saw a chance of promotion, and the club applied for admission to the more prestigious VFL in 1908, but Richmond and University were admitted instead. North was kicked out of the VFA during the 1907/08 offseason as a result of applying to join the VFL, before the local community reestablished the North Melbourne Football Club under a new committee, successfully enabling the club to play in the VFA in the 1908 season. The reformation of the Club necessitated a massive clean out of the team, leaving only two players remaining from the previous season. The 1910 season was marked by one of the most sensational transfers in Victorian football history, when Andy Curran masterminded the clearance of Carlton's famed "Big Four" of 'Mallee' Johnson, Fred Jinks, Charlie Hammond and Frank 'Silver' Caine to North Melbourne. These signings secured the Northerners' third premiership in 1910. The 1912 finals series was one of the most amazing ever, with the semi-final having to be played three times, after North and Brunswick drew twice. North was eventually victorious and moved on to the final, but lost the game by a mere four points with the last kick of the day. The next few years were punctuated by "The Invincibles". In the Northerners' most illustrious period ever, the club went undefeated from 1914 to 1919, collecting premierships in 1914, 1915 and 1918 – the league was in recess in 1916 and 1917 due to World War I. As well as this, the club won the championship in both 1915 and 1918 for finishing on top of the ladder, and accounted for VFL side St Kilda comfortably. During this period the club won 58 consecutive matches including 49 successive premiership matches, a record that has remained unmatched in Association or League history since. Despite being rejected from the VFL in both 1896 and 1907, North persisted in trying to gain admission into the League. On 30 June 1921, North told its players it would disband and try to gain entry to the VFL by the 'back-door'. Essendon League Football Club had lost its playing ground at East Melbourne and had decided to acquire the North Melbourne Recreation Reserve as a new playing ground. North accepted their proposal in the idea that the clubs would amalgamate. All of North's players were urged to join the Essendon League Club to help facilitate the amalgamation. The amalgamation was foiled when some members of the VFA launched a successful legal challenge. As a result, the Essendon League Club moved instead to the Essendon Oval, replacing the ground's original occupants, Essendon Association. North was now without a playing team and the Essendon Association Club was now without a ground, so as a matter of convenience the two clubs amalgamated so they could compete in the 1922 season. As it had after the merger with West Melbourne, North once again managed to avert its destruction. After three attempts, 29 years of waiting and numerous other applications to enter the VFL, finally North was rewarded for its persistence with admittance to the League in 1925, along with Footscray and Hawthorn. Even then, the opportunity was almost lost as the League delegates debated into the early hours of the morning on which clubs should be invited to join the intake. It was only after much deliberation that North Melbourne's name was eventually substituted for Prahran's making North "the lucky side" of the invitees that included Footscray and Hawthorn. North Melbourne was forced to change its uniform to avoid a clash when it joined the VFL. North Melbourne were cellar dwellers for its first twenty-five years of VFL membership and struggled to win matches in the superior VFL competition, with the only bright note being Sel Murray winning the VFL Leading Goalkicker Medal in 1941 with 88 goals. By the late 1940s, North Melbourne had developed a strong list and significant supporter base. In 1949 North secured the VFL Minor Premiership, finishing top of the ladder at the end of the home-and-away season with 14 wins and 5 losses. They failed to make the Grand Final that year (eventually won by Essendon), but in 1950 they did reach the final, defeated by a more efficient Essendon. It was in this year that the club adopted the "Kangaroos" mascot. In February 1965, North Melbourne moved its playing and training base from the Arden Street Oval to Coburg Oval, signing a seven-year lease with the City of Coburg after initially negotiating long-term leases for up to 40 years. The club came to an arrangement to merge with the VFA's Coburg Football Club, whom it was displacing from the ground; fourteen Coburg committeemen joined the North Melbourne committee, but the merger was never completed after Coburg established a rival committee which remained loyal to the VFA. The lease at Coburg lasted only eight months; the Coburg council was hesitant to build a new grandstand without the security of a long-term lease, and neither party made the returns they expected, so it was terminated by mutual agreement in September 1965 and North Melbourne returned to the Arden Street Oval. Onfield, the 1950s and 1960s were lean years for North Melbourne, though the club did secure two consecutive Night Premierships in 1965 and 1966. Allen Aylett was a brilliant player in the late 1950s and early 1960s (and captain between 1961 and 1964), as was Noel Teasdale, who lost the Brownlow Medal on a countback in 1965 (he was later awarded a retrospective medal when the counting system was amended). In the late 1960s, under the leadership of Allen Aylett, North Melbourne began its climb to supremacy. As part of a major recruitment drive North secured the services of several big-name stars, including Barry Davis from Essendon, Doug Wade from Geelong, John Rantall from South Melbourne, and Barry Cable from Perth. In a major coup, the great Ron Barassi was appointed coach in 1973. Barrassi reversed the club's playing fortunes, taking a struggling team that was once regarded as the traditional cellar dwellers of the competition through to a golden era of success that transformed North Melbourne into one of the powerhouses of the VFL. Barassi took North to a Grand Final (losing to Richmond by 41 points) in 1974 and brought success in his 1975 and 1977 seasons. North made five consecutive Grand Finals from 1974–1978) and defeated Norwood in the 1975 national championship and thus declared Champions of Australia. In 1973 and 1974, North's wingman Keith Greig (recruited from Brunswick Football Club, Victoria) won consecutive Brownlow Medals; forward Malcolm Blight (recruited from Woodville Football Club, South Australia) then won the award in 1978. Doug Wade (recruited from Geelong Football Club, Geelong) won the Coleman medal in 1974 with his 103 goals for the season. Barassi remained team coach until 1980, but only a Night Premiership in that year resulted in him leaving Arden Street. North then entered another period of decline, though Malcolm Blight kicked 103 goals to take out the Coleman medal in 1982, and another Brownlow win came through the talented Ross Glendinning in 1983. In that year, North Melbourne won a third Minor Premiership with 16 wins and 6 losses for the season, but they failed to make the Grand Final. The capable coaching of John Kennedy aside, the 1980s and early 1990s were lean years for the Kangaroos. However, the rebuilding of the club was taking place. The Krakouer brothers (Jim and Phil) brought a spark into the side and lifted many hopes for North supporters and the excitement to the general football public. The innovative idea of night games was instigated by the club and meeting the challenges, the club survived. One major highlight was the recruitment of forward John Longmire in 1989, who topped the club goalkicking over five consecutive seasons (1990–1994) and won the Coleman medal in 1990 with 98 goals. At the beginning of the 1993 season, in a dramatic and controversial move, the board of the club sacked coach and long-time player Wayne Schimmelbusch, and appointed Denis Pagan in his place. Results were immediate, as North reached the finals for the first time in nearly a decade. Pagan was instrumental in appointing young centre half-forward Wayne Carey as the club's youngest-ever captain. Carey had been recruited at the same time as Longmire, but had taken longer to develop as a player. Over the next nine seasons, Carey came to be regarded as the standout player in the league, and was known as 'the King'. North Melbourne became a powerhouse through the 1990s under Pagan and Carey, and finished in the top four from 1994 until 2000. After being eliminated in the preliminary finals in 1994 and 1995, North went on to defeat the Sydney Swans in the 1996 Grand Final to take out the club's third premiership, and the gold centenary AFL cup; Glenn Archer won the Norm Smith Medal. The club was again eliminated in the preliminary final in 1997. In 1998, as the club won both the pre-season Ansett Cup and topped the ladder with 16 wins and 6 losses, but went on to lose the 1998 Grand Final to Adelaide, not helped by an inaccurate goalkicking performance of 8.22 (70) to Adelaide's 15.15 (105). In 1999, the Kangaroos finished in second position on the ladder, and went on to defeat Carlton in the Grand Final, winning the club's fourth VFL/AFL premiership; former Sydney midfielder Shannon Grant taking out the Norm Smith Medal. The club was eliminated in the preliminary finals in 2000 against Melbourne. In 1996, the club was in advanced talks with the Fitzroy Football Club to create the North Fitzroy Kangaroos Football Club, which was in a terminal financial condition, to a merger between the two clubs; however, Fitzroy ultimately merged with the Brisbane Bears instead. Seeking new markets and greater financial security in an increasingly corporatized AFL environment, the title "North Melbourne" was officially dropped from the logo in 1999, from which time the team played only as the "Kangaroos". During the successful 1999 season, North Melbourne played home games in Sydney with a view of becoming a second team in New South Wales; however, the experiment was not successful, with crowds averaging only 12,000. The 21st century did not begin well for North Melbourne. Its decade-long onfield potency was in decline, questions were raised about its financial position and long-term sustainability. Furthermore, three of the people most important to the club's success in the 1990s left the club under acrimonious circumstances: CEO Greg Miller left the club, captain Wayne Carey left prior to the 2002 season following an extramarital affair with the wife of teammate and vice captain Anthony Stevens, coach Denis Pagan was lured to Carlton at the end of 2002. Pagan was replaced by 1996 premiership player Dean Laidley, who had previously been an Assistant Coach at Collingwood from 1999 until the end of season 2002. On a post-season holiday, several players were caught in the 2002 Bali bombing terrorist attack, notably defender Jason McCartney, who suffered second-degree burns to over 50% of his body while carrying others to safety and nearly died during surgery after being flown back to Melbourne. In what is regarded as one of the most inspirational stories of Australian rules football and Australian sport in general, McCartney successfully returned to action on 6 June 2003 against Richmond at Docklands Stadium. Playing at full forward, he took a mark in the final quarter, scored a goal from the resulting set shot and set up Leigh Harding's winning goal with two minutes remaining. McCartney retired immediately after the game, citing that his recovery had left him spent, and he was chaired from the ground. McCartney wore the numbers "88" and "202" on the front of his long-sleeved for the match, signifying the Australian and total number of victims of the Bali bombings, while many in the crowd bore signs reading "Bali 88/202". Onfield, the club reached the elimination finals in 2002 and 2005, but otherwise failed to reach the finals from 2001 until 2006. After two seasons of finals North Melbourne dropped to 13th in 2009, and coach Dean Laidley announced his resignation with Darren Crocker acting as caretaker coach for the rest of the season, to eventually be replaced by ex-Brisbane Lions premiership player and Collingwood assistant coach Brad Scott. A$15 million redevelopment of the Arden Street, which had started in 2006, was completed in 2009, giving the club top-class training facilities. North Melbourne struggled in its first two years under Brad Scott, finishing 9th in both 2010 and 2011. In 2012, the club returned to the finals for the first time since 2008, finishing the season in 8th place, but would go down to the West Coast Eagles by 96 points in an elimination final. In 2012, the club began a three-year deal to play two games each year at Blundstone Arena in Hobart, Tasmania. The club finished 10th in 2013 in a season full of close losses. Nick Dal Santo signed with the club at the end of the 2013 season as a restricted free agent. In 2014, North Melbourne finished 6th at the end of the home and away season and reached 40,000 members for the first time in the club's history. In September, North Melbourne went on to defeat Essendon by 12 points in the 2nd Elimination Final, only taking the lead in the last quarter. The following week, North Melbourne beat Geelong in the 2nd Semi-final by 6 points advancing them through to their first preliminary final since 2007. They lost to Sydney by 71 points. In 2015 the club made history by becoming the first team to qualify for a preliminary final from 8th spot, losing to the West Coast Eagles by 25 points. In 2016, North Melbourne won its first nine matches, which is the club's best start to a season in its VFL/AFL history. On 27 July 2016, the club announced it had surpassed 45,000 members for the first time in the club's history. In 2016, the Kangaroos fielded what was the oldest team in AFL history. Unfortunately after the midpoint of the season they fell away and struggled against some of the worst teams in competition. In the mid season of 2019 Brad Scott made the decision to leave NMFC after a great 10 years at the club taking them to the finals on multiple occasions. Rhyce Shaw took over as caretaker coach in the interim and was later awarded the position as head coach for the following season. The club was widely known as the "Shinboners" for much of its early history. The origins of the nickname are unknown but it may have had something to do with the club's reputation for targeting the shinbones of opposition players, or to do with local butchers who showed their support for North by dressing up beef leg-bones in the club colours. By 1926, the club was known as the "Blue Birds", but this nickname did not last. It was Phonse Tobin, North president from 1953 to 1956, who oversaw the club adopting the kangaroo emblem in 1954; Tobin found the image of a shinbone unsavoury and wanted the club to have a mascot it could show with pride. In selecting a new name, he wanted something characteristically Australian and was inspired by a large kangaroo he saw on display outside a city store. The official name of the club is North Melbourne, but the club has gone under several other aliases over the years. The club was founded as the "North Melbourne Football Club", but changed to "North Melbourne cum Albert Park" after merging with Albert Park in 1876. Following the reformation of the club in 1877, it was known as the "Hotham Football Club" but later took the name "North Melbourne" again in 1888. In 1998 the club proposed changing its name to the "Northern Kangaroos", but it was rejected by the AFL. From 1999 to 2007, the club traded without much success as "The Kangaroos" in a bid to increase its appeal nationally; this decision was reversed at the end of 2007 and the club has again reverted to the name "North Melbourne". "Join in the Chorus" is the official anthem of the North Melbourne Football Club. It is sung to the tune of a Scottish folk song from around 1911, "A Wee Deoch an Doris". The song is generally sung, in accordance to common football tradition, after a victory. It is also played before every match. "Join in the Chorus" is believed to be the oldest club anthem of any AFL club and has been associated with North from its early VFA days. The preamble of the song originates from a score of a theatre musical called "Australia: Heart to Heart and Hand to Hand", written by Toso Taylor in the 1890s in pre-federation Australia. The second verse is unknown in origin and was presumably added later by members of the club when the song was chosen. The chorus was appropriated from a song written and performed by Scottish musician Harry Lauder. The recording currently used by the club was performed by the Fable Singers in April 1972 and only includes the choruses. The song has a strong Victorian heritage and has been traditionally sung by the Victorian State Football and Victorian Cricket teams respectively. The lyrics have occasionally been changed, including updating the year in the song (e.g. "North Melbourne will be premiers in 1993"), or to remove the words "North Melbourne" during the period when the club was competing only as the Kangaroos. For the 2015 premiership season, You Am I's lead singer, Tim Rogers, a North Melbourne supporter, announced that he would assist in an updated version of the song including the two verses. This version is only played at North home games as the team runs onto the ground. The term "Shinboner spirit" is often used to refer to camaraderie and determination of players or members of the North Melbourne Football Club. The term persists to the modern day, despite North Melbourne having switched its official nickname from the Shinboners to the Kangaroos in the 1950s. Because it relates to the club's original nickname, Shinboner spirit is often associated with the complete history of the club. In 2005, to celebrate the club's 80th anniversary of senior competition in the VFL and the 30th anniversary of its first VFL premiership, the Kangaroos held a "Shinboner Spirit" gala event attended by almost the entire surviving players. In the awards ceremony, the key Shinboners of the past 80 years were acknowledged and Glenn Archer was named the "Shinboner of the Century". The North Melbourne Football Club has a long history of wearing various designs in the colours of royal blue and white. Most of the club's earliest jumpers were long-sleeved and not the sleeveless design common today. In their early years the club sported a hooped design when they took to the field. This changed at the behest of the VFA in 1884 who insisted that Hotham change their jumpers to vertical stripes to provide a visible contrast between Hotham and Geelong. After 1884 the vertical top was worn more often, usually in the lace-up design in the gallery below. After the merger with West Melbourne, North used a composite jumper that incorporated West Melbourne's red sash for the 1908 season. The merger was in reality, a takeover. The red sash was a token gesture and was removed the following season. In the early 1920s North experimented with an NMFC monogram design, following League clubs like Carlton and South Melbourne. Upon promotion to the VFL in 1925, North Melbourne was forced to abandon its royal blue and white striped jumper as it was deemed the jumper design clashed with other clubs. During this period a jumper with a V design was used for several years, before the club returned to using its striped jumper combination of royal blue and white which has been used continuously since 1932. North Melbourne's guernsey since entering the VFL in 1925 consists of white and royal blue vertical stripes. Up until 2016, North Melbourne's home jumper was predominantly white, but that has recently become the away design and a more predominantly blue design has been made the home guernsey. Changes in the North Melbourne uniform through the years: VFA: VFL/AFL: North Melbourne has experienced 7 logo changes since its introduction, with 5 of them featuring a bounding kangaroo behind a shield of blue and white stripes. In 2016, North Melbourne introduced a new logo that featured a much fiercer looking kangaroo, with its head only, sitting on top of the words 'North Melbourne' inside a shield. The change was welcomed as the previous logo (2007-2016) didn't seem to represent what they stood for or the direction they were heading. The new kangaroo looks slightly to the right, indicating that it is looking into the future. Arden Street Oval was home to the Kangaroos between 1882 and 1985. The oval is currently owned by the City of Melbourne and leased by the North Melbourne Football Club for social, administration and training facilities. The grandstands were removed because VFL/AFL matches are no longer played there. `The North Melbourne Football Club is a non-profit organisation limited by guarantee. Members of the club serve as the guarantees of capital and have full voting rights at AGMs to elect directors to the club's board. The club's board of directors has nine members, with each director serving a three-year term before their position is put up for re-election at an AGM. Only one-third of the board is contested at each AGM due to the rolling structure of the terms of the directors. This structure safeguards the entire board from being ousted at a single AGM and has made North Melbourne immune to a lot of the in-house fighting witnessed at other AFL football clubs. The board governs the club as well as selecting a chairman to head the club through a majority vote of directors. North Melbourne is unique in its structure, because from 1986 to 2006 the club was privately owned and limited by shares. The club was floated in 1986 through a membership vote led by then chairman Bob Ansett. At the meeting, members were encouraged to buy into the club by purchasing shares. The float ended up raising over $3 million and helped to keep the club solvent through the next decade. In 1991, the John Elliott-led Carlton Football Club attempted a hostile take over North Melbourne by purchasing a large parcel of shares formerly owned by Bob Ansett. The Blues acquired 20 per cent of the capital but that stake was eventually bought back by John Magowan, the former head of Merrill Lynch Australia, in 2001. The resulting melodrama saw the formation of B-Class shareholders who had the effective power of veto over any attempt to merge or relocate the club. Further takeover attempts were made in the first decade of the 21st century by the Southport Sharks. Then chairman Allan Aylett knocked back a proposal from the Sharks that would have seen them gain a majority stake in the club in exchange for an injection of capital. In early 2006, another proposal from Sharks to underwrite the Kangaroos' games on the Gold Coast, in exchange for a slice of the shareholder structure at the club was knocked back after AFL intervention. Due to an Australian Taxation Office ruling in 2006, the club proposed a shareholder restructure that would have seen the B Class shareholders power reduced significantly and some voting rights returned to members. This was done to avoid extraordinary taxes being placed on the club, but the move was blocked in December by Bob Ansett and his proxies who feared that the restructure would make the club vulnerable to further takeover bids. On 28 February 2007, another meeting was called to resolve the shareholder issue. A motion was passed that would return see some voting rights return to members and stop any future tax increments. In April 2007 it was revealed the AFL was attempting to buy out the shareholders of the club in a bid to gain full ownership, and force a relocation of the club to the Gold Coast. During October 2007, a group called We Are North Melbourne emerged and launched a public campaign, calling for ordinary members to be given the final say on the relocation issue. While the group became synonymous with the push to keep the club in Melbourne, its first priority was to see the club's shareholder structure wound-up and control returned to ordinary members. North Melbourne reverted to public company in November 2008. A moratorium was passed at an extraordinary general meeting that will allow James Brayshaw's board to serve unopposed until 2010, so as to allow his ticket the maximum time to enact their policies to make the North Melbourne Football Club financially viable. On 20 November 2016, former Aussie Rules footballer and Football Federation Australia chairman Ben Buckley replaced James Brayshaw as the new chairman of the club. In Round 1,1985, North Melbourne pioneered the concept of playing football on Friday nights. Since then, North Melbourne has played the most Friday night games of any AFL club. Friday night matches later became the most lucrative timeslot for televised games, and North Melbourne's relatively low supporter base resulted in fewer Friday night matches. Between 2010 and 2014, North Melbourne had hosted an annual Friday night match against Carlton in recognition of its pioneering role in the concept. After years of campaigning to play on Good Friday, the AFL announced on 25 October 2016 that North Melbourne will play the Western Bulldogs on Good Friday 2017. Good Friday in Australia is also considered as a day where people raise money for the Royal Childrens Hospital, and North Melbourne announced on 7 March 2017 that $5 from each ticket will go to the charity. Also, North Melbourne will play Essendon on Good Friday 19 April 2019. North Melbourne has a strong history of supporting Aboriginal footballers and fostering Aboriginal talent in the VFL and AFL. The first indigenous footballer to play for the club was Percy Johnson in the 1950s, and was followed by other fan favorites like Bertie Johnson, Barry Cable and the Krakouer brothers in the following decades. The following is a list of Indigenous footballers to have played senior football at the club: †: Aboriginality uncertain The following footballers who were killed in action during the World Wars played senior football for North Melbourne. In 2014, North Melbourne put forward a proposal in which the away team's club songs would no longer be played at their home matches, however, this was quickly rejected by the Australian Football League. Such a move would have placed the AFL in line with other leagues such as the National Rugby League and overseas competitions in creating a true home ground advantage for the home side. Essendon – North and Essendon have a chequered history that dates back to the late 19th century; firstly in 1896, Essendon had North excluded from the VFL because both clubs drew supporters from the same area. North supporters have long been bitter with Essendon for excluding them from the VFL, and have blamed that for their small supporter base in comparison to Essendon's. North's first VFL Grand Final was against Essendon in 1950. The rivalry was reignited in the 90s as both teams were constantly in premiership contention. In 1998, following comments by Essendon coach labelling Kangaroos executives Greg Miller and Mark Dawson "marshmallows", a reference to their softness, North supporters threw marshmallows at Sheedy after the opening Qualifying Final. In 2014 North Melbourne versed Essendon in Elimination Final 2 Essendon was leading by 30 odd points then North came back to win by 12 points Hawthorn – North and Hawthorn have a fierce rivalry that dates back to the 1970s when they played off against each other in three Grand Finals in the space of four years. From 1974 to 1978 the two clubs played against each other in ten finals, and took each other on for the Australian Championship in Adelaide in 1976. During the 1980s Hawthorn dominated North, and during the 90s the results were reversed with North dominating Hawthorn. North Melbourne defeated Hawthorn in the 1975 Grand Final by 55 points. However, Hawthorn defeated North Melbourne in the 1976 Grand Final by 30 points and in the 1978 Grand Final by 18 points. The rivalry re-ignited in 2014 following a choking incident involving Brian Lake having North Melbourne forward Drew Petrie in a choking hold during a clash between the two sides at Docklands Stadium and reached fever pitch in 2015 following several fights including an all in during the first term of their round 5 clash. At a special function in August 2001, the North Melbourne Team of the Century was announced. There was no minimum number of games set for selection. Wayne Carey was named as captain and Denis Pagan as coach. The selection panel was Geoff Poulter (journalist), Father Gerard Dowling (club historian), Keith McKenzie (former coach), Lloyd Holyoak (former president), Max Ritchie (former player and chairman of selectors) and Greg Miller (chief executive). On 18 March 2005, the North Melbourne football club held a special gala dinner entitled the "North Story" to celebrate the 80th anniversary of North's admission to the VFL, and the 30th anniversary of the club's first VFL premiership. Over 3500 people attended the historic event held at the Royal Exhibition Building, including almost all surviving North Melbourne players. Glenn Archer was voted the Shinboner of the Century by his peers as the player who most represents the 'Shinboner Spirit'. The following players were voted 'Shinboners' of their era: To commemorate the 150th year of the founding of the North Melbourne Football Club a 150th Year Celebration was organised for the first weekend of August 2019 which commenced with a Friday Night blockbuster against arch rivals Hawthorn. Starting from 27 points down in the first quarter, the Roos fought back heroically against the Hawks to triumph as 22 point winners, and get the weekend celebrations underway. The following day the Roos VFL side took on Box Hill and similarly won in a heroic performance coming back from 31 points down at three quarter time to win by 2 points with just minutes to spare. To cap off the weekend, a 150th Year Celebration Dinner was held at the Melbourne Convention Centre where the 150 greatest ever North Melbourne players were announced with a top 10 greatest North Melbourne players announced on the night from the results of an expert panel. Top 10 Greatest North Players North Melbourne operated its own seconds/reserves team from 1925 until 1999, and will begin to do so again from the 2018 season. From 1919 to 1991 the VFL/AFL operated a reserves competition, and from 1992 to 1999 a "de facto" AFL reserves competition was run by the Victorian State Football League, and North Melbourne fielded its reserves team in both of these competitions while it was in the VFL/AFL, allowing players who were not selected for the senior team to play for North Melbourne in the lower grade. During that time, the North Melbourne reserves team won seven premierships (1947, 1957, 1967, 1978, 1979, 1995, 1996). Following the demise of the AFL reserves competition, North Melbourne's reserves team was dissolved, and over the following eighteen years the club entered reserves affiliations with a range of Victorian Football League clubs. Under the affiliations, reserves players for North Melbourne play VFL football with one of the affiliated clubs. The club had five different affiliation arrangements over that time: Starting in the 2018 VFL season, North Melbourne re-established its own reserves team, which will play in the VFL. It played its home games at Chirnside Park in Werribee until mid-2019, and then began playing at Arden Street Oval. In 2017, following the inaugural AFL Women's (AFLW) season, North Melbourne was among eight clubs that applied for licences to enter the competition from 2019 onwards. In September 2017, the club was announced as one of two clubs, along with , to receive a licence to join the competition in 2019. In April 2018, the club announced the signing of midfielder Emma Kearney, who had just won the AFL Women's best and fairest and a premiership and club best-and-fairest with the Bulldogs. North Melbourne began fielding a team in the second-tier VFL Women's league in 2020. AFL Women's ^ Denotes the ladder was split into two or more conferences. These numbers refer to the club's overall finishing position that season. VFL Women's Sources: Club historical data and VFLW stats
https://en.wikipedia.org/wiki?curid=22106
Treaty on the Non-Proliferation of Nuclear Weapons The Treaty on the Non-Proliferation of Nuclear Weapons, commonly known as the Non-Proliferation Treaty or NPT, is an international treaty whose objective is to prevent the spread of nuclear weapons and weapons technology, to promote cooperation in the peaceful uses of nuclear energy, and to further the goal of achieving nuclear disarmament and general and complete disarmament. Between 1965 and 1968, the treaty was negotiated by the Eighteen Nation Committee on Disarmament, a United Nations-sponsored organization based in Geneva, Switzerland. Opened for signature in 1968, the treaty entered into force in 1970. As required by the text, after twenty-five years, NPT Parties met in May 1995 and agreed to extend the treaty indefinitely. More countries are parties to the NPT than any other arms limitation and disarmament agreement, a testament to the treaty's significance. As of August 2016, 191 states have become parties to the treaty, though North Korea, which acceded in 1985 but never came into compliance, announced its withdrawal from the NPT in 2003, following detonation of nuclear devices in violation of core obligations. Four UN member states have never accepted the NPT, three of which possess or are thought to possess nuclear weapons: India, Israel, and Pakistan. In addition, South Sudan, founded in 2011, has not joined. The treaty defines nuclear-weapon states as those that have built and tested a nuclear explosive device before 1 January 1967; these are the United States, Russia, the United Kingdom, France, and China. Four other states are known or believed to possess nuclear weapons: India, Pakistan, and North Korea have openly tested and declared that they possess nuclear weapons, while Israel is deliberately ambiguous regarding its nuclear weapons status. The NPT is often seen to be based on a central bargain: the NPT non-nuclear-weapon states agree never to acquire nuclear weapons and the NPT nuclear-weapon states in exchange agree to share the benefits of peaceful nuclear technology and to pursue nuclear disarmament aimed at the ultimate elimination of their nuclear arsenals. The treaty is reviewed every five years in meetings called Review Conferences of the Parties to the Treaty of Non-Proliferation of Nuclear Weapons. Even though the treaty was originally conceived with a limited duration of 25 years, the signing parties decided, by consensus, to unconditionally extend the treaty indefinitely during the Review Conference in New York City on 11 May 1995, in the culmination of U.S. government efforts led by Ambassador Thomas Graham Jr. At the time the NPT was proposed, there were predictions of 25–30 nuclear weapon states within 20 years. Instead, over forty years later, five states are not parties to the NPT, and they include the only four additional states believed to possess nuclear weapons. Several additional measures have been adopted to strengthen the NPT and the broader nuclear nonproliferation regime and make it difficult for states to acquire the capability to produce nuclear weapons, including the export controls of the Nuclear Suppliers Group and the enhanced verification measures of the International Atomic Energy Agency (IAEA) Additional Protocol. Critics argue that the NPT cannot stop the proliferation of nuclear weapons or the motivation to acquire them. They express disappointment with the limited progress on nuclear disarmament, where the five authorized nuclear weapons states still have 22,000 warheads in their combined stockpile and have shown a reluctance to disarm further. Several high-ranking officials within the United Nations have said that they can do little to stop states using nuclear reactors to produce nuclear weapons. The NPT consists of a preamble and eleven articles. Although the concept of "pillars" is not expressed anywhere in the NPT, the treaty is nevertheless sometimes interpreted as a "three-pillar" system, with an implicit balance among them: These pillars are interrelated and mutually reinforcing. An effective nonproliferation regime whose members comply with their obligations provides an essential foundation for progress on disarmament and makes possible greater cooperation on the peaceful use of nuclear energy. With the right to access the benefits of peaceful nuclear technology comes the responsibility of nonproliferation. Progress on disarmament reinforces efforts to strengthen the nonproliferation regime and to enforce compliance with obligations, thereby also facilitating peaceful nuclear cooperation. The "pillars" concept has been questioned by some who believe that the NPT is, as its name suggests, principally about nonproliferation, and who worry that "three pillars" language misleadingly implies that the three elements have equivalent importance. Under Article I of the NPT, nuclear-weapon states pledge not to transfer nuclear weapons or other nuclear explosive devices to any recipient or in any way assist, encourage or induce any non-nuclear-weapon state in the manufacture or acquisition of a nuclear weapon. Under Article II of the NPT, non-nuclear-weapon states pledge not to acquire or exercise control over nuclear weapons or other nuclear explosive devices and not to seek or receive assistance in the manufacture of such devices. Under Article III of the Treaty, non-nuclear-weapon states pledge to accept IAEA safeguards to verify that their nuclear activities serve only peaceful purposes. Five states are recognized by NPT as nuclear weapon states (NWS): China (signed 1992), France (1992), the Soviet Union (1968; obligations and rights now assumed by the Russian Federation), the United Kingdom (1968), and the United States (1968), which also happen to be the five permanent members of the United Nations Security Council. These five NWS agree not to transfer "nuclear weapons or other nuclear explosive devices" and "not in any way to assist, encourage, or induce" a non-nuclear weapon state (NNWS) to acquire nuclear weapons (Article I). NNWS parties to the NPT agree not to "receive", "manufacture", or "acquire" nuclear weapons or to "seek or receive any assistance in the manufacture of nuclear weapons" (Article II). NNWS parties also agree to accept safeguards by the International Atomic Energy Agency (IAEA) to verify that they are not diverting nuclear energy from peaceful uses to nuclear weapons or other nuclear explosive devices (Article III). The five NWS parties have made undertakings not to use their nuclear weapons against a non-NWS party except in response to a nuclear attack, or a conventional attack in alliance with a Nuclear Weapons State. However, these undertakings have not been incorporated formally into the treaty, and the exact details have varied over time. The U.S. also had nuclear warheads targeted at North Korea, a non-NWS, from 1959 until 1991. The previous United Kingdom Secretary of State for Defence, Geoff Hoon, has also explicitly invoked the possibility of the use of the country's nuclear weapons in response to a non-conventional attack by "rogue states". In January 2006, President Jacques Chirac of France indicated that an incident of state-sponsored terrorism on France could trigger a small-scale nuclear retaliation aimed at destroying the "rogue state's" power centers. Under Article VI of the NPT, all Parties undertake to pursue good-faith negotiations on effective measures relating to cessation of the nuclear arms race, to nuclear disarmament, and to general and complete disarmament. Article VI of the NPT represents the only binding commitment in a multilateral treaty to the goal of disarmament by the nuclear-weapon states. The NPT's preamble contains language affirming the desire of treaty signatories to ease international tension and strengthen international trust so as to create someday the conditions for a halt to the production of nuclear weapons, and treaty on general and complete disarmament that liquidates, in particular, nuclear weapons and their delivery vehicles from national arsenals. The wording of the NPT's Article VI arguably imposes only a vague obligation on all NPT signatories to move in the general direction of nuclear and total disarmament, saying, "Each of the Parties to the Treaty undertakes to pursue negotiations in good faith on effective measures relating to cessation of the nuclear arms race at an early date and to nuclear disarmament, and on a treaty on general and complete disarmament." Under this interpretation, Article VI does not strictly require all signatories to actually conclude a disarmament treaty. Rather, it only requires them "to negotiate in good faith." On the other hand, some governments, especially non-nuclear-weapon states belonging to the Non-Aligned Movement, have interpreted Article VI's language as being anything but vague. In their view, Article VI constitutes a formal and specific obligation on the NPT-recognized nuclear-weapon states to disarm themselves of nuclear weapons, and argue that these states have failed to meet their obligation. The International Court of Justice (ICJ), in its advisory opinion on the Legality of the Threat or Use of Nuclear Weapons, issued 8 July 1996, unanimously interprets the text of Article VI as implying that There exists an obligation to pursue in good faith and bring to a conclusion negotiations leading to nuclear disarmament in all its aspects under strict and effective international control. The ICJ opinion notes that this obligation involves all NPT parties (not just the nuclear weapon states) and does not suggest a specific time frame for nuclear disarmament. Critics of the NPT-recognized nuclear-weapon states (the United States, Russia, China, France, and the United Kingdom) sometimes argue that what they view as the failure of the NPT-recognized nuclear weapon states to disarm themselves of nuclear weapons, especially in the post–Cold War era, has angered some non-nuclear-weapon NPT signatories of the NPT. Such failure, these critics add, provides justification for the non-nuclear-weapon signatories to quit the NPT and develop their own nuclear arsenals. Other observers have suggested that the linkage between proliferation and disarmament may also work the other way, i.e., that the failure to resolve proliferation threats in Iran and North Korea, for instance, will cripple the prospects for disarmament. No current nuclear weapons state, the argument goes, would seriously consider eliminating its last nuclear weapons without high confidence that other countries would not acquire them. Some observers have even suggested that the very progress of disarmament by the superpowers—which has led to the elimination of thousands of weapons and delivery systems—could eventually make the possession of nuclear weapons more attractive by increasing the perceived strategic value of a small arsenal. As one U.S. official and NPT expert warned in 2007, "logic suggests that as the number of nuclear weapons decreases, the 'marginal utility' of a nuclear weapon as an instrument of military power increases. At the extreme, which it is precisely disarmament's hope to create, the strategic utility of even one or two nuclear weapons would be huge." NPT Article IV acknowledges the right of all Parties to develop nuclear energy for peaceful purposes and to benefit from international cooperation in this area, in conformity with their nonproliferation obligations. Article IV also encourages such cooperation. This so-called third pillar provides for the transfer of nuclear technology and materials to NPT Parties for peaceful purposes in the development of civilian nuclear energy programs in those countries, subject to IAEA safeguards to demonstrate that their nuclear programs are not being used for the development of nuclear weapons. As the commercially popular light water reactor nuclear power station uses enriched uranium fuel, it follows that states must be able either to enrich uranium or purchase it on an international market. Mohamed ElBaradei, then Director General of the International Atomic Energy Agency, has called the spread of enrichment and reprocessing capabilities the "Achilles' heel" of the nuclear nonproliferation regime. As of 2007 13 states have an enrichment capability. During the 1960s and 1970s many states, almost 60, were supplied with research reactors fuelled by weapon grade highly enriched uranium (HEU) through the United States Atoms for Peace program and a similar Soviet Union program. In the 1980s a program to convert HEU research reactors to use low enriched fuel was started in the United States due to proliferation concerns. However 26 states possessed more that 1 kg of civilian HEU in 2015, and as of 2016 the stocks of HEU for civilian research were 60 tonnes, with 74 research reactors still using HEU. Because the availability of fissile material has long been considered the principal obstacle to, and "pacing element" for, a country's nuclear weapons development effort, it was declared a major emphasis of U.S. policy in 2004 to prevent the further spread of uranium enrichment and plutonium reprocessing (a.k.a. "ENR") technology. Countries possessing ENR capabilities, it is feared, have what is in effect the option of using this capability to produce fissile material for weapons use on demand, thus giving them what has been termed a "virtual" nuclear weapons program. The degree to which NPT members have a "right" to ENR technology notwithstanding its potentially grave proliferation implications, therefore, is at the cutting edge of policy and legal debates surrounding the meaning of Article IV and its relation to Articles I, II, and III of the treaty. Countries that have become Parties to the NPT as non-nuclear-weapon States have a strong record of not building nuclear weapons, although some tried and one eventually left the NPT and acquired nuclear weapons. Iraq was found by the IAEA to have violated its safeguards obligations and subject to punitive sanctions by the UN Security Council. North Korea never came into compliance with its NPT safeguards agreement and was cited repeatedly for these violations, and later withdrew from the NPT and tested multiple nuclear devices. Iran was found in non-compliance with its NPT safeguards obligations in an unusual non-consensus decision because it "failed in a number of instances over an extended period of time" to report aspects of its enrichment program. In 1991, Romania reported previously undeclared nuclear activities by the former regime and the IAEA reported this non-compliance to the Security Council for information only. Libya pursued a clandestine nuclear weapons program before abandoning it in December 2003. The IAEA reported Syria's safeguards non-compliance to the UN Security Council, which did not take action. In some regions, the fact that all neighbors are verifiably free of nuclear weapons reduces any pressure individual states might feel to build those weapons themselves, even if neighbors are known to have peaceful nuclear energy programs that might otherwise be suspicious. In this, the treaty works as designed. In 2004, Mohamed ElBaradei said that by some estimates thirty-five to forty states could have the knowledge to develop nuclear weapons. "Article I": Each nuclear-weapons state (NWS) undertakes not to transfer, to any recipient, nuclear weapons, or other nuclear explosive devices, and not to assist any non-nuclear weapon state to manufacture or acquire such weapons or devices. "Article II": Each non-NWS party undertakes not to receive, from any source, nuclear weapons, or other nuclear explosive devices; not to manufacture or acquire such weapons or devices; and not to receive any assistance in their manufacture. "Article III": Each non-NWS party undertakes to conclude an agreement with the IAEA for the application of its safeguards to all nuclear material in all of the state's peaceful nuclear activities and to prevent diversion of such material to nuclear weapons or other nuclear explosive devices. "Article IV": 1. Nothing in this Treaty shall be interpreted as affecting the inalienable right of all the Parties to the Treaty to develop research, production and use of nuclear energy for peaceful purposes without discrimination and in conformity with Articles I and II of this Treaty. 2. All the Parties to the Treaty undertake to facilitate, and have the right to participate in, the fullest possible exchange of equipment, materials and scientific and technological information for the peaceful uses of nuclear energy. Parties to the Treaty in a position to do so shall also co-operate in contributing alone or together with other States or international organizations to the further development of the applications of nuclear energy for peaceful purposes, especially in the territories of non-nuclear-weapon States Party to the Treaty, with due consideration for the needs of the developing areas of the world. "Article VI": Each party "undertakes to pursue negotiations in good faith on effective measures relating to cessation of the nuclear arms race at an early date and to nuclear disarmament, and on a Treaty on general and complete disarmament under strict and effective international control". "Article IX": "For the purposes of this Treaty, a nuclear-weapon State is one which has manufactured and exploded a nuclear weapon or other nuclear explosive device prior to 1 January 1967." "Article X": Establishes the right to withdraw from the Treaty giving 3 months' notice. It also establishes the duration of the Treaty (25 years before 1995 Extension Initiative). The impetus behind the NPT was concern for the safety of a world with many nuclear weapon states. It was recognized that the Cold War deterrent relationship between just the United States and the Soviet Union was fragile. Having more nuclear-weapon states would reduce security for all, multiplying the risks of miscalculation, accidents, unauthorized use of weapons, escalation in tensions, and nuclear conflict. Moreover, since the use of nuclear weapons in Hiroshima and Nagasaki in 1945, it has been apparent that the development of nuclear capabilities by States could enable them to divert technology and materials for weapons purposes. Thus, the problem of preventing such diversions became a central issue in discussions on peaceful uses of nuclear energy. Initial efforts, which began in 1946, to create an international system enabling all States to have access to nuclear technology under appropriate safeguards, were terminated in 1949 without the achievement of this objective, due to serious political differences between the major Powers. By then, both the United States and the former Soviet Union had tested nuclear weapons, and were beginning to build their stockpiles. In December 1953, US President Dwight D. Eisenhower in his "Atoms for Peace" proposal, presented to the eighth session of the United Nations General Assembly, urged that an international organization be established to disseminate peaceful nuclear technology, while guarding against development of weapons capabilities in additional countries. His proposal resulted in 1957 in the establishment of the International Atomic Energy Agency (IAEA), which was charged with the dual responsibility for promotion and control of nuclear technology. IAEA technical activities began in 1958. An interim safeguards system for small nuclear reactors, put in place in 1961, was replaced in 1964 by a system covering larger installations and, over the following years, was expanded to include additional nuclear facilities. In recent years, efforts to strengthen the effectiveness and improve the efficiency of the IAEA safeguards system culminated in the approval of the Model Additional Protocol by the IAEA Board of Governors in May 1997. Within the framework of the United Nations, the principle of nuclear non-proliferation was addressed in negotiations as early as 1957. The NPT process was launched by Frank Aiken, Irish Minister for External Affairs, in 1958. The NPT gained significant momentum in the early 1960s. The structure of a treaty to uphold nuclear non-proliferation as a norm of international behaviour had become clear by the mid-1960s, and by 1968 final agreement had been reached on a Treaty that would prevent the proliferation of nuclear weapons, enable cooperation for the peaceful use of nuclear energy, and further the goal of achieving nuclear disarmament. It was opened for signature in 1968, with Finland the first State to sign. Accession became nearly universal after the end of the Cold War and of South African apartheid. In 1992, The People's Republic of China and France acceded to the NPT, the last of the five nuclear powers recognized by the treaty to do so. The treaty provided, in article X, for a conference to be convened 25 years after its entry into force to decide whether the treaty should continue in force indefinitely, or be extended for an additional fixed period or periods. Accordingly, at the NPT Review and Extension Conference in May 1995, state parties to the treaty agreed-without a vote-on the treaty's indefinite extension, and decided that review conferences should continue to be held every five years. After Brazil acceded to the NPT in 1998, the only remaining non-nuclear-weapon state which had not signed was Cuba, which joined the NPT (and the Treaty of Tlatelolco NWFZ) in 2002. Several NPT states parties have given up nuclear weapons or nuclear weapons programs. South Africa undertook a nuclear weapons program, but has since renounced it and acceded to the treaty in 1991 after destroying its small nuclear arsenal; after this, the remaining African countries signed the treaty. The former Soviet Republics where nuclear weapons had been based, namely Ukraine, Belarus and Kazakhstan, transferred those weapons to Russia and joined the NPT by 1994 following the signature of the Budapest Memorandum on Security Assurances. Successor states from the breakups of Yugoslavia and Czechoslovakia also joined the treaty soon after their independence. Montenegro and East Timor were the last countries to accede to the treaty on their independence in 2006 and 2003; the only other country to accede in the 21st century was Cuba in 2002. The three Micronesian countries in Compact of Free Association with the USA joined the NPT in 1995, along with Vanuatu. Major South American countries Argentina, Chile, and Brazil joined in 1995 and 1998. Arabian Peninsula countries included Saudi Arabia and Bahrain in 1988, Qatar and Kuwait in 1989, UAE in 1995, and Oman in 1997. The European states of Monaco and Andorra joined in 1995–6. Also acceding in the 1990s were Myanmar in 1992 and Guyana in 1993. At the time the treaty was being negotiated, NATO had in place secret nuclear weapons sharing agreements whereby the United States provided nuclear weapons to be deployed by, and stored in, other NATO states. Some argue this is an act of proliferation violating Articles I and II of the treaty. A counter-argument is that the U.S. controlled the weapons in storage within the NATO states, and that no transfer of the weapons or control over them was intended "unless and until a decision were made to go to war, at which the treaty would no longer be controlling", so there is no breach of the NPT. These agreements were disclosed to a few of the states, including the Soviet Union, negotiating the treaty, but most of the states that signed the NPT in 1968 would not have known about these agreements and interpretations at that time. As of 2005, it is estimated that the United States still provides about 180 tactical B61 nuclear bombs for use by Belgium, Germany, Italy, the Netherlands and Turkey under these NATO agreements. Many states, and the Non-Aligned Movement, now argue this violates Articles I and II of the treaty, and are applying diplomatic pressure to terminate these agreements. They point out that the pilots and other staff of the "non-nuclear" NATO states practice handling and delivering the U.S. nuclear bombs, and non-U.S. warplanes have been adapted to deliver U.S. nuclear bombs which must have involved the transfer of some technical nuclear weapons information. NATO believes its "nuclear forces continue to play an essential role in war prevention, but their role is now more fundamentally political". U.S. nuclear sharing policies were originally designed to help prevent the proliferation of nuclear weapons—not least by persuading the then West Germany not to develop an independent nuclear capability by assuring it that West Germany would be able, in the event of war with the Warsaw Pact, to wield (U.S.) nuclear weapons in self-defense. (Until that point of all-out war, however, the weapons themselves would remain in U.S. hands.) The point was to limit the spread of countries having their own nuclear weapons programs, helping ensure that NATO allies would not choose to go down the proliferation route. (West Germany was discussed in U.S. intelligence estimates for a number of years as being a country with the potential to develop nuclear weapons capabilities of its own if officials in Bonn were not convinced that their defense against the Soviet Union and its allies could otherwise be met.) Four states—India, Israel, Pakistan, and South Sudan—have never signed the treaty. India and Pakistan have publicly disclosed their nuclear weapon programs, and Israel has a long-standing policy of deliberate ambiguity with regards to its nuclear program (see List of states with nuclear weapons). India has detonated nuclear devices, first in 1974 and again in 1998. India is estimated to have enough fissile material for more than 150 warheads. India was among the few countries to have a no first use policy, a pledge not to use nuclear weapons unless first attacked by an adversary using nuclear weapons, however India's former NSA Shivshankar Menon signaled "a significant shift from "no first use" to "no first use against non-nuclear weapon states"" in a speech on the occasion of Golden Jubilee celebrations of the National Defence College in New Delhi on 21 October 2010, a doctrine Menon said reflected India's "strategic culture, with its emphasis on minimal deterrence". India argues that the NPT creates a club of "nuclear haves" and a larger group of "nuclear have-nots" by restricting the legal possession of nuclear weapons to those states that tested them before 1967, but the treaty never explains on what ethical grounds such a distinction is valid. India's then External Affairs Minister Pranab Mukherjee said during a visit to Tokyo in 2007: "If India did not sign the NPT, it is not because of its lack of commitment for non-proliferation, but because we consider NPT as a flawed treaty and it did not recognize the need for universal, non-discriminatory verification and treatment." Although there have been unofficial discussions on creating a South Asian nuclear weapons free zone, including India and Pakistan, this is considered to be highly unlikely for the foreseeable future. In early March 2006, India and the United States finalized an agreement, in the face of criticism in both countries, to restart cooperation on civilian nuclear technology. Under the deal India has committed to classify 14 of its 22 nuclear power plants as being for civilian use and to place them under IAEA safeguards. Mohamed ElBaradei, then Director General of the IAEA, welcomed the deal by calling India "an important partner in the non-proliferation regime." In December 2006, United States Congress approved the United States-India Peaceful Atomic Energy Cooperation Act, endorsing a deal that was forged during Prime Minister Singh's visit to the United States in July 2005 and cemented during President Bush's visit to India earlier in 2006. The legislation allows for the transfer of civilian nuclear material to India. Despite its status outside the Nuclear Non-Proliferation Treaty, nuclear cooperation with India was permitted on the basis of its clean non-proliferation record, and India's need for energy fueled by its rapid industrialization and a billion-plus population. On 1 August 2008, the IAEA approved the India Safeguards Agreement and on 6 September 2008, India was granted the waiver at the Nuclear Suppliers Group (NSG) meeting held in Vienna, Austria. The consensus was arrived after overcoming misgivings expressed by Austria, Ireland and New Zealand and is an unprecedented step in giving exemption to a country, which has not signed the NPT and the Comprehensive Test Ban Treaty (CTBT). While India could commence nuclear trade with other willing countries. The U.S. Congress approved this agreement and President Bush signed it on 8 October 2008. When China announced expanded nuclear cooperation with Pakistan in 2010, proponents of arms control denounced both the deals, claiming that they weakened the NPT by facilitating nuclear programmes in states which are not parties to the NPT. , Australia, a top three uranium producer and home to world's largest known reserves, had continued its refusal to export Uranium to India despite diplomatic pressure from India. In November 2011 the Australian Prime Minister announced a desire to allow exports to India, a policy change which was authorized by her party's national conference in December. On 4 December 2011, Prime Minister Julia Gillard overturned Australia's long-standing ban on exporting uranium to India. She further said "We should take a decision in the national interest, a decision about strengthening our strategic partnership with India in this the Asian century," and said that any agreement to sell uranium to India would include strict safeguards to ensure it would only be used for civilian purposes, and not end up in nuclear weapons. On Sep 5, 2014; Australian Prime Minister Tony Abbott sealed a civil nuclear deal to sell uranium to India. "We signed a nuclear cooperation agreement because Australia trusts India to do the right thing in this area, as it has been doing in other areas," Abbott told reporters after he and Indian Prime Minister Narendra Modi signed a pact to sell uranium for peaceful power generation. In May 1998, following India's nuclear tests earlier that month, Pakistan conducted two sets of nuclear tests, the Chagai-I and Chagai-II. Although there is little confirmed information in public, as of 2015, Pakistan was estimated to have as many as 120 warheads. According to analyses of the Carnegie Endowment for International Peace and the Stimson Center, Pakistan has enough fissile material for 350 warheads. Pakistani officials argue that the NPT is discriminatory. When asked at a briefing in 2015 whether Islamabad would sign the NPT if Washington requested it, Foreign Secretary Aizaz Ahmad Chaudhry was quoted as responding "It is a discriminatory treaty. Pakistan has the right to defend itself, so Pakistan will not sign the NPT. Why should we?" Until 2010, Pakistan had always maintained the position that it would sign the NPT if India did so. In 2010, Pakistan abandoned this historic position and stated that it would join the NPT only as a recognized nuclear-weapon state. The NSG Guidelines currently rule out nuclear exports by all major suppliers to Pakistan, with very narrow exceptions, since it does not have full-scope IAEA safeguards (i.e. safeguards on all its nuclear activities). Pakistan has sought to reach an agreement similar to that with India, but these efforts have been rebuffed by the United States and other NSG members, on the grounds that Pakistan's track record as a nuclear proliferator makes it impossible for it to have any sort of nuclear deal in the near future. By 2010, China reportedly signed a civil nuclear agreement with Pakistan, using the justification that the deal was "peaceful." The British government criticized this, on the grounds that 'the time is not yet right for a civil nuclear deal with Pakistan'. China did not seek formal approval from the nuclear suppliers group, and claimed instead that its cooperation with Pakistan was "grandfathered" when China joined the NSG, a claim that was disputed by other NSG members. Pakistan applied for membership on 19 May 2016, supported by Turkey and China However, many NSG members opposed Pakistan's membership bid due to its track record, including the illicit procurement network of Pakistani scientist A.Q. Khan, which aided the nuclear programs of Iran, Libya and North Korea. Pakistani officials reiterated the request in August 2016. Israel has a long-standing policy of deliberate ambiguity with regards to its nuclear program (see List of countries with nuclear weapons). Israel has been developing nuclear technology at its Dimona site in the Negev since 1958, and some nonproliferation analysts estimate that Israel may have stockpiled between 100 and 200 warheads using reprocessed plutonium. The position on the NPT is explained in terms of "Israeli exceptionality", a term coined by Professor Gerald M. Steinberg, in reference to the perception that the country's small size, overall vulnerability, as well as the history of deep hostility and large-scale attacks by neighboring states, require a deterrent capability. The Israeli government refuses to confirm or deny possession of nuclear weapons, although this is now regarded as an open secret after Israeli junior nuclear technician Mordechai Vanunu—subsequently arrested and sentenced for treason by Israel—published evidence about the program to the British "Sunday Times" in 1986. On 18 September 2009 the General Conference of the International Atomic Energy Agency called on Israel to open its nuclear facilities to IAEA inspection and adhere to the non-proliferation treaty as part of a resolution on "Israeli nuclear capabilities," which passed by a narrow margin of 49–45 with 16 abstentions. The chief Israeli delegate stated that "Israel will not co-operate in any matter with this resolution." However, similar resolutions were defeated in 2010, 2013, 2014, and 2015. As with Pakistan, the NSG Guidelines currently rule out nuclear exports by all major suppliers to Israel. North Korea acceded to the treaty on 12 December 1985, but gave notice of withdrawal from the treaty on 10 January 2003 following U.S. allegations that it had started an illegal enriched uranium weapons program, and the U.S. subsequently stopping fuel oil shipments under the Agreed Framework which had resolved plutonium weapons issues in 1994. The withdrawal became effective 10 April 2003 making North Korea the first state ever to withdraw from the treaty. North Korea had once before announced withdrawal, on 12 March 1993, but suspended that notice before it came into effect. On 10 February 2005, North Korea publicly declared that it possessed nuclear weapons and pulled out of the six-party talks hosted by China to find a diplomatic solution to the issue. "We had already taken the resolute action of pulling out of the Nuclear Non-Proliferation Treaty and have manufactured nuclear arms for self-defence to cope with the Bush administration's evermore undisguised policy to isolate and stifle the DPRK [Democratic People's Republic of Korea]," a North Korean Foreign Ministry statement said regarding the issue. Six-party talks resumed in July 2005. On 19 September 2005, North Korea announced that it would agree to a preliminary accord. Under the accord, North Korea would scrap all of its existing nuclear weapons and nuclear production facilities, rejoin the NPT, and readmit IAEA inspectors. The difficult issue of the supply of light water reactors to replace North Korea's indigenous nuclear power plant program, as per the 1994 Agreed Framework, was left to be resolved in future discussions. On the next day North Korea reiterated its known view that until it is supplied with a light water reactor it will not dismantle its nuclear arsenal or rejoin the NPT. On 2 October 2006, the North Korean foreign minister announced that his country was planning to conduct a nuclear test "in the future", although it did not state when. On Monday, 9 October 2006 at 01:35:28 (UTC) the United States Geological Survey detected a magnitude 4.3 seismic event north of Kimchaek, North Korea indicating a nuclear test. The North Korean government announced shortly afterward that they had completed a successful underground test of a nuclear fission device. In 2007, reports from Washington suggested that the 2002 CIA reports stating that North Korea was developing an enriched uranium weapons program, which led to North Korea leaving the NPT, had overstated or misread the intelligence. On the other hand, even apart from these press allegations, there remains some information in the public record indicating the existence of a uranium effort. Quite apart from the fact that North Korean First Vice Minister Kang Sok Ju at one point admitted the existence of a uranium enrichment program, Pakistan's then-President Musharraf revealed that the A.Q. Khan proliferation network had provided North Korea with a number of gas centrifuges designed for uranium enrichment. Additionally, press reports have cited U.S. officials to the effect that evidence obtained in dismantling Libya's WMD programs points toward North Korea as the source for Libya's uranium hexafluoride (UF6) – which, if true, would mean that North Korea has a uranium conversion facility for producing feedstock for centrifuge enrichment. Iran is a party to the NPT since 1970 but was found in non-compliance with its NPT safeguards agreement, and the status of its nuclear program remains in dispute. In November 2003 IAEA Director General Mohamed ElBaradei reported that Iran had repeatedly and over an extended period failed to meet its safeguards obligations under the NPT with respect to: After about two years of EU3-led diplomatic efforts and Iran temporarily suspending its enrichment program, the IAEA Board of Governors, acting under Article XII.C of the IAEA Statute, found in a rare non-consensus decision with 12 abstentions that these failures constituted non-compliance with the IAEA safeguards agreement. This was reported to the UN Security Council in 2006, after which the Security Council passed a resolution demanding that Iran suspend its enrichment. Instead, Iran resumed its enrichment program. The IAEA has been able to verify the non-diversion of declared nuclear material in Iran, and is continuing its work on verifying the absence of undeclared activities. In February 2008, the IAEA also reported that it was working to address "alleged studies" of weaponization, based on documents provided by certain Member States, which those states claimed originated from Iran. Iran rejected the allegations as "baseless" and the documents as "fabrications." In June 2009, the IAEA reported that Iran had not "cooperated with the Agency in connection with the remaining issues ... which need to be clarified to exclude the possibility of military dimensions to Iran's nuclear program." The United States concluded that Iran violated its Article III NPT safeguards obligations, and further argued based on circumstantial evidence that Iran's enrichment program was for weapons purposes and therefore violated Iran's Article II nonproliferation obligations. The November 2007 US National Intelligence Estimate (NIE) later concluded that Iran had halted an active nuclear weapons program in the fall of 2003 and that it had remained halted as of mid-2007. The NIE's "Key Judgments," however, also made clear that what Iran had actually stopped in 2003 was only "nuclear weapon design and weaponization work and covert uranium conversion-related and uranium enrichment-related work"-namely, those aspects of Iran's nuclear weapons effort that had not by that point already been leaked to the press and become the subject of IAEA investigations. Since Iran's uranium enrichment program at Natanz—and its continuing work on a heavy water reactor at Arak that would be ideal for plutonium production—began secretly years before in conjunction with the very weaponization work the NIE discussed and for the purpose of developing nuclear weapons, many observers find Iran's continued development of fissile material production capabilities distinctly worrying. Particularly because fissile material availability has long been understood to be the principal obstacle to nuclear weapons development and the primary "pacing element" for a weapons program, the fact that Iran has reportedly suspended weaponization work may not mean very much. As The Bush Administration's Director of National Intelligence (DNI) Mike McConnell hds put it in 2008, the aspects of its work that Iran allegedly suspended were thus "probably the least significant part of the program." Iran stated it has a legal right to enrich uranium for peaceful purposes under the NPT, and further says that it had "constantly complied with its obligations under the NPT and the Statute of the International Atomic Energy Agency". Iran also stated that its enrichment program has been part of its civilian nuclear energy program, which is allowed under Article IV of the NPT. The Non-Aligned Movement has welcomed the continuing cooperation of Iran with the IAEA and reaffirmed Iran's right to the peaceful uses of nuclear technology. Early during his tenure as United Nations Secretary General, between 2007 and 2016, Ban Ki-moon welcomed the continued dialogue between Iran and the IAEA. He urged a peaceful resolution of the issue. In April 2010, during the signing of the U.S.-Russia New START Treaty, President Obama said that the United States, Russia, and other nations were demanding that Iran face consequences for failing to fulfill its obligations under the Nuclear Non-Proliferation Treaty, saying "We will not tolerate actions that flout the NPT, risk an arms race in a vital region, and threaten the credibility of the international community and our collective security." In 2015, Iran negotiated a nuclear deal with the P5+1, a group of countries that consisted of the five permanent members of the UN Security Council (China, France, Russia, the United Kingdom, and the United States) plus Germany. On July 14 2015, the P5+1 and Iran concluded the Joint Comprehensive Plan of Action, lifting sanctions on Iran in exchange for constraints and on Iran's nuclear activities and increased verification by the IAEA. On May 8 2018, the United States withdrew from the JCPOA and reimposed sanctions on Iran. South Africa is the only country that developed nuclear weapons by itself and later dismantled them – unlike the former Soviet states Ukraine, Belarus and Kazakhstan, which inherited nuclear weapons from the former USSR and also acceded to the NPT as non-nuclear weapon states. During the days of apartheid, the South African government developed a deep fear of both a black uprising and the threat of communism. This led to the development of a secret nuclear weapons program as an ultimate deterrent. South Africa has a large supply of uranium, which is mined in the country's gold mines. The government built a nuclear research facility at Pelindaba near Pretoria where uranium was enriched to fuel grade for the Koeberg Nuclear Power Station as well as weapon grade for bomb production. In 1991, after international pressure and when a change of government was imminent, South African Ambassador to the United States Harry Schwarz signed the Nuclear Non-Proliferation Treaty. In 1993, the then president Frederik Willem de Klerk openly admitted that the country had developed a limited nuclear weapon capability. These weapons were subsequently dismantled before South Africa acceded to the NPT and opened itself up to IAEA inspection. In 1994, the IAEA completed its work and declared that the country had fully dismantled its nuclear weapons program. Libya had signed (in 1968) and ratified (in 1975) the Nuclear Non-Proliferation Treaty and was subject to IAEA nuclear safeguards inspections, but undertook a secret nuclear weapons development program in violation of its NPT obligations, using material and technology provided by the A.Q. Khan proliferation network—including actual nuclear weapons designs allegedly originating in China. Libya began secret negotiations with the United States and the United Kingdom in March 2003 over potentially eliminating its WMD programs. In October 2003, Libya was embarrassed by the interdiction of a shipment of Pakistani-designed centrifuge parts sent from Malaysia, also as part of A. Q. Khan's proliferation ring. In December 2003, Libya announced that it had agreed to eliminate all its WMD programs, and permitted U.S. and British teams (as well as IAEA inspectors) into the country to assist this process and verify its completion. The nuclear weapons designs, gas centrifuges for uranium enrichment, and other equipment—including prototypes for improved SCUD ballistic missiles—were removed from Libya by the United States. (Libyan chemical weapons stocks and chemical bombs were also destroyed on site with international verification, with Libya joining the Chemical Weapons Convention.) Libya's non-compliance with its IAEA safeguards was reported to the U.N. Security Council, but with no action taken, as Libya's return to compliance with safeguards and Article II of the NPT was welcomed. In 2011, the Libyan government of Muammar al-Gaddafi was overthrown in the Libyan Civil War with the assistance of a military intervention by NATO forces acting under the auspices of UN Security Council Resolution 1973. Gaddafi's downfall 8 years after the disarmament of Libya, in which Gaddafi agreed to eliminate Libya's nuclear weapons program, has been repeatedly cited by North Korea, which views Gaddafi's fate as a "cautionary tale" that influences North Korea's decision to maintain and intensify its nuclear weapons program and arsenal despite pressure to denuclearize. Syria is a state party to the NPT since 1969 and has a limited civil nuclear program. Before the advent of the Syrian Civil War it was known to operate only one small Chinese-built research reactor, SRR-1. Despite being a proponent of a Weapons of Mass Destruction Free Zone in the Middle East the country was accused of pursuing a military nuclear program with a reported nuclear facility in a desert Syrian region of Deir ez-Zor. The reactor's components had likely been designed and manufactured in North Korea, with the reactor's striking similarity in shape and size to the North Korean Yongbyon Nuclear Scientific Research Center. That information alarmed Israeli military and intelligence to such a degree that the idea of a targeted airstrike was conceived. It resulted in Operation Orchard, that took place on 6 September 2007 and saw as many as eight Israeli aircraft taking part. The Israeli government is said to have bounced the idea of the operation off of the US Bush administration, although the latter declined to participate. The nuclear reactor was destroyed in the attack, which also killed about ten North Korean workers. The attack did not cause an international outcry or any serious Syrian retaliatory moves as both parties tried to keep it secret: Despite a half-century state of war declared by surrounding states, Israel did not want publicity as regards its breach of the ceasefire, while Syria was not willing to acknowledge its clandestine nuclear program. Article X allows a state to leave the treaty if "extraordinary events, related to the subject matter of this Treaty, have jeopardized the supreme interests of its country", giving three months' (ninety days') notice. The state is required to give reasons for leaving the NPT in this notice. NATO states argue that when there is a state of "general war" the treaty no longer applies, effectively allowing the states involved to leave the treaty with no notice. This is a necessary argument to support the NATO nuclear weapons sharing policy, but a troubling one for the logic of the treaty. NATO's argument is based on the phrase "the consequent need to make every effort to avert the danger of such a war" in the treaty preamble, inserted at the behest of U.S. diplomats, arguing that the treaty would at that point have failed to fulfill its function of prohibiting a general war and thus no longer be binding. Many states do not accept this argument. See United States–NATO nuclear weapons sharing above. North Korea has also caused an uproar by its use of this provision of the treaty. Article X.1 only requires a state to give three months' notice in total, and does not provide for other states to question a state's interpretation of "supreme interests of its country". In 1993, North Korea gave notice to withdraw from the NPT. However, after 89 days, North Korea reached agreement with the United States to freeze its nuclear program under the Agreed Framework and "suspended" its withdrawal notice. In October 2002, the United States accused North Korea of violating the Agreed Framework by pursuing a secret uranium enrichment program, and suspended shipments of heavy fuel oil under that agreement. In response, North Korea expelled IAEA inspectors, disabled IAEA equipment, and, on 10 January 2003, announced that it was ending the suspension of its previous NPT withdrawal notification. North Korea said that only one more day's notice was sufficient for withdrawal from the NPT, as it had given 89 days before. The IAEA Board of Governors rejected this interpretation. Most countries held that a new three-months withdrawal notice was required, and some questioned whether North Korea's notification met the "extraordinary events" and "supreme interests" requirements of the treaty. The Joint Statement of 19 September 2005 at the end of the Fourth Round of the Six-Party Talks called for North Korea to "return" to the NPT, implicitly acknowledging that it had withdrawn. The main outcome of the 2000 Conference was the adoption by consensus of a comprehensive Final Document, which included among other things "practical steps for the systematic and progressive efforts" to implement the disarmament provisions of the NPT, commonly referred to as the Thirteen Steps. On 18 July 2005, US President George W. Bush met Indian Prime Minister Manmohan Singh and declared that he would work to change US law and international rules to permit trade in US civilian nuclear technology with India. At the time, British columnist George Monbiot argued that the U.S.-India nuclear deal, in combination with US attempts to deny Iran (an NPT signatory) civilian nuclear fuel-making technology, might destroy the NPT regime. In the first half of 2010, it was strongly believed that China had signed a civilian nuclear deal with Pakistan claiming that the deal was "peaceful". Arms control advocates criticised the reported China-Pakistan deal as they did in case of U.S.-India deal claiming that both the deals violate the NPT by facilitating nuclear programmes in states which are not parties to the NPT. Some reports asserted that the deal was a strategic move by China to balance US influence in South-Asia. According to a report published by U.S. Department of Defense in 2001, China had provided Pakistan with nuclear materials and has given critical technological assistance in the construction of Pakistan's nuclear weapons development facilities, in violation of the Nuclear Non-Proliferation Treaty, of which China even then was a signatory. At the Seventh Review Conference in May 2005, there were stark differences between the United States, which wanted the conference to focus on non-proliferation, especially on its allegations against Iran, and most other countries, who emphasized the lack of serious nuclear disarmament by the nuclear powers. The non-aligned countries reiterated their position emphasizing the need for nuclear disarmament. The 2010 Review Conference was held in May 2010 in New York City, and adopted a final document that included a summary by the Review Conference President, Ambassador Libran Capactulan of the Philippines, and an Action Plan that was adopted by consensus. The 2010 conference was generally considered a success because it reached consensus where the previous Review Conference in 2005 ended in disarray, a fact that many attributed to the U.S. President Barack Obama's commitment to nuclear nonproliferation and disarmament. Some have warned that this success raised unrealistically high expectations that could lead to failure at the next Review Conference in 2015. The "Global Summit on Nuclear Security" took place 12–13 April 2010. The summit was proposed by President Obama in Prague and was intended to strengthen the Nuclear Non-Proliferation Treaty in conjunction with the Proliferation Security Initiative and the Global Initiative to Combat Nuclear Terrorism. Forty seven states and three international organizations took part in the summit, which issued a communiqué and a work plan. For further information see 2010 Nuclear Security Summit. In a major policy speech at the Brandenburg Gate in Berlin on 19 June 2013, United States President Barack Obama outlined plans to further reduce the number of warheads in the U.S. nuclear arsenal. According to "Foreign Policy", Obama proposed a "one-third reduction in strategic nuclear warheads—on top of the cuts already required by the New START treaty—bringing the number of deployed warheads to about 1,000". Obama is seeking to "negotiate these reductions with Russia to continue to move beyond Cold War nuclear postures," according to briefing documents provided to "Foreign Policy". In the same speech, Obama emphasized his administration's efforts to isolate any nuclear weapons capabilities emanating from Iran and North Korea. He also called for a renewed bipartisan effort in the United States Congress to ratify the Comprehensive Nuclear-Test-Ban Treaty and called on countries to negotiate a new treaty to end the production of fissile material for nuclear weapons. On 24 April 2014, it was announced that the nation of the Marshall Islands has brought suit in The Hague against the United States, the former Soviet Union, the United Kingdom, France, China, India, Pakistan, North Korea and Israel seeking to have the disarmament provisions of the NNPT enforced. The 2015 Review Conference of the Parties to the Treaty on the Non-Proliferation of Nuclear Weapons (NPT) was held at the United Nations in New York from 27 April to 22 May 2015 and presided over by Ambassador Taous Feroukhi of Algeria. The Treaty, particularly article VIII, paragraph 3, envisages a review of the operation of the Treaty every five years, a provision which was reaffirmed by the States parties at the 1995 NPT Review and Extension Conference and the 2000 NPT Review Conference. At the 2015 NPT Review Conference, States parties examined the implementation of the Treaty's provisions since 2010. Despite intensive consultations, the Conference was not able to reach agreement on the substantive part of the draft Final Document. Over the years the NPT has come to be seen by many Third World states as "a conspiracy of the nuclear 'haves' to keep the nuclear 'have-nots' in their place". This argument has roots in Article VI of the treaty which "obligates the nuclear weapons states to liquidate their nuclear stockpiles and pursue complete disarmament. The non-nuclear states see no signs of this happening". Some argue that the NWS have not fully complied with their disarmament obligations under Article VI of the NPT. Some countries such as India have criticized the NPT, because it "discriminated against states not possessing nuclear weapons on January 1, 1967," while Iran and numerous Arab states have criticized Israel for not signing the NPT. There has been disappointment with the limited progress on nuclear disarmament, where the five authorized nuclear weapons states still have 22,000 warheads among them and have shown a reluctance to disarm further. As noted , the International Court of Justice, in its advisory opinion on the Legality of the Threat or Use of Nuclear Weapons, stated that "there exists an obligation to pursue in good faith and bring to a conclusion negotiations leading to nuclear disarmament in all its aspects under strict and effective international control. Some critics of the nuclear-weapons states contend that they have failed to comply with Article VI by failing to make disarmament the driving force in national planning and policy with respect to nuclear weapons, even while they ask other states to plan for their security without nuclear weapons. The United States responds to criticism of its disarmament record by pointing out that, since the end of the Cold War, it has eliminated over 13,000 nuclear weapons, and eliminated over 80% of its deployed strategic warheads and 90% of non-strategic warheads deployed to NATO, in the process eliminating whole categories of warheads and delivery systems and reducing its reliance on nuclear weapons. U.S. officials have also pointed out the ongoing U.S. work to dismantle nuclear warheads. By the time accelerated dismantlement efforts ordered by President George W. Bush were completed, the U.S. arsenal was less than a quarter of its size at the end of the Cold War, and smaller than it had been at any point since the Eisenhower administration, well before the drafting of the NPT. The United States has also purchased many thousands of weapons' worth of uranium formerly in Soviet nuclear weapons for conversion into reactor fuel. As a consequence of this latter effort, it has been estimated that the equivalent of one lightbulb in every ten in the United States is powered by nuclear fuel removed from warheads previously targeted at the United States and its allies during the Cold War. The U.S. Special Representative for Nuclear Nonproliferation agreed that nonproliferation and disarmament are linked, noting that they can be mutually reinforcing but also that growing proliferation risks create an environment that makes disarmament more difficult. The United Kingdom, France and Russia likewise defend their nuclear disarmament records, and the five NPT NWS issued a joint statement in 2008 reaffirming their Article VI disarmament commitments. According to Thomas Reed and Danny Stillman, the "NPT has one giant loophole": Article IV gives each non-nuclear weapon state the 'inalienable right' to pursue nuclear energy for the generation of power. A "number of high-ranking officials, even within the United Nations, have argued that they can do little to stop states using nuclear reactors to produce nuclear weapons". A 2009 United Nations report said that: The revival of interest in nuclear power could result in the worldwide dissemination of uranium enrichment and spent fuel reprocessing technologies, which present obvious risks of proliferation as these technologies can produce fissile materials that are directly usable in nuclear weapons. According to critics, those states which possess nuclear weapons, but are not authorized to do so under the NPT, have not paid a significant price for their pursuit of weapons capabilities. Also, the NPT has been explicitly weakened by a number of bilateral deals made by NPT signatories, notably the United States. Based on concerns over the slow pace of nuclear disarmament and the continued reliance on nuclear weapons in military and security concepts, doctrines and policies, the Treaty on the Prohibition of Nuclear Weapons was adopted in July 2017 and was subsequently opened for signature on 20 September 2017. Upon entry into force, it will prohibit each state party from the development, testing, production, stockpiling, stationing, transfer, use and threat of use of nuclear weapons, as well as assistance to those activities. It reaffirms in its preamble the vital role of the full and effective implementation of the NPT.
https://en.wikipedia.org/wiki?curid=22107
Nikolai Bukharin Nikolai Ivanovich Bukharin ( – 15 March 1938) was a Bolshevik revolutionary, Soviet Union politician, Marxist philosopher and prolific author on revolutionary theory. As a young man, he spent six years in exile working closely with fellow exiles Vladimir Lenin and Leon Trotsky. After the revolution of February 1917, he returned to Moscow, where his Bolshevik credentials earned him a high rank in the Bolshevik party and after the October Revolution became editor of the party newspaper "Pravda." Within the Bolshevik Party, Bukharin was initially a left communist, but gradually moved from the left to the right from 1921. His strong support for and defence of the New Economic Policy (NEP) eventually saw him lead the Right Opposition. By late 1924, this stance had positioned Bukharin favourably as Joseph Stalin's chief ally, with Bukharin soon elaborating Stalin's new theory and policy of Socialism in One Country. Together, Bukharin and Stalin ousted Trotsky, Grigory Zinoviev and Lev Kamenev from the party at the XVth Communist Party Congress in December 1927. From 1926 to 1929, Bukharin enjoyed great power as General Secretary of the Comintern's executive committee. However, Stalin's decision to proceed with collectivisation drove the two men apart, and Bukharin was expelled from the Politburo in 1929. When the Great Purge began in 1936, Stalin looked for any pretext to liquidate his former allies and rivals for power, and some of Bukharin's letters, conversations and tapped phone-calls indicated disloyalty. Arrested in February 1937, Bukharin was charged with conspiring to overthrow the Soviet state. After a show trial that alienated many Western communist sympathisers, he was executed in March 1938. Nikolai Bukharin was born on 27 September (9 October, new style), 1888, in Moscow. He was the second son of two schoolteachers, Ivan Gavrilovich Bukharin and Liubov Ivanovna Bukharina. His childhood is vividly recounted in his mostly autobiographic novel "How It All Began". Bukharin's political life began at the age of sixteen, with his lifelong friend Ilya Ehrenburg, when they participated in student activities at Moscow University related to the Russian Revolution of 1905. He joined the Russian Social Democratic Labour Party in 1906, becoming a member of the Bolshevik faction. With Grigori Sokolnikov, Bukharin convened the 1907 national youth conference in Moscow, which was later considered the founding of Komsomol. By age twenty, he was a member of the Moscow Committee of the party. The committee was widely infiltrated by the Tsarist secret police, the Okhrana. As one of its leaders, Bukharin quickly became a person of interest to them. During this time, he became closely associated with Valerian Obolensky and Vladimir Smirnov. He also met his future first wife, Nadezhda Mikhailovna Lukina, his cousin and the sister of Nikolai Lukin, who was also a member of the party. They married in 1911, soon after returning from internal exile. In 1911, after a brief imprisonment, Bukharin was exiled to Onega in Arkhangelsk, but he soon escaped to Hanover. He stayed in Germany for a year before visiting Kraków (now in Poland) in 1912 to meet Vladimir Lenin for the first time. During the exile, he continued his education and wrote several books that established him in his 20s as a major Bolshevik theorist. His work, "" influenced Lenin, who freely borrowed from it in his larger and better-known work, "Imperialism, the Highest Stage of Capitalism". He and Lenin also often had hot disputes on theoretical issues, as well as Bukharin's closeness with the European Left and his anti-statist tendencies. Bukharin developed an interest in the works of Austrian Marxists and non-Marxist economic theorists, such as Aleksandr Bogdanov, who deviated from Leninist positions. Also, while in Vienna in 1913, he helped the Georgian Bolshevik Joseph Stalin write an article, "Marxism and the National Question," at Lenin's request. In October 1916, while based in New York City, Bukharin edited the newspaper "Novy Mir" ("New World") with Leon Trotsky and Alexandra Kollontai. When Trotsky arrived in New York in January 1917, Bukharin was the first of the emigrés to greet him. (Trotsky's wife recalled, "with a bear hug and immediately began to tell them about a public library which stayed open late at night and which he proposed to show us at once" dragging the tired Trotskys across town "to admire his great discovery"). At the news of the Russian Revolution of February 1917, exiled revolutionaries from around the world began to flock back to the homeland. Trotsky left New York on 27 March 1917, sailing for St. Petersburg. Bukharin left New York in early April and returned to Russia by way of Japan (where he was temporarily detained by local police), arriving in Moscow in early May 1917. Politically, the Bolsheviks in Moscow were a minority in relation to the Mensheviks and Social Democrats. As more people began to be attracted to Lenin's promise to bring peace by withdrawing from the Great War, membership in the Bolshevik faction began to increase dramatically —from 24,000 members in February 1917 to 200,000 members in October 1917. Upon his return to Moscow, Bukharin resumed his seat on the Moscow City Committee and also became a member of the Moscow Regional Bureau of the party. To complicate matters further, the Bolsheviks themselves were divided into a right wing and a left wing. The right-wing of the Bolsheviks, including Aleksei Rykov and Viktor Nogin, controlled the Moscow Committee, while the younger left-wing Bolsheviks, including Vladimir Smirnov, Valerian Osinsky, Georgii Lomov, Nikolay Yakovlev, Ivan Kizelshtein and Ivan Stukov, were members of the Moscow Regional Bureau. On 10 October 1917, Bukharin, along with two other Moscow Bolsheviks: Andrei Bubnov and Grigori Sokolnikov were elected to the Central Committee. This strong representation on the Central Committee was a direct recognition of the fact that the Moscow Bureau had grown in importance. Whereas the Bolsheviks had previously been a minority in Moscow behind the Mensheviks and the Socialist Revolutionaries, by September 1917 the Bolsheviks were in the majority in Moscow. Furthermore, the Moscow Regional Bureau was formally responsible for the party organizations in each of the thirteen (13) central provinces around Moscow—which accounted for 37% of the whole population of Russia and 20% of the Bolshevik membership.While no one dominated revolutionary politics in Moscow during the October Revolution as Trotsky did in St. Petersburg, Bukharin certainly was the most prominent leader in Moscow. During the October Revolution, Bukharin drafted, introduced, and defended the revolutionary decrees of the Moscow Soviet. Bukharin then represented the Moscow Soviet in their report to the revolutionary government in Petrograd. Following the October Revolution, Bukharin became the editor of the party's newspaper, "Pravda". Bukharin believed passionately in the promise of world revolution. In the Russian turmoil near the end of World War I, when a negotiated peace with the Central Powers was looming, he demanded a continuance of the war, fully expecting to incite all the foreign proletarian classes to arms. Even as he was uncompromising toward Russia's battlefield enemies, he also rejected any fraternization with the capitalist Allied powers: he reportedly wept when he learned of official negotiations for assistance. Bukharin emerged as the leader of the Left Communists in bitter opposition to Lenin's decision to sign the Treaty of Brest-Litovsk. In this wartime power struggle, he was urged by some of his more fiery allies to have Lenin arrested. He rejected this idea immediately, but the issue would later become the basis of Stalinist charges against him, culminating in the show trial of 1938. After the ratification of the treaty, Bukharin resumed his responsibilities within the party. In March 1919, he became a member of the Comintern's executive committee and a candidate member of the Politburo. During the Civil War period, he published several theoretical economic works, including the popular primer "The ABC of Communism" (with Yevgeni Preobrazhensky, 1919), and the more academic "Economics of the Transitional Period" (1920) and "Historical Materialism" (1921). By 1921, he changed his position and accepted Lenin's emphasis on the survival and strengthening of the Soviet state as the bastion of the future world revolution. He became the foremost supporter of the New Economic Policy (NEP), to which he was to tie his political fortunes. Considered by the left communists as a retreat from socialist policies, the NEP reintroduced money and allowed private ownership and capitalistic practices in agriculture, retail trade, and light industry while the state retained control of heavy industry. While some have criticized Bukharin for this apparent U-turn, his change of emphasis can be partially explained by the necessity for peace and stability following seven years of war in Russia, and the failure of communist revolutions in Central and Eastern Europe, which ended the prospect of worldwide revolution. After Lenin's death in 1924, Bukharin became a full member of the Politburo. In the subsequent power struggle among Leon Trotsky, Grigory Zinoviev, Lev Kamenev and Stalin, Bukharin allied himself with Stalin, who positioned himself as centrist of the Party and supported the NEP against the Left Opposition, which wanted more rapid industrialization, escalation of class struggle against the kulaks (wealthier peasants), and agitation for world revolution. It was Bukharin who formulated the thesis of "Socialism in One Country" put forth by Stalin in 1924, which argued that socialism (in Marxist theory, the transitional stage from capitalism to communism) could be developed in a single country, even one as underdeveloped as Russia. This new theory stated that socialist gains could be consolidated in a single country, without that country relying on simultaneous successful revolutions across the world. The thesis would become a hallmark of Stalinism. Trotsky, the prime force behind the Left Opposition, was defeated by a triumvirate formed by Stalin, Zinoviev, and Kamenev, with the support of Bukharin. At the Fourteenth Party Congress in December 1925, Stalin openly attacked Kamenev and Zinoviev, revealing that they had asked for his aid in expelling Trotsky from the Party. By 1926, the Stalin-Bukharin alliance ousted Zinoviev and Kamenev from the Party leadership, and Bukharin enjoyed the highest degree of power during the 1926–1928 period. He emerged as the leader of the Party's right wing, which included two other Politburo members (Alexei Rykov, Lenin's successor as Chairman of the Council of People's Commissars and Mikhail Tomsky, head of trade unions) and he became General Secretary of the Comintern's executive committee in 1926. However, prompted by a grain shortage in 1928, Stalin reversed himself and proposed a program of rapid industrialization and forced collectivization because he believed that the NEP was not working fast enough. Stalin felt that in the new situation the policies of his former foes—Trotsky, Zinoviev, and Kamenev—were the right ones. Bukharin was worried by the prospect of Stalin's plan, which he feared would lead to "military-feudal exploitation" of the peasantry. Bukharin did want the Soviet Union to achieve industrialization but he preferred the more moderate approach of offering the peasants the opportunity to become prosperous, which would lead to greater grain production for sale abroad. Bukharin pressed his views throughout 1928 in meetings of the Politburo and at the Communist Party Congress, insisting that enforced grain requisition would be counterproductive, as War Communism had been a decade earlier. Bukharin's support for the continuation of the NEP was not popular with higher Party cadres, and his slogan to peasants, "Enrich yourselves!" and proposal to achieve socialism "at snail's pace" left him vulnerable to attacks first by Zinoviev and later by Stalin. Stalin attacked Bukharin's views, portraying them as capitalist deviations and declaring that the revolution would be at risk without a strong policy that encouraged rapid industrialization. Having helped Stalin achieve unchecked power against the Left Opposition, Bukharin found himself easily outmaneuvered by Stalin. Yet Bukharin played to Stalin's strength by maintaining the appearance of unity within the Party leadership. Meanwhile, Stalin used his control of the Party machine to replace Bukharin's supporters in the Rightist power base in Moscow, trade unions, and the Comintern.Bukharin attempted to gain support from earlier foes including Kamenev and Zinoviev who had fallen from power and held mid-level positions within the Communist party. The details of his meeting with Kamenev, to whom he confided that Stalin was "Genghis Khan" and changed policies to get rid of rivals, were leaked by the Trotskyist press and subjected him to accusations of factionalism. Eventually, Bukharin lost his position in the Comintern and the editorship of "Pravda" in April 1929 and he was expelled from the Politburo on 17 November of that year. Bukharin was forced to renounce his views under pressure. He wrote letters to Stalin pleading for forgiveness and rehabilitation, but through wiretaps of Bukharin's private conversations with Stalin's enemies, Stalin knew Bukharin's repentance was insincere. International supporters of Bukharin, Jay Lovestone of the Communist Party USA among them, were also expelled from the Comintern. They formed an international alliance to promote their views, calling it the "International Communist Opposition", though it became better known as the Right Opposition, after a term used by the Trotskyist Left Opposition in the Soviet Union to refer to Bukharin and his supporters there. In the brief period of thaw in 1934–1936, Bukharin was politically rehabilitated and was made editor of "Izvestia" in 1934. There, he consistently highlighted the dangers of fascist regimes in Europe and the need for "proletarian humanism". One of his first decisions as editor was to invite Boris Pasternak to contribute to the newspaper and sit in on editorial meetings. Pasternak described Bukharin as "a wonderful, historically extraordinary man, but fate has not been kind to him." They first met during the lying-in-state of the Soviet police chief, Vyacheslav Menzhinsky in May 1934, when Pasternak was seeking help for his fellow poet, Osip Mandelstam, who had been arrested - though at that time neither Pasternak nor Bukharin knew why. Bukharin had acted as Mandelstam's political protector since 1922. According to Mandelstam's wife, Nadezhda, "M. owed him all the pleasant things in his life. His 1928 volume of poetry would never have come out without the active intervention of Bukharin. The journey to Armenia, our apartment and ration cards, contracts for future volumes – all this was arranged by Bukharin." Bukharin wrote to Stalin, pleading clemency for Mandelstam, and appealed personally to the head of the NKVD, Genrikh Yagoda. It was Yagoda who told him about Mandelstam's Stalin Epigram, after which he refused to have any further contact with Nadezhda Mandelstam, who had lied to him by denying that her husband had written "anything rash". – but continued to befriend Pasternak. Soon after Mandelstam's arrest, Bukharin was delegated to prepare the official report on poetry for the First Soviet Writers' Congress, in August 1934. He could not any longer risk mentioning Mandelstam in his speech to the congress, but did devote a large section of his to Pasternak, whom he described as "remote from current affairs...a singer of the old intelligensia...delicate and subtle...a wounded and easily vulnerable soul. He is the embodiment of chaste but self-absorbed laboratory craftsmanship..." His speech was greeted with wild applause, though it greatly offended some of the listeners, such as the communist poet Semyon Kirsanov, who complained: "according to Bukharin, all the poets who have used their verses to participate in political life are out of date, but the others are not out of date, the so-called pure (and not so pure) lyric poets." When Bukharin was arrested two years later, Boris Pasternak displayed extraordinary courage by having a letter delivered to Bukharin's wife saying that he was convinced of his innocence. Stalin's collectivization policy proved to be as disastrous as Bukharin predicted, but Stalin had by then achieved unchallenged authority in the party leadership. However, there were signs that moderates among Stalin's supporters sought to end official terror and bring a general change in policy, after mass collectivization was largely completed and the worst was over. Although Bukharin had not challenged Stalin since 1929, his former supporters, including Martemyan Ryutin, drafted and clandestinely circulated an anti-Stalin platform, which called Stalin the "evil genius of the Russian Revolution". However, Sergey Kirov, First Secretary of the Leningrad Regional Committee was assassinated in Leningrad in December 1934, and his death was used by Stalin as a pretext to launch the Great Purge, in which about a million people were to perish as Stalin eliminated all past and potential opposition to his authority. Some historians believe that Kirov's assassination in 1934 was arranged by Stalin himself or at least that there is sufficient evidence to plausibly posit such a conclusion. After Kirov's assassination, the NKVD charged an ever-growing group of former oppositionists with Kirov's murder and other acts of treason, terrorism, sabotage, and espionage. In February 1936, shortly before the purge started in earnest, Bukharin was sent to Paris by Stalin to negotiate the purchase of the Marx and Engels archives, held by the German Social Democratic Party (SPD) before its dissolution by Hitler. He was joined by his young wife Anna Larina, which therefore opened the possibility of exile, but he decided against it, saying that he could not live outside the Soviet Union. Bukharin, who had been forced to follow the Party line since 1929, confided to his old friends and former opponents his real view of Stalin and his policy. His conversations with Boris Nicolaevsky, a Menshevik leader who held the manuscripts on behalf of the SPD, formed the basis of "Letter of an Old Bolshevik", which was very influential in contemporary understanding of the period (especially the Ryutin Affair and the Kirov murder), although there are doubts about its authenticity. According to Nicolaevsky, Bukharin spoke of "the mass annihilation of completely defenseless men, with women and children" under forced collectivization and liquidation of kulaks as a class that dehumanized the Party members with "the profound psychological change in those communists who took part in the campaign. Instead of going mad, they accepted terror as a normal administrative method and regarded obedience to all orders from above as a supreme virtue. ... They are no longer human beings. They have truly become the cogs in a terrible machine." Yet to another Menshevik leader, Fyodor Dan, he confided that Stalin became "the man to whom the Party granted its confidence" and "is a sort of a symbol of the Party" even though he "is not a man, but a devil." In Dan's account, Bukharin's acceptance of the Soviet Union's new direction was thus a result of his utter commitment to Party solidarity. To André Malraux, he also confided, "Now he is going to kill me". To his boyhood friend, Ilya Ehrenburg, he expressed the suspicion that the whole trip was a trap set up by Stalin. Indeed, his contacts with Mensheviks during this trip were to feature prominently in his trial. Following the trial and execution of Zinoviev, Kamenev, and other leftist Old Bolsheviks in 1936, Bukharin and Rykov were arrested on 27 February 1937 following a plenum of the Central Committee and were charged with conspiring to overthrow the Soviet state. Bukharin was tried in the Trial of the Twenty One on 2–13 March 1938 during the Great Purge, along with ex-premier Alexei Rykov, Christian Rakovsky, Nikolai Krestinsky, Genrikh Yagoda, and 16 other defendants alleged to belong to the so-called "Bloc of Rightists and Trotskyites". In a trial meant to be the culmination of previous show trials, it was alleged that Bukharin and others sought to assassinate Lenin and Stalin from 1918, murder Maxim Gorky by poison, partition the Soviet Union and hand out her territories to Germany, Japan, and Great Britain. Even more than earlier Moscow show trials, Bukharin's trial horrified many previously sympathetic observers as they watched allegations become more absurd than ever and the purge expand to include almost every living Old Bolshevik leader except Stalin. For some prominent communists such as Bertram Wolfe, Jay Lovestone, Arthur Koestler, and Heinrich Brandler, the Bukharin trial marked their final break with communism and even turned the first three into passionate anti-Communists eventually. While Anastas Mikoyan and Vyacheslav Molotov later claimed that Bukharin was never tortured and his letters from prison do not give the suggestion that he was tortured, it is also known that his interrogators were instructed with the order: "beating permitted". Bukharin held out for three months, but threats to his young wife and infant son, combined with "methods of physical influence" wore him down. But when he read his confession amended and corrected personally by Stalin, he withdrew his whole confession. The examination started all over again, with a double team of interrogators. Bukharin's confession and his motivation became subject of much debate among Western observers, inspiring Koestler's acclaimed novel "Darkness at Noon" and a philosophical essay by Maurice Merleau-Ponty in "Humanism and Terror." His confessions were somewhat different from others in that while he pleaded guilty to the "sum total of crimes," he denied knowledge when it came to specific crimes. Some astute observers noted that he would allow only what was in the written confession and refuse to go any further. There are several interpretations of Bukharin's motivations (besides being coerced) in the trial. Koestler and others viewed it as a true believer's last service to the Party (while preserving the little amount of personal honor left) whereas Bukharin biographer Stephen Cohen and Robert Tucker saw traces of Aesopian language, with which Bukharin sought to turn the table into an anti-trial of Stalinism (while keeping his part of the bargain to save his family). While his letters to Stalin – he wrote 34 very emotional and desperate letters tearfully protesting his innocence and professing his loyalty – suggest a complete capitulation and acceptance of his role in the trial, it contrasts with his actual conduct in the trial. Bukharin himself speaks of his "peculiar duality of mind" in his last plea, which led to "semi-paralysis of the will" and Hegelian "unhappy consciousness", which likely stemmed not only from his knowledge of the ruinous reality of Stalinism (although he could not of course say so in the trial) but also of the impending threat of fascism. The result was a curious mix of fulsome confessions (of being a "degenerate fascist" working for the "restoration of capitalism") and subtle criticisms of the trial. After disproving several charges against him (one observer noted that he "proceeded to demolish or rather showed he could very easily demolish the whole case.") and saying that "the confession of the accused is not essential. The confession of the accused is a medieval principle of jurisprudence" in a trial that was solely based on confessions, he finished his last plea with the words: the monstrousness of my crime is immeasurable especially in the new stage of struggle of the U.S.S.R. May this trial be the last severe lesson, and may the great might of the U.S.S.R. become clear to all. The state prosecutor Vyshinsky characterized Bukharin as an "accursed crossbreed of fox and pig" who supposedly committed a "whole nightmare of vile crimes". While in prison, he wrote at least four book-length manuscripts including a lyrical autobiographical novel, "How It All Began", philosophical treatise "Philosophical Arabesques", a collection of poems, and "Socialism and Its Culture" – all of which were found in Stalin's archive and published in the 1990s. Among other intercessors, the French author and Nobel laureate Romain Rolland wrote to Stalin seeking clemency, arguing that "an intellect like that of Bukharin is a treasure for his country." He compared Bukharin's situation to that of the great chemist Antoine Lavoisier who was guillotined during the French Revolution: "We in France, the most ardent revolutionaries... still profoundly grieve and regret what we did. ... I beg you to show clemency." He had earlier written to Stalin in 1937, "For the sake of Gorky I am asking you for mercy, even if he may be guilty of something," to which Stalin noted: "We must not respond." Bukharin was shot on 15 March 1938 at the Kommunarka shooting ground, but the announcement of his death was overshadowed by the Nazi Anschluss of Austria. According to Zhores and Roy Medvedev in "The Unknown Stalin" (2006), Bukharin's last message to Stalin stated "Koba, why do you need me to die?", which was written in a note to Stalin just before his execution. "Koba" was Stalin's "nom de guerre", and Bukharin's use of it was a sign of how close the two had once been. The note was allegedly found still in Stalin's desk after his death in 1953. Despite the promise to spare his family, Bukharin's wife, Anna Larina, was sent to a labor camp, but she survived to see her husband officially rehabilitated by the Soviet state under Mikhail Gorbachev in 1988. Bukharin was immensely popular within the party throughout the twenties and thirties, even after his fall from power. In his testament, Lenin portrayed him as the Golden Boy of the party, writing: Speaking of the young C.C. members, I wish to say a few words about Bukharin and Pyatakov. They are, in my opinion, the most outstanding figures (among the youngest ones), and the following must be borne in mind about them: Bukharin is not only a most valuable and major theorist of the Party; he is also rightly considered the favourite of the whole Party, but his theoretical views can be classified as fully Marxist only with great reserve, for there is something scholastic about him (he has never made a study of the dialectics, and, I think, never fully understood it) ... Both of these remarks, of course, are made only for the present, on the assumption that both these outstanding and devoted Party workers fail to find an occasion to enhance their knowledge and amend their one-sidedness. Bukharin made several notable contributions to Marxist–Leninist thought, most notably "The Economics of the Transition Period" (1920) and his prison writings, "Philosophical Arabesques", as well as being a founding member of the Soviet Academy of Arts and Sciences, and a keen botanist. His primary contributions to economics were his critique of marginal utility theory, his analysis of imperialism, and his writings on the transition to communism in the Soviet Union. His ideas, especially in economics and the question of market-socialism, later became highly influential in Chinese market-socialism and Deng Xiaoping's reforms. British author Martin Amis argues that Bukharin was perhaps the only major Bolshevik to acknowledge "moral hesitation" by questioning, even in passing, the violence and sweeping reforms of the early Soviet Union. Amis writes that Bukharin said "during the Civil War he had seen 'things that I would not want even my enemies to see'." Bukharin was a cartoonist who left many cartoons of contemporary Soviet politicians. The renowned artist Konstantin Yuon once told him: "Forget about politics. There is no future in politics for you. Painting is your real calling." His cartoons are sometimes used to illustrate the biographies of Soviet officials. Russian historian Yury Zhukov stated that Nikolai Bukharin's portraits of Joseph Stalin were the only ones drawn from the original, not from a photograph.
https://en.wikipedia.org/wiki?curid=22109
Nasal consonant In phonetics, a nasal, also called a nasal occlusive or nasal stop in contrast with an oral stop or nasalized consonant, is an occlusive consonant produced with a lowered velum, allowing air to escape freely through the nose. The vast majority of consonants are oral consonants. Examples of nasals in English are , and , in words such as "nose", "bring" and "mouth". Nasal occlusives are nearly universal in human languages. There are also other kinds of nasal consonants in some languages. Nearly all nasal consonants are nasal occlusives, in which air escapes through the nose but not through the mouth, as it is blocked (occluded) by the lips or tongue. The oral cavity still acts as a resonance chamber for the sound. Rarely, non-occlusive consonants may be nasalized. Most nasals are voiced, and in fact, the nasal sounds and are among the most common sounds cross-linguistically. Voiceless nasals occur in a few languages such as Burmese, Welsh, Icelandic and Guaraní. (Compare oral stops, which block off the air completely, and fricatives, which obstruct the air with a narrow channel. Both stops and fricatives are more commonly voiceless than voiced, and are known as obstruents.) In terms of acoustics, nasals are sonorants, which means that they do not significantly restrict the escape of air (as it can freely escape out the nose). However, nasals are also obstruents in their articulation because the flow of air through the mouth is blocked. This duality, a sonorant airflow through the nose along with an obstruction in the mouth, means that nasal occlusives behave both like sonorants and like obstruents. For example, nasals tend to pattern with other sonorants such as and , but in many languages, they may develop from or into stops. Acoustically, nasals have bands of energy at around 200 and 2,000 Hz. 1. The symbol is commonly used to represent the dental nasal as well, rather than , as it is rarely distinguished from the alveolar nasal. Examples of languages containing nasal occlusives: The voiced retroflex nasal is is a common sound in Languages of India. The voiced palatal nasal is a common sound in European languages, such as: Spanish , French and Italian , Catalan and Hungarian , Czech and Slovak , Polish , Occitan and Portuguese , and (before a vowel) Modern Greek . Many Germanic languages, including German, Dutch, English and Swedish, as well as varieties of Chinese such as Mandarin and Cantonese, have , and . Tamil has a six-fold distinction between , , , , and . Catalan, Occitan, Spanish, and Italian have , , as phonemes, and and as allophones. Nevertheless, in several American dialects of Spanish, there is no palatal nasal but only a palatalized nasal, , as in English "canyon". In Brazilian Portuguese and Angolan Portuguese , written , is typically pronounced as , a nasal palatal approximant, a nasal glide (in Polish, this feature is also possible as an allophone). Semivowels in Portuguese often nasalize before and always after nasal vowels, resulting in and . What would be coda nasal occlusives in other West Iberian languages is only slightly pronounced before dental consonants. Outside this environment the nasality is spread over the vowel or become a nasal diphthong ("mambembe" , outside the final, only in Brazil, and "mantém" in all Portuguese dialects). The Japanese syllabary kana ん, typically romanized as "n" and occasionally "m", can manifest as one of several different nasal consonants depending on what consonant follows it; this allophone, colloquially written in IPA as , is known as the moraic nasal, per the language's moraic structure. Welsh has a set of voiceless nasals, [m̥], [n̥] and [ŋ̊], which occur predominantly as a result of nasal mutation of their voiced counterparts ([m], [n] and [ŋ]). The Mapos Buang language of New Guinea has a phonemic uvular nasal, [ɴ], which contrasts with a velar nasal. It is extremely rare for a language to have [ɴ] as a phoneme. Yanyuwa is highly unusual in that it has a seven-way distinction between [m], [n̪], [n], [ɳ], [ṉ] (palato-alveolar), [ŋ̟] (front velar), and [ŋ̠] (back velar). This may be the only language in existence that contrasts nasals at seven distinct points of articulation. The term 'nasal occlusive' (or 'nasal stop') is generally abbreviated to "nasal". However, there are also nasalized fricatives, nasalized flaps, nasal glides, and nasal vowels, as in French, Portuguese, and Polish. In the , nasal vowels and nasalized consonants are indicated by placing a tilde (~) over the vowel or consonant in question: French "sang" , Portuguese "bom" . A few languages have phonemic voiceless nasal occlusives. Among them are Icelandic, Faroese, Burmese, Jalapa Mazatec, Kildin Sami, Welsh, and Central Alaskan Yup'ik. Iaai of New Caledonia has an unusually large number of them, with , along with a number of voiceless approximants. Ladefoged and Maddieson (1996) distinguish purely nasal consonants, the nasal occlusives such as "m n ng" in which the airflow is purely nasal, from partial nasal consonants such as prenasalized consonants and nasal pre-stopped consonants, which are nasal for only part of their duration, as well as from nasalized consonants, which have simultaneous oral and nasal airflow. In some languages, such as Portuguese, a nasal consonant may have occlusive and non-occlusive allophones. In general, therefore, a nasal consonant may be: A few languages, perhaps 2%, contain no phonemically distinctive nasals. This led Ferguson (1963) to assume that all languages have at least one primary nasal occlusive. However, there are exceptions. When a language is claimed to lack nasals altogether, as with several Niger–Congo languages or the Pirahã language of the Amazon, nasal and non-nasal or prenasalized consonants usually alternate allophonically, and it is a theoretical claim on the part of the individual linguist that the nasal is not the basic form of the consonant. In the case of some Niger–Congo languages, for example, nasals occur before only nasal vowels. Since nasal vowels are phonemic, it simplifies the picture somewhat to assume that nasalization in occlusives is allophonic. There is then a second step in claiming that nasal vowels nasalize oral occlusives, rather than oral vowels denasalizing nasal occlusives, that is, whether are phonemically without full nasals, or without prenasalized stops. Postulating underlying oral or prenasalized stops rather than true nasals helps to explain the apparent instability of nasal correspondences throughout Niger–Congo compared with, for example, Indo-European. This analysis comes at the expense, in some languages, of postulating either a single nasal consonant that can only be syllabic, or a larger set of nasal vowels than oral vowels, both typologically odd situations. The way such a situation could develop is illustrated by a Jukunoid language, Wukari. Wukari allows oral vowels in syllables like "ba, mba" and nasal vowels in "bã, mã", suggesting that nasals become prenasalized stops before oral vowels. Historically, however, *mb became **mm before nasal vowels, and then reduced to *m, leaving the current asymmetric distribution. In older speakers of the Tlingit language, and are allophones. Tlingit is usually described as having an unusual, perhaps unique lack of despite having five lateral obstruents; the older generation could be argued to have but at the expense of having no nasals. Several of languages surrounding Puget Sound, such as Quileute (Chimakuan family), Lushootseed (Salishan family), and Makah (Wakashan family), are truly without any nasalization whatsoever, in consonants or vowels, except in special speech registers such as baby talk or the archaic speech of mythological figures (and perhaps not even that in the case of Quileute). This is an areal feature, only a few hundred years old, where nasals became voiced stops ( became , etc.) after colonial contact. For example, "Snohomish" is currently pronounced "sdohobish", but was transcribed with nasals in the first English-language records. The only other places in the world where this is known to occur is in Melanesia. In the central dialect of the Rotokas language of Bougainville Island, nasals are only used when imitating foreign accents. (A second dialect has a series of nasals.) The Lakes Plain languages of West Irian are similar. The unconditioned loss of nasals, as in Puget Sound, is unusual. However, currently in Korean, word-initial and are shifting to and . This started out in nonstandard dialects and was restricted to the beginning of prosodic units (a common position for fortition), but has expanded to many speakers of the standard language to the beginnings of common words even within prosodic units.
https://en.wikipedia.org/wiki?curid=22110
Nuvistor The nuvistor is a type of vacuum tube announced by RCA in 1959. Most nuvistors are basically thimble-shaped, but somewhat smaller than a thimble, and much smaller than conventional tubes of the day, almost approaching the compactness of early discrete transistor casings. Triodes and a few tetrodes were made. The tube is made entirely of metal and ceramic. Making nuvistors requires special equipment, since there is no intubation to pump gases out of the envelope. Instead, the entire structure is assembled, inserted into its metal envelope, sealed and processed in a large vacuum chamber with simple robotic devices. Nuvistors are among the highest performing small signal receiving tubes. They feature excellent VHF and UHF performance plus low noise figures, and were widely used throughout the 1960s in television sets (beginning with RCA's "New Vista" line of color sets in 1961 with the CTC-11 chassis), radio and high-fidelity equipment primarily in RF sections, and oscilloscopes. They competed with the solid state revolution, and along with GE's Compactron, probably held it at bay for a few years. RCA discontinued their use in television tuners for its product line in late 1971. One famous application was in the Ampex MR-70, a costly studio tape recorder whose entire electronics section was based on nuvistors. Another limited application of this very small tube was in studio-grade microphones from that era, the AKG/Norelco C12a, which employed the 7586, being a good example. It was also later found that, with minor circuit modification, the nuvistor made a sufficient replacement for the obsolete Telefunken VF14 tube, used in the famed Neumann U 47 studio microphone.. Tektronix also used nuvistors in several of its high end oscilloscopes of the 1960s, before replacing them later with JFET transistors.
https://en.wikipedia.org/wiki?curid=22111
No Logo No Logo: Taking Aim at the Brand Bullies is a book by the Canadian author Naomi Klein. First published by Knopf Canada and Picador in December 1999, shortly after the 1999 WTO Ministerial Conference protests in Seattle had generated media attention around such issues, it became one of the most influential books about the alter-globalization movement and an international bestseller. The book focuses on branding and often makes connections with the anti-globalization movement. Throughout the four parts ("No Space", "No Choice", "No Jobs", and "No Logo"), Klein writes about issues such as sweatshops in the Americas and Asia, culture jamming, corporate censorship, and Reclaim the Streets. She pays special attention to the deeds and misdeeds of Nike, The Gap, McDonald's, Shell, and Microsoft – and of their lawyers, contractors, and advertising agencies. Many of the ideas in Klein's book derive from the influence of the Situationists, an art/political group founded in the late 1950s. However, while globalization appears frequently as a recurring theme, Klein rarely addresses the topic of globalization itself, and when she does, it is usually indirectly. She goes on to discuss globalization in much greater detail in her book, "Fences and Windows" (2002). The book comprises four sections: "No Space", "No Choice", "No Jobs", and "No Logo". The first three deal with the negative effects of brand-oriented corporate activity, while the fourth discusses various methods people have taken in order to fight back. The book begins by tracing the history of brands. Klein argues that there has been a shift in the usage of branding and gives examples of this shift to "anti-brand" branding. Early examples of brands were often used to put a recognizable face on factory-produced products. These slowly gave way to the idea of selling lifestyles. According to Klein, in response to an economic crash in the 1980s (due to the Latin American debt crisis, Black Monday (1987), the savings and loan crisis, and the Japanese asset price bubble), corporations began to seriously rethink their approach to marketing and to target the youth demographic, as opposed to the baby boomers, who had previously been considered a much more valuable segment. The book discusses how brand names such as Nike or Pepsi expanded beyond the mere products which bore their names, and how these names and logos began to appear everywhere. As this happened, the brands' obsession with the youth market drove them to further associate themselves with whatever the youth considered "cool". Along the way, the brands attempted to associate their names with everything from movie stars and athletes to grassroots social movements. Klein argues that large multinational corporations consider the marketing of a brand name to be more important than the actual manufacture of products; this theme recurs in the book, and Klein suggests that it helps explain the shift to production in Third World countries in such industries as clothing, footwear, and computer hardware. This section also looks at ways in which brands have "muscled" their presence into the school system, and how in doing so, they have pipelined advertisements into the schools and used their position to gather information about the students. Klein argues that this is part of a trend toward targeting younger and younger consumers. In the second section, Klein discusses how brands use their size and clout to limit the number of choices available to the public – whether through market dominance (e.g., Wal-Mart) or through aggressive invasion of a region (e.g., Starbucks). Klein argues that each company's goal is to become the dominant force in its respective field. Meanwhile, other corporations, such as Sony or Disney, simply open their own chains of stores, preventing the competition from even putting their products on the shelves. This section also discusses the way that corporations merge with one another in order to add to their ubiquity and provide greater control over their image. ABC News, for instance, is allegedly under pressure not to air any stories that are overly critical of Disney, its parent company. Other chains, such as Wal-Mart, often threaten to pull various products off their shelves, forcing manufacturers and publishers to comply with their demands. This might mean driving down manufacturing costs or changing the artwork or content of products like magazines or albums so they better fit with Wal-Mart's image of family friendliness. Also discussed is the way that corporations abuse copyright laws in order to silence anyone who might attempt to criticize their brand. In this section, the book takes a darker tone and looks at the way in which manufacturing jobs move from local factories to foreign countries, and particularly to places known as export processing zones. Such zones often have no labor laws, leading to dire working conditions. The book then shifts back to North America, where the lack of manufacturing jobs has led to an influx of work in the service sector, where most of the jobs are for minimum wage and offer no benefits. The term "McJob" is introduced, defined as a job with poor compensation that does not keep pace with inflation, inflexible or undesirable hours, little chance of advancement, and high levels of stress. Meanwhile, the public is being sold the perception that these jobs are temporary employment for students and recent graduates, and therefore need not offer living wages or benefits. All of this is set against a backdrop of massive profits and wealth being produced within the corporate sector. The result is a new generation of employees who have come to resent the success of the companies they work for. This resentment, along with rising unemployment, labour abuses abroad, disregard for the environment, and the ever-increasing presence of advertising breeds a new disdain for corporations. The final section of the book discusses various movements that have sprung up during the 1990s. These include "Adbusters" magazine and the culture-jamming movement, as well as Reclaim the Streets and the McLibel trial. Less radical protests are also discussed, such as the various movements aimed at putting an end to sweatshop labour. Klein concludes by contrasting consumerism and citizenship, opting for the latter. "When I started this book," she writes, "I honestly didn't know whether I was covering marginal atomized scenes of resistance or the birth of a potentially broad-based movement. But as time went on, what I clearly saw was a movement forming before my eyes." After the book's release, Klein was heavily criticized by the newspaper "The Economist", leading to a broadcast debate with Klein and the magazine's writers, dubbed "No Logo vs. Pro Logo". The 2004 book "The Rebel Sell" (published as "Nation of Rebels" in the United States) specifically criticized "No Logo", stating that turning the improving quality of life in the working class into a fundamentally anti-market ideology is shallow. In this book, Klein criticized Nike so severely that Nike published a point-by-point response. In 2000, "No Logo" was short-listed for the "Guardian" First Book Award in 2000. In 2001, the book won the following awards: Several imprints of "No Logo" exist, including a hardcover first edition, a subsequent hardcover edition, and a paperback. A 10th anniversary edition was published by Fourth Estate that includes a new introduction by the author. Translations from the original English into several other languages have also been published. The subtitle, "Taking Aim at the Brand Bullies", was dropped in some later editions. Naomi Klein explains her ideas in the 40-minute video "No Logo – Brands, Globalization & Resistance" (2003), directed by Sut Jhally. Members of the English rock group Radiohead have stated that the book influenced them particularly during the making of their fourth and fifth albums, "Kid A" (2000) and "Amnesiac" (2001), respectively. (The albums were recorded over the same sessions.) The band recommended the book to fans on their website and considered calling the album "Kid A" "No Logo" for a time. Argentine artist Indio Solari wrote a song for his first solo album named "Nike es la cultura" ("Nike is the culture"), in which he says, "You shout No Logo! Or you doesn't shout No Logo! Or you shout No Logo No!" in reference to this book. Dhani Harrison, son of George Harrison and front-man of English electronic/alternative rock group Thenewno2, has stated that "No Logo" had a large influence on their release, "You Are Here" (2008). Argentine-American rock singer Kevin Johansen wrote a song, "Logo", inspired by Klein's book. A copy of "No Logo" is even used in the official video for the song. Rapper MC Lars's album "This Gigantic Robot Kills" contains a track entitled "No Logo", a satirical analysis of anti-government youth, partially inspired by the book.
https://en.wikipedia.org/wiki?curid=22113
National War College The National War College (NWC) of the United States is a school in the National Defense University. It is housed in Roosevelt Hall on Fort Lesley J. McNair, Washington, D.C., the third-oldest Army post still active. The National War College (NWC) was officially established on July 1, 1946, as an upgraded replacement for the Army-Navy Staff College, which operated from June 1943 to July 1946. The college was one of James Forrestal's favorite causes. According to Lt. Gen. Leonard T. Gerow, President of the Board which recommended its formation: Mid-level and senior military officers who are likely to be promoted to the most senior ranks are selected to study at the War College in preparation for higher staff and command positions. About 75 percent of the student body is composed of equal representation from the land, air, and sea (including Marine and Coast Guard) services. The remaining 25 percent are drawn from the Department of State and other federal departments and agencies. In addition, international fellows from a number of countries join the student body. The curriculum is based upon critical analysis of strategic problem solving with emphasis on strategic leadership. Starting with the 2014–2015 academic year, the curriculum will be based upon a core standard throughout National Defense University. Because of NWC's privileged location close to the White House, the Supreme Court, and Capitol Hill, it has been able throughout its history to call upon an extraordinarily well-connected array of speakers to animate its discussions. All lectures at the National War College are conducted under a strict "no quotation nor attribution" policy which has facilitated discussion on some of the most difficult issues of the day. Graduates of the National War College include numerous current and former flag officers, general officers, and U.S. ambassadors. Notable graduates include: Roosevelt Hall (built 1903–1907) is a Beaux Arts–style building housing the NWC since its inception in 1946. Designed by the New York architectural firm McKim, Mead and White, it is now designated a National Historical Landmark and is listed on the National Register of Historic Places.
https://en.wikipedia.org/wiki?curid=22115
Nürburgring The Nürburgring is a 150,000 person capacity motorsports complex located in the town of Nürburg, Rhineland-Palatinate, Germany. It features a Grand Prix race track built in 1984, and a much longer "Nordschleife" "North loop" track which was built in the 1920s around the village and medieval castle of Nürburg in the Eifel mountains. The north loop is long and has more than 300 metres (1,000 feet) of elevation change from its lowest to highest points. Jackie Stewart nicknamed the old track "The Green Hell". Originally, the track featured four configurations: the -long "Gesamtstrecke" ("Whole Course"), which in turn consisted of the "Nordschleife" ("North Loop"), and the "Südschleife" ("South Loop"). There also was a warm-up loop called "Zielschleife" ("Finish Loop") or "Betonschleife" ("Concrete Loop"), around the pit area. Between 1982 and 1983 the start/finish area was demolished to create a new "GP-Strecke", and this is used for all major and international racing events. However, the shortened "Nordschleife" is still in use for racing, testing and public access. In 1907, the first Eifelrennen race was held on the one-off Taunus circuit, a 117 km (73 mi) made up of public roads starting between the towns of Wehrleim and Saalburg just north of Frankfurt. In the early 1920s, ADAC Eifelrennen races were held on the twisty 33.2 km (20.6 mi) Nideggen public road circuit near Cologne and Bonn. Sometime around 1925, the construction of a dedicated race track was proposed just south of the Nideggen circuit around the ancient castle of the town of Nürburg, following the examples of Italy's Monza and Targa Florio courses, and Berlin's AVUS, yet with a different character. The layout of the circuit in the mountains was similar to the Targa Florio event, one of the most important motor races at that time. The original Nürburgring was to be a showcase for German automotive engineering and racing talent. Construction of the track, designed by the "Eichler Architekturbüro" from Ravensburg (led by architect Gustav Eichler), began in September 1925. The track was completed in spring of 1927, and the ADAC Eifelrennen races were continued there. The first races to take place on 18 June 1927 showed motorcycles and sidecars. The first motorcycle race was won by Toni Ulmen on an English 350 cc Velocette. The cars followed a day later, and Rudolf Caracciola was the winner of the over 5000 cc class in a Mercedes-Benz Compressor. In addition, the track was opened to the public in the evenings and on weekends, as a one-way toll road. The whole track consisted of 174 bends (prior to 1971 changes), and averaged in width. The fastest time ever around the full "Gesamtstrecke" was by Louis Chiron, at an average speed of 112.31 km/h (72 mph) in his Bugatti. In 1929 the full Nürburgring was used for the last time in major racing events, as future Grands Prix would be held only on the "Nordschleife". Motorcycles and minor races primarily used the shorter and safer "Südschleife". Memorable pre-war races at the circuit featured the talents of early "Ringmeister" (Ringmasters) such as Rudolf Caracciola, Tazio Nuvolari and Bernd Rosemeyer. After World War II, racing resumed in 1947 and in 1951, the "Nordschleife" of the Nürburgring again became the main venue for the German Grand Prix as part of the Formula One World Championship (with the exception of 1959, when it was held on the AVUS in Berlin). A new group of "Ringmeister" arose to dominate the race – Alberto Ascari, Juan Manuel Fangio, Stirling Moss, Jim Clark, John Surtees, Jackie Stewart and Jacky Ickx. On 5 August 1961, during practice for the 1961 German Grand Prix, Phil Hill became the first person to complete a lap of the "Nordschleife" in under 9 minutes, with a lap of 8 minutes 55.2 seconds (153.4 km/h or 95.3 mph) in the Ferrari 156 "Sharknose" Formula One car. Over half a century later, even the highest performing road cars still have difficulty breaking 8 minutes without a professional race driver or one very familiar with the track. Also, several rounds of the German motorcycle Grand Prix were held, mostly on the "Südschleife", but the Hockenheimring and the Solitudering were the main sites for Grand Prix motorcycle racing. In 1953, the ADAC 1000 km Nürburgring race was introduced, an Endurance race and Sports car racing event that counted towards the World Sportscar Championship for decades. The 24 Hours Nürburgring for touring car racing was added in 1970. By the late 1960s, the "Nordschleife" and many other tracks were becoming increasingly dangerous for the latest generation of F1 cars. In 1967, a chicane was added before the start/finish straight, called "Hohenrain", in order to reduce speeds at the pit lane entry. This made the track longer. Even this change, however, was not enough to keep Stewart from nicknaming it "The Green Hell" following his victory in the 1968 German Grand Prix amid a driving rainstorm and thick fog. In 1970, after the fatal crash of Piers Courage at Zandvoort, the F1 drivers decided at the French Grand Prix to boycott the Nürburgring unless major changes were made, as they did at Spa the year before. The changes were not possible on short notice, and the German GP was moved to the Hockenheimring, which had already been modified. In accordance with the demands of the F1 drivers, the "Nordschleife" was reconstructed by taking out some bumps, smoothing out some sudden jumps (particularly at Brünnchen), and installing Armco safety barriers. The track was made straighter, following the race line, which reduced the number of corners. The German GP could be hosted at the Nürburgring again, and was for another six years from 1971 to 1976. In 1973 the entrance into the dangerous and bumpy Kallenhard corner was made slower by adding another left-hand corner after the fast Metzgesfeld sweeping corner. Safety was improved again later on, e.g. by removing the jumps on the long main straight and widening it, and taking away the bushes right next to the track at the main straight, which had made that section of the Nürburgring dangerously narrow. A second series of three more F1 races was held until 1976. However, primarily due to its length of over , and the lack of space due to its situation on the sides of the mountains, increasing demands by the F1 drivers and the FIA's CSI commission were too expensive or impossible to meet. For instance, by the 1970s the German Grand Prix required five times the marshals and medical staff as a typical F1 race, something the German organizers were unwilling to provide. Additionally, even with the 1971 modifications it was still possible for cars to become airborne off the track. The Nürburgring was also unsuitable for the burgeoning television market; its vast expanse made it almost impossible to effectively cover a race there. As a result, early in the season it was decided that the 1976 race would be the last to be held on the old circuit. Niki Lauda, the reigning world champion and only person ever to lap the full "Nordschleife" in under seven minutes (6:58.6, 1975), proposed to the other drivers that they boycott the circuit in 1976. Lauda was not only concerned about the safety arrangements and the lack of marshals around the circuit, he also did not like the prospect of running the race in another rainstorm. Usually when that happened, some parts of the circuit were wet and other parts were dry, which is what the conditions of the circuit were for that race. The other drivers voted against the idea and the race went ahead. Lauda crashed in his Ferrari coming out of the left-hand kink before Bergwerk after a new magnesium component on his Ferrari's rear suspension failed. He was badly burned as his car was still loaded with fuel in lap 2. Lauda was saved by the combined actions of fellow drivers Arturo Merzario, Guy Edwards, Brett Lunger, and Harald Ertl, rather than by the ill-equipped track marshals. The crash also showed that the track's distances were too long for regular fire engines and ambulances, even though the "ONS-Staffel" was equipped with a Porsche 911 rescue car, marked (R). The old Nürburgring never hosted another F1 race again, as the German Grand Prix was moved to the Hockenheimring for 1977. The German motorcycle Grand Prix was held for the last time on the old Nürburgring in 1980, also permanently moving to Hockenheim. By its very nature, the "Nordschleife" was impossible to make safe in its old configuration. It soon became apparent that it would have to be completely overhauled if there was any prospect of Formula One returning there- the Nurburgring's administration and race organizers were not willing to provide the enormous expense of providing the amount of marshals needed for a Grand Prix- up to 6 times the amount that most other circuits needed. With this in mind, in 1981 work began on a -long new circuit, which was built on and around the old pit area. At the same time, a bypass shortened the "Nordschleife" to , and with an additional small pit lane, this version was used for races in 1983, e.g. the 1000km Nürburgring endurance race, while construction work was going on nearby. During qualifying for that race, the late Stefan Bellof set a lap of 6:11.13 for the "Nordschleife" in his Porsche 956, or on average. This lap held the all-time record for 35 years (partially because no major racing has taken place there since 1984) until it was surpassed by Timo Bernhard in the Porsche 919 Hybrid Evo, which ran the slightly longer version of the circuit in 5:19.546- averaging on June 29, 2018. Meanwhile, more run-off areas were added at corners like Aremberg and Brünnchen, where originally there were just embankments protected by Armco barriers. The track surface was made safer in some spots where there had been nasty bumps and jumps. Racing line markers were added to the corners all around the track as well. Also, bushes and hedges at the edges of corners were taken out and replaced with Armco and grass. The former "Südschleife" had not been modified in 1970/71 and was abandoned a few years later in favour of the improved "Nordschleife". It is now mostly gone (in part due to the construction of the new circuit) or converted to a normal public road, but since 2005 a vintage car event has been hosted on the old track layout, including part of the parking area. The new track was completed in 1984 and named "GP-Strecke" (: literally, ""Grand Prix Course""). It was built to meet the highest safety standards. However, it was considered in character a mere shadow of its older sibling. Some fans, who had to sit much farther away from the track, called it "Eifelring", "Ersatzring", "Grünering" or similar nicknames, believing it did not deserve to be called Nürburgring. Like many circuits of the time, it offered few overtaking opportunities. Prior to the 2013 German Grand Prix both Mark Webber and Lewis Hamilton said they like the track. Webber described the layout as "an old school track" before adding, "It’s a beautiful little circuit for us to still drive on so I think all the guys enjoy driving here." While Hamilton said "It’s a fantastic circuit, one of the classics and it hasn’t lost that feel of an old classic circuit." To celebrate its opening, an exhibition race was held, on 12 May, featuring an array of notable drivers. Driving identical Mercedes 190E 2.3–16's, the line-up was Elio de Angelis, Jack Brabham (Formula 1 World Champion 1959, 1960, 1966), Phil Hill (1961), Denis Hulme (1967), James Hunt (1976), Alan Jones (1980), Jacques Laffite, Niki Lauda (1975, 1977)*, Stirling Moss, Alain Prost*, Carlos Reutemann, Keke Rosberg (1982), Jody Scheckter (1979), Ayrton Senna*, John Surtees (1964) and John Watson. Senna won ahead of Lauda, Reutemann, Rosberg, Watson, Hulme and Jody Scheckter, being the only one to resist Lauda's performance who – having missed the qualifying – had to start from the last row and overtook all the others except Senna. There were nine former and two future Formula 1 World Champions competing, in a field of 20 cars with 16 Formula 1 drivers; the other four were local drivers: Klaus Ludwig, Manfred Schurti, Udo Schütz and Hans Herrmann. Besides other major international events, the Nürburgring has seen the brief return of Formula One racing, as the 1984 European Grand Prix was held at the track, followed by the 1985 German Grand Prix. As F1 did not stay, other events were the highlights at the new Nürburgring, including the 1000km Nürburgring, DTM, motorcycles, and newer types of events, like truck racing, vintage car racing at the AvD "Oldtimer Grand Prix", and even the "Rock am Ring" concerts. Following the success and first world championship of Michael Schumacher, a second German F1 race was held at the Nürburgring between 1995 and 2006, called the European Grand Prix, or in 1997 and 1998, the Luxembourg Grand Prix. For 2002, the track was changed, by replacing the former "Castrol-chicane" at the end of the start/finish straight with a sharp right-hander (nicknamed "Haug-Hook"), in order to create an overtaking opportunity. Also, a slow Omega-shaped section was inserted, on the site of the former kart track. This extended the GP track from , while at the same time, the Hockenheimring was shortened from . Both the Nürburgring and the Hockenheimring events have been losing money due to high and rising Formula One license fees charged by Bernie Ecclestone and low attendance due to high ticket prices; starting with the 2007 Formula One season, Hockenheim and Nürburgring will alternate for hosting of the German GP. In Formula One, Ralf Schumacher collided with his teammate Giancarlo Fisichella and his brother at the start of the 1997 race which was won by Jacques Villeneuve. In 1999, in changing conditions, Johnny Herbert managed to score the only win for the team of former "Ringmeister" Jackie Stewart. One of the highlights of the 2005 season was Kimi Räikkönen's spectacular exit while in the last lap of the race, when his suspension gave way after being rattled lap after lap by a flat-spotted tyre that was not changed due to the short-lived rule. Prior to the 2007 European Grand Prix, the "Audi S" (turns 8 and 9) was renamed "Michael Schumacher S" after Michael Schumacher. Schumacher had retired from Formula One the year before, but returned in 2010, and in 2011 became the second Formula One driver to drive through a turn named after them (after Ayrton Senna driving his "S for Senna" at Autódromo José Carlos Pace). In 2007, the FIA announced that Hockenheimring and Nürburgring would alternate with the German Grand Prix with Nürburgring hosting in 2007. Due to name-licensing problems, it was held as the European Grand Prix that year. In 2014, the new owners of the Nürburgring were unable to secure a deal to continue hosting the German Grand Prix in the odd numbered years, so the 2015 and 2017 German Grands Prix were cancelled. While it is unusual for deaths to occur during sanctioned races, there are many accidents and several deaths each year during public sessions. It is common for the track to be closed several times a day for cleanup, repair, and medical intervention. While track management does not publish any official figures, several regular visitors to the track have used police reports to estimate the number of fatalities at somewhere between 3 and 12 in a full year. Jeremy Clarkson noted in "Top Gear" in 2004 that "over the years this track has claimed over 200 lives". Several touring car series still compete on the "Nordschleife", using either only the simple version with its separate small pit lane, or a combined -long track that uses a part of the original modern F1 track (without the Mercedes Arena section, which is often used for support pits) plus its huge pit facilities. Entry-level competition requires a regularity test (GLP) for street-legal cars. Two racing series (RCN/CHC and VLN) compete on 15 Saturdays each year, for several hours. The annual highlight is the 24 Hours Nürburgring weekend, held usually in mid-May, featuring 220 cars – from small cars to Turbo Porsches or factory race cars built by BMW, Opel, Audi, and Mercedes-Benz, over 700 drivers (amateurs and professionals), and up to 290,000 spectators. As of 2015 the World Touring Car Championship holds the FIA WTCC Race of Germany at the Nordschleife as a support category to the 24 Hours. Automotive media outlets and manufacturers use the "Nordschleife" as a standard to publish their lap times achieved with their production vehicles. BMW Sauber’s Nick Heidfeld made history on 28 April 2007 as the first driver in over thirty years to tackle the Nürburgring "Nordschleife" track in a contemporary Formula One car. Heidfeld’s three laps in an F1.06 were part of festivities celebrating BMW’s contribution to motorsport. About 45,000 spectators showed up for the main event, the third four-hour VLN race of the season. Conceived largely as a photo opportunity, the lap times were not as fast as the car was capable of, BMW instead choosing to run the chassis at a particularly high ride height to allow for the "Nordschleife"'s abrupt gradient changes and to limit maximum speeds accordingly. Former F1 driver Hans-Joachim Stuck was injured during the race when he crashed his BMW Z4. As part of the festivities before the 2013 24 Hours Nürburgring race, Michael Schumacher and other Mercedes-Benz drivers took part in a promotional event which saw Schumacher complete a demonstration lap of the Nordschleife at the wheel of a 2011 Mercedes W02. As with Heidfeld's lap, and also partly due to Formula One's strict in-season testing bans, the lap left many motorsport fans underwhelmed. Since its opening in 1927, the track has been used by the public for the so-called "Touristenfahrten", i.e. anyone with a road-legal car or motorcycle, as well as tour buses, motor homes, or cars with trailers. It is opened every day of the week, except when races take place. The track may be closed for weeks during the winter months, depending on weather conditions and maintenance work. Passing on the right is prohibited, and some sections have speed limits; the normal traffic rules (StVO in German) apply also here. The Nürburgring is a popular attraction for many driving enthusiasts and riders from all over the world, partly because of its history and the challenge it provides. The lack of oncoming traffic and intersections sets it apart from regular roads, and the absence of a blanket speed limit is a further attraction. Normal ticket buyers on tourist days cannot quite complete a full lap of the "Nordschleife", which bypasses the modern "GP-Strecke", as they are required to slow down and pass through a "pit lane" section where toll gates are installed. On busier days, a mobile ticket barrier is installed on the main straight in order to reduce the length of queues at the fixed barriers. This is open to all ticket holders. On rare occasions, it is possible to drive both the "Nordschleife" and the Grand Prix circuit combined. Drivers interested in lap times often time themselves from the first bridge after the barriers to the last gantry (aka Bridge-to-Gantry or BTG time) before the exit. However, the track's general conditions state that any form of racing, including speed record attempts, is forbidden. The driver's insurance coverage may consequently be voided, leaving the driver fully liable for damage. Normal, non-racing, non-timed driving accidents might be covered by driver's insurance, but it is increasingly common for UK insurers especially to insert exclusion clauses that mean drivers and riders on the Nürburgring only have third-party coverage or are not covered at all. Drivers who have crashed into the barriers, suffered mechanical failure or been otherwise required to be towed off track during "Touristenfahrten" sessions are referred to as having joined the "Bongard Club". This nickname is derived from the name of the company which operates the large yellow recovery flatbed trucks which ferry those unfortunate drivers and their vehicles to the nearest exit. Due to the high volume of traffic, there is an emphasis on quickly clearing and repairing any compromised safety measures so the track can be immediately re-opened for use. Additionally, those found responsible for damage to the track or safety barriers are required to pay for repairs, along with the time and cost associated with personnel and equipment to address those damages, making any accident or breakdown a potentially expensive incident. Because it is technically operated as a public toll road, failing to report an accident or instance where track surfaces are affected is considered unlawfully leaving the scene of an accident. This is all part of the rules and regulations which aim to ensure a safe experience for all visitors to the track. One of the original purposes of the "Nordschleife" was as a test track for auto manufacturers, and its demanding layout had been traditionally used as a proving ground. Weekdays are often booked for so-called "Industriefahrten" for auto makers and the media. With the advent of the Internet, awareness of the "Nordschleife" has risen in Germany and abroad, in addition to publicity in print media. In 1999, Porsche reported that their new 996 GT3 had lapped the Nürburgring in under eight minutes, and in subsequent years, manufacturers from overseas also showed up to test cars. Some high-performance models are promoted with videotaped laps published on the web, and the claimed lap times are generating discussion. Few of these supercars are actually entered in racing where the claims could be backed up. For sixteen weeks per year, the industry pool () rents exclusive daytime use of the track for automotive development, and endurance testing. the industry pool consisted of approximately 30 car manufacturers, associations, and component suppliers. By 2019, the track was being rented by the industry pool for 18 weeks per year. The TV Series "Top Gear" has also used the "Nordschleife" for its challenges, often involving Sabine Schmitz. In addition, during series 17 (summer 2011) of Top Gear, James May was very critical of the ride quality of cars whose development processes included testing on the "Nordschleife", saying that cars which were tested at Nordschleife got ruined. Multiple layouts of the Nürburgring have been featured in video games, such as the "Gran Turismo" series, the "Forza Motorsport" series, "iRacing" and "Assetto Corsa". "Grand Prix Legends", a historic racing simulator also included the Nürburgring on its roster of default Grand Prix tracks. Other pastimes are hosted at the Nürburgring, such as the Rock am Ring, Germany's biggest rock music festival, attracting close to 100,000 rock fans each year since 1985. Since 1978, the "Nordschleife" is also the venue of a major running event ("Nürburgring-Lauf/Run am Ring"). In 2003, a major bicycling event ("Rad am Ring") was added and it became the multi-sports event "Rad & Run am Ring". In 2009, new commercial areas opened, including a hotel and shopping mall. In the summer of 2009, ETF Ride Systems opened a new interactive dark ride application called "Motor Mania" at the racetrack, in collaboration with Lagotronics B.V. The roller coaster "ring°racer" was scheduled to open in 2011 but never started its operations due to technical failures. In 2012, the track was preparing to file for bankruptcy as a result of nearly $500 million in debts and the inability to secure financing. On 1 August 2012, the government of Rheinland-Pfalz guaranteed $312 million to allow the track to meet its debt obligations. In 2013, the Nürburgring was for sale for US$165 million (€127.3 million). The sale process was by sealed-bid auction with an expected completion date of "Late Summer". This meant there was to be a new owner in 2013, unencumbered by the debts of the previous operation, with the circuit expected to return to profitability. On March 11, 2014 it was reported that the Nürburgring was sold for 77 million euros ($106.8 million). Düsseldorf-based Capricorn Development was the buyer. The company was to take full ownership of the Nürburgring on January 1, 2015. But in October 2014, Russian billionaire, the chairman of Moscow-based Pharmstandard, Viktor Kharitonin, bought a majority stake in the Nürburgring. In May 2015, the Nürburgring was set to hold the first Grüne Hölle Rock festival as a replacement for the Rock am Ring festival, but the project did not take place. Grüne Hölle Rock has changed their name to Rock im Revier and will be held in the Schalke area. The "Nordschleife" operates in a clockwise direction, and was formerly known for its abundance of sharp crests, causing fast-moving, firmly-sprung racing cars to jump clear off the track surface at many locations. Although by no means the most fearsome, "Flugplatz" is perhaps the most aptly (although coincidentally) named and widely remembered. The name of this part of the track comes from a small airfield, which in the early years was located close to the track in this area. The track features a very short straight that climbs sharply uphill for a short time, then suddenly drops slightly downhill, and this is immediately followed by two very fast right-hand kinks. Chris Irwin's career was ended following a massive accident at "Flugplatz", in a Ford 3L GT sports car in 1968. Manfred Winkelhock flipped his March Formula Two car at the same corner in 1980. This section of the track was renovated in 2016 after an accident in which a Nissan GTR flew over the fence and killed a spectator. The "Flugplatz" is one of the most important parts of the Nürburgring because after the two very fast right handers comes what is possibly the fastest part of the track: a downhill straight called "Kottenborn", into a very fast curve called "Schwedenkreuz" (Swedish Cross). Drivers are flat out for some time here. Right before "Flugplatz" is "Quiddelbacher-Höhe" (peak, as in "mountain summit"), where the track crosses a bridge over the Bundesstraße 257. The "Fuchsröhre" is soon after the very fast downhill section succeeding the "Flugplatz". After negotiating a long right hand corner called "Aremberg" (which is after "Schwedenkreuz") the road goes slightly uphill, under a bridge and then it plunges downhill, and the road switches back left and right and finding a point of reference for the racing line is difficult. This whole sequence is flat out and then, the road climbs sharply uphill. The road then turns left and levels out at the same time; this is one of the many jumps of the Nürburgring where the car goes airborne. This leads to the "Adenauer Forst" (forest) turns. The "Fuchsröhre" is one of the fastest and most dangerous parts of the Nürburgring because of the extremely high speeds in such a tight and confined place; this part of the Nürburgring goes right through a forest and there is only about 7–8 feet of grass separating the track from Armco barrier, and beyond the barriers is a wall of trees. Perhaps the most notorious corner on the long circuit, "Bergwerk" has been responsible for some serious and sometimes fatal accidents. A tight right-hand corner, coming just after a long, fast section and a left-hand kink on a small crest, was where Carel Godin de Beaufort fatally crashed. The fast kink was also the scene of Niki Lauda's infamous fiery accident during the 1976 German Grand Prix. This left kink is often referred to as the "Lauda Links" (Lauda left). The "Bergwerk", along with the "Breidscheid"/"Adenauer" Bridge corners before it, are one of the series of corners that make or break one's lap time around the Nürburgring because of the fast, lengthy uphill section called "Kesselchen" (Little Valley) that comes after the "Bergwerk". Although being one of the slower corners on the "Nordschleife", the "Karussell" is perhaps its most famous and one of its most iconic- it is one of two berm-style, banked corners on the track. Soon after the driver has negotiated the long uphill section after "Bergwerk" and gone through a section called "Klostertal" (Monastery Valley), the driver turns right through a long hairpin, past an abandoned section called "Steilstrecke" (Steep Route) and then goes up another hill towards the "Karrusell". The entrance to the corner is blind, although Juan Manuel Fangio is reputed to have advised a young driver to "aim for the tallest tree," a feature that was also built into the rendering of the circuit in the "Gran Turismo 4" and "Grand Prix Legends" video games. Once the driver has reached the top of the hill, the road then becomes sharply banked on one side and level on the other- this banking drops off, rather than climbing up like most bankings on circuits. The sharply banked side has a concrete surface, and there is a foot-wide tarmac surface on the bottom of the banking for cars to get extra grip through the very rough concrete banking. Cars drop into the concrete banking, and keep the car in the corner (which is 210 degrees, much like a hairpin bend) until the road levels out and the concrete surface becomes tarmac again. This corner is very hard on the driver's wrists and hands because of the prolonged bumpy cornering the driver must do while in the "Karrusell". Usually cars come out of the top of the end of the banking to hit the apex that comes right after the end of the "Karrusell". The combination of a recognisable corner, slow-moving cars, and the variation in viewing angle as cars rotate around the banking, means that this is one of the circuit's most popular locations for photographers. It is named after German pre-WWII racing driver Rudolf Caracciola, who reportedly made the corner his own by hooking the inside tires into a drainage ditch to help his car "hug" the curve. As more concrete was uncovered and more competitors copied him, the trend took hold. At a later reconstruction, the corner was remade with real concrete banking, as it remains to this day. Shortly after the "Karussell" is a steep section, with gradients in excess of 16%, leading to a right-hander called "Hohe Acht", which is some 300 m higher in altitude than "Breidscheid". A favourite spectator vantage point, the "Brünnchen" section is composed of two right-hand corners and a very short straight. The first corner goes sharply downhill and the next, after the very short downhill straight, goes uphill slightly. This is a section of the track where on public days, accidents happen particularly at the blind uphill right-hand corner. Like almost every corner at the Nürburgring, both right-handers are blind. The short straight used to have a steep and sudden drop-off that caused cars to take off and a bridge that went over a pathway; these were taken out and smoothed over when the circuit was rebuilt in 1970 and 1971. The "Pflanzgarten", which is soon after the "Brünnchen", is one of the fastest, trickiest and most difficult sections of the Nürburgring. It is full of jumps, including two huge ones, one of which is called "Sprunghügel" (Hill Jump). This very complex section is unique in that it is made up of two different sections; getting the entire "Pflanzgarten" right is crucial to a good lap time around the Nürburgring. This section was the scene of Briton Peter Collins's fatal accident during the 1958 German Grand Prix, and the scene of a number of career-ending accidents in Formula One in the 1970s —Britons Mike Hailwood and Ian Ashley were two victims of the Pflanzgarten. "Pflanzgarten 1" is made up of a slightly banked, downhill left hander which then suddenly switches back left, then right. Then immediately, giving the driver nearly no time to react (knowledge of this section is key) the road drops away twice: the first jump is only slight, then right after (somewhat like a staircase) the road drops away very sharply which usually causes almost all cars to go airborne at this jump; the drop is so sudden. Then, immediately after the road levels out very shortly after the jump and the car touches the ground again, the road immediately and suddenly goes right very quickly and then right again; this is what makes up the end of the first "Pflanzgarten"- a very fast multiple apex sequence of right hand corners. The road then goes slightly uphill and then through another jump; the road suddenly drops away and levels out and at the same time, the road turns through a flat-out left hander. Then, the road drops away again very suddenly, which is the second huge jump of the "Pflanzgarten" known as the "Sprunghügel". The road then goes downhill then quickly levels out, then it goes through a flat-out right hander and this starts the "Stefan Bellof S" (named as such because Bellof crashed a Porsche 956 there during the 1983 Nurburgring 1000 km), which was known as "Pflanzgarten 2" prior to 2013. The "Stefan Bellof S" is very tricky because the road quickly switches back left and right—a car is going so fast through here that it is like walking on a tightrope. It is very difficult to find the racing line here because the curves come up so quickly, so it is hard to find any point of reference. Then, after a jump at the end of the switchback section, it goes through a flat-out, top gear right hander and into a short straight that leads into two very fast curves called the "Schwalbenschwanz" (Swallow's Tail). The room for error on every part of the consistently high-speed "Pflanzgarten" and the "Stefan Bellof S" is virtually non-existent (much like the entire track itself). The road and the surface of the "Pflanzgarten" and the "Stefan Bellof S" moves around unpredictably; knowledge of this section is key to getting through cleanly. The "Schwalbenschwanz" is a sequence of very fast sweepers located after the "Stefan Bellof S". After a short straight, there is a very fast right hand sweeper that progressively goes uphill, and this leads into a blind left-hander that is a bit slower. The apex is completely blind, and the corner then changes gradient a bit; it goes up then down, which leads into a short straight that ends at the "Kleines Karussell". Originally, this part had a bridge that went over a stream and was very bumpy; this bridge was taken out and replaced with a culvert (large industrial pipe) so that the road could be smoothed over. The "Kleines Karussell" is similar to its bigger brother, except that it is a 90 degree corner instead of 210 degrees, and is faster and slightly less banked. Once this part of the track is dealt with, the drivers are near the end of the lap; with two more corners to negotiate before the 2.135 km long "Döttinger Höhe" straight. Lap times recorded on the Nürburgring "Nordschleife" are published by several manufacturers. They are published and discussed in print media, and online. The Nürburgring is known for its changeable weather. The near-fatal accident of Niki Lauda in 1976 was accompanied by poor weather conditions and also the 2007 Grand Prix race saw an early deluge take several cars out through aquaplaning, with Vitantonio Liuzzi making a lucky escape, hitting a retrieving truck with the rear wing first, rather than the fatal accident that befell Jules Bianchi seven years later at Suzuka. In spite of this reputation, the Nürburg weather station only recorded an average of between 1981–2010. Contrasting this, the relatively nearby Ardennes racetrack of Spa-Francorchamps in Wallonia, Belgium has a much rainier climate, seen by data from the village hosting the track called Stavelot and the village of Malmedy that circuit passes by. Nürburg has a semi-continental climate with both oceanic and continental tendencies. It does however land in the former category (Köppen "Cfb"). Due to the Nordschleife's varied terrain and elevation, weather may be completely different on either end of the track. The elevation shift also makes thermal differences a strong possibility. The modern Grand Prix circuit also has sizeable elevation changes between the start-finish straight and the lowest point on the opposite end of the track, but the geographical distance and actual elevation gain between the two are lower. Annual sunshine is in the 1500s, which is low by European standards, but only slightly gloomier than the nearest large city of Cologne located on a plain. Contrasting that, Nürburg has cooler weather year round due to the higher elevation of the Eifel Mountains than the Rhine Valley.
https://en.wikipedia.org/wiki?curid=22122
Noun class In linguistics, a noun class is a particular category of nouns. A noun may belong to a given class because of the characteristic features of its referent, such as gender, animacy, shape, but such designations are often clearly conventional. Some authors use the term "grammatical gender" as a synonym of "noun class", but others consider these different concepts. Noun classes should not be confused with noun classifiers. There are three main ways by which natural languages categorize nouns into noun classes: Usually, a combination of the three types of criteria is used, though one is more prevalent. Noun classes form a system of grammatical agreement. A noun in a given class may require: In Modern English, countable and uncountable nouns are distinguished by the choice of "many"/"much". The choice between the relative pronoun "who" (persons) and "which" (non-persons) may also be considered a form of agreement with a semantic noun class. A few nouns also exhibit vestigial noun classes, such as "stewardess", where the suffix "-ess" added to "steward" denotes a female person. This type of noun affixation is not very frequent in English, but quite common in languages which have the true grammatical gender, including most of the Indo-European family, to which English belongs. When noun class is expressed on other parts of speech in addition to nouns and pronouns, the language is said to have grammatical gender. In languages without inflectional noun classes, nouns may still be extensively categorized by independent particles called noun classifiers. Common criteria that define noun classes include: See Swahili for the semantic motivations for an elaborate noun-class system. The Ojibwe language and other members of the Algonquian languages distinguish between animate and inanimate classes. Some sources argue that the distinction is between things which are powerful and things which are not. All living things, as well as sacred things and things connected to the Earth are considered powerful and belong to the animate class. Still, the assignment is somewhat arbitrary, as "raspberry" is animate, but "strawberry" is inanimate. In Navajo (Southern Athabaskan) nouns are classified according to their animacy, shape, and consistency. Morphologically, however, the distinctions are not expressed on the nouns themselves, but on the verbs of which the nouns are the subject or direct object. For example, in the sentence ' "My shirt is lying on the bed", the verb "lies" is used because the subject ' "my shirt" is a flat, flexible object. In the sentence "My belt is lying on the bed", the verb ' "lies" is used because the subject ' "my belt" is a slender, flexible object. Koyukon (Northern Athabaskan) has a more intricate system of classification. Like Navajo, it has classificatory verb stems that classify nouns according to animacy, shape, and consistency. However, in addition to these verb stems, Koyukon verbs have what are called "gender prefixes" that further classify nouns. That is, Koyukon has two different systems that classify nouns: (a) a classificatory verb system and (b) a gender system. To illustrate, the verb stem "-tonh" is used for enclosed objects. When "-tonh" is combined with different gender prefixes, it can result in "daaltonh" which refers to objects enclosed in boxes or "etltonh" which refers to objects enclosed in bags. The Dyirbal language is well known for its system of four noun classes, which tend to be divided along the following semantic lines: The class usually labeled "feminine", for instance, includes the word for fire and nouns relating to fire, as well as all dangerous creatures and phenomena. (This inspired the title of the George Lakoff book "Women, Fire, and Dangerous Things".) The Ngangikurrunggurr language has noun classes reserved for canines and hunting weapons. The Anindilyakwa language has a noun class for things that reflect light. The Diyari language distinguishes only between female and other objects. Perhaps the most noun classes in any Australian language are found in Yanyuwa, which has 16 noun classes, including nouns associated with food, trees and abstractions, in addition to separate classes for men and masculine things, women and feminine things. In the men's dialect, the classes for men and for masculine things have simplified to a single class, marked the same way as the women's dialect marker reserved exclusively for men. In Basque there are two classes, animate and inanimate; however, the only difference is in the declension of locative cases (inessive, locative genitive, allative, terminal allative, ablative and directional ablative). There are a few words with both masculine and feminine forms, generally words for relatives (cousin: lehengusu (m)/lehengusina (f)) or words borrowed from Latin ("king": "errege", from the Latin word "rex"; "queen": "erregina", from "regina"). In names for familiar relatives, where both genders are taken into account, either the words for each gender are put together ("son": "seme"; "daughter": "alaba"; "children"(meaning son(s) and daughter(s)): "seme-alaba(k)") or there is a noun that includes both: "father": "aita"; "mother": "ama"; "parent": "guraso". Some members of the Northwest Caucasian family, and almost all of the Northeast Caucasian languages, manifest noun class. In the Northeast Caucasian family, only Lezgian, Udi, and Aghul do not have noun classes. Some languages have only two classes, whereas Bats has eight. The most widespread system, however, has four classes: male, female, animate beings and certain objects, and finally a class for the remaining nouns. The Andi language has a noun class reserved for insects. Among Northwest Caucasian languages, only Abkhaz and Abaza have noun class, making use of a human male/human female/non-human distinction. In all Caucasian languages that manifest class, it is not marked on the noun itself but on the dependent verbs, adjectives, pronouns and prepositions. Atlantic–Congo languages can have ten or more noun classes, defined according to non-sexual criteria. Certain nominal classes are reserved for humans. The Fula language has about 26 noun classes (exact number varies slightly by dialect). According to Steven Pinker, the Kivunjo language has 16 noun classes including classes for precise locations and for general locales, classes for clusters or pairs of objects and classes for the objects that come in pairs or clusters, and classes for abstract qualities. According to Carl Meinhof, the Bantu languages have a total of 22 noun classes called nominal classes (this notion was introduced by W. H. J. Bleek). While no single language is known to express all of them, most of them have at least 10 noun classes. For example, by Meinhof's numbering, Shona has 20 classes, Swahili has 15, Sotho has 18 and Ganda has 17. Additionally, there are polyplural noun classes. A polyplural noun class is a plural class for more than one singular class. For example, Proto-Bantu class 10 contains plurals of class 9 nouns and class 11 nouns, while class 6 contains plurals of class 5 nouns and class 15 nouns. Classes 6 and 10 are inherited as polyplural classes by most surviving Bantu languages, but many languages have developed new polyplural classes that are not widely shared by other languages. Specialists in Bantu emphasize that there is a clear difference between genders (such as known from Afro-Asiatic and Indo-European) and nominal classes (such as known from Niger–Congo). Languages with nominal classes divide nouns formally on the base of hyperonymic meanings. The category of nominal class replaces not only the category of gender, but also the categories of number and case. Critics of the Meinhof's approach notice that his numbering system of nominal classes counts singular and plural numbers of the same noun as belonging to separate classes. This seems to them to be inconsistent with the way other languages are traditionally considered, where number is orthogonal to gender (according to the critics, a Meinhof-style analysis would give Ancient Greek 9 genders). If one follows broader linguistic tradition and counts singular and plural as belonging to the same class, then Swahili has 8 or 9 noun classes, Sotho has 11 and Ganda has 10. The Meinhof numbering tends to be used in scientific works dealing with comparisons of different Bantu languages. For instance, in Swahili the word "rafiki" ‘friend’ belongs to the class 9 and its "plural form" is "marafiki" of the class 6, even if most nouns of the 9 class have the plural of the class 10. For this reason, noun classes are often referred to by combining their singular and plural forms, e.g., "rafiki" would be classified as "9/6", indicating that it takes class 9 in the singular, and class 6 in the plural. However not all Bantu languages have these exceptions. In Ganda each singular class has a corresponding plural class (apart from one class which has no singular–plural distinction; also some plural classes correspond to more than one singular class) and there are no exceptions as there are in Swahili. For this reason Ganda linguists use the orthogonal numbering system when discussing Ganda grammar (other than in the context of Bantu comparative linguistics), giving the 10 traditional noun classes of that language. The distinction between genders and nominal classes is blurred still further by Indo-European languages that have nouns that behave like Swahili's "rafiki". Italian, for example, has a group of nouns deriving from Latin neuter nouns that acts as masculine in the singular but feminine in the plural: "il braccio"/"le braccia"; "l'uovo"/"le uova". (These nouns are still placed in a neuter gender of their own by some grammarians.) Here is a complete list of nominal classes in Swahili: "Ø-" means no prefix. Some classes are homonymous (esp. 9 and 10). The Proto-Bantu class 12 disappeared in Swahili, class 13 merged with 7, and 14 with 11. Class prefixes appear also on adjectives and verbs, e.g.: The class markers which appear on the adjectives and verbs may differ from the noun prefixes: In this example, the verbal prefix a- and the pronominal prefix wa- are in concordance with the noun prefix m-: they all express class 1 despite of their different forms. The Zande language distinguishes four noun classes: There are about 80 inanimate nouns which are in the animate class, including nouns denoting heavenly objects (moon, rainbow), metal objects (hammer, ring), edible plants (sweet potato, pea), and non-metallic objects (whistle, ball). Many of the exceptions have a round shape, and some can be explained by the role they play in Zande mythology. The term gender, as used by some linguists, refers to a noun-class system composed with 2, 3, or 4 classes, particularly if the classification is semantically based on a distinction between masculine and feminine. Genders are then considered a sub-class of noun classes. Not all linguists recognize a distinction between noun-classes and genders, however, and instead use either the term "gender" or "noun class" for both. Some languages, such as Japanese, Chinese and the Tai languages, have elaborate systems of particles that go with nouns based on shape and function, but are free morphemes rather than affixes. Because the classes defined by these classifying words are not generally distinguished in other contexts, there are many linguists who take the view that they do not create noun classes.
https://en.wikipedia.org/wiki?curid=22130
Natural gas Natural gas (also called fossil gas; sometimes just gas), is a naturally occurring hydrocarbon gas mixture consisting primarily of methane, but commonly including varying amounts of other higher alkanes, and sometimes a small percentage of carbon dioxide, nitrogen, hydrogen sulfide, or helium. It is formed when layers of decomposing plant and animal matter are exposed to intense heat and pressure under the surface of the Earth over millions of years. The energy that the plants originally obtained from the sun is stored in the form of chemical bonds in the gas. Natural gas is a non-renewable hydrocarbon used as a source of energy for heating, cooking, and electricity generation. It is also used as a fuel for vehicles and as a chemical feedstock in the manufacture of plastics and other commercially important organic chemicals. Natural gas has an effect on climate change; it is a greenhouse gas itself, and creates carbon dioxide during oxidation. Natural gas can be used in place of coal, for example, to generate electricity, which may lower net carbon dioxide emissions in absence of carbon capture. Natural gas is found in deep underground rock formations or associated with other hydrocarbon reservoirs in coal beds and as methane clathrates. Petroleum is another resource and fossil fuel found close to and with natural gas. Most natural gas was created over time by two mechanisms: biogenic and thermogenic. Biogenic gas is created by methanogenic organisms in marshes, bogs, landfills, and shallow sediments. Deeper in the earth, at greater temperature and pressure, thermogenic gas is created from buried organic material. In petroleum production, gas is sometimes burned as flare gas. Before natural gas can be used as a fuel, most, but not all, must be processed to remove impurities, including water, to meet the specifications of marketable natural gas. The by-products of this processing include ethane, propane, butanes, pentanes, and higher molecular weight hydrocarbons, hydrogen sulfide (which may be converted into pure sulfur), carbon dioxide, water vapor, and sometimes helium and nitrogen. Natural gas is sometimes informally referred to simply as "gas", especially when it is being compared to other energy sources, such as oil or coal. However, it is not to be confused with gasoline, which is often shortened in colloquial usage to "gas", especially in North America. Natural gas was discovered accidentally in ancient China, as it resulted from the drilling for brines. Natural gas was first used by the Chinese in about 500 BC (possibly even 1000 BC). They discovered a way to transport gas seeping from the ground in crude pipelines of bamboo to where it was used to boil salt water to extract the salt, in the Ziliujing District of Sichuan. The discovery and identification of natural gas in the Americas happened in 1626. In 1821, William Hart successfully dug the first natural gas well at Fredonia, New York, United States, which led to the formation of the Fredonia Gas Light Company. The city of Philadelphia created the first municipally owned natural gas distribution venture in 1836. By 2009, 66 000 km³ (or 8%) had been used out of the total 850 000 km³ of estimated remaining recoverable reserves of natural gas. Based on an estimated 2015 world consumption rate of about 3400 km³ of gas per year, the total estimated remaining economically recoverable reserves of natural gas would last 250 years at current consumption rates. An annual increase in usage of 2–3% could result in currently recoverable reserves lasting significantly less, perhaps as few as 80 to 100 years. In the 19th century, natural gas was primarily obtained as a by-product of producing oil. The small, light gas carbon chains came out of solution as the extracted fluids underwent pressure reduction from the reservoir to the surface, similar to uncapping a soft drink bottle where the carbon dioxide effervesces. The gas was often viewed as a by-product, a hazard, and a disposal problem in active oil fields. The large volumes produced could not be utilized until relatively expensive pipeline and storage facilities were constructed to deliver the gas to consumer markets. Until the early part of the 20th century, most natural gas associated with oil was either simply released or burned off at oil fields. Gas venting and production flaring are still practised in modern times, but efforts are ongoing around the world to retire them, and to replace them with other commercially viable and useful alternatives. Unwanted gas (or stranded gas without a market) is often returned to the reservoir with 'injection' wells while awaiting a possible future market or to re-pressurize the formation, which can enhance oil extraction rates from other wells. In regions with a high natural gas demand (such as the US), pipelines are constructed when it is economically feasible to transport gas from a wellsite to an end consumer. In addition to transporting gas via pipelines for use in power generation, other end uses for natural gas include export as liquefied natural gas (LNG) or conversion of natural gas into other liquid products via gas to liquids (GTL) technologies. GTL technologies can convert natural gas into liquids products such as gasoline, diesel or jet fuel. A variety of GTL technologies have been developed, including Fischer–Tropsch (F–T), methanol to gasoline (MTG) and syngas to gasoline plus (STG+). F–T produces a synthetic crude that can be further refined into finished products, while MTG can produce synthetic gasoline from natural gas. STG+ can produce drop-in gasoline, diesel, jet fuel and aromatic chemicals directly from natural gas via a single-loop process. In 2011, Royal Dutch Shell's per day F–T plant went into operation in Qatar. Natural gas can be "associated" (found in oil fields), or "non-associated" (isolated in natural gas fields), and is also found in coal beds (as coalbed methane). It sometimes contains a significant amount of ethane, propane, butane, and pentane—heavier hydrocarbons removed for commercial use prior to the methane being sold as a consumer fuel or chemical plant feedstock. Non-hydrocarbons such as carbon dioxide, nitrogen, helium (rarely), and hydrogen sulfide must also be removed before the natural gas can be transported. Natural gas extracted from oil wells is called casinghead gas (whether or not truly produced up the annulus and through a casinghead outlet) or associated gas. The natural gas industry is extracting an increasing quantity of gas from challenging resource types: sour gas, tight gas, shale gas, and coalbed methane. There is some disagreement on which country has the largest proven gas reserves. Sources that consider that Russia has by far the largest proven reserves include the US CIA (47 600 km³), the US Energy Information Administration (47 800 km³), and OPEC (48 700 km³). However, BP credits Russia with only 32 900 km³, which would place it in second place, slightly behind Iran (33 100 to 33 800 km³, depending on the source). With Gazprom, Russia is frequently the world's largest natural gas extractor. Major proven resources (in cubic kilometers) are world 187 300 (2013), Iran 33 600 (2013), Russia 32 900 (2013), Qatar 25 100 (2013), Turkmenistan 17 500 (2013) and the United States 8500 (2013). It is estimated that there are about 900 000 km³ of "unconventional" gas such as shale gas, of which 180 000 km³ may be recoverable. In turn, many studies from MIT, Black & Veatch and the DOE predict that natural gas will account for a larger portion of electricity generation and heat in the future. The world's largest gas field is the offshore South Pars / North Dome Gas-Condensate field, shared between Iran and Qatar. It is estimated to have of natural gas and of natural gas condensates. Because natural gas is not a pure product, as the reservoir pressure drops when non-associated gas is extracted from a field under supercritical (pressure/temperature) conditions, the higher molecular weight components may partially condense upon isothermic depressurizing—an effect called retrograde condensation. The liquid thus formed may get trapped as the pores of the gas reservoir get depleted. One method to deal with this problem is to re-inject dried gas free of condensate to maintain the underground pressure and to allow re-evaporation and extraction of condensates. More frequently, the liquid condenses at the surface, and one of the tasks of the gas plant is to collect this condensate. The resulting liquid is called natural gas liquid (NGL) and has commercial value. Shale gas is natural gas produced from shale. Because shale has matrix permeability too low to allow gas to flow in economical quantities, shale gas wells depend on fractures to allow the gas to flow. Early shale gas wells depended on natural fractures through which gas flowed; almost all shale gas wells today require fractures artificially created by hydraulic fracturing. Since 2000, shale gas has become a major source of natural gas in the United States and Canada. Because of increased shale gas production the United States was in 2014 the number one natural gas producer in the world. Following the increased production in the United States, shale gas exploration is beginning in countries such as Poland, China, and South Africa. Town gas is a flammable gaseous fuel made by the destructive distillation of coal. It contains a variety of calorific gases including hydrogen, carbon monoxide, methane, and other volatile hydrocarbons, together with small quantities of non-calorific gases such as carbon dioxide and nitrogen, and is used in a similar way to natural gas. This is a historical technology and is not usually economically competitive with other sources of fuel gas today. Most town "gashouses" located in the eastern US in the late 19th and early 20th centuries were simple by-product coke ovens that heated bituminous coal in air-tight chambers. The gas driven off from the coal was collected and distributed through networks of pipes to residences and other buildings where it was used for cooking and lighting. (Gas heating did not come into widespread use until the last half of the 20th century.) The coal tar (or asphalt) that collected in the bottoms of the gashouse ovens was often used for roofing and other waterproofing purposes, and when mixed with sand and gravel was used for paving streets. Methanogenic "Archaea" are responsible for almost all biological sources of methane, though methylphosphonate-degrading "Bacteria" produce an as-yet not fully quantified fraction of biogenic methane, particularly in the oceans. Some live in symbiotic relationships with other life forms, including termites, ruminants, and cultivated crops. Other sources of methane, the principal component of natural gas, include landfill gas, biogas, and methane hydrate. When methane-rich gases are produced by the anaerobic decay of organic matter (biomass), these are referred to as biogas (or natural biogas). Sources of biogas include swamps, marshes, and landfills, as well as agricultural waste materials such as sewage sludge and manure by way of anaerobic digesters, in addition to enteric fermentation, particularly in cattle. Landfill gas is created by decomposition of waste in landfill sites. Excluding water vapor, about half of landfill gas is methane and most of the rest is carbon dioxide, with small amounts of nitrogen, oxygen, and hydrogen, and variable trace amounts of hydrogen sulfide and siloxanes. If the gas is not removed, the pressure may get so high that it works its way to the surface, causing damage to the landfill structure, unpleasant odor, vegetation die-off, and an explosion hazard. The gas can be vented to the atmosphere, flared or burned to produce electricity or heat. Biogas can also be produced by separating organic materials from waste that otherwise goes to landfills. This method is more efficient than just capturing the landfill gas it produces. Anaerobic lagoons produce biogas from manure, while biogas reactors can be used for manure or plant parts. Like landfill gas, biogas is mostly methane and carbon dioxide, with small amounts of nitrogen, oxygen and hydrogen. However, with the exception of pesticides, there are usually lower levels of contaminants. Landfill gas cannot be distributed through utility natural gas pipelines unless it is cleaned up to less than 3% , and a few parts per million , because and corrode the pipelines. The presence of will lower the energy level of the gas below requirements for the pipeline. Siloxanes in the gas will form deposits in gas burners and need to be removed prior to entry into any gas distribution or transmission system. Consequently, it may be more economical to burn the gas on site or within a short distance of the landfill using a dedicated pipeline. Water vapor is often removed, even if the gas is burned on site. If low temperatures condense water out of the gas, siloxanes can be lowered as well because they tend to condense out with the water vapor. Other non-methane components may also be removed to meet emission standards, to prevent fouling of the equipment or for environmental considerations. Co-firing landfill gas with natural gas improves combustion, which lowers emissions. Biogas, and especially landfill gas, are already used in some areas, but their use could be greatly expanded. Systems have been established for use in parts of Hertfordshire, UK and Lyon in France. Using materials that would otherwise generate no income, or even cost money to get rid of, improves the profitability and energy balance of biogas production. Gas generated in sewage treatment plants is commonly used to generate electricity. For example, the Hyperion sewage plant in Los Angeles burns of gas per day to generate power New York City utilizes gas to run equipment in the sewage plants, to generate electricity, and in boilers. Using sewage gas to make electricity is not limited to large cities. The city of Bakersfield, California, uses cogeneration at its sewer plants. California has 242 sewage wastewater treatment plants, 74 of which have installed anaerobic digesters. The total biopower generation from the 74 plants is about 66 MW. Huge quantities of natural gas (primarily methane) exist in the form of hydrates under sediment on offshore continental shelves and on land in arctic regions that experience permafrost, such as those in Siberia. Hydrates require a combination of high pressure and low temperature to form. In 2010, the cost of extracting natural gas from crystallized natural gas was estimated to be as much as twice the cost of extracting natural gas from conventional sources, and even higher from offshore deposits. In 2013, Japan Oil, Gas and Metals National Corporation (JOGMEC) announced that they had recovered commercially relevant quantities of natural gas from methane hydrate. The image below is a schematic block flow diagram of a typical natural gas processing plant. It shows the various unit processes used to convert raw natural gas into sales gas pipelined to the end user markets. The block flow diagram also shows how processing of the raw natural gas yields byproduct sulfur, byproduct ethane, and natural gas liquids (NGL) propane, butanes and natural gasoline (denoted as pentanes +). As of 2018, natural gas production in the US has peaked twice, with current levels exceeding both previous peaks. It reached 24.1 Trillion cubic feet per year in 1973, followed by a decline, and reached 24.5 trillion cubic feet in 2001. After a brief drop, withdrawals have been increasing nearly every year since 2006, with 2017 production at 33.4 Trillion cubic feet. Because of its low density, it is not easy to store natural gas or to transport it by vehicle. Natural gas pipelines are impractical across oceans, since the gas needs to be cooled down and compressed, as the friction in the pipeline causes the gas to heat up. Many existing pipelines in America are close to reaching their capacity, prompting some politicians representing northern states to speak of potential shortages. The large trade cost implies that natural gas markets are globally much less integrated, causing significant price differences across countries. In Western Europe, the gas pipeline network is already dense. New pipelines are planned or under construction in Eastern Europe and between gas fields in Russia, Near East and Northern Africa and Western Europe. Whenever gas is bought or sold at custody transfer points, rules and agreements are made regarding the gas quality. These may include the maximum allowable concentration of , and . Usually sales quality gas that has been treated to remove contamination is traded on a "dry gas" basis and is required to be commercially free from objectionable odours, materials, and dust or other solid or liquid matter, waxes, gums and gum forming constituents, which might damage or adversely affect operation of equipment downstream of the custody transfer point. LNG carriers transport liquefied natural gas (LNG) across oceans, while tank trucks can carry liquefied or compressed natural gas (CNG) over shorter distances. Sea transport using CNG carrier ships that are now under development may be competitive with LNG transport in specific conditions. Gas is turned into liquid at a liquefaction plant, and is returned to gas form at regasification plant at the terminal. Shipborne regasification equipment is also used. LNG is the preferred form for long distance, high volume transportation of natural gas, whereas pipeline is preferred for transport for distances up to over land and approximately half that distance offshore. CNG is transported at high pressure, typically above . Compressors and decompression equipment are less capital intensive and may be economical in smaller unit sizes than liquefaction/regasification plants. Natural gas trucks and carriers may transport natural gas directly to end-users, or to distribution points such as pipelines. In the past, the natural gas which was recovered in the course of recovering petroleum could not be profitably sold, and was simply burned at the oil field in a process known as flaring. Flaring is now illegal in many countries. Additionally, higher demand in the last 20–30 years has made production of gas associated with oil economically viable. As a further option, the gas is now sometimes re-injected into the formation for enhanced oil recovery by pressure maintenance as well as miscible or immiscible flooding. Conservation, re-injection, or flaring of natural gas associated with oil is primarily dependent on proximity to markets (pipelines), and regulatory restrictions. Natural gas can be indirectly exported through the absorption in other physical output. A recent study suggests that the expansion of shale gas production in the US has caused prices to drop relative to other countries. This has caused a boom in energy intensive manufacturing sector exports, whereby the average dollar unit of US manufacturing exports has almost tripled its energy content between 1996 and 2012. A "master gas system" was invented in Saudi Arabia in the late 1970s, ending any necessity for flaring. Satellite observation, however, shows that flaring and venting are still practiced in some gas-extracting countries. Natural gas is used to generate electricity and heat for desalination. Similarly, some landfills that also discharge methane gases have been set up to capture the methane and generate electricity. Natural gas is often stored underground inside depleted gas reservoirs from previous gas wells, salt domes, or in tanks as liquefied natural gas. The gas is injected in a time of low demand and extracted when demand picks up. Storage nearby end users helps to meet volatile demands, but such storage may not always be practicable. With 15 countries accounting for 84% of the worldwide extraction, access to natural gas has become an important issue in international politics, and countries vie for control of pipelines. In the first decade of the 21st century, Gazprom, the state-owned energy company in Russia, engaged in disputes with Ukraine and Belarus over the price of natural gas, which have created concerns that gas deliveries to parts of Europe could be cut off for political reasons. The United States is preparing to export natural gas. Floating liquefied natural gas (FLNG) is an innovative technology designed to enable the development of offshore gas resources that would otherwise remain untapped due to environmental or economic factors which currently make them impractical to develop via a land-based LNG operation. FLNG technology also provides a number of environmental and economic advantages: Many gas and oil companies are considering the economic and environmental benefits of floating liquefied natural gas (FLNG). There are currently projects underway to construct five FLNG facilities. Petronas is close to completion on their FLNG-1 at Daewoo Shipbuilding and Marine Engineering and are underway on their FLNG-2 project at Samsung Heavy Industries. Shell Prelude is due to start production 2017. The Browse LNG project will commence FEED in 2019. Natural gas is primarily used in the northern hemisphere. North America and Europe are major consumers. Often well head gases require removal of various hydrocarbon molecules contained within the gas. Some of these gases include heptane, pentane, propane and other hydrocarbons with molecular weights above methane (). The natural gas transmission lines extend to the natural gas processing plant or unit which removes the higher molecular weighted hydrocarbons to produce natural gas with energy content between . The processed natural gas may then be used for residential, commercial and industrial uses. Natural gas flowing in the distribution lines is called mid-stream natural gas and is often used to power engines which rotate compressors. These compressors are required in the transmission line to pressurize and repressurize the mid-stream natural gas as the gas travels. Typically, natural gas powered engines require natural gas to operate at the rotational name plate specifications. Several methods are used to remove these higher molecular weighted gases for use by the natural gas engine. A few technologies are as follows: Natural gas is a major source of electricity generation through the use of cogeneration, gas turbines and steam turbines. Natural gas is also well suited for a combined use in association with renewable energy sources such as wind or solar and for alimenting peak-load power stations functioning in tandem with hydroelectric plants. Most grid peaking power plants and some off-grid engine-generators use natural gas. Particularly high efficiencies can be achieved through combining gas turbines with a steam turbine in combined cycle mode. Natural gas burns more cleanly than other fuels, such as oil and coal. Because burning natural gas produces both water and carbon dioxide, it produces less carbon dioxide per unit of energy released than coal, which produces mostly carbon dioxide. Burning natural gas produces only about half the carbon dioxide per kilowatt-hour (kWh) that coal does. For transportation, burning natural gas produces about 30% less carbon dioxide than burning petroleum. The US Energy Information Administration reports the following emissions in million metric tons of carbon dioxide in the world for 2012: Coal-fired electric power generation emits around of carbon dioxide for every megawatt-hour (MWh) generated, which is almost double the carbon dioxide released by natural gas-fired generation. Because of this higher carbon efficiency of natural gas generation, as the fuel mix in the United States has changed to reduce coal and increase natural gas generation, carbon dioxide emissions have unexpectedly fallen. Those measured in the first quarter of 2012 were the lowest of any recorded for the first quarter of any year since 1992. Combined cycle power generation using natural gas is currently the cleanest available source of power using hydrocarbon fuels, and this technology is widely and increasingly used as natural gas can be obtained at increasingly reasonable costs. Fuel cell technology may eventually provide cleaner options for converting natural gas into electricity, but as yet it is not price-competitive. Locally produced electricity and heat using natural gas powered Combined Heat and Power plant (CHP or Cogeneration plant) is considered energy efficient and a rapid way to cut carbon emissions. Natural gas generated power has increased from 740 TWh in 1973 to 5140 TWh in 2014, generating 22% of the worlds total electricity. Approximately half as much as generated with coal. Efforts around the world to reduce the use of coal has led some regions to switch to natural gas. Natural gas dispensed in a residential setting can generate temperatures in excess of making it a powerful domestic cooking and heating fuel. In much of the developed world it is supplied through pipes to homes, where it is used for many purposes including ranges and ovens, gas-heated clothes dryers, heating/cooling, and central heating. Heaters in homes and other buildings may include boilers, furnaces, and water heaters. Both North America and Europe are major consumers of natural gas. Domestic appliances, furnaces, and boilers use low pressure, usually 6 to 7 inches of water (6" to 7" WC), which is about 0.25 psig. The pressures in the supply lines vary, either utilization pressure (UP, the aforementioned 6" to 7" WC) or elevated pressure (EP), which may be anywhere from 1 psig to 120 psig. Systems using EP have a regulator at the service entrance to step down the pressure to UP. In the US compressed natural gas (CNG) is available in some rural areas as an alternative to less expensive and more abundant LPG (liquefied petroleum gas), the dominant source of rural gas. It is used in homes lacking direct connections to public utility provided gas, or to fuel portable grills. Natural gas is also supplied by independent natural gas suppliers through Natural Gas Choice programs throughout the United States. CNG is a cleaner and also cheaper alternative to other automobile fuels such as gasoline (petrol). By the end of 2014 there were over 20 million natural gas vehicles worldwide, led by Iran (3.5 million), China (3.3 million), Pakistan (2.8 million), Argentina (2.5 million), India (1.8 million), and Brazil (1.8 million). The energy efficiency is generally equal to that of gasoline engines, but lower compared with modern diesel engines. Gasoline/petrol vehicles converted to run on natural gas suffer because of the low compression ratio of their engines, resulting in a cropping of delivered power while running on natural gas (10–15%). CNG-specific engines, however, use a higher compression ratio due to this fuel's higher octane number of 120–130. Besides use in road vehicles, CNG can also be used in aircraft. Compressed natural gas has been used in some aircraft like the Aviat Aircraft Husky 200 CNG and the Chromarat VX-1 KittyHawk LNG is also being used in aircraft. Russian aircraft manufacturer Tupolev for instance is running a development program to produce LNG- and hydrogen-powered aircraft. The program has been running since the mid-1970s, and seeks to develop LNG and hydrogen variants of the Tu-204 and Tu-334 passenger aircraft, and also the Tu-330 cargo aircraft. Depending on the current market price for jet fuel and LNG, fuel for an LNG-powered aircraft could cost 5,000 rubles (US$100) less per tonne, roughly 60%, with considerable reductions to carbon monoxide, hydrocarbon and nitrogen oxide emissions. The advantages of liquid methane as a jet engine fuel are that it has more specific energy than the standard kerosene mixes do and that its low temperature can help cool the air which the engine compresses for greater volumetric efficiency, in effect replacing an intercooler. Alternatively, it can be used to lower the temperature of the exhaust. Natural gas is a major feedstock for the production of ammonia, via the Haber process, for use in fertilizer production. Natural gas can be used to produce hydrogen, with one common method being the hydrogen reformer. Hydrogen has many applications: it is a primary feedstock for the chemical industry, a hydrogenating agent, an important commodity for oil refineries, and the fuel source in hydrogen vehicles. Protein rich animal and fish feed is produced by feeding natural gas to Methylococcus capsulatus bacteria on commercial scale. Natural gas is also used in the manufacture of fabrics, glass, steel, plastics, paint, and other products. Human activity is responsible for about 60% of all methane emissions and for most of the resulting increase in atmospheric methane. Natural gas is intentionally released or is otherwise known to leak during the extraction, storage, transportation, and distribution of fossil fuels. Globally, this accounts for an estimated 33% of anthropogenic emissions in year 2020. The decomposition of municipal solid waste (a source of landfill gas) and wastewater account for an additional 18% of such emissions. These estimates include substantial uncertainties which should be reduced in the near future with improved satellite measurements, such as those planned for MethaneSAT. After release to the atmosphere, methane is removed by gradual oxidation to carbon dioxide and water by hydroxyl radicals () formed in the troposphere or stratosphere, giving the overall chemical reaction + 2 → + 2. While the lifetime of atmospheric methane is relatively short when compared to carbon dioxide, with a half-life of about 7 years, it is more efficient at trapping heat in the atmosphere, so that a given quantity of methane has 84 times the global-warming potential of carbon dioxide over a 20-year period and 28 times over a 100-year period. Natural gas is thus a potent greenhouse gas due to the strong radiative forcing of methane in the short term, and the continuing effects of carbon dioxide in the longer term. Targeted efforts to reduce warming quickly by reducing anthropogenic methane emissions is a climate change mitigation strategy supported by the Global Methane Initiative. When refined and burned, natural gas can produce 25–30% less carbon dioxide per joule delivered than oil, and 40–45% less than coal.. It can also produce potentially fewer toxic pollutants than other hydrocarbon fuels. In absolute terms, natural gas use comprises nearly one quarter of human carbon emissions, and this contribution is growing rapidly. Globally, natural gas use emitted about 6.7 billion tons of during year 2017, while coal and oil use emitted 11.4 and 14.5 billion tons, respectively. According to an updated version of the Special Report on Emissions Scenario by 2030, natural gas would be the source of 11 billion tons a year because demand is increasing 1.9% per year. The continued financing and construction of new gas pipelines indicates that huge emissions of fossil greenhouse gases could be locked-in for 40 to 50 years into the future. In the U.S. state of Texas alone, five new long-distance gas pipelines have been under construction, with the first entering service in Q3 2019, and the others scheduled to come online during 2020-2022. To reduce its greenhouse emissions, the government of the Netherlands is subsidizing a transition away from natural gas for all homes in the country by 2050. In Amsterdam, no new residential gas accounts are allowed as of July 1, 2018, and all homes in the city are expected to be converted by 2040 to use the excess heat from adjacent industrial buildings and operations. Natural gas produces far lower amounts of sulfur dioxide and nitrogen oxides than other fossil fuels. The pollutants due to natural gas combustion are listed below: Natural gas extraction also produces radioactive isotopes of polonium (Po-210), lead (Pb-210) and radon (Rn-220). Radon is a gas with initial activity from 5 to 200,000 becquerels per cubic meter of gas. It decays rapidly to Pb-210 which can build up as a thin film in gas extraction equipment. The natural gas extraction workforce face unique health and safety challenges and is recognized by the National Institute for Occupational Safety and Health (NIOSH) as a priority industry sector in the National Occupational Research Agenda (NORA) to identify and provide intervention strategies regarding occupational health and safety issues. Some gas fields yield sour gas containing hydrogen sulfide (), a toxic compound when inhaled. Amine gas treating, an industrial scale process which removes acidic gaseous components, is often used to remove hydrogen sulfide from natural gas. Extraction of natural gas (or oil) leads to decrease in pressure in the reservoir. Such decrease in pressure in turn may result in subsidence, sinking of the ground above. Subsidence may affect ecosystems, waterways, sewer and water supply systems, foundations, and so on. Releasing natural gas from subsurface porous rock formations may be accomplished by a process called hydraulic fracturing or "fracking". It is estimated that hydraulic fracturing will eventually account for nearly 70% of natural gas development in North America. Since the first commercial hydraulic fracturing operation in 1949, approximately one million wells have been hydraulically fractured in the United States. The production of natural gas from hydraulically fractured wells has utilized the technological developments of directional and horizontal drilling, which improved access to natural gas in tight rock formations. Strong growth in the production of unconventional gas from hydraulically fractured wells occurred between 2000–2012. In hydraulic fracturing, well operators force water mixed with a variety of chemicals through the wellbore casing into the rock. The high pressure water breaks up or "fracks" the rock, which releases gas from the rock formation. Sand and other particles are added to the water as a proppant to keep the fractures in the rock open, thus enabling the gas to flow into the casing and then to the surface. Chemicals are added to the fluid to perform such functions as reducing friction and inhibiting corrosion. After the "frack," oil or gas is extracted and 30–70% of the frack fluid, i.e. the mixture of water, chemicals, sand, etc., flows back to the surface. Many gas-bearing formations also contain water, which will flow up the wellbore to the surface along with the gas, in both hydraulically fractured and non-hydraulically fractured wells. This produced water often has a high content of salt and other dissolved minerals that occur in the formation. The volume of water used to hydraulically fracture wells varies according to the hydraulic fracturing technique. In the United States, the average volume of water used per hydraulic fracture has been reported as nearly 7,375 gallons for vertical oil and gas wells prior to 1953, nearly 197,000 gallons for vertical oil and gas wells between 2000–2010, and nearly 3 million gallons for horizontal gas wells between 2000–2010. Determining which fracking technique is appropriate for well productivity depends largely on the properties of the reservoir rock from which to extract oil or gas. If the rock is characterized by low-permeability — which refers to its ability to let substances, i.e. gas, pass through it, then the rock may be considered a source of tight gas. Fracking for shale gas, which is currently also known as a source of unconventional gas, involves drilling a borehole vertically until it reaches a lateral shale rock formation, at which point the drill turns to follow the rock for hundreds or thousands of feet horizontally. In contrast, conventional oil and gas sources are characterized by higher rock permeability, which naturally enables the flow of oil or gas into the wellbore with less intensive hydraulic fracturing techniques than the production of tight gas has required. The decades in development of drilling technology for conventional and unconventional oil and gas production has not only improved access to natural gas in low-permeability reservoir rocks, but also posed significant adverse impacts on environmental and public health. The US EPA has acknowledged that toxic, carcinogenic chemicals, i.e. benzene and ethylbenzene, have been used as gelling agents in water and chemical mixtures for high volume horizontal fracturing (HVHF). Following the hydraulic fracture in HVHF, the water, chemicals, and frack fluid that return to the well's surface, called flowback or produced water, may contain radioactive materials, heavy metals, natural salts, and hydrocarbons which exist naturally in shale rock formations. Fracking chemicals, radioactive materials, heavy metals, and salts that are removed from the HVHF well by well operators are so difficult to remove from the water they're mixed with, and would so heavily pollute the water cycle, that most of the flowback is either recycled into other fracking operations or injected into deep underground wells, eliminating the water that HVHF required from the hydrologic cycle. Natural gas in its native state is colorless and almost odorless. In order to assist consumers in detecting leaks, an odorizer with a scent similar to rotten eggs, tert-Butylthiol (t-butyl mercaptan), is added. Sometimes a related compound, thiophane, may be used in the mixture. Situations in which an odorant that is added to natural gas can be detected by analytical instrumentation, but cannot be properly detected by an observer with a normal sense of smell, have occurred in the natural gas industry. This is caused by odor masking, when one odorant overpowers the sensation of another. As of 2011, the industry is conducting research on the causes of odor masking. Explosions caused by natural gas leaks occur a few times each year. Individual homes, small businesses and other structures are most frequently affected when an internal leak builds up gas inside the structure. Frequently, the blast is powerful enough to significantly damage a building but leave it standing. In these cases, the people inside tend to have minor to moderate injuries. Occasionally, the gas can collect in high enough quantities to cause a deadly explosion, disintegrating one or more buildings in the process. Many building codes now forbid the installation of gas pipes inside cavity walls and/or below floor boards to mitigate against this risk. The gas usually dissipates readily outdoors, but can sometimes collect in dangerous quantities if flow rates are high enough. From 1994 through 2013, the United States had 745 serious incidents with gas distribution, causing 278 fatalities and 1059 injuries, with $110,658,083 in property damage. However, considering the tens of millions of structures that use the fuel, the individual risk of using natural gas is very low. Natural gas heating systems may cause carbon monoxide poisoning if unvented or poorly vented. In 2011, natural gas furnaces, space heaters, water heaters and stoves were blamed for 11 carbon monoxide deaths in the US. Another 22 deaths were attributed to appliances running on liquified petroleum gas, and 17 deaths on gas of unspecified type. Improvements in natural gas furnace designs have greatly reduced CO poisoning concerns. Detectors are also available that warn of carbon monoxide and/or explosive gas (methane, propane, etc.). Quantities of natural gas are measured in normal cubic meters (cubic meter of gas at "normal" temperature and pressure ) or standard cubic feet (cubic foot of gas at "standard" temperature and pressure ), . The gross heat of combustion of commercial quality natural gas is around , but this can vary by several percent. This is about (assuming a density of , an approximate value). Gas prices for end users vary greatly across the EU. A single European energy market, one of the key objectives of the EU, should level the prices of gas in all EU member states. Moreover, it would help to resolve supply and global warming issues, as well as strengthen relations with other Mediterranean countries and foster investments in the region. In US units, of natural gas produces around . The actual heating value when the water formed does not condense is the net heat of combustion and can be as much as 10% less. In the United States, retail sales are often in units of therms (th); 1 therm = 100,000 BTU. Gas sales to domestic consumers are often in units of 100 standard cubic feet (scf). Gas meters measure the volume of gas used, and this is converted to therms by multiplying the volume by the energy content of the gas used during that period, which varies slightly over time. The typical annual consumption of a single family residence is 1,000 therms or one Residential Customer Equivalent (RCE). Wholesale transactions are generally done in decatherms (Dth), thousand decatherms (MDth), or million decatherms (MMDth). A million decatherms is a trillion BTU, roughly a billion cubic feet of natural gas. The price of natural gas varies greatly depending on location and type of consumer. In 2007, a price of $7 per 1000 cubic feet () was typical in the United States. The typical caloric value of natural gas is roughly 1,000 BTU per cubic foot, depending on gas composition. This corresponds to around $7 per million BTU or around $7 per gigajoule (GJ). In April 2008, the wholesale price was $10 per 1000 cubic feet ($10/million BTU). The residential price varies from 50% to 300% more than the wholesale price. At the end of 2007, this was $12–$16 per 1000 cubic feet (). Natural gas in the United States is traded as a futures contract on the New York Mercantile Exchange. Each contract is for 10,000 million BTU or . Thus, if the price of gas is $10/million BTU on the NYMEX, the contract is worth $100,000. Canada uses metric measure for internal trade of petrochemical products. Consequently, natural gas is sold by the gigajoule (GJ), cubic meter (m3) or thousand cubic meters (E3m3). Distribution infrastructure and meters almost always meter volume (cubic foot or cubic meter). Some jurisdictions, such as Saskatchewan, sell gas by volume only. Other jurisdictions, such as Alberta, gas is sold by the energy content (GJ). In these areas, almost all meters for residential and small commercial customers measure volume (m3 or ft3), and billing statements include a multiplier to convert the volume to energy content of the local gas supply. A gigajoule (GJ) is a measure approximately equal to half a barrel (250 lbs) of oil, or 1 million BTUs, or of gas. The energy content of gas supply in Canada can vary from depending on gas supply and processing between the wellhead and the customer. Outside of the European Union, the U.S., and Canada, natural gas is sold in gigajoule retail units. LNG (liquefied natural gas) and LPG (liquefied petroleum gas) are traded in metric tonnes (1,000 kg) or million BTU as spot deliveries. Long term natural gas distribution contracts are signed in cubic meters, and LNG contracts are in metric tonnes. The LNG and LPG is transported by specialized transport ships, as the gas is liquified at cryogenic temperatures. The specification of each LNG/LPG cargo will usually contain the energy content, but this information is in general not available to the public. In the Russian Federation, Gazprom sold approximately of natural gas in 2008. In 2013 they produced of natural and associated gas. Gazprom supplied Europe with of gas in 2013. In August 2015, possibly the largest natural gas discovery in history was made and notified by an Italian gas company ENI. The energy company indicated that it has unearthed a "supergiant" gas field in the Mediterranean Sea covering about . It was also reported that the gas field could hold a potential of natural gas. ENI said that the energy is about. The field was found in the deep waters off the northern coast of Egypt and ENI claims that it will be the largest ever in the Mediterranean and even the world. Research conducted by the World Pensions Council (WPC) suggests that large US and Canadian pension funds and Asian and MENA area SWF investors have become particularly active in the fields of natural gas and natural gas infrastructure, a trend started in 2005 by the formation of Scotia Gas Networks in the UK by OMERS and Ontario Teachers' Pension Plan. Natural gas may be stored by adsorbing it to the porous solids called sorbents. The optimal condition for methane storage is at room temperature and atmospheric pressure. Pressures up to 4 MPa (about 40 times atmospheric pressure) will yield greater storage capacity. The most common sorbent used for ANG is activated carbon (AC), primarily in three forms: Activated Carbon Fiber (ACF), Powdered Activated Carbon (PAC), activated carbon monolith. General:
https://en.wikipedia.org/wiki?curid=22131
Nuclear chain reaction Nuclear chain reaction occurs when one single nuclear reaction causes an average of one or more subsequent nuclear reactions, thus leading to the possibility of a self-propagating series of these reactions. The specific nuclear reaction may be the fission of heavy isotopes (e.g., uranium-235, 235U). The nuclear chain reaction releases several million times more energy per reaction than any chemical reaction. Chemical chain reactions were first proposed by German chemist Max Bodenstein in 1913, and were reasonably well understood before nuclear chain reactions were proposed. It was understood that chemical chain reactions were responsible for exponentially increasing rates in reactions, such as produced in chemical explosions. The concept of a nuclear chain reaction was reportedly first hypothesized by Hungarian scientist Leó Szilárd on September 12, 1933. Szilárd that morning had been reading in a London paper of an experiment in which protons from an accelerator had been used to split lithium-7 into alpha particles, and the fact that much greater amounts of energy were produced by the reaction than the proton supplied. Ernest Rutherford commented in the article that inefficiencies in the process precluded use of it for power generation. However, the neutron had been discovered in 1932, shortly before, as the product of a nuclear reaction. Szilárd, who had been trained as an engineer and physicist, put the two nuclear experimental results together in his mind and realized that if a nuclear reaction produced neutrons, which then caused further similar nuclear reactions, the process might be a self-perpetuating nuclear chain-reaction, spontaneously producing new isotopes and power without the need for protons or an accelerator. Szilárd, however, did not propose fission as the mechanism for his chain reaction, since the fission reaction was not yet discovered, or even suspected. Instead, Szilárd proposed using mixtures of lighter known isotopes which produced neutrons in copious amounts. He filed a patent for his idea of a simple nuclear reactor the following year. In 1936, Szilárd attempted to create a chain reaction using beryllium and indium, but was unsuccessful. Nuclear fission was discovered by Otto Hahn and Fritz Strassmann in December 1938 and explained theoretically in January 1939 by Lise Meitner and her nephew Otto Robert Frisch. A few months later, Frédéric Joliot-Curie, H. Von Halban and L. Kowarski in Paris searched for, and discovered, neutron multiplication in uranium, proving that a nuclear chain reaction by this mechanism was indeed possible. On May 4, 1939, Joliot-Curie, Halban, and Kowarski filed three patents. The first two described power production from a nuclear chain reaction, the last one called "Perfectionnement aux charges explosives" was the first patent for the atomic bomb and is filed as patent No. 445686 by the Caisse nationale de Recherche Scientifique. In parallel, Szilárd and Enrico Fermi in New York made the same analysis. This discovery prompted the letter from Szilárd and signed by Albert Einstein to President Franklin D. Roosevelt, warning of the possibility that Nazi Germany might be attempting to build an atomic bomb. On December 2, 1942, a team led by Fermi (and including Szilárd) produced the first artificial self-sustaining nuclear chain reaction with the Chicago Pile-1 (CP-1) experimental reactor in a racquets court below the bleachers of Stagg Field at the University of Chicago. Fermi's experiments at the University of Chicago were part of Arthur H. Compton's Metallurgical Laboratory of the Manhattan Project; the lab was later renamed Argonne National Laboratory, and tasked with conducting research in harnessing fission for nuclear energy. In 1956, Paul Kuroda of the University of Arkansas postulated that a natural fission reactor may have once existed. Since nuclear chain reactions may only require natural materials (such as water and uranium, if the uranium has sufficient amounts of 235U), it was possible to have these chain reactions occur in the distant past when uranium-235 concentrations were higher than today, and where there was the right combination of materials within the Earth's crust. Kuroda's prediction was verified with the discovery of evidence of natural self-sustaining nuclear chain reactions in the past at Oklo in Gabon, Africa, in September 1972. Fission chain reactions occur because of interactions between neutrons and fissile isotopes (such as 235U). The chain reaction requires both the release of neutrons from fissile isotopes undergoing nuclear fission and the subsequent absorption of some of these neutrons in fissile isotopes. When an atom undergoes nuclear fission, a few neutrons (the exact number depends on uncontrollable and unmeasurable factors; the expected number depends on several factors, usually between 2.5 and 3.0) are ejected from the reaction. These free neutrons will then interact with the surrounding medium, and if more fissile fuel is present, some may be absorbed and cause more fissions. Thus, the cycle repeats to give a reaction that is self-sustaining. Nuclear power plants operate by precisely controlling the rate at which nuclear reactions occur. Nuclear weapons, on the other hand, are specifically engineered to produce a reaction that is so fast and intense it cannot be controlled after it has started. When properly designed, this uncontrolled reaction will lead to an explosive energy release. Nuclear weapons employ high quality, highly enriched fuel exceeding the critical size and geometry (critical mass) necessary in order to obtain an explosive chain reaction. The fuel for energy purposes, such as in a nuclear fission reactor, is very different, usually consisting of a low-enriched oxide material (e.g. UO2). There are two primary isotopes used for fission reactions inside of nuclear reactors. The first and most common is U-235 or uranium-235. This is the fissile isotope of uranium and it makes up approximately 0.7% of all naturally occurring uranium. Because of the small amount of uranium-235 that exists, it is considered a non-renewable energy source despite being found in rock formations around the world. U-235 cannot be used as fuel in its base form for energy production. It must undergo a process known as enrichment to produce the compound UO2 or uranium dioxide. The uranium dioxide is then pressed and formed into ceramic pellets, which can subsequently be placed into fuel rods. This is when the compound uranium dioxide can be used for nuclear power production. The second most common isotope used in nuclear fission is Pu-239 or plutonium-239. This is due to its ability to become fissile with slow neutron interaction. This isotope is formed inside nuclear reactors through exposing U-238 to the neutrons released by the radioactive U-235 isotope. This neutron capture causes beta particle decay that enables U-238 to transform into Pu-239. Plutonium was once found naturally in the earth's crust but only trace amounts remain. The only way it is accessible in large quantities for energy production is through the neutron capture method. The fissile isotope uranium-235 in its natural state is unfit for nuclear reactors. In order to be prepared for use as fuel in energy production, it must be enriched. The enrichment process does not apply to plutonium. Reactor-grade plutonium is created as a byproduct of neutron interaction between two different isotopes of uranium. The first step to enriching uranium begins by converting uranium oxide (created through the uranium milling process) into a gaseous form. This gas is known as uranium hexafluoride, which is created by combining hydrogen fluoride, fluorine gas, and uranium oxide. Uranium dioxide is also present in this process and it is sent off to be used in reactors not requiring enriched fuel. The remaining uranium hexafluoride compound is drained into strong metal cylinders where it solidifies. The next step is separating the uranium hexafluoride from the depleted U-235 left over. This is typically done with centrifuges that spin fast enough to allow for the 1% mass difference in uranium isotopes to separate themselves. A laser is then used to enrich the hexafluoride compound. The final step involves reconverting the now enriched compound back into uranium oxide, leaving the final product: enriched uranium oxide. This form of UO2 can now be used in fission reactors inside power plants to produce energy. When a fissile atom undergoes nuclear fission, it breaks into two or more fission fragments. Also, several free neutrons, gamma rays, and neutrinos are emitted, and a large amount of energy is released. The sum of the rest masses of the fission fragments and ejected neutrons is less than the sum of the rest masses of the original atom and incident neutron (of course the fission fragments are not at rest). The mass difference is accounted for in the release of energy according to the equation "E"=Δ"mc"2: Due to the extremely large value of the speed of light, "c", a small decrease in mass is associated with a tremendous release of active energy (for example, the kinetic energy of the fission fragments). This energy (in the form of radiation and heat) carries the missing mass, when it leaves the reaction system (total mass, like total energy, is always conserved). While typical chemical reactions release energies on the order of a few eVs (e.g. the binding energy of the electron to hydrogen is 13.6 eV), nuclear fission reactions typically release energies on the order of hundreds of millions of eVs. Two typical fission reactions are shown below with average values of energy released and number of neutrons ejected: Note that these equations are for fissions caused by slow-moving (thermal) neutrons. The average energy released and number of neutrons ejected is a function of the incident neutron speed. Also, note that these equations exclude energy from neutrinos since these subatomic particles are extremely non-reactive and, therefore, rarely deposit their energy in the system. The prompt neutron lifetime, "l", is the average time between the emission of neutrons and either their absorption in the system or their escape from the system. The neutrons that occur directly from fission are called "prompt neutrons," and the ones that are a result of radioactive decay of fission fragments are called "delayed neutrons". The term lifetime is used because the emission of a neutron is often considered its "birth," and the subsequent absorption is considered its "death". For thermal (slow-neutron) fission reactors, the typical prompt neutron lifetime is on the order of 10−4 seconds, and for fast fission reactors, the prompt neutron lifetime is on the order of 10−7 seconds. These extremely short lifetimes mean that in 1 second, 10,000 to 10,000,000 neutron lifetimes can pass. The "average" (also referred to as the "adjoint unweighted") prompt neutron lifetime takes into account all prompt neutrons regardless of their importance in the reactor core; the "effective" prompt neutron lifetime (referred to as the "adjoint weighted" over space, energy, and angle) refers to a neutron with average importance. The mean generation time, Λ, is the average time from a neutron emission to a capture that results in fission. The mean generation time is different from the prompt neutron lifetime because the mean generation time only includes neutron absorptions that lead to fission reactions (not other absorption reactions). The two times are related by the following formula: In this formula, k is the effective neutron multiplication factor, described below. The six factor formula effective neutron multiplication factor, "k", is the average number of neutrons from one fission that cause another fission. The remaining neutrons either are absorbed in non-fission reactions or leave the system without being absorbed. The value of "k" determines how a nuclear chain reaction proceeds: When describing kinetics and dynamics of nuclear reactors, and also in the practice of reactor operation, the concept of reactivity is used, which characterizes the deflection of reactor from the critical state: ρ = ("k" − 1)/"k". InHour (from "inverse of an hour", sometimes abbreviated ih or inhr) is a unit of reactivity of a nuclear reactor. In a nuclear reactor, "k" will actually oscillate from slightly less than 1 to slightly more than 1, due primarily to thermal effects (as more power is produced, the fuel rods warm and thus expand, lowering their capture ratio, and thus driving "k" lower). This leaves the average value of "k" at exactly 1. Delayed neutrons play an important role in the timing of these oscillations. In an infinite medium, the multiplication factor may be described by the four factor formula; in a non-infinite medium, the multiplication factor may be described by the six factor formula. Not all neutrons are emitted as a direct product of fission; some are instead due to the radioactive decay of some of the fission fragments. The neutrons that occur directly from fission are called "prompt neutrons," and the ones that are a result of radioactive decay of fission fragments are called "delayed neutrons". The fraction of neutrons that are delayed is called β, and this fraction is typically less than 1% of all the neutrons in the chain reaction. The delayed neutrons allow a nuclear reactor to respond several orders of magnitude more slowly than just prompt neutrons would alone. Without delayed neutrons, changes in reaction rates in nuclear reactors would occur at speeds that are too fast for humans to control. The region of supercriticality between "k" = 1 and "k" = 1/(1 − β) is known as delayed supercriticality (or delayed criticality). It is in this region that all nuclear power reactors operate. The region of supercriticality for "k" > 1/(1 − β) is known as prompt supercriticality (or prompt criticality), which is the region in which nuclear weapons operate. The change in "k" needed to go from critical to prompt critical is defined as a dollar. Nuclear fission weapons require a mass of fissile fuel that is prompt supercritical. For a given mass of fissile material the value of "k" can be increased by increasing the density. Since the probability per distance travelled for a neutron to collide with a nucleus is proportional to the material density, increasing the density of a fissile material can increase "k". This concept is utilized in the implosion method for nuclear weapons. In these devices, the nuclear chain reaction begins after increasing the density of the fissile material with a conventional explosive. In the gun-type fission weapon, two subcritical pieces of fuel are rapidly brought together. The value of "k" for a combination of two masses is always greater than that of its components. The magnitude of the difference depends on distance, as well as the physical orientation. The value of "k" can also be increased by using a neutron reflector surrounding the fissile material Once the mass of fuel is prompt supercritical, the power increases exponentially. However, the exponential power increase cannot continue for long since k decreases when the amount of fission material that is left decreases (i.e. it is consumed by fissions). Also, the geometry and density are expected to change during detonation since the remaining fission material is torn apart from the explosion. Detonation of a nuclear weapon involves bringing fissile material into its optimal supercritical state very rapidly. During part of this process, the assembly is supercritical, but not yet in an optimal state for a chain reaction. Free neutrons, in particular from spontaneous fissions, can cause the device to undergo a preliminary chain reaction that destroys the fissile material before it is ready to produce a large explosion, which is known as predetonation. To keep the probability of predetonation low, the duration of the non-optimal assembly period is minimized and fissile and other materials are used that have low spontaneous fission rates. In fact, the combination of materials has to be such that it is unlikely that there is even a single spontaneous fission during the period of supercritical assembly. In particular, the gun method cannot be used with plutonium (see nuclear weapon design). Chain reactions naturally give rise to reaction rates that grow (or shrink) exponentially, whereas a nuclear power reactor needs to be able to hold the reaction rate reasonably constant. To maintain this control, the chain reaction criticality must have a slow enough time scale to permit intervention by additional effects (e.g., mechanical control rods or thermal expansion). Consequently, all nuclear power reactors (even fast-neutron reactors) rely on delayed neutrons for their criticality. An operating nuclear power reactor fluctuates between being slightly subcritical and slightly delayed-supercritical, but must always remain below prompt-critical. It is impossible for a nuclear power plant to undergo a nuclear chain reaction that results in an explosion of power comparable with a nuclear weapon, but even low-powered explosions due to uncontrolled chain reactions (that would be considered "fizzles" in a bomb) may still cause considerable damage and meltdown in a reactor. For example, the Chernobyl disaster involved a runaway chain reaction but the result was a low-powered steam explosion from the relatively small release of heat, as compared with a bomb. However, the reactor complex was destroyed by the heat, as well as by ordinary burning of the graphite exposed to air. Such steam explosions would be typical of the very diffuse assembly of materials in a nuclear reactor, even under the worst conditions. In addition, other steps can be taken for safety. For example, power plants licensed in the United States require a negative void coefficient of reactivity (this means that if water is removed from the reactor core, the nuclear reaction will tend to shut down, not increase). This eliminates the possibility of the type of accident that occurred at Chernobyl (which was due to a positive void coefficient). However, nuclear reactors are still capable of causing smaller explosions even after complete shutdown, such as was the case of the Fukushima Daiichi nuclear disaster. In such cases, residual decay heat from the core may cause high temperatures if there is loss of coolant flow, even a day after the chain reaction has been shut down (see SCRAM). This may cause a chemical reaction between water and fuel that produces hydrogen gas, which can explode after mixing with air, with severe contamination consequences, since fuel rod material may still be exposed to the atmosphere from this process. However, such explosions do not happen during a chain reaction, but rather as a result of energy from radioactive beta decay, after the fission chain reaction has been stopped.
https://en.wikipedia.org/wiki?curid=22133
Nichiren Nichiren (日蓮; born as , Dharma name: "Rencho", 16 February 1222 – 13 October 1282) was a Japanese Buddhist priest of the Kamakura period (1185–1333), who developed the teachings of Nichiren Buddhism, a branch school of Mahayana Buddhism. Nichiren declared that the Lotus Sutra alone contains the highest truth of Buddhist teachings suited for the Third Age of Buddhism. He advocated the repeated recitation of its title, "Nam(u)-myoho-renge-kyo" and held that Shakyamuni Buddha and all other Buddhist deities were extraordinary manifestations of a particular Buddha-nature termed “"Myoho—Renge"” that is equally accessible to all. He declared that believers of the Sutra must propagate it even under persecution. Nichiren was a prolific writer and his biography, temperament, and the evolution of his beliefs has been gleaned primarily from his own writings. After his death, he was bestowed the title "Nichiren Dai-Bosatsu" (日蓮大菩薩) ("Great Bodhisattva Nichiren") by Emperor Go-Kōgon (1358) and the title "Risshō Daishi" (立正大師) ("Great Teacher of Rectification") was conferred posthumously in year 1922 by imperial edict. Today, Nichiren Buddhism includes traditional temple schools such as Nichiren-shu and Nichiren Shōshū, as well as lay movements such as Soka Gakkai, Risshō Kōsei Kai, Reiyūkai, Kenshōkai, Honmon Butsuryū-shū, Kempon Hokke, and Shōshinkai among many others. Each group has varying views of Nichiren's teachings with claims and interpretations of Nichiren's identity ranging from the rebirth of Bodhisattva Visistacaritra to the Primordial or "True Buddha" (本仏: "Honbutsu") of the Third Age of Buddhism. The main narrative of Nichiren's life has been constructed from extant letters and treatises he wrote, counted in one collection as 523 complete writings and 248 fragments. Aside from historical documents stored in the repositories of various Nichiren sects, the first extensive non-religious biographical account of Nichiren did not appear until more than 200 years after his death. He launched his teachings in 1253, advocating an exclusive return to the Lotus Sutra as based on its original Tendai interpretations. His 1260 treatise "Risshō Ankoku Ron" (立正安国論) ("On Establishing the Correct Teaching for the Peace of the Land") argued that a nation that embraces the Lotus Sutra will experience peace and prosperity whereas rulers who support inferior religious teachings invite disorder and disaster into their realms. In a 1264 essay, he stated that the title of the Lotus Sutra, ""Nam(u)-myoho-renge-kyo,"" encompasses all Buddhist teachings and its recitation leads to enlightenment. As a result of his adamant stance, he experienced severe persecution imposed by the Kamakura Shogunate and consequently began to see himself as "bodily reading the Lotus Sutra ("Jpn. Hokke shikidoku")." In some of his writings during a second exile (1271-1274) he began to identify himself with the key Lotus Sutra characters Sadāparibhūta and Visistacaritra and saw himself in the role of leading a vast outpouring of Bodhisattvas of the Earth. In 1274, after his two predictions of foreign invasion and political strife were seemingly actualized by the first attempted Mongol invasion of Japan along with an unsuccessful coup within the Hōjō clan, Nichiren was pardoned by the Shogunate authorities and his advice was sought but not heeded. The "Risshō Ankoku Ron" in which he first predicted foreign invasion and civil disorder is now considered by Japanese historians to be a literary classic illustrating the apprehensions of that period. Several hagiographies about Nichiren and are reflected in various pieces of artwork about incidents in his life. Nichiren remains a controversial figure among scholars who cast him as either a fervent nationalist or a social reformer with a transnational religious vision. Critical scholars have used words such as intolerant, nationalistic, militaristic, and self-righteous to portray him. On the other hand, Nichiren has been presented as a revolutionary, a classic reformer, and as a prophet. Nichiren is often compared to other religious figures who shared similar rebellious and revolutionary drives to reform degeneration in their respective societies or schools. According to the lunar Chinese calendar, Nichiren was born on 27th of the first month in 1222, which is 16 February in the Gregorian calendar. Nichiren was born in the village of Kominato (today part of the city of Kamogawa), Nagase District, Awa Province (within present-day Chiba Prefecture). Accounts of his lineage vary. Nichiren described himself as "the son of a "Sendara" ("Skt: chandala", despised outcast), "a son born of the lowly people living on a rocky strand of the out-of-the-way sea," and "the son of a sea-diver." In contrast, Hōnen, Shinran, Dōgen, and Eisai, the other founders of religious schools who predated Nichiren, were all born in the Kyoto region and came from noble or samurai backgrounds. Although his writings reflect a fierce pride of his lowly birth, followers after his death began to ascribe to him a more noble lineage, perhaps to attract more adherents. Some have claimed his father was a rōnin, a manorial functionary ("shokan"), or a political refugee. Nichiren's father was Mikuni-no-Tayu Shigetada, also known as Nukina Shigetada Jiro (d. 1258) and his mother was Umegiku-nyo (d. 1267). On his birth, his parents named him which has variously been translated into English as "Splendid Sun" and "Virtuous Sun Boy" among others. The exact site of Nichiren's birth is believed to be currently submerged off the shore from present-day Kominato-zan Tanjō-ji (小湊山 誕生寺) near a temple in Kominato that commemorates his birth. Between the years 1233 and 1253 Nichiren engaged in an intensive study of all of the ten schools of Buddhism prevalent in Japan at that time as well as the Chinese classics and secular literature. During these years, he became convinced of the preeminence of the "Lotus Sutra" and in 1253 returned to the temple where he first studied to present his findings. In a 1271 letter Nichiren outlined his rationale for deeply studying Buddhism: At the age of 12 he began his Buddhist study at a temple of the Tendai school, Seichō-ji (清澄寺, also called Kiyosumi-dera). He was formally ordained at sixteen years old and took the Buddhist name , "Renchō" meaning "Lotus Growth." He left Seichō-ji for Kamakura where he studied Pure Land Buddhism, a school that stressed salvation through nianfo (Japanese "nembutsu") or the invocation of Amitābha (Japanese "Amida"), and then studied Zen which had been growing in popularity in both Kamakura and Kyoto. He next traveled to Mount Hiei, the center of Japanese Tendai Buddhism, where he scrutinized the school's original doctrines and its subsequent incorporation of the theories and practices of Pure Land and Esoteric Buddhism. In the final stage of this twenty-year period he traveled to Mount Kōya, the center of Shingon esoteric Buddhism, and to Nara where he studied its six established schools, especially the Ritsu sect which emphasized strict monastic discipline. According to one of his letters, Nichiren returned to Seicho-ji Temple on 28 April 1253 to lecture on his twenty years of scholarship. What followed was his first public declaration of "Nam(u) Myoho Renge Kyo" atop Mount Kiyosumi. This marked the start of his campaign to return Tendai to the exclusive reliance of the Lotus Sutra and his efforts to convert the entire Japanese nation to this belief. This declaration also marks the start of his efforts to make profound Buddhist theory practical and actionable so an ordinary person could manifest Buddhahood within his or her own lifetime in the midst of day-to-day realities. At the same event, according to his own account and subsequent hagiography, he changed his name to "Nichiren", an abbreviation of "Nichi" (日 "Sun") and "Ren" (蓮 "Lotus"). "Nichi" represents both the light of truth and the sun goddess Amaterasu, symbolizing Japan itself. "Ren" signifies the Lotus Sutra. Nichiren envisioned Japan as the country where the true teaching of Buddhism would be revived and the starting point for its worldwide spread. At his lecture, it is construed, Nichiren vehemently attacked Honen, the founder of Pure Land Buddhism, and its practice of chanting the Nembutsu, "Nam(u) Amida Butsu". It is likely he also denounced the core teachings of Seicho-ji which had incorporated non-exclusive Lotus Sutra teachings and practices. In so doing he earned the animosity of the local steward, Hojo Kagenobu, who attempted to have Nichiren killed. Modern scholarship suggests that events unfolded not in a single day but over a longer period of time and had social, and political dimensions. Nichiren then developed a base of operation in Kamakura where he converted several Tendai priests, directly ordained others, and attracted lay disciples who were drawn mainly from the strata of the lower and middle samurai class. Their households provided Nichiren with economic support and became the core of Nichiren communities in several locations in the Kanto region of Japan. Nichiren arrived in Kamakura in 1254. Between 1254 and 1260 half of the population had perished due to a tragic succession of calamities that included drought, earthquakes, epidemics, famine, fires, and storms. Nichiren sought scriptual references to explain the unfolding of natural disasters and then wrote a series of works which, based on the Buddhist theory of the non-duality of the human mind and the environment, attributed the sufferings to the weakened spiritual condition of people, thereby causing the "Kami" (protective forces or traces of the Buddha) to abandon the nation. The root cause of this, he argued, was the widespread decline of the Dharma due to the mass adoption of the Pure Land teachings. The most renowned of these works, considered his first major treatise, was the "Risshō Ankoku Ron" (立正安国論), "On Securing the Peace of the Land through the Propagation of True Buddhism." Nichiren submitted it to Hōjō Tokiyori, the "de facto" leader of the Kamakura shogunate, as a political move to effectuate radical reform. In it he argued the necessity for "the Sovereign to recognize and accept the singly true and correct form of Buddhism (i.e., 立正: "risshō") as the only way to achieve peace and prosperity for the land and its people and end their suffering (i.e., 安国: "ankoku")." Using a dialectic form well-established in China and Japan, the treatise is a 10-segment fictional dialogue between a Buddhist wise man, presumably Nichiren, and a visitor who together lament the tragedies that have beleaguered the nation. The wise man answers the guest's questions and, after a heated exchange, gradually leads him to enthusiastically embrace the vision of a country grounded firmly on the ideals of the Lotus Sutra. In this writing Nichiren displays a skill in using analogy, anecdote, and detail to persuasively appeal to an individual's unique psychology, experiences, and level of understanding. The teacher builds his argument by quoting extensively from a set of Buddhist sutras and commentaries. In his future writings Nichiren continued to draw from the same sutras and commentaries, effectively forming Nichiren's canon of sources out of the Buddhist library which he deemed supportive of the Lotus Sutra including the Konkomyo, Daijuku, Ninno, Yakushi, and Nirvana sutras. They share in common apocalyptic or nation-protecting teachings and prophecies. The "Risshō Ankoku Ron" concludes with an urgent appeal to the ruler to cease all financial support for Buddhist schools promoting inferior teachings. Otherwise, Nichiren warns, as predicted by the sutras, the continued influence of inferior teachings would invite even more natural disasters as well as the outbreak of civil strife and foreign invasion. Nichiren submitted his treatise on 16 July 1260 but it drew no official response. It did, however, prompt a severe backlash from the Buddhist priests of other schools. Nichiren was challenged to a religious debate with leading Kamakura prelates in which, by his account, they were swiftly dispatched. Their lay followers, however, attempted to kill him at his dwelling which forced him to flee Kamakura. His critics had influence with key governmental figures and spread slanderous rumors about him. One year after he submitted the "Rissho Ankoku Ron" the authorities had him arrested and exiled to the Izu peninsula. Nichiren's Izu exile lasted two years. In his extant writings from this time period, Nichiren began to strongly draw from chapters 10-22 of the Lotus Sutra, what Tanabe calls its "third realm" "(daisan hōmon)". Nichiren began to emphasize the purpose of human existence as being the practice of the bodhisattva ideal in the real world which entails undertaking struggle and manifesting endurance. He suggested that he is a model of this behavior, a "votary" ("gyōja") of the Lotus Sutra. Upon being pardoned in 1263 Nichiren returned to Kamakura. In November 1264 he was ambushed and nearly killed at Komatsubara in Awa Province by a force led by Lord Tōjō Kagenobu. For the next few years he preached in provinces outside of Kamakura but returned in 1268. At this point the Mongols sent envoys to Japan demanding tribute and threatening invasion. Nichiren sent 11 letters to influential leaders reminding them about his predictions in the "Rissho Ankoku Ron". The threat and execution of Mongol invasion was the worst crisis in pre-modern Japanese history. In 1269 Mongol envoys again arrived to demand Japanese submission to their hegemony and the "bakufu" responded by mobilizing military defenses. The role of Buddhism in "nation-protection" ("chingo kokka") was long established in Japan at this time and the government galvanized prayers from Buddhist schools for this purpose. Nichiren and his followers, however, felt emboldened that the predictions he had made in 1260 of foreign invasion seemingly were being fulfilled and more people joined their movement. Daring a rash response from the "bakufu", Nichiren vowed in letters to his followers that he was giving his life to actualize the Lotus Sutra. He accelerated his polemics against the non-Lotus teachings the government had been patronizing at the very time it was attempting to solidify national unity and resolve. In a series of letters to prominent leaders he directly provoked the major prelates of Kamakura temples that the Hojo family patronized, criticized the principles of Zen which was popular among the samurai class, critiqued the esoteric practices of Shingon just as the government was invoking them, and condemned the ideas underlying Risshū as it was enjoying a revival. His actions at that time have been described by modern scholars either as a high form of altruisim or the ravings of a fanatic and madman. His claims drew the ire of the influential religious figures of the time and their followers, especially the Shingon priest Ryōkan (良観). In September 1271, after a fiery exchange of letters between the two, Nichiren was arrested by a band of soldiers and tried by Hei no Saemon (平の左衛門, also called 平頼綱 "Taira no Yoritsuna"), the deputy chief of the Hojo clan's Board of Retainers. Nichiren considered this as his second remonstration to the government. According to Nichiren's own account, he was sentenced to exile but was brought to Tatsunukuchi beach in Shichirigahama for execution. At the final moment an astronomical phenomenon, "a brilliant orb as bright as the moon," arced over the execution grounds, terrifying Nichiren's executioners into inaction. Some scholars have proposed alternative narratives for this story. Regardless of the account, Nichiren's life was spared and he was exiled to Sado Island. The incident has become known as the "Tatsunokuchi Persecution" and was regarded by Nichiren as a death-and-resurrection turning point. In the Nichiren tradition this is called his moment of "Hosshaku kenpon" (発迹顕本), translated as "casting off the transient and revealing the true" or "outgrowing the provisional and revealing the essential." After the failed execution authorities carried out Nichiren's original sentence of exile to Sado Island in the Sea of Japan. Upon arriving, he was dispatched to a small dilapidated temple located in a graveyard. Nichiren was accompanied by a few disciples and in the first winter they endured terrible cold, food deprivation, and threats from local inhabitants. Nichiren scholars describe a clear shift in both tone and message in letters written before his Sado exile and those written during and after. Initially, Nichiren's urgent concern was to rally his followers in Kamakura. The tactics of the "bakufu" suppression of the Nichiren community included exile, imprisonment, land confiscation, or ousting from clan membership. Apparently a majority of his disciples abandoned their faith and others questioned why they and Nichiren were facing such adversity in light of the Lotus Sutra's promise of "peace and security in the present life." In response he began to identify himself with Sadāparibhūta, a key figure in the Lotus Sutra, who in the 20th chapter invited repeated persecution in his efforts to propagate the sutra. Such hardship, Nichiren argued, fulfilled and validated the Lotus Sutra. He also identified himself with the bodhisattva Visistacaritra to whom Shakyamuni entrusted the future propagation of the Lotus Sutra, seeing himself in the role of leading a vast outpouring of Bodhisattvas of the Earth who pledged to liberate the oppressed. The numerous letters and minor treatises he wrote in Sado include what is considered his two most significant works, the "Kanjin no Honzon Shō" (観心本尊抄: "The Object of Devotion for Observing the Mind") and the "Kaimoku Shō" (開目抄: "On the Opening of the Eyes"). In the latter he stated that facing adversity should be regarded as a matter of course and that the resolve to carry on with the mission to propagate the sutra was for him more important than guarantees of protection: "Let Heaven forsake me. Let ordeals confront me. I will not begrudge bodily life... . No matter what trials we may encounter, so long as we do not have a mind of doubt, I and my disciples will naturally achieve the Buddha realm." He concluded this work with the vow to be the "pillar of Japan, the eyes of Japan, the great ship of Japan." At the end of the 1271-1272 winter Nichiren's conditions had improved. He had attracted a small band of followers in Sado who provided him with support and disciples from the mainland began visiting him and providing supplies. In 1272 there was an attempted coup in Kamakura and Kyoto, seemingly fulfilling the prediction he had made in the "Rissho Ankoku Ron" of rebellion in the domain. At this point Nichiren was transferred to much better accommodations. While on Sado island, Nichiren inscribed the first Mandala "Gohonzon" (御本尊). Although there is evidence of a Gohonzon in embryonic form as far back as the days right before his exile, the first in full form is dated to July 8, 1273 and includes the inscription of "Nichiren inscribes this for the first time." His writings on Sado provide his rationale for a calligraphic mandala depicting the assembly at Eagle Peak which was to be used as an object of devotion or worship. By increasingly associating himself with Visistacaritra he implied a direct link to the original and universal Buddha. He read in the 16th ("Life span") chapter of the Lotus Sutra a three-fold "secret Dharma" of the "daimoku", the object of worship ("honzon"), and the ordination platform ("kaidan"). These became the means for people to directly access the Buddha's enlightenment. At the bottom of each mandala he wrote: "This is the great mandala never before revealed in Jambudvipa during the more than 2,200 years since the Buddha's nirvana." He inscribed many Mandala Gohonzon during the rest of his life. More than a hundred Mandala Gohonzon preserved today are attributed to Nichiren's own hand. Nichiren was pardoned on February 14, 1274 and returned to Kamakura one month later on March 26. Nichiren wrote that his innocence and the accuracy of his predictions caused the regent Hōjō Tokimune to intercede on his behalf. Scholars have suggested that some of his well-connected followers might have had influence on the government's decision to release him. On April 8 he was summoned by Hei no Saemon, who inquired about the timing of the next Mongol invasion. Nichiren predicted that it would occur within the year. He used the audience as yet another opportunity to remonstrate with the government. Claiming that reliance on prayers based on esoteric rituals would invite further calamity, he urged the "bakufu" to ground itself exclusively on the Lotus Sutra. Deeply disappointed by the government's refusal to heed his advice, Nichiren left Kamakura one month later, on May 12, determined to become a solitary wayfarer. Five days later, however, on a visit to the residence of Lord Hakii Sanenaga of Mt. Minobu, he learned that followers in nearby regions had held steadfast during his exile. Despite severe weather and deprivation, Nichiren remained in Minobu for the rest of his career. During his self-imposed exile at Mount Minobu, a location 100 miles west of Kamakura, Nichiren led a widespread movement of followers in Kanto and Sado mainly through his prolific letter-writing. During the so-called "Atsuhara affair" of 1279 when governmental attacks were aimed at Nichiren's followers rather than himself, Nichiren's letters reveal an assertive and well-informed leader who provided detailed instructions through a sophisticated network of disciples serving as liaisons between Minobu and other affected areas in Japan. He also showed the ability to provide a compelling narrative of events that gave his followers a broad perspective of what was unfolding. More than half of the extant letters of Nichiren were written during his years at Minobu. Some consisted of moving letters to followers expressing appreciation for their assistance, counseling on personal matters, and explaining his teachings in more understandable terms. Two of his works from this period, the "Senji Shō" (撰時抄: "The Selection of the Time") and the "Hōon Shō" (報恩抄: "On Repaying Debts of Gratitude") constitute, along with his "Risshō Ankoku Ron" (立正安国論: "On Establishing the Correct Teaching for the Peace of the Land"), "Kaimoku Shō" ("The Opening of the Eyes"), and "Kanjin no Honzon Shō" ("The Object of Devotion for Observing the Mind"), what is commonly regarded as his five major writings. During his years at Minobu Nichiren intensified his attacks on mystical and esoteric practices (mikkyō 密教) that had been incorporated into the Japanese Tendai school. It becomes clear at this point that he understood that he was creating his own form of Lotus Buddhism. Nichiren and his disciples completed the Myō-hōkke-in Kuon-ji Temple (久遠寺) in 1281. In the 19th century this structure burned down to be replaced by a new structure completed in the second half of the Meiji era. While at Minobu Nichiren also inscribed numerous Mandala Gohonzon for bestowal upon specific disciples and lay believers. Nichiren Shoshu believers claim that after the execution of the three Atsuhara farmers he inscribed the "Dai Gohonzon" on October 12, 1279, a Gohonzon specifically addressed to all humanity. This assertion has been disputed by other schools as historically and textually incorrect. It is enshrined currently at the "Tahō Fuji Dai-Nichirenge-Zan Taiseki-ji", informally known as the "Head Temple Taiseki-ji" of the Nichiren Shōshū Order of Buddhism, located at the foot of Mount Fuji in Fujinomiya, Shizuoka. Several of these Mandala Gohonzon are prominently retained by the Nichiren-shū in Yamanashi Prefecture. Others survive today in the repositories of Nichiren Shōshū temples such as Taiseki-ji (大石寺) in Fujinomiya, Shizuoka, which has a particularly large collection of scrolls that is publicly aired once a year. It is apparent that Nichiren took great care in deciding which of his disciples were eligible to receive a Gohonzon inscribed by him. In the case of a letter written to Lady Niiama he took great care to explain why he would not inscribe a Gohonzon despite a deep personal bond. Among the Gohonzon he inscribed were several that were quite large and perhaps intended for congregational use in chapels maintained by some lay followers. In 1282, after years of privation, Nichiren fell ill. His followers encouraged him to travel to the hot springs for their medicinal benefits. En route, unable to travel further, he stopped at the home of a disciple in Ikegami, outside of present-day Tokyo, and died on 13 October 1282. According to legend, he died in the presence of fellow disciples after having spent several days lecturing from his sickbed on the Lotus Sutra, writing a final letter, and leaving instructions for the future of his movement after his death, namely the designation of the six senior disciples. His funeral and cremation took place the following day. His disciples left Ikegami with Nichiren's ashes on October 21, reaching back to Minobu on October 25. Nichiren's teachings developed over the course of his career and their evolution can be seen through the study of his writings as well as in the annotations he made in his personal copy of the Lotus Sutra, the so-called "Chū-hokekyō". Some scholars set a clear demarcation in his teachings at the time he arrived at Sado Island whereas others see a threefold division of thought: up to and through the Izu exile, from his return to Kamakura through the Sado Island exile, and during his years at Minobu. According to Anesaki, Nichiren, upon his arrival at Minobu, quickly turned his attention to consolidating his teachings toward their perpetuation. The scope of his thinking was outlined in an essay "Hokke Shuyō-shō" [法華取要抄, "Choosing the Heart of the Lotus Sutra"], considered by Nikkō Shōnin as one of Nichiren's ten major writings. Anesaki also claims that later during his Minobu years, in lectures he is said to have transmitted to his disciples, Nichiren summarized the key ideas of his teachings in one paragraph: Buddhahood is eternal, all people can and should manifest it in their lives; Nichiren is the personage in the Lotus Sutra whose mission it is to enable people to realize their enlightenment; his followers who share his vow are the Bodhisattvas of the Earth. This requires a spiritual and moral unity among followers based on their inherent Buddhahood; Nichiren established the seeds of this community and his followers to come must extend it globally. Thus the enlightened individual, country, and world are different expressions of the ideal of the Buddha land; and the enlightened heart of the individual plays out its role with the world and cosmos as its stage. This is Nichiren's vision of "Kosen-rufu", a time when the teachings of the Lotus Sutra would be widely spread throughout the world. Nichiren set a precedent for Buddhist social activism centuries before its emergence in other Buddhist schools. The uniqueness of his teachings was his attempt to move Buddhism from the theoretical to the actualizable. He held adamantly that his teachings would permit a nation to right itself and ultimately lead to world peace. Some of his religious thinking was derived from the Tendai understanding of the Lotus Sutra, syncretic beliefs that were deeply rooted in the culture of his times, and new perspectives that were products of Kamakura Buddhism. Other ideas were completely original and unique to him. Nichiren was a product of his times and some of his teachings were drawn from existing schools of thought or from emerging ideas in Kamakura Buddhism. Nichiren appropriated and expanded on these ideas. Nichiren stressed the concept of immanence, meaning that the Buddha's pure land is to be found in this present world ("shaba soku jakkōdo"). Related concepts such as attaining enlightenment in one's current form ("sokushin jōbutsu") and the belief that enlightenment is not attained but is originally existing within all people ("hongaku") had been introduced by Kūkai and Saicho several centuries earlier. These concepts were based on Chih-i's cosmology of the unity and interconnectedness of the universe called Three Thousand Realms in a Single Moment of Life ("ichinen sanzen"). Nichiren advanced these concepts by declaring that they were actualizable rather than theoretical. Cause and effect were simultaneous instead of linear. Contemplation of one's mind ("kanjin") took place within the singular belief in and commitment to the Lotus Sutra. According to Nichiren these phenomena manifest when a person chants the title of the Lotus Sutra ("date") and shares its validity with others, even at the cost of one's life if need be. Nichiren constructed a triad relationship between faith, practice, and study. Faith meant embracing his new paradigm of the Lotus Sutra. It was something that needed to be continually deepened. "To accept ("ju") [faith in the sutra] is easy," he explained to a follower, "to uphold it ("ji") is difficult. But the realization of Buddhahood lies in upholding [faith]." This could only be manifested by the practice of chanting the "daimoku" as well as teaching others to do the same, and study. Consequently, Nichiren consistently and vehemently objected to the perspective of the Pure Land school that stressed an other-worldly aspiration to some pure land. Behind his assertion is the concept of the nonduality of the subjective realm (the individual) and the objective realm (the land that the individual inhabits) which indicates that when the individual taps buddhahood, his or her present world becomes peaceful and harmonious. For Nichiren the widespread propagation of the Lotus Sutra and consequent world peace (""kosen-rufu"") was achievable and inevitable and tasked his future followers with a mandate to accomplish it. The Kamakura period of 13th century Japan was characterized by a sense of foreboding. Nichiren, as well as the others of this time, believed that they had entered the Latter Day of the Law ("Mappō"), the time which Shakyamuni predicted his teachings would lose their efficacy. Indeed, Japan had entered an era of extreme natural disasters, internal strife and political conflict. Although Nichiren attributed the turmoils and disasters in society to the widespread practice of what he deemed inferior Buddhist teachings that were under government sponsorship, he was enthusiastically upbeat about the portent of the age. He asserted, in contrast to other Mahayana schools, this was the best possible moment to be alive, the era in which the Lotus Sutra was to spread, and the time in which the Bodhisattvas of the Earth would appear to propagate it. "It is better to be a leper who chants Nam(u)-myōhō-renge-kyō than be a chief abbot of the Tendai school." The tradition of conducting open and sustained debate to clarify matters of fundamental Buddhist principles has deep-seated roots in Tibet, China, and Korea. This tradition was also quite pronounced in Japan. In addition to formalized religious debates, the Kamakura period was marked by flourishing and competitive oral religious discourse. Temples began to compete for the patronage of the wealthy and powerful through oratorical sermonizing and temple lecturers ("kōshi") faced pressure to attract crowds. Sermonizing spread from within the confines of temples to homes and the streets as wandering mendicants ("shidōso", "hijiri", or "inja") preached to both the educated and illiterate in exchange for alms. In order to teach principles of faith preachers incorporated colorful storytelling, music, vaudeville, and drama—which later evolved into Noh. A predominant topic of debate in Kamakura Buddhism was the concept of rebuking "slander of the Dharma." The Lotus Sutra itself strongly warns about slander of the Dharma. Hōnen, in turn, employed harsh polemics instructing people to “discard” ("sha" 捨), “close” ("hei" 閉), “put aside” ("kaku" 閣), and “abandon” ("hō" 抛) the Lotus Sutra and other non-Pure Land teachings. His ideas were vociferously attacked by many including Nichiren. Nichiren, however, elevated countering slander of the Dharma into a pillar of Buddhist practice. In fact, far more of his extant writings deal with the clarification of what constitutes the essence of Buddhist teachings than expositions of how to meditate. At age 32, Nichiren began a career of denouncing other Mahayana Buddhist schools of his time and declaring what he asserted was the correct teaching, the Universal Dharma ("Nam(u)-Myōhō-Renge-Kyō"), and chanting its words as the only path for both personal and social salvation. The first target of his polemics was Pure Land Buddhism which had begun to gain ascendancy among the leaders and populace and even had established itself within the Tendai school. Nichiren's detailed rationale is most famously articulated in his : "Treatise On Establishing the Correct Teaching for the Peace of the Land," his first major treatise and the first of his three remonstrations with the "bakufu" authorities. Although his times were harsh and permeated by "bakufu" culture, Nichiren always chose the power of language over bearing arms or resorting to violence. He didn't mince his words and was relentless to pursue dialogue whether in the form of debate, conversations, or correspondence. His spirit of engaging in discourse is captured in his statement, "Whatever obstacles I may encounter, as long as men [persons] of wisdom do not prove my teachings to be false, I will never yield." Hōnen introduced the concept of "single practice" Buddhism. Basing himself on the writings of the Chinese Buddhist Shandao, he advocated the singular practice of Nianfo, the recitation of the Buddha Amida's name. This practice was revolutionary because it was accessible to all and minimalized the monopolistic role of the entire monastic establishment. Nichiren appropriated the structure of a universally accessible single practice but substituted the Nianfo with the daimoku of "Nam(u)-myōhō-renge-kyō". This constituted renouncing the principle of aspirating to a Pure Land after death and asserting instead the Lotus perspective of attaining Buddhahood in one's present form in this lifetime. Japan had a long-established system of folk beliefs that existed outside of and parallel to the schools of the Buddhist establishment. Many of these beliefs had an influence on the various religious schools which, in turn, influenced each other, a phenomenon known as syncretism. Among these beliefs were the existence of "kami", indigenous gods and goddesses or protective forces, that influenced human and natural occurrences in a holistic universe. Some beliefs ascribed "kami" to traces of the Buddha. The belief in "kami" was deeply embedded in the episteme of the time. Human agency through prayers and rituals could summon forth "kami" who would engage in nation-protection ("chingo kokka"). According to some of his accounts, Nichiren undertook his study of Buddhism to largely understand why the "kami" had seemingly abandoned Japan, as witnessed by the decline of the imperial court. Because the court and the people had turned to teachings that had weakened their minds and resolve, he came to conclude, both people of wisdom and the protective forces had abandoned the nation. By extension, he argued, through proper prayer and action his troubled society would transform into an ideal world in which peace and wisdom prevail and "the wind will not thrash the branches nor the rain fall hard enough to break clods." From Nichiren's corpus appear several lines of unique Buddhist thought. Developed during his Izu exile, the Five Guides ("gogi") are five criteria through which Buddhist teachings can be evaluated and ranked. They are the quality of the teaching (kyō), the innate human capacity ("ki") of the people, the time ("ji"), the characteristic of the land or country ("koku"), and the sequence of dharma propagation ("kyōhō rufu no zengo"). From these five interrelated perspectives Nichiren declared his interpretation of the Lotus Sutra as the supreme teaching. Throughout his career Nichiren harshly denounced Buddhist practices other than his own as well as the existing social and political system. The tactic he adopted was "shakubuku," conversion, in which he shocked his adversaries with his denunciations while attracting followers through his outward display of supreme confidence. Modern detractors criticize his exclusivist single-truth perspective as intolerant. Apologists argue his arguments should be understood in the context of his samurai society and not through post-modern lenses such as tolerance. Both of them may be regarded as having seized an aspect of the truth, namely that Nichiren, rather like Dogen, was not less brilliantly original for being a rigid dogmatist in doctrine. As his career advanced, Nichiren's vehement polemics against Pure Land teachings came to include sharp criticisms of the Shingon, Zen, and Ritsu schools of Buddhism. Collectively his criticisms have become known as "the Four Denunciations." Later in his career he critiqued the Japanese Tendai school for its appropriation of Shingon elements. Reliance on Shingon rituals, he claimed, was magic and would decay the nation. He held that Zen was devilish in its belief that attaining enlightenment was possible without relying on the Buddha's words; Ritsu was thievery because it hid behind token deeds such as public works. In modern parlance the Four Denunciations rebuked thinking that demoralized and disengaged people by encouraging resignation and escapism. Nichiren deemed the world to be in a degenerative age and believed that people required a simple and effective means to rediscover the core of Buddhism and thereby restore their spirits and times. He described his Three Great Secret Laws ("Sandai hiho") as this very means. In a writing entitled "Sandai Hiho Sho", or "On the Transmission of the Three Great Secret Laws", Nichiren delineated three teachings in the heart of the 16th chapter of the Lotus Sutra which are secret because he claimed he received them as the leader of the Bodhisattvas of the Earth through a silent transmission from Shakyamuni. They are the invocation ("daimoku"), the object of worship ("honzon"), and the platform of ordination or place of worship ("kaidan"). The "daimoku", the rhythmic chanting of Nam(u)-myōhō-renge-kyō is the means to discover that one's own life, the lives of others, and the environment is the essence of the Buddha of absolute freedom. The chanting is to be done while contemplating the "honzon". At the age of 51, Nichiren inscribed his own Mandala Gohonzon, the object of veneration or worship in his Buddhism, "never before known," as he described it. The Gohonzon is a calligraphic representation of the cosmos and chanting "daimoku" to it is Nichiren's method of meditation to experience the truth of Buddhism. He believed this practice was efficacious, simple to perform, and suited to the capacity of the people and the time. Nichiren describes the first two secret laws in numerous other writings but the reference to the platform of ordination appears only in the "Sandai Hiho Sho", a work whose authenticity has been questioned by some scholars. Nichiren apparently left the fulfillment of this secret Dharma to his successors and its interpretation has been a matter of heated debate. Some state that it refers to the construction of a physical national ordination platform sanctioned by the emperor; others contend that the ordination platform is the community of believers ("sangha") or, simply, the place where practitioners of the Lotus Sutra live and make collective efforts to realize the ideal of establishing the true Dharma in order to establish peace to the land ("rissho ankoku"). The latter conception entails a robust interplay between religion and secular life and an egalitarian structure in which people are dedicated to perfecting an ideal society. According to Nichiren, practicing the Three Secret Laws results in the "Three Proofs" which verify their validity. The first proof is "documentary," whether the religion's fundamental texts, here the writings of Nichiren, make a lucid case for the eminence of the religion. "Theoretical proof" is an intellectual standard of whether a religion's teachings reasonably clarify the mysteries of life and death. "Actual proof," deemed the most important by Nichiren, demonstrates the validity of the teaching through the actual improvements achieved by practitioners in their daily lives. Nichiren was deeply aware of the karmic struggles his followers faced in their day-to-day existence and assured them that they could "cross the sea of suffering" ("Shiji Shiro-dono gosho"). Through prevailing over such challenges, he taught, they would establish a sense of inner freedom, peace of mind, and understanding of the Dharma that persisted independent of the ups and downs of sentient existence. He accepted prevailing Buddhist notions about karma that taught that a person's current conditions were the cumulative effect of past thoughts, words, and actions. He showed little concern, however, for attributing current circumstances to supposed past deeds. Rather, he viewed karma through the lens of the teachings of the Lotus Sutra which could enable all people to become Buddhas, even the ignorant and evil people of the Latter Day of the Law. When confronting karmic situations, the act of chanting Nam(u)-myoho-renge-kyo would open the wisdom of the Buddha, transforming karma into mission and a creative and joy-filled way of life. Beyond the sphere of a single individual's life, the process would awaken a person's concern for the broader society and sense of social responsibility. Nichiren introduced the term "votary of the Lotus Sutra" ("Hokekyō no gyōja") to describe himself. The Lotus Sutra itself speaks of the great trials that will be faced by individuals who base themselves on its teachings and attempt to spread it. Nichiren claims he read the sutra "bodily" ("shikidoku"), voluntarily inviting the entailing hardships it predicts rather than just reciting or meditating on its words. Through challenging these persecutions Nichiren claimed to have discovered his personal mission and felt great joy even when experiencing the harshness of exile. His sufferings became, in his thinking, redemptive opportunities to change his karma and give his life transcendent meaning. In enduring severe persecutions Nichiren claimed that the negative karma he had accumulated from the past could be eradicated quickly in his current life. He was an active agent in this process, not a victim. He even expressed appreciation to his tormentors for giving him the opportunity to serve as an envoy of the Buddha. In letters to some of his followers Nichiren extended the concept of meeting persecution for the sake of propagating the Dharma to experiencing tribulations in life such as problems with family discord or illness. He encouraged these followers to take ownership of such life events, view them as opportunities to repay karmic debts and mitigate them in shorter periods of time than would otherwise be the case. Naturally, he did not use language redolent of twentieth century concepts within, e.g., psychotherapy, hyperliberalism or the Human Potential Movement, but his insistence on personal responsibility can be reimagined in such terms. Nichiren reached a state of conviction that offered a new perspective on karma. He express that his resolve to carry out his mission was paramount in importance and that the Lotus Sutra's promise of a peaceful and secure existence meant finding joy and validation in the process of overcoming karma. According to Stone, in confronting karma Nichiren "demonstrated an attitude that wastes little energy in railing against it but unflinchingly embraces it, interpreting it in whatever way appears meaningful at the moment so as to use that suffering for one's own development and to offer it on behalf of others." Nichiren's teachings are replete with vows he makes for himself and asks his followers to share as well. Some are personal in nature such as frequent admonitions for people to transform their inner lives. "You must quickly reform the tenets you hold in your heart," he stated in his treatise "Rissho Ankoku Ron". He urged his followers to attain "treasures of the heart" and to reflect on their behavior as human beings. These vows are "this-worldly" rather than theoretical and are matched with an easily accessible practice. Nichiren also made a "great vow" of a political dimension. He and his followers to come would create the conditions that lead to a just nation and world which the Lotus Sutra describes as "Kosen-rufu". In earlier Japanese Buddhism the concept of "nation" was equated with Tennō (天皇), or imperial rule and "peace of the land" was associated with the stability of the regime. Nichiren's teachings, however, fully embraced a newly emerging viewpoint in medieval Japan that "nation" referred to the land and the people. Nichiren was unique among his contemporaries in charging the actual government in power, in this case the "bakufu" rather than the throne, with the peace of the land as well as the thriving of the Dharma. In his teachings based on the Lotus Sutra, all human beings are equal, whether the nation's sovereign or an unknown commoner. Enlightenment is not restricted to an individual' s inner life but is actualized by efforts toward the transformation of the land and the realization of an ideal society. This entails an urgent mandate. Nichiren links the great vow of personages in the Lotus Sutra to raise all people to the consciousness of the Buddha, to his own single-minded struggles to teach the Law despite the great persecutions he, Nichiren himself, encountered, to his injunction to future disciples to create the Buddha land in the "saha" world over the course of the myriad years to follow. Nichiren was a charismatic leader who attracted many followers during both his missionary trips and his exiles. Most of these followers were warriors and feudal lords. He maintained to his women followers that they were equally able to attain enlightenment. He set a high standard of leadership and, in his writings, shared his rationale and strategies with them, openly urging them to share his conviction and struggles. He left the fulfillment of the "kaidan", the third of his Three Secret Dharmas, to his disciples. His many extant letters demonstrate the scope and breadth of his relationship with them and his expectations for them. They recognized and trusted his charismatic leadership and his understanding of Buddhism. Many sought his guidance to overcome personal problems. Many were actively involved with supporting him financially and protecting his community of followers. Several of disciples were praised by him for sharing in his privations and a few lost their lives in these situations. Although over the centuries the movement he established was fraught with divisions, his followers sustained his teachings and example and various times gained considerable influence. Today his followers are found in influential lay movements as well as traditional Nichiren schools. The relationship between Nichiren and his disciples has been called "shitei funi", the oneness of mentor and disciple. Although the functions of the mentor and disciple may vary, they share the same goals and the same responsibility. Nichiren claimed the precedent for "shitei funi" is a core theme of the Lotus Sutra, especially in chapters 21 and 22 where the Buddha entrusts the future propagation of the sutra to the gathered bodhisattvas. After Nichiren's death, his teachings were interpreted in different ways. As a result, Nichiren Buddhism encompasses several major branches and schools, each with its own doctrine and set of interpretations of Nichiren's teachings. Many of Nichiren's writings still exist in his original handwriting, some as complete works and some as fragments. Other documents survive as copies made by his immediate disciples. Nichiren's existing works number over 700 manuscripts in total, including transcriptions of orally delivered lectures, letters of remonstration and illustrations. Scholars have divided the writings attributed to Nichiren into three categories: those whose authenticity are universally accepted, those generally designated as written by someone else after his death, and a third category in which the veracity of works is still being debated. In addition to treatises written in formal Classical Chinese ("kanbun" 漢文), Nichiren also wrote expositories and letters to followers in mixed kanji-kana vernacular as well as letters in simplified kana for believers such as children who could not read the more formal styles. Some of Nichiren's "kanbun" works, especially the "Risshō Ankoku Ron", are considered exemplary of the "kanbun" style, while many of his letters show unusual empathy and understanding for the downtrodden of his day. Among his "kanbun" treatises, five are generally accepted by Nichiren schools as his major works: Nichiren Shōshū adds an additional five writings to comprise a set of ten major writings. Other Nichiren sects dispute these selections as being either of secondary importance or as apocryphal: Among the collection of his extant writings are numerous letters to his follows in the form of thank you notes, messages of condolence, responses to questions, and spiritual counseling for trying moments in his followers' lives. Collectively these letters demonstrate that Nichiren was a master of providing both comfort and challenge befitting the unique personalities and situations of each individual. Many of these letters use tales drawn from Indian, Chinese, and Japanese traditions as well as historical anecdotes and stories from the Buddhist canon. Nichiren incorporated several hundred of these anecdotes and took liberty to freely embellish some of them; a few of the stories he provided do not appear in other collections and could be original. Another category of his letters follow the genres of Japanese "zuihitsu", lyrical and loosely organized essays that combine personal reflection and poetic language, or personal diaries ("nikki") Nichiren was a master of this genre and these colloquial works reveal his highly personal and charismatic method of proselytization as well as his deep caring for his followers. Nichiren used his letters as a means to inspire key supporters. About one hundred followers are identified as recipients and several received between 5 and 20 of them. The recipients tended to be of the warrior class and only scattered references appear about his lower status followers, many of whom were illiterate. The series of letters he wrote his followers during the "Atsuhara affair" of 1279 provide a case study of how he used personal written communications to direct a response to the government's actions and to keep his followers steadfast during the ordeal. Against a backdrop of earlier Buddhist teachings that deny the possibility of enlightenment to women or reserve that possibility for life after death, Nichiren is highly sympathetic to women. Based on various passages from the Lotus Sutra, Nichiren asserts that "Other sutras are written for men only. This sutra is for everyone." Ninety of his extant letters, nearly a fifth of the total, were addressed to female correspondents. Nichiren Shu published separate volumes of those writings. In these letters Nichiren plays particular attention to the instantaneous attainment of enlightenment of the Dragon King's daughter in the "Devadatta" (Twelfth) chapter of the Lotus Sutra and displays deep concern for the fears and worries of his female disciples. In many of his letters to female believers he often expressed compliments for their in-depth questions about Buddhism while encouraging them in their efforts to attain enlightenment in this lifetime.
https://en.wikipedia.org/wiki?curid=22135
Nichiren Buddhism Nichiren Buddhism () is a branch of Mahayana Buddhism based on the teachings of the 13th-century Japanese Buddhist priest Nichiren (1222–1282) and is one of the Kamakura Buddhism schools. Its teachings derive from some 300–400 extant letters and treatises attributed to Nichiren. Nichiren Buddhism focuses on the Lotus Sutra doctrine that all people have an innate Buddha-nature and are therefore inherently capable of attaining enlightenment in their current form and present lifetime. There are three essential aspects to Nichiren Buddhism, the undertaking of faith, the practice of chanting "Nam Myoho Renge Kyo" accompanied by selected recitations of the Lotus Sutra, and the study of Nichiren's scriptural writings, called "Gosho". The Nichiren "Gohonzon" is a calligraphic image which is prominently displayed in the home or temple buildings of its believers. The "Gohonzon" used in Nichiren Buddhism is composed of the names of key bodhisattvas and Buddhas in the Lotus Sutra as well as "Namu-Myoho-Renge-Kyo" written in large characters down the center. After his death, Nichiren left to his followers the mandate to widely propagate the "Gohonzon" and "Daimoku" in order to secure the peace and prosperity of society. Traditional Nichiren Buddhist temple groups are commonly associated with Nichiren Shōshū and various Nichiren-shū schools. There are also lay groups not affiliated with temples such as Soka Gakkai and Soka Gakkai International, Kenshokai, Shoshinkai, Risshō Kōsei Kai, and Honmon Butsuryū-shū. Several Japanese new religions are Nichiren-inspired lay groups. With the advent, and proselytizing efforts, of the Soka Gakkai International, called "the most prominent Japanese 'export' religion to draw significant numbers of non-Japanese converts", Nichiren Buddhism has spread throughout the world. Nichiren proposed a classification system that ranks the quality of religions and various Nichiren schools can be either accommodating or vigorously opposed to any other forms of Buddhism or religious beliefs. Within Nichiren Buddhism there are two major divisions which fundamentally differ over whether Nichiren should be regarded as a bodhisattva of the earth, a saint, great teacher—or the actual Buddha of the third age of Buddhism. It is practiced worldwide, with practitioners throughout the United States, Brazil and Europe, as well as in South Korea and southeast Asia. The largest groups are Soka Gakkai International, Nichiren Shu, and Nichiren Shōshū. Nichiren's teachings encompass a significant number of concepts. Briefly, the basic practice of Nichiren Buddhism is chanting the invocation "Nam-myoho-renge-kyo" to a mandala inscribed by Nichiren, called the "Gohonzon". Embracing "Nam-myoho-renge-kyo" entails both chanting and having the mind of faith ("shinjin"). Both the invocation and the Gohonzon, as taught by Nichiren, embody the title and essence of the Lotus Sutra, which he taught as the only valid scripture for the Latter Day of the Law, as well as the life state of Buddhahood inherent in all life. Nichiren considered that in the Latter Day of the Law – a time of human strife and confusion, when Buddhism would be in decline – Buddhism had to be more than the theoretical or meditative practice it had become, but was meant to be practiced "with the body", that is, in one's actions and the consequent results that are manifested. More important than the formality of ritual, he claimed, was the substance of the practitioner's life in which the spiritual and material aspects are interrelated. He considered conditions in the world to be a reflection of the conditions of the inner lives of people; the premise of his first major remonstrance, Rissho Ankoku Ron (Establishing The Correct Teaching for the Peace of The Land), is that if a nation abandons heretical forms of Buddhism and adopts faith in the Lotus Sutra, the nation will know peace and security. He considered his disciples the "Bodhisattvas of the Earth" who appeared in the Lotus Sutra with the vow to spread the correct teaching and thereby establish a peaceful and just society. For Nichiren, enlightenment is not limited to one's inner life, but is "something that called for actualization in endeavors toward the transformation of the land, toward the realization of an ideal society." The specific task to be pursued by Nichiren's disciples was the widespread propagation of his teachings (the invocation and the "Gohonzon") in a way that would effect actual change in the world's societies so that the sanctuary, or seat, of Buddhism could be built. Nichiren saw this sanctuary as a specific seat of his Buddhism, but there is thought that he also meant it in a more general sense, that is, wherever his Buddhism would be practiced. This sanctuary, along with the invocation and "Gohonzon", comprise "the three great secret laws (or dharmas)" found in the Lotus Sutra. Nichiren Buddhism originated in 13th-century feudal Japan. It is one of six new forms of "Shin Bukkyo" (English: "New Buddhism") of "Kamakura Buddhism." The arrival of these new schools was a response to the social and political upheaval in Japan during this time as power passed from the nobility to a shogunate military dictatorship led by the Minamoto clan and later to the Hōjō clan. A prevailing pessimism existed associated with the perceived arrival of the Age of the Latter Day of the Law. The era was marked by an intertwining relationship between Buddhist schools and the state which included clerical corruption. By Nichiren's time the Lotus Sūtra was firmly established in Japan. From the ninth century, Japanese rulers decreed that the Lotus Sūtra be recited in temples for its "nation-saving" qualities. It was the most frequently read and recited sutra by the literate lay class and its message was disseminated widely through art, folk tales, music, and theater. It was commonly held that it had powers to bestow spiritual and worldly benefits to individuals. However, even Mount Hiei, the seat of Tiantai Lotus Sutra devotion, had come to adopt an eclectic assortment of esoteric rituals and Pure Land practices as "expedient means" to understand the sutra itself. Nichiren developed his thinking in this midst of confusing Lotus Sutra practices and a competing array of other "Old Buddhism" and "New Buddhism" schools. The biographical development of his thinking is sourced almost entirely from his extant writings as there is no documentation about him in the public records of his times. Modern scholarship on Nichiren's life tries to provide sophisticated textual and sociohistorical analyses to cull longstanding myths about Nichiren that accrued over time from what is actually concretized. It is clear that from an early point in his studies Nichiren came to focus on the Lotus Sutra as the culmination and central message of Shakyamuni. As his life unfolded he engaged in a "circular hermeneutic" in which the interplay of the Lotus Sutra text and his personal experiences verified and enriched each other in his mind. As a result, there are significant turning points as his teachings reach full maturity. Scholar Yoshirō Tamura categorizes the development of Nichiren's thinking into three periods: For more than 20 years Nichiren examined Buddhist texts and commentaries at Mount Hiei's Enryaku-ji temple and other major centers of Buddhist study in Japan. In later writings he claimed he was motivated by four primary questions: (1) What were the essentials of the competing Buddhist sects so they could be ranked according to their merits and flaws? (2) Which of the many Buddhist scriptures that had reached Japan represented the essence of Shakyamuni's teaching? (3) How could he be assured of the certainty of his own enlightenment? (4) Why was the Imperial house defeated by the Kamakura regime in 1221 despite the prayers and rituals of Tendai and Shingon priests? He eventually concluded that the highest teachings of Shakyamuni Buddha ( – ) were to be found in the Lotus Sutra. Throughout his career Nichiren carried his personal copy of the Lotus Sutra which he continually annotated. The mantra he expounded on 28 April 1253, known as the "Daimoku" or "Odaimoku", Namu Myōhō Renge Kyō, expresses his devotion to the Lotus Sutra. From this early stage of his career, Nichiren started to engage in fierce polemics criticizing the teachings of Buddhism taught by the other sects of his day, a practice that continued and expanded throughout his life. Although Nichiren accepted the Tendai theoretical constructs of "original enlightenment" ("hongaku shisō") and "attaining Buddhahood in one's present form" ("sokushin jobutsu") he drew a distinction, insisting both concepts should be seen as practical and realizable amidst the concrete realities of daily life. He took issue with other Buddhist schools of his time that stressed transcendence over immanence. Nichiren's emphasis on "self-power" (Jpn. "ji-riki") led him to harshly criticize Honen and his Pure Land Buddhism school because of its exclusive reliance on Amida Buddha for salvation which resulted in "other-dependence." (Jpn. "ta-riki") In addition to his critique of Pure Land Buddhism, he later expanded his polemics to criticisms of the Zen, Shingon, and Ritsu sects. These four critiques were later collectively referred to as his "four dictums." Later in his writings, Nichiren referred to his early exegeses of the Pure Land teachings as just the starting point for his polemics against the esoteric teachings, which he had deemed as a far more significant matter of concern. Adding to his criticisms of esoteric Shingon, Nichiren wrote detailed condemnations about the Tendai school which had abandoned its Lotus Sutra-exclusiveness and incorporated esoteric doctrines and rituals as well as faith in the soteriological power of Amida Buddha. The target of his tactics expanded during the early part of his career. Between 1253 and 1259 he proselytized and converted individuals, mainly attracting mid- to lower-ranking samurai and local landholders and debated resident priests in Pure Land temples. In 1260, however, he attempted to directly reform society as a whole by submitting a treatise entitled ""Risshō Ankoku Ron"" (""Establishment of the Legitimate Teaching for the Protection of the Country"") to Hōjō Tokiyori, the "de facto" leader of the nation. In it he cites passages from the Ninnō, Yakushi, Daijuku, and Konkōmyō sutras. Drawing on Tendai thinking about the non duality of person and land, Nichiren argued that the truth and efficacy of the people's religious practice will be expressed in the outer conditions of their land and society. He thereby associated the natural disasters of his age with the nation's attachment to inferior teachings, predicted foreign invasion and internal rebellion, and called for the return to legitimate dharma to protect the country. Although the role of Buddhism in "nation-protection" ("chingo kokka") was well-established in Japan at this time, in this thesis Nichiren explicitly held the leadership of the country directly responsible for the safety of the land. During the middle stage of his career, in refuting other religious schools publicly and vociferously, Nichiren provoked the ire of the country's rulers and of the priests of the sects he criticized. As a result, he was subjected to persecution which included two assassination attempts, an attempted beheading and two exiles. His first exile, to Izu Peninsula (1261–1263), convinced Nichiren that he was "bodily reading the Lotus Sutra ("Jpn. Hokke shikidoku")," fulfilling the predictions on the 13th chapter ("Fortitude") that votaries would be persecuted by ignorant lay people, influential priests, and their friends in high places. Nichiren began to argue that through "bodily reading the Lotus Sutra," rather than just studying its text for literal meaning, a country and its people could be protected. According to Habito, Nichiren argued that bodily reading the Lotus Sutra entails four aspects: His three-year exile to Sado Island proved to be another key turning point in Nichiren's thinking. Here he began inscribing the "Gohonzon" and wrote several major theses in which he claimed that he was functioning, at first, in the role of Bodhisattva Never Disparaging of the 20th chapter of the Lotus Sutra and, later, as Bodhisattva Superior Practices, the leader of the Bodhisattvas of the Earth. In his work "The True Object of Worship", he identified himself as functioning as the primordial Buddha, one and the same as the eternal Law represented by the mantra "Namu Myōhō Renge Kyō" which he physically embodied as the "Gohonzon" mandala. This has been described as embodying the same condition or state he attained in a physical object of devotion worship so that others could attain that equivalent condition of enlightenment. During this time the "daimoku" becomes the means to directly access the Buddha's enlightenment. He concludes his work "The Opening of the Eyes" with the declaration "I will be the pillar of Japan; I will be the eyes of Japan; I will be the vessel of Japan. Inviolable shall remain these vows!" His thinking now went beyond theories of karmic retribution or guarantees of the Lotus Sutra as a protective force. Rather, he expressed a resolve to fulfill his mission despite the consequences. All of his disciples, he asserted, should emulate his spirit and work just like him in helping all people open their innate Buddha lives even though this means entails encountering enormous challenges. Nichiren's teachings reached their full maturity between the years 1274 and 1282 while he resided in primitive settings at Mount Minobu located in today's Yamanashi Prefecture. During this time he devoted himself to training disciples, produced most of the "Gohonzon" which he sent to followers, and authored works constituting half of his extant writings including six treatises that were categorized by his follower Nikkō as among his ten most important. In 1278 the “Atsuhara Affair” (“Atsuhara Persecution”) occurred, culminating three years later. In the prior stage of his career, between 1261 and 1273, Nichiren endured and overcame numerous trials that were directed at him personally including assassination attempts, an attempted execution, and two exiles, thereby “bodily reading the Lotus Sutra” ("shikidoku" 色読). In so doing, according to him, he validated the 13th ("Fortitude") chapter of the Lotus Sutra in which a host of bodhisattvas promise to face numerous trials that follow in the wake of upholding and spreading the sutra in the evil age following the death of the Buddha: slander and abuse; attack by swords and staves; enmity from kings, ministers, and respected monks; and repeated banishment. On two occasions, however, the persecution was aimed at his followers. First, in 1271, in conjunction with the arrest and attempted execution of Nichiren and his subsequent exile to Sado, many of his disciples were arrested, banished, or had lands confiscated by the government. At that time, Nichiren stated, most recanted their faith in order to escape the government's actions. In contrast, during the Atsuhara episode twenty lay peasant-farmer followers were arrested on questionable charges and tortured; three were ultimately executed. This time none recanted their faith. Some of his prominent followers in other parts of the country were also being persecuted but maintained their faith as well. Although Nichiren was situated in Minobu, far from the scene of the persecution, the Fuji district of present-day Shizuoka Prefecture, Nichiren held his community together in the face of significant oppression through a sophisticated display of legal and rhetorical responses. He also drew on a wide array of support from the network of leading monks and lay disciples he had raised, some of whom were also experiencing persecution at the hands of the government. Throughout the events he wrote many letters to his disciples in which he gave context to the unfolding events by asserting that severe trials have deep significance. According to Stone, “By standing firm under interrogation, the Atsuhara peasants had proved their faith in Nichiren’s eyes, graduating in his estimation from ‘ignorant people’ to devotees meriting equally with himself the name of ‘practitioners of the Lotus Sutra.’” During this time Nichiren inscribed 114 mandalas that are extant today, 49 of which have been identified as being inscribed for individual lay followers and which may have served to deepen the bond between teacher and disciple. In addition, a few very large mandalas were inscribed, apparently intended for use at gathering places, suggesting the existence of some type of conventicle structure. The Atsuhara Affair also gave Nichiren the opportunity to better define what was to become Nichiren Buddhism. He stressed that meeting great trials was a part of the practice of the Lotus Sutra; the great persecutions of Atsuhara were not results of karmic retribution but were the historical unfolding of the Buddhist Dharma. The vague “single good of the true vehicle” which he advocated in the "Risshō ankoku ron" now took final form as chanting the Lotus Sutra's "daimoku" or title which he described as the heart of the “origin teaching” ("honmon" 本門) of the Lotus Sutra. This, he now claimed, lay hidden in the depths of the 16th (“The Life Span of the Tathāgata”) chapter, never before being revealed, but intended by the Buddha solely for the beginning of the Final Dharma Age. A prolific writer, Nichiren's personal communiques among his followers as well as numerous treatises detail his view of the correct form of practice for the "Latter Day of the Law" ("mappō"); lay out his views on other Buddhist schools, particularly those of influence during his lifetime; and elucidate his interpretations of Buddhist teachings that preceded his. These writings are collectively known as "Gosho" (御書) or "Nichiren ibun" (日蓮遺文). Out of 162 historically identified followers of Nichiren, 47 were women. Many of his writings were to women followers in which he displays strong empathy for their struggles, and continually stressed the Lotus Sutra's teaching that all people, men and women equally, can become enlightened just as they are. His voice is sensitive and kind which differs from the strident picture painted about him by critics. Which of these writings, including the "Ongi Kuden" (orally transmitted teachings), are deemed authentic or apocryphal is a matter of debate within the various schools of today's Nichiren Buddhism. His "Rissho Ankoku Ron", preserved at Shochuzan Hokekyo-ji, is one of the National Treasures of Japan. After Nichiren's death in 1282 the Kamakura shogunate weakened largely due to financial and political stresses resulting from defending the country from the Mongols. It was replaced by the Ashikaga shogunate (1336–1573), which in turn was succeeded by the Azuchi–Momoyama period (1573–1600), and then the Tokugawa shogunate (1600–1868). During these time periods, collectively comprising Japan's medieval history, Nichiren Buddhism experienced considerable fracturing, growth, turbulence and decline. A prevailing characteristic of the movement in medieval Japan was its lack of understanding of Nichiren's own spiritual realization. Serious commentaries about Nichiren's theology did not appear for almost two hundred years. This contributed to divisive doctrinal confrontations that were often superficial and dogmatic. This long history of foundings, divisions, and mergers have led to today's 37 legally incorporated Nichiren Buddhist groups. In the modern period, Nichiren Buddhism experienced a revival, largely initiated by lay people and lay movements. Several denominations comprise the umbrella term "Nichiren Buddhism" which was known at the time as the "Hokkeshū" (Lotus School) or "Nichirenshū" (Nichiren School). The splintering of Nichiren's teachings into different schools began several years after Nichiren's passing. Despite their differences, however, the Nichiren groups shared commonalities: asserting the primacy of the Lotus Sutra, tracing Nichiren as their founder, centering religious practice on chanting "Namu-myoho-renge-kyo", using the "Gohonzon" in meditative practice, insisting on the need for propagation, and participating in remonstrations with the authorities. The movement was supported financially by local warlords or stewards ("jitõ") who often founded tightly-organized clan temples ("ujidera") that were frequently led by sons who became priests. Most Nichiren schools point to the founding date of their respective head or main temple (for example, Nichiren Shū the year 1281, Nichiren Shōshū the year 1288, and Kempon Hokke Shu the year 1384) although they did not legally incorporate as religious bodies until the late 19th and early 20th century. A last wave of temple mergers took place in the 1950s. The roots of this splintering can be traced to the organization of the Nichiren community during his life. In 1282, one year before his death, Nichiren named "six senior priests" ("rokurōsō") disciple to lead his community: Nikkō Shonin (日興), Nisshō (日昭), Nichirō (日朗), Nikō (日向), Nitchō (日頂), and Nichiji (日持). Each had led communities of followers in different parts of the Kanto region of Japan and these groups, after Nichiren's death, ultimately morphed into lineages of schools. Nikkō Shonin, Nichirō, and Nisshō were the core of the Minobu (also known as the Nikō or Kuon-ji) "monryu" or school. Nikō became the second chief abbot of Minobu (Nichiren is considered by this school to be the first). Nichirō's direct lineage was called the Nichirō or Hikigayatsu "monryu". Nisshō's lineage became the Nisshō or Hama "monryu". Nitchō formed the Nakayama lineage but later returned to become a follower of Nikkō. Nichiji, originally another follower of Nikkō, eventually traveled to the Asian continent (ca. 1295) on a missionary journey and some scholarship suggests he reached northern China, Manchuria, and possibly Mongolia. Kuon-ji Temple in Mount Minobu eventually became the head temple of today's Nichiren Shū, the largest branch among traditional schools, encompassing the schools and temples tracing their origins to Nikō, Nichirō, Nisshō, Nitchō, and Nichiji. The lay and/or new religious movements Reiyūkai, Risshō Kōsei Kai, and Nipponzan-Myōhōji-Daisanga stem from this lineage. Nikkō left Kuon-ji in 1289 and became the founder of what was to be called the Nikkō "monryu" or lineage. He founded a center at the foot of Mount Fuji which would later be known as the Taisekiji temple of Nichiren Shōshū. Soka Gakkai is the largest independent lay organization that shares roots with this lineage. Fault lines between the various Nichiren groups crystallized over several issues: The cleavage between Nichiren groups has also been classified by the so-called "Itchi" (meaning unity or harmony) and "Shoretsu" (a contraction of two words meaning superior/inferior) lineages. Although there were rivalries and unique interpretations among the early Hokkeshũ lineages, none were as deep and distinct as the divide between the Nikkō or Fuji school and the rest of the tradition. Animosity and discord among the six senior disciples started after the second death anniversary of Nichiren's 100th Day Memorial ceremony (23 January 1283) when the rotation system as agreed upon the ""Shuso Gosenge Kiroku"" (English: Record document of founder's demise) and "Rimbo Cho" (English: Rotation Wheel System) to clean and maintain Nichiren's grave. By the third anniversary of Nichiren's passing (13 October 1284), these arrangements seemed to have broken down. Nikkō claimed that the other five senior priests no longer returned to Nichiren's tomb in Mount Minobu, citing signs of neglect at the gravesite. He took up residency and overall responsibility for Kuonji temple while Nikō served as its doctrinal instructor. Before long tensions grew between the two concerning the behavior of Hakii Nanbu Rokurō Sanenaga, the steward of the Minobu district and the temple's patron. Nikkō accused Sanenaga of unorthodox practices deemed to be heretical such as crafting a standing statue of Shakyamuni Buddha as an object of worship, providing funding for the construction of a Pure Land "stupa" in Fuji, and visiting and worshiping at the Mishima Taisha Shinto shrine which was an honorary shrine of the Hōjō clan shogunate. Nikkō regarded the latter as a violation of Nichiren's "Rissho ankoku ron". In addition, Nikkō made accusatory charges that after Nichiren's death, other disciples slowly began to gradually deviate from what Nikkō viewed as Nichiren's orthodox teachings. Chief among these complaints was the syncretic practices of some of the disciples to worship images of Shakyamuni Buddha. Nikkō admonished other disciple priests for signing their names "Tendai Shamon" (of the Tendai Buddhist school) in documents they sent to the Kamakura government. Furthermore, Nikkō alleged that the other disciples disregarded some of Nichiren's writings written in Katakana rather than in Classical Chinese syllabary. Sanenaga defended his actions, claiming that it was customary for his political family to provide monetary donations and make homage to the Shinto shrine of the Kamakura shogunate. Nikō tolerated Sanenaga's acts, claiming that similar incidents occurred previously with the knowledge of Nichiren. Sanenaga sided with Nikō and Nikkō departed in 1289 from Minobu. He returned to his home in Suruga Province and established two temples: Taiseki-ji in the Fuji district and Honmonji in Omosu district. He spent most of his life at the latter, where he trained his followers. According to Stone, it is not absolutely clear that Nikkō intended to completely break from the other senior disciples and start his own school. However, his followers claimed that he was the only one of the six senior disciples who maintained the purity of Nichiren's legacy. Two documents appeared, first mentioned and discovered by Taiseki-ji High Priest Nikkyo Shonin in 1488, claiming Nichiren transferred his teaching exclusively to Nikkō but their authenticity has been questioned. Taiseki-ji does not dispute that the original documents are missing but holds that certified copies are preserved in their repositories. In contrast, other Nichiren sects vehemently claim them as forgeries since they are not in the original handwriting of Nichiren or Nikkō, holding they were copied down by Nikkō's disciples after his death." In addition to using the letters to defend its claim to othodoxy, the documents may have served to justify Taiseki-ji's claimed superiority over other Nikkō temples, especially Ikegami Honmon-ji, the site of Nichiren's tomb. Even though there had been efforts by temples of the Nikkō lineage in the late 19th century to unify into one single separate Nichiren school the "Kommon-ha", today's Nichiren Shōshū comprises only the Taiseki-ji temple and its dependent temples. It is not identical to the historical Nikkō or Fuji lineage. Parts of the "Kommon-ha", the "Honmon-Shu", eventually became part of Nichiren Shu in the 1950s. New religious movements like Sōka Gakkai, Shōshinkai, and Kenshōkai trace their origins to the Nichiren Shōshū school. In the early 14th century Hokkeshū followers spread the teachings westward and established congregations (Jpn. "shū") into the imperial capital of Kyoto and as far as Bizen and Bitchu. During this time there is documentation of face-to-face public debates between Hokkeshū and Nembutsu adherents. By the end of the century Hokkeshū temples had been founded all over Kyoto, only being outnumbered by Zen temples. The demographic base of support in Kyoto were members of the merchant class (Jpn. "machishū"), some of whom had acquired great wealth. Tanabe hypothesizes they were drawn to this faith because of Nichiren's emphasis on the "third realm" (Jpn. "daisan hōmon") of the Lotus Sutra, staked out in chapters 10-22, which emphasize practice in the mundane world. In the 15th century, the political and social order began to collapse and Hokkeshū followers armed themselves. The "Hokke-ikki" was an uprising in 1532 of Hokke followers against the followers of the Pure Land school in 1532. Initially successful it became the most powerful religious group in Kyoto but its fortunes were reversed in 1536 when Mt. Hiei armed forces destroyed twenty-one Hokkeshū temples and killed some 58,000 of its followers. In 1542 permission was granted by the government to rebuild the destroyed temples and the Hokke "machishū" played a crucial role in rebuilding the commerce, industry, and arts in Kyoto. Their influence in the arts and literature continued through the Momoyama (1568–1615) and Edo (1615–1868) periods and many of the most famous artists and literati were drawn from their ranks. Although the various sects of Nichiren Buddhism were administratively independent, there is evidence of cooperation between them. For example, in 1466 the major Hokke temples in Kyoto signed the Kanshō-era accord (Kanshō "meiyaku") to protect themselves against threats from Mt. Hiei. Despite strong sectarian differences, there is also evidence of interactions between Hokkeshū and Tendai scholar-monks. During the Edo period, with the consolidation of power by the Tokugawa shogunate, increased pressure was placed major Buddhist schools and Nichiren temples to conform to governmental policies. Some Hokkeshū adherents, the followers of the so-called Fuju-fuse lineage, adamantly bucked this policy based on their readings of Nichiren's teachings to neither take ("fuju") nor give ("fuse") offerings from non-believers. Suppressed, adherents often held their meetings clandestinely which led to the Fuju-fuse persecution and numerous executions of believers in 1668. During this time of persecution, most likely to prevent young priests from adopting a passion for propagation, Nichiren seminaries emphasized Tendai studies with only a few top-ranking students permitted to study some of Nichiren's writings. During the Edo period the majority of Hokkeshū temples were subsumed into the shogunate's Danka system, an imposed nationwide parish system designed to ensure religious peace and root out Christianity. In this system Buddhist temples, in addition to their ceremonial duties, were forced to carry out state administrative functions. Thereby they became agents of the government and were prohibited to engage in any missionary activities. Hokkeshū temples were now obligated, just like those of other Buddhist schools, to focus on funeral and memorial services ("Sōshiki bukkyō") as their main activity. Stagnation was often the price for the protected status. Nichiren Buddhism was deeply influenced by the transition from the Tokugawa (1600–1868) to Meiji (1868–1912) periods in nineteenth-century Japan. The changeover from early modern ("kinsei") to modern ("kindai") was marked by the transformation of late-feudal institutions into modern ones as well as the political transition from shogunal to imperial rule and the economic shift from national isolation to integration in the world economy. This entailed creating a centralized state, stitching together some 260 feudal domains ruled by hereditary leaders ("daimyō"), and moving from a caste social system to a meritocracy based on educational achievement. Although commonly perceived as a singular event called the Meiji Restoration, the transition was full of twists and turns that began in the later Tokugawa years and continued decades after the 1867–1868 demise of the shogunate and launch of imperial rule. By this time Japanese Buddhism was often characterized by syncretism in which local nativistic worship was incorporated into Buddhist practice. For example, Tendai, Shingon, Jodō, and Nichiren temples often had chapels within them dedicated to Inari Shinto worship. Within Nichiren Buddhism there was a phenomenon of "Hokke Shintō" (Lotus Shinto), closely influenced by Yoshida Shintō. Anti-Buddhist sentiment had been building throughout the latter part of the Tokugawa period (1603–1868). Scholars such as Tominaga Nakamoto and Hirata Atsutane attacked the theoretical roots of Buddhism. Critics included promoters of Confucianism, nativism, Shinto-inspired Restorationists, and modernizers. Buddhism was critiqued as a needless drain on public resources and also as an insidious foreign influence that had obscured the indigenous Japanese spirit. Under attack by two policies of the day, "shinbutsu bunri" (Separation of Shinto Deities and Buddhas) and "haibutsu kishaku" (Eradication of Buddhism), Japanese Buddhism during the Tokugawa-to-Meiji transition proved to be a crisis of survival. The new government promoted policies that reduced the material resources available to Buddhist temples and downgraded their role in the religious, political, and social life of the nation. The policies of "shibutsu bunri" were implemented at the local level throughout Japan but were particularly intense in three domains that were the most active in the Restoration: Satsuma, Choshii, and Tosa. In Satsuma, for example, by 1872 all of its 1000+ Buddhist temples had been abolished, their monks laicized, and their landholdings confiscated. Throughout the country thousands of Buddhist temples and, at a minimum, tens of thousands of Buddhist sutras, paintings, statues, temple bells and other ritual objects were destroyed, stolen, lost, or sold during the early years of the restoration. Starting in the second decade of the restoration, pushback against these policies came from Western powers interested in providing a safe harbor for Christianity and Buddhist leaders who proposed an alliance of Shinto and Buddhism to resist Christianity. As part of this accommodation, Buddhist priests were forced to promote key teachings of Shinto and provide support for national policies. Nichiren Buddhism, like the other Buddhist schools, struggled between accommodation and confrontation. The Nichiren scholar Udana-in Nichiki (1800–1859) argued for a policy of co-existence with other schools of Buddhism, Confucianism, Nativism, and European religions. His disciple Arai Nissatsu (1830–1888) forged an alliance of several Nichiren branches and became the first superintendent of the present Nichiren Shū which was incorporated in 1876. Nissatsu was active in Buddhist intersect cooperation to resist the government's hostile policies, adopted the government's "Great Teaching" policy that was Shinto-derived, and promoted intersectarian understanding. In the process, however, he reinterpreted some of Nichiren's important teachings. Among those arguing against accommodation were Nichiren scholar and lay believer Ogawa Taidō (1814–1878) and the cleric Honda Nisshō (1867–1931) of the Kempon Hokke denomination. After the above events and centuries of splintering based on dogma and institutional histories, the following major Nichiren temple schools, according to Matsunaga, were officially recognized in the Meiji era: Nichiren Buddhism went through many reforms in the Meiji Period during a time of persecution, Haibutsu kishaku (廃仏毀釈), when the government attempted to eradicate mainstream Japanese Buddhism. As a part of the Meiji Restoration, the interdependent Danka system between the state and Buddhist temples was dismantled which left the latter without its funding. Buddhist institutions had to align themselves to the new nationalistic agenda or perish. Many of these reform efforts were led by lay people. The trend toward lay centrality was prominent in Nichiren Buddhism as well, predating the Meiji period. Some Nichiren reformers in the Meiji period attempted to inject a nationalistic interpretation of Nichiren's teachings; others called for globalist perspectives. According to Japanese researcher "Yoshiro Tamura", the term "Nichirenism" applies broadly to the following three categories: Both Nichiren and his followers have been associated with fervent Japanese nationalism specifically identified as Nichirenism between the Meiji period and the conclusion of World War II. The nationalistic interpretation of Nichiren's teachings were inspired by lay Buddhist movements like Kokuchūkai and resulted in violent historical events such as the May 15 Incident and the League of Blood Incident. Among the key proponents of this interpretation are Chigaku Tanaka who founded the Kokuchūkai (English: Nation's Pillar Society). Tanaka was charismatic and through his writings and lecturers attracted many followers such as Kanji Ishiwara. Nisshō Honda advocated the unification of Japanese Buddhists to support the imperial state. Other ultra-nationalist activists who based their ideas on Nichiren were Ikki Kita and Nisshō Inoue. Nichirenism also includes several intellectuals and activists who reacted against the prewar ultranationalistic interpretations and argued for an egalitarian and socialist vision of society based on Nichiren's teachings and the Lotus Sutra. These figures ran against the growing tide of Japanese militarism and were subjected to political harassment and persecution. A leading figure in this group was Girō Seno who formed the New Buddhist Youth League ("Shinkō Bukkyō Seinen Dōmei"). Originally influenced by the ideals of Tanaka and Honda, Giro Seno came to reject ultra-nationalism and argued for humanism, socialism, pacifism, and democracy as a new interpretation of Nichiren's beliefs. He was imprisoned for two years under the National Security Act. The same fate was also endured by Tsunesaburo Makiguchi, who refused the religious dictum of Shinto display accepted by Nichiren Shoshu for the "Soka Kyoiku Gakkai", his lay organization composed of primarily secretaries and teachers until it grew to become Soka Gakkai after World War II. Several Nichiren-inspired religious movements arose and appealed primarily to this segment of society with a message of alleviating suffering salvation for many poor urban workers. Honmon Butsuryū-shū, an early example of lay-based religious movements of the modern period inspired by Nichiren, was founded several years before the Meiji Restoration. Reiyukai, Rissho Koseikai stemming from Nichiren Shu while Kenshokai and Soka Gakkai, once affiliated with Nichiren Shoshu, are more recent examples of lay-inspired movements drawing from Nichiren's teachings and life. Nichiren Buddhism has had a major impact on Japan's literary and cultural life. Japanese literary figure Takayama Chogyū and children's author Kenji Miyazawa praised Nichiren's teachings. A prominent researcher, Masaharu Anesaki, was encouraged to study Nichiren which led to the work "Nichiren: The Buddhist Prophet" which introduced Nichiren to the West. Non-Buddhist Japanese individuals such as Uchimura Kanzō listed Nichiren as one of five historical figures who best represented Japan, while Tadao Yanaihara described Nichiren as one of the four historical figures he most admired. While various sects and organizations have had a presence in nations outside Japan for over a century, the ongoing expansion of Nichiren Buddhism overseas started in 1960 when Soka Gakkai president Daisaku Ikeda initiated his group's worldwide propagation efforts growing from a few hundred transplanted Japanese to over 3500 families just by 1962. Nichiren Buddhism is now practiced in many countries outside of Japan. In the United States, religious studies scholar Charles S. Prebish coined the typology of "two Buddhisms" to delineate the divide between forms of Buddhism that appealed either primarily to people of the Asian diaspora or to Euro-American converts. Nattier, on the other hand, proposes a three-way typology. "Import" or "elite" Buddhism refers to a class of people who have the time and means to seek Buddhist teachers to appropriate certain Buddhist techniques such as meditation. "Export or evangelical" Buddhism refers to groups that actively proselytize for new members in their local organizations. "Baggage" or "ethnic" Buddhism refers to diaspora Buddhists, usually of a single ethnic group, who have relocated more for social and economic advancement than for evangelical purposes. Another taxonomy divides Western Buddhist groups into three different categories: evangelical, church-like, and meditational. Nichiren Shu has been classified into the church-like category. One of several Japanese Buddhist schools that followed in the wake of Japanese military conquest and colonization, Nichiren Shu opened a temple in Pusan, Korea in 1881. Its fortunes rose and diminished with the political tides but eventually failed. It also established missions in Sakhalin, Manchuria, and Taiwan. A Nichiren Shu mission was established in Hawaii in 1900. By 1920 it established temples at Pahala, Honolulu, Wailuku and Maui. In 1955 it officially started a mission in Brazil. In 1991 it established the Nichiren Buddhist International Center in 1991 and in 2002 built a center in Hayward, California, to help overseas missions. However, Nichiren Shu does not widely propagate in the West. Some have characterized the Soka Gakkai as evangelical but others claim that it broke out of the "Two Buddhisms" paradigm. It is quite multi-ethnic and it has taken hold among native populations in locations including Korea, Malaysia, Brazil, Europe, parts of Africa, India, and North America. The growth of the Soka Gakkai was sparked by repeated missionary trips beginning in the early 1960s by Daisaku Ikeda, its third president. In 1975 the Soka Gakkai International was launched in Guam. In the United States it has attracted a diverse membership including a significant demographic of African Americans. Since the 1970s it has created institutions, publications and exhibitions to support its overall theme of "peace, culture, and education." There is academic research on various national organizations affiliated with this movement: the United States, the United Kingdom, Italy, Canada, Brazil, Scotland, Southeast Asia, Germany, and Thailand. The Rissho Kosei Kai focuses on using its teachings to promote a culture of religiosity through inter-religious dialogue. In 1967, it launched the "Faith to All Men Movement" to awaken a globalized religiosity. It has over 2 million members and 300 Dharma centers in 20 countries throughout the world including Frankfurt and Moorslede. It is active in interfaith organizations, including the International Association for Religious Freedom (IARF) and Religions for Peace (WCRP). It has consultative states with the United Nations and since 1983 issues an annual Peace Prize to individuals or organizations worldwide that work for peace and development and promote interreligious cooperation. The Reiyukai conducts more typical missionary activities in the West. It has a membership of between five hundred and one thousand members in Europe, concentrated in Italy, Spain, England and France. The approximately 1,500 members of the Nihonzan Myohoji have built peace pagodas, conducted parades beating the drum while chanting the daimoku, and encouraged themselves and others to create world peace. Nichiren Shoshu has six temples in the United States led by Japanese priests and supported by lay Asians and non-Asians. There is one temple in Brazil and the residing priest serves as a "circuit rider" to attend to other locations. The following lists are based on English-language Wikipedia articles and the Japanese Wikipedia article on . In alphabetical order (Japanese characters preceded by "ja:" link to articles in the Japanese Wikipedia). In alphabetical order (Japanese characters preceded by "ja:" link to articles in the Japanese Wikipedia):
https://en.wikipedia.org/wiki?curid=22137
Newport News Shipbuilding Newport News Shipbuilding (NNS), a division of Huntington Ingalls Industries, is the largest industrial employer in Virginia, and sole designer, builder and refueler of United States Navy aircraft carriers and one of two providers of U.S. Navy submarines. Founded as the Chesapeake Dry Dock and Construction Co. in 1886, Newport News Shipbuilding has built more than 800 ships, including both naval and commercial ships. Located in the city of Newport News, its facilities span more than , strategically positioned in one of the great harbors of the East Coast. The shipyard is a major employer, not only for the lower Virginia Peninsula, but also portions of Hampton Roads south of the James River and the harbor, portions of the Middle Peninsula region, and even some northeastern counties of North Carolina. The shipyard is building the aircraft carriers and . In 2013, Newport News Shipbuilding began the deactivation of the first nuclear-powered aircraft carrier, , which it also built. Newport News Shipbuilding also performs refueling and complex overhaul (RCOH) work on s. This is a four-year vessel renewal program that not only involves refueling of the vessel's nuclear reactors but also includes modernization work. The yard has completed RCOH for four "Nimitz"-class carriers (, , and ). As of May 2016 this work was underway for the fifth "Nimitz"-class vessel, . As of November 2017 this work was underway for the sixth "Nimitz"-class vessel, . Industrialist Collis P. Huntington (1821–1900) provided crucial funding to complete the Chesapeake and Ohio Railroad (C&O) from Richmond, Virginia to the Ohio River in the early 1870s. Although originally built for general commerce, this C&O rail link to the midwest was soon also being used to transport bituminous coal from the previously isolated coalfields, adjacent to the New River and the Kanawha River in West Virginia. In 1881, the Peninsula Extension of the C&O was built from Richmond down the Virginia Peninsula to reach a new coal pier on Hampton Roads in Warwick County near the small unincorporated community of Newport News Point. However, building the railroad and coal pier was only the first part of Huntington's dreams for Newport News. In 1886, Huntington built a shipyard to repair ships servicing this transportation hub. In 1891 Newport News Shipbuilding and Drydock Company delivered its first ship, the tugboat "Dorothy". By 1897 NNS had built three warships for the US Navy: , and . When Collis died in 1900, his nephew Henry E. Huntington inherited much of his uncle's fortune. He also married Collis' widow Arabella Huntington, and assumed Collis' leadership role with Newport News Shipbuilding and Drydock Company. Under Henry Huntington's leadership, growth continued. In 1906 the revolutionary launched a great naval race worldwide. Between 1907 and 1923, Newport News built six of the US Navy's total of 22 dreadnoughts – , , , , and . All but the first were in active service in World War II. In 1907 President Theodore Roosevelt sent the Great White Fleet on its round-the-world voyage. NNS had built seven of its 16 battleships. In 1914 NNS built SS "Medina" for the Mallory Steamship Company; as she was until 2009 the world's oldest active ocean-faring passenger ship. In the early years, leaders of the Newport News community and those of the shipyard were virtually interchangeable. Shipyard president Walter A. Post served from March 9, 1911 to February 12, 1912, when he died. Earlier, he had come to the area as one of the builders of the C&O Railway's terminals, and had served as the first mayor of Newport News after it became an independent city in 1896. It was on March 14, 1914 that Albert Lloyd Hopkins, a young New Yorker trained in engineering, succeeded Post as president of the company. In May 1915 while traveling to England on shipyard business aboard , Albert L. Hopkins tenure and life ended prematurely when that ship was torpedoed and sunk by a German U-boat off Queenstown on the Irish coast. His assistant, Frederic Gauntlett, was also on board, but was able to swim to safety. Homer Lenoir Ferguson was company vice president when Hopkins died, and assumed the presidency the following August. He saw the company through both world wars, became a noted community leader, and was a co-founder of the Mariners' Museum with Archer Huntington. He served until July 31, 1946, after World War II had ended on both the European and Pacific fronts. Just northwest of the shipyard, Hilton Village, one of the first planned communities in the country, was built by the federal government to house shipyard workers in 1918. The planners met with the wives of shipyard workers. Based on their input 14 house plans were designed for the projected 500 English-village-style homes. After the war, in 1922, Henry Huntington acquired it from the government, and helped facilitate the sale of the homes to shipyard employees and other local residents. Three streets there were named after Post, Hopkins, and Ferguson. The "Lusitania" incident was among the events that brought the United States into World War I. Between 1918 and 1920 NNS delivered 25 destroyers, and after the war it began building aircraft carriers. was delivered in 1934, and NNS went on to build and . After World War I NNS completed a major reconditioning and refurbishment of the ocean liner . Before the war she had been the German liner "Vaterland", but the start of hostilities found her laid up in New York Harbor and she had been seized by the US Government in 1917 and converted into a troopship. War duty and age meant that all wiring, plumbing, and interior layouts were stripped and redesigned while her hull was strengthened and her boilers converted from coal to oil while being refurbished. Virtually a new ship emerged from NNS in 1923, and SS "Leviathan" became the flagship of United States Lines. In 1927 NNS launched the world's first significant turbo-electric ocean liner: Panama Pacific Line's . At the time she was also the largest merchant ship yet built in the United States, although she was a modest size compared with the biggest European liners of her era. NNS launched "California"s sister ships "Virginia" in 1928 and "Pennsylvania" in 1929. NNS followed them by launching two even larger turbo-electric liners for Dollar Steamship Company: the in 1930, followed by her sister in 1931. was launched in 1939 and entered service with United States lines shortly before World War II but soon returned to the shipyard for conversion to a troopship, USS "West Point". By 1940 the Navy had ordered a battleship, seven more aircraft carriers and four cruisers. During World War II, NNS built ships as part of the U.S. Government's Emergency Shipbuilding Program, and swiftly filled requests for "Liberty ships" that were needed during the war. It founded the North Carolina Shipbuilding Company, an emergency yard on the banks of the Cape Fear River and launched its first Liberty ship before the end of 1941, building 243 ships in all, including 186 Libertys. For its contributions during the war, the Navy awarded the company its "E" pennant for excellence in shipbuilding. NNS ranked 23rd among United States corporations in the value of wartime production contracts. In the post-war years NNS built the famous passenger liner , which set a transatlantic speed record that still stands today. In 1954 NNS, Westinghouse and the Navy developed and built a prototype nuclear reactor for a carrier propulsion system. NNS designed in 1960. In 1959 NNS launched its first nuclear-powered submarine, . In the 1970s, NNS launched two of the largest tankers ever built in the western hemisphere and also constructed three liquefied natural gas carriers – at over 390,000 deadweight tons, the largest ever built in the United States. NNS and Westinghouse Electric Company jointly form Offshore Power Systems to build floating nuclear power plants for Public Service Electric and Gas Company. In the 1980s, NNS produced a variety of Navy products, including nuclear aircraft carriers and nuclear attack submarines. Since 1999 the shipyard has produced only warships for the Navy. In 2007, the US Navy found that workers had used incorrect metal to fuse together pipes and joints on submarines under construction and this could have led to cracking and leaks. In 2009 it was found that bolts and fasteners in weapons-handling systems on four Navy submarines, , , , and , were installed incorrectly, delaying the launching of the boats while the problems were corrected. In 1968, Newport News merged with Tenneco Corporation. In 1996, Tenneco initiated a spinoff of Newport News into an independent company (Newport News Shipbuilding). The company was purchased by Northrop Grumman in 2001 for $2.6 billion and renamed "Northrop Grumman Newport News". This division was merged with Northrop Grumman Ship Systems in 2008 and given the name "Northrop Grumman Shipbuilding". Three years later, the company was spun off as Huntington Ingalls Industries, Inc., which trades under the symbol HII on the New York Stock Exchange. Other ships built at the Newport News yard include:
https://en.wikipedia.org/wiki?curid=22141
Newton's method In numerical analysis, Newton's method, also known as the Newton–Raphson method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots (or zeroes) of a real-valued function. The most basic version starts with a single-variable function defined for a real variable , the function's derivative , and an initial guess for a root of . If the function satisfies sufficient assumptions and the initial guess is close, then is a better approximation of the root than . Geometrically, is the intersection of the -axis and the tangent of the graph of at : that is, the improved guess is the unique root of the linear approximation at the initial point. The process is repeated as until a sufficiently precise value is reached. This algorithm is first in the class of Householder's methods, succeeded by Halley's method. The method can also be extended to complex functions and to systems of equations. The idea is to start with an initial guess which is reasonably close to the true root, then to approximate the function by its tangent line using calculus, and finally to compute the -intercept of this tangent line by elementary algebra. This -intercept will typically be a better approximation to the original function's root than the first guess, and the method can be iterated. More formally, suppose is a differentiable function defined on the interval with values in the real numbers , and we have some current approximation . Then we can derive the formula for a better approximation, by referring to the diagram on the right. The equation of the tangent line to the curve at is where denotes the derivative. The -intercept of this line (the value of which makes ) is taken as the next approximation, , to the root, so that the equation of the tangent line is satisfied when formula_4: Solving for gives We start the process with some arbitrary initial value . (The closer to the zero, the better. But, in the absence of any intuition about where the zero might lie, a "guess and check" method might narrow the possibilities to a reasonably small interval by appealing to the intermediate value theorem.) The method will usually converge, provided this initial guess is close enough to the unknown zero, and that . Furthermore, for a zero of multiplicity 1, the convergence is at least quadratic (see rate of convergence) in a neighbourhood of the zero, which intuitively means that the number of correct digits roughly doubles in every step. More details can be found in the analysis section below. Householder's methods are similar but have higher order for even faster convergence. However, the extra computations required for each step can slow down the overall performance relative to Newton's method, particularly if or its derivatives are computationally expensive to evaluate. The name "Newton's method" is derived from Isaac Newton's description of a special case of the method in "De analysi per aequationes numero terminorum infinitas" (written in 1669, published in 1711 by William Jones) and in "De metodis fluxionum et serierum infinitarum" (written in 1671, translated and published as "Method of Fluxions" in 1736 by John Colson). However, his method differs substantially from the modern method given above: Newton applies the method only to polynomials. He does not compute the successive approximations , but computes a sequence of polynomials, and only at the end arrives at an approximation for the root . Finally, Newton views the method as purely algebraic and makes no mention of the connection with calculus. Newton may have derived his method from a similar but less precise method by Vieta. The essence of Vieta's method can be found in the work of the Persian mathematician Sharaf al-Din al-Tusi, while his successor Jamshīd al-Kāshī used a form of Newton's method to solve to find roots of (Ypma 1995). A special case of Newton's method for calculating square roots was known since ancient times and is often called the Babylonian method. Newton's method was used by 17th-century Japanese mathematician Seki Kōwa to solve single-variable equations, though the connection with calculus was missing. Newton's method was first published in 1685 in "A Treatise of Algebra both Historical and Practical" by John Wallis. In 1690, Joseph Raphson published a simplified description in "Analysis aequationum universalis". Raphson again viewed Newton's method purely as an algebraic method and restricted its use to polynomials, but he describes the method in terms of the successive approximations instead of the more complicated sequence of polynomials used by Newton. Finally, in 1740, Thomas Simpson described Newton's method as an iterative method for solving general nonlinear equations using calculus, essentially giving the description above. In the same publication, Simpson also gives the generalization to systems of two equations and notes that Newton's method can be used for solving optimization problems by setting the gradient to zero. Arthur Cayley in 1879 in "The Newton–Fourier imaginary problem" was the first to notice the difficulties in generalizing Newton's method to complex roots of polynomials with degree greater than 2 and complex initial values. This opened the way to the study of the theory of iterations of rational functions. Newton's method is an extremely powerful technique—in general the convergence is quadratic: as the method converges on the root, the difference between the root and the approximation is squared (the number of accurate digits roughly doubles) at each step. However, there are some difficulties with the method. Newton's method requires that the derivative can be calculated directly. An analytical expression for the derivative may not be easily obtainable or could be expensive to evaluate. In these situations, it may be appropriate to approximate the derivative by using the slope of a line through two nearby points on the function. Using this approximation would result in something like the secant method whose convergence is slower than that of Newton's method. It is important to review the proof of quadratic convergence of Newton's method before implementing it. Specifically, one should review the assumptions made in the proof. For situations where the method fails to converge, it is because the assumptions made in this proof are not met. If the first derivative is not well behaved in the neighborhood of a particular root, the method may overshoot, and diverge from that root. An example of a function with one root, for which the derivative is not well behaved in the neighborhood of the root, is for which the root will be overshot and the sequence of will diverge. For , the root will still be overshot, but the sequence will oscillate between two values. For , the root will still be overshot but the sequence will converge, and for the root will not be overshot at all. In some cases, Newton's method can be stabilized by using successive over-relaxation, or the speed of convergence can be increased by using the same method. If a stationary point of the function is encountered, the derivative is zero and the method will terminate due to division by zero. A large error in the initial estimate can contribute to non-convergence of the algorithm. To overcome this problem one can often linearise the function that is being optimized using calculus, logs, differentials, or even using evolutionary algorithms, such as the stochastic funnel algorithm. Good initial estimates lie close to the final globally optimal parameter estimate. In nonlinear regression, the sum of squared errors (SSE) is only "close to" parabolic in the region of the final parameter estimates. Initial estimates found here will allow the Newton–Raphson method to quickly converge. It is only here that the Hessian matrix of the SSE is positive and the first derivative of the SSE is close to zero. In a robust implementation of Newton's method, it is common to place limits on the number of iterations, bound the solution to an interval known to contain the root, and combine the method with a more robust root finding method. If the root being sought has multiplicity greater than one, the convergence rate is merely linear (errors reduced by a constant factor at each step) unless special steps are taken. When there are two or more roots that are close together then it may take many iterations before the iterates get close enough to one of them for the quadratic convergence to be apparent. However, if the multiplicity formula_8 of the root is known, the following modified algorithm preserves the quadratic convergence rate: This is equivalent to using successive over-relaxation. On the other hand, if the multiplicity of the root is not known, it is possible to estimate formula_8 after carrying out one or two iterations, and then use that value to increase the rate of convergence. Suppose that the function has a zero at , i.e., , and is differentiable in a neighborhood of . If is continuously differentiable and its derivative is nonzero at , then there exists a neighborhood of such that for all starting values in that neighborhood, the sequence will converge to . If the function is continuously differentiable and its derivative is not 0 at and it has a second derivative at then the convergence is quadratic or faster. If the second derivative is not 0 at then the convergence is merely quadratic. If the third derivative exists and is bounded in a neighborhood of , then: where If the derivative is 0 at , then the convergence is usually only linear. Specifically, if is twice continuously differentiable, and , then there exists a neighborhood of such that for all starting values in that neighborhood, the sequence of iterates converges linearly, with rate 1/2 Alternatively if and for ,  in a neighborhood of , being a zero of multiplicity , and if then there exists a neighborhood of such that for all starting values in that neighborhood, the sequence of iterates converges linearly. However, even linear convergence is not guaranteed in pathological situations. In practice these results are local, and the neighborhood of convergence is not known in advance. But there are also some results on global convergence: for instance, given a right neighborhood of , if is twice differentiable in and if , in , then, for each in the sequence is monotonically decreasing to . According to Taylor's theorem, any function which has a continuous second derivative can be represented by an expansion about a point that is close to a root of . Suppose this root is . Then the expansion of about is: where the Lagrange form of the Taylor series expansion remainder is where is in between and . Since is the root, () becomes: Dividing equation () by and rearranging gives Remembering that is defined by one finds that That is, Taking absolute value of both sides gives Equation () shows that the rate of convergence is at least quadratic if the following conditions are satisfied: The term "sufficiently" close in this context means the following: \right | for some ; Finally, () can be expressed in the following way: where is the supremum of the variable coefficient of on the interval defined in condition 1, that is: The initial point has to be chosen such that conditions 1 to 3 are satisfied, where the third condition requires that . The disjoint subsets of the basins of attraction—the regions of the real number line such that within each region iteration from any point leads to one particular root—can be infinite in number and arbitrarily small. For example, for the function , the following initial conditions are in successive basins of attraction: Newton's method is only guaranteed to converge if certain conditions are satisfied. If the assumptions made in the proof of quadratic convergence are met, the method will converge. For the following subsections, failure of the method to converge indicates that the assumptions made in the proof were not met. In some cases the conditions on the function that are necessary for convergence are satisfied, but the point chosen as the initial point is not in the interval where the method converges. This can happen, for example, if the function whose root is sought approaches zero asymptotically as goes to or . In such cases a different method, such as bisection, should be used to obtain a better estimate for the zero to use as an initial point. Consider the function: It has a maximum at and solutions of at . If we start iterating from the stationary point (where the derivative is zero), will be undefined, since the tangent at (0,1) is parallel to the -axis: The same issue occurs if, instead of the starting point, any iteration point is stationary. Even if the derivative is small but not zero, the next iteration will be a far worse approximation. For some functions, some starting points may enter an infinite cycle, preventing convergence. Let and take 0 as the starting point. The first iteration produces 1 and the second iteration returns to 0 so the sequence will alternate between the two without converging to a root. In fact, this 2-cycle is stable: there are neighborhoods around 0 and around 1 from which all points iterate asymptotically to the 2-cycle (and hence not to the root of the function). In general, the behavior of the sequence can be very complex (see Newton fractal). The real solution of this equation is …. If the function is not continuously differentiable in a neighborhood of the root then it is possible that Newton's method will always diverge and fail, unless the solution is guessed on the first try. A simple example of a function where Newton's method diverges is trying to find the cube root of zero. The cube root is continuous and infinitely differentiable, except for , where its derivative is undefined: For any iteration point , the next iteration point will be: The algorithm overshoots the solution and lands on the other side of the -axis, farther away than it initially was; applying Newton's method actually doubles the distances from the solution at each iteration. In fact, the iterations diverge to infinity for every , where . In the limiting case of (square root), the iterations will alternate indefinitely between points and , so they do not converge in this case either. If the derivative is not continuous at the root, then convergence may fail to occur in any neighborhood of the root. Consider the function Its derivative is: Within any neighborhood of the root, this derivative keeps changing sign as approaches 0 from the right (or from the left) while for . So is unbounded near the root, and Newton's method will diverge almost everywhere in any neighborhood of it, even though: In some cases the iterates converge but do not converge as quickly as promised. In these cases simpler methods converge just as quickly as Newton's method. If the first derivative is zero at the root, then convergence will not be quadratic. Let then and consequently So convergence is not quadratic, even though the function is infinitely differentiable everywhere. Similar problems occur even when the root is only "nearly" double. For example, let Then the first few iterations starting at are it takes six iterations to reach a point where the convergence appears to be quadratic. If there is no second derivative at the root, then convergence may fail to be quadratic. Let Then And except when where it is undefined. Given , which has approximately times as many bits of precision as has. This is less than the 2 times as many which would be required for quadratic convergence. So the convergence of Newton's method (in this case) is not quadratic, even though: the function is continuously differentiable everywhere; the derivative is not zero at the root; and is infinitely differentiable except at the desired root. When dealing with complex functions, Newton's method can be directly applied to find their zeroes. Each zero has a basin of attraction in the complex plane, the set of all starting values that cause the method to converge to that particular zero. These sets can be mapped as in the image shown. For many complex functions, the boundaries of the basins of attraction are fractals. In some cases there are regions in the complex plane which are not in any of these basins of attraction, meaning the iterates do not converge. For example, if one uses a real initial condition to seek a root of , all subsequent iterates will be real numbers and so the iterations cannot converge to either root, since both roots are non-real. In this case almost all real initial conditions lead to chaotic behavior, while some initial conditions iterate either to infinity or to repeating cycles of any finite length. Curt McMullen has shown that for any possible purely iterative algorithm similar to Newton's method, the algorithm will diverge on some open regions of the complex plane when applied to some polynomial of degree 4 or higher. However, McMullen gave a generally convergent algorithm for polynomials of degree 3. One may also use Newton's method to solve systems of (nonlinear) equations, which amounts to finding the zeroes of continuously differentiable functions . In the formulation given above, one then has to left multiply with the inverse of the Jacobian matrix instead of dividing by : Rather than actually computing the inverse of the Jacobian matrix, one can save time by solving the system of linear equations for the unknown . The -dimensional variant of Newton's method can be used to solve systems of greater than (nonlinear) equations as well if the algorithm uses the generalized inverse of the non-square Jacobian matrix instead of the inverse of . If the nonlinear system has no solution, the method attempts to find a solution in the non-linear least squares sense. See Gauss–Newton algorithm for more information. Another generalization is Newton's method to find a root of a functional defined in a Banach space. In this case the formulation is where is the Fréchet derivative computed at . One needs the Fréchet derivative to be boundedly invertible at each in order for the method to be applicable. A condition for existence of and convergence to a root is given by the Newton–Kantorovich theorem. In -adic analysis, the standard method to show a polynomial equation in one variable has a -adic root is Hensel's lemma, which uses the recursion from Newton's method on the -adic numbers. Because of the more stable behavior of addition and multiplication in the -adic numbers compared to the real numbers (specifically, the unit ball in the -adics is a ring), convergence in Hensel's lemma can be guaranteed under much simpler hypotheses than in the classical Newton's method on the real line. The Newton–Fourier method is Joseph Fourier's extension of Newton's method to provide bounds on the absolute error of the root approximation, while still providing quadratic convergence. Assume that is twice continuously differentiable on and that contains a root in this interval. Assume that on this interval (this is the case for instance if , , and , and on this interval). This guarantees that there is a unique root on this interval, call it . If it is concave down instead of concave up then replace by since they have the same roots. Let be the right endpoint of the interval and let be the left endpoint of the interval. Given , define which is just Newton's method as before. Then define where the denominator is and not . The iterations will be strictly decreasing to the root while the iterations will be strictly increasing to the root. Also, so that distance between and decreases quadratically. When the Jacobian is unavailable or too expensive to compute at every iteration, a quasi-Newton method can be used. Newton's method can be generalized with the q-analog of the usual derivative. A nonlinear equation has multiple solutions in general. But if the initial value is not appropriate, Newton's method may not converge to the desired solution or may converge to the same solution found earlier. When we have already found N solutions of formula_38, then the next root can be found by applying Newton's method to the next equation: This method is applied to obtain zeros of the Bessel function of the second kind. Hirano's modified Newton method is a modification conserving the convergence of Newton method and avoiding unstableness. It is developed to solve complex polynomials. Combining Newton's method with interval arithmetic is very useful in some contexts. This provides a stopping criterion that is more reliable than the usual ones (which are a small value of the function or a small variation of the variable between consecutive iterations). Also, this may detect cases where Newton's method converges theoretically but diverges numerically because of an insufficient floating-point precision (this is typically the case for polynomials of large degree, where a very small change of the variable may change dramatically the value of the function; see Wilkinson's polynomial). Consider formula_40, where formula_41 is a real interval, and suppose that we have an interval extension formula_42 of formula_43, meaning that formula_42 takes as input an interval formula_45 and outputs an interval formula_46 such that: We also assume that formula_48, so in particular formula_49 has at most one root in formula_41. We then define the interval Newton operator by: where formula_52. Note that the hypothesis on formula_42 implies that formula_54 is well defined and is an interval (see interval arithmetic for further details on interval operations). This naturally leads to the following sequence: The mean value theorem ensures that if there is a root of formula_49 in formula_57, then it is also in formula_58. Moreover, the hypothesis on formula_42 ensures that formula_58 is at most half the size of formula_57 when formula_8 is the midpoint of formula_63, so this sequence converges towards formula_64, where formula_65 is the root of formula_49 in formula_41. If formula_68 strictly contains formula_69, the use of extended interval division produces a union of two intervals for formula_70 ; multiple roots are therefore automatically separated and bounded. Newton's method can be used to find a minimum or maximum of a function formula_71. The derivative is zero at a minimum or maximum, so local minima and maxima can be found by applying Newton's method to the derivative. The iteration becomes: An important application is Newton–Raphson division, which can be used to quickly find the reciprocal of a number , using only multiplication and subtraction, that is to say the number such that . We can rephrase that as finding the zero of . We have . Newton's iteration is Therefore, Newton's iteration needs only two multiplications and one subtraction. This method is also very efficient to compute the multiplicative inverse of a power series. Many transcendental equations can be solved using Newton's method. Given the equation with and/or a transcendental function, one writes The values of that solve the original equation are then the roots of , which may be found via Newton's method. Newton's method is applied to the ratio of Bessel functions in order to obtain its root. A numerical verification for solutions of nonlinear equations has been established by using Newton's method multiple times and forming a set of solution candidates. An iterative Newton-Raphson procedure was employed in order to impose a stable Dirichlet boundary condition in CFD, as a quite general strategy to model current and potential distribution for electrochemical cell stacks. Consider the problem of finding the square root of a number "a", that is to say the positive number such that . Newton's method is one of many methods of computing square roots. We can rephrase that as finding the zero of . We have . For example, for finding the square root of 612 with an initial guess , the sequence given by Newton's method is: where the correct digits are underlined. With only a few iterations one can obtain a solution accurate to many decimal places. Rearranging the formula as follows yields the Babylonian method of finding square roots: i.e. the arithmetic mean of the guess, and . Consider the problem of finding the positive number with . We can rephrase that as finding the zero of . We have . Since for all and for , we know that our solution lies between 0 and 1. For example, with an initial guess , the sequence given by Newton's method is (note that a starting value of 0 will lead to an undefined result, showing the importance of using a starting point that is close to the solution): The correct digits are underlined in the above example. In particular, is correct to 12 decimal places. We see that the number of correct digits after the decimal point increases from 2 (for ) to 5 and 10, illustrating the quadratic convergence. The following is an implementation example of the Newton's method in the Julia programming language for finding a root of a function codice_1 which has derivative codice_2. The initial guess will be and the function will be so that . Each new iteration of Newton's method will be denoted by codice_3. We will check during the computation whether the denominator (codice_4) becomes too small (smaller than codice_5), which would be the case if , since otherwise a large amount of error could be introduced. x0 = 1 # The initial guess f(x) = x^2 - 2 # The function whose root we are trying to find fprime(x) = 2 * x # The derivative of the function tolerance = 10^(-7) # 7 digit accuracy is desired epsilon = 10^(-14) # Do not divide by a number smaller than this maxIterations = 20 # Do not allow the iterations to continue indefinitely solutionFound = false # Have not converged to a solution yet for i = 1:maxIterations end if solutionFound else end
https://en.wikipedia.org/wiki?curid=22145
New Order (band) New Order are an English rock band formed in 1980 by vocalist and guitarist Bernard Sumner, bassist Peter Hook, and drummer Stephen Morris. The band formed after the demise of Joy Division, following the suicide of lead singer Ian Curtis; they were joined by Gillian Gilbert on keyboards later that year. New Order's integration of post-punk with electronic and dance music made them one of the most acclaimed and influential bands of the 1980s. They were the flagship band for Manchester-based independent record label Factory Records and its nightclub The Haçienda, and worked in long-term collaboration with graphic designer Peter Saville. While the band's early years were overshadowed by the legacy of Joy Division, their experience of the early 1980s New York club scene saw them increasingly incorporate dance rhythms and electronic instrumentation into their work. Their 1983 hit "Blue Monday" became the best-selling 12-inch single of all time and a popular club track. In the 1980s, they released successful albums such as "Power, Corruption & Lies" (1983), "Technique" (1989), and the singles compilation "Substance" (1987). They briefly disbanded in 1993 to work on individual projects before reuniting in 1998. In the years since, New Order has gone through various hiatuses and personnel changes, most prominently the departure of Hook in 2007. They released their tenth studio album, "Music Complete", in 2015. Between 1977 and 1980, Ian Curtis, Peter Hook, Stephen Morris, and Bernard Sumner were members of the post-punk band Joy Division, often featuring heavy production input from producer Martin Hannett. Curtis took his own life on 18 May 1980, the day before Joy Division were scheduled to depart for their first American tour, and prior to the release of the band's second album, "Closer". The rest of the band decided soon after Curtis's death that they would carry on. Prior to his death, the members of Joy Division had agreed not to continue under the Joy Division name should any one member leave. On 29 July 1980, the still unnamed trio debuted live at Manchester's Beach Club. Rob Gretton, the band's manager for over twenty years, is credited for having found the name "New Order" in an article in "The Guardian" entitled "The People's New Order of Kampuchea". The band adopted this name, despite its previous use for former Stooge Ron Asheton's band The New Order. The group states that the name New Order (as was also the case with "Joy Division") does not draw a direct line to National Socialism or Fascism. The band rehearsed with each member taking turns on vocals. Sumner ultimately took the role, as he could sing when he wasn't playing his guitar. They wanted to complete the line-up with someone they knew well and whose musical skill and style was compatible with their own. Gretton suggested Morris's girlfriend Gillian Gilbert, and she was invited to join the band in early October 1980, as keyboardist and guitarist. Her first live performance with the band occurred at The Squat in Manchester on 25 October 1980. The initial release as New Order was the single "Ceremony", backed with "In a Lonely Place". These two songs were written in the weeks before Curtis took his own life. With the release of "Movement" in November 1981, New Order initially started on a similar route as their previous incarnation, performing dark, melodic songs, albeit with an increased use of synthesisers. The band viewed the period as a low point, as they were still reeling from Curtis' death. Hook commented that the only positive thing to come out of the "Movement" sessions was that producer Martin Hannett had showed the band how to use a mixing board, which allowed them to produce records by themselves from then on. More recently, Hook indicated a change of heart: "I think "Movement" gets a raw deal in general really – for me, when you consider the circumstances in which it was written, it is a fantastic record." New Order visited New York City again in 1981, where the band were introduced to post-disco, freestyle and electro. The band had taken to listening to Italian disco to cheer themselves up, while Morris taught himself drum programming. The singles that followed, "Everything's Gone Green" and "Temptation", saw a change in direction toward dance music. The Haçienda, Factory Records' own nightclub (largely funded by New Order) opened in May 1982 in Manchester and was even issued a Factory catalogue number: FAC51. The opening of UK's first ever superclub was marked by a nearly 23-minute instrumental piece originally entitled "Prime 5 8 6", but released 15 years later as "Video 5 8 6". Composed primarily by Sumner and Morris, "Prime 5 8 6"/"Video 5 8 6" was an early version of "5 8 6" that contained rhythm elements that would later surface on "Blue Monday" and "Ultraviolence". "Power, Corruption & Lies", released in May 1983, was a synthesiser-based outing and a dramatic change in sound from Joy Division and the preceding album, although the band had been hinting at the increased use of technology during the music-making process for a number of years then, including their work as Joy Division. Starting from what earlier singles had hinted, this was where the band had found their footing, mixing early techno music with their earlier guitar-based sound and showing the strong influence of acts like Kraftwerk and Giorgio Moroder. Even further in this direction was the electronically sequenced, four-on-the-floor single "Blue Monday". Inspired by Klein + M.B.O.'s "Dirty Talk" and Sylvester's disco classic, "You Make Me Feel (Mighty Real)", "Blue Monday" became the best-selling independent 12" single of all time in the UK; however, (much to the chagrin of the buying public) it was not on the track list of "Power, Corruption & Lies". This resulted in a sticker being applied to unsold copies of "Power, Corruption & Lies" album saying, "DOES NOT CONTAIN BLUE MONDAY". The song was included however on the cassette format in some countries, such as Australia and New Zealand, and on the original North American CD release of the album, alongside its B-side, "The Beach". "Blue Monday" was also included on the 2008 collector's edition of "Power, Corruption & Lies". The 1983 single "Confusion" firmly established the group as a dance music force, inspiring many musicians in subsequent years. In 1984 they followed the largely synthesised single "Thieves Like Us" with the heavy guitar-drum-bass rumble of "Murder", a not-too-distant cousin of "Ecstasy" from the "Power, Corruption & Lies" album. KROQ Los Angeles DJ Jed The Fish claims New Order had more to do with the emergence of house music than the Warehouse music of Chicago and “Frankie Knuckles and the whole so-called House music scene. Unless you were actually from regional Chicago, had you ever heard of House music until New Order? Be real, now.” 1985's "Low-Life" refined and sometimes mixed the two styles, brandishing "The Perfect Kiss"—the video for which was filmed by Jonathan Demme—and "Sub-culture". In February 1986, the soundtrack album to "Pretty in Pink" featuring "Shellshock" was released on A&M Records. An instrumental version of "Thieves Like Us" and the instrumental "Elegia" appeared in the film but were not on the soundtrack album. Later that summer, New Order headlined a line-up that included the Smiths, the Fall, and A Certain Ratio during the Festival of the Tenth Summer at Manchester's G-Mex. "Brotherhood" (1986) divided the two approaches onto separate album sides. The album notably featured "Bizarre Love Triangle" and "Angel Dust" (of which a remixed instrumental version is available on the UK "True Faith" CD video single, under the title "Evil Dust"), a track which marries a synth break beat with "Low-Life"-era guitar effects. While New Order toured North America with friends Echo & the Bunnymen, the summer of 1987 saw the release of the compilation "Substance", which featured the new single "True Faith". "Substance" was an important album in collecting the group's 12-inch singles onto CD for the first time and featured new versions of "Temptation" and "Confusion"—referred to as "Temptation '87" and "Confusion '87". A second disc featured several of the B-sides from the singles on the first disc, as well as additional A-sides "Procession" and "Murder". The single, "True Faith", with its surreal video, became a hit on MTV and the band's first American top 40 hit. The single's B-side, "1963"—originally planned on being the A-side until the group's label convinced them to release "True Faith" instead—would later be released as a single in its own right several years later, with two new versions. In December 1987, the band released a further single, "Touched by the Hand of God", with a Kathryn Bigelow-directed video parodying glam-metal. The song was one of four new tracks the band had recorded for the American comedy film "Salvatation!", and reached number 20 on the UK Singles Chart and number 1 in the UK Independent Singles chart. However, it would not appear on an album until the 1994 compilation "The Best of New Order". By this time, the group was heavily influenced by the Balearic sounds of Ibiza, which were making their way into the Haçienda. Partly recorded at Mediterranean Sound studios on Ibiza, "Technique" was released in February 1989. The album entered the charts at number one in the UK and contained a mix of acid house influence (as on opening track "Fine Time") and a more traditional rock sound (as on the single "Run 2"). The album is a blend of upbeat, accessible music coupled with blunt, poignant lyrics. During the summer of 1989, New Order supported "Technique" by touring with Public Image Ltd, Throwing Muses and The Sugarcubes across the United States and Canada in what the press dubbed the "Monsters of Alternative Rock" tour. Around this time, band members also began side projects including Electronic (Sumner with Johnny Marr) and Revenge (Hook with Davyth Hicks). Morris and Gilbert began to work together on outside TV theme production work. In 1991, the band were sued by the publishing company of American singer John Denver, who alleged that the guitar break in "Run 2" was similar to his song "Leaving on a Jet Plane". The case was settled out of court and the song has since been credited to both New Order and John Denver. In 1990, New Order recorded the official song of the England national football team's 1990 World Cup campaign, "World in Motion", under the ad-hoc band name EnglandNewOrder. The song, co-written with comedian Keith Allen, was the band's sole number one UK hit. The song was originally planned to be titled "E for England", however the Football Association vetoed the title upon realising that this was a reference to ecstasy; a drug heavily associated with the Haçienda. (Allen claimed that his original draft lyrics included "E is for England, England starts with E / We'll all be smiling when we're in Italy.") The song also featured chanting from members of the England team and Allen, and a guest rap from left winger John Barnes. It was again produced by Stephen Hague, who the band chose to produce their next album. The band's next album "Republic" was shadowed by the collapse of their longtime label Factory Records. The label had been ailing due to financial difficulties, and was forced to declare bankruptcy in 1992. New Order never had a formal contract with Factory. Although unusual for a major group, this was Factory's standard practice until the mid-1980s. Because of this, the band, rather than Factory Records, legally owned all of their recordings. This has been cited by Wilson himself as the main reason London Records' 1992 offer to buy the ailing label fell through. Following Factory's collapse, New Order signed with London, as did Morris and Gilbert separately for their side project The Other Two, whose debut album was originally intended for release on Factory. "Republic", released around the world in 1993, spawned the singles "Regret"—New Order's highest-charting single in the US—"Ruined in a Day", "World", and "Spooky". Following the release and promotion of "Republic", the band put New Order on hold while focusing on side projects; with The Other Two's debut album released in 1993. In 1994, a second singles collection was released, entitled "The Best of New Order". It featured all of the band's singles since "Substance" as well as a few extra tracks: "Vanishing Point" (from 1989's "Technique"), "The Perfect Kiss", "Thieves Like Us", "Shellshock", and new recordings of "True Faith", "Bizarre Love Triangle", "1963", and "Round & Round". The new versions of "True Faith" and "1963" (the latter as a more guitar-oriented version produced by Arthur Baker) were released as singles to promote the album. In the US, the track listing was altered to set it apart from "Substance" as well as the UK release of "The Best of New Order" which had been available months prior. This collection was followed by a remix album, "The Rest of New Order", featuring a selection of existing and newly commissioned mixes of classic New Order tracks. Some versions contained an extra disc or cassette composed entirely of remixes of "Blue Monday". "Blue Monday" was released as a single for a third time to promote the collection. The group reconvened in 1998 at the suggestion of Rob Gretton. Nearly five years had passed since they had last seen each other. Sumner said, "We decided before we agreed to doing any gig, to have a meeting, and if anyone had any grudges to bear, to iron them out." By the second meeting everyone agreed to continue playing, scheduling their reunion gig for the Phoenix Festival that same year. In addition to rarer songs, New Order also decided to begin playing Joy Division songs again. When the Phoenix Festival was cancelled due to low ticket sales, New Order instead played the last night of that year's Reading Festival. Their 2001 release "Get Ready" largely departed from their more electronic style and focused on more guitar oriented music. According to Sumner, ""Get Ready" was guitar-heavy simply because we felt that we'd left that instrument alone for a long time." Longtime fan Billy Corgan of The Smashing Pumpkins played guitar and sang back-up on the track "Turn My Way," and in 2001 toured with the band on dates in the UK, US, and Japan for a short period of time. Phil Cunningham (formerly of Marion) joined the band in a live capacity, deputising for Gilbert who declined to tour in favour of caring for her and Morris' children. Primal Scream's Bobby Gillespie provided vocals on the track "Rock the Shack". Singles from the album included "Crystal," "60 Miles an Hour" and Someone Like You." In 2002, "Q" featured New Order on their list of the "50 Bands to See Before You Die", although this was as part of a sub-list of "5 Bands That Could Go Either Way". Both New Order and Joy Division were portrayed in the Michael Winterbottom film "24 Hour Party People", depicting the rise and fall of Factory Records as seen through the eyes of label founder Tony Wilson. Cameos by Wilson himself, along with Mark E. Smith of The Fall and former members of Happy Mondays and Inspiral Carpets, lent a degree of legitimacy to the proceedings. The film touched on some of Factory's other artists, including Happy Mondays and The Durutti Column. The soundtrack featured the new track "Here to Stay," produced by the Chemical Brothers, which was released as a single. The single's music video highlighted scenes taken from the film. The band released a new album on 27 March 2005, titled "Waiting for the Sirens' Call", their first with new member Phil Cunningham. Cunningham replaced Gilbert (now married to Morris) so she could look after their children. Singles from this album were "Krafty", "Jetstream" (which features guest vocals by Ana Matronic from Scissor Sisters), and the title track. At the 2005 NME Awards, New Order and Joy Division received the award for "Godlike Geniuses" (for lifetime achievement). Previous winners include Ozzy Osbourne, The Clash, and Happy Mondays. In 2006 the album track "Guilt Is a Useless Emotion" was nominated for a Grammy Award in the category of Best Dance Recording. In the autumn of 2005, the group released another greatest hits compilation, in the form of "Singles". The two-disc release was an updated version of the "Substance" collection and contained every single released from their 1981 debut all the way through to "Waiting for the Sirens' Call". However, unlike "Substance", which focused almost exclusively on the 12" versions of the group's singles, "Singles" collected the 7" versions, many of which (like "Ceremony", "Temptation" and "Confusion") had never been released on CD. The album was accompanied by a two-disc DVD set, titled "Item", that collected the extended UK version of "NewOrderStory" with a DVD of all New Order music videos as well as two newly commissioned videos for "Temptation '87" and "Ceremony". The "New Order: Live in Glasgow" DVD was recorded at the Glasgow Academy in 2006 and features 18 tracks, including 4 Joy Division songs. Next to that, the release also contains a bonus disc of footage from the band's personal archive including 1980s footage from Glastonbury, Rome, Cork, Rotterdam and Toronto. In 2006, the band played several one-off live dates as well as short tours in the UK, Brazil and Argentina. After their Buenos Aires show in November 2006, Peter Hook suggested that the band should stop touring. In early May 2007, Hook was interviewed by British radio station XFM – originally to talk about his contribution to the debut album of Jane's Addiction singer Perry Farrell's new band Satellite Party – and stated that "Me and Bernard aren't working together." Further complicating the news, NewOrderOnline, a website with support from New Order management, reported that according to "a source close to the band", "The news about the split is false... New Order still exists despite what [Hook] said … Peter Hook can leave the band, but this doesn't mean the end of New Order." However, Sumner revealed in 2009 that he no longer wished to make music as New Order. In September 2011, the band announced that they would perform for the first time since 2006, at the Ancienne Belgique, Brussels on 17 October and at the Bataclan, Paris on 18 October. The band's line-up included keyboardist Gillian Gilbert, who returned to the band after a ten-year break, and Bad Lieutenant bassist Tom Chapman in place of Peter Hook. They played subsequent shows in London and South America in December. In December 2011, New Order released "Live at the London Troxy", a live album from their performance of 10 December 2011 at The Troxy in London. This release featured the new lineup and their first show in London in over five years. They continued to tour throughout 2012, including a short tour of New Zealand and Australia in February/March. They played at the 'T in the Park' festival in Scotland on 3 and 4 July 2012 and at the EXIT Festival in Novi Sad Serbia on 13 July 2012. New Order performed at Hyde Park with Blur and The Specials to celebrate the 2012 Summer Olympics closing ceremony. "Lost Sirens" was released in the United Kingdom on 14 January 2013. "Lost Sirens" is an eight-track album of tracks left out of "Waiting for the Sirens' Call". The album was discussed by Gillian Gilbert in a Brazilian interview to promote the band's appearance in São Paulo. She acknowledged issues with former member Peter Hook, and stated there was "a lot going on behind the scenes on the copyright" delaying the release. The band debuted their first newly-written song since the "Waiting for the Sirens' Call" sessions, titled "Singularity", during Lollapalooza Chile in March 2014. In July, the group toured North America, where they debuted the song "Plastic". On 2 September it was announced that the band decided to release their new album through Mute Records. The New Order catalogue remains with Warner Music. On 22 September 2015, the band released a new album, "Music Complete", their first without Peter Hook. The album was produced mostly by the band themselves, except "Singularity" and "Unlearn This Hatred", both produced by Tom Rowlands, while "Superheated" features additional production by Stuart Price. In November 2015, Peter Hook sued Bernard Sumner, Stephen Morris and Gillian Gilbert. In an objection, he claimed that they set up a new company behind his back and it has generated an income of £7.8 million in four years while he received only a fraction of that. The three members insisted they had treated Hook fairly and that his stake in the band's royalties was reasonable. The judge ruled that there was "at least a reasonable prospect" of Hook proving that he was not getting a fair share of royalties and other income. He was willing to hear the case but urged the parties to come to an agreement rather than suffer legal costs of around £900,000. On 13 July 2017, New Order played a concert at Manchester International Festival with Liam Gillick. The performance was then released as a live album titled "∑(No,12k,Lg,17Mif) New Order + Liam Gillick: So it goes.. (Live at MIF)" in July 12, 2019. On 20 September 2017, a posting on New Order's official website announced that a full and final settlement had been reached in the long running disputes with their former bassist Peter Hook. On 23 August 2018, the band played the first date of a North American tour at the Palace Theatre in St. Paul Minnesota, US. According to the band's official website, other stops on the tour included Cleveland, Ohio; Washington, DC; Toronto, Ontario; Long Beach, California; Kahului, Hawaii; and Honolulu, Hawaii. As of 24 August 2018, a single show in Santiago, Chile had been announced for 21 November 2018. In January 2020, the band played a four night residency in Florida, and in February 2020, they announced a co-headlining tour in North America with the Pet Shop Boys, and that the only concert in the UK in 2020 would be at The O2 on 10 October. In 1988, Bernard Sumner teamed up with former Smiths guitarist Johnny Marr to form the group Electronic, also enlisting the help of Neil Tennant and Chris Lowe of the Pet Shop Boys. Electronic regrouped in 1996 for "Raise the Pressure", which also featured Karl Bartos (formerly of Kraftwerk). The project's third album "Twisted Tenderness" was released in 1999 after which the band dissolved. In June 2009, Bernard Sumner formed a new band called Bad Lieutenant with Phil Cunningham (guitar) and Jake Evans (guitar and vocals). Their album "Never Cry Another Tear" was released on 5 October 2009. In addition to Cunningham and Evans the album also features appearances by Stephen Morris (drums), Jack Mitchell (drums), Tom Chapman (bass) and Alex James (bass). The live band included Morris on drums and Tom Chapman on bass. Hook has been involved with several other projects. In the 1990s, Hook recorded with Killing Joke with a view to joining the band. However, original bassist Martin 'Youth' Glover instead returned to the band. In 1995 he toured with The Durutti Column. He has recorded one album with the band Revenge with Davyth Hicks and Chris Jones and two with Monaco (both as bassist, keyboardist and lead vocalist) with David Potts. Monaco scored a club and alternative radio hit with "What Do You Want From Me?" in 1997. Hook also formed a band called Freebass with fellow bass players Mani (The Stone Roses) and Andy Rourke (The Smiths) and vocalist Gary Briggs, which was active from 2007 to 2010. He also contributed to Perry Farrell's Satellite Party. Hook's current band Peter Hook and The Light is touring Joy Division and New Order albums in their entirety. In 1990 Gilbert and Morris formed their own band, The Other Two. The Other Two released its first single "Tasty Fish" in 1991 and released two albums, "The Other Two & You" in 1993 and "Super Highways" in 1999. They have also been involved in scoring television soundtracks. In 2007, Gilbert and Morris remixed two tracks for the Nine Inch Nails remixes album "Year Zero Remixed". "BeMusic" was a name the band used for their publishing company (the LP label for "Movement" says "B Music" in large letters, though using an italic ß for the letter B). All four members of the band used the name for production work for other artists' recordings between 1982 and 1985. The first BeMusic credit was for Peter Hook producing Stockholm Monsters in 1982. Other artists with producer or musician credit for "BeMusic" were 52nd Street, Section 25, Marcel King, Quando Quango, Paul Haig, Thick Pigeon, Nyam Nyam and Life. Their production work as BeMusic was collected on two LTM Recordings compilation CDs, "Cool As Ice: The BeMusic Productions" and "Twice As Nice" (which also included production work by Donald Johnson, of A Certain Ratio, and Arthur Baker). New Order's music mixes rock with dance music, as can be seen on signature tracks such as 1982's "Temptation", 1983's "Blue Monday" and 1987's "True Faith". Founding member Hook stated that the band's shift from playing cold dark tracks from 1981 to producing electro/rock tracks from 1982 was inspired by the music of German electronic group Kraftwerk, US rock band Sparks who had produced disco/electro-rock music with producer Giorgio Moroder on their "No. 1 in Heaven" album, and also the Moroder/Donna Summer collaboration on "I Feel Love". New Order's collaboration with New York DJ Arthur Baker was inspired by the records' sounds of Grandmaster Flash and the Furious Five and Afrika Bambaataa & the Soulsonic Force. The group's album art earned them the status of icons in the alternative community, and has shown considerable longevity. According to an unsigned "Allmusic" text, the band are also regarded as "the first alternative dance" music group with their fusion of "used icy, gloomy post-punk with Kraftwerk-style synth-pop" and were also labeled as synth-pop, post-punk, new wave, dance-rock and electronica. They have heavily influenced techno, rock, and pop musicians including the Pet Shop Boys, 808 State, Arcade Fire, and Moby, and were themselves influenced by the likes of David Bowie, Neu!, Kraftwerk, Cabaret Voltaire and Giorgio Moroder. They have also significantly influenced electro, freestyle and house. New Order's Kraftwerk influence was acknowledged by their single "Krafty", which had cover art referencing "Autobahn". Bassist Peter Hook contributed to New Order's sound by developing an idiosyncratic bass guitar technique. He often used the bass as a lead instrument, playing melodies on the high strings with a signature heavy chorus effect, leaving the actual basslines to keyboards or sequencers. This has often been cited as a defining characteristic of the New Order sound. Drummer Stephen Morris plays a mixture of acoustic and electronic drums, and in many cases plays along seamlessly with sequenced parts. All the band members could and did switch instruments throughout gigs, as evidenced on Jonathan Demme's video for "The Perfect Kiss" and the concert videos "Taras Shevchenko" (recorded in New York, November 1981) and "Pumped Full of Drugs" (Tokyo, May 1985). During such live gigs, Sumner alternated between guitar, keyboards, melodica and (on the track "Confusion") bass; Gilbert switched between keyboards and guitar, Morris between drums and keyboards, and Hook played both bass and electronic drums. "Taras Shevchenko" is also notable for the fact all four members of the group have left the stage before the final song, "Temptation", comes to a complete end. Both New Order and Joy Division were among the most successful artists on the Factory Records label, run by Granada television personality Tony Wilson, and partnered with Factory in the financing of the Manchester club The Haçienda. The band rarely gave interviews in the 1980s, later ascribing this to not wanting to discuss Ian Curtis. The band became more open in the 1990s; for example, the aforementioned "NewOrderStory" (and in particular the longer UK version) featured extensive personal interviews. Speaking in 2009, fellow synthpop musician Phil Oakey described New Order's slow-burn career as cult musicians as being unusually prolonged and effective: "If you want to make a lot of money out of pop, be number 3 a lot. Like New Order did." Despite this, the band have commented on the unplanned nature of their career and the considerable expense lost in supporting the Haçienda. Almost all New Order recordings bear minimalist packaging, art directed by Peter Saville. The group's record sleeves bucked the 1980s trend by rarely showing the band members (with the exception of the "Low-Life" album) or even providing basic information such as the band name or title of the release. Song names were often hidden within the shrink wrapped package, either on the disc itself (such as the "Blue Monday" single), on an inconspicuous part of an inner sleeve ("The Perfect Kiss" single), or written in a cryptic colour code invented by Saville ("Power, Corruption & Lies"). Saville said his intention was to sell the band as a "mass-produced secret" of sorts, and that the minimalist style was enough to allow fans to identify the band's products without explicit labelling. Saville frequently sent the artwork straight to the printer, unreviewed by either the band or the label. ! Year !! Awards !! Work !! Category !! Result Former members
https://en.wikipedia.org/wiki?curid=22146
Niccolò Fontana Tartaglia Niccolò Fontana Tartaglia (; 1499/1500, Brescia – 13 December 1557, Venice) was an Italian mathematician, engineer (designing fortifications), a surveyor (of topography, seeking the best means of defense or offense) and a bookkeeper from the then-Republic of Venice (now part of Italy). He published many books, including the first Italian translations of Archimedes and Euclid, and an acclaimed compilation of mathematics. Tartaglia was the first to apply mathematics to the investigation of the paths of cannonballs, known as ballistics, in his "Nova Scientia" ("A New Science"); his work was later partially validated and partially superseded by Galileo's studies on falling bodies. He also published a treatise on retrieving sunken ships. Niccolò Fontana was the son of Michele Fontana, a dispatch rider who travelled to neighboring towns to deliver mail. But in 1506, Michele was murdered by robbers, and Niccolò, his two siblings, and his mother were left impoverished. Niccolò experienced further tragedy in 1512 when the King Louis XII's troops invaded Brescia during the War of the League of Cambrai against Venice. The militia of Brescia defended their city for seven days. When the French finally broke through, they took their revenge by massacring the inhabitants of Brescia. By the end of battle, over 45,000 residents were killed. During the massacre, Niccolò and his family sought sanctuary in the local cathedral. But the French entered and a soldier sliced Niccolò's jaw and palate with a saber and left him for dead. His mother nursed him back to health but the young boy was left with a speech impediment, prompting the nickname "Tartaglia" ("stammerer"). After this he would never shave, and grew a beard to camouflage his scars. Tartaglia's biographer Arnoldo Masotti writes that: Tartaglia moved to Verona around 1517, then to Venice in 1534, a major European commercial hub and one of the great centers of the Italian renaissance at this time. Also relevant is Venice's place at the forefront of European printing culture in the sixteenth century, making early printed texts available even to poor scholars if sufficiently motivated or well-connected — Tartaglia knew of Archimedes' work on the quadrature of the parabola, for example, from Guarico's Latin edition of 1503, which he had found "in the hands of a sausage-seller in Verona in 1531" ("in mano di un salzizaro in Verona, l'anno 1531" in his words). Tartaglia eked out a living teaching practical mathematics in abacus schools and earned a penny where he could: "Nova scientia" (1537) was Tartaglia's first published work, described by Matteo Valleriani as: Then dominant Aristotelian physics preferred categories like "heavy" and "natural" and "violent" to describe motion, generally eschewing mathematical explanations. Tartaglia brought mathematical models to the fore, "eviscerat[ing] Aristotelian terms of projectile movement" in the words Mary J. Henninger-Voss. One of his findings was that the maximum range of a projectile was achieved by directing the cannon at a 45° angle to the horizon. Tartaglia's model for a cannonball's flight was that it proceeded from the cannon in a straight line, then after a while started to arc towards the earth along a circular path, then finally dropped in another straight line directly towards the earth. At the end of Book 2 of "Nova scientia", Tartaglia proposes to find the length of that initial rectilinear path for a projectile fired at an elevation of 45°, engaging in a Euclidean-style argument, but one with numbers attached to line segments and areas, and eventually proceeds algebraically to find the desired quantity ("procederemo per algebra" in his words). Mary J. Henninger-Voss notes that "Tartaglia's work on military science had an enormous circulation throughout Europe", being a reference for common gunners into the eighteenth century, sometimes through unattributed translations. He influenced Galileo as well, who owned "richly annotated" copies of his works on ballistics as he set about solving the projectile problem once and for all. Archimedes' works began to be studied outside the universities in Tartaglia's day as exemplary of the notion that mathematics is the key to understanding physics, Federigo Commandino reflecting this notion when saying in 1558 that "with respect to geometry no one of sound mind could deny that Archimedes was some god". Tartaglia published a 71-page Latin edition of Archimedes in 1543, "Opera Archimedis Syracusani philosophi et mathematici ingeniosissimi", containing Archimedes' works on the parabola, the circle, centers of gravity, and floating bodies. Guarico had published Latin editions of the first two in 1503, but the works on centers of gravity and floating bodies had not been published before. Tartaglia published Italian versions of some Archimedean texts later in life, his executor continuing to publish his translations after his death. Galileo probably learned of Archimedes' work through these widely disseminated editions. Tartaglia's Italian edition of Euclid in 1543, "Euclide Megarense philosopho", was especially significant as the first translation of the "Elements" into any modern European language. For two centuries Euclid had been taught from two Latin translations taken from an Arabic source; these contained errors in Book V, the Eudoxian theory of proportion, which rendered it unusable. Tartaglia's edition was based on Zamberti's Latin translation of an uncorrupted Greek text, and rendered Book V correctly. He also wrote the first modern and useful commentary on the theory. This work went through many editions in the sixteenth century and helped diffuse knowledge of mathematics to a non-academic but increasingly well-informed literate and numerate public in Italy. The theory became an essential tool for Galileo, as it had been for Archimedes. Tartaglia exemplified and eventually transcended the abacco tradition that had flourished in Italy since the twelfth century, a tradition of concrete commercial mathematics taught at abacus schools maintained by communities of merchants. "Maestros d'abaco" like Tartaglia taught not with the abacus but with paper-and-pen, inculcating algorithms of the type found in grade schools today. Tartaglia's masterpiece was the "General Trattato di Numeri et Misure" ("General Treatise on Number and Measure"), a 1500-page encyclopedia in six parts written in the Venetian dialect, the first three coming out in 1556 about the time of Tartaglia's death and the last three published posthumously by his literary executor and publisher Curtio Troiano in 1560. David Eugene Smith wrote of the "General Trattato" that it was: Part I is 554 pages long and constitutes essentially a commercial arithmetic, taking up such topics as basic operations with the complex currencies of the day (ducats, soldi, pizolli, and so on), exchanging currencies, calculating interest, and dividing profits in joint companies. The book is replete with worked examples with much emphasis on methods and rules (that is, algorithms), all ready to use virtually as is. Part II takes up more general arithmetic problems, including progressions, powers, binomial expansions, Tartaglia's Triangle (as "Pascal's Triangle" is still called. in Italy), calculations with roots, and proportions / fractions. Part IV concerns triangles, regular polygons, the Platonic solids, and Archimedean topics like the quadrature of the circle and circumscribing a cylinder around a sphere. Tartaglia was proficient with binomial expansions and included many worked examples in Part II of the "General Trattato", one a detailed explanation of how to calculate the summands of formula_1, including the appropriate binomial coefficients. Tartaglia knew of Pascal's Triangle one hundred years before Pascal, as shown in this image from the "General Trattato". His examples are numeric, but he thinks about it geometrically, the horizontal line formula_3 at the top of the triangle being broken into two segments formula_4 and formula_5, where point formula_6 is the apex of the triangle. Binomial expansions amount to taking formula_7 for exponents formula_8 as you go down the triangle. The symbols along the outside represent powers at this early stage of algebraic notation: formula_9, and so on. He writes explicitly about the additive formation rule, that (for example) the adjacent 15 and 20 in the fifth row add up to 35, which appears beneath them in the sixth row. Tartaglia is perhaps best known today for his conflicts with Gerolamo Cardano. Cardano cajoled Tartaglia into revealing his solution to the cubic equations by promising not to publish them. Tartaglia divulged the secrets of the solutions of three different forms of the cubic equation in verse. Several years later, Cardano happened to see unpublished work by Scipione del Ferro who independently came up with the same solution as Tartaglia. As the unpublished work was dated before Tartaglia's, Cardano decided his promise could be broken and included Tartaglia's solution in his next publication. Even though Cardano credited his discovery, Tartaglia was extremely upset and a famous public challenge match resulted between himself and Cardano's student, Ludovico Ferrari. Widespread stories that Tartaglia devoted the rest of his life to ruining Cardano, however, appear to be completely fabricated. Mathematical historians now credit both Cardano and Tartaglia with the formula to solve cubic equations, referring to it as the "Cardano–Tartaglia formula". Tartaglia was a prodigious calculator and master of solid geometry. In Part IV of the "General Trattato" he shows by example how to calculate the height of a pyramid on a triangular base, that is, an irregular tetrahedron. The base of the pyramid is a formula_10 triangle formula_11, with edges of length formula_12, and formula_13 rising up to the apex formula_14 from points formula_15, formula_6, and formula_17 respectively. Base triangle formula_11 partitions into formula_19 and formula_20 triangles by dropping the perpendicular from point formula_17 to side formula_22. He proceeds to erect a triangle in the plane perpendicular to line formula_22 through the pyramid's apex, point formula_14, calculating all three sides of this triangle and noting that its height is the height of the pyramid. At the last step, he applies what amounts to this formula for the height formula_25 of a triangle in terms of its sides formula_26 (the height from side formula_27 to its opposite vertex): a formula deriving from the Law of Cosines (not that he cites any justification in this section of the "General Trattato"). Tartaglia drops a digit early in the calculation, taking formula_29 as formula_30, but his method is sound. The final (correct) answer is: The volume of the pyramid is easily gotten after that (not that Tartaglia gives it): Simon Stevin invented decimal fractions later in the sixteenth century, so the last figure would have been foreign to Tartaglia, who always used fractions. All the same, his approach is in some ways a modern one, suggesting by example an algorithm for calculating the height of most or all irregular tetrahedra, but (as usual for him) he gives no explicit formula.
https://en.wikipedia.org/wiki?curid=22148
Nagarjuna Nāgārjuna (c. 150 – c. 250 CE), (Tibetan: mGon-po kLu-grub) is widely considered one of the most important Buddhist philosophers. Along with his disciple Āryadeva, he is considered the founder of the Madhyamaka school of Mahāyāna Buddhism. Nāgārjuna is also credited with developing the philosophy of the Prajñāpāramitā sūtras and, by some sources, with having revealed these scriptures to the world after recovering them from the nāgas. He is traditionally thought to have written many treatises on rasayana, as well as serving a term as the head of Nālandā. Very little is reliably known of the life of Nāgārjuna, since the surviving accounts were written in Chinese and Tibetan centuries after his death. According to some accounts, Nāgārjuna was originally from South India. Some scholars believe that Nāgārjuna was an advisor to a king of the Satavahana dynasty. Archaeological evidence at Amarāvatī indicates that if this is true, the king may have been Yajña Śrī Śātakarṇi, who ruled between 167 and 196 CE. On the basis of this association, Nāgārjuna is conventionally placed at around 150–250 CE. According to a 4th/5th-century biography translated by Kumārajīva, Nāgārjuna was born into a Brahmin family in Vidarbha (a region of Maharashtra) and later became a Buddhist. Some sources claim that in his later years, Nāgārjuna lived on the mountain of Śrīparvata near the city that would later be called Nāgārjunakoṇḍa ("Hill of Nāgārjuna"). The ruins of Nāgārjunakoṇḍa are located in Guntur district, Andhra Pradesh. The Caitika and Bahuśrutīya nikāyas are known to have had monasteries in Nāgārjunakoṇḍa. The archaeological finds at Nagarjunakonda have not resulted in any evidence that the site was associated with Nagarjuna. The name "Nagarjunakonda" dates from the medieval period, and the 3rd-4th century inscriptions found at the site make it clear that it was known as "Vijayapuri" in the ancient period. There exist a number of influential texts attributed to Nāgārjuna though, as there are many pseudepigrapha attributed to him, lively controversy exists over which are his authentic works. The "Mūlamadhyamakakārikā" is Nāgārjuna's best-known work. It is "not only a grand commentary on the Buddha's discourse to Kaccayana, the only discourse cited by name, but also a detailed and careful analysis of most of the important discourses included in the Nikayas and the agamas, especially those of the "Atthakavagga" of the "Sutta-nipata". In the "Mūlamadhyamakakārikā", "[A]ll experienced phenomena are empty ("sunya"). This did not mean that they are not experienced and, therefore, non-existent; only that they are devoid of a permanent and eternal substance ("svabhava") because, like a dream, they are mere projections of human consciousness. Since these imaginary fictions are experienced, they are not mere names ("prajnapti")." According to David Seyfort Ruegg, the "Madhyamakasastrastuti" attributed to Candrakirti (c. 600 – c. 650) refers to eight texts by Nagarjuna:the "(Madhyamaka)karikas", the "Yuktisastika", the "Sunyatasaptati", the "Vigrahavyavartani", the "Vidala" (i.e. "Vaidalyasutra/Vaidalyaprakarana"), the "Ratnavali", the "Sutrasamuccaya", and "Samstutis ("Hymns")". This list covers not only much less than the grand total of works ascribed to Nagarjuna in the Chinese and Tibetan collections, but it does not even include all such works that Candrakirti has himself cited in his writings.According to one view, that of Christian Lindtner, the works definitely written by Nāgārjuna are: The Tibetan historian Buston considers the first six to be the main treatises of Nāgārjuna (this is called the "yukti corpus", "rigs chogs"), while according to Tāranātha only the first five are the works of Nāgārjuna. TRV Murti considers Ratnaavali, Pratitya Samutpaada Hridaya and Sutra Samuccaya to be works of Nāgārjuna as the first two are quoted profusely by Chandrakirti and the third by Shantideva. In addition to works mentioned above, several others are attributed to Nāgārjuna. There is an ongoing, lively controversy over which of those works are authentic. Contemporary research suggest that some these works belong to a significantly later period, either to late 8th or early 9th century CE, and hence can not be authentic works of Nāgārjuna. Several works considered important in esoteric Buddhism are attributed to Nāgārjuna and his disciples by traditional historians like Tāranātha from 17th century Tibet. These historians try to account for chronological difficulties with various theories. For example, apropagation of later writings via mystical revelation. For a useful summary of this tradition, see Wedemeyer 2007. According to Ruegg, "three collections of stanzas on the virtues of intelligence and moral conduct ascribed to Nagarjuna are extant in Tibetan translation": "Prajñasatakaprakarana", "Nitisastra-Jantuposanabindu" and "Niti-sastra-Prajñadanda." Other works are extant only in Chinese, one of these is the "Shih-erh-men-lun" or 'Twelve-topic treatise' (*"Dvadasanikaya" or *"Dvadasamukha-sastra"); one of the three basic treatises of the Sanlun school (East Asian Madhyamaka). Lindtner considers that the "Mahāprajñāpāramitāupadeśa" ("Ta-chih-tu-lun", Taisho 1509, "Commentary on the great prajñaparamita") which has been influential in Chinese Buddhism, is not a genuine work of Nāgārjuna. This work is also only attested in a Chinese translation by Kumārajīva and is unknown in the Tibetan and Indian traditions. There is much discussion as to whether this is a work of Nāgārjuna, or someone else. Étienne Lamotte, who translated one third of the work into French, felt that it was the work of a North Indian bhikṣu of the Sarvāstivāda school who later became a convert to the Mahayana. The Chinese scholar-monk Yin Shun felt that it was the work of a South Indian and that Nāgārjuna was quite possibly the author. These two views are not necessarily in opposition and a South Indian Nāgārjuna could well have studied the northern Sarvāstivāda. Neither of the two felt that it was composed by Kumārajīva, which others have suggested. Other attributed works include: From studying his writings, it is clear that Nāgārjuna was conversant with many of the Śrāvaka philosophies and with the Mahāyāna tradition. However, determining Nāgārjuna's affiliation with a specific nikāya is difficult, considering much of this material has been lost. If the most commonly accepted attribution of texts (that of Christian Lindtner) holds, then he was clearly a Māhayānist, but his philosophy holds assiduously to the Śrāvaka "Tripiṭaka", and while he does make explicit references to Mahāyāna texts, he is always careful to stay within the parameters set out by the Śrāvaka canon. Nāgārjuna may have arrived at his positions from a desire to achieve a consistent exegesis of the Buddha's doctrine as recorded in the āgamas. In the eyes of Nāgārjuna, the Buddha was not merely a forerunner, but the very founder of the Madhyamaka system. David Kalupahana sees Nāgārjuna as a successor to Moggaliputta-Tissa in being a champion of the middle-way and a reviver of the original philosophical ideals of the Buddha. Nāgārjuna assumes a knowledge of the definitions of the sixteen categories as given in the Nyaya Sutras, the chief text of the Hindu Nyaya school, and wrote a treatise on the pramanas where he reduced the syllogism of five members into one of three. In the Vigrahavyavartani Karika, Nāgārjuna criticises the Nyaya theory of pramanas (means of knowledge) Nāgārjuna was fully acquainted with the classical Hindu philosophies of Samkhya and even the Vaiseshika. Because of the high degree of similarity between Nāgārjuna's philosophy and Pyrrhonism, particularly the surviving works of Sextus Empiricus, Thomas McEvilley suspects that Nāgārjuna was influenced by Greek Pyrrhonists texts imported into India. But according to others, Pyrrho of Elis (c. 360-c. 270 BCE), usually credited with founding this school of sceptical philosophy, was himself influenced by Indian philosophy, when he travelled to India with Alexander the Great's army and studied with the gymnosophists. According to Christopher I. Beckwith, Pyrrho's teachings are based on Buddhism, because "adiaphora", "astathmēta" and "anepikrita" in the "Aristocles Passage" resemble the Buddhist three marks of existence. According to him, the key innovative tenets of Pyrrho's skepticism were only found in Indian philosophy at the time and not in Greece. Nāgārjuna's major thematic focus is the concept of śūnyatā (translated into English as "emptiness") which brings together other key Buddhist doctrines, particularly anātman "not-self" and pratītyasamutpāda "dependent origination", to refute the metaphysics of some of his contemporaries. For Nāgārjuna, as for the Buddha in the early texts, it is not merely sentient beings that are "selfless" or non-substantial; all phenomena (dhammas) are without any svabhāva, literally "own-being", "self-nature", or "inherent existence" and thus without any underlying essence. They are "empty" of being independently existent; thus the heterodox theories of svabhāva circulating at the time were refuted on the basis of the doctrines of early Buddhism. This is so because all things arise always dependently: not by their own power, but by depending on conditions leading to their coming into existence, as opposed to being. Nāgārjuna means by real any entity which has a nature of its own (svabhāva), which is not produced by causes (akrtaka), which is not dependent on anything else (paratra nirapeksha). Chapter 24 verse 14 of the Mūlamadhyamakakārikā provides one of Nāgārjuna's most famous quotations on emptiness and co-arising: As part of his analysis of the emptiness of phenomena in the Mūlamadhyamakakārikā, Nāgārjuna critiques svabhāva in several different concepts. He discusses the problems of positing any sort of inherent essence to causation, movement, change and personal identity. Nāgārjuna makes use of the Indian logical tool of the tetralemma to attack any essentialist conceptions. Nāgārjuna's logical analysis is based on four basic propositions: To say that all things are 'empty' is to deny any kind of ontological foundation; therefore Nāgārjuna's view is often seen as a kind of ontological anti-foundationalism or a metaphysical anti-realism. Understanding the nature of the emptiness of phenomena is simply a means to an end, which is nirvana. Thus Nāgārjuna's philosophical project is ultimately a soteriological one meant to correct our everyday cognitive processes which mistakenly posits svabhāva on the flow of experience. Some scholars such as Fyodor Shcherbatskoy and T.R.V. Murti held that Nāgārjuna was the inventor of the Shunyata doctrine; however, more recent work by scholars such as Choong Mun-keat, Yin Shun and Dhammajothi Thero has argued that Nāgārjuna was not an innovator by putting forth this theory, but that, in the words of Shi Huifeng, "the connection between emptiness and dependent origination is not an innovation or creation of Nāgārjuna". Nāgārjuna was also instrumental in the development of the two truths doctrine, which claims that there are two levels of truth in Buddhist teaching, the ultimate truth ("paramārtha satya") and the conventional or superficial truth ("saṃvṛtisatya"). The ultimate truth to Nāgārjuna is the truth that everything is empty of essence, this includes emptiness itself ('the emptiness of emptiness'). While some (Murti, 1955) have interpreted this by positing Nāgārjuna as a neo-Kantian and thus making ultimate truth a metaphysical noumenon or an "ineffable ultimate that transcends the capacities of discursive reason", others such as Mark Siderits and Jay L. Garfield have argued that Nāgārjuna's view is that "the ultimate truth is that there is no ultimate truth" (Siderits) and that Nāgārjuna is a "semantic anti-dualist" who posits that there are only conventional truths. Hence according to Garfield: Suppose that we take a conventional entity, such as a table. We analyze it to demonstrate its emptiness, finding that there is no table apart from its parts […]. So we conclude that it is empty. But now let us analyze that emptiness […]. What do we find? Nothing at all but the table’s lack of inherent existence. […]. To see the table as empty […] is to see the table as conventional, as dependent. In articulating this notion in the "Mūlamadhyamakakārikā", Nāgārjuna drew on an early source in the "Kaccānagotta Sutta", which distinguishes definitive meaning ("nītārtha") from interpretable meaning ("neyārtha"): The version linked to is the one found in the nikayas, and is slightly different from the one found in the "Samyuktagama". Both contain the concept of teaching via the middle between the extremes of existence and non-existence. Nagarjuna does not make reference to "everything" when he quotes the agamic text in his "Mūlamadhyamakakārikā". Jay L. Garfield describes that Nāgārjuna approached causality from the four noble truths and dependent origination. Nāgārjuna distinguished two dependent origination views in a causal process, that which causes effects and that which causes conditions. This is predicated in the two truth doctrine, as conventional truth and ultimate truth held together, in which both are empty in existence. The distinction between effects and conditions is controversial. In Nāgārjuna's approach, cause means an event or state that has power to bring an effect. Conditions, refer to proliferating causes that bring a further event, state or process; without a metaphysical commitment to an occult connection between explaining and explanans. He argues nonexistent causes and various existing conditions. The argument draws from unreal causal power. Things conventional exist and are ultimately nonexistent to rest in the middle way in both causal existence and nonexistence as casual emptiness within the Mūlamadhyamakakārikā doctrine. Although seeming strange to Westerners, this is seen as an attack on a reified view of causality. Nāgārjuna also taught the idea of relativity; in the Ratnāvalī, he gives the example that shortness exists only in relation to the idea of length. The determination of a thing or object is only possible in relation to other things or objects, especially by way of contrast. He held that the relationship between the ideas of "short" and "long" is not due to intrinsic nature (svabhāva). This idea is also found in the Pali Nikāyas and Chinese Āgamas, in which the idea of relativity is expressed similarly: "That which is the element of light ... is seen to exist on account of [in relation to] darkness; that which is the element of good is seen to exist on account of bad; that which is the element of space is seen to exist on account of form." Nāgārjuna is often depicted in composite form comprising human and nāga characteristics. Often the nāga-aspect forms a canopy crowning and shielding his human head. The notion of the nāga is found throughout Indian religious culture, and typically signifies an intelligent serpent or dragon, who is responsible for the rains, lakes and other bodies of water. In Buddhism, it is a synonym for a realised arhat, or wise person in general.
https://en.wikipedia.org/wiki?curid=22149
Nuclear reactor A nuclear reactor, formerly known as an atomic pile, is a device used to initiate and control a self-sustained nuclear chain reaction. Nuclear reactors are used at nuclear power plants for electricity generation and in nuclear marine propulsion. Heat from nuclear fission is passed to a working fluid (water or gas), which in turn runs through steam turbines. These either drive a ship's propellers or turn electrical generators' shafts. Nuclear generated steam in principle can be used for industrial process heat or for district heating. Some reactors are used to produce isotopes for medical and industrial use, or for production of weapons-grade plutonium. As of early 2019, the IAEA reports there are 454 nuclear power reactors and 226 nuclear research reactors in operation around the world. Just as conventional thermal power stations generate electricity by harnessing the thermal energy released from burning fossil fuels, nuclear reactors convert the energy released by controlled nuclear fission into thermal energy for further conversion to mechanical or electrical forms. When a large fissile atomic nucleus such as uranium-235 or plutonium-239 absorbs a neutron, it may undergo nuclear fission. The heavy nucleus splits into two or more lighter nuclei, (the fission products), releasing kinetic energy, gamma radiation, and free neutrons. A portion of these neutrons may be absorbed by other fissile atoms and trigger further fission events, which release more neutrons, and so on. This is known as a nuclear chain reaction. To control such a nuclear chain reaction, Control rods containing neutron poisons and neutron moderators can change the portion of neutrons that will go on to cause more fission. Nuclear reactors generally have automatic and manual systems to shut the fission reaction down if monitoring detects unsafe conditions. The reactor core generates heat in a number of ways: A kilogram of uranium-235 (U-235) converted via nuclear processes releases approximately three million times more energy than a kilogram of coal burned conventionally (7.2 × 1013 joules per kilogram of uranium-235 versus 2.4 × 107 joules per kilogram of coal). A nuclear reactor coolant — usually water but sometimes a gas or a liquid metal (like liquid sodium or lead) or molten salt — is circulated past the reactor core to absorb the heat that it generates. The heat is carried away from the reactor and is then used to generate steam. Most reactor systems employ a cooling system that is physically separated from the water that will be boiled to produce pressurized steam for the turbines, like the pressurized water reactor. However, in some reactors the water for the steam turbines is boiled directly by the reactor core; for example the boiling water reactor. The rate of fission reactions within a reactor core can be adjusted by controlling the quantity of neutrons that are able to induce further fission events. Nuclear reactors typically employ several methods of neutron control to adjust the reactor's power output. Some of these methods arise naturally from the physics of radioactive decay and are simply accounted for during the reactor's operation, while others are mechanisms engineered into the reactor design for a distinct purpose. The fastest method for adjusting levels of fission-inducing neutrons in a reactor is via movement of the control rods. Control rods are made of neutron poisons and therefore absorb neutrons. When a control rod is inserted deeper into the reactor, it absorbs more neutrons than the material it displaces—often the moderator. This action results in fewer neutrons available to cause fission and reduces the reactor's power output. Conversely, extracting the control rod will result in an increase in the rate of fission events and an increase in power. The physics of radioactive decay also affects neutron populations in a reactor. One such process is delayed neutron emission by a number of neutron-rich fission isotopes. These delayed neutrons account for about 0.65% of the total neutrons produced in fission, with the remainder (termed "prompt neutrons") released immediately upon fission. The fission products which produce delayed neutrons have half-lives for their decay by neutron emission that range from milliseconds to as long as several minutes, and so considerable time is required to determine exactly when a reactor reaches the critical point. Keeping the reactor in the zone of chain reactivity where delayed neutrons are "necessary" to achieve a critical mass state allows mechanical devices or human operators to control a chain reaction in "real time"; otherwise the time between achievement of criticality and nuclear meltdown as a result of an exponential power surge from the normal nuclear chain reaction, would be too short to allow for intervention. This last stage, where delayed neutrons are no longer required to maintain criticality, is known as the prompt critical point. There is a scale for describing criticality in numerical form, in which bare criticality is known as "zero dollars" and the prompt critical point is "one dollar", and other points in the process interpolated in cents. In some reactors, the coolant also acts as a neutron moderator. A moderator increases the power of the reactor by causing the fast neutrons that are released from fission to lose energy and become thermal neutrons. Thermal neutrons are more likely than fast neutrons to cause fission. If the coolant is a moderator, then temperature changes can affect the density of the coolant/moderator and therefore change power output. A higher temperature coolant would be less dense, and therefore a less effective moderator. In other reactors the coolant acts as a poison by absorbing neutrons in the same way that the control rods do. In these reactors power output can be increased by heating the coolant, which makes it a less dense poison. Nuclear reactors generally have automatic and manual systems to scram the reactor in an emergency shut down. These systems insert large amounts of poison (often boron in the form of boric acid) into the reactor to shut the fission reaction down if unsafe conditions are detected or anticipated. Most types of reactors are sensitive to a process variously known as xenon poisoning, or the iodine pit. The common fission product Xenon-135 produced in the fission process acts as a neutron poison that absorbs neutrons and therefore tends to shut the reactor down. Xenon-135 accumulation can be controlled by keeping power levels high enough to destroy it by neutron absorption as fast as it is produced. Fission also produces iodine-135, which in turn decays (with a half-life of 6.57 hours) to new xenon-135. When the reactor is shut down, iodine-135 continues to decay to xenon-135, making restarting the reactor more difficult for a day or two, as the xenon-135 decays into cesium-135, which is not nearly as poisonous as xenon-135, with a half-life of 9.2 hours. This temporary state is the "iodine pit." If the reactor has sufficient extra reactivity capacity, it can be restarted. As the extra xenon-135 is transmuted to xenon-136, which is much less a neutron poison, within a few hours the reactor experiences a "xenon burnoff (power) transient". Control rods must be further inserted to replace the neutron absorption of the lost xenon-135. Failure to properly follow such a procedure was a key step in the Chernobyl disaster. Reactors used in nuclear marine propulsion (especially nuclear submarines) often cannot be run at continuous power around the clock in the same way that land-based power reactors are normally run, and in addition often need to have a very long core life without refueling. For this reason many designs use highly enriched uranium but incorporate burnable neutron poison in the fuel rods. This allows the reactor to be constructed with an excess of fissionable material, which is nevertheless made relatively safe early in the reactor's fuel burn cycle by the presence of the neutron-absorbing material which is later replaced by normally produced long-lived neutron poisons (far longer-lived than xenon-135) which gradually accumulate over the fuel load's operating life. The energy released in the fission process generates heat, some of which can be converted into usable energy. A common method of harnessing this thermal energy is to use it to boil water to produce pressurized steam which will then drive a steam turbine that turns an alternator and generates electricity. The neutron was discovered in 1932 by British physicist James Chadwick. The concept of a nuclear chain reaction brought about by nuclear reactions mediated by neutrons was first realized shortly thereafter, by Hungarian scientist Leó Szilárd, in 1933. He filed a patent for his idea of a simple reactor the following year while working at the Admiralty in London. However, Szilárd's idea did not incorporate the idea of nuclear fission as a neutron source, since that process was not yet discovered. Szilárd's ideas for nuclear reactors using neutron-mediated nuclear chain reactions in light elements proved unworkable. Inspiration for a new type of reactor using uranium came from the discovery by Lise Meitner, Fritz Strassmann and Otto Hahn in 1938 that bombardment of uranium with neutrons (provided by an alpha-on-beryllium fusion reaction, a "neutron howitzer") produced a barium residue, which they reasoned was created by the fissioning of the uranium nuclei. Subsequent studies in early 1939 (one of them by Szilárd and Fermi) revealed that several neutrons were also released during the fissioning, making available the opportunity for the nuclear chain reaction that Szilárd had envisioned six years previously. On 2 August 1939 Albert Einstein signed a letter to President Franklin D. Roosevelt (written by Szilárd) suggesting that the discovery of uranium's fission could lead to the development of "extremely powerful bombs of a new type", giving impetus to the study of reactors and fission. Szilárd and Einstein knew each other well and had worked together years previously, but Einstein had never thought about this possibility for nuclear energy until Szilard reported it to him, at the beginning of his quest to produce the Einstein-Szilárd letter to alert the U.S. government. Shortly after, Hitler's Germany invaded Poland in 1939, starting World War II in Europe. The U.S. was not yet officially at war, but in October, when the Einstein-Szilárd letter was delivered to him, Roosevelt commented that the purpose of doing the research was to make sure "the Nazis don't blow us up." The U.S. nuclear project followed, although with some delay as there remained skepticism (some of it from Fermi) and also little action from the small number of officials in the government who were initially charged with moving the project forward. The following year the U.S. Government received the Frisch–Peierls memorandum from the UK, which stated that the amount of uranium needed for a chain reaction was far lower than had previously been thought. The memorandum was a product of the MAUD Committee, which was working on the UK atomic bomb project, known as Tube Alloys, later to be subsumed within the Manhattan Project. Eventually, the first artificial nuclear reactor, Chicago Pile-1, was constructed at the University of Chicago, by a team led by Italian physicist Enrico Fermi, in late 1942. By this time, the program had been pressured for a year by U.S. entry into the war. The Chicago Pile achieved criticality on 2 December 1942 at 3:25 PM. The reactor support structure was made of wood, which supported a pile (hence the name) of graphite blocks, embedded in which was natural uranium oxide 'pseudospheres' or 'briquettes'. Soon after the Chicago Pile, the U.S. military developed a number of nuclear reactors for the Manhattan Project starting in 1943. The primary purpose for the largest reactors (located at the Hanford Site in Washington), was the mass production of plutonium for nuclear weapons. Fermi and Szilard applied for a patent on reactors on 19 December 1944. Its issuance was delayed for 10 years because of wartime secrecy. "World's first nuclear power plant" is the claim made by signs at the site of the EBR-I, which is now a museum near Arco, Idaho. Originally called "Chicago Pile-4", it was carried out under the direction of Walter Zinn for Argonne National Laboratory. This experimental LMFBR operated by the U.S. Atomic Energy Commission produced 0.8 kW in a test on 20 December 1951 and 100 kW (electrical) the following day, having a design output of 200 kW (electrical). Besides the military uses of nuclear reactors, there were political reasons to pursue civilian use of atomic energy. U.S. President Dwight Eisenhower made his famous Atoms for Peace speech to the UN General Assembly on 8 December 1953. This diplomacy led to the dissemination of reactor technology to U.S. institutions and worldwide. The first nuclear power plant built for civil purposes was the AM-1 Obninsk Nuclear Power Plant, launched on 27 June 1954 in the Soviet Union. It produced around 5 MW (electrical). After World War II, the U.S. military sought other uses for nuclear reactor technology. Research by the Army and the Air Force never came to fruition; however, the U.S. Navy succeeded when they steamed the USS "Nautilus" (SSN-571) on nuclear power 17 January 1955. The first commercial nuclear power station, Calder Hall in Sellafield, England was opened in 1956 with an initial capacity of 50 MW (later 200 MW). The first portable nuclear reactor "Alco PM-2A" was used to generate electrical power (2 MW) for Camp Century from 1960 to 1963. All commercial power reactors are based on nuclear fission. They generally use uranium and its product plutonium as nuclear fuel, though a thorium fuel cycle is also possible. Fission reactors can be divided roughly into two classes, depending on the energy of the neutrons that sustain the fission chain reaction: In principle, fusion power could be produced by nuclear fusion of elements such as the deuterium isotope of hydrogen. While an ongoing rich research topic since at least the 1940s, no self-sustaining fusion reactor for power generation has ever been built. Used by thermal reactors: In 2003, the French Commissariat à l'Énergie Atomique (CEA) was the first to refer to "Gen II" types in Nucleonics Week. The first mentioning of "Gen III" was in 2000, in conjunction with the launch of the Generation IV International Forum (GIF) plans. "Gen IV" was named in 2000, by the United States Department of Energy (DOE) for developing new plant types. More than a dozen advanced reactor designs are in various stages of development. Some are evolutionary from the PWR, BWR and PHWR designs above, some are more radical departures. The former include the advanced boiling water reactor (ABWR), two of which are now operating with others under construction, and the planned passively safe Economic Simplified Boiling Water Reactor (ESBWR) and AP1000 units (see Nuclear Power 2010 Program). Rolls-Royce aims to sell nuclear reactors for the production of synfuel for aircraft. Generation IV reactors are a set of theoretical nuclear reactor designs currently being researched. These designs are generally not expected to be available for commercial construction before 2030. Current reactors in operation around the world are generally considered second- or third-generation systems, with the first-generation systems having been retired some time ago. Research into these reactor types was officially started by the Generation IV International Forum (GIF) based on eight technology goals. The primary goals being to improve nuclear safety, improve proliferation resistance, minimize waste and natural resource utilization, and to decrease the cost to build and run such plants. Generation V reactors are designs which are theoretically possible, but which are not being actively considered or researched at present. Though some generation V reactors could potentially be built with current or near term technology, they trigger little interest for reasons of economics, practicality, or safety. Controlled nuclear fusion could in principle be used in fusion power plants to produce power without the complexities of handling actinides, but significant scientific and technical obstacles remain. Several fusion reactors have been built, but only recently reactors have been able to release more energy than the amount of energy used in the process. Despite research having started in the 1950s, no commercial fusion reactor is expected before 2050. The ITER project is currently leading the effort to harness fusion power. Thermal reactors generally depend on refined and enriched uranium. Some nuclear reactors can operate with a mixture of plutonium and uranium (see MOX). The process by which uranium ore is mined, processed, enriched, used, possibly reprocessed and disposed of is known as the nuclear fuel cycle. Under 1% of the uranium found in nature is the easily fissionable U-235 isotope and as a result most reactor designs require enriched fuel. Enrichment involves increasing the percentage of U-235 and is usually done by means of gaseous diffusion or gas centrifuge. The enriched result is then converted into uranium dioxide powder, which is pressed and fired into pellet form. These pellets are stacked into tubes which are then sealed and called fuel rods. Many of these fuel rods are used in each nuclear reactor. Most BWR and PWR commercial reactors use uranium enriched to about 4% U-235, and some commercial reactors with a high neutron economy do not require the fuel to be enriched at all (that is, they can use natural uranium). According to the International Atomic Energy Agency there are at least 100 research reactors in the world fueled by highly enriched (weapons-grade/90% enrichment) uranium. Theft risk of this fuel (potentially used in the production of a nuclear weapon) has led to campaigns advocating conversion of this type of reactor to low-enrichment uranium (which poses less threat of proliferation). Fissile U-235 and non-fissile but fissionable and fertile U-238 are both used in the fission process. U-235 is fissionable by thermal (i.e. slow-moving) neutrons. A thermal neutron is one which is moving about the same speed as the atoms around it. Since all atoms vibrate proportionally to their absolute temperature, a thermal neutron has the best opportunity to fission U-235 when it is moving at this same vibrational speed. On the other hand, U-238 is more likely to capture a neutron when the neutron is moving very fast. This U-239 atom will soon decay into plutonium-239, which is another fuel. Pu-239 is a viable fuel and must be accounted for even when a highly enriched uranium fuel is used. Plutonium fissions will dominate the U-235 fissions in some reactors, especially after the initial loading of U-235 is spent. Plutonium is fissionable with both fast and thermal neutrons, which make it ideal for either nuclear reactors or nuclear bombs. Most reactor designs in existence are thermal reactors and typically use water as a neutron moderator (moderator means that it slows down the neutron to a thermal speed) and as a coolant. But in a fast breeder reactor, some other kind of coolant is used which will not moderate or slow the neutrons down much. This enables fast neutrons to dominate, which can effectively be used to constantly replenish the fuel supply. By merely placing cheap unenriched uranium into such a core, the non-fissionable U-238 will be turned into Pu-239, "breeding" fuel. In thorium fuel cycle thorium-232 absorbs a neutron in either a fast or thermal reactor. The thorium-233 beta decays to protactinium-233 and then to uranium-233, which in turn is used as fuel. Hence, like uranium-238, thorium-232 is a fertile material. The amount of energy in the reservoir of nuclear fuel is frequently expressed in terms of "full-power days," which is the number of 24-hour periods (days) a reactor is scheduled for operation at full power output for the generation of heat energy. The number of full-power days in a reactor's operating cycle (between refueling outage times) is related to the amount of fissile uranium-235 (U-235) contained in the fuel assemblies at the beginning of the cycle. A higher percentage of U-235 in the core at the beginning of a cycle will permit the reactor to be run for a greater number of full-power days. At the end of the operating cycle, the fuel in some of the assemblies is "spent", having spent 4 to 6 years in the reactor producing power. This spent fuel is discharged and replaced with new (fresh) fuel assemblies. Though considered "spent," these fuel assemblies contain a large quantity of fuel. In practice it is economics that determines the lifetime of nuclear fuel in a reactor. Long before all possible fission has taken place, the reactor is unable to maintain 100%, full output power, and therefore, income for the utility lowers as plant output power lowers. Most nuclear plants operate at a very low profit margin due to operating overhead, mainly regulatory costs, so operating below 100% power is not economically viable for very long. The fraction of the reactor's fuel core replaced during refueling is typically one-third, but depends on how long the plant operates between refueling. Plants typically operate on 18 month refueling cycles, or 24 month refueling cycles. This means that 1 refueling, replacing only one-third of the fuel, can keep a nuclear reactor at full power for nearly 2 years. The disposition and storage of this spent fuel is one of the most challenging aspects of the operation of a commercial nuclear power plant. This nuclear waste is highly radioactive and its toxicity presents a danger for thousands of years. After being discharged from the reactor, spent nuclear fuel is transferred to the on-site spent fuel pool. The spent fuel pool is a large pool of water that provides cooling and shielding of the spent nuclear fuel. Once the energy has decayed somewhat (approximately 5 years), the fuel can be transferred from the fuel pool to dry shielded casks, that can be safely stored for thousands of years. After loading into dry shielded casks, the casks are stored on-site in a specially guarded facility in impervious concrete bunkers. On-site fuel storage facilities are designed to withstand the impact of commercial airliners, with little to no damage to the spent fuel. An average on-site fuel storage facility can hold 30 years of spent fuel in a space smaller that a football field. Not all reactors need to be shut down for refueling; for example, pebble bed reactors, RBMK reactors, molten salt reactors, Magnox, AGR and CANDU reactors allow fuel to be shifted through the reactor while it is running. In a CANDU reactor, this also allows individual fuel elements to be situated within the reactor core that are best suited to the amount of U-235 in the fuel element. The amount of energy extracted from nuclear fuel is called its burnup, which is expressed in terms of the heat energy produced per initial unit of fuel weight. Burn up is commonly expressed as megawatt days thermal per metric ton of initial heavy metal. Nuclear safety covers the actions taken to prevent nuclear and radiation accidents and incidents or to limit their consequences. The nuclear power industry has improved the safety and performance of reactors, and has proposed new safer (but generally untested) reactor designs but there is no guarantee that the reactors will be designed, built and operated correctly. Mistakes do occur and the designers of reactors at Fukushima in Japan did not anticipate that a tsunami generated by an earthquake would disable the backup systems that were supposed to stabilize the reactor after the earthquake, despite multiple warnings by the NRG and the Japanese nuclear safety administration. According to UBS AG, the Fukushima I nuclear accidents have cast doubt on whether even an advanced economy like Japan can master nuclear safety. Catastrophic scenarios involving terrorist attacks are also conceivable. An interdisciplinary team from MIT has estimated that given the expected growth of nuclear power from 2005–2055, at least four serious nuclear accidents would be expected in that period. Some serious nuclear and radiation accidents have occurred. Nuclear power plant accidents include the SL-1 accident (1961), the Three Mile Island accident (1979), Chernobyl disaster (1986), and the Fukushima Daiichi nuclear disaster (2011). Nuclear-powered submarine mishaps include the K-19 reactor accident (1961), the K-27 reactor accident (1968), and the K-431 reactor accident (1985). Nuclear reactors have been launched into Earth orbit at least 34 times. A number of incidents connected with the unmanned nuclear-reactor-powered Soviet RORSAT radar satellite program resulted in spent nuclear fuel reentering the Earth's atmosphere from orbit. Almost two billion years ago a series of self-sustaining nuclear fission "reactors" self-assembled in the area now known as Oklo in Gabon, West Africa. The conditions at that place and time allowed a natural nuclear fission to occur with circumstances that are similar to the conditions in a constructed nuclear reactor. Fifteen fossil natural fission reactors have so far been found in three separate ore deposits at the Oklo uranium mine in Gabon. First discovered in 1972 by French physicist Francis Perrin, they are collectively known as the Oklo Fossil Reactors. Self-sustaining nuclear fission reactions took place in these reactors approximately 1.5 billion years ago, and ran for a few hundred thousand years, averaging 100 kW of power output during that time. The concept of a natural nuclear reactor was theorized as early as 1956 by Paul Kuroda at the University of Arkansas. Such reactors can no longer form on Earth in its present geologic period. Radioactive decay of formerly more abundant uranium-235 over the time span of hundreds of millions of years has reduced the proportion of this naturally occurring fissile isotope to below the amount required to sustain a chain reaction with only plain water as a moderator. The natural nuclear reactors formed when a uranium-rich mineral deposit became inundated with groundwater that acted as a neutron moderator, and a strong chain reaction took place. The water moderator would boil away as the reaction increased, slowing it back down again and preventing a meltdown. The fission reaction was sustained for hundreds of thousands of years, cycling on the order of hours to a few days. These natural reactors are extensively studied by scientists interested in geologic radioactive waste disposal. They offer a case study of how radioactive isotopes migrate through the Earth's crust. This is a significant area of controversy as opponents of geologic waste disposal fear that isotopes from stored waste could end up in water supplies or be carried into the environment. Nuclear reactors produce tritium as part of normal operations, which is eventually released into the environment in trace quantities. As an isotope of hydrogen, tritium (T) frequently binds to oxygen and forms T2O. This molecule is chemically identical to H2O and so is both colorless and odorless, however the additional neutrons in the hydrogen nuclei cause the tritium to undergo beta decay with a half-life of 12.3 years. Despite being measurable, the tritium released by nuclear power plants is minimal. The United States NRC estimates that a person drinking water for one year out of a well contaminated by what they would consider to be a significant tritiated water spill would receive a radiation dose of 0.3 millirem. For comparison, this is an order of magnitude less than the 4 millirem a person receives on a round trip flight from Washington, D.C. to Los Angeles, a consequence of less atmospheric protection against highly energetic cosmic rays at high altitudes. The amounts of strontium-90 released from nuclear power plants under normal operations is so low as to be undetectable above natural background radiation. Detectable strontium-90 in ground water and the general environment can be traced to weapons testing that occurred during the mid-20th century (accounting for 99% of the Strontium-90 in the environment) and the Chernobyl accident (accounting for the remaining 1%).
https://en.wikipedia.org/wiki?curid=22151
Edsger W. Dijkstra Edsger Wybe Dijkstra (; ; 11 May 1930 – 6 August 2002) was a Dutch computer scientist, programmer, software engineer, systems scientist, science essayist, and pioneer in computing science. A theoretical physicist by training, he worked as a programmer at the Mathematisch Centrum (Amsterdam) from 1952 to 1962. A university professor for much of his life, Dijkstra held the Schlumberger Centennial Chair in Computer Sciences at the University of Texas at Austin from 1984 until his retirement in 1999. He was a professor of mathematics at the Eindhoven University of Technology (1962–1984) and a research fellow at the Burroughs Corporation (1973–1984). One of the most influential figures of computing science's founding generation, Dijkstra helped shape the new discipline from both an engineering and a theoretical perspective. His fundamental contributions cover diverse areas of computing science, including compiler construction, operating systems, distributed systems, sequential and concurrent programming, programming paradigm and methodology, programming language research, program design, program development, program verification, software engineering principles, graph algorithms, and philosophical foundations of computer programming and computer science. Many of his papers are the source of new research areas. Several concepts and problems that are now standard in computer science were first identified by Dijkstra and/or bear names coined by him. As a foremost opponent of the mechanizing view of computing science, he refuted the use of the concepts of 'computer science' and 'software engineering' as umbrella terms for academic disciplines. Until the mid-1960s computer programming was considered more an art (or a craft) than a scientific discipline. In Harlan Mills's words (1986), "programming [before the 1970s] was regarded as a private, puzzle-solving activity of writing computer instructions to work as a program". In the late 1960s, computer programming was in a state of crisis. Dijkstra was one of a small group of academics and industrial programmers who advocated a new programming style to improve the quality of programs. Dijkstra, who had a background in mathematics and physics, was one of the driving forces behind the acceptance of computer programming as a scientific discipline. He coined the phrase "structured programming" and during the 1970s this became the new programming orthodoxy. His ideas about structured programming helped lay the foundations for the birth and development of the professional discipline of software engineering, enabling programmers to organize and manage increasingly complex software projects. As Bertrand Meyer (2009) noted, "The revolution in views of programming started by Dijkstra's iconoclasm led to a movement known as structured programming, which advocated a systematic, rational approach to program construction. Structured programming is the basis for all that has been done since in programming methodology, including object-oriented programming." The academic study of concurrent computing started in the 1960s, with Dijkstra (1965) credited with being the first paper in this field, identifying and solving the mutual exclusion problem. He was also one of the early pioneers of the research on principles of distributed computing. His foundational work on concurrency, semaphores, mutual exclusion, deadlock (deadly embrace), finding shortest paths in graphs, fault-tolerance, self-stabilization, among many other contributions comprises many of the pillars upon which the field of distributed computing is built. Shortly before his death in 2002, he received the ACM PODC Influential-Paper Award in distributed computing for his work on self-stabilization of program computation. This annual award was renamed the Dijkstra Prize (Edsger W. Dijkstra Prize in Distributed Computing) the following year. As the prize, sponsored jointly by the ACM Symposium on Principles of Distributed Computing (PODC) and the EATCS International Symposium on Distributed Computing (DISC), recognizes that "No other individual has had a larger influence on research in principles of distributed computing". Edsger W. Dijkstra was born in Rotterdam. His father was a chemist who was president of the Dutch Chemical Society; he taught chemistry at a secondary school and was later its superintendent. His mother was a mathematician, but never had a formal job. Dijkstra had considered a career in law and had hoped to represent the Netherlands in the United Nations. However, after graduating from school in 1948, at his parents' suggestion he studied mathematics and physics and then theoretical physics at the University of Leiden. In the early 1950s, electronic computers were a novelty. Dijkstra stumbled on his career quite by accident, and through his supervisor, Professor A. Haantjes, he met Adriaan van Wijngaarden, the director of the Computation Department at the Mathematical Center in Amsterdam, who offered Dijkstra a job; he officially became the Netherlands' first "programmer" in March 1952. For some time Dijkstra remained committed to physics, working on it in Leiden three days out of each week. With increasing exposure to computing, however, his focus began to shift. As he recalled: When Dijkstra married Maria (Ria) C. Debets in 1957, he was required as a part of the marriage rites to state his profession. He stated that he was a programmer, which was unacceptable to the authorities, there being no such profession then in The Netherlands. In 1959, he received his PhD from the University of Amsterdam for a thesis entitled 'Communication with an Automatic Computer', devoted to a description of the assembly language designed for the first commercial computer developed in the Netherlands, the X1. His thesis supervisor was Van Wijngaarden. From 1952 until 1962, Dijkstra worked at the Mathematisch Centrum in Amsterdam, where he worked closely with Bram Jan Loopstra and Carel S. Scholten, who had been hired to build a computer. Their mode of interaction was disciplined: They would first decide upon the interface between the hardware and the software, by writing a programming manual. Then the hardware designers would have to be faithful to their part of the contract, while Dijkstra, the programmer, would write software for the nonexistent machine. Two of the lessons he learned from this experience were the importance of clear documentation, and that program debugging can be largely avoided through careful design. Dijkstra formulated and solved the shortest path problem for a demonstration at the official inauguration of the ARMAC computer in 1956. Because of the absence of journals dedicated to automatic computing, he did not publish the result until 1959. At the Mathematical Center, Dijkstra and his colleague developed the first compiler for the programming language ALGOL 60 by August 1960, more than a year before a compiler was produced by another group. It had a profound influence on his later thinking on programming as a scientific activity. In 1962, Dijkstra moved to Eindhoven, and later to Nuenen, in the south of the Netherlands, where he became a professor in the Mathematics Department at the Eindhoven University of Technology. The university did not have a separate computer science department and the culture of the mathematics department did not particularly suit him. Dijkstra tried to build a group of computer scientists who could collaborate on solving problems. This was an unusual model of research for the Mathematics Department. In the late 1960s he built the THE operating system (named for the university, then known as Technische Hogeschool Eindhoven), which has influenced the designs of subsequent operating systems through its use of software based paged virtual memory. Dijkstra joined Burroughs Corporation, a company known then for producing computers based on an innovative hardware architecture, as its research fellow in August 1973. His duties consisted of visiting some of the firm's research centers a few times a year and carrying on his own research, which he did in the smallest Burroughs research facility, namely, his study on the second floor of his house in Nuenen. In fact, Dijkstra was the only research fellow of Burroughs and worked for it from home, occasionally travelling to its branches in the United States. As a result, he reduced his appointment at the university to one day a week. That day, Tuesday, soon became known as the day of the famous 'Tuesday Afternoon Club', a seminar during which he discussed with his colleagues scientific articles, looking at all aspects: notation, organisation, presentation, language, content, etc. Shortly after he moved in 1984 to the University of Texas at Austin (USA), a new 'branch' of the Tuesday Afternoon Club emerged in Austin, Texas. The Burroughs years saw him at his most prolific in output of research articles. He wrote nearly 500 documents in the EWD series (described below), most of them technical reports, for private circulation within a select group. Dijkstra accepted the Schlumberger Centennial Chair in the Computer Science Department at the University of Texas at Austin in 1984. Dijkstra worked in Austin until his retirement in November 1999. To mark the occasion and to celebrate his forty-plus years of seminal contributions to computing science, the Department of Computer Sciences organized a symposium, which took place on his 70th birthday in May 2000. Dijkstra and his wife returned from Austin to his original house in Nuenen (Netherlands) where he found that he had only months to live. He said that he wanted to retire in Austin, Texas, but to die in the Netherlands. Dijkstra died on 6 August 2002 after a long struggle with cancer. He and his wife Maria (Ria) Debets were survived by their three children: Marcus, Femke, and the computer scientist Rutger M. Dijkstra. As an early pioneer in many research areas of computing science, Dijkstra helped shape the new discipline from both an engineering and an academic perspective. Many of his papers are the source of new research areas. Many concepts that are now standard in computer science were first identified by Dijkstra and/or bear names coined by him. Several important problems were also first formulated and solved by him. A 1994 survey of over a thousand professors of computer science was conducted to obtain a list of 38 most influential scholarly papers in the field, and Dijkstra is the author of five papers. During his forty-plus years as a computing scientist, which included positions in both academia and industry, Dijkstra made numerous seminal contributions to many areas of computing science, including compiler construction, operating systems, concurrent programming (concurrent computing), distributed programming (distributed computing), programming paradigm and methodology, programming language research, program design, program development, program verification, software engineering principles, algorithm design, and philosophical foundations of computer programming and computer science. In addition, Dijkstra was intensely interested in teaching computer science, and in the relationships between academic computing science and the software industry. Dijkstra's algorithmic work (especially graph algorithms, concurrent algorithms, and distributed algorithms) plays an important role in many areas of computing science. According to Leslie Lamport (2002), Dijkstra "started the field of concurrent and distributed algorithms with his 1965 CACM paper "Solution of a Problem in Concurrent Programming Control", in which he first stated and solved the mutual exclusion problem." As Lamport explains, "that paper is probably why PODC exists (...). It remains to this day the most influential paper in the field. That it did not win a PODC Influential Paper Award reflects an artificial separation between concurrent and distributed algorithms–a separation that has never existed in Dijkstra's work." In 1959 Dijkstra published in a 3-page article 'A note on two problems in connexion with graphs' the algorithm to find the shortest path in a graph between any two given nodes, now called Dijkstra's algorithm. Its impact over the next 40 years is summarised from the article of Mikkel Thorup, 'Undirected Single Source Shortest Paths with Positive Integer Weights in Linear Time' (1999): "Since 1959, all theoretical developments in SSSP [Single-Source Shortest Paths] for general directed and undirected graphs have been based on Dijkstra's algorithm." Dijkstra's algorithm is used in SPF, Shortest Path First, which is used in the routing protocols OSPF and IS-IS. Various modifications to Dijkstra's algorithm have been proposed by many authors using heuristics to reduce the run time of shortest path search. One of the most used heuristic algorithms is the A* search algorithm (first described by Peter Hart, Nils Nilsson and Bertram Raphael of Stanford Research Institute in 1968), the main goal is to reduce the run time by reducing the search space. Dijkstra thought about the shortest path problem when working at the Mathematical Center in Amsterdam in 1956 as a programmer to demonstrate capabilities of a new computer named ARMAC. His objective was to choose both a problem as well as an answer (that would be produced by computer) that non-computing people could understand. He designed the shortest path algorithm in about 20 minutes without aid of paper and pen and later implemented it for ARMAC for a slightly simplified transportation map of 64 cities in the Netherlands (so that 6 bits would suffice to represent the city in the algorithm). As he recalled, in an interview published in 2001: A year later, he came across another problem from hardware engineers working on the institute's next computer: minimize the amount of wire needed to connect the pins on the back panel of the machine. As a solution, he rediscovered the algorithm known as Prim's minimal spanning tree algorithm. The Prim's algorithm was originally developed in 1930 by Czech mathematician Vojtěch Jarník. and later independently rediscovered and republished by Robert C. Prim in 1957, and Dijkstra in 1959. Therefore, it is also sometimes called the DJP algorithm. In 1961 Dijkstra first described the shunting-yard algorithm, a method for parsing mathematical expressions specified in infix notation, in the Mathematisch Centrum report. It can be used to produce output in Reverse Polish notation (RPN) or as an abstract syntax tree (AST). The algorithm was named the "shunting yard" algorithm because its operation resembles that of a railroad shunting yard. The shunting-yard algorithm is commonly used to implement operator-precedence parsers. In 1962 or 1963 Dijkstra proposed the semaphore mechanism for mutual exclusion algorithm for n processes (a generalization of Dekker's algorithm), which was probably the first published concurrent algorithm and which introduced a new area of algorithmic research. He also identified the deadlock problem and proposed the banker's algorithm that prevents deadlock. In 1974 Dijkstra presented three self-stabilizing algorithms for mutual exclusion on a ring. Dijkstra's work is considered to be the first to introduce and demonstrate the self-stabilization concept. In the mid-1970s Dijkstra (together with other authors) introduced two useful abstractions (mutator and collector) to the study of garbage collection. The mutator abstracts the process that performs the computation, including allocation of a new storage cell. The collector is the process that automatically reclaims garbage. Furthermore, this paper gives a formalization of "tri-color marking" that is basic to incremental garbage collection. In the early 1980s Dijkstra and Carel S. Scholten proposed the Dijkstra–Scholten algorithm for detecting termination in distributed systems. In 1981 Dijkstra developed smoothsort, a comparison-based sorting algorithm and a variation of heapsort. Dijkstra was known to be a fan of the programming language ALGOL, and worked on the team that implemented the first compiler for ALGOL 60, which he was closely involved in developing, realising, and popularising. As discussed by Peter Naur in the article 'The European side of the last phase of the development of ALGOL 60', in the "Proceedings of the First ACM SIGPLAN Conference on History of Programming Languages", January 1978, Dijkstra took part in the period 1958–1959 in a number of meetings that culminated in the publication of the report defining the ALGOL 60 language. Dijkstra's name does not appear in the list of 13 authors of the final report, but he was an official member of the International Federation for Information Processing (IFIP) IFIP Working Group 2.1, which supports and maintains ALGOL 60 and ALGOL 68. He eventually left the committee, apparently because he could not agree with the majority opinions. Still, while at the Mathematisch Centrum (Amsterdam), he wrote jointly with Jaap Zonneveld the first ALGOL 60 compiler. Dijkstra and Zonneveld, who collaborated on the compiler, agreed not to shave until the project was completed. Zonneveld shaved shortly thereafter, but Dijkstra kept his beard for the rest of his life. ALGOL was the result of a collaboration of American and European committees. ALGOL 60 (short for ALGOrithmic Language 1960) is a member of the ALGOL family of computer programming languages. It followed on from ALGOL 58 and inspired many languages that followed it. It gave rise to many other languages, including BCPL, B, Pascal, Simula, and C. Algol 60 was a sophisticatedly designed computer language and it provided a large number of hitherto unknown implementation challenges. As Bjarne Stroustrup notes, "one problem with Algol60 was that no one knew how to implement it." A major new challenge in Algol 60 implementation was the run-time allocation and management of data. In 1960, Dijkstra and Zonneveld showed how recursive procedures could be executed using a run-time stack of activation records, and how to efficiently access identifiers from statically enclosing scopes using the so-called "display". The ALGOL 60 compiler was one of the first to support recursion employing a novel method to do so. Dijkstra's short book "Primer of Algol 60 Programming", originally published in 1962, was the standard reference for the language for several years. Computer programming in the 1950s to 1960s was not recognized as an academic discipline and unlike mature sciences there were no theoretical concepts or coding systems. Programming as a professional activity was poorly understood in those years. In the late 1960s computer programming was in state of crisis. Software crisis is a term used in the early days of computing science for the difficulty of writing useful and efficient computer programs in the required time. The software crisis was due to the rapid increases in computer power and the complexity of the problems that could be tackled. With the increase in the complexity of the software, many software problems arose because existing methods were insufficient. The term "software crisis" was coined by some attendees at the first NATO Software Engineering Conference in 1968 at Garmisch, Germany. His 1972 ACM Turing Award Lecture makes reference to this same problem: "The major cause of the software crisis is that the machines have become several orders of magnitude more powerful! To put it quite bluntly: as long as there were no machines, programming was no problem at all; when we had a few weak computers, programming became a mild problem, and now we have gigantic computers, programming has become an equally gigantic problem." While Dijkstra had programmed extensively in machine code in the 1950s, he came to the conclusion that in high-level languages frequent use of the GOTO statement was usually symptomatic of poor structure. In 1968 he wrote a private paper "A Case against the GO TO Statement", which was then published as a letter in CACM. Editor Niklaus Wirth gave this letter the heading "Go To Statement Considered Harmful", which introduced the phrase "considered harmful" into computing. Dijkstra argued that the programming statement GOTO, found in many high-level programming languages, is a major source of errors, and should therefore be eliminated. This letter caused a huge debate in the programming community. Some went to the length of equating good programming with the elimination of GO TO. Dijkstra refused to mention the debate, or even the GO TO statement, in his article "Notes on Structured Programming". The debate has long since died down; programming languages provide alternatives to the GO TO, few programmers today use it liberally, and most never use it at all. Dijkstra's thesis was that departures from linear control flow were clearer if allowed only in disciplined higher-level structures such as the if-then-else statement and the while loop. This methodology was developed into structured programming movement, the title of his 1972 book, coauthored with C.A.R. Hoare and Ole-Johan Dahl. Considered by many as the first significant movement in history of computer programming, structured programming became the new programming orthodoxy during the 1970s. Bertrand Meyer remarked that, "The revolution in views of programming started by Dijkstra's iconoclasm led to a movement known as structured programming, which advocated a systematic, rational approach to program construction. Structured programming is the basis for all that has been done since in programming methodology, including object-oriented programming." Structured programming is often regarded as "goto-less programming". But as Bertrand Meyer notes, "As the first book on the topic ["Structured Programming" by Dijkstra, Dahl, and Hoare] shows, structured programming is about much more than control structures and the goto. Its principal message is that programming should be considered a scientific discipline based on mathematical rigor." , structured programming – especially in the 1970s and 1980s – significantly influenced the birth of many modern programming languages such as Pascal, C, Modula-2, and Ada. The Fortran 77 version which incorporates the concepts of structured programming, was released in 1978. The C++ language was a considerably extended and enhanced version of the popular C (see also: list of C-based programming languages). Since C++ was developed from a more traditional , it is a 'hybrid language', rather than a pure object-oriented programming language. In his article "Structured Programming: Retrospect and Prospect" (1986), Harlan Mills writes, "Edsger W. Dijkstra's 1969 "Structured Programming" article precipitated a decade of intense focus on programming techniques that has fundamentally altered human expectations and achievements in software development. Before this decade of intense focus, programming was regarded as a private, puzzle-solving activity of writing computer instructions to work as a program. After this decade, programming could be regarded as a public, mathematics-based activity of restructuring specifications into programs. Before, the challenge was in getting programs to run at all, and then in getting them further debugged to do the right things. After, programs could be expected to both run and do the right things with little or no debugging. Before, it was common wisdom that no sizable program could be error-free. After, many sizable programs have run a year or more with no errors detected. These expectations and achievements are not universal because of the inertia of industrial practices. But they are well-enough established to herald fundamental change in software development." The book "Concise Encyclopedia of Computer Science" (2004), edited by Edwin D. Reilly, notes that "the major contributions of structured programming have been twofold—the elevation of programming technique to something less of an art and more of a science, and the demonstration that carefully structured programs can be creative works of sufficient literary merit to deserve being read by humans and not just by computer." Dijkstra's ideas about programming methodology (especially the structured programming movement) helped lay the foundations for the birth and development of the professional discipline of software engineering (in particular the software design and development), enabling programmers to organize and manage increasingly complex software projects. In the late 1960s Dijkstra discussed the concept of program families. And in the mid 1970s David Parnas and others clarified the idea and showed how to apply it in software engineering principles. The rise of the structured programming movement led to many other "structured" approaches applied to software design. The techniques of structured analysis and structured design are outgrowths of structured programming concepts and techniques, and of the early ideas about modular design. Principles of modularity were strengthened by Larry Constantine's concepts of coupling (to be minimized between modules) and cohesion (to be maximized within modules), by David Parnas's techniques of information hiding, and by abstract data types. A number of tools and methods employing structured concepts were developed, such as Structured Design, Jackson's Structured Programming, Ross' Structured Analysis and Design Technique (SADT), Yourdon's Structured Method, Structured Systems Analysis and Design Method (SSADM), and James Martin's information technology engineering. The field of software metrics is often considered as a direct influence of the structured programming movement on software engineering in the 1970s. Separation of concerns (SoC), one of the basic principles in software engineering, is a design principle for separating a computer program into distinct sections, such that each section addresses a separate concern. The term "separation of concerns" was coined by Dijkstra in his 1974 paper "On the role of scientific thought". In the 1960s Dijkstra and his colleagues in Eindhoven designed and implemented THE (standing for 'Technische Hogeschool Eindhoven') operating system, which was organised into clearly identified abstraction layers. His 1968 article on this subject provided the foundation for subsequent designs of the operating systems. The IEEE Computer Society's David Alan Grier writes, "We generally trace the idea of building computer systems in layers back to a 1967 paper that the Dutch computer scientist Edsger Dijkstra gave to a joint IEEE Computer Society/ACM conference. Prior to this paper, engineers had struggled with the problem of how to organize software. If you look at early examples of programs, and you can find many in the electronic library of the Computer Society, you will find that most code of that era is complicated, difficult to read, hard to modify, and challenging to reuse. In his 1967 paper, Dijkstra described how software could be constructed in layers and gave an example of a simple operating system that used five layers. He admitted that this system might not be a realistic test of his ideas but he argued that the "larger the project, the more essential the structuring!" The idea of using layers to control complexity has become a mainstay of software architecture. We see it in many forms and apply it to many problems. We see it in the hierarchy of classes in object-oriented programming and in the structure of service-oriented architecture (SOA). SOA is a relatively recent application of layering in computer science. It was articulated in 2007 as a means of controlling complexity in business systems, especially distributed systems that make substantial use of the Internet. Like Dijkstra's plan for system development, its layering system is called the SOA Solution Stack or S3. The S3's nine layers are: 1) operational systems, 2) service components, 3) services, 4) business processes, 5) consumer actions, 6) system integration, 7) quality control and assurance, 8) information architecture, and 9) system governance and policies." Dijkstra organized the design of the system in layers in order to reduce the overall complexity of the software. Though the term 'architecture' had not yet been used to describe software design, this was certainly considered the first glimpse of software architecture. It introduced a number of design principles which have become part of the working vocabulary of every professional programmer: levels of abstraction, programming in layers, the semaphore, and cooperating sequential processes. His original paper on the THE operating system was reprinted in the 25th Anniversary issue of Communications of the ACM, in January 1983. By way of introduction, the Editor-in-Chief says, "This project initiated a long line of research in multilevel systems architecture — a line that continues to the present day because hierarchical modularity is a powerful approach to organizing large systems." In a one-page paper from 1965 Dijkstra introduced the 'mutual exclusion problem' for n processes and discussed a solution to it. It was probably the first published concurrent algorithm. The notion, standard by now, of a 'critical section' was also introduced in this paper. Per Brinch Hansen, a pioneer in the field of concurrent computing, considers Dijkstra's "Cooperating Sequential Processes" (1965) to be the first classic paper in concurrent programming. As Brinch Hansen notes, 'Dijkstra lays the conceptual foundation for abstract concurrent programming' with that paper. In 1968 Dijkstra published his seminal paper 'Cooperating sequential processes', a 70-page essay that originated the field of concurrent programming. He discussed in it the notion of mutual exclusion (mutex) and the criteria a satisfactory solution should satisfy. He also redressed the historical perspective left out of his 1965 paper by including the first known correct solution to the mutual exclusion problem, for two processes, due to Theodorus Dekker. Dijkstra subsequently generalized Dekker's solution to n processes. Further, he proposed the first synchronisation mechanism for concurrent processes, the semaphore with its two operations, P and V. He also identified the 'deadlock problem' (called there 'the problem of the deadly embrace') and proposed an elegant 'Banker's algorithm' that prevents deadlock. The deadlock detection and prevention became perennial research problems in the field of concurrent programming. The dining philosophers problem is an example problem often used in concurrent algorithm design to illustrate synchronization issues and techniques for resolving them. It was originally formulated in 1965 by Dijkstra as a student exam exercise, presented in terms of computers competing for access to tape drive peripherals. Soon after, Tony Hoare gave the problem its present formulation. The sleeping barber problem is also attributed to Dijkstra. In his book "Concurrent Programming: Algorithms, Principles, and Foundations" Michel Raynal writes, "Since the early work of E.W. Dijkstra (1965), who introduced the mutual exclusion problem, the concept of a process, the semaphore object, the notion of a weakest precondition, and guarded commands (among many other contributions), synchronization is no longer a catalog of tricks but a domain of computing science with its own concepts, mechanisms, and techniques whose results can be applied in many domains. This means that process synchronization has to be a major topic of any computer science curriculum." John W. McCormick et al. (2011) notes, "The notion of the concurrent program as a means for writing parallel programs without regard for the underlying hardware was first introduced by Edsger Dijkstra (1968). Moti Ben-Ari (1982) elegantly summed up Dijkstra's idea in three sentences: ‘Concurrent programming is the name given to programming notation and techniques for expressing potential parallelism and solving the resulting synchronization and communication problems. Implementation of parallelism is a topic in computer systems (hardware and software) that is essentially independent of concurrent programming. Concurrent programming is important because it provides an abstract setting in which to study parallelism without getting bogged down in the implementation details.’" Dijkstra was one of the very early pioneers of the research on principles of distributed computing. As the citation for the Dijkstra Prize recognizes, "no other individual has had a larger influence on research in principles of distributed computing." Some of his papers are even considered to be those that established the field. Dijkstra's 1965 paper, "Solution of a Problem in Concurrent Programming Control" was the first to present the correct solution to the mutual exclusion problem. Leslie Lamport writes that this work "is probably why PODC exists" and it "started the field of concurrent and distributed algorithms". In particular, his paper "Self-stabilizing Systems in Spite of Distributed Control" (1974) started the sub-field of self-stabilization. It is also considered as the first scientific examination of fault-tolerant systems. Dijkstra's paper was not widely noticed until Leslie Lamport's invited talk at the ACM Symposium on Principles of Distributed Computing (PODC) in 1983. In his report on Dijkstra's work on self-stabilizing distributed systems, Lamport regard it to be 'a milestone in work on fault tolerance' and 'a very fertile field for research'. From the 1970s, Dijkstra's chief interest was formal verification. In 1976 Dijkstra published a seminal book, "A Discipline of Programming", which put forward his method of systematic development of programs together with their correctness proofs. In his exposition he used his 'Guarded Command Language'. The language, with its reliance on non-determinism, the adopted weakest precondition semantics and the proposed development method has had a considerable impact on the field to this day. The refinement calculus, originally proposed by Ralph-Johan Back and developed by Carroll Morgan, is an extension of Dijkstra's weakest precondition calculus, where program statements are modeled as predicate transformers. In 1984, to add further support to this approach to programming, he published jointly with Wim Feijen an introductory textbook for first-year students of computer science. The book, first published in Dutch, was entitled "Een methode van programmeren". The English edition appeared in 1988 as "A Method of Programming". Many of his opinions on computer science and programming have become widespread. For example, the programming phrase "two or more, use a for" (a rule of thumb when to use a loop) is sometimes attributed to him. He was the first to make the claim that programming is so inherently complex that, in order to manage it successfully, programmers need to harness every trick and abstraction possible. Dijkstra was one of the most famous opponents of the engineering view of computing science. Like Peter Naur and Kristen Nygaard, Dijkstra disliked the very term 'computer science'. Computer science, as Dijkstra pointed out, deserves a better name. He suggests it can be called 'computing science'. Instead of the computer, or computing technology, Dijkstra wanted to emphasize the abstract mechanisms that computing science uses to master complexity. When expressing the abstract nature of computing science, he wrote, In "The Humble Programmer" (1972), Dijkstra wrote: "We must not forget that it is not our [computing scientists'] business to make programs, it is our business to design classes of computations that will display a desired behaviour." Dijkstra also opposed the inclusion of software engineering under the umbrella of academic computer science. He wrote that, "As economics is known as "The Miserable Science", software engineering should be known as "The Doomed Discipline", doomed because it cannot even approach its goal since its goal is self-contradictory." And "software engineering has accepted as its charter "How to program if you cannot."." In the world of computing science, Dijkstra is well known as a "character". In the preface of his book "A Discipline of Programming" (1976) he stated the following: "For the absence of a bibliography I offer neither explanation nor apology." In fact, most of his articles and books have no references at all. This approach to references was deplored by some researchers. But Dijkstra chose this way of working to preserve his self-reliance. As a university professor for much of his life, Dijkstra saw teaching not just as a required activity but as a serious research endeavor. His approach to teaching was unconventional. His lecturing style has been described as idiosyncratic. When lecturing, the long pauses between sentences have often been attributed to the fact that English is not Dijkstra's first language. However the pauses also served as a way for him to think on his feet and he was regarded as a quick and deep thinker while engaged in the act of lecturing. His courses for students in Austin had little to do with computer science but they dealt with the presentation of mathematical proofs. At the beginning of each semester he would take a photo of each of the students, in order to memorize their names. He never followed a textbook, with the possible exception of his own while it was under preparation. When lecturing, he would write proofs in chalk on a blackboard rather than using overhead foils. He invited the students to suggest ideas, which he then explored, or refused to explore because they violated some of his tenets. He assigned challenging homework problems, and would study his students' solutions thoroughly. He conducted his final examinations orally, over a whole week. Each student was examined in Dijkstra's office or home, and an exam lasted several hours. He was also highly original in his way of assessing people's capacity for a job. When Vladimir Lifschitz came to Austin in 1990 for a job interview, Dijkstra gave him a puzzle. Vladimir solved it and has been working in Austin since then. Despite having invented much of the technology of software, Dijkstra eschewed the use of computers in his own work for many decades. Even after he succumbed to his UT colleagues' encouragement and acquired a Macintosh computer, he used it only for e-mail and for browsing the World Wide Web. Dijkstra never wrote his articles using a computer. He preferred to rely on his typewriter and later on his Montblanc pen. Dijkstra's favorite writing instrument was the Montblanc Meisterstück fountain pen. He repeatedly tried other pens, but none ever displaced the Montblanc. He had no use for word processors, believing that one should be able to write a letter or article without rough drafts, rewriting, or any significant editing. He would work it all out in his head before putting pen to paper, and once mentioned that when he was a physics student he would solve his homework problems in his head while walking the streets of Leiden. Most of Dijkstra's publications were written by him alone. He never had a secretary and took care of all his correspondence alone. When colleagues prepared a Festschrift for his sixtieth birthday, published by Springer-Verlag, he took the trouble to thank each of the 61 contributors separately, in a hand-written letter. Throughout Dijkstra's career, his work was characterized by elegance and economy. A prolific writer (especially as an essayist), Dijkstra authored more than 1,300 papers, many written by hand in his precise script. They were essays and parables; fairy tales and warnings; comprehensive explanation and pedagogical pretext. Most were about mathematics and computer science; others were trip reports that are more revealing about their author than about the people and places visited. It was his habit to copy each paper and circulate it to a small group of colleagues who would copy and forward the papers to another limited group of scientists. His love affair with simplicity came at an early age and under his mother's guidance. He once said he had asked his mother whether trigonometry was a difficult topic. She replied that he must learn all the formulas and that furthermore if he required more than five lines to prove something, he was on the wrong track. Dijkstra was famous for his wit, eloquence, rudeness, abruptness and often cruelty to fellow professionals, and way with words, such as in his remark, "The question of whether Machines Can Think (…) is about as relevant as the question of whether Submarines Can Swim."; his advice to a promising researcher, who asked how to select a topic for research, "Do only what only you can do". Dijkstra was also known for his vocal criticism and absence of social skills when interacting with colleagues. As an outspoken and critical visionary, he strongly opposed the teaching of BASIC. In many of his more humorous essays, Dijkstra described a fictional company of which he served as chairman. The company was called Mathematics, Inc., a company that he imagined having commercialized the production of mathematical theorems in the same way that software companies had commercialized the production of computer programs. He invented a number of activities and challenges of Mathematics Inc. and documented them in several papers in the EWD series. The imaginary company had produced a proof of the Riemann Hypothesis but then had great difficulties collecting royalties from mathematicians who had proved results assuming the Riemann Hypothesis. The proof itself was a trade secret. Many of the company's proofs were rushed out the door and then much of the company's effort had to be spent on maintenance. A more successful effort was the Standard Proof for Pythagoras' Theorem, that replaced the more than 100 incompatible existing proofs. Dijkstra described Mathematics Inc. as "the most exciting and most miserable business ever conceived". EWD 443 (1974) describes his fictional company as having over 75 percent of the world's market share. Dijkstra was well known for his habit of carefully composing manuscripts with his fountain pen. The manuscripts are called EWDs, since Dijkstra numbered them with "EWD", his initials, as a prefix. According to Dijkstra himself, the EWDs started when he moved from the Mathematical Centre in Amsterdam to the Eindhoven University of Technology (then Technische Hogeschool Eindhoven). After going to Eindhoven, Dijkstra experienced a writer's block for more than a year. Dijkstra distributed photocopies of a new EWD among his colleagues. Many recipients photocopied and forwarded their copies, so the EWDs spread throughout the international computer science community. The topics were computer science and mathematics, and included trip reports, letters, and speeches. These short articles span a period of 40 years. Almost all EWDs appearing after 1972 were hand-written. They are rarely longer than 15 pages and are consecutively numbered. The last one, No. 1318, is from 14 April 2002. Within computer science they are known as the EWD reports, or, simply the EWDs. More than 1300 EWDs have been scanned, with a growing number transcribed to facilitate search, and are available online at the Dijkstra archive of the University of Texas. Dijkstra's self-confidence went together with a remarkably modest lifestyle, to the point of being spartan. His and his wife's house in Nuenen was simple, small and unassuming. He did not own a TV, a VCR or a mobile telephone, and did not go to the movies. In contrast, he played the piano well and, while in Austin, liked to go to concerts. An enthusiastic listener of classical music, Dijkstra's favorite composer was Mozart. Dijkstra died on 6 August 2002. According to officials at the University of Texas, the cause of death was cancer. In 1972 the Association for Computing Machinery (ACM) acknowledged Dijkstra's seminal contributions to the field by awarding him the distinguished Turing Award. The citation for the award reads: The introduction given at the awards ceremony is a tribute to Dijkstra: In the words of Sir Tony Hoare, FRS, delivered by him at Dijkstra's funeral: In March 2003, the following email was sent to the distributed computing community: Former ACM President Peter J. Denning wrote about Dijkstra: Among Dijkstra's awards and honors are: The Distinguished Fellowship of the British Computer Society (BCS) is awarded under bylaw 7 of the BCS's Royal Charter. The award was first approved in 1969 and the first election was made in 1971 to Dijkstra. On the occasion of Dijkstra's 60th birthday in 1990, The Department of Computer Science (UTCS) at the University of Texas at Austin organized a two-day seminar in his honor. Speakers came from all over the United States and Europe, and a group of computer scientists contributed research articles which were edited into a book. In 2002, the C&C Foundation of Japan recognized Dijkstra "for his pioneering contributions to the establishment of the scientific basis for computer software through creative research in basic software theory, algorithm theory, structured programming, and semaphores." Dijkstra was alive to receive notice of the award, but it was accepted by his family in an award ceremony after his death. Shortly before his death in 2002, Dijkstra received the ACM PODC Influential-Paper Award in distributed computing for his work on self-stabilization of program computation. This annual award was renamed the Dijkstra Prize (Edsger W. Dijkstra Prize in Distributed Computing) the following year, in his honor. The Dijkstra Award for Outstanding Academic Achievement in Computer Science (Loyola University Chicago, Department of Computer Science) is named for Edsger W. Dijkstra. Beginning in 2005, this award recognizes the top academic performance by a graduating computer science major. Selection is based on GPA in all major courses and election by department faculty. The Department of Computer Science (UTCS) at the University of Texas at Austin hosted the inaugural Edsger W. Dijkstra Memorial Lecture on 12 October 2010. Tony Hoare, Emeritus Professor at Oxford and Principal Researcher at Microsoft Research, was the speaker for the event. This lecture series was made possible by a generous grant from Schlumberger to honor the memory of Dijkstra.
https://en.wikipedia.org/wiki?curid=10018
Educational perennialism Educational perennialism also infrequently referred to as Universal Curriculum is a normative educational philosophy. Perennialists believe that one should teach the things that are of everlasting pertinence to all people everywhere, and that the emphasis should be on principles, not facts. Since people are human, one should teach first about humans, rather than machines or techniques, and about liberal, rather than vocational, topics. Although perennialism may appear similar to essentialism, perennialism focuses first on personal development, while essentialism focuses first on essential skills. Essentialist curricula thus tend to be much more vocational and fact-based, and far less liberal and principle-based. Both philosophies are typically considered to be teacher-centered, as opposed to student-centered philosophies of education such as progressivism. However, since the teachers associated with perennialism are in a sense the authors of the Western masterpieces themselves, these teachers may be open to student criticism through the associated Socratic method, which, if carried out as true dialogue, involves a balance between teacher activity and student activity, with the teacher promoting discussion. The word perennial in secular perennialism suggests something that lasts an indefinitely long time, recurs again and again, or is self-renewing. As promoted primarily by Robert Hutchins and Mortimer Adler, a universal curriculum based upon the common and essential nature of all human beings is recommended. This form of perennialism comprises the humanist and scientific traditions. Hutchins and Adler implemented these ideas with great success at the University of Chicago, where they still strongly influence the curriculum in the form of the undergraduate Common Core. Other notable figures in the movement include Stringfellow Barr and Scott Buchanan (who together initiated the Great Books program at St. John's College in Annapolis, Maryland), Mark Van Doren, Alexander Meiklejohn, and Sir Richard Livingstone, an English classicist with an American following. Secular perennialists espouse the idea that education should focus on the historical development of a continually developing common oriented base of human knowledge and art, the timeless value of classic thought on central human issues by landmark thinkers, and revolutionary ideas critical to historical paradigm shifts or changes in world view. A program of studies which is highly general, nonspecialized, and nonvocational is advocated. They firmly believe that exposure of all citizens to the development of thought by those most responsible for the evolution of the occidental oriented tradition is integral to the survival of the freedoms, human rights and responsibilities inherent to a true Democracy. Adler states: ... our political democracy depends upon the reconstitution of our schools. Our schools are not turning out young people prepared for the high office and the duties of citizenship in a democratic republic. Our political institutions cannot thrive, they may not even survive, if we do not produce a greater number of thinking citizens, from whom some statesmen of the type we had in the 18th century might eventually emerge. We are, indeed, a nation at risk, and nothing but radical reform of our schools can save us from impending disaster... Whatever the price... the price we will pay for not doing it will be much greater. Hutchins writes in the same vein: The business of saying ... that people are not capable of achieving a good education is too strongly reminiscent of the opposition of every extension of democracy. This opposition has always rested on the allegation that the people were incapable of exercising the power they demanded. Always the historic statement has been verified: you cannot expect the slave to show the virtues of the free man unless you first set him free. When the slave has been set free, he has, in the passage of time, become indistinguishable from those who have always been free ... There appears to be an innate human tendency to underestimate the capacity of those who do not belong to "our" group. Those who do not share our background cannot have our ability. Foreigners, people who are in a different economic status, and the young seem invariably to be regarded as intellectually backward ... As with the essentialists, perennialists are educationally conservative in the requirement of a curriculum focused upon fundamental subject areas, but stress that the overall aim should be exposure to history's finest thinkers as models for discovery. The student should be taught such basic subjects as English, languages, history, mathematics, natural science, philosophy, and fine arts. Adler states: "The three R's, which always signified the formal disciplines, are the essence of liberal or general education." Secular perennialists agree with progressivists that memorization of vast amounts of factual information and a focus on second-hand information in textbooks and lectures does not develop rational thought. They advocate learning through the development of meaningful conceptual thinking and judgement by means of a directed reading list of the profound, aesthetic, and meaningful great books of the Western canon. These books, secular perennialists argue, are written by the world's finest thinkers, and cumulatively comprise the "Great Conversation" of humanity with regard to the central human questions. Their basic argument for the use of original works (abridged translations being acceptable as well) is that these are the products of "genius". Hutchins remarks: Great books are great teachers; they are showing us every day what ordinary people are capable of. These books come out of ignorant, inquiring humanity. They are usually the first announcements for success in learning. Most of them were written for, and addressed to, ordinary people. It is important to note that the Great Conversation is not static, which is the impression that one might obtain from some descriptions of perennialism, a confusion with religious perennialism, or even the term perennialism itself. The Great Conversation and the set of related great books changes as the representative thought of man changes or progresses, and is therefore representative of an evolution of thought, but is not based upon the whim or fancy of the latest cultural fads. Hutchins makes this point very clear: In the course of history... new books have been written that have won their place in the list. Books once thought entitled to belong to it have been superseded; and this process of change will continue as long as men can think and write. It is the task of every generation to reassess the tradition in which it lives, to discard what it cannot use, and to bring into context with the distant and intermediate past the most recent contributions to the Great Conversation. ...the West needs to recapture and reemphasize and bring to bear upon its present problems the wisdom that lies in the works of its greatest thinkers and in the name of love Perennialism was a solution proposed in response to what was considered by many to be a failing educational system. Again Hutchins writes: The products of American high schools are illiterate; and a degree from a famous college or university is no guarantee that the graduate is in any better case. One of the most remarkable features of American society is that the difference between the "uneducated" and the "educated" is so slight. In this regard John Dewey and Hutchins were in agreement. Hutchins's book "The Higher Learning in America" deplored the "plight of higher learning" that had turned away from cultivation of the intellect and toward anti-intellectual practicality due in part, to a lust for money. In a highly negative review of the book, Dewey wrote a series of articles in "The Social Frontier" which began by applauding Hutchins' attack on "the aimlessness of our present educational scheme. Perennialists believe that reading is to be supplemented with mutual investigations (between the teacher and the student) and minimally-directed discussions through the Socratic method in order to develop a historically oriented understanding of concepts. They argue that accurate, independent reasoning distinguishes the developed or educated mind and they thus stress the development of this faculty. A skilled teacher would keep discussions on topic and correct errors in reasoning, but it would be the class, not the teacher, who would reach the conclusions. While not directing or leading the class to a conclusion, the teacher may work to accurately formulate problems within the scope of the texts being studied. While the standard argument for utilizing a modern text supports distillation of information into a form relevant to modern society, perennialists argue that many of the historical debates and the development of ideas presented by the great books are relevant to any society, at any time, and thus that the suitability of the great books for instructional use is unaffected by their age. Perennialists freely acknowledge that any particular selection of great books will disagree on many topics; however, they see this as an advantage, rather than a detriment. They believe that the student must learn to recognize such disagreements, which often reflect current debates. The student becomes responsible for thinking about the disagreements and reaching a reasoned, defensible conclusion. This is a major goal of the Socratic discussions. They do not advocate teaching a settled scholarly interpretation of the books, which would cheat the student of the opportunity to learn rational criticism and to know his own mind. Perennialism was originally religious in nature, developed first by Thomas Aquinas in the thirteenth century in his work "De Magistro" ("On the Teacher"). In the nineteenth century, John Henry Newman presented a defense of religious perennialism in "The Idea of a University". Discourse 5 of that work, "Knowledge Its Own End", is a recent statement of a Christian educational perennialism. There are several epistemological options, which affect the pedagogical options. The possibilities may be surveyed by considering four extreme positions, as indicated in the following table:
https://en.wikipedia.org/wiki?curid=10021
MDMA 3,4-Methylenedioxymethamphetamine (MDMA), commonly known as ecstasy (E) or molly, is a psychoactive drug primarily used for recreational purposes. The desired effects include altered sensations, increased energy, empathy, as well as pleasure. When taken by mouth, effects begin in 30 to 45 minutes and last 3 to 6 hours. Adverse effects include addiction, memory problems, paranoia, difficulty sleeping, teeth grinding, blurred vision, sweating and a rapid heartbeat. Deaths have been reported due to increased body temperature and dehydration. Following use people often feel depressed and tired. MDMA acts primarily by increasing the activity of the neurotransmitters serotonin, dopamine and noradrenaline in parts of the brain. It belongs to the substituted amphetamine classes of drugs and has stimulant and hallucinogenic effects. MDMA is illegal in most countries and, as of 2018, has no approved medical uses. Limited exceptions are sometimes made for research. Researchers are investigating whether MDMA may assist in treating severe, treatment-resistant posttraumatic stress disorder (PTSD) with phase 3 clinical trials to look at effectiveness and safety expected to begin in 2018. In 2017, the FDA granted MDMA a breakthrough therapy designation for PTSD, meaning that if studies show promise, a review for potential medical use could occur more quickly. MDMA was first developed in 1912 by Merck. It was used to enhance psychotherapy beginning in the 1970s and became popular as a street drug in the 1980s. MDMA is commonly associated with dance parties, raves, and electronic dance music. It may be mixed with other substances such as ephedrine, amphetamine, and methamphetamine. In 2016, about 21 million people between the ages of 15 and 64 used ecstasy (0.3% of the world population). This was broadly similar to the percentage of people who use cocaine or amphetamines, but lower than for cannabis or opioids. In the United States, as of 2017, about 7% of people have used MDMA at some point in their lives and 0.9% have used it in the last year. In general, MDMA users report feeling the onset of subjective effects within 30 to 60 minutes of oral consumption and reaching peak effect at 75 to 120 minutes, which then plateaus for about 3.5 hours. The desired short-term psychoactive effects of MDMA have been reported to include: The experience elicited by MDMA depends on the dose, setting, and user. The variability of the induced altered state is lower compared to other psychedelics. For example, MDMA used at parties is associated with high motor activity, reduced sense of identity, and poor awareness of surroundings. Use of MDMA individually or in small groups in a quiet environment and when concentrating, is associated with increased lucidity, concentration, sensitivity to aesthetic aspects of the environment, enhanced awareness of emotions, and improved capability of communication. In psychotherapeutic settings, MDMA effects have been characterized by infantile ideas, mood lability, and memories and moods connected with childhood experiences. MDMA has been described as an "empathogenic" drug because of its empathy-producing effects. Results of several studies show the effects of increased empathy with others. When testing MDMA for medium and high doses, it showed increased hedonic and arousal continuum. The effect of MDMA increasing sociability is consistent, while its effects on empathy have been more mixed. MDMA is often considered the drug of choice within the rave culture and is also used at clubs, festivals, and house parties. In the rave environment, the sensory effects of music and lighting are often highly synergistic with the drug. The psychedelic amphetamine quality of MDMA offers multiple appealing aspects to users in the rave setting. Some users enjoy the feeling of mass communion from the inhibition-reducing effects of the drug, while others use it as party fuel because of the drug's stimulatory effects. MDMA is used less often than other stimulants, typically less than once per week. MDMA is sometimes taken in conjunction with other psychoactive drugs such as LSD, psilocybin mushrooms, and ketamine, an act called "candy-flipping". , MDMA has no accepted medical indications. Before it was widely banned, it saw limited use in psychotherapy. A few psychotherapists continue to use MDMA in therapy despite the drug's legal status. Small doses of MDMA are used by some religious practitioners as an entheogen to enhance prayer or meditation. MDMA has been used as an adjunct to New Age spiritual practices. MDMA has become widely known as ecstasy (shortened "E", "X", or "XTC"), usually referring to its tablet form, although this term may also include the presence of possible adulterants or diluents. The UK term "mandy" and the US term "molly" colloquially refer to MDMA in a crystalline powder form that is thought to be free of adulterants. MDMA is also sold in the form of the hydrochloride salt, either as loose crystals or in gelcaps. Partly due to the global supply shortage of sassafras oil—a problem largely assuaged by use of improved or alternative modern methods of synthesis—the purity of substances sold as molly have been found to vary widely. Some of these substances contain methylone, ethylone, MDPV, mephedrone, or any other of the group of compounds commonly known as bath salts, in addition to, or in place of, MDMA. Powdered MDMA ranges from pure MDMA to crushed tablets with 30–40% purity. MDMA tablets typically have low purity due to bulking agents that are added to dilute the drug and increase profits (notably lactose) and binding agents. Tablets sold as ecstasy sometimes contain 3,4-methylenedioxyamphetamine (MDA), 3,4-methylenedioxyethylamphetamine (MDEA), other amphetamine derivatives, caffeine, opiates, or painkillers. Some tablets contain little or no MDMA. The proportion of seized ecstasy tablets with MDMA-like impurities has varied annually and by country. The average content of MDMA in a preparation is 70 to 120 mg with the purity having increased since the 1990s. MDMA is usually consumed by mouth. It is also sometimes snorted. Acute adverse effects are usually the result of high or multiple doses, although single dose toxicity can occur in susceptible individuals. The most serious short-term physical health risks of MDMA are hyperthermia and dehydration. Cases of life-threatening or fatal hyponatremia (excessively low sodium concentration in the blood) have developed in MDMA users attempting to prevent dehydration by consuming excessive amounts of water without replenishing electrolytes. The immediate adverse effects of MDMA use can include: Other adverse effects that may occur or persist for up to a week following cessation of moderate MDMA use include: , the long-term effects of MDMA on human brain structure and function have not been fully determined. However, there is consistent evidence of structural and functional deficits in MDMA users with high lifetime exposure. There is no evidence of structural or functional changes in MDMA users with only a moderate (<50 doses used and <100 tablets consumed) lifetime exposure. Nonetheless, MDMA in moderate use may still be neurotoxic. Furthermore, it is not clear yet whether "typical" users of MDMA (1 to 2 pills of 75 to 125 mg MDMA or analogue every 1 to 4 weeks) will develop neurotoxic brain lesions. Long-term exposure to MDMA in humans has been shown to produce marked neurodegeneration in striatal, hippocampal, prefrontal, and occipital serotonergic axon terminals. Neurotoxic damage to serotonergic axon terminals has been shown to persist for more than two years. Elevations in brain temperature from MDMA use are positively correlated with MDMA-induced neurotoxicity. However, most studies on MDMA and serotonergic neurotoxicity in humans focus on the heaviest users, those who consume more than seven times the average. It is therefore possible that no serotonergic neurotoxicity is present in most casual users. Adverse neuroplastic changes to brain microvasculature and white matter also occur in humans using low doses of MDMA. Reduced gray matter density in certain brain structures has also been noted in human MDMA users. Global reductions in gray matter volume, thinning of the parietal and orbitofrontal cortices, and decreased hippocampal activity have been observed in long term users. The effects established so far for recreational use of ecstasy lie in the range of moderate to severe effects for serotonin transporter reduction. Impairments in multiple aspects of cognition, including attention, learning, memory, visual processing, and sleep have been found in regular MDMA users. The magnitude of these impairments is correlated with lifetime MDMA usage and are partially reversible with abstinence. Several forms of memory are impaired by chronic ecstasy use; however, the effects for memory impairments in ecstasy users are generally small overall. MDMA use is also associated with increased impulsivity and depression. Serotonin depletion following MDMA use can cause depression in subsequent days. In some cases, depressive symptoms persist for longer periods. Some studies indicate repeated recreational use of ecstasy is associated with depression and anxiety, even after quitting the drug. Depression is one of the main reasons for cessation of use. At high doses, MDMA induces a neuroimmune response that, through several mechanisms, increases the permeability of the blood-brain barrier, thereby making the brain more susceptible to environmental toxins and pathogens. In addition, MDMA has immunosuppressive effects in the peripheral nervous system and pro-inflammatory effects in the central nervous system. MDMA is a moderately teratogenic drug (i.e., it is toxic to the fetus). In utero exposure to MDMA is associated with a neuro- and cardiotoxicity and impaired motor functioning. Motor delays may be temporary during infancy or long-term. The severity of these developmental delays increases with heavier MDMA use. Approximately 60% of MDMA users experience withdrawal symptoms when they stop taking MDMA. Some of these symptoms include fatigue, loss of appetite, depression, and trouble concentrating. Tolerance to some of the desired and adverse effects of MDMA is expected to occur with consistent MDMA use. A 2007 analysis estimated MDMA to have a psychological dependence and physical dependence potential roughly three-fourths to four-fifths that of cannabis. MDMA has been shown to induce ΔFosB in the nucleus accumbens. Because MDMA releases dopamine in the striatum, the mechanisms by which it induces ΔFosB in the nucleus accumbens are analogous to other dopaminergic psychostimulants. Therefore, chronic use of MDMA at high doses can result in altered brain structure and drug addiction that occur as a consequence of ΔFosB overexpression in the nucleus accumbens. MDMA is less addictive than other stimulants such as methamphetamine and cocaine. Compared with amphetamine, MDMA and its metabolite MDA are less reinforcing. One study found approximately 15% of chronic MDMA users met the DSM-IV diagnostic criteria for substance dependence. However, there is little evidence for a specific diagnosable MDMA dependence syndrome because MDMA is typically used relatively infrequently. There are currently no medications to treat MDMA addiction. A 2007 UK study ranked MDMA 18th in harmfulness out of 20 recreational drugs. Rankings for each drug were based on the risk for acute physical harm, the propensity for physical and psychological dependency on the drug, and the negative familial and societal impacts of the drug. The authors did not evaluate or rate the negative impact of ecstasy on the cognitive health of ecstasy users (e.g. impaired memory and concentration). MDMA overdose symptoms vary widely due to the involvement of multiple organ systems. Some of the more overt overdose symptoms are listed in the table below. The number of instances of fatal MDMA intoxication is low relative to its usage rates. In most fatalities, MDMA was not the only drug involved. Acute toxicity is mainly caused by serotonin syndrome and sympathomimetic effects. MDMA's toxicity in overdose may be exacerbated by caffeine, with which it is frequently cut in order to increase volume. A scheme for management of acute MDMA toxicity has been published focusing on treatment of hyperthermia, hyponatraemia, serotonin syndrome, and multiple organ failure. A number of drug interactions can occur between MDMA and other drugs, including serotonergic drugs. MDMA also interacts with drugs which inhibit CYP450 enzymes, like ritonavir (Norvir), particularly CYP2D6 inhibitors. Concurrent use of MDMA high dosages with another serotonergic drug can result in a life-threatening condition called serotonin syndrome. Severe overdose resulting in death has also been reported in people who took MDMA in combination with certain monoamine oxidase inhibitors, such as phenelzine (Nardil), tranylcypromine (Parnate), or moclobemide (Aurorix, Manerix). MDMA acts primarily as a presynaptic releasing agent of serotonin, norepinephrine, and dopamine, which arises from its activity at trace amine-associated receptor 1 (TAAR1) and vesicular monoamine transporter 2 (VMAT2). MDMA is also a monoamine transporter substrate (i.e., a substrate for DAT, NET, and SERT), so it enters monoamine neurons via these neuronal membrane transport proteins; by acting as a monoamine transporter substrate, MDMA produces competitive reuptake inhibition at the neuronal membrane transporters (i.e., it competes with endogenous monoamines for reuptake). MDMA inhibits both vesicular monoamine transporters (VMATs), the second of which (VMAT2) is highly expressed within monoamine neurons at vesicular membranes. Once inside a monoamine neuron, MDMA acts as a VMAT2 inhibitor and a TAAR1 agonist. Inhibition of VMAT2 by MDMA results in increased concentrations of the associated neurotransmitter (serotonin, norepinephrine, or dopamine) in the cytosol of a monoamine neuron. Activation of TAAR1 by MDMA triggers protein kinase A and protein kinase C signaling events which then phosphorylates the associated monoamine transporters – DAT, NET, or SERT – of the neuron. In turn, these phosphorylated monoamine transporters either reverse transport direction – i.e., move neurotransmitters from the cytosol to the synaptic cleft – or withdraw into the neuron, respectively producing neurotransmitter efflux and noncompetitive reuptake inhibition at the neuronal membrane transporters. The actions increase the synaptic concentrations of monoamine neurotransmitters. MDMA has ten times more affinity for uptake at serotonin transporters compared to dopamine and norepinephrine transporters and consequently has mainly serotonergic effects. In summary, MDMA enters monoamine neurons by acting as a monoamine transporter substrate. MDMA activity at VMAT2 moves neurotransmitters out from synaptic vesicles and into the cytosol; MDMA activity at TAAR1 moves neurotransmitters out of the cytosol and into the synaptic cleft. MDMA also has weak agonist activity at postsynaptic serotonin receptors 5-HT1 and 5-HT2 receptors, and its more efficacious metabolite MDA likely augments this action. Cortisol, prolactin, and oxytocin quantities in serum are increased by MDMA. MDMA is a ligand at both sigma receptor subtypes, though its efficacies at the receptors have not yet been elucidated. The MDMA concentration in the blood stream starts to rise after about 30 minutes, and reaches its maximal concentration in the blood stream between 1.5 and 3 hours after ingestion. It is then slowly metabolized and excreted, with levels of MDMA and its metabolites decreasing to half their peak concentration over the next several hours. The duration of action of MDMA is usually four to six hours, after which serotonin levels in the brain are depleted. Serotonin levels typically return to normal within 24–48 hours. Metabolites of MDMA that have been identified in humans include 3,4-methylenedioxyamphetamine (MDA), 4-hydroxy-3-methoxymethamphetamine (HMMA), 4-hydroxy-3-methoxyamphetamine (HMA), 3,4-dihydroxyamphetamine (DHA) (also called alpha-methyldopamine (α-Me-DA)), 3,4-methylenedioxyphenylacetone (MDP2P), and 3,4-Methylenedioxy-N-hydroxyamphetamine (MDOH). The contributions of these metabolites to the psychoactive and toxic effects of MDMA are an area of active research. 80% of MDMA is metabolised in the liver, and about 20% is excreted unchanged in the urine. MDMA is known to be metabolized by two main metabolic pathways: (1) "O"-demethylenation followed by catechol-"O"-methyltransferase (COMT)-catalyzed methylation and/or glucuronide/sulfate conjugation; and (2) "N"-dealkylation, deamination, and oxidation to the corresponding benzoic acid derivatives conjugated with glycine. The metabolism may be primarily by cytochrome P450 (CYP450) enzymes CYP2D6 and CYP3A4 and COMT. Complex, nonlinear pharmacokinetics arise via autoinhibition of CYP2D6 and CYP2D8, resulting in zeroth order kinetics at higher doses. It is thought that this can result in sustained and higher concentrations of MDMA if the user takes consecutive doses of the drug. MDMA and metabolites are primarily excreted as conjugates, such as sulfates and glucuronides. MDMA is a chiral compound and has been almost exclusively administered as a racemate. However, the two enantiomers have been shown to exhibit different kinetics. The disposition of MDMA may also be stereoselective, with the S-enantiomer having a shorter elimination half-life and greater excretion than the R-enantiomer. Evidence suggests that the area under the blood plasma concentration versus time curve (AUC) was two to four times higher for the ("R")-enantiomer than the ("S")-enantiomer after a 40 mg oral dose in human volunteers. Likewise, the plasma half-life of ("R")-MDMA was significantly longer than that of the ("S")-enantiomer (5.8 ± 2.2 hours vs 3.6 ± 0.9 hours). However, because MDMA excretion and metabolism have nonlinear kinetics, the half-lives would be higher at more typical doses (100 mg is sometimes considered a typical dose). MDMA is in the substituted methylenedioxyphenethylamine and substituted amphetamine classes of chemicals. As a free base, MDMA is a colorless oil insoluble in water. The most common salt of MDMA is the hydrochloride salt; pure MDMA hydrochloride is water-soluble and appears as a white or off-white powder or crystal. There are numerous methods available to synthesize MDMA via different intermediates. The original MDMA synthesis described in Merck's patent involves brominating safrole to 1-(3,4-methylenedioxyphenyl)-2-bromopropane and then reacting this adduct with methylamine. Most illicit MDMA is synthesized using MDP2P (3,4-methylenedioxyphenyl-2-propanone) as a precursor. MDP2P in turn is generally synthesized from piperonal, safrole or isosafrole. One method is to isomerize safrole to isosafrole in the presence of a strong base, and then oxidize isosafrole to MDP2P. Another method uses the Wacker process to oxidize safrole directly to the MDP2P intermediate with a palladium catalyst. Once the MDP2P intermediate has been prepared, a reductive amination leads to racemic MDMA (an equal parts mixture of ("R")-MDMA and ("S")-MDMA). Relatively small quantities of essential oil are required to make large amounts of MDMA. The essential oil of "Ocotea cymbarum", for example, typically contains between 80 and 94% safrole. This allows 500 ml of the oil to produce between 150 and 340 grams of MDMA. MDMA and MDA may be quantitated in blood, plasma or urine to monitor for use, confirm a diagnosis of poisoning or assist in the forensic investigation of a traffic or other criminal violation or a sudden death. Some drug abuse screening programs rely on hair, saliva, or sweat as specimens. Most commercial amphetamine immunoassay screening tests cross-react significantly with MDMA or its major metabolites, but chromatographic techniques can easily distinguish and separately measure each of these substances. The concentrations of MDA in the blood or urine of a person who has taken only MDMA are, in general, less than 10% those of the parent drug. MDMA was first synthesized in 1912 by Merck chemist Anton Köllisch. At the time, Merck was interested in developing substances that stopped abnormal bleeding. Merck wanted to avoid an existing patent held by Bayer for one such compound: hydrastinine. Köllisch developed a preparation of a hydrastinine analogue, methylhydrastinine, at the request of fellow lab members, Walther Beckh and Otto Wolfes. MDMA (called methylsafrylamin, safrylmethylamin or N-Methyl-a-Methylhomopiperonylamin in Merck laboratory reports) was an intermediate compound in the synthesis of methylhydrastinine. Merck was not interested in MDMA itself at the time. On 24 December 1912, Merck filed two patent applications that described the synthesis and some chemical properties of MDMA and its subsequent conversion to methylhydrastinine. Merck records indicate its researchers returned to the compound sporadically. A 1920 Merck patent describes a chemical modification to MDMA. In 1927, Max Oberlin studied the pharmacology of MDMA while searching for substances with effects similar to adrenaline or ephedrine, the latter being structurally similar to MDMA. Compared to ephedrine, Oberlin observed that it had similar effects on vascular smooth muscle tissue, stronger effects at the uterus, and no "local effect at the eye". MDMA was also found to have effects on blood sugar levels comparable to high doses of ephedrine. Oberlin concluded that the effects of MDMA were not limited to the sympathetic nervous system. Research was stopped "particularly due to a strong price increase of safrylmethylamine", which was still used as an intermediate in methylhydrastinine synthesis. Albert van Schoor performed simple toxicological tests with the drug in 1952, most likely while researching new stimulants or circulatory medications. After pharmacological studies, research on MDMA was not continued. In 1959, Wolfgang Fruhstorfer synthesized MDMA for pharmacological testing while researching stimulants. It is unclear if Fruhstorfer investigated the effects of MDMA in humans. Outside of Merck, other researchers began to investigate MDMA. In 1953 and 1954, the United States Army commissioned a study of toxicity and behavioral effects in animals injected with mescaline and several analogues, including MDMA. Conducted at the University of Michigan in Ann Arbor, these investigations were declassified in October 1969 and published in 1973. A 1960 Polish paper by Biniecki and Krajewski describing the synthesis of MDMA as an intermediate was the first published scientific paper on the substance. MDMA may have been in non-medical use in the western United States in 1968. An August 1970 report at a meeting of crime laboratory chemists indicates MDMA was being used recreationally in the Chicago area by 1970. MDMA likely emerged as a substitute for its analog methylenedioxyamphetamine (MDA), a drug at the time popular among users of psychedelics which was made a Schedule 1 substance in the United States in 1970. American chemist and psychopharmacologist Alexander Shulgin reported he synthesized MDMA in 1965 while researching methylenedioxy compounds at Dow Chemical Company, but did not test the psychoactivity of the compound at this time. Around 1970, Shulgin sent instructions for N-methylated MDA (MDMA) synthesis to the founder of a Los Angeles chemical company who had requested them. This individual later provided these instructions to a client in the Midwest. Shulgin may have suspected he played a role in the emergence of MDMA in Chicago. Shulgin first heard of the psychoactive effects of N-methylated MDA around 1975 from a young student who reported "amphetamine-like content". Around 30 May 1976, Shulgin again heard about the effects of N-methylated MDA, this time from a graduate student in a medicinal chemistry group he advised at San Francisco State University who directed him to the University of Michigan study. She and two close friends had consumed 100 mg of MDMA and reported positive emotional experiences. Following the self-trials of a colleague at the University of San Francisco, Shulgin synthesized MDMA and tried it himself in September and October 1976. Shulgin first reported on MDMA in a presentation at a conference in Bethesda, Maryland in December 1976. In 1978, he and David E. Nichols published a report on the drug's psychoactive effect in humans. They described MDMA as inducing "an easily controlled altered state of consciousness with emotional and sensual overtones" comparable "to marijuana, to psilocybin devoid of the hallucinatory component, or to low levels of MDA". While not finding his own experiences with MDMA particularly powerful, Shulgin was impressed with the drug's disinhibiting effects and thought it could be useful in therapy. Believing MDMA allowed users to strip away habits and perceive the world clearly, Shulgin called the drug "window". Shulgin occasionally used MDMA for relaxation, referring to it as "my low-calorie martini", and gave the drug to friends, researchers, and others who he thought could benefit from it. One such person was Leo Zeff, a psychotherapist who had been known to use psychedelic substances in his practice. When he tried the drug in 1977, Zeff was impressed with the effects of MDMA and came out of his semi-retirement to promote its use in therapy. Over the following years, Zeff traveled around the United States and occasionally to Europe, eventually training an estimated four thousand psychotherapists in the therapeutic use of MDMA. Zeff named the drug "Adam", believing it put users in a state of primordial innocence. Psychotherapists who used MDMA believed the drug eliminated the typical fear response and increased communication. Sessions were usually held in the home of the patient or the therapist. The role of the therapist was minimized in favor of patient self-discovery accompanied by MDMA induced feelings of empathy. Depression, substance abuse, relationship problems, premenstrual syndrome, and autism were among several psychiatric disorders MDMA assisted therapy was reported to treat. According to psychiatrist George Greer, therapists who used MDMA in their practice were impressed by the results. Anecdotally, MDMA was said to greatly accelerate therapy. According to David Nutt, MDMA was widely used in the western US in couples counseling, and was called "empathy". Only later was the term "ecstasy" used for it, coinciding with rising opposition to its use. In the late 1970s and early 1980s, "Adam" spread through personal networks of psychotherapists, psychiatrists, users of psychedelics, and yuppies. Hoping MDMA could avoid criminalization like LSD and mescaline, psychotherapists and experimenters attempted to limit the spread of MDMA and information about it while conducting informal research. Early MDMA distributors were deterred from large scale operations by the threat of possible legislation. Between the 1970s and the mid-1980s, this network of MDMA users consumed an estimated 500,000 doses. A small recreational market for MDMA developed by the late 1970s, consuming perhaps 10,000 doses in 1976. By the early 1980s MDMA was being used in Boston and New York City nightclubs such as Studio 54 and Paradise Garage. Into the early 1980s, as the recreational market slowly expanded, production of MDMA was dominated by a small group of therapeutically minded Boston chemists. Having commenced production in 1976, this "Boston Group" did not keep up with growing demand and shortages frequently occurred. Perceiving a business opportunity, Michael Clegg, the Southwest distributor for the Boston Group, started his own "Texas Group" backed financially by Texas friends. In 1981, Clegg had coined "Ecstasy" as a slang term for MDMA to increase its marketability. Starting in 1983, the Texas Group mass-produced MDMA in a Texas lab or imported it from California and marketed tablets using pyramid sales structures and toll-free numbers. MDMA could be purchased via credit card and taxes were paid on sales. Under the brand name "Sassyfras", MDMA tablets were sold in brown bottles. The Texas Group advertised "Ecstasy parties" at bars and discos, describing MDMA as a "fun drug" and "good to dance to". MDMA was openly distributed in Austin and Dallas–Fort Worth area bars and nightclubs, becoming popular with yuppies, college students, and gays. Recreational use also increased after several cocaine dealers switched to distributing MDMA following experiences with the drug. A California laboratory that analyzed confidentially submitted drug samples first detected MDMA in 1975. Over the following years the number of MDMA samples increased, eventually exceeding the number of MDA samples in the early 1980s. By the mid-1980s, MDMA use had spread to colleges around the United States. In an early media report on MDMA published in 1982, a Drug Enforcement Administration (DEA) spokesman stated the agency would ban the drug if enough evidence for abuse could be found. By mid-1984, MDMA use was becoming more noticed. Bill Mandel reported on "Adam" in a 10 June San Francisco Chronicle article, but misidentified the drug as methyloxymethylenedioxyamphetamine (MMDA). In the next month, the World Health Organization identified MDMA as the only substance out of twenty phenethylamines to be seized a significant number of times. After a year of planning and data collection, MDMA was proposed for scheduling by the DEA on 27 July 1984 with a request for comments and objections. The DEA was surprised when a number of psychiatrists, psychotherapists, and researchers objected to the proposed scheduling and requested a hearing. In a Newsweek article published the next year, a DEA pharmacologist stated that the agency had been unaware of its use among psychiatrists. An initial hearing was held on 1 February 1985 at the DEA offices in Washington, D.C. with administrative law judge Francis L. Young presiding. It was decided there to hold three more hearings that year: Los Angeles on 10 June, Kansas City, Missouri on 10–11 July, and Washington, D.C. on 8–11 October. Sensational media attention was given to the proposed criminalization and the reaction of MDMA proponents, effectively advertising the drug. In response to the proposed scheduling, the Texas Group increased production from 1985 estimates of 30,000 tablets a month to as many as 8,000 per day, potentially making two million ecstasy tablets in the months before MDMA was made illegal. By some estimates the Texas Group distributed 500,000 tablets per month in Dallas alone. According to one participant in an ethnographic study, the Texas Group produced more MDMA in eighteen months than all other distribution networks combined across their entire histories. By May 1985, MDMA use was widespread in California, Texas, southern Florida, and the northeastern United States. According to the DEA there was evidence of use in twenty-eight states and Canada. Urged by Senator Lloyd Bentsen, the DEA announced an emergency Schedule I classification of MDMA on 31 May 1985. The agency cited increased distribution in Texas, escalating street use, and new evidence of MDA (an analog of MDMA) neurotoxicity as reasons for the emergency measure. The ban took effect one month later on 1 July 1985 in the midst of Nancy Reagan's "Just Say No" campaign. As a result of several expert witnesses testifying that MDMA had an accepted medical usage, the administrative law judge presiding over the hearings recommended that MDMA be classified as a Schedule III substance. Despite this, DEA administrator John C. Lawn overruled and classified the drug as Schedule I. Later Harvard psychiatrist Lester Grinspoon sued the DEA, claiming that the DEA had ignored the medical uses of MDMA, and the federal court sided with Grinspoon, calling Lawn's argument "strained" and "unpersuasive", and vacated MDMA's Schedule I status. Despite this, less than a month later Lawn reviewed the evidence and reclassified MDMA as Schedule I again, claiming that the expert testimony of several psychiatrists claiming over 200 cases where MDMA had been used in a therapeutic context with positive results could be dismissed because they weren't published in medical journals. No double blind studies had yet been conducted as to the efficacy of MDMA for therapy. In 2017 the FDA granted breakthrough therapy designation for its use with psychotherapy for PTSD. While engaged in scheduling debates in the United States, the DEA also pushed for international scheduling. In 1985 the World Health Organization's Expert Committee on Drug Dependence recommended that MDMA be placed in Schedule I of the 1971 United Nations Convention on Psychotropic Substances. The committee made this recommendation on the basis of the pharmacological similarity of MDMA to previously scheduled drugs, reports of illicit trafficking in Canada, drug seizures in the United States, and lack of well-defined therapeutic use. While intrigued by reports of psychotherapeutic uses for the drug, the committee viewed the studies as lacking appropriate methodological design and encouraged further research. Committee chairman Paul Grof dissented, believing international control was not warranted at the time and a recommendation should await further therapeutic data. The Commission on Narcotic Drugs added MDMA to Schedule I of the convention on 11 February 1986. The use of MDMA in Texas clubs declined rapidly after criminalization, although by 1991 the drug remained popular among young middle-class whites and in nightclubs. In 1985, MDMA use became associated with Acid House on the Spanish island of Ibiza. Thereafter in the late 1980s, the drug spread alongside rave culture to the UK and then to other European and American cities. Illicit MDMA use became increasingly widespread among young adults in universities and later, in high schools. Since the mid-1990s, MDMA has become the most widely used amphetamine-type drug by college students and teenagers. MDMA became one of the four most widely used illicit drugs in the US, along with cocaine, heroin, and cannabis. According to some estimates as of 2004, only marijuana attracts more first time users in the US. After MDMA was criminalized, most medical use stopped, although some therapists continued to prescribe the drug illegally. Later, Charles Grob initiated an ascending-dose safety study in healthy volunteers. Subsequent legally-approved MDMA studies in humans have taken place in the US. in Detroit (Wayne State University), Chicago (University of Chicago), San Francisco (UCSF and California Pacific Medical Center), Baltimore (NIDA–NIH Intramural Program), and South Carolina, as well as in Switzerland (University Hospital of Psychiatry, Zürich), the Netherlands (Maastricht University), and Spain (Universitat Autònoma de Barcelona). "Molly", short for 'molecule', was recognized as a slang term for crystalline or powder MDMA in the 2000s. In 2010, the BBC reported that use of MDMA had decreased in the UK in previous years. This may be due to increased seizures during use and decreased production of the precursor chemicals used to manufacture MDMA. Unwitting substitution with other drugs, such as mephedrone and methamphetamine, as well as legal alternatives to MDMA, such as BZP, MDPV, and methylone, are also thought to have contributed to its decrease in popularity. In 2017 it was found that some pills being sold as MDMA contained pentylone, which can cause very unpleasant agitation and paranoia. According to David Nutt, when safrole was restricted by the United Nations in order to reduce the supply of MDMA, producers in China began using anethole instead, but this gives para-methoxyamphetamine (PMA, also known as "Dr Death"), which is much more toxic than MDMA and can cause overheating, muscle spasms, seizures, unconsciousness, and death. People wanting MDMA are sometimes sold PMA instead. MDMA is legally controlled in most of the world under the UN Convention on Psychotropic Substances and other international agreements, although exceptions exist for research and limited medical use. In general, the unlicensed use, sale or manufacture of MDMA are all criminal offences. In Australia, MDMA was declared an illegal substance in 1986 because of its harmful effects and potential for abuse. It is classed as a Schedule 9 Prohibited Substance in the country, meaning it is available for scientific research purposes only. Any other type of sale, use or manufacture is strictly prohibited by law. Permits for research uses on humans must be approved by a recognized ethics committee on human research. In Western Australia under the Misuse of Drugs Act 1981 4.0g of MDMA is the amount required determining a court of trial, 2.0g is considered a presumption with intent to sell or supply and 28.0g is considered trafficking under Australian law. In the United Kingdom, MDMA was made illegal in 1977 by a modification order to the existing Misuse of Drugs Act 1971. Although MDMA was not named explicitly in this legislation, the order extended the definition of Class A drugs to include various ring-substituted phenethylamines. The drug is therefore illegal to sell, buy, or possess without a licence in the UK. Penalties include a maximum of seven years and/or unlimited fine for possession; life and/or unlimited fine for production or trafficking. Some researchers such as David Nutt have criticized the current scheduling of MDMA, which he determines to be a relatively harmless drug. An editorial he wrote in the Journal of Psychopharmacology, where he compared the risk of harm for horse riding (1 adverse event in 350) to that of ecstasy (1 in 10,000) resulted in his dismissal as well as the resignation of his colleagues from the ACMD. In the United States, MDMA is currently placed in Schedule I of the Controlled Substances Act. In a 2011 federal court hearing, the American Civil Liberties Union successfully argued that the sentencing guideline for MDMA/ecstasy is based on outdated science, leading to excessive prison sentences. Other courts have upheld the sentencing guidelines. The United States District Court for the Eastern District of Tennessee explained its ruling by noting that "an individual federal district court judge simply cannot marshal resources akin to those available to the Commission for tackling the manifold issues involved with determining a proper drug equivalency." In the Netherlands, the Expert Committee on the List (Expertcommissie Lijstensystematiek Opiumwet) issued a report in June 2011 which discussed the evidence for harm and the legal status of MDMA, arguing in favor of maintaining it on List I. In Canada, MDMA is listed as a Schedule 1 as it is an analogue of amphetamine. The CDSA was updated as a result of the Safe Streets and Communities Act changing amphetamines from Schedule III to Schedule I in March 2012. In 2014, 3.5% of 18 to 25 year-olds had used MDMA in the United States. In the European Union as of 2018, 4.1% of adults (15–64 years old) have used MDMA at least once in their life, and 0.8% had used it in the last year. Among young adults, 1.8% had used MDMA in the last 1.8% year. In Europe, an estimated 37% of regular club-goers aged 14 to 35 used MDMA in the past year according to the 2015 European Drug report. The highest one-year prevalence of MDMA use in Germany in 2012 was 1.7% among people aged 25 to 29 compared with a population average of 0.4%. Among adolescent users in the United States between 1999 and 2008, girls were more likely to use MDMA than boys. In 2008 the European Monitoring Centre for Drugs and Drug Addiction noted that although there were some reports of tablets being sold for as little as €1, most countries in Europe then reported typical retail prices in the range of €3 to €9 per tablet, typically containing 25–65 mg of MDMA. By 2014 the EMCDDA reported that the range was more usually between €5 and €10 per tablet, typically containing 57–102 mg of MDMA, although MDMA in powder form was becoming more common. The United Nations Office on Drugs and Crime stated in its 2014 World Drug Report that US ecstasy retail prices range from US$1 to $70 per pill, or from $15,000 to $32,000 per kilogram. A new research area named Drug Intelligence aims to automatically monitor distribution networks based on image processing and machine learning techniques, in which an Ecstasy pill picture is analyzed to detect correlations among different production batches. These novel techniques allow police scientists to facilitate the monitoring of illicit distribution networks. , most of the MDMA in the United States is produced in British Columbia, Canada and imported by Canada-based Asian transnational criminal organizations. The market for MDMA in the United States is relatively small compared to methamphetamine, cocaine, and heroin. In the United States, about 0.9 million people used ecstasy in 2010. MDMA is particularly expensive in Australia, costing A$15–A$30 per tablet. In terms of purity data for Australian MDMA, the average is around 34%, ranging from less than 1% to about 85%. The majority of tablets contain 70–85 mg of MDMA. Most MDMA enters Australia from the Netherlands, the UK, Asia, and the US. A number of ecstasy manufacturers brand their pills with a logo, often being the logo of an unrelated corporation. Some pills depict logos of products or shows popular with children, such as Shaun the Sheep. The Multidisciplinary Association for Psychedelic Studies (MAPS) is funding pilot studies and clinical trials investigating the use of MDMA in psychotherapy to treat posttraumatic stress disorder (PTSD), social anxiety in autistic adults, and anxiety in terminal illness. In November 2016, the United States Food and Drug Administration (FDA) approved large-scale phase 3 clinical trials involving the use of MDMA for the treatment of PTSD in individuals who do not respond to traditional prescription drugs or psychotherapy. MDMA has also been proposed as an adjunct to substance abuse treatment. In 2017, doctors in the UK began the first clinical study of MDMA in alcohol addiction. The potential for MDMA to be used as a rapid-acting antidepressant has been studied in clinical trials, but as of 2017 the evidence on efficacy and safety were insufficient to reach a conclusion. A 2014 review of the safety and efficacy of MDMA as a treatment for various disorders, particularly PTSD, indicated that MDMA has therapeutic efficacy in some patients; however, it emphasized that issues regarding the controlability of MDMA-induced experiences and neurochemical recovery must be addressed. The author noted that oxytocin and -cycloserine are potentially safer co-drugs in PTSD treatment, albeit with limited evidence of efficacy. This review and a second corroborating review by a different author both concluded that, because of MDMA's demonstrated potential to cause lasting harm in humans (e.g., serotonergic neurotoxicity and persistent memory impairment), "considerably more research must be performed" on its efficacy in PTSD treatment to determine if the potential treatment benefits outweigh its potential to harm to a patient.
https://en.wikipedia.org/wiki?curid=10024
Flag of Europe The Flag of Europe or European Flag is an official symbol used by two separate organisations, the Council of Europe (CoE) as a symbol representing Europe, and the European Union (EU). It consists of a circle of twelve five-pointed stars on a blue field. The design was conceived in 1955, and officially adopted later that year by the Council of Europe as a symbol for the whole of Europe. The Council of Europe urged it to be adopted by other European organisations, and in 1985 the European Communities (EC) adopted it. The EU also inherited the emblem's use when it was formed in 1993, being the successor organisation to the EC starting from 1 December 2009 (date of entry into force of the Lisbon Treaty). It has been in wide official use by the EC since the 1990s, but it has never been given official status in any of the EU's treaties. Its adoption as an official symbol of the EU was planned as part of the proposed 2004 European Constitution, which failed to be ratified in 2005. The flag is used by different European organisations as well as by unified European sporting teams under the name of Team Europe. The blazon given by the EU in 1996 describe the design as: "On an azure field a circle of twelve golden mullets, their points not touching." The flag used is the Flag of Europe, which consists of a circle of 12 golden stars on a blue background. Originally designed in 1955 for the Council of Europe, the flag was adopted by the European Communities, the predecessors of the present Union, in 1986. The Council of Europe gave the flag a symbolic description in the following terms, though the official symbolic description adopted by the EU omits the reference to the "Western world": Other symbolic interpretations have been offered based on the account of its design by Paul M. Levy. The five-pointed star is used on many national flags and represents aspiration and education. Their golden colour is that of the sun, which is said to symbolise glory and enlightenment. Their arrangement in a circle represents the constellation of Corona Borealis and can be seen as a crown and the stability of government. The blue background resembles the sky and symbolises truth and the intellect. It is also the colour traditionally used to represent the Virgin Mary. In many paintings of the Virgin Mary as Stella Maris she is crowned with a circle of twelve stars. In 1987, following the adoption of the flag by the EC, Arsène Heitz (1908–1989), one of the designers who had submitted proposals for the flag's design, suggested a religious inspiration for it. He claimed that the circle of stars was based on the iconographic tradition of showing the Blessed Virgin Mary as the Woman of the Apocalypse, wearing a "crown of twelve stars". The French satirical magazine "Le Canard enchaîné" reacted to Heitz's statement with an article entitled "L’Europe violée par la Sainte Vierge" ("Europe Raped by the Blessed Virgin") in the edition of 20 December 1989. Heitz also made a connection to the date of the flag's adoption, 8 December 1955, coinciding with the Catholic Feast of the Immaculate Conception of the Blessed Virgin Mary. Paul M. G. Lévy, then Director of Information at the Council of Europe responsible for designing the flag, in a 1989 statement maintained that he had not been aware of any religious connotations. In an interview given 26 February 1998, Lévy denied not only awareness of the "Marian" connection, but also denied that the final design of a circle of twelve stars was Heitz's. To the question "Who really designed the flag?" Lévy replied: Carlo Curti Gialdino (2005) has reconstructed the design process to the effect that Heitz's proposal contained varying numbers of stars, from which the version with twelve stars was chosen by the Committee of Ministers meeting at Deputy level in January 1955 as one out of two remaining candidate designs. Lévy's 1998 interview apparently gave rise to a new variant of the "Marian" anecdote. An article published in "" in August 1998 alleged that it was Lévy himself who was inspired to introduce a "Marian" element as he walked past a statue of the Blessed Virgin Mary. An article posted in "La Raison" in February 2000 further connected the donation of a stained glass window for Strasbourg Cathedral by the Council of Europe on 21 October 1956. This window, a work by Parisian master Max Ingrand, shows a blessing Madonna underneath a circle of 12 stars on dark blue ground. The overall design of the Madonna is inspired by the banner of the cathedral's "Congrégation Mariale des Hommes", and the twelve stars are found on the statue venerated by this congregation inside the cathedral (twelve is also the number of members of the congregation's council). The Regional Office for Cultural Affairs describe this stained glass window called "Le vitrail de l'Europe de Max Ingrand" (The Glass Window of Europe of Max Ingrand). According to graphical specifications published online by the Council of Europe in 2004, the flag is rectangular with 2:3 proportions: its fly (width) is one and a half times the length of its hoist (height). Twelve yellow stars are centred in a circle (the radius of which is a third of the length of the hoist) upon a blue background. All the stars are upright (one point straight up), have five points and are spaced equally, like the hour positions on the face of a clock. The diameter of each star is equal to one-ninth of the height of the hoist. The colours are regulated in the 1996 guide by the EC, and equivalently in the 2004 guide by the Council of Europe. The base colour of the flag is defined as Pantone "Reflex Blue", while the golden stars are portrayed in Pantone "Yellow": The 2013 logo of the Council of Europe has the colours: The twelve-star "flag of Europe" was designed in 1950 and officially adopted by the Council of Europe in 1955. The same flag was adopted by the European Parliament in 1983. The European Council adopted it was an "emblem" for the European Communities in 1985. Its status in the European Communities was inherited by the European Union upon its formation in 1993. The proposal to adopt it as official flag of the European Union failed with the ratification of the European Constitution in 2005, and mention of all emblems suggesting statehood was removed from the Treaty of Lisbon of 2007, although sixteen member states signed a declaration supporting the continued use of the flag. In 2007, the European Parliament officially adopted the flag for its own use. The Council of Europe in 1950 appointed a committee to study the question of adopting a symbol. Numerous proposals were looked into. Among the unsuccessful proposals was the flag of Richard von Coudenhove-Kalergi's International Paneuropean Union, which he had himself recently adopted for the European Parliamentary Union. The design was a blue field with a red cross inside an orange circle at the centre. Kalergi was very committed to defending the cross as "the great symbol of Europe's moral unity", the Red Cross in particular being "recognized by the whole world, by Christian and non-Christian nations[,] as a symbol of international charity and of the brotherhood of man", but the proposal was rejected by Turkey (a member of the Council of Europe since 1949) on grounds of its religious associations in spite of Kalergi's suggestion of adding a crescent alongside the cross to overcome the Muslim objections. Other proposals included the flag was the European Movement, which had a large green E on a white background, a design was based on the Olympic rings, eight silver rings on a blue background, rejected due to the rings' similarity with "dial", "chain" and "zeros", or a large yellow star on a blue background, rejected due to its similarity with the so-called Burnet flag and the flag of the Belgian Congo. The Consultative Assembly narrowed their choice to two designs. One was by Salvador de Madariaga, the founder of the College of Europe, who suggested a constellation of stars on a blue background (positioned according to capital cities, with a large star for Strasbourg, the seat of the Council). He had circulated his flag round many European capitals and the concept had found favour. The second was a variant by Arsène Heitz, who worked for the Council's postal service and had submitted dozens of designs; the design of his that was accepted by the Assembly was similar to Salvador de Madariaga's, but rather than a constellation, the stars were arranged in a circle. In 1987, Heitz would claim that his inspiration had been the crown of twelve stars of the Woman of the Apocalypse, often found in Marian iconography (see below). On 25 September 1953, the Consultative Assembly of the Council of Europe recommended that a blue flag with fifteen gold stars be adopted as an emblem for the organisation, the number fifteen reflecting the number of states of the Council of Europe. West Germany objected to the fifteen-star design, as one of the members was Saar Protectorate, and to have its own star would imply sovereignty for the region. The Committee of Ministers (the Council's main decision making body) agreed with the Assembly that the flag should be a circle of stars, but opted for a fixed number of twelve stars, "representing perfection and entirety". The Parliamentary Assembly of the Council of Europe on 25 October 1955 agreed to this. Paul M. G. Lévy drew up the exact design of the new flag. Officially adopted on 8 December 1955, the flag was unveiled at the Château de la Muette in Paris on 13 December 1955. Following Expo 58 in Brussels, the flag caught on and the Council of Europe lobbied for other European organisations to adopt the flag as a sign of European unity. The European Parliament took the initiative in seeking a flag to be adopted by the European Communities. Shortly after the first direct elections in 1979 a draft resolution was put forward on the issue. The resolution proposed that the Communities' flag should be that of the Council of Europe and it was adopted by the Parliament on 11 April 1983. The June 1984 European Council (the Communities' leaders) summit in Fontainebleau stressed the importance of promoting a European image and identity to citizens and the world. The European Council appointed an "ad hoc" committee, named "Committee for 'a People's Europe'" (Adonnino Committee). This committee submitted a substantial report, including wide-ranging suggestions, from organising a "European lottery" to campaigning for the introduction of local voting rights for foreign nationals throughout Europe. Under the header of "strengthening of the Community's image and identity", the Committee suggested the introduction of "a flag and an emblem", recommending a design based on the Council of Europe flag, but with the addition of "a gold letter E" in the center of the circle of stars. The European Council held in Milan on 28/29 June 1985 largely followed the recommendations of the Adonnino Committee. But as the adoption of a flag was strongly reminiscent of a national flag representing statehood and was extremely controversial with some member states (in particular the United Kingdom), the Council of Europe's "flag of Europe" design was adopted only with the official status of a "logo". This compromise was widely disregarded from the beginning, and the "European logo", in spite of the explicit language of giving it the status of a "logo", was referred to as the "Community flag" or even "European flag" from the outset. The Communities began to use the "emblem" as its "de facto" flag from 1986, with it being raised outside the Berlaymont building (the seat of the European Commission) for the first time on 29 May 1986. The European Union, which was established by the Maastricht Treaty in 1992 to replace the European Communities and encompass its functions, has retained "de facto" use of the "Community logo" of the EC. Technically and officially, the "European flag" as used by the European Union remains not a "flag" but "a Community 'logo' — or 'emblem' — [...] eligible to be reproduced on rectangular pieces of fabric". In 1997, the "Central and Eastern Eurobarometer" poll included a section intending to "discover the level of public awareness of the European Union" in what were then candidate countries in Central and Eastern Europe. Interviewees were shown "a sticker of the European flag" and asked to identify it. Responses considered correct were: the European Union, the European Community, the Common Market, and "Europe in general". 52% of those interviewed gave one of the correct answers, 15% gave a wrong answer (naming another institution, such as NATO or the United Nations), and 35% could or would not identify it. In 2002, Dutch architect Rem Koolhaas designed a new flag, dubbed the "barcode", as it displayed the colours of the national flags of the EU member states in vertical stripes. It was never officially adopted by the EU or any organisation, but it was used as the logo of the Austrian EU Presidency in 2006. The official status of the emblem as the flag of the European Union was to be formalised as part of the Constitution of the European Union. However, as the proposed constitution failed ratification, the mention of all state-like emblems, including the flag, were removed from the replacement Treaty of Lisbon of 2007. Instead, a separate declaration by sixteen Member States was included in the final act of the Treaty of Lisbon stating that the flag, the anthem, the motto and the currency and Europe Day "will for them continue as symbols to express the sense of community of the people in the European Union and their allegiance to it." In reaction to the removal of the flag from the treaty, the European Parliament, which had supported the inclusion of such symbols, backed a proposal to use these symbols "more often" on behalf of the Parliament itself; Jo Leinen, MEP for Germany, suggested that the Parliament should take "an "avant-garde" role" in their use. In September 2008, the Parliament's Committee on Constitutional Affairs proposed a formal change in the institution's rules of procedure to make "better use of the symbols". Specifically, the flag would be present in all meeting rooms (not just the hemicycle) and at all official events. The proposal was passed on 8 October 2008 by 503 votes to 96 (15 abstentions). In 2015, a set of commemorative Euro coin was issued on the occasion of the 30th anniversary of the adoption of the emblem by the European Communities. In April 2004, the European flag was flown on behalf of the European Space Agency, by Dutch astronaut André Kuipers while on board the International Space Station, in reference to the Framework Agreement establishing the legal basis for co-operation between the European Space Agency and the European Union. Following the 2004 Summer Olympics, President Romano Prodi expressed his hope "to see the EU Member State teams in Beijing [viz., the 2008 games] carry the flag of the European Union alongside their own national flag as a symbol of our unity". Use of the flag has also been reported as representing the European team at the Ryder Cup golf competition in the early 2000s, although most European participants preferred to use their own national flags. The flag has been widely used by advocates of European integration since the late 1990s or early 2000s. It is often displayed in the context of Europe Day, on 9 May. Outside the EU, it was used in the context of several of the "colour revolutions" during the 2000s. In Belarus, it was used on protest marches alongside the banned former national flag and flags of opposition movements during the protests of 2004–2006. The flag was used widely in a 2007 pro-EU march in Minsk. Similar uses were reported from Moldova. In Georgia, the flag has been on most government buildings since the coming to power of Mikheil Saakashvili (2007), who used it during his inauguration, stating: "[the European] flag is Georgia's flag as well, as far as it embodies our civilisation, our culture, the essence of our history and perspective, and our vision for the future of Georgia." It was used in 2008 by pro-western Serbian voters ahead of an election. The flag became a symbol of European integration of Ukraine in the 2010s, particularly after Euromaidan. Ukraine is not a part of the EU but is a member of the Council of Europe. The flag is used by the Cabinet of Ukraine, Prime Minister of Ukraine, and MFA UA during official meetings. It was flown during the 2013 Euromaidan protests in Ukraine, and in 2016 by the pro-EU faction in the EU membership referendum campaigns in the United Kingdom. The flag has also been adopted as a symbol for EU policies and expansionism by EU-sceptics. In an early instance, Macedonian protesters burned "the flag of the EU" in the context of EU involvement in the 2001 insurgency in the Republic of Macedonia. In the 2005 Islamic protests against the "Jyllands-Posten Muhammad" cartoons, the Danish flag was most frequently burned, but (as the cartoons were reprinted in many European countries), some protesters opted for burning "the EU flag" instead. Protesters during the Greek government-debt crisis of 2012 "burned the EU flag and shouted 'EU out' ". Burning of the EU flag has been reported from other anti-EU rallies since. By the 2010s, the association of the emblem with the EU had become so strong that the Council of Europe saw it necessary to design a new logo, to "avoid confusion", officially adopted in 2013. The EU emblem ("EU flag") is depicted on the euro banknotes. Euro coins also display a circle of twelve stars on both the national and common sides. It is also depicted on many driving licences and vehicle registration plates issued in the Union. Diplomatic missions of EU member states fly the EU flag alongside their national flag. In October 2000, the then-new British Embassy in Berlin sparked controversy between the UK and Germany and the EU when the embassy did not have a second external flagpole for the EU flag. After diplomatic negotiations, it was agreed that the outside flagpole would have the diplomatic Union Flag while inside the embassy, the EU flag would accompany the UK flag. Some member states' national airlines such as Lufthansa have the EU flag alongside their national flags on aircraft as part of their aircraft registration codes, but this is not an EU-mandated directive. A number of logos used by EU institutions, bodies and agencies are derived from the design and colours of the EU emblem. Other emblems make reference to the European flag, such as the EU organic food label that uses the twelve stars but reorders them into the shape of a leaf on a green background. The original logo of the European Broadcasting Union used the twelve stars on a blue background adding ray beams to connect the countries. There was a proposal in 2003 to deface national civil ensigns with the EU emblem. The proposal was rejected by Parliament in 2004. The flag is usually flown by the government of the country holding the rotating presidency Council of Ministers. In 2009, Czech President Václav Klaus, a eurosceptic, refused to fly the flag from his castle. In response, Greenpeace projected an image of the flag onto the castle and attempted to fly the flag from the building themselves. Extraordinary flying of the flag is common on Europe Day, celebrated annually on 9 May. On Europe Day 2008, the flag was flown for the first time above the German Reichstag. The flag has also been displayed in the context of EU military operations (EUFOR Althea). Sixteen out of twenty-seven member states in 2007 signed the declaration recognising "the flag with a circle of twelve golden stars on a blue background" as representing "the sense of community of the people in the European Union and their allegiance to it." In 2017, president of France Emmanuel Macron signed a declaration endorsing the 2007 statement, so that, as of 2018, 17 out of 27 members have recognised the emblem as a flag representing "allegiance to the EU": Austria, Belgium, Bulgaria, Cyprus, France, Germany, Greece, Hungary, Italy, Lithuania, Luxembourg, Malta, Portugal, Romania, Slovakia, Slovenia and Spain. Italy has incorporated the EU flag into its flag code. According to an Italian law passed in 2000, it is mandatory for most public offices and buildings to hoist the European Flag alongside the Italian national flag (Law 22/1998 and Presidential Decree 121/2000). Outside official use, the flag may not be used for "aims incompatible with European values". The 2000 Italian flag code expressly replaces the Italian flag with the European flag in precedence when dignitaries from other EU countries visit – for example the EU flag would be in the middle of a group of three flags rather than the Italian flag. In Germany, the federal flag code of 1996 is only concerned with the German flag, but some of the states have legislated additional provisions for the European flag, such as Bavaria in its flag regulation of 2001, which mandates that the European flag take the third order of precedence, after the federal and state flags, except on Europe Day, where it is to take the first order of precedence. In Ireland on occasions of "European Union Events" (for example, at a European Council meeting), where the European flag is flown alongside all national flags of member states, the national flags are placed in alphabetical order (according to their name in the main language of that state) with the European flag either at the head, or the far-right, of the order of flags. In most member states, use of the EU flag is only "de facto" and not regulated by legislation, and as such subject to "ad hoc" revision. In national usage, national protocol usually demands the national flag takes precedence over the European flag (which is usually displayed to the right of the national flag from the observer's perspective). In November 2014, the speaker of the Hungarian Parliament László Kövér ordered the removal of the EU flag from the parliament building, following an incident in which a member of parliament had "defenestrated" two EU flags from a fourth story window. In November 2015, the newly-elected Polish government under Beata Szydło removed the EU flag from government press conferences. For the flag of the European Council, many stylistic proposals were made in regards to colours and symbolism. These first proposals were made 19 January 1950 by Paul Levy in a letter to the Secrétarie Géneral. He proposed that the flag should contain a cross for several reasons. Firstly, the cross symbolizes roads crossing, and also represents the east, the west, the north, and the south with its arms. Furthermore, the cross appears in most of the European Council members' flags, and it is the oldest and most noble symbol in Europe. Moreover, the cross depicted christianity. As far as the colours are concerned, he proposed them to be white and green, colours of the European Movement, which was of great significance since 1947. Green also depicted hope, and the green cross over a white background was a design that had not been used yet. Finally, Levy proposed that the arms of Strasbourg was an important element to be added as it represented where the Council would be, and being located in the heart of the cross meant that the Council was the point where the European roads met. Shortly after this design considerations by Paul Levy, on 27 July 1950, Richard Coudenhove-Kalergi, president of the Pan-European movement wrote a memorandum which contained some rules that a flag for such union should follow. The rules he stated where: After this statements, Coudenhove-Kalergi proposed that the Pan-European movement flag would be the perfect one to fit this criteria 15 July 1951, the consultative assembly put forward a final memorandum on the European flag. The symbols proposed where the following Furthermore, several colours were also proposed: In the end, the flag of europe was chosen to have 12 five-pointed golden stars in a circle over a blue background, probably inspired by the Pan-European flag and other designs such as Salvador de Madriaga's and Arsène Heitz's proposals. The design of the European flag has been used in a variation, such as that of the Council of Europe mentioned above, and also to a greater extent such as the flag of the Western European Union (WEU; now defunct), which uses the same colours and the stars but has a number of stars based on membership and in a semicircle rather than a circle. It is also defaced with the initials of the former Western European Union in two languages. The European Parliament used its own flag from 1973, but never formally adopted it. It fell out of use with the adoption of the twelve-star flag by the Parliament in 1983. The flag followed the yellow and blue colour scheme however instead of twelve stars there were the letters EP and PE (initials of the European Parliament in the six community languages at the time) surrounded by a wreath. Sometime later, the Parliament chose to use a logo consisting of a stylised hemicycle and the EU flag at the bottom right. The flag of Bosnia and Herzegovina, imposed by High Representative Carlos Westendorp, after the country's parliament failed to agree on a design, is reminiscent of the symbolism of the EU flag, using the same blue and yellow colours, and the stars, although of a different number and colour, are a direct reference to those of the European flag. Likewise, Kosovo uses blue, yellow and stars in its flag, which has been mocked as "a none too subtle nod to the flag of the European Union, which is about to become Kosovo's new best friend as it takes over protector status from the United Nations". The flag of the Brussels-Capital Region (introduced in 2016) consists of a yellow iris with a white outline upon a blue background. Its colours are based on the colours of the Flag of Europe, because Brussels is considered the unofficial capital of the EU. The coat of arms of the Chairman of the European Union Military Committee (CEUMC), the highest ranking officer within the EU's Common Security and Defence Policy (CSDP), depicts the European emblem's blazon heraldically rather than vexillologically. Several EU publications related to the CSDP generally, and its prospective development as a defence arm, have also displayed the European emblem in this manner, albeit as a graphical design element rather than an official symbol.
https://en.wikipedia.org/wiki?curid=10025
Timeline of the evolutionary history of life This timeline of the evolutionary history of life represents the current scientific theory outlining the major events during the development of life on planet Earth. In biology, evolution is any change across successive generations in the heritable characteristics of biological populations. Evolutionary processes give rise to diversity at every level of biological organization, from kingdoms to species, and individual organisms and molecules, such as DNA and proteins. The similarities between all present day organisms indicate the presence of a common ancestor from which all known species, living and extinct, have diverged through the process of evolution. More than 99 percent of all species, amounting to over five billion species, that ever lived on Earth are estimated to be extinct. Estimates on the number of Earth's current species range from 10 million to 14 million, of which about 1.2 million have been documented and over 86 percent have not yet been described. However, a May 2016 scientific report estimates that 1 trillion species are currently on Earth, with only one-thousandth of one percent described. While the dates given in this article are estimates based on scientific evidence, there has been controversy between more traditional views of increased biodiversity through a cone of diversity with the passing of time and the view that the basic pattern on Earth has been one of annihilation and diversification and that in certain past times, such as the Cambrian explosion, there was great diversity. Species go extinct constantly as environments change, as organisms compete for environmental niches, and as genetic mutation leads to the rise of new species from older ones. Occasionally biodiversity on Earth takes a hit in the form of a mass extinction in which the extinction rate is much higher than usual. A large extinction-event often represents an accumulation of smaller extinction- events that take place in a relatively brief period of time. The first known mass extinction in earth's history was the Great Oxygenation Event 2.4 billion years ago. That event led to the loss of most of the planet's obligate anaerobes. Researchers have identified five major extinction events in earth's history since: Smaller extinction-events have occurred in the periods between these larger catastrophes, with some standing at the delineation points of the periods and epochs recognized by scientists in geologic time. The Holocene extinction event is currently under way. Factors in mass extinctions include continental drift, changes in atmospheric and marine chemistry, volcanism and other aspects of mountain formation, changes in glaciation, changes in sea level, and impact events. In this timeline, Ma (for "megaannum") means "million years ago," ka (for "kiloannum") means "thousand years ago," and ya means "years ago." 4000 Ma and earlier. 4000 Ma – 2500 Ma 2500 Ma – 542 Ma. Contains the Palaeoproterozoic, Mesoproterozoic and Neoproterozoic eras. 542 Ma – present The Phanerozoic Eon, literally the "period of well-displayed life," marks the appearance in the fossil record of abundant, shell-forming and/or trace-making organisms. It is subdivided into three eras, the Paleozoic, Mesozoic and Cenozoic, which are divided by major mass extinctions. 542 Ma – 251.0 Ma and contains the Cambrian, Ordovician, Silurian, Devonian, Carboniferous and Permian periods. From 251.4 Ma to 66 Ma and containing the Triassic, Jurassic and Cretaceous periods. 66 Ma – present
https://en.wikipedia.org/wiki?curid=10029
Edmund Burke Edmund Burke (; 12 January [NS] 1729 – 9 July 1797) was an Irish statesman and philosopher. Born in Dublin, Burke served as a member of parliament (MP) between 1766 and 1794 in the House of Commons of Great Britain with the Whig Party after moving to London in 1750. Burke was a proponent of underpinning virtues with manners in society and of the importance of religious institutions for the moral stability and good of the state. These views were expressed in his "A Vindication of Natural Society". He criticized the actions of the British government towards the American colonies, including its taxation policies. Burke also supported the rights of the colonists to resist metropolitan authority, although he opposed the attempt to achieve independence. He is remembered for his support for Catholic emancipation, the impeachment of Warren Hastings from the East India Company, and his staunch opposition to the French Revolution. In his "Reflections on the Revolution in France", Burke asserted that the revolution was destroying the fabric of good society and traditional institutions of state and society and condemned the persecution of the Catholic Church that resulted from it. This led to his becoming the leading figure within the conservative faction of the Whig Party which he dubbed the Old Whigs as opposed to the pro-French Revolution New Whigs led by Charles James Fox. In the 19th century, Burke was praised by both conservatives and liberals. Subsequently in the 20th century, he became widely regarded as the philosophical founder of modern conservatism. Burke was born in Dublin, Ireland. His mother Mary, "née" Nagle (c. 1702–1770), was a Roman Catholic who hailed from a déclassé County Cork family and a cousin of the Catholic educator Nano Nagle whereas his father Richard (died 1761), a successful solicitor, was a member of the Church of Ireland. It remains unclear whether this is the same Richard Burke who converted from Catholicism. The Burke dynasty descends from an Anglo-Norman knight surnamed de Burgh (Latinised as "de Burgo"), who arrived in Ireland in 1185 following Henry II of England's 1171 invasion of Ireland and is among the chief Gall or Old English families that assimilated into Gaelic society, becoming "more Irish than the Irish themselves". Burke adhered to his father's faith and remained a practising Anglican throughout his life, unlike his sister Juliana who was brought up as and remained a Roman Catholic. Later, his political enemies repeatedly accused him of having been educated at the Jesuit College of St. Omer, near Calais, France; and of harbouring secret Catholic sympathies at a time when membership of the Catholic Church would disqualify him from public office (see Penal Laws in Ireland). As Burke told Frances Crewe: Mr. Burke's Enemies often endeavoured to convince the World that he had been bred up in the Catholic Faith, & that his Family were of it, & that he himself had been educated at St. Omer—but this was false, as his father was a regular practitioner of the Law at Dublin, which he could not be unless of the Established Church: & it so happened that though Mr. B—was twice at Paris, he never happened to go through the Town of St. Omer. After being elected to the House of Commons, Burke was required to take the oath of allegiance and abjuration, the oath of supremacy and declare against transubstantiation. Although never denying his Irishness, Burke often described himself as "an Englishman". According to the historian J. C. D. Clark, this was in an age "before 'Celtic nationalism' sought to make Irishness and Englishness incompatible". As a child, Burke sometimes spent time away from the unhealthy air of Dublin with his mother's family in the Blackwater Valley in County Cork. He received his early education at a Quaker school in Ballitore, County Kildare, some from Dublin; and possibly like his cousin Nano Nagle at a Hedge school. He remained in correspondence with his schoolmate from there, Mary Leadbeater, the daughter of the school's owner, throughout his life. In 1744, Burke started at Trinity College Dublin, a Protestant establishment which up until 1793 did not permit Catholics to take degrees. In 1747, he set up a debating society Edmund Burke's Club which in 1770 merged with TCD's Historical Club to form the College Historical Society, the oldest undergraduate society in the world. The minutes of the meetings of Burke's Club remain in the collection of the Historical Society. Burke graduated from Trinity in 1748. Burke's father wanted him to read Law and with this in mind he went to London in 1750, where he entered the Middle Temple, before soon giving up legal study to travel in Continental Europe. After eschewing the Law, he pursued a livelihood through writing. The late Lord Bolingbroke's "Letters on the Study and Use of History" was published in 1752 and his collected works appeared in 1754. This provoked Burke into writing his first published work, "A Vindication of Natural Society: A View of the Miseries and Evils Arising to Mankind", appearing in Spring 1756. Burke imitated Bolingbroke's style and ideas in a "reductio ad absurdum" of his arguments for atheistic rationalism in order to demonstrate their absurdity. Burke claimed that Bolingbroke's arguments against revealed religion could apply to all social and civil institutions as well. Lord Chesterfield and Bishop Warburton as well as others initially thought that the work was genuinely by Bolingbroke rather than a satire. All the reviews of the work were positive, with critics especially appreciative of Burke's quality of writing. Some reviewers failed to notice the ironic nature of the book which led to Burke stating in the preface to the second edition (1757) that it was a satire. Richard Hurd believed that Burke's imitation was near-perfect and that this defeated his purpose, arguing that an ironist "should take care by a constant exaggeration to make the "ridicule" shine through the Imitation. Whereas this "Vindication" is everywhere enforc'd, not only in the language, and on the principles of L. Bol., but with so apparent, or rather so real an earnestness, that half his purpose is sacrificed to the other". A minority of scholars have taken the position that in fact Burke did write the "Vindication" in earnest, later disowning it only for political reasons. In 1757, Burke published a treatise on aesthetics titled "A Philosophical Enquiry into the Origin of Our Ideas of the Sublime and Beautiful" that attracted the attention of prominent Continental thinkers such as Denis Diderot and Immanuel Kant. It was his only purely philosophical work and when asked by Sir Joshua Reynolds and French Laurence to expand it thirty years later, Burke replied that he was no longer fit for abstract speculation (Burke had written it before he was nineteen years of age). On 25 February 1757, Burke signed a contract with Robert Dodsley to write a "history of England from the time of Julius Caesar to the end of the reign of Queen Anne", its length being eighty quarto sheets (640 pages), nearly 400,000 words. It was to be submitted for publication by Christmas 1758. Burke completed the work to the year 1216 and stopped; it was not published until after Burke's death, in an 1812 collection of his works, "An Essay Towards an Abridgement of the English History". G. M. Young did not value Burke's history and claimed that it was "demonstrably a translation from the French". On commenting on the story that Burke stopped his history because David Hume published his, Lord Acton said "it is ever to be regretted that the reverse did not occur". During the year following that contract, Burke founded with Dodsley the influential "Annual Register", a publication in which various authors evaluated the international political events of the previous year. The extent to which Burke contributed to the "Annual Register" is unclear. In his biography of Burke, Robert Murray quotes the "Register" as evidence of Burke's opinions, yet Philip Magnus in his biography does not cite it directly as a reference. Burke remained the chief editor of the publication until at least 1789 and there is no evidence that any other writer contributed to it before 1766. On 12 March 1757, Burke married Jane Mary Nugent (1734–1812), daughter of Dr. Christopher Nugent, a Catholic physician who had provided him with medical treatment at Bath. Their son Richard was born on 9 February 1758 while an elder son, Christopher, died in infancy. Burke also helped raise a ward, Edmund Nagle (later Admiral Sir Edmund Nagle), the son of a maternal cousin orphaned in 1763. At about this same time, Burke was introduced to William Gerard Hamilton (known as "Single-speech Hamilton"). When Hamilton was appointed Chief Secretary for Ireland, Burke accompanied him to Dublin as his private secretary, a position he held for three years. In 1765, Burke became private secretary to the liberal Whig politician Charles, Marquess of Rockingham, then Prime Minister of Great Britain, who remained Burke's close friend and associate until his untimely death in 1782. Rockingham also introduced Burke as a Freemason. In December 1765, Burke entered the House of Commons of the British Parliament as Member for Wendover in Buckinghamshire, a pocket borough in the gift of Lord Fermanagh, later 2nd Earl Verney and a close political ally of Rockingham. After Burke delivered his maiden speech, William Pitt the Elder said he had "spoken in such a manner as to stop the mouths of all Europe" and that the Commons should congratulate itself on acquiring such a Member. The first great subject Burke addressed was the controversy with the American colonies which soon developed into war and ultimate separation. In reply to the 1769 Grenvillite pamphlet "The Present State of the Nation", he published his own pamphlet titled "Observations on a Late State of the Nation". Surveying the finances of France, Burke predicts "some extraordinary convulsion in that whole system". During the same year, with mostly borrowed money, Burke purchased "Gregories", a estate near Beaconsfield. Although the estate included saleable assets such as art works by Titian, "Gregories" proved a heavy financial burden in the following decades and Burke was never able to repay its purchase price in full. His speeches and writings, having made him famous, led to the suggestion that he was the author of the "Letters of Junius". At about this time, Burke joined the circle of leading intellectuals and artists in London of whom Samuel Johnson was the central luminary. This circle also included David Garrick, Oliver Goldsmith and Joshua Reynolds. Edward Gibbon described Burke as "the most eloquent and rational madman that I ever knew". Although Johnson admired Burke's brilliance, he found him a dishonest politician. Burke took a leading role in the debate regarding the constitutional limits to the executive authority of the King. He argued strongly against unrestrained royal power and for the role of political parties in maintaining a principled opposition capable of preventing abuses, either by the monarch, or by specific factions within the government. His most important publication in this regard was his "Thoughts on the Cause of the Present Discontents" of 23 April 1770. Burke identified the "discontents" as stemming from the "secret influence" of a neo-Tory group he labelled as the "king's friends", whose system "comprehending the exterior and interior administrations, is commonly called, in the technical language of the Court, "Double Cabinet"". Britain needed a party with "an unshaken adherence to principle, and attachment to connexion, against every allurement of interest". Party divisions, "whether operating for good or evil, are things inseparable from free government". During 1771, Burke wrote a bill that would have given juries the right to determine what was libel, if passed. Burke spoke in favour of the bill, but it was opposed by some, including Charles James Fox, not becoming law. When introducing his own bill in 1791 in opposition, Fox repeated almost verbatim the text of Burke's bill without acknowledgement. Burke was prominent in securing the right to publish debates held in Parliament. Speaking in a parliamentary debate on the prohibition on the export of grain on 16 November 1770, Burke argued in favour of a free market in corn: "There are no such things as a high, & a low price that is encouraging, & discouraging; there is nothing but a natural price, which grain brings at an universal market". In 1772, Burke was instrumental in the passing of the Repeal of Certain Laws Act 1772 which repealed various old laws against dealers and forestallers in corn. In the "Annual Register" for 1772 (published in July 1773), Burke condemned the partition of Poland. He saw it as "the first very great breach in the modern political system of Europe" and as upsetting the balance of power in Europe. On 4 November 1774, Burke was elected Member for Bristol, at the time "England's second city" and a large constituency with a genuine electoral contest. At the conclusion of the poll, he made his "Speech to the Electors of Bristol at the Conclusion of the Poll", a remarkable disclaimer of the constituent-imperative form of democracy, for which he substituted his statement of the "representative mandate" form. He failed to win re-election for that seat. In May 1778, Burke supported a parliamentary motion revising restrictions on Irish trade. His constituents, citizens of the great trading city of Bristol, urged Burke to oppose free trade with Ireland. Burke resisted their protestations and said: "If, from this conduct, I shall forfeit their suffrages at an ensuing election, it will stand on record an example to future representatives of the Commons of England, that one man at least had dared to resist the desires of his constituents when his judgment assured him they were wrong". Burke published "Two Letters to Gentlemen of Bristol on the Bills relative to the Trade of Ireland" in which he espoused "some of the chief principles of commerce; such as the advantage of free intercourse between all parts of the same kingdom, [...] the evils attending restriction and monopoly, [...] and that the gain of others is not necessarily our loss, but on the contrary an advantage by causing a greater demand for such wares as we have for sale". Burke also supported the attempts of Sir George Savile to repeal some of the penal laws against Catholics. Burke also called capital punishment "the Butchery which we call justice" in 1776 and in 1780 condemned the use of the pillory for two men convicted for attempting to practice sodomy. This support for unpopular causes, notably free trade with Ireland and Catholic emancipation, led to Burke losing his seat in 1780. For the remainder of his parliamentary career, Burke represented Malton, another pocket borough under the Marquess of Rockingham's patronage. Burke expressed his support for the grievances of the American Thirteen Colonies under the government of King George III and his appointed representatives. On 19 April 1774, Burke made a speech, "On American Taxation" (published in January 1775), on a motion to repeal the tea duty: Again and again, revert to your old principles—seek peace and ensue it; leave America, if she has taxable matter in her, to tax herself. I am not here going into the distinctions of rights, nor attempting to mark their boundaries. I do not enter into these metaphysical distinctions; I hate the very sound of them. Leave the Americans as they anciently stood, and these distinctions, born of our unhappy contest, will die along with it. [...] Be content to bind America by laws of trade; you have always done it... Do not burthen them with taxes... But if intemperately, unwisely, fatally, you sophisticate and poison the very source of government by urging subtle deductions, and consequences odious to those you govern, from the unlimited and illimitable nature of supreme sovereignty, you will teach them by these means to call that sovereignty itself in question. [...] If that sovereignty and their freedom cannot be reconciled, which will they take? They will cast your sovereignty in your face. No body of men will be argued into slavery. On 22 March 1775, Burke delivered in the House of Common a speech (published during May 1775) on reconciliation with America. Burke appealed for peace as preferable to civil war and reminded the House of Commons of America's growing population, its industry and its wealth. He warned against the notion that the Americans would back down in the face of force since most Americans were of British descent: [T]he people of the colonies are descendants of Englishmen. [...] They are therefore not only devoted to liberty, but to liberty according to English ideas and on English principles. The people are Protestants, [...] a persuasion not only favourable to liberty, but built upon it. [...] My hold of the colonies is in the close affection which grows from common names, from kindred blood, from similar privileges, and equal protection. These are ties which, though light as air, are as strong as links of iron. Let the colonies always keep the idea of their civil rights associated with your government—they will cling and grapple to you, and no force under heaven will be of power to tear them from their allegiance. But let it be once understood that your government may be one thing and their privileges another, that these two things may exist without any mutual relation—the cement is gone, the cohesion is loosened, and everything hastens to decay and dissolution. As long as you have the wisdom to keep the sovereign authority of this country as the sanctuary of liberty, the sacred temple consecrated to our common faith, wherever the chosen race and sons of England worship freedom, they will turn their faces towards you. The more they multiply, the more friends you will have; the more ardently they love liberty, the more perfect will be their obedience. Slavery they can have anywhere. It is a weed that grows in every soil. They may have it from Spain, they may have it from Prussia. But, until you become lost to all feeling of your true interest and your natural dignity, freedom they can have from none but you. Burke prized peace with America above all else, pleading with the House of Commons to remember that the interest by way of money received from the American colonies was far more attractive than any sense of putting the colonists in their place: The proposition is peace. Not peace through the medium of war, not peace to be hunted through the labyrinth of intricate and endless negotiations, not peace to arise out of universal discord. [...] [I]t is simple peace, sought in its natural course and in its ordinary haunts. It is peace sought in the spirit of peace, and laid in principles purely pacific. Burke was not merely presenting a peace agreement to Parliament, but rather he stepped forward with four reasons against using force, carefully reasoned. He laid out his objections in an orderly manner, focusing on one before moving to the next. His first concern was that the use of force would have to be temporary and that the uprisings and objections to British governance in Colonial America would not be. Second, Burke worried about the uncertainty surrounding whether Britain would win a conflict in America. "An armament", Burke said, "is not a victory". Third, Burke brought up the issue of impairment, stating that it would do the British government no good to engage in a scorched earth war and have the object they desired (America) become damaged or even useless. The American colonists could always retreat into the mountains, but the land they left behind would most likely be unusable, whether by accident or design. The fourth and final reason to avoid the use of force was experience as the British had never attempted to rein in an unruly colony by force and they did not know if it could be done, let alone accomplished thousands of miles away from home. Not only were all of these concerns reasonable, but some turned out to be prophetic—the American colonists did not surrender, even when things looked extremely bleak and the British were ultimately unsuccessful in their attempts to win a war fought on American soil. It was not temporary force, uncertainty, impairment, or even experience that Burke cited as the number one reason for avoiding war with the American colonies. Rather, it was the character of the American people themselves: "In this character of Americans, a love of freedom is the predominating feature which marks and distinguishes the whole. [...] [T]his fierce spirit of liberty is stronger in the English colonies, probably, than in any other people of the earth. [...] [The] men [are] acute, inquisitive, dextrous, prompt in attack, ready in defence, full of resources". Burke concludes with another plea for peace and a prayer that Britain might avoid actions which in Burke's words "may bring on the destruction of this Empire". Burke proposed six resolutions to settle the American conflict peacefully: Had they been passed, the effect of these resolutions can never be known. Unfortunately, Burke delivered this speech just less than a month before the explosive conflict at Concord and Lexington. As these resolutions were not enacted, little was done that would help to dissuade conflict. Among the reasons this speech was so greatly admired was its passage on Lord Bathurst (1684–1775) in which Burke describes an angel in 1704 prophesying to Bathurst the future greatness of England and also of America: "Young man, There is America—which at this day serves little more than to amuse you with stories of savage men, and uncouth manners; yet shall, before you taste of death, shew itself equal to the whole of that commerce which now attracts the envy of the world". Samuel Johnson was so irritated at hearing it continually praised that he made a parody of it, where the devil appears to a young Whig and predicts that in short time Whiggism will poison even the paradise of America. The administration of Lord North (1770–1782) tried to defeat the colonist rebellion by military force. British and American forces clashed in 1775 and in 1776 came the American Declaration of Independence. Burke was appalled by celebrations in Britain of the defeat of the Americans at New York and Pennsylvania. He claimed the English national character was being changed by this authoritarianism. Burke wrote: "As to the good people of England, they seem to partake every day more and more of the Character of that administration which they have been induced to tolerate. I am satisfied, that within a few years there has been a great Change in the National Character. We seem no longer that eager, inquisitive, jealous, fiery people, which we have been formerly". In Burke's view, the British government was fighting "the American English" ("our English Brethren in the Colonies"), with a Germanic king employing "the hireling sword of German boors and vassals" to destroy the English liberties of the colonists. On American independence, Burke wrote: "I do not know how to wish success to those whose Victory is to separate from us a large and noble part of our Empire. Still less do I wish success to injustice, oppression and absurdity". During the Gordon Riots in 1780, Burke became a target of hostility and his home was placed under armed guard by the military. The fall of North led to Rockingham being recalled to power in March 1782. Burke was appointed Paymaster of the Forces and a Privy Counsellor, but without a seat in Cabinet. Rockingham's unexpected death in July 1782 and replacement with Shelburne as Prime Minister put an end to his administration after only a few months, but Burke did manage to introduce two Acts. The Paymaster General Act 1782 ended the post as a lucrative sinecure. Previously, Paymasters had been able to draw on money from HM Treasury at their discretion. Instead, now they were required to put the money they had requested to withdraw from the Treasury into the Bank of England, from where it was to be withdrawn for specific purposes. The Treasury would receive monthly statements of the Paymaster's balance at the Bank. This Act was repealed by Shelburne's administration, but the Act that replaced it repeated verbatim almost the whole text of the Burke Act. The Civil List and Secret Service Money Act 1782 was a watered down version of Burke's original intentions as outlined in his famous "Speech on Economical Reform" of 11 February 1780. However, he managed to abolish 134 offices in the royal household and civil administration. The third Secretary of State and the Board of Trade were abolished and pensions were limited and regulated. The Act was anticipated to save £72,368 a year. In February 1783, Burke resumed the post of Paymaster of the Forces when Shelburne's government fell and was replaced by a coalition headed by North that included Charles James Fox. That coalition fell in 1783 and was succeeded by the long Tory administration of William Pitt the Younger which lasted until 1801. Accordingly, having supported Fox and North, Burke was in opposition for the remainder of his political life. In 1774, Burke's "Speech to the Electors at Bristol at the Conclusion of the Poll" was noted for its defence of the principles of representative government against the notion that elected officials should merely be delegates: [I]t ought to be the happiness and glory of a representative to live in the strictest union, the closest correspondence, and the most unreserved communication with his constituents. Their wishes ought to have great weight with him; their opinion, high respect; their business, unremitted attention. It is his duty to sacrifice his repose, his pleasures, his satisfactions, to theirs; and above all, ever, and in all cases, to prefer their interest to his own. But his unbiased opinion, his mature judgment, his enlightened conscience, he ought not to sacrifice to you, to any man, or to any set of men living. These he does not derive from your pleasure; no, nor from the law and the constitution. They are a trust from Providence, for the abuse of which he is deeply answerable. Your representative owes you, not his industry only, but his judgment; and he betrays, instead of serving you, if he sacrifices it to your opinion. Political scientist Hanna Pitkin points out that Burke linked the interest of the district with the proper behaviour of its elected official, explaining: "Burke conceives of broad, relatively fixed interest, few in number and clearly defined, of which any group or locality has just one. These interests are largely economic or associated with particular localities whose livelihood they characterize, in his over-all prosperity they involve". Burke was a leading sceptic with respect to democracy. While admitting that theoretically in some cases it might be desirable, he insisted a democratic government in Britain in his day would not only be inept, but also oppressive. He opposed democracy for three basic reasons. First, government required a degree of intelligence and breadth of knowledge of the sort that occurred rarely among the common people. Second, he thought that if they had the vote, common people had dangerous and angry passions that could be aroused easily by demagogues, fearing that the authoritarian impulses that could be empowered by these passions would undermine cherished traditions and established religion, leading to violence and confiscation of property. Third, Burke warned that democracy would create a tyranny over unpopular minorities, who needed the protection of the upper classes. Burke proposed a bill to ban slaveholders from being able to sit in the British House of Commons claiming they were a danger incompatible with British liberty. For years, Burke pursued impeachment efforts against Warren Hastings, formerly Governor-General of Bengal, that resulted in the trial during 1786. His interaction with the British dominion of India began well before Hastings' impeachment trial. For two decades prior to the impeachment, Parliament had dealt with the Indian issue. This trial was the pinnacle of years of unrest and deliberation. In 1781, Burke was first able to delve into the issues surrounding the East India Company when he was appointed Chairman of the Commons Select Committee on East Indian Affairs—from that point until the end of the trial, India was Burke's primary concern. This committee was charged "to investigate alleged injustices in Bengal, the war with Hyder Ali, and other Indian difficulties". While Burke and the committee focused their attention on these matters, a second secret committee was formed to assess the same issues. Both committee reports were written by Burke. Among other purposes, the reports conveyed to the Indian princes that Britain would not wage war on them, along with demanding that the East India Company should recall Hastings. This was Burke's first call for substantive change regarding imperial practices. When addressing the whole House of Commons regarding the committee report, Burke described the Indian issue as one that "began 'in commerce' but 'ended in empire'". On 28 February 1785, Burke delivered a now-famous speech, "The Nabob of Arcot's Debts", wherein he condemned the damage to India by the East India Company. In the province of the Carnatic, the Indians had constructed a system of reservoirs to make the soil fertile in a naturally dry region, and centred their society on the husbandry of water: These are the monuments of real kings, who were the fathers of their people; testators to a posterity which they embraced as their own. These are the grand sepulchres built by ambition; but by the ambition of an insatiable benevolence, which, not contented with reigning in the dispensation of happiness during the contracted term of human life, had strained, with all the reachings and graspings of a vivacious mind, to extend the dominion of their bounty beyond the limits of nature, and to perpetuate themselves through generations of generations, the guardians, the protectors, the nourishers of mankind. Burke held that the advent of British dominion, in particular the conduct of the East India Company, had destroyed much that was good in these traditions and that as a consequence of this and the lack of new customs to replace them the Indians were suffering. He set about establishing a set of British expectations, whose moral foundation would in his opinion warrant the empire. On 4 April 1786, Burke presented the House of Commons with the "Article of Charge of High Crimes and Misdemeanors" against Hastings. The impeachment in Westminster Hall which did not begin until 14 February 1788 would be the "first major public discursive event of its kind in England", bringing the morality and duty of imperialism to the forefront of public perception. Burke was already known for his eloquent rhetorical skills and his involvement in the trial only enhanced its popularity and significance. Burke's indictment, fuelled by emotional indignation, branded Hastings a "captain-general of iniquity" who never dined without "creating a famine", whose heart was "gangrened to the core" and who resembled both a "spider of Hell" and a "ravenous vulture devouring the carcasses of the dead". The House of Commons eventually impeached Hastings, but subsequently the House of Lords acquitted him of all charges. Initially, Burke did not condemn the French Revolution. In a letter of 9 August 1789, he wrote: "England gazing with astonishment at a French struggle for Liberty and not knowing whether to blame or to applaud! The thing indeed, though I thought I saw something like it in progress for several years, has still something in it paradoxical and Mysterious. The spirit it is impossible not to admire; but the old Parisian ferocity has broken out in a shocking manner". The events of 5–6 October 1789, when a crowd of Parisian women marched on Versailles to compel King Louis XVI to return to Paris, turned Burke against it. In a letter to his son Richard Burke dated 10 October, he said: "This day I heard from Laurence who has sent me papers confirming the portentous state of France—where the Elements which compose Human Society seem all to be dissolved, and a world of Monsters to be produced in the place of it—where Mirabeau presides as the Grand Anarch; and the late Grand Monarch makes a figure as ridiculous as pitiable". On 4 November, Charles-Jean-François Depont wrote to Burke, requesting that he endorse the Revolution. Burke replied that any critical language of it by him should be taken "as no more than the expression of doubt", but he added: "You may have subverted Monarchy, but not recover'd freedom". In the same month, he described France as "a country undone". Burke's first public condemnation of the Revolution occurred on the debate in Parliament on the army estimates on 9 February 1790 provoked by praise of the Revolution by Pitt and Fox: Since the House had been prorogued in the summer much work was done in France. The French had shewn themselves the ablest architects of ruin that had hitherto existed in the world. In that very short space of time they had completely pulled down to the ground, their monarchy; their church; their nobility; their law; their revenue; their army; their navy; their commerce; their arts; and their manufactures. [...] [There was a danger of] an imitation of the excesses of an irrational, unprincipled, proscribing, confiscating, plundering, ferocious, bloody and tyrannical democracy. [...] [In religion] the danger of their example is no longer from intolerance, but from Atheism; a foul, unnatural vice, foe to all the dignity and consolation of mankind; which seems in France, for a long time, to have been embodied into a faction, accredited, and almost avowed. In January 1790, Burke read Richard Price's sermon of 4 November 1789 entitled "A Discourse on the Love of Our Country" to the Revolution Society. That society had been founded to commemorate the Glorious Revolution of 1688. In this sermon, Price espoused the philosophy of universal "Rights of Men". Price argued that love of our country "does not imply any conviction of the superior value of it to other countries, or any particular preference of its laws and constitution of government". Instead, Price asserted that Englishmen should see themselves "more as citizens of the world than as members of any particular community". A debate between Price and Burke ensued that was "the classic moment at which two fundamentally different conceptions of national identity were presented to the English public". Price claimed that the principles of the Glorious Revolution included "the right to choose our own governors, to cashier them for misconduct, and to frame a government for ourselves". Immediately after reading Price's sermon, Burke wrote a draft of what eventually became "Reflections on the Revolution in France". On 13 February 1790, a notice in the press said that shortly Burke would publish a pamphlet on the Revolution and its British supporters, but he spent the year revising and expanding it. On 1 November, he finally published the "Reflections" and it was an immediate best-seller. Priced at five shillings, it was more expensive than most political pamphlets, but by the end of 1790 it had gone through ten printings and sold approximately 17,500 copies. A French translation appeared on 29 November and on 30 November the translator Pierre-Gaëton Dupont wrote to Burke saying 2,500 copies had already been sold. The French translation ran to ten printings by June 1791. What the Glorious Revolution had meant was as important to Burke and his contemporaries as it had been for the last one hundred years in British politics. In the "Reflections", Burke argued against Price's interpretation of the Glorious Revolution and instead, gave a classic Whig defence of it. Burke argued against the idea of abstract, metaphysical rights of humans and instead advocated national tradition: The Revolution was made to preserve our "antient" indisputable laws and liberties, and that "antient" constitution of government which is our only security for law and liberty... The very idea of the fabrication of a new government, is enough to fill us with disgust and horror. We wished at the period of the Revolution, and do now wish, to derive all we possess as "an inheritance from our forefathers". Upon that body and stock of inheritance we have taken care not to inoculate any cyon [scion] alien to the nature of the original plant. [...] Our oldest reformation is that of Magna Charta. You will see that Sir Edward Coke, that great oracle of our law, and indeed all the great men who follow him, to Blackstone, are industrious to prove the pedigree of our liberties. They endeavour to prove that the ancient charter... were nothing more than a re-affirmance of the still more ancient standing law of the kingdom. [...] In the famous law [...] called the "Petition of Right", the parliament says to the king, "Your subjects have "inherited" this freedom", claiming their franchises not on abstract principles "as the rights of men", but as the rights of Englishmen, and as a patrimony derived from their forefathers. Burke said: "We fear God, we look up with awe to kings; with affection to parliaments; with duty to magistrates; with reverence to priests; and with respect to nobility. Why? Because when such ideas are brought before our minds, it is "natural" to be so affected". Burke defended this prejudice on the grounds that it is "the general bank and capital of nations, and of ages" and superior to individual reason, which is small in comparison. "Prejudice", Burke claimed, "is of ready application in the emergency; it previously engages the mind in a steady course of wisdom and virtue, and does not leave the man hesitating in the moment of decision, sceptical, puzzled, and unresolved. Prejudice renders a man's virtue his habit". Burke criticised social contract theory by claiming that society is indeed a contract, although it is "a partnership not only between those who are living, but between those who are living, those who are dead, and those who are to be born". The most famous passage in Burke's "Reflections" was his description of the events of 5–6 October 1789 and the part of Marie-Antoinette in them. Burke's account differs little from modern historians who have used primary sources. His use of flowery language to describe it provoked both praise and criticism. Philip Francis wrote to Burke saying that what he wrote of Marie-Antoinette was "pure foppery". Edward Gibbon reacted differently: "I adore his chivalry". Burke was informed by an Englishman who had talked with the Duchesse de Biron that when Marie-Antoinette was reading the passage she burst into tears and took considerable time to finish reading it. Price had rejoiced that the French king had been "led in triumph" during the October Days, but to Burke this symbolised the opposing revolutionary sentiment of the Jacobins and the natural sentiments of those who shared his own view with horror—that the ungallant assault on Marie-Antoinette—was a cowardly attack on a defenceless woman. Louis XVI translated the "Reflections" "from end to end" into French. Fellow Whig MPs Richard Sheridan and Charles James Fox disagreed with Burke and split with him. Fox thought the "Reflections" to be "in very bad taste" and "favouring Tory principles". Other Whigs such as the Duke of Portland and Earl Fitzwilliam privately agreed with Burke, but they did not wish for a public breach with their Whig colleagues. Burke wrote on 29 November 1790: "I have received from the Duke of Portland, Lord Fitzwilliam, the Duke of Devonshire, Lord John Cavendish, Montagu (Frederick Montagu MP), and a long et cetera of the old Stamina of the Whiggs a most full approbation of the principles of that work and a kind indulgence to the execution". The Duke of Portland said in 1791 that when anyone criticised the "Reflections" to him, he informed them that he had recommended the book to his sons as containing the true Whig creed. In the opinion of Paul Langford, Burke crossed something of a Rubicon when he attended a levee on 3 February 1791 to meet the King, later described by Jane Burke as follows: On his coming to Town for the Winter, as he generally does, he went to the Levee with the Duke of "Portland", who went with Lord William to kiss hands on his going into the Guards—while Lord William was kissing hands, The King was talking to The Duke, but his Eyes were fixed on [Burke] who was standing in the Crowd, and when He said His say to The Duke, without waiting for [Burke]'s coming up in his turn, The King went up to him, and, after the usual questions of how long have you been in Town and the weather, He said you have been very much employed of late, and very much confined. [Burke] said, no, Sir, not more than usual—You have and very well employed too, but there are none so deaf as those that w'ont hear, and none so blind as those that w'ont see—[Burke] made a low bow, Sir, I certainly now understand you, but was afraid my vanity or presumption might have led me to imagine what Your Majesty has said referred to what I have done—You cannot be vain—You have been of "use to us all", it is a general opinion, is it not so Lord Stair? who was standing near. It is said Lord Stair;—Your Majesty's adopting it, Sir, will make the opinion general, said [Burke]—I know it is the general opinion, and I know that there is no Man who calls himself a Gentleman that must not think himself obliged to you, for you have supported the cause of the Gentlemen—You know the tone at Court is a whisper, but The King said all this loud, so as to be heard by every one at Court. Burke's "Reflections" sparked a pamphlet war. Mary Wollstonecraft was one of the first into print, publishing "A Vindication of the Rights of Men" a few weeks after Burke. Thomas Paine followed with the "Rights of Man" in 1791. James Mackintosh, who wrote "Vindiciae Gallicae", was the first to see the "Reflections" as "the manifesto of a Counter Revolution". Mackintosh later agreed with Burke's views, remarking in December 1796 after meeting him that Burke was "minutely and accurately informed, to a wonderful exactness, with respect to every fact relating to the French Revolution". Mackintosh later said: "Burke was one of the first thinkers as well as one of the greatest orators of his time. He is without parallel in any age, excepting perhaps Lord Bacon and Cicero; and his works contain an ampler store of political and moral wisdom than can be found in any other writer whatever". In November 1790, François-Louis-Thibault de Menonville, a member of the National Assembly of France, wrote to Burke, praising "Reflections" and requesting more "very refreshing mental food" that he could publish. This Burke did in April 1791 when he published "A Letter to a Member of the National Assembly". Burke called for external forces to reverse the Revolution and included an attack on the late French philosopher Jean-Jacques Rousseau as being the subject of a personality cult that had developed in revolutionary France. Although Burke conceded that Rousseau sometimes showed "a considerable insight into human nature", he mostly was critical. Although he did not meet Rousseau on his visit to Britain in 1766–1767, Burke was a friend of David Hume, with whom Rousseau had stayed. Burke said Rousseau "entertained no principle either to influence of his heart, or to guide his understanding—but "vanity""—which he "was possessed to a degree little short of madness". He also cited Rousseau's "Confessions" as evidence that Rousseau had a life of "obscure and vulgar vices" that was not "chequered, or spotted here and there, with virtues, or even distinguished by a single good action". Burke contrasted Rousseau's theory of universal benevolence and his having sent his children to a foundling hospital, stating that he was "a lover of his kind, but a hater of his kindred". These events and the disagreements that arose from them within the Whig Party led to its break-up and to the rupture of Burke's friendship with Fox. In debate in Parliament on Britain's relations with Russia, Fox praised the principles of the Revolution, although Burke was not able to reply at this time as he was "overpowered by continued cries of question from his own side of the House". When Parliament was debating the Quebec Bill for a constitution for Canada, Fox praised the Revolution and criticised some of Burke's arguments such as hereditary power. On 6 May 1791, Burke used the opportunity to answer Fox during another debate in Parliament on the Quebec Bill and condemn the new French Constitution and "the horrible consequences flowing from the French idea of the Rights of Man". Burke asserted that those ideas were the antithesis of both the British and the American constitutions. Burke was interrupted and Fox intervened, saying that Burke should be allowed to carry on with his speech. However, a vote of censure was moved against Burke for noticing the affairs of France which was moved by Lord Sheffield and seconded by Fox. Pitt made a speech praising Burke and Fox made a speech—both rebuking and complimenting Burke. He questioned the sincerity of Burke, who seemed to have forgotten the lessons he had learned from him, quoting from Burke's own speeches of fourteen and fifteen years before. Burke's response was as follows: It certainly was indiscreet at any period, but especially at his time of life, to parade enemies, or give his friends occasion to desert him; yet if his firm and steady adherence to the British constitution placed him in such a dilemma, he would risk all, and, as public duty and public experience taught him, with his last words exclaim, "Fly from the French Constitution". At this point, Fox whispered that there was "no loss of friendship". "I regret to say there is", Burke replied, "I have indeed made a great sacrifice; I have done my duty though I have lost my friend. There is something in the detested French constitution that envenoms every thing it touches". This provoked a reply from Fox, yet he was unable to give his speech for some time since he was overcome with tears and emotion. Fox appealed to Burke to remember their inalienable friendship, but he also repeated his criticisms of Burke and uttered "unusually bitter sarcasms". This only aggravated the rupture between the two men. Burke demonstrated his separation from the party on 5 June 1791 by writing to Fitzwilliam, declining money from him. Burke was dismayed that some Whigs, instead of reaffirming the principles of the Whig Party he laid out in the "Reflections", had rejected them in favour of "French principles" and that they criticised Burke for abandoning Whig principles. Burke wanted to demonstrate his fidelity to Whig principles and feared that acquiescence to Fox and his followers would allow the Whig Party to become a vehicle for Jacobinism. Burke knew that many members of the Whig Party did not share Fox's views and he wanted to provoke them into condemning the French Revolution. Burke wrote that he wanted to represent the whole Whig Party "as tolerating, and by a toleration, countenancing those proceedings" so that he could "stimulate them to a public declaration of what every one of their acquaintance privately knows to be [...] their sentiments". On 3 August 1791, Burke published his "Appeal from the New to the Old Whigs" in which he renewed his criticism of the radical revolutionary programmes inspired by the French Revolution and attacked the Whigs who supported them as holding principles contrary to those traditionally held by the Whig Party. Burke owned two copies of what has been called "that practical compendium of Whig political theory", namely "The Tryal of Dr. Henry Sacheverell" (1710). Burke wrote of the trial: "It rarely happens to a party to have the opportunity of a clear, authentic, recorded, declaration of their political tenets upon the subject of a great constitutional event like that of the [Glorious] Revolution". Writing in the third person, Burke asserted in his "Appeal": [The] foundations laid down by the Commons, on the trial of Doctor Sacheverel, for justifying the revolution of 1688, are the very same laid down in Mr. Burke's Reflections; that is to say,—a breach of the "original contract", implied and expressed in the constitution of this country, as a scheme of government fundamentally and inviolably fixed in King, Lords and Commons.—That the fundamental subversion of this antient constitution, by one of its parts, having been attempted, and in effect accomplished, justified the Revolution. That it was justified "only" upon the "necessity" of the case; as the "only" means left for the recovery of that "antient" constitution, formed by the "original contract" of the British state; as well as for the future preservation of the "same" government. These are the points to be proved. Burke then provided quotations from Paine's "Rights of Man" to demonstrate what the New Whigs believed. Burke's belief that Foxite principles corresponded to Paine's was genuine. Finally, Burke denied that a majority of "the people" had, or ought to have, the final say in politics and alter society at their pleasure. People had rights, but also duties and these duties were not voluntary. According to Burke, the people could not overthrow morality derived from God. Although Whig grandees such as Portland and Fitzwilliam privately agreed with Burke's "Appeal", they wished he had used more moderate language. Fitzwilliam saw the "Appeal" as containing "the doctrines I have sworn by, long and long since". Francis Basset, a backbench Whig MP, wrote to Burke that "though for reasons which I will not now detail I did not then deliver my sentiments, I most perfectly differ from Mr. Fox & from the great Body of opposition on the French Revolution". Burke sent a copy of the "Appeal" to the King and the King requested a friend to communicate to Burke that he had read it "with great Satisfaction". Burke wrote of its reception: "Not one word from one of our party. They are secretly galled. They agree with me to a title; but they dare not speak out for fear of hurting Fox. [...] They leave me to myself; they see that I can do myself justice". Charles Burney viewed it as "a most admirable book—the best & most useful on political subjects that I have ever seen", but he believed the differences in the Whig Party between Burke and Fox should not be aired publicly. Eventually, most of the Whigs sided with Burke and gave their support to William Pitt the Younger's Tory government which in response to France's declaration of war against Britain declared war on France's Revolutionary Government in 1793. In December 1791, Burke sent government ministers his "Thoughts on French Affairs" where he put forward three main points, namely that no counter-revolution in France would come about by purely domestic causes; that the longer the Revolutionary Government exists, the stronger it becomes; and that the Revolutionary Government's interest and aim is to disturb all of the other governments of Europe. As a Whig, Burke did not wish to see an absolute monarchy again in France after the extirpation of Jacobinism. Writing to an "émigré" in 1791, Burke expressed his views against a restoration of the "Ancien Régime": When such a complete convulsion has shaken the State, and hardly left any thing whatsoever, either in civil arrangements, or in the Characters and disposition of men's minds, exactly where it was, whatever shall be settled although in the former persons and upon old forms, will be in some measure a new thing and will labour under something of the weakness as well as other inconveniences of a Change. My poor opinion is that you mean to establish what you call 'L'ancien Régime,' If any one means that system of Court Intrigue miscalled a Government as it stood, at Versailles before the present confusions as the thing to be established, that I believe will be found absolutely impossible; and if you consider the Nature, as well of persons, as of affairs, I flatter myself you must be of my opinion. That was tho' not so violent a State of Anarchy as well as the present. If it were even possible to lay things down exactly as they stood, before the series of experimental politicks began, I am quite sure that they could not long continue in that situation. In one Sense of L'Ancien Régime I am clear that nothing else can reasonably be done. Burke delivered a speech on the debate of the Aliens Bill on 28 December 1792. He supported the Bill as it would exclude "murderous atheists, who would pull down Church and state; religion and God; morality and happiness". The peroration included a reference to a French order for 3,000 daggers. Burke revealed a dagger he had concealed in his coat and threw it to the floor: "This is what you are to gain by an alliance with France". Burke picked up the dagger and continued: When they smile, I see blood trickling down their faces; I see their insidious purposes; I see that the object of all their cajoling is—blood! I now warn my countrymen to beware of these execrable philosophers, whose only object it is to destroy every thing that is good here, and to establish immorality and murder by precept and example—'Hic niger est hunc tu Romane caveto' ['Such a man is evil; beware of him, Roman'. Horace, "Satires" I. 4. 85.]. Burke supported the war against Revolutionary France, seeing Britain as fighting on the side of the royalists and "émigres" in a civil war, rather than fighting against the whole nation of France. Burke also supported the royalist uprising in La Vendée, describing it on 4 November 1793 in a letter to William Windham as "the sole affair I have much heart in". Burke wrote to Henry Dundas on 7 October urging him to send reinforcements there as he viewed it as the only theatre in the war that might lead to a march on Paris, but Dundas did not follow Burke's advice. Burke believed the British government was not taking the uprising seriously enough, a view reinforced by a letter he had received from the Prince Charles of France ("S.A.R. le comte d'Artois"), dated 23 October, requesting that he intercede on behalf of the royalists to the government. Burke was forced to reply on 6 November: "I am not in His Majesty's Service; or at all consulted in his Affairs". Burke published his "Remarks on the Policy of the Allies with Respect to France", begun in October, where he said: "I am sure every thing has shewn us that in this war with France, one Frenchman is worth twenty foreigners. La Vendée is a proof of this". On 20 June 1794, Burke received a vote of thanks from the House of Commons for his services in the Hastings Trial and he immediately resigned his seat, being replaced by his son Richard. A tragic blow fell upon Burke with the loss of Richard in August 1794, to whom he was tenderly attached and in whom he saw signs of promise which were not patent to others and which in fact appear to have been non-existent, although this view may have rather reflected the fact that his son Richard had worked successfully in the early battle for Catholic emancipation. King George III, whose favour he had gained by his attitude on the French Revolution, wished to create him Earl of Beaconsfield, but the death of his son deprived the opportunity of such an honour and all its attractions, so the only award he would accept was a pension of £2,500. Even this modest reward was attacked by the Duke of Bedford and the Earl of Lauderdale, to whom Burke replied in his "Letter to a Noble Lord" (1796): "It cannot at this time be too often repeated; line upon line; precept upon precept; until it comes into the currency of a proverb, "To innovate is not to reform"". He argued that he was rewarded on merit, but the Duke of Bedford received his rewards from inheritance alone, his ancestor being the original pensioner: "Mine was from a mild and benevolent sovereign; his from Henry the Eighth". Burke also hinted at what would happen to such people if their revolutionary ideas were implemented and included a description of the British Constitution: But as to "our" country and "our" race, as long as the well compacted structure of our church and state, the sanctuary, the holy of holies of that ancient law, defended by reverence, defended by power, a fortress at once and a temple, shall stand inviolate on the brow of the British Sion—as long as the British Monarchy, not more limited than fenced by the orders of the State, shall, like the proud Keep of Windsor, rising in the majesty of proportion, and girt with the double belt of its kindred and coeval towers, as long as this awful structure shall oversee and guard the subjected land—so long as the mounds and dykes of the low, fat, Bedford level will have nothing to fear from all the pickaxes of all the levellers of France. Burke's last publications were the "Letters on a Regicide Peace" (October 1796), called forth by negotiations for peace with France by the Pitt government. Burke regarded this as appeasement, injurious to national dignity and honour. In his "Second Letter", Burke wrote of the French Revolutionary government: "Individuality is left out of their scheme of government. The State is all in all. Everything is referred to the production of force; afterwards, everything is trusted to the use of it. It is military in its principle, in its maxims, in its spirit, and in all its movements. The State has dominion and conquest for its sole objects—dominion over minds by proselytism, over bodies by arms". This is held to be the first explanation of the modern concept of totalitarian state. Burke regarded the war with France as ideological, against an "armed doctrine". He wished that France would not be partitioned due to the effect this would have on the balance of power in Europe and that the war was not against France, but against the revolutionaries governing her. Burke said: "It is not France extending a foreign empire over other nations: it is a sect aiming at universal empire, and beginning with the conquest of France". In November 1795, there was a debate in Parliament on the high price of corn and Burke wrote a memorandum to Pitt on the subject. In December, Samuel Whitbread MP introduced a bill giving magistrates the power to fix minimum wages and Fox said he would vote for it. This debate probably led Burke to editing his memorandum as there appeared a notice that Burke would soon publish a letter on the subject to the Secretary of the Board of Agriculture Arthur Young, but he failed to complete it. These fragments were inserted into the memorandum after his death and published posthumously in 1800 as "Thoughts and Details on Scarcity". In it, Burke expounded "some of the doctrines of political economists bearing upon agriculture as a trade". Burke criticised policies such as maximum prices and state regulation of wages and set out what the limits of government should be: That the State ought to confine itself to what regards the State, or the creatures of the State, namely, the exterior establishment of its religion; its magistracy; its revenue; its military force by sea and land; the corporations that owe their existence to its fiat; in a word, to every thing that is "truly and properly" public, to the public peace, to the public safety, to the public order, to the public prosperity. The economist Adam Smith remarked that Burke was "the only man I ever knew who thinks on economic subjects exactly as I do, without any previous communications having passed between us". Writing to a friend in May 1795, Burke surveyed the causes of discontent: "I think I can hardly overrate the malignity of the principles of Protestant ascendency, as they affect Ireland; or of Indianism [i.e. corporate tyranny, as practiced by the British East Indies Company], as they affect these countries, and as they affect Asia; or of Jacobinism, as they affect all Europe, and the state of human society itself. The last is the greatest evil". By March 1796, Burke had changed his mind: "Our Government and our Laws are beset by two different Enemies, which are sapping its foundations, Indianism, and Jacobinism. In some Cases they act separately, in some they act in conjunction: But of this I am sure; that the first is the worst by far, and the hardest to deal with; and for this amongst other reasons, that it weakens discredits, and ruins that force, which ought to be employed with the greatest Credit and Energy against the other; and that it furnishes Jacobinism with its strongest arms against all "formal" Government". For more than a year prior to his death, Burke knew that his stomach was "irrecoverably ruind". After hearing that Burke was nearing death, Fox wrote to Mrs. Burke enquiring after him. Fox received the reply the next day: Mrs. Burke presents her compliments to Mr. Fox, and thanks him for his obliging inquiries. Mrs. Burke communicated his letter to Mr. Burke, and by his desire has to inform Mr. Fox that it has cost Mr. Burke the most heart-felt pain to obey the stern voice of his duty in rending asunder a long friendship, but that he deemed this sacrifice necessary; that his principles continue the same; and that in whatever of life may yet remain to him, he conceives that he must live for others and not for himself. Mr. Burke is convinced that the principles which he has endeavoured to maintain are necessary to the welfare and dignity of his country, and that these principles can be enforced only by the general persuasion of his sincerity. Burke died in Beaconsfield, Buckinghamshire, on 9 July 1797 and was buried there alongside his son and brother. Burke is regarded by most political historians in the English-speaking world as a liberal conservative and the father of modern British conservatism. Burke was utilitarian and empirical in his arguments while Joseph de Maistre, a fellow conservative from the Continent, was more providentialist and sociological and deployed a more confrontational tone in his arguments. Burke believed that property was essential to human life. Because of his conviction that people desire to be ruled and controlled, the division of property formed the basis for social structure, helping develop control within a property-based hierarchy. He viewed the social changes brought on by property as the natural order of events which should be taking place as the human race progressed. With the division of property and the class system, he also believed that it kept the monarch in check to the needs of the classes beneath the monarch. Since property largely aligned or defined divisions of social class, class too was seen as natural—part of a social agreement that the setting of persons into different classes, is the mutual benefit of all subjects. Concern for property is not Burke's only influence. Christopher Hitchens summarises as follows: "If modern conservatism can be held to derive from Burke, it is not just because he appealed to property owners in behalf of stability but also because he appealed to an everyday interest in the preservation of the ancestral and the immemorial". Burke's support for Irish Catholics and Indians often led him to be criticised by Tories. His opposition to British imperialism in Ireland and India and his opposition to French imperialism and radicalism in Europe made it difficult for Whig or Tory to accept Burke wholly as their own. In the 19th century, Burke was praised by both liberals and conservatives. Burke's friend Philip Francis wrote that Burke "was a man who truly & prophetically foresaw all the consequences which would rise from the adoption of the French principles", but because Burke wrote with so much passion, people were doubtful of his arguments. William Windham spoke from the same bench in the House of Commons as Burke had when he had separated from Fox and an observer said Windham spoke "like the ghost of Burke" when he made a speech against peace with France in 1801. William Hazlitt, a political opponent of Burke, regarded him as amongst his three favourite writers (the others being Junius and Rousseau) and made it "a test of the sense and candour of any one belonging to the opposite party, whether he allowed Burke to be a great man". William Wordsworth was originally a supporter of the French Revolution and attacked Burke in "A Letter to the Bishop of Llandaff" (1793), but by the early 19th century he had changed his mind and came to admire Burke. In his "Two Addresses to the Freeholders of Westmorland", Wordsworth called Burke "the most sagacious Politician of his age", whose predictions "time has verified". He later revised his poem "The Prelude" to include praise of Burke ("Genius of Burke! forgive the pen seduced/By specious wonders") and portrayed him as an old oak. Samuel Taylor Coleridge came to have a similar conversion as he had criticised Burke in "The Watchman", but in his "Friend" (1809–1810) had defended Burke from charges of inconsistency. Later in his "Biographia Literaria" (1817), Coleridge hails Burke as a prophet and praises Burke for referring "habitually to "principles". He was a "scientific" statesman; and therefore a "seer"". Henry Brougham wrote of Burke that "all his predictions, save one momentary expression, had been more than fulfilled: anarchy and bloodshed had borne sway in France; conquest and convulsion had desolated Europe. [...] [T]he providence of mortals is not often able to penetrate so far as this into futurity". George Canning believed that Burke's "Reflections" "has been justified by the course of subsequent events; and almost every prophecy has been strictly fulfilled". In 1823, Canning wrote that he took Burke's "last works and words [as] the manual of my politics". The Conservative Prime Minister Benjamin Disraeli "was deeply penetrated with the spirit and sentiment of Burke's later writings". The 19th-century Liberal Prime Minister William Ewart Gladstone considered Burke "a magazine of wisdom on Ireland and America" and in his diary recorded: "Made many extracts from Burke—"sometimes almost divine"". The Radical MP and anti-Corn Law activist Richard Cobden often praised Burke's "Thoughts and Details on Scarcity". The Liberal historian Lord Acton considered Burke one of the three greatest Liberals, along with Gladstone and Thomas Babington Macaulay. Lord Macaulay recorded in his diary: "I have now finished reading again most of Burke's works. Admirable! The greatest man since Milton". The Gladstonian Liberal MP John Morley published two books on Burke (including a biography) and was influenced by Burke, including his views on prejudice. The Cobdenite Radical Francis Hirst thought Burke deserved "a place among English libertarians, even though of all lovers of liberty and of all reformers he was the most conservative, the least abstract, always anxious to preserve and renovate rather than to innovate. In politics he resembled the modern architect who would restore an old house instead of pulling it down to construct a new one on the site". Burke's "Reflections on the Revolution in France" was controversial at the time of its publication, but after his death it was to become his best known and most influential work and a manifesto for Conservative thinking. Two contrasting assessments of Burke also were offered long after his death by Karl Marx and Winston Churchill. In "Das Kapital", Marx wrote: The sycophant—who in the pay of the English oligarchy played the romantic "laudator temporis acti" against the French Revolution just as, in the pay of the North American colonies at the beginning of the American troubles, he had played the liberal against the English oligarchy—was an out-and-out vulgar bourgeois. "The laws of commerce are the laws of Nature, and therefore the laws of God." (E. Burke, l.c., pp. 31, 32) No wonder that, true to the laws of God and Nature, he always sold himself in the best market. In "Consistency in Politics", Churchill wrote: On the one hand [Burke] is revealed as a foremost apostle of Liberty, on the other as the redoubtable champion of Authority. But a charge of political inconsistency applied to this life appears a mean and petty thing. History easily discerns the reasons and forces which actuated him, and the immense changes in the problems he was facing which evoked from the same profound mind and sincere spirit these entirely contrary manifestations. His soul revolted against tyranny, whether it appeared in the aspect of a domineering Monarch and a corrupt Court and Parliamentary system, or whether, mouthing the watch-words of a non-existent liberty, it towered up against him in the dictation of a brutal mob and wicked sect. No one can read the Burke of Liberty and the Burke of Authority without feeling that here was the same man pursuing the same ends, seeking the same ideals of society and Government, and defending them from assaults, now from one extreme, now from the other. The historian Piers Brendon asserts that Burke laid the moral foundations for the British Empire, epitomised in the trial of Warren Hastings, that was ultimately to be its undoing. When Burke stated that "[t]he British Empire must be governed on a plan of freedom, for it will be governed by no other", this was "an ideological bacillus that would prove fatal. This was Edmund Burke's paternalistic doctrine that colonial government was a trust. It was to be so exercised for the benefit of subject people that they would eventually attain their birthright—freedom". As a consequence of this opinion, Burke objected to the opium trade which he called a "smuggling adventure" and condemned "the great Disgrace of the British character in India". A Royal Society of Arts blue plaque commemorates Burke at 37 Gerrard Street now in London's Chinatown. One of Burke's largest and most developed critics was the American political theorist Leo Strauss. In his book "Natural Right and History", Strauss makes a series of points in which he somewhat harshly evaluates Burke's writings. One of the topics that he first addresses is the fact that Burke creates a definitive separation between happiness and virtue and explains that "Burke, therefore, seeks the foundation of government 'in a conformity to our duties' and not in 'imaginary rights of man" Strauss views Burke as believing that government should focus solely on the duties that a man should have in society as opposed to trying to address any additional needs or desires. Government is simply a practicality to Burke and not necessarily meant to function as a tool to help individuals live their best lives. Strauss also argues that in a sense Burke's theory could be seen as opposing the very idea of forming such philosophies. Burke expresses the view that theory cannot adequately predict future occurrences and therefore men need to have instincts that cannot be practised or derived from ideology. This leads to an overarching criticism that Strauss holds regarding Burke which is his rejection of the use of logic. Burke dismisses a widely held view amongst theorists that reason should be the primary tool in the forming of a constitution or contract. Burke instead believes that constitutions should be made based on natural processes as opposed to rational planning for the future. However, Strauss points out that criticising rationality actually works against Burke's original stance of returning to traditional ways because some amount human reason is inherent and therefore is in part grounded in tradition. In regards to this formation of legitimate social order, Strauss does not necessarily support Burke's opinion—that order cannot be established by individual wise people, but exclusively by a culmination of individuals with historical knowledge of past functions to use as a foundation. Strauss notes that Burke would oppose more newly formed republics due to this thought, although Lenzner adds the fact that he did seem to believe that America's constitution could be justified given the specific circumstances. On the other hand, France's constitution was much too radical as it relied too heavily on enlightened reasoning as opposed to traditional methods and values. Burke's religious writing comprises published works and commentary on the subject of religion. Burke's religious thought was grounded in the belief that religion is the foundation of civil society. He sharply criticised deism and atheism and emphasised Christianity as a vehicle of social progress. Born in Ireland to a Catholic mother and a Protestant father, Burke vigorously defended the Anglican Church, but he also demonstrated sensitivity to Catholic concerns. He linked the conservation of a state-established religion with the preservation of citizens' constitutional liberties and highlighted Christianity's benefit not only to the believer's soul, but also to political arrangements. The statement that "The only thing necessary for the triumph of evil is for good men to do nothing" is often attributed to Burke despite the debated origin of this quote. In 1770, it is known that Burke wrote in "Thoughts on the Cause of the Present Discontents": In 1867, John Stuart Mill later made a similar statement in an inaugural address delivered before the University of St. Andrews:
https://en.wikipedia.org/wiki?curid=10030
Early music Early music generally comprises Medieval music (500–1400) and Renaissance music (1400–1600), but can also include Baroque music (1600–1750). Early music is a broad musical era in the history of Western art music. Interpretations of historical scope of "early music" vary. The original Academy of Ancient Music formed in 1726 defined "Ancient" music as works written by composers who lived before the end of the 16th century. Johannes Brahms and his contemporaries would have understood Early music to range from the High Renaissance and Baroque, while some scholars consider that Early music should include the music of ancient Greece or Rome before 500 AD (a period that is generally covered by the term Ancient music). Music critic Michael Kennedy excludes Baroque, defining Early music as "musical compositions from [the] earliest times up to and including music of [the] Renaissance period". Musicologist Thomas Forrest Kelly considers that the essence of Early music is the revival of "forgotten" musical repertoire and that the term is intertwined with the rediscovery of old performance practice. According to the UK's National Centre for Early Music, the term "early music" refers to both a repertory (European music written between 1250 and 1750 embracing Medieval, Renaissance and the Baroque) – and a historically informed approach to the performance of that music. Today, the understanding of "Early music" has come to include "any music for which a historically appropriate style of performance must be reconstructed on the basis of surviving scores, treatises, instruments and other contemporary evidence." In the later 20th century there was a resurgence of interest in the performance of music from the Medieval and Renaissance eras, and a number of instrumental consorts and choral ensembles specialising in Early music repertoire were formed. Groups such as the Tallis Scholars, the Early Music Consort and the Taverner Consort and Players have been influential in bringing Early music to modern audiences through performances and popular recordings. The revival of interest in Early music has given rise to a scholarly approach to the performance of music. Through academic musicological research of music treatises, urtext editions of musical scores and other historical evidence, performers attempt to be faithful to the performance style of the musical era in which a work was originally conceived. Additionally, there has been a rise in the use of original or reproduction period instruments as part of the performance of Early music, such as the revival of the harpsichord or the viol. The practice of "historically informed performance" is nevertheless dependent on stylistic inference. According to Margaret Bent, Renaissance notation is not as prescriptive as modern scoring, and there is much that was left to the performer's interpretation: "Renaissance notation is under-prescriptive by our standards; when translated into modern form it acquires a prescriptive weight that overspecifies and distorts its original openness. Accidentals … may or may not have been notated, but what modern notation requires would then have been perfectly apparent without notation to a singer versed in counterpoint".
https://en.wikipedia.org/wiki?curid=10033
Euroscepticism Euroscepticism, also known as EU-scepticism, means criticism of the European Union (EU) and European integration. It ranges from those who oppose some EU institutions and policies and seek reform (soft Euroscepticism), to those who oppose EU membership outright and see the EU as unreformable (hard Euroscepticism or anti-European Unionism/anti-EUism). The opposite of Euroscepticism is known as pro-Europeanism, or European Unionism. Euroscepticism should not be confused with anti-Europeanism, which is a dislike of European culture and European ethnic groups by non-Europeans. The main sources of Euroscepticism have been beliefs that integration undermines national sovereignty and the nation state, that the EU is elitist and lacks democratic legitimacy and transparency, that it is too bureaucratic and wasteful, that it encourages high levels of migration, or perceptions that it is a neoliberal organisation serving the business elite at the expense of the working class, responsible for austerity and driving privatization. Euroscepticism is found in groups across the political spectrum, both left-wing and right-wing and is often found in populist parties. Although they criticise the EU for many of the same reasons, Eurosceptic left-wing populists focus more on economic issues (such as the European debt crisis and TTIP) while Eurosceptic right-wing populists focus more on nationalism and immigration (such as the European migrant crisis). The rise in radical right-wing parties since the 2000s is strongly linked to a rise in Euroscepticism. Eurobarometer surveys of EU citizens show that trust in the EU and its institutions has declined strongly since a peak in 2007. Since then, it has been consistently below 50%. A 2009 survey showed that support for EU membership was lowest in the United Kingdom (UK), Latvia and Hungary. By 2016, the countries viewing the EU most unfavourably were the UK, Greece, France and Spain. A referendum on continued EU membership was held in the UK in 2016 which resulted in a 51.9% vote in favour of leaving the EU. The UK officially left the EU on 31 January 2020. Since 2015, trust in the EU has risen slightly in most EU countries as a result of falling unemployment rates and accelerating economic growth. Post 2019 election survey "Eurobarometer" report showed that 68% citizens support the European Union, the highest level since 1983; at the same time sentiment among Europeans that things are not going in the right direction in both the EU and in their own countries had increased to 50%. While having some overlaps, Euroscepticism and anti-Europeanism are different. Anti-Europeanism has always had a strong influence in American culture and American exceptionalism, which sometimes sees Europe on the decline or as a rising rival power, or both. Some aspects of Euroscepticism in the United Kingdom have been mirrored by U.S. authors. There can be considered to be several different types of Eurosceptic thought, which differ in the extent to which adherents reject integration between member states of the European Union (EU) and in their reasons for doing so. Aleks Szczerbiak and Paul Taggart described two of these as hard and soft Euroscepticism. At the same time, some have said that there is no clear line between the presumed 'hard' and 'soft' Euroscepticism. Kopecky and Mudde have said that if the demarcation line is the number of and which policies a party opposes, then the question arises of how many must a party oppose and which ones should a party oppose that makes them 'hard' Eurosceptic instead of 'soft'. According to Taggart and Szczerbiak, hard Euroscepticism (also called anti-EU-ism) is "a principled opposition to the EU and European integration and therefore can be seen in parties who think that their countries should withdraw from membership, or whose policies towards the EU are tantamount to being opposed to the whole project of European integration as it is currently conceived." The Europe of Freedom and Direct Democracy group in the European Parliament, typified by such parties as the Brexit Party and UK Independence Party (UKIP), displays hard Euroscepticism. In western European EU member countries, hard Euroscepticism is currently a characteristic of many anti-establishment parties. Some hard Eurosceptics prefer to call themselves "Eurorealists" rather than "sceptics", and regard their position as pragmatic rather than ‘in principle’. Additionally, Tony Benn, a left-wing Labour Party MP who fought against European integration in 1975 by opposing membership of the European Communities in that year's referendum on the issue, emphasised his opposition to xenophobia and his support of democracy, saying: "My view about the European Union has always been not that I am hostile to foreigners, but that I am in favour of democracy [...] I think they're building an empire there, they want us to be a part of their empire and I don't want that." The Czech president Václav Klaus rejected the term ’Euroscepticism’ for its purported negative undertones, saying (at a meeting in April 2012) that the expressions for a Eurosceptic and their opponent should be "a Euro-realist" and someone who is "Euro-naïve", respectively. François Asselineau of the French Popular Republican Union has criticised the use of the term 'sceptic' to describe hard Eurosceptics, and would rather advocate the use of the term "Euro opponent". He believes the use of the term 'sceptic' for soft Eurosceptics to be correct, since other Eurosceptic parties in France are "merely criticising" the EU without taking into account the fact that the Treaty of Rome can only be modified with a unanimous agreement of all the EU member states, something he considers impossible to achieve. Soft Euroscepticism is support for the existence of, and membership of, a form of European Union, but with opposition to specific EU policies; or, in Taggart's and Szczerbiak's words, "where there is NOT a principled objection to European integration or EU membership but where concerns on one (or a number) of policy areas lead to the expression of qualified opposition to the EU, or where there is a sense that 'national interest' is currently at odds with the EU's trajectory." The European Conservatives and Reformists group, typified by centre-right parties such as Czech Civic Democratic Party, along with the European United Left–Nordic Green Left which is an alliance of the left-wing parties in the European Parliament, display soft Euroscepticism. Some scholars consider the gradual difference in terminology between 'hard' and 'soft' Euroscepticism inadequate to accommodate the large differences in terms of political agenda. Therefore, "hard Euroscepticism" has also been referred to as "Europhobia" as opposed to mere "Euroscepticism". Other alternative names for 'hard' and 'soft' Euroscepticism include, respectively, "withdrawalist" and "reformist" Euroscepticism. A survey in , conducted by TNS Opinion and Social on behalf of the European Commission, showed that, across the EU as a whole, those with a positive image of the EU are down from a high of 52% in 2007 to 37% in autumn 2015; this compares with 23% with a negative image of the EU, and 38% with a neutral image. About 43% of Europeans thought things were "going in the wrong direction” in the EU, compared with 23% who thought things were going "in the right direction" (11% "don't know"). About 32% of EU citizens tend to trust the EU as an institution, and about 55% do not tend to trust it (13% "don't know"). Distrust of the EU was highest in Greece (81%), Cyprus (72%), Austria (65%), France (65%) and Germany, the United Kingdom (UK) and the Czech Republic (all 63%). Overall, more respondents distrusted their own government (66%) than they distrusted the EU (55%). Distrust of national government was highest in Greece (82%), Slovenia (80%), Portugal (79%), Cyprus (76%) and France (76%). A Eurobarometer survey carried out four days prior to and six days after the U.S. presidential election in November 2016 revealed that the surprise victory of Donald Trump caused an increase in the popularity of the European Union in Europe. The increase was strongest among the political right and among respondents who perceived their country as economically struggling. A survey carried out in April 2018 for the European Parliament by Kantar Public consulting found that support for the EU was "the highest score ever measured since 1983". Support for the EU was up in 26 out of 28 EU countries, the exceptions being Germany and the UK, where support had dropped by about 2% since the previous survey. Almost half (48%) of the 27,601 EU citizens surveyed agreed that their voice counted in the EU, up from 37% in 2016, whereas 46% disagreed with this statement. Two-thirds (67%) of respondents felt that their country had benefited from EU membership and 60% said that being part of the bloc was a good thing, as opposed to 12% who felt the opposite. At the height of the EU's financial and economic crises in 2011, just 47% had been of the view that EU membership was a good thing. Support for EU membership was greatest in Malta (93%), Ireland (91%), Lithuania (90%), Poland (88%), Luxembourg (88%), Estonia (86%) and Denmark (84%) and lowest in Greece (57%), Bulgaria (57%), Cyprus (56%), Austria (54%), the UK (53%) and Italy (44%). When asked which issues should be a priority for the European Parliament, survey respondents picked terrorism as the most pressing topic of discussion, ahead of youth unemployment and immigration. Not all countries shared the same priorities. Immigration topped the list in Italy (66% of citizens surveyed considered it a priority issue), Malta (65%) and Hungary (62%) but fighting youth unemployment and support for economic growth were top concerns in Spain, Greece, Portugal, Cyprus and Croatia. Social protection of citizens was the top concern for Dutch, Swedish and Danish respondents. The April 2019 Eurobarometer showed that despite the challenges of the past years – and in cases such as the ongoing debate surrounding Brexit, possibly even because of it – the European sense of togetherness had not weakened, with 68% of respondents across the EU27 believing that their countries have benefited from being part of the EU, a historically high level since 1983. On the other hand, more Europeans (27%) were uncertain and saw the EU as "neither a good thing nor a bad thing" – an increase in 19 countries. Despite the overall positive attitude towards the EU, but in line with the uncertainty expressed by a growing number of Europeans, the feeling that things were not going in the right direction in both the EU and in their own countries had increased to 50% on EU average since September 2018. A study analysed voting records of the Fifth European Parliament and ranked groups, concluding: "Towards the top of the figure are the more pro-European parties (PES, EPP-ED, and ALDE), whereas towards the bottom of the figure are the more anti-European parties (EUL/NGL, G/EFA, UEN and EDD)." In 2004, 37 Members of the European Parliament (MEPs) from the UK, Poland, Denmark and Sweden founded a new European Parliament group called "Independence and Democracy" from the old Europe of Democracies and Diversities (EDD) group. The main goal of the ID group was to reject the proposed Treaty establishing a constitution for Europe. Some delegations within the group, notably that from UKIP, also advocated the complete withdrawal of their country from the EU, while others only wished to limit further European integration. The elections of 2009 saw a significant fall in support in some areas for Eurosceptic parties, with all such MEPs from Poland, Denmark and Sweden losing their seats. In the UK, the Eurosceptic UKIP achieved second place in the election, finishing ahead of the governing Labour Party, and the British National Party (BNP) won its first-ever two MEPs. Although new members joined the ID group from Greece and the Netherlands, it was unclear whether the group would reform in the new parliament. The ID group did reform, as the Europe of Freedom and Democracy (EFD) and is represented by 32 MEPs from nine countries. The elections of 2014 saw a big anti-establishment vote in favour of Eurosceptic parties, which took around a quarter of the seats available. Those that won their national elections included: UKIP in the UK (the first time since 1906 that a party other than Labour or the Conservatives had won a national vote), the National Front in France, the People's Party in Denmark and Syriza in Greece. Second places were taken by Sinn Féin in Ireland and the Five Star Movement in Italy. Herman Van Rompuy, the President of the European Council, agreed following the election to re-evaluate the economic area's agenda and to launch consultations on future policy areas with the 28 member states. The elections of 2019 saw the centre-left and centre-right parties suffer significant losses including losing their overall majority, while green, pro-EU liberal, and some Eurosceptic right wing parties saw significant gains. Those that won their national elections included: The Brexit Party in the UK (which was only launched on 12 April 2019 by former UKIP leader Nigel Farage), the National Rally of France (formerly the National Front party until June 2018), Fidesz in Hungary, Lega in Italy, and Law and Justice in Poland. There were also notable falls in support for the Danish People's Party (previously topped the 2014 European election). Whilst Vox got elected with 3 seats, Spain's first Eurosceptic party and Belgium's Vlaams Belang rallied to gain second place after its poor 2014 result. The Freiheitliche Partei Österreichs (FPÖ), established in 1956, is a right-wing populist party that mainly attracts support from young people and workers. In 1989, it changed its stance over the EU to Euroscepticism. It opposed Austria joining the EU in 1994, and opposed the introduction of the euro in 1998. The party would like to leave the EU if it threatens to develop into a country, or if Turkey joins. The FPÖ received 20–27% of the national vote in the 1990s, and more recently received 17.5% in 2008. Following the 2017 Austrian legislative election, it has 51/183 National Council seats, 16/62 Federal Council seats, and 4/19 European Parliament seats. The Bündnis Zukunft Österreich (BZÖ), established in 2005, is a socially conservative party that has always held Eurosceptic elements. In 2011 the party openly supported leaving the eurozone, and in 2012 it announced that it supported a full withdrawal from the European Union. The party has also called upon a referendum on the Lisbon Treaty. In polls it currently receives around 10%–15%, although in one state it did receive 45% of the vote in 2009. Since the 2017 election, it has 0/183 National Council seats, 0/62 Federal Council seats, and 0/19 European Parliament seats. Team Stronach, established in 2012, has campaigned to reform the European Union, as well as to replace the euro with an Austrian Euro. In 2012, it regularly received 8–10% support in national polls. Politicians from many different parties (including the Social Democratic Party and the BZÖ) as well as previous independents switched their allegiances to the new party upon creation. In two local elections in March 2013, it won 11% of the vote in Carinthia, and 10% of the vote in Lower Austria. It dissolved in 2017 Ewald Stadler, a former member of FPÖ (and later of BZÖ) was very Eurosceptic, but in 2011 became a member of the European Parliament due to the Lisbon Treaty. Before Stadler accepted the seat, this led to heavy critics by Jörg Leichtfried (SPÖ) "Stadler wants to just rescue his political career" because Stadler before mentioned he would never accept a seat as MEP if this was only due to the Lisbon Treaty. On 23 December 2013 he founded a conservative and Eurosceptic party called The Reform Conservatives, although it has been inactive since June 2016. In the 2014 European Parliament election, the FPÖ increased its vote to 19.72% (up 7.01%), gaining 2 new MEPs, making a total of 4; the party came third, behind the ÖVP and the SPÖ. EU-STOP (the electoral alliance of the EU Withdrawal Party and the Neutral Free Austria Federation) polled 2.76%, gaining no seats, and the Reform Conservatives 1.18%, with Team Stronach putting up no candidates. In the 2019 European Parliament election, the FPÖ came 3rd with 17.2% of the vote which was only slightly down on 2014 despite a scandal allegedly promising public contracts to a woman posing as a Russian backer. This precipitated the collapse of the ruling coalition and a new election being called. According to Eurostat, in the fall of 2018, 44% of Belgium people stated that they did not trust the European Union. The main Eurosceptic party in Belgium is Vlaams Belang which is active in the Dutch-speaking part of Belgium. In the 2014 European Parliament election, Belgium's Vlaams Belang lost over half of its previous vote share, polling 4.26% (down 5.59%) and losing 1 of its 2 members of the European Parliament. Despite the presence of Eurosceptic parties in Belgium, their weight is relatively low, as Belgium is predominantly Europeanist. In 2019, Belgium's Vlaams Belang stated in its program for the European election that it opposes the creation of a European State, would like to change the Economic and monetary union, and to end the Schengen Area, and refuses the accession of Turkey to the European Union. More widely, the euro-sceptic arguments of the Vlaams Belang are based on four pillars: During the 2019 European election and the 2019 Belgian Federal Election "Vlaams Belang" made substantial gains in both and polled in second place in Flemish region. At the beginning of 2019 the Party was enrolled in the group of European Alliance of People and Nations in the European Parliament. The New Flemish Alliance (N-VA) is a soft euro-sceptic party in the Dutch-speaking region of Belgium which has obtained 26.83% of the votes or 4 seats of the Dutch-language college out of 12 (21 MEPs for Belgium) in the 2014 European Parliament election. In April 2019 it stood in European Conservatives and Reformists of the European Parliament, and can be considered a moderate Eurosceptic party. In the French-speaking part of Belgium there are four Eurosceptic parties. The first one is Nation Movement, a far-right party which was a member of the Alliance for Peace and Freedom in the European Parliament. The second one is National Front, also a far right party which criticizes the European bureaucracy, intends to guarantee and preserve national independence and freedom in a liberated Europe; it also reaffirms the Christian roots of Europe. The third one is the People's Party, classified as right or extreme right. In its program for the European election of 2019 the People's Party proposes to abolish the European Commission, reduce the number of European parliamentarians and fight against the worker-posted directive. For this party, European Union must be led by a president elected by universal suffrage with clear but limited competences. It also wants to renegotiate the European Union treaties, restrict the judicial activism of the European Court of Human Rights. It declares itself against the Global Compact for Migration. The last one is the . In early 2019, the Party aims to reduce the powers of the European Commission, to abolish the Common Agricultural Policy, to abandon common defense projects, to simplify the exit procedure of the European Union, to reject federalism and to forbid the European Union to direct economic, fiscal or social policy, Finally, the Worker's Party of Belgium is an electoral and unitary party. It also intends to revise the European treaties considered too liberal. One of the Party's currencies is "The left that stings, against the Europe of money". Parties with mainly Eurosceptic views are Union of Communists in Bulgaria, NFSB, Attack, and VMRO – BND, which is a member of the Eurosceptic European Conservatives and Reformists). Bulgaria's Minister of Finance, Simeon Djankov, stated in 2011 that ERM II membership to enter the Euro zone would be postponed until after the Eurozone crisis had stabilised. In the 2014 European Parliament election Bulgaria remained overwhelmingly pro-EU, with the Eurosceptic Attack party receiving 2.96% of the vote, down 9%, with the splinter group National Front for the Salvation of Bulgaria taking 3.05%; neither party secured any MEPs. A coalition between VMRO – BND and Bulgaria Without Cencorship secured an MEP position for Angel Dzhambazki from IMRO, who is a soft Eurosceptic. Followers of Eurosceptic Attack tore down and trampled the European flag on 3 March 2016 at a meeting of the party in the Bulgarian capital Sofia, dedicated to the commemoration of the 138th anniversary of the liberation of Bulgaria from the Ottoman Empire. In the 2019 European Parliament election, Bulgaria remained overwhelmingly pro-EU with the ruling centre-right Gerb party winning with 30.6%, against 26% for the socialist BSP Parties with Eurosceptic views are mainly small right-wing parties like Croatian Party of Rights, Croatian Party of Rights dr. Ante Starčević, Croatian Pure Party of Rights, Autochthonous Croatian Party of Rights, Croatian Christian Democratic Party and Only Croatia – Movement for Croatia. The only parliamentary party that is vocally Eurosceptic is the Human Shield that won 5 out of 151 seats at the 2016 parliamentary election. Their position is generally considered to waver between hard and soft Euroscepticism; it requests thorough reform of the EU so that all member states would be perfectly equal. In the 2019 European Parliament election, the Human Shield gained its first seat in the European Parliament with 5.7% of the vote putting it in 5th place. In May 2010, the Czech president Václav Klaus said that they "needn't hurry to enter the Eurozone". Petr Mach, an economist, a close associate of president Václav Klaus and a member of the Civic Democratic Party between 1997 and 2007, founded the Free Citizens Party in 2009. The party aims to mainly attract dissatisfied Civic Democratic Party voters. At the time of the Lisbon Treaty ratification, they were actively campaigning against it, supported by the president Vaclav Klaus, who demanded opt-outs such as were granted to the United Kingdom and Poland, unlike the governing Civic Democratic Party, who endorsed it in the Chamber of Deputies. After the treaty has been ratified, Mach's party is in favour of withdrawing from the European Union completely. In the 2014 European Parliament election, the Free Citizens Party won one mandate and allied with UKIP in the Europe of Freedom and Direct Democracy (EFD). The 2017 Czech legislative election brought into Parliament two soft eurosceptic parties: centre-right Civic Democratic Party (ODS) (11%) and far-left Communist Party of Bohemia and Moravia (KSČM) (8%). And one hard eurosceptic party, far-right Freedom and Direct Democracy (SPD) (11%). An April 2016 survey by the CVVM Institute indicated that 25% of Czechs were satisfied with EU membership, down from 32% the previous year. Dividends worth CZK 270 billion were paid to the foreign owners of Czech companies in 2017, which has become a political issue in the Czech Republic. In the 2019 European Parliament election, the Civic Democratic Party saw its vote share rise to 14.5% and its seats doubled from 2 to 4. The Freedom and Direct Democracy party took 2 seats with 9.14% of the vote. KSČM dropped 2 seats leaving it with only one and a vote share of 6.9% Parties with mainly Eurosceptic views in Cyprus are New Internationalist Left, the Progressive Party of Working People, Committee for a Radical Left Rally and ELAM. In the 2019 European Parliament election, there was little change politically - the conservatives won narrowly, the ruling DISY taking two seats with 29%, followed by socialist AKEL (27.5%, two seats) with no seats taken by Eurosceptic parties. The People's Movement against the EU only takes part in European Parliament elections and has one member in the European Parliament. The soft Eurosceptic June Movement, originally a split-off from the People's Movement against the EU, existed from 1992 to 2009. In the Danish Parliament, the Red-Green Alliance has withdrawal from the EU as a policy. The Danish People's Party also advocates withdrawal, but says it supports some EU structures such as the internal market, and supported the EU-positive Liberal-Conservative coalition between 2001 and 2011 and again from 2015 to present. The Socialist People's Party, minorities within the Social Liberal Party and Social Democratic Party, and some smaller parties were against accession to the European Union in 1972. Still in 1986, these parties advocated a no vote in the Single European Act referendum. Later, the Social Liberal Party changed to a strongly EU-positive party, and EU opposition within the Social Democratic Party faded. The Socialist People's Party were against the Amsterdam Treaty in 1998 and Denmark's joining the euro in 2000, but has become increasingly EU-positive, for example when MEP Margrete Auken left the European United Left–Nordic Green Left and joined The Greens–European Free Alliance in 2004. In the 2014 European Parliament election, the Danish People's Party came first by a large margin with 26.6% of the vote, gaining 2 extra seats for a total of 4 MEPs. The People's Movement against the EU polled 8.1%, retaining its single MEP. In the 2019 European Parliament election, the Danish People's Party lost around two-thirds of their previous vote share dropping from 4 seats to just 1. The People's Movement against the EU lost their seat and the Red-Green Alliance got one seat. The Independence Party and Centre Party were against accession to the EU, but only the Independence Party still wants Estonia to withdraw from the EU. The Conservative People's Party (EKRE) also has some Eurosceptic policies and increased its vote share from 4% in 2014 to 12.7% in the 2019 European Elections winning one seat. The largest Eurosceptic party in Finland is the Finns Party. In the European Parliament election, 2014, the Finns Party increased their vote share by 3.1% to 12.9%, adding a second MEP. With their 39 seats, the Finns Party are also the second-biggest party in the 200-seat Finnish Eduskunta. In Eurobarometer 77 (fieldwork in Spring 2012), 41% of Finns trusted the European Union (EU-27 average: 31%), 51% trusted The European Parliament (EU-27average: 40%), and 74% were in favour of the euro currency (EU-27 average: 52%). In the European Parliament election, 2019, the Finns Party increased their vote share slightly from 12.9% to 13.8% and retained their 2 seats. In France there are multiple parties that are Eurosceptic to different degrees, varying from advocating less EU intervention in national affairs, to advocating outright withdrawal from the EU and the Eurozone. These parties belong to all sides of the political spectrum, so the reasons for their Euroscepticism may differ. In the past many French people appeared to be uninterested in such matters, with only 40% of the French electorate voting in the 2009 European Parliament elections. Right-wing Eurosceptic parties include the Gaullist "Debout la République", and "Mouvement pour la France", which was part of Libertas, a pan-European Eurosceptic party. In the 2009 European Parliament elections, Debout la République received 1.77% of the national vote, and Libertas 4.8%. In a similar way to some moderate parties, the French right and far-right in general are naturally opposed to the EU, as they criticise France's loss of political and economic sovereignty to a supranational entity. Some of these hard Eurosceptic parties include the Popular Republican Union and the Front National (FN). Front National and Popular Republican Union both seek France's withdrawal from the EU and the euro, although Popular Republican Union also seeks France's withdrawal from NATO. The FN received 33.9% of the votes in the 2017 French presidential election, making it the largest Eurosceptic party in France. Eurosceptic parties on the left in France tend to criticise what they see as the neoliberal agenda of the EU, as well as the elements of its structure which are undemocatic and seen as top-down. These parties include the "Parti de Gauche" and the French Communist Party, which formed the "Front de Gauche" for the 2009 European Parliament elections and received 6.3% of the votes. The leader of the Left Front defends a complete reform of the Monetary Union, rather than the withdrawal of France from the Eurozone. Some of the major far-left Eurosceptic parties in France include the New Anticapitalist Party which received 4.8% and Lutte Ouvrière which received 1.2%. The Citizen and Republican Movement, a left-wing Eurosceptic and souverainist party, have not participated in any elections for the European Parliament. The party "Chasse, Pêche, Nature & Traditions", is an agrarianist Eurosceptic party that says it is neither left nor right. In the European Parliament election, 2014, the National Front won the elections with 24.85% of the vote, a swing of 18.55%, winning 24 seats, up from 3 previously. The former French President François Hollande had called for the EU to be reformed and for a scaling back of its power. In the European Parliament election, 2019, the renamed National Rally won the elections with 23.31% of the vote, winning 22 seats, down from 23 previously when their vote share was 24.86%. The Alternative for Germany (AfD) is Germany's largest Eurosceptic party. It has been elected into the German Parliament with 94 seats in September 2017. Initially the AfD was a soft Eurosceptic party, that considered itself pro-Europe and pro-EU, but opposed the euro, which it believed had undermined European integration. In the European Parliament election, 2014, the Alternative for Germany came 5th with 7% of the vote, winning 7 seats and is a member of the Eurosceptic European Conservatives and Reformists. The Alternative for Germany went on to take seats in three state legislatures in the Autumn of 2014. The party became purely Eurosceptic in 2015, when a split occurred in the party, leading to Frauke Petry's leadership and a more hard-line approach to the European Union. In July 2015 a split from AfD created a new soft Eurosceptic party called Alliance for Progress and Renewal. In the European Parliament election, 2019, the Alternative for Germany increased their vote share from 7.04% and 7 seats to 10.79% and 11 seats. Golden Dawn, Communist Party of Greece (KKE), Greek Solution, ANEL, Course of Freedom, Popular Unity, and LAOS have been the main Eurosceptic parties in Greece. According to the London School of Economics, Greece used to be the second most Eurosceptic country in the European Union, with 50% of Greeks thinking that their country has not benefited at all from the EU (only behind the UK). Meanwhile, 33% of Greeks viewed Greek membership in EU as a good thing, marginally ahead of the UK. 81% of Greeks felt that the EU was going in the wrong direction. These figures represented a major increase in Euroscepticism in Greece since 2009. In June 2012, the Eurosceptic parties in Greece that were represented in the parliament before the Election in January 2015 (ANEL, Golden Dawn, KKE) got 45.8% of the votes and 40.3% of the seats in the parliament. In the legislative election of January 2015 the pro-European (left and right-wing) parties (ND, PASOK, Potami, KIDISO, EK and Prasinoi-DIMAR) got 43.28% of the votes. The Eurosceptic parties got 54.64%. The Eurosceptic left (KKE, ANTARSYA-MARS and KKE (M–L)/M–L KKE) got 42.58% of the votes and the Eurosceptic right (Golden Dawn, ANEL and LAOS) got 12.06% of the votes, with Syriza ahead with 36.34%. The Eurosceptic parties got 194 seats in the new parliament and the pro-EU parties got 106 seats. According to the polls conducted in June and July 2015 (12 polls), the Eurosceptic left would get on average 48.03% (excluding extraparliamentary parties as ANTARSYA-MARS and KKE (m–l)/ML-KKE), the parliamentary pro-EU parties (Potami, New Democracy and PASOK) would get 33.82%, the extra-parliamentary (not represented in the Hellenic Parliament) pro-EU parties (KIDISO and EK) would get 4.44% and the Eurosceptic right would get 10.2% (excluding extraparliamentary parties, such as LAOS, not displayed on recent opinion polls). The soft Eurosceptic parties would get 42.31%, the hard Eurosceptic parties (including KKE, ANEL and Golden Dawn) would get 15.85%, and the pro-EU parties (including extra-parliamentary parties displayed on opinion polls) would get 38.27% of the votes. In the European Parliament election, 2014, Syriza won the election with 26.58% of the vote (a swing of 21.88%) taking 6 seats (up 5), with Golden Dawn coming 3rd taking 3 seats, the Communist Party taking 2 seats and the Independent Greeks gaining their first ever seat. Syriza's leader Tsipras said he's not anti-European and does not want to leave the euro. According to "The Economist", Tsipras is willing to negotiate with Greece's European partners, and it is believed a Syriza victory could encourage radical leftist parties across Europe. Alexis Tsipras vowed to reverse many of the austerity measures adopted by Greece since a series of bailouts began in 2010, at odds with the Eurogroup's positions. The government coalition in Greece was composed by Syriza and ANEL (right-wing hard Eurosceptic party, led by Panos Kammenos, who is the current Minister of Defence). Euroscepticism has softened in Greece as the economy improved. According to a research in early 2018, 68% of Greeks judge as positive the participation of Greece in the EU (instead of 53.5% in 2017). In the European Parliament election, 2019, the New Democracy movement, beat the ruling left-wing Syriza formation with 33.12% and 23.76% of the vote respectively, maintaining Syriza's 6 seats and prompting the Prime Minister Alexis Tsipras to call a legislative election on 7 July 2019. In this election, which was won by ND, the pro-European parties (ND, SYRIZA, KINAL, MeRA25, and the extra-parliamentary Union of Centrists and Recreate Greece) got 84.9% of the vote and the Eurosceptic parties (KKE, Greek Solution, the extraparliamentary Golden Dawn and a host of other small mainly left-wing parties) got 15.1%. That drastic change in the balance is mostly the result of SYRIZA abandoning Euroscepticism. Viktor Orbán is the soft Eurosceptic Prime Minister of Hungary for the national-conservative Fidesz Party. Another Eurosceptic party in Hungary is Jobbik, a radical, xenophobic and far-right party. In Hungary 39% of the population have a positive image of the EU, 20% have a negative image, and 40% neutral (1% "Don't know"). In the 2014 Hungarian parliamentary election, Fidesz got 44.54% of the votes, Jobbik got 20.54% of the votes and the communist Hungarian Workers' Party got 0.58% of the votes. Thus, Eurosceptic parties in Hungary obtained 65.66% of the votes, one of the highest figures in Europe. The green-liberal Politics Can Be Different (Lehet Más a Politika, LMP) classifies as a soft or reformist Eurosceptic party given its self-professed "euro-critical" stance. During the European parliamentary campaign of 2014 party Co-President András Schiffer described LMP as having a pronounced pro-integration position on environmental, wage and labour policy as supporting member state autonomy on the self-determination of local communities concerning land resources. So as to combat the differentiated integration of the multi-speed Europe which discriminates against Eastern and Southern member states, LMP would like to initiate an eco-social market economy within the union. In the European Parliament election, 2019, Fidesz consolidated their position by increasing their vote share to 51.48% and adding a seat to take their tally to 13. Eurosceptic Jobbik dropped to 6.34% of the votes, losing 2 of its 3 seats. The Momentum Movement, a newly created pro-European party, came 3rd with 9.93% of the vote, with the strongly pro-European Democratic Coalition coming second with 16.05% of the vote. Euroscepticism is a minority view in Ireland, with opinion polls from 2016 to 2018 indicating growing support for EU membership, moving from 70% to 92% in that time. The Irish people initially voted against ratifying the Nice and Lisbon Treaties. Following renegotiations, second referendums on both were passed with approximately 2:1 majorities in both cases. Some commentators and smaller political groups questioned the validity of the Irish Government's decision to call second referendums. The left-wing Irish republican party Sinn Féin expresses soft Eurosceptic positions on the current structure of the European Union and the direction in which it is moving. The party expresses, "support for Europe-wide measures that promote and enhance human rights, equality and the all-Ireland agenda", but has a "principled opposition" to a European superstate. In its manifesto for the 2015 UK general election, Sinn Féin pledged that the party would campaign for the UK to stay within the EU. In the last European Parliament election in 2014, Sinn Féin won 3 seats coming second in seats and third in votes with 19.5% of the vote up 8.3%. The Trotskyist organisation, the Socialist Party, supports Ireland leaving the EU and supported the Brexit result. It argues that the European Union is institutionally capitalist and neoliberal. The Socialist Party campaigned against the Lisbon and Nice Treaties and favours the foundation of an alternative Socialist European Union. The strongly pro-European Fine Gael won the 2019 European election increasing their vote share to 29.6% and increasing their seats from 3 to 4. The Five Star Movement (M5S), an anti-establishment movement founded by comedian Beppe Grillo, originally set itself out as a Eurosceptic party. The M5S received 25.5% of vote in the 2013 general election, becoming the largest anti-establishment and Eurosceptic party in Europe. The party used to advocate a non-binding referendum on the withdrawal of Italy from the Eurozone (but not from the European Union) and the return to the lira. This policy was rejected in 2018, and the M5S's leader has since stated that the "European Union is the Five Star Movement's home", clarifying that the party wants Italy to stay in the EU, even though it remains critical of some of its treaties. The M5S's popular support is distributed all across Italy: in the 2018 general election the party won 32.7% of the popular vote nation-wide, and was particularly successful in central and southern Italy. A party that retains a Eurosceptic identity is Lega Nord, a regionalist movement led by Matteo Salvini favouring Italy's exit from the Eurozone and the re-introduction of the lira. When in government, LN approved the Treaty of Lisbon. The party won 6.2% of the vote in the 2014 European Parliament elections, but two of its leading members are presidents of Lombardy and Veneto (where LN gained 40.9% of the vote in 2015). In the 2014 European Parliament election the Five Star Movement came second, with 17 seats and 21.2% of the vote after contesting EP seats for the first time. Lega Nord had five seats and The Other Europe with Tsipras had three seats. Other minor Eurosceptic organizations include right-wing political parties (e.g., Brothers of Italy, Tricolour Flame, New Force, National Front, CasaPound, National Movement for Sovereignty, the No Euro Movement), far-left political parties (e.g., the Communist Party of Marco Rizzo, the Italian Communist Party) and other political movements (e.g., the Sovereignist Front, MMT Italy). In addition, the European Union is criticized (especially for the austerity and the creation of the euro) by some left-wing thinkers, like the trade unionist Giorgio Cremaschi and the journalist Paolo Barnard, and some academics, such as the economists Alberto Bagnai and Vladimiro Giacchè, the philosopher Diego Fusaro and the mathematician Marino Badiale. According to the Standard Eurobarometer 87 conducted by the European Commission in spring 2017, 48% of Italians tend not to trust the European Union compared to 36% of Italians who do. In the 2019 European election, the Italian Eurosceptic-souverainist right-wing, represented in large part by Lega Nord, obtained negative results and no presidency in the committees of the European Parliament; on the contrary pro-EU Italian Democratic Party member, David Sassoli, was elected President of the European Parliament. Despite its political alliance with Lega for Conte Cabinet, Five Star Movement voted Ursula von der Leyen, member of pro-EU Christian Democratic Union of Germany, as President of the European Commission. The COVID-19 pandemic also prompted a eurosceptic surge in Italy, as Italians became dissatisfied with the EU's coronavirus response. The National Alliance (For Fatherland and Freedom/LNNK/All for Latvia!), Union of Greens and Farmers and For Latvia from the Heart are parties that are described by some political commentators as bearing soft Eurosceptic views. A small hard Eurosceptic party exists, but it has failed to gain any administrative seats throughout history of its existence. The Order and Justice party has mainly Eurosceptic views. The Alternative Democratic Reform Party is a soft Eurosceptic party. It is a member of the Alliance of European Conservatives and Reformists. The Labour Party was not in favour of Malta entering the European Union. It was in favour of a partnership with the EU. After a long battle, the Nationalist Party led by Eddie Fenech Adami won the referendum and the following election, making Malta one of the states to enter the European Union on 1 May 2004. The party is now pro-European. Historically, the Netherlands have been a very pro-European country, being one of the six founding members of the European Coal and Steel Community in 1952, and campaigning with much effort to include the United Kingdom into the Community in the 1970s and others after that. It has become slightly more Eurosceptic in the 2000s, rejecting the European Constitution in 2005 and complaining about the relatively high financial investment into the Union or the democratic deficit amongst other issues. A prominent former Eurosceptic party in the Netherlands was the Pim Fortuyn List (LPF) established by politician and academic Pim Fortuyn in 2002. The party campaigned to reduce Dutch financial contributions to the EU, was against Turkish membership and opposed what it saw as the excessive bureaucracy and threat to national sovereignty posed by the EU. During the 2002 general election, the LPF polled in second place with 17% of the vote. Following the assassination of Fortuyn in the run-up to the election, support for the party declined soon after and it was disbanded in 2008 with many of is former supporters transferring to the Party for Freedom. Despite these concerns, in 2014 the majority of the Dutch electorate continued to support parties that favour ongoing European integration: the Social Democrats, the Christian Democrats, the Liberals, but most of all the (Liberal) Democrats. In 2016, a substantial majority in a low-turnout referendum rejected the ratification of an EU trade and association treaty with Ukraine. In the 2019 European Parliament election, Eurosceptic parties had mixed results with Geert Wilders' Party for Freedom losing all 4 of its seats taking only 3.53% of the vote. The new Forum for Democracy established in late 2016 took 10.96% of the vote and entered the European Parliament with 3 seats. The main party with Eurosceptic views is the Confederation Freedom and Independence. Former president of Poland Lech Kaczyński resisted giving his signature on behalf of Poland to the Treaty of Lisbon, objecting specifically to the Charter of Fundamental Rights of the European Union. Subsequently, Poland got an opt-out from this charter. As Polish President, Kaczyński also opposed the Polish government's intentions to join the euro. In 2015, it was reported that Euroscepticism was growing in Poland, which was thought to be due to the "economic crisis, concern over perceived interference from Brussels and migration". Polish president Andrzej Duda indicated that he wished for Poland to step back from further EU integration. He suggested that the country should "hold a referendum on joining the euro, resist further integration and fight the EU’s green policies", despite getting the largest share of EU cash. In the 2019 European Parliament election, Law and Justice won the election with a vote share increase up from 31.78% to 45.38%, increasing its seats from 19 to 27. The main Eurosceptic parties in Portugal are National Renovator Party (PNR), Portuguese Communist Party (PCP), and Left Bloc (BE). Opinion polling in Portugal in 2015 indicated that 48 per cent tended not to trust the EU, while 79 per cent tended not to trust the Portuguese government (then led by Portugal Ahead). Eurosceptic political parties hold a combined total of 34 seats out of 230 in Portugal's parliament (BE 19, PCP 15, PNR 0) and a combined total of 4 out of Portugal's 21 seats in the European Parliament (PCP 3, BE 1, PNR 0). In the last 2014 European Parliament election, the Portuguese Communist Party won three seats and the Left Bloc won one seat. In the 2019 European Parliament election, Left Bloc took 9.83% and gained 1 seat, Portuguese Communist Party working in coalition with Ecologist Party "The Greens" took 6.88% and 2 seats and National Renovator Party (PNR) polled just 0.49%, with no seats. Several parties espousing Eurosceptic views exist on the right, such as the New Republic, the Greater Romania Party and Noua Dreaptă, but as of June 2020 none of these parties are represented in European Parliament. Euroscepticism is relatively unpopular in Romania; all mainstream political parties are pro-European and a 2015 survey found 65% of Romanians had a positive view of the country's EU membership. The Eurosceptic parties remained unrepresented in the 2019 European Parliament election. Parties with primarily hard Eurosceptic views represented in the National Council are People's Party - Our Slovakia and We Are Family. Prominent Slovak Eurosceptic politicians include Marian Kotleba, Boris Kollár, Andrej Danko and Richard Sulík . Soft Eurosceptic views are represented in Slovak National Party, New Majority and Freedom and Solidarity. In the 2019 European Parliament election, People's Party - Our Slovakia came 3rd securing 12.07% and winning their first 2 seats in the European Parliament. We Are Family took just over 3% and no seats. Parties with mainly Eurosceptic views are Slovenian National Party and United Left. Neither won seats in the 2019 European Parliament election in Slovenia. The process of Europeanization changed during the years in Spain. In 1986 Spain entered in the European Community. Since then, Spain has been one of the most Europeanist countries. Therefore, when Spain became part of the European Community, the country had a strong pro-Europeanist feeling, according to Eurobarometer, as it reflected a 60% of the population. In Spain different reasons explain its entrance to the European Community. On the one hand, democracy has just been established in Spain after Francisco Franco dictatorship. On the other hand, the main objectives of Spain were to achieve economic development, and also a social modernization. Spain was one of the few countries to vote Yes for the European Constitution in a referendum in February 2005, though by a lower margin in Catalonia and the Basque Country. In 2008, after the financial crisis reached Spain, the percentage of pro European persons started to fall. Thus, during the five years of the economic crisis, the Eurobarometer shows how the trust in the EU increasingly fell in Spain, and the confidence of the Spanish citizens in the European Union decreased for more than 50 points. Spain became one of the most Eurosceptic countries among all European Union Members, as it happened in pretty much European countries, where nationalist and eurosceptic characterised parties became stronger. The historical two-parties system, composed by the conservative Partido Popular and the social-democratic Partido Socialista Obrero Español, collapsed. In the 2000s, the liberal Ciudadanos and leftist party Podemos became part of the political context, gaining electoral consensus, followed years later by conservative party Vox. The new parties were the effect of the disaffection of most Spaniards towards politics and politicians, that increased for several reasons: firstly, corruption at all political levels, reaching the Royal Family too; secondly, recession intensified distrust of the population towards national government; thirdly, a phase of renovation of the autonomous regions which extended the distance between the National government and the Regional ones. Candidatura d'Unitat Popular, a left-wing to far-left political party with about 1,300 members advocates independence for Catalonia outside of the European Union. Up to 2014 European elections, there are no Spanish parties present in the Eurosceptic groups at the European Parliament. In the 2015 Spanish general election, Podemos became the first left-wing Eurosceptic political party to win seats in the Congress of Deputies, obtaining 69 seats, and in the 2019 Spanish general election, Vox became the first right-wing Eurosceptic political party to win seats in the Congress of Deputies, obtaining 24 seats. The Left Party of Sweden is against accession to the eurozone and wants Sweden to leave the European Union. The right-wing populist party Sweden Democrats are also strongly against the Union but not to withdraw EEA and Schengen Agreement. The June List, a Eurosceptic list consisting of members from both the political right and left won three seats in the 2004 Elections to the European Parliament and sat in the EU-critical IND/DEM group in the European Parliament. The movement favours a withdrawal from the EU. Around 75% of the Riksdag members represent parties that officially supports the Sweden membership. In the European Parliament election, 2014, the Sweden Democrats gained 2 seats with 9.67% of the vote, up 6.4%, and the Left Party took one seat with 6.3% of the vote. In the European Parliament election, 2019, the Sweden Democrats increased from 2 to 3 seats with 15.34% of the vote, up from 9.67%, and the Left Party retained its one seat with 6.8% of the vote. In winter 2019–2020, in connection with the request from "poor" member countries of much higher membership fees for "rich" member countries, for the reason of keeping support levels so "poor" countries wouldn't suffer from Brexit, where a "rich" country left the union in part due to high membership fees, a media and social media debate for a "Swexit" increased. This was still rejected by parties representing a majority of the parliament, with the COVID-19 pandemic quickly taking over the debate. Prosperous Armenia represents the main Eurosceptic party in Armenia. Following the 2018 Armenian parliamentary election, the party gained 26 seats in the National Assembly. The party is a member of the Alliance of Conservatives and Reformists in Europe. Georgian March is the main Eurosceptic party in Georgia. The party supports a slight distancing of Georgia from the West, as well as rejecting the country's entrance into NATO. The three main Eurosceptic parties in Iceland are the Independence Party, Left-Green Movement and the Progressive Party. The Independence Party and the Progressive Party won the parliamentary election in April 2013 and they have halted the current negotiations with the European Union regarding Icelandic membership and tabled a parliamentary resolution on 21 February 2014 to withdraw the application completely. In 2017, Iceland's newly elected government announced that it would hold a vote in parliament on whether to hold a referendum on resuming EU membership negotiations. In November 2017 that government was replaced by a coalition of the Independence Party, the Left Green Movement and the Progressive Party; all of whom oppose membership. Only 11 out of 63 MPs are in favour of EU membership. The two main Eurosceptic parties in Moldova are the left-wing Party of Socialists of the Republic of Moldova, which officially declared its main purpose to be the integration of Moldova in the Eurasian Economic Union and the Party of Communists of the Republic of Moldova, even if nowadays its leader speech became more soft on the issue of Euroscepticism. As of November 2014 both parties are represented in Moldovan Parliament, with 45 MPs out of a total of 101 MPs. All parliamentary parties in Montenegro officially support the country's bid for accession to the European Union. The only party that rejected the European integration and instead publicly advocates a tighter political and economic integration with Russia was the non-parliamentary far-right party Serb List. Since having come into national opposition, and amid disagreements in the Macedonia naming dispute, the VMRO-DPMNE - which does not endorse the name of "North Macedonia", instead continuing to refer to "Republic of Macedonia" without qualifiers - no longer supports the country's candidacy for EU membership. It has subsequently declared its aspirations towards Russia. Norway has rejected EU membership in two referendums, 1972 and 1994. The Centre Party, Christian Democratic Party, Socialist Left Party and Liberal Party were against EU membership in both referendums. The Centre Party, Socialist Left Party, Capitalist Party, Christians and Red Party are also against Norway's current membership of the European Economic Area. Parties with mainly Eurosceptic views are the ruling United Russia, and opposition parties the Communist Party of the Russian Federation and Liberal Democratic Party of Russia. Following the 2014 Crimean crisis, the European Union issued sanctions on the Russian Federation "in response to the illegal annexation of Crimea and deliberate destabilisation of a neighbouring sovereign country". In response to this, Alexey Borodavkin – Russia's permanent representative with the UN – said "The EU is committing a direct violation of human rights by its actions against Russia. The unilateral sanctions introduced against us are not only illegitimate according to international law, they also undermine Russian citizens' freedom of travel, freedom of development, freedom of work and others". In the same year, Russian president Vladimir Putin said: "What are the so-called European values? Maintaining the coup, the armed seizure of power and the suppression of dissent with the help of the armed forces?" A February 2014 poll conducted by the Levada Center, Russia's largest independent polling organization, found that nearly 80% of Russian respondents had a "good" impression of the EU. This changed dramatically in 2014 with the Ukrainian crisis resulting in 70% taking a hostile view of the EU compared to 20% viewing it positively. A Levada poll released in August 2018 found that 68% of Russians polled believe that Russia needs to dramatically improve relations with Western countries. 42% of Russian respondents said they had a positive view of the EU, up from 28% in May 2018. A referendum was held in the landlocked microstate on 20 October 2013 in which the citizens were asked whether the country should submit an application to join the European Union. The proposal was rejected because of a low turnout, even though 50.3% of voters approved it. The "Yes" campaign was supported by the main left-wing parties (Socialist Party, United Left) and the Union for the Republic whereas the Sammarinese Christian Democratic Party suggested voting with a blank ballot, the Popular Alliance declared itself neutral, and We Sammarinese and the RETE movement supported the "No" campaign. The Citizens' Rights Directive, which defines the right of free movement for the European citizens, may have been an important reason for those voting no. Parties with mainly Eurosceptic views are Serbian Radical Party, Democratic Party of Serbia, Dveri, DJB and the Serbian People's Party of Nenad Popović. Switzerland has long been known for its neutrality in international politics. Swiss voters rejected EEA membership in 1992, and EU membership in 2001. Despite the passing of several referendums calling for closer relations between Switzerland and the European Union such as the adoption of bilateral treaties and the joining of the Schengen Area, a second referendum of the joining of the EEA or the EU is not expected, and the general public remains opposed to joining. In February 2014, the Swiss voters narrowly approved a referendum limiting the freedom of movement of EU citizens to Switzerland. Eurosceptic political parties include the Swiss People's Party, which is the largest political party in Switzerland, with 29.4% of the popular vote as of the 2015 federal election. Smaller Eurosceptic parties include, but are not limited to, the Federal Democratic Union, the Ticino League, and the Geneva Citizens' Movement, all of which are considered right-wing parties. In addition, the Campaign for an Independent and Neutral Switzerland is a political organisation in Switzerland that is strongly opposed to Swiss membership of or further integration otherwise with the European Union. Regionally, the German-speaking majority as well as the Italian-speaking areas are the most Eurosceptic, while French-speaking Switzerland tends to be more pro-European integration. In the 2001 referendum, the majority of French-speakers voted against EU membership. According to a 2016 survey conducted by M.I.S Trend and published in "L'Hebdo", 69 percent of the Swiss population supports systematic border controls, and 53 percent want restrictions on the EU accord of the free movements of peoples and 14 percent want it completely abolished. 54% of the Swiss population said that if necessary, they would ultimately keep the freedom of movement of people's accord. The two main Eurosceptic parties are the far-right ultranationalist, Nationalist Movement Party (MHP), which secured 11.10% of votes, and 49 seats in the Parliament at the last election, and the Felicity Party (Saadet Partisi), a far-right Sunni Islamist party, which has no seats in the Parliament, as it only secured 0.68% of the votes in the last election, far below the 10% threshold necessary to be represented in the Parliament. Many left-wing nationalist and far-left parties hold no seats at parliament but they control many activist and student movements in Turkey. The Patriotic Party (formerly called Workers' Party) consider the European Union as a frontrunner of global imperialism. Parties with mainly Eurosceptic views are Opposition Platform — For life, Communist Party of Ukraine and Right Sector. The far-right Ukrainian group Right Sector opposes joining the European Union. It regards the EU as an "oppressor" of European nations. Euroscepticism in the United Kingdom has been a significant element in British politics since the inception of the European Economic Community (EEC), the predecessor to the EU. The European Union strongly divides the British public, political parties, media and civil society. The UK Independence Party has backed the idea of the UK unilaterally leaving the European Union (Brexit) since its inception in 1993. During the 23 June 2016 referendum on the issue, the Conservatives had no official position on the issue; although its leader David Cameron was in favour of remaining in the EU, the party was split on the issue. The Labour Party officially supported remaining in the EU, although party leader Jeremy Corbyn did suggest early on in the campaign that he would consider withdrawal; which he advocated years previously. The Liberal Democrats were the most adamantly pro-EU party, and since the referendum, pro-Europeanism has been their main policy. The referendum resulted in an overall vote to leave the EU, as opposed to remaining an EU member, by 51.9% to 48.1%, on a turnout of 72.2%. The vote was split between the constituent countries of the United Kingdom, with a majority in England and Wales voting to leave, and a majority in Scotland and Northern Ireland, as well as Gibraltar (a British Overseas Territory), voting to remain. As a result of the referendum, the government notified the EU of its intention to withdraw on 29 March 2017 by invoking Article 50 of the Lisbon Treaty. On 12 April 2019 a new Eurosceptic party, the Brexit Party was officially launched by former UK Independence Party Leader Nigel Farage. In the 2019 European Parliament election, the Brexit Party topped the national poll by a large margin with 31.69% gaining 29 seats by running on a single policy of leaving the EU, versus the second-placed Liberal Democrats with 18.53% and 16 seats who promoted themselves as the party of Remain (the total vote for Remain-supporting parties was approximately the same as that for parties supporting a 'no-deal' Brexit). The Conservative party suffered their lowest ever national vote share of 9.1% with just 4 seats following 3 years of Theresa May's unsuccessful Brexit negotiations. The Labour party's ambiguous position on Brexit led to their vote share dropping significantly to 14.08% resulting in the loss of half their seats, down from 20 to 10. The rapid growth of the Brexit Party was a contributing factor to Theresa May announcing on 24 May that she would step down as Prime Minister on 7 June 2019. After the elections, the Eurosceptic Blue Collar Conservative caucus of Conservative MPs was formed. On 23 January 2020, the United Kingdom parliament ratified a withdrawal agreement from the European Union, which was ratified by the EU parliament on 30 January. On 31 January the United Kingdom officially left the European Union.
https://en.wikipedia.org/wiki?curid=10037
Estimator In statistics, an estimator is a rule for calculating an estimate of a given quantity based on observed data: thus the rule (the estimator), the quantity of interest (the estimand) and its result (the estimate) are distinguished. There are point and interval estimators. The point estimators yield single-valued results, although this includes the possibility of single vector-valued results and results that can be expressed as a single function. This is in contrast to an interval estimator, where the result would be a range of plausible values (or vectors or functions). Estimation theory is concerned with the properties of estimators; that is, with defining properties that can be used to compare different estimators (different rules for creating estimates) for the same quantity, based on the same data. Such properties can be used to determine the best rules to use under given circumstances. However, in robust statistics, statistical theory goes on to consider the balance between having good properties, if tightly defined assumptions hold, and having less good properties that hold under wider conditions. An "estimator" or "point estimate" is a statistic (that is, a function of the data) that is used to infer the value of an unknown parameter in a statistical model. The parameter being estimated is sometimes called the "estimand". It can be either finite-dimensional (in parametric and semi-parametric models), or infinite-dimensional (semi-parametric and non-parametric models). If the parameter is denoted formula_1 then the estimator is traditionally written by adding a circumflex over the symbol: formula_2. Being a function of the data, the estimator is itself a random variable; a particular realization of this random variable is called the "estimate". Sometimes the words "estimator" and "estimate" are used interchangeably. The definition places virtually no restrictions on which functions of the data can be called the "estimators". The attractiveness of different estimators can be judged by looking at their properties, such as unbiasedness, mean square error, consistency, asymptotic distribution, etc. The construction and comparison of estimators are the subjects of the estimation theory. In the context of decision theory, an estimator is a type of decision rule, and its performance may be evaluated through the use of loss functions. When the word "estimator" is used without a qualifier, it usually refers to point estimation. The estimate in this case is a single point in the parameter space. There also exists another type of estimator: interval estimators, where the estimates are subsets of the parameter space. The problem of density estimation arises in two applications. Firstly, in estimating the probability density functions of random variables and secondly in estimating the spectral density function of a time series. In these problems the estimates are functions that can be thought of as point estimates in an infinite dimensional space, and there are corresponding interval estimation problems. Suppose a fixed "parameter" formula_3 needs to be estimated. Then an "estimator" is a function that maps the sample space to a set of "sample estimates". An estimator of formula_3 is usually denoted by the symbol formula_2. It is often convenient to express the theory using the algebra of random variables: thus if "X" is used to denote a random variable corresponding to the observed data, the estimator (itself treated as a random variable) is symbolised as a function of that random variable, formula_6. The estimate for a particular observed data value formula_7 (i.e. for formula_8) is then formula_9, which is a fixed value. Often an abbreviated notation is used in which formula_2 is interpreted directly as a random variable, but this can cause confusion. The following definitions and attributes are relevant. For a given sample formula_11, the "error" of the estimator formula_2 is defined as where formula_14 is the parameter being estimated. The error, "e", depends not only on the estimator (the estimation formula or procedure), but also on the sample. The "mean squared error" of formula_2 is defined as the expected value (probability-weighted average, over all samples) of the squared errors; that is, It is used to indicate how far, on average, the collection of estimates are from the single parameter being estimated. Consider the following analogy. Suppose the parameter is the bull's-eye of a target, the estimator is the process of shooting arrows at the target, and the individual arrows are estimates (samples). Then high MSE means the average distance of the arrows from the bull's-eye is high, and low MSE means the average distance from the bull's-eye is low. The arrows may or may not be clustered. For example, even if all arrows hit the same point, yet grossly miss the target, the MSE is still relatively large. However, if the MSE is relatively low then the arrows are likely more highly clustered (than highly dispersed) around the target. For a given sample formula_11, the "sampling deviation" of the estimator formula_2 is defined as where formula_20 is the expected value of the estimator. The sampling deviation, "d", depends not only on the estimator, but also on the sample. The "variance" of formula_2 is simply the expected value of the squared sampling deviations; that is, formula_22. It is used to indicate how far, on average, the collection of estimates are from the "expected value" of the estimates. (Note the difference between MSE and variance.) If the parameter is the bull's-eye of a target, and the arrows are estimates, then a relatively high variance means the arrows are dispersed, and a relatively low variance means the arrows are clustered. Even if the variance is low, the cluster of arrows may still be far off-target, and even if the variance is high, the diffuse collection of arrows may still be unbiased. Finally, even if all arrows grossly miss the target, if they nevertheless all hit the same point, the variance is zero. The "bias" of formula_2 is defined as formula_24. It is the distance between the average of the collection of estimates, and the single parameter being estimated. The bias of formula_2 is a function of the true value of formula_3 so saying that the bias of formula_2 is formula_28 means that for every formula_3 the bias of formula_2 is formula_28. The bias also is the expected value of the error, since formula_32. If the parameter is the bull's-eye of a target, and the arrows are estimates, then a relatively high absolute value for the bias means the average position of the arrows is off-target, and a relatively low absolute bias means the average position of the arrows is on target. They may be dispersed, or may be clustered. The relationship between bias and variance is analogous to the relationship between accuracy and precision. The estimator formula_2 is an "unbiased estimator" of formula_1 if and only if formula_35. Bias is a property of the estimator, not of the estimate. Often, people refer to a "biased estimate" or an "unbiased estimate," but they really are talking about an "estimate from a biased estimator," or an "estimate from an unbiased estimator." Also, people often confuse the "error" of a single estimate with the "bias" of an estimator. That the error for one estimate is large, does not mean the estimator is biased. In fact, even if all estimates have astronomical absolute values for their errors, if the expected value of the error is zero, the estimator is unbiased. Also, an estimator's being biased does not preclude the error of an estimate from being zero in a particular instance. The ideal situation is to have an unbiased estimator with low variance, and also try to limit the number of samples where the error is extreme (that is, have few outliers). Yet unbiasedness is not essential. Often, if just a little bias is permitted, then an estimator can be found with lower MSE and/or fewer outlier sample estimates. An alternative to the version of "unbiased" above, is "median-unbiased", where the median of the distribution of estimates agrees with the true value; thus, in the long run half the estimates will be too low and half too high. While this applies immediately only to scalar-valued estimators, it can be extended to any measure of central tendency of a distribution: see median-unbiased estimators. A consistent sequence of estimators is a sequence of estimators that converge in probability to the quantity being estimated as the index (usually the sample size) grows without bound. In other words, increasing the sample size increases the probability of the estimator being close to the population parameter. Mathematically, a sequence of estimators } is a consistent estimator for parameter "θ" if and only if, for all , no matter how small, we have The consistency defined above may be called weak consistency. The sequence is "strongly consistent", if it converges almost surely to the true value. An estimator that converges to a "multiple" of a parameter can be made into a consistent estimator by multiplying the estimator by a scale factor, namely the true value divided by the asymptotic value of the estimator. This occurs frequently in estimation of scale parameters by measures of statistical dispersion. An asymptotically normal estimator is a consistent estimator whose distribution around the true parameter "θ" approaches a normal distribution with standard deviation shrinking in proportion to formula_43 as the sample size "n" grows. Using formula_44 to denote convergence in distribution, "tn" is asymptotically normal if for some "V". In this formulation "V/n" can be called the "asymptotic variance" of the estimator. However, some authors also call "V" the "asymptotic variance". Note that convergence will not necessarily have occurred for any finite "n", therefore this value is only an approximation to the true variance of the estimator, while in the limit the asymptotic variance (V/n) is simply zero. To be more specific, the distribution of the estimator "tn" converges weakly to a dirac delta function centered at formula_3. The central limit theorem implies asymptotic normality of the sample mean formula_47 as an estimator of the true mean. More generally, maximum likelihood estimators are asymptotically normal under fairly weak regularity conditions — see the asymptotics section of the maximum likelihood article. However, not all estimators are asymptotically normal; the simplest examples are found when the true value of a parameter lies on the boundary of the allowable parameter region. Two naturally desirable properties of estimators are for them to be unbiased and have minimal mean squared error (MSE). These cannot in general both be satisfied simultaneously: a biased estimator may have lower mean squared error (MSE) than any unbiased estimator; see estimator bias. Among unbiased estimators, there often exists one with the lowest variance, called the minimum variance unbiased estimator (MVUE). In some cases an unbiased efficient estimator exists, which, in addition to having the lowest variance among unbiased estimators, satisfies the Cramér–Rao bound, which is an absolute lower bound on variance for statistics of a variable. Concerning such "best unbiased estimators", see also Cramér–Rao bound, Gauss–Markov theorem, Lehmann–Scheffé theorem, Rao–Blackwell theorem.
https://en.wikipedia.org/wiki?curid=10043
Emerald Emerald is a gemstone and a variety of the mineral beryl (Be3Al2(SiO3)6) colored green by trace amounts of chromium and sometimes vanadium. Beryl has a hardness of 7.5–8 on the Mohs scale. Most emeralds are highly included, so their toughness (resistance to breakage) is classified as generally poor. Emerald is a cyclosilicate. The word "emerald" is derived (via and ), from Vulgar Latin: "esmaralda"/"esmaraldus", a variant of Latin "smaragdus", which was borrowed via (smaragdos; "green gem") from a Semitic language. Emeralds, like all colored gemstones, are graded using four basic parameters–the four "C"s of connoisseurship: "color", "clarity," "cut" and "carat weight". Normally, in the grading of colored gemstones, color is by far the most important criterion. However, in the grading of emeralds, clarity is considered a close second. A fine emerald must possess not only a pure verdant green hue as described below, but also a high degree of transparency to be considered a top gemstone. In the 1960s, the American jewelry industry changed the definition of "emerald" to include the green vanadium-bearing beryl. As a result, "vanadium emeralds" purchased as emeralds in the United States are not recognized as such in the UK and Europe. In America, the distinction between traditional emeralds and the new vanadium kind is often reflected in the use of terms such as "Colombian emerald". In gemology, color is divided into three components: "hue", "saturation", and "tone". Emeralds occur in hues ranging from yellow-green to blue-green, with the primary hue necessarily being green. Yellow and blue are the normal secondary hues found in emeralds. Only gems that are medium to dark in tone are considered emeralds; light-toned gems are known instead by the species name "green beryl". The finest emeralds are approximately 75% tone on a scale where 0% tone is colorless and 100% is opaque black. In addition, a fine emerald will be saturated and have a hue that is bright (vivid). Gray is the normal saturation modifier or mask found in emeralds; a grayish-green hue is a dull-green hue. Emeralds tend to have numerous inclusions and surface-breaking fissures. Unlike diamonds, where the loupe standard, i.e. 10× magnification, is used to grade clarity, emeralds are graded by eye. Thus, if an emerald has no visible inclusions to the eye (assuming normal visual acuity) it is considered flawless. Stones that lack surface breaking fissures are extremely rare and therefore almost all emeralds are treated ("oiled", see below) to enhance the apparent clarity. The inclusions and fissures within an emerald are sometime described as "jardin" (French for "garden"), because of their mossy appearance. Imperfections are unique for each emerald and can be used to identify a particular stone. Eye-clean stones of a vivid primary green hue (as described above), with no more than 15% of any secondary hue or combination (either blue or yellow) of a medium-dark tone, command the highest prices. The relative non-uniformity motivates the cutting of emeralds in cabochon form, rather than faceted shapes. Faceted emeralds are most commonly given an oval cut, or the signature emerald cut, a rectangular cut with facets around the top edge. Most emeralds are oiled as part of the post-lapidary process, in order to fill in surface-reaching cracks so that clarity and stability are improved. Cedar oil, having a similar refractive index, is often used in this widely adopted practice. Other liquids, including synthetic oils and polymers with refractive indexes close to that of emeralds, such as "Opticon", are also used. These treatments are typically applied in a vacuum chamber under mild heat, to open the pores of the stone and allow the fracture-filling agent to be absorbed more effectively. The U.S. Federal Trade Commission requires the disclosure of this treatment when an oil treated emerald is sold. The use of oil is traditional and largely accepted by the gem trade, although oil treated emeralds are worth much less than un-treated emeralds of similar quality. Other treatments, for example the use of green-tinted oil, are not acceptable in the trade. Gems are graded on a four-step scale; "none", "minor", "moderate" and "highly" enhanced. These categories reflect levels of enhancement, not "clarity". A gem graded "none" on the enhancement scale may still exhibit visible inclusions. Laboratories apply these criteria differently. Some gemologists consider the mere presence of oil or polymers to constitute enhancement. Others may ignore traces of oil if the presence of the material does not improve the look of the gemstone. Emeralds in antiquity were mined in Egypt at locations on Mount Smaragdus since 1500 BCE, and India, and Austria since at least the 14th century CE. The Egyptian mines were exploited on an industrial scale by the Roman and Byzantine Empires, and later by Islamic conquerors. Mining ceased with the discovery of the Colombian deposits; only ruins remain. Colombia is by far the world's largest producer of emeralds, constituting 50–95% of the world production, with the number depending on the year, source and grade. Emerald production in Colombia has increased drastically in the last decade, increasing by 78% from 2000 to 2010. The three main emerald mining areas in Colombia are Muzo, Coscuez, and Chivor. Rare "trapiche" emeralds are found in Colombia, distinguished by ray-like spokes of dark impurities. Zambia is the world's second biggest producer, with its Kafubu River area deposits (Kagem Mines) about southwest of Kitwe responsible for 20% of the world's production of gem-quality stones in 2004. In the first half of 2011, the Kagem Mines produced 3.74 tons of emeralds. Emeralds are found all over the world in countries such as Afghanistan, Australia, Austria, Brazil, Bulgaria, Cambodia, Canada, China, Egypt, Ethiopia, France, Germany, India, Italy, Kazakhstan, Madagascar, Mozambique, Namibia, Nigeria, Norway, Pakistan, Russia, Somalia, South Africa, Spain, Switzerland, Tanzania, the United States, Zambia, and Zimbabwe. In the US, emeralds have been found in Connecticut, Montana, Nevada, North Carolina, and South Carolina. In Canada, in 1997 emeralds were discovered in the Yukon. Since the onset of concerns regarding diamond origins, research has been conducted to determine if the mining location could be determined for an emerald already in circulation. Traditional research used qualitative guidelines such as an emerald's color, style and quality of cutting, type of fracture filling, and the anthropological origins of the artifacts bearing the mineral to determine the emerald's mine location. More recent studies using energy dispersive X-ray spectroscopy methods have uncovered trace chemical element differences between emeralds, including ones mined in close proximity to one another. American gemologist David Cronin and his colleagues have extensively examined the chemical signatures of emeralds resulting from fluid dynamics and subtle precipitation mechanisms, and their research demonstrated the chemical homogeneity of emeralds from the same mining location and the statistical differences that exist between emeralds from different mining locations, including those between the three locations: Muzo, Coscuez, and Chivor, in Colombia, South America. Both hydrothermal and "flux-growth" synthetics have been produced, and a method has been developed for producing an emerald overgrowth on colorless beryl. The first commercially successful emerald synthesis process was that of Carroll Chatham, likely involving a lithium vanadate flux process, as Chatham's emeralds do not have any water and contain traces of vanadate, molybdenum and vanadium. The other large producer of flux emeralds was Pierre Gilson Sr., whose products have been on the market since 1964. Gilson's emeralds are usually grown on natural colorless beryl seeds, which are coated on both sides. Growth occurs at the rate of 1 mm per month, a typical seven-month growth run produces emerald crystals 7 mm thick. Hydrothermal synthetic emeralds have been attributed to IG Farben, Nacken, Tairus, and others, but the first satisfactory commercial product was that of Johann Lechleitner of Innsbruck, Austria, which appeared on the market in the 1960s. These stones were initially sold under the names "Emerita" and "Symeralds", and they were grown as a thin layer of emerald on top of natural colorless beryl stones. Later, from 1965 to 1970, the Linde Division of Union Carbide produced completely synthetic emeralds by hydrothermal synthesis. According to their patents (attributable to E.M. Flanigen), acidic conditions are essential to prevent the chromium (which is used as the colorant) from precipitating. Also, it is important that the silicon-containing nutrient be kept away from the other ingredients to prevent nucleation and confine growth to the seed crystals. Growth occurs by a diffusion-reaction process, assisted by convection. The largest producer of hydrothermal emeralds today is Tairus, which has succeeded in synthesizing emeralds with chemical composition similar to emeralds in alkaline deposits in Colombia, and whose products are thus known as “Colombian created emeralds” or “Tairus created emeralds”. Luminescence in ultraviolet light is considered a supplementary test when making a natural versus synthetic determination, as many, but not all, natural emeralds are inert to ultraviolet light. Many synthetics are also UV inert. Synthetic emeralds are often referred to as "created", as their chemical and gemological composition is the same as their natural counterparts. The U.S. Federal Trade Commission (FTC) has very strict regulations as to what can and what cannot be called a "synthetic" stone. The FTC says: "§ 23.23(c) It is unfair or deceptive to use the word "laboratory-grown", "laboratory-created", "[manufacturer name]-created", or "synthetic" with the name of any natural stone to describe any industry product unless such industry product has essentially the same optical, physical, and chemical properties as the stone named." Emerald is regarded as the traditional birthstone for May as well as the traditional gemstone for the astrological sign of Cancer. One of the quainter anecdotes about emeralds was told by the 16th-century historian Brantôme, who referred to the many impressive emeralds the Spanish under Cortez had brought back to Europe from Latin America. On one of Cortez's most notable emeralds he had the text engraved, "Inter Natos Mulierum non sur-rexit mayor" ("Among those born of woman there hath not arisen a greater," Matthew 11:11) which referred to John the Baptist. Brantôme considered engraving such a beautiful and simple product of nature sacrilegious and considered this act the cause for Cortez's loss of an extremely precious pearl (to which he dedicated a work, "A beautiful and incomparable pearl"), and even for the death of King Charles IX of France, who died soon afterward. The chief deity of one of India's most famous temples, the Meenakshi Amman Temple in Madurai, is the goddess Meenakshi, whose idol is traditionally thought to be made of emerald.
https://en.wikipedia.org/wiki?curid=10045
Erie Canal The Erie Canal is a canal in New York, United States that is part of the east–west, cross-state route of the New York State Canal System (formerly known as the New York State Barge Canal). Originally, it ran from the Hudson River in Albany to Lake Erie in Buffalo. It was built to create a navigable water route from New York City and the Atlantic Ocean to the Great Lakes. When completed in 1825, it was the second longest canal in the world (after the Grand Canal in China) and greatly enhanced the development and economy of New York, New York City, and the United States. The canal was first proposed in the 1780s, then re-proposed in 1807. A survey was authorized, funded, and executed in 1808. Proponents of the project gradually wore down opponents; its construction began in 1817. The canal has 34 numbered locks starting with Black Rock Lock and ending downstream with the Troy Federal Lock. Both are owned by the federal government. It has an elevation difference of about . It opened on October 26, 1825. In a time when bulk goods were limited to pack animals (a maximum), and there were no railways, water was the most cost-effective way to ship bulk goods. Political opponents to the canal and to New York Governor DeWitt Clinton denigrated it as "Clinton's Folly" or "Clinton's Big Ditch".
https://en.wikipedia.org/wiki?curid=10046
Ethanol Ethanol (also called ethyl alcohol, grain alcohol, drinking alcohol, or simply alcohol) is a chemical compound, a simple alcohol with the chemical formula . Its formula can be also written as −− or (an ethyl group linked to a hydroxyl group), and is often abbreviated as EtOH. Ethanol is a volatile, flammable, colorless liquid with a slight characteristic odor. It is a psychoactive substance and is the principal active ingredient found in alcoholic drinks. Ethanol is naturally produced by the fermentation of sugars by yeasts or via petrochemical processes, and is commonly consumed as a popular recreational drug. It also has medical applications as an antiseptic and disinfectant. The compound is widely used as a chemical solvent, either for scientific chemical testing or in synthesis of other organic compounds, and is a vital substance used across many different kinds of manufacturing industries. Ethanol is also used as an alternative fuel source. "Ethanol" is the systematic name defined by the International Union of Pure and Applied Chemistry (IUPAC) for a compound consisting of alkyl group with two carbon atoms (prefix “eth-”), having a single bond between them (infix “-an-”), attached functional group −OH group (suffix “-ol”). The “eth-” prefix and the qualifier “ethyl” in “ethyl alcohol” originally come from the name “ethyl” assigned in 1834 to the group − by Justus Liebig. He coined the word from the German name "Aether" of the compound −O− (commonly called “ether” in English, more specifically called “diethyl ether”). According to the "Oxford English Dictionary", "Ethyl" is a contraction of the Ancient Greek αἰθήρ (', “upper air”) and the Greek word ὕλη (', “substance”). The name "ethanol" was coined as a result of a resolution that was adopted at the International Conference on Chemical Nomenclature that was held in April 1892 in Geneva, Switzerland.
https://en.wikipedia.org/wiki?curid=10048
Eric Clapton Eric Patrick Clapton, (born 1945) is an English rock and blues guitarist, singer, and songwriter. He is the only three-time inductee to the Rock and Roll Hall of Fame: once as a solo artist and separately as a member of the Yardbirds and of Cream. Clapton has been referred to as one of the most important and influential guitarists of all time. Clapton ranked second in "Rolling Stone"s list of the "100 Greatest Guitarists of All Time" and fourth in Gibsons "Top 50 Guitarists of All Time". He was also named number five in "Time" magazine's list of "The 10 Best Electric Guitar Players" in 2009. In the mid-1960s Clapton left the Yardbirds to play with John Mayall & the Bluesbreakers. Immediately after leaving Mayall, Clapton formed the power trio Cream with drummer Ginger Baker and bassist Jack Bruce, in which Clapton played sustained blues improvisations and "arty, blues-based psychedelic pop". After Cream broke up, he formed blues rock band Blind Faith with Baker, Steve Winwood, and Ric Grech. Clapton's solo career began in the 1970s, where his work bore the influence of the mellow style of J. J. Cale and the reggae of Bob Marley. His version of Marley's "I Shot the Sheriff" helped reggae reach a mass market. Two of his most popular recordings were "Layla", recorded with Derek and the Dominos; and Robert Johnson's "Crossroads", recorded with Cream. Following the death of his son Conor in 1991, Clapton's grief was expressed in the song "Tears in Heaven", which appeared on his "Unplugged" album. Clapton has been the recipient of 18 Grammy Awards, and the Brit Award for Outstanding Contribution to Music. In 2004 he was awarded a CBE at Buckingham Palace for services to music. He has received four Ivor Novello Awards from the British Academy of Songwriters, Composers and Authors, including the Lifetime Achievement Award. In his solo career, Clapton has sold more than 100 million records worldwide, making him one of the best-selling musicians of all time. In 1998, Clapton, a recovering alcoholic and drug addict, founded the Crossroads Centre on Antigua, a medical facility for recovering substance abusers. Clapton was born on 30 March 1945 in Ripley, Surrey, England, to 16-year-old Patricia Molly Clapton ( 1929 – March 1999) and Edward Walter Fryer ( 1920 – 1985), a 25-year-old soldier from Montreal, Quebec. Fryer shipped off to war prior to Clapton's birth and then returned to Canada. Clapton grew up believing that his grandmother, Rose, and her second husband, Jack Clapp, Patricia's stepfather, were his parents, and that his mother was actually his older sister. The similarity in surnames gave rise to the erroneous belief that Clapton's real surname is Clapp (Reginald Cecil Clapton was the name of Rose's first husband, Eric Clapton's maternal grandfather). Years later, his mother married another Canadian soldier and moved to Germany, leaving young Eric with his grandparents in Surrey. Clapton received an acoustic Hoyer guitar, made in Germany, for his thirteenth birthday, but the inexpensive steel-stringed instrument was difficult to play and he briefly lost interest. Two years later Clapton picked it up again and started playing consistently. Clapton was influenced by the blues from an early age, and practised long hours to learn the chords of blues music by playing along to the records. He preserved his practice sessions using his portable Grundig reel-to-reel tape recorder, listening to them over and over until he felt he'd got it right. In 1961, after leaving Hollyfield School in Surbiton, Clapton studied at the Kingston College of Art but was dismissed at the end of the academic year because his focus remained on music rather than art. His guitar playing was so advanced that, by the age of 16, he was getting noticed. Around this time, Clapton began busking around Kingston, Richmond, and the West End. In 1962, Clapton started performing as a duo with fellow blues enthusiast David Brock in pubs around Surrey. When he was seventeen years old, Clapton joined his first band, an early British R&B group, the Roosters, whose other guitarist was Tom McGuinness. He stayed with this band from January until August 1963. In October of that year, Clapton did a seven-gig stint with Casey Jones & the Engineers. In October 1963, Clapton joined the Yardbirds, a blues-influenced rock and roll band, and stayed with them until March 1965. Synthesising influences from Chicago blues and leading blues guitarists such as Buddy Guy, Freddie King, and B.B. King, Clapton forged a distinctive style and rapidly became one of the most talked-about guitarists in the British music scene. The band initially played Chess/Checker/Vee-Jay blues numbers and began to attract a large cult following when they took over the Rolling Stones' residency at the Crawdaddy Club in Richmond. They toured England with American bluesman Sonny Boy Williamson II; a joint LP album, recorded in December 1963, was issued in 1965. Yardbirds' rhythm guitarist, Chris Dreja, recalled that whenever Clapton broke a guitar string during a concert, he would stay on stage and replace it. The English audiences would wait out the delay by doing what is called a "slow handclap". Clapton's nickname of "Slowhand" came from Giorgio Gomelsky, a pun on the slow handclapping that ensued when Clapton stopped playing while he replaced a string. In December 1964, Clapton made his first appearance at the Royal Albert Hall, London, with the Yardbirds. Since then, Clapton has performed at the Hall over 200 times, and has stated that performing at the venue is like "playing in my front room". In March 1965, Clapton and the Yardbirds had their first major hit, "For Your Love", written by songwriter Graham Gouldman, who also wrote hit songs for Herman's Hermits and the Hollies (and would later achieve success of his own as a member of 10cc). In part because of its success, the Yardbirds elected to move toward a pop-oriented sound, much to the annoyance of Clapton, who was devoted to the blues and not commercial success. He left the Yardbirds on the day that "For Your Love" went public, a move that left the band without its lead guitarist and most accomplished member. Clapton suggested fellow guitarist Jimmy Page as his replacement, but Page declined out of loyalty to Clapton, putting Jeff Beck forward. Beck and Page played together in the Yardbirds for a while, but Beck, Page, and Clapton were never in the group together. They first appeared together on the 12-date benefit tour for Action for Research into multiple sclerosis in 1983 with the first date taking place on 23 September at the Royal Albert Hall. Clapton joined John Mayall & the Bluesbreakers in April 1965, only to quit a few months later. In June, Clapton was invited to jam with Jimmy Page, recording a number of tracks that would be retroactively credited to The Immediate All-Stars. In the summer of 1965 he left for Greece with a band called the Glands, which included his old friend Ben Palmer on piano. In November 1965 he rejoined John Mayall. In March 1966, while still a member of the Bluesbreakers, Clapton briefly collaborated on a side project with Jack Bruce and Steve Winwood among others, recording only a few tracks under the name Eric Clapton and the Powerhouse. During his second Bluesbreakers stint, Clapton gained a reputation as the best blues guitarist on the club circuit. Although Clapton gained world fame for his playing on the influential album, "Blues Breakers – John Mayall – With Eric Clapton", this album was not released until he had left the band for the last time in July 1966. Having swapped his Fender Telecaster and Vox AC30 amplifier for a 1960 Gibson Les Paul Standard guitar and Marshall amplifier, Clapton's sound and playing inspired the famous slogan "Clapton is God", spray-painted by an unknown admirer on a wall in Islington in 1967. The graffiti was captured in a now-famous photograph, in which a dog is urinating on the wall. Clapton is reported to have been embarrassed by the slogan, saying in his "The South Bank Show" profile in 1987, "I never accepted that I was the greatest guitar player in the world. I always "wanted" to be the greatest guitar player in the world, but that's an ideal, and I accept it as an ideal". Clapton left the Bluesbreakers in July 1966 (replaced by Peter Green) and was invited by drummer Ginger Baker to play in his newly formed band Cream, one of the earliest supergroups, with Jack Bruce on bass (Bruce was previously of the Bluesbreakers, the Graham Bond Organisation and Manfred Mann). Before the formation of Cream, Clapton was not well known in the United States; he left the Yardbirds before "For Your Love" hit the US top ten, and had yet to perform there. During his time with Cream, Clapton began to develop as a singer, songwriter and guitarist, though Bruce took most of the lead vocals and wrote the majority of the material with lyricist Pete Brown. Cream's first gig was an unofficial performance at the Twisted Wheel Club in Manchester on 1966 before their full debut two nights later at the National Jazz and Blues Festival in Windsor. Cream established its enduring legend with the high-volume blues jamming and extended solos of their live shows. By early 1967, fans of the emerging blues-rock sound in the UK had begun to portray Clapton as Britain's top guitarist; however, he found himself rivalled by the emergence of Jimi Hendrix, an acid rock-infused guitarist who used wailing feedback and effects pedals to create new sounds for the instrument. Hendrix attended a performance of the newly formed Cream at the Central London Polytechnic on 1966, during which he sat in on a double-timed version of "Killing Floor". Top UK stars, including Clapton, Pete Townshend and members of the Rolling Stones and the Beatles, avidly attended Hendrix's early club performances. Hendrix's arrival had an immediate and major effect on the next phase of Clapton's career. Clapton first visited the United States while touring with Cream. In March 1967, Cream performed a nine-show stand at the RKO Theater in New York. Clapton's 1964 painted Gibson SG guitar – The Fool – a "psychedelic fantasy", according to Clapton, made its debut at the RKO Theater. Clapton used the guitar for most of Cream's recordings after "Fresh Cream", particularly on "Disraeli Gears", until the band broke up in 1968. One of the world's best-known guitars, it symbolises the psychedelic era. They recorded "Disraeli Gears" in New York from 11 to 15 May 1967. Cream's repertoire varied from hard rock ("I Feel Free") to lengthy blues-based instrumental jams ("Spoonful"). "Disraeli Gears" contained Clapton's searing guitar lines, Bruce's soaring vocals and prominent, fluid bass playing, and Baker's powerful, polyrhythmic jazz-influenced drumming. Together, Cream's talents secured them as an influential power trio. Clapton's voice can be heard on Frank Zappa's album "We're Only in It for the Money", on the tracks "Are You Hung Up?" and "Nasal Retentive Calliope Music". In 28 months, Cream had become a commercial success, selling millions of records and playing throughout the US and Europe. They redefined the instrumentalist's role in rock and were one of the first blues-rock bands to emphasise musical virtuosity and lengthy jazz-style improvisation sessions. Their US hit singles include "Sunshine of Your Love" (No. 5, 1968), "White Room" (No. 6, 1968) and "Crossroads" (No. 28, 1969) – a live version of Robert Johnson's "Cross Road Blues". Though Cream were hailed as one of the greatest groups of its day, and the adulation of Clapton as a guitar legend reached new heights, the supergroup was short-lived. Drug and alcohol use escalated tension between the three members, and conflicts between Bruce and Baker eventually led to Cream's demise. A strongly critical "Rolling Stone" review of a concert of the group's second headlining US tour was another significant factor in the trio's demise, and it affected Clapton profoundly. Clapton has also credited "Music from Big Pink", the debut album of The Band, and its revolutionary Americana sound as influencing his decision to leave Cream. Cream's farewell album, "Goodbye", comprising live performances recorded at The Forum, Los Angeles, on 1968, was released shortly after Cream disbanded. It also spawned the studio single "Badge", co-written by Clapton and George Harrison. Clapton met Harrison and became close friends with him after the Beatles shared a bill with the Clapton-era Yardbirds at the London Palladium. In 1968, Clapton played the lead guitar solo on Harrison's "While My Guitar Gently Weeps", from the Beatles' self-titled double album (also known as the "White Album"). Harrison's debut solo album, "Wonderwall Music" (1968), became the first of many Harrison solo records to include Clapton on guitar. Clapton went largely uncredited for his contributions to Harrison's albums due to contractual restraints, and Harrison was credited as "L'Angelo Misterioso" for his contributions to the song "Badge" on "Goodbye". The pair would often play live together as each other's guest. A year after Harrison's death in 2001, Clapton was musical director for the Concert for George. In January 1969, when the Beatles were recording and filming what became "Let It Be", tensions became so acute that Harrison quit the group for several days, prompting John Lennon to suggest they complete the project with Clapton if Harrison did not return. Michael Lindsay-Hogg, television director of the recording sessions for "Let It Be", later recalled: "I was there when John mentioned Clapton – but that wasn't going to happen. Would Eric have become a Beatle? No. Paul [McCartney] didn't want to go there. He didn't want them to break up. Then George came back." Clapton was on good terms with all four of the Beatles; in December 1968 he had played with Lennon at "The Rolling Stones Rock and Roll Circus" as part of the one-off group the Dirty Mac. Cream briefly reunited in 1993 to perform at the ceremony inducting them into the Rock and Roll Hall of Fame. A full reunion took place in May 2005, with Clapton, Bruce and Baker playing four sold-out concerts at London's Royal Albert Hall, and three shows at New York's Madison Square Garden that October. Recordings from the London shows, "Royal Albert Hall London May 2-3-5-6, 2005", were released on CD, LP and DVD in late 2005. Clapton's next group, Blind Faith, formed in 1969, was composed of Cream drummer Ginger Baker, Steve Winwood of Traffic, and Ric Grech of Family, and yielded one LP and one arena-circuit tour. The supergroup debuted before 100,000 fans in London's Hyde Park on 1969. They performed several dates in Scandinavia and began a sold-out American tour in July before their only album was released. The LP "Blind Faith" consisted of just six songs, one of them the hit "Can't Find My Way Home". Another, "Presence of the Lord", is the first song credited solely to Clapton. The album's jacket image of a topless pubescent girl was deemed controversial in the US and was replaced by a photograph of the band. Blind Faith dissolved after less than seven months. Clapton subsequently toured as a sideman for an act that had opened for Blind Faith, Delaney and Bonnie and Friends. He also performed as a member of Lennon's Plastic Ono Band at the Toronto Rock and Roll Revival in September 1969, a recording from which was released as the album "Live Peace in Toronto 1969". On 30 September, Clapton played lead guitar on Lennon's second solo single, "Cold Turkey". On 15 December that year, Clapton performed with Lennon, Harrison and others as the Plastic Ono Supergroup at a fundraiser for UNICEF in London. Delaney Bramlett encouraged Clapton in his singing and writing. Using the Bramletts' backing group and an all-star cast of session players (including Leon Russell and Stephen Stills), Clapton recorded his first solo album during two brief tour hiatuses, titled "Eric Clapton". Delaney Bramlett co-wrote six of the songs with Clapton, also producing the LP, and Bonnie Bramlett co-wrote "Let It Rain". The album yielded the unexpected US No. 18 hit, J. J. Cale's "After Midnight". Clapton also worked with much of Delaney and Bonnie's band to record George Harrison's "All Things Must Pass" in spring 1970. During this period, Clapton also recorded with artists such as Dr. John, Leon Russell, Billy Preston, Ringo Starr and Dave Mason. With Chicago blues artist Howlin' Wolf, he recorded "The London Howlin' Wolf Sessions", that also included long-time Wolf guitarist Hubert Sumlin and members of the Rolling Stones, Winwood and Starr. Despite the superstar line-up, critic Cub Koda noted: "Even Eric Clapton, who usually welcomes any chance to play with one of his idols, has criticized this album repeatedly in interviews, which speaks volumes in and of itself." Other noted recordings from this period include Clapton's guitar work on "Go Back Home" from Stephen Stills' self-titled first solo album. With the intention of counteracting the "star" cult faction that had begun to form around him, Clapton assembled a new band composed of Delaney and Bonnie's former rhythm section, Bobby Whitlock as keyboardist and vocalist, Carl Radle as the bassist, and drummer Jim Gordon, with Clapton playing guitar. It was his intention to show that he need not fill a starring role, and functioned well as a member of an ensemble. During this period, Clapton was increasingly influenced by The Band and their 1968 album "Music from Big Pink", saying: "What I appreciated about the Band was that they were more concerned with songs and singing. They would have three- and four-part harmonies, and the guitar was put back into perspective as being accompaniment. That suited me well, because I had gotten so tired of the virtuosity – or "pseudo"-virtuosity – thing of long, boring guitar solos just because they were expected. The Band brought things back into perspective. The priority was the song." The band was originally called "Eric Clapton and Friends". The eventual name was a fluke that occurred when the band's provisional name of "Del and the Dynamos" was misread as Derek and the Dominos. Clapton's biography states that Tony Ashton of Ashton, Gardner and Dyke told Clapton to call the band "Del and the Dominos", since "Del" was his nickname for Eric Clapton. Del and Eric were combined and the final name became "Derek and the Dominos". Clapton's close friendship with George Harrison brought him into contact with Harrison's wife, Pattie Boyd, with whom he became deeply infatuated. When she spurned his advances, Clapton's unrequited affections prompted most of the material for the Dominos' album "Layla and Other Assorted Love Songs" (1970). Heavily blues-influenced, the album features the twin lead guitars of Clapton and Duane Allman, with Allman's slide guitar as a key ingredient of the sound. Working at Criteria Studios in Miami with Atlantic Records producer Tom Dowd, who had worked with Clapton on Cream's "Disraeli Gears", the band recorded a double album. The album contained the hit love song "Layla", inspired by the classical poet of Persian literature, Nizami Ganjavi's "The Story of Layla and Majnun", a copy of which Ian Dallas had given to Clapton. The book moved Clapton profoundly, as it was the tale of a young man who fell hopelessly in love with a beautiful, unavailable woman and went crazy because he could not marry her. The two parts of "Layla" were recorded in separate sessions: the opening guitar section was recorded first, and for the second section, laid down a few weeks later, drummer Jim Gordon played the piano part for the melody, which he claimed to have written (though Bobby Whitlock stated that Rita Coolidge wrote it). The "Layla" LP was actually recorded by a five-piece version of the group, thanks to the unforeseen inclusion of guitarist Duane Allman of the Allman Brothers Band. A few days into the Layla sessions, Dowd – who was also producing the Allmans – invited Clapton to an Allman Brothers outdoor concert in Miami. The two guitarists met first on stage, then played all night in the studio, and became friends. Duane first added his slide guitar to "Tell the Truth" and "Nobody Knows You When You're Down and Out". In four days, the five-piece Dominos recorded "Key to the Highway", "Have You Ever Loved a Woman" (a blues standard popularised by Freddie King and others) and "Why Does Love Got to be So Sad?" In September, Duane briefly left the sessions for gigs with his own band, and the four-piece Dominos recorded "I Looked Away", "Bell Bottom Blues" and "Keep on Growing". Allman returned to record "I Am Yours", "Anyday" and "It's Too Late". On 9 September, they recorded Hendrix's "Little Wing" and the title track. The following day, the final track, "It's Too Late", was recorded. Tragedy dogged the group throughout its brief career. During the sessions, Clapton was devastated by news of the death of Jimi Hendrix; eight days previously the band had cut a cover of "Little Wing" as a tribute. On 1970, one day before Hendrix's death, Clapton had purchased a left-handed Fender Stratocaster that he had planned to give to Hendrix as a birthday gift. Adding to Clapton's woes, "Layla" received only lukewarm reviews upon release. The shaken group undertook a US tour without Allman, who had returned to the Allman Brothers Band. Despite Clapton's later admission that the tour took place amid a blizzard of drugs and alcohol, it resulted in the live double album "In Concert". Recording of a second Dominos studio album was underway when a clash of egos took place and Clapton walked out, thus disbanding the group. Allman was killed in a motorcycle accident on 1971. Clapton wrote later in his autobiography that he and Allman were inseparable during the "Layla" sessions in Florida; he talked about Allman as the "musical brother I'd never had but wished I did". Although Radle remained Clapton's bass player until the summer of 1979 (Radle died in May 1980 from the effects of alcohol and narcotics), it was not until 2003 that Clapton and Whitlock appeared together again; Clapton guested on Whitlock's appearance on the "Later with Jools Holland" show. Another tragic footnote to the Dominos story was the fate of drummer Jim Gordon, who was an undiagnosed schizophrenic and years later murdered his mother during a psychotic episode. Gordon was confined to 16-years-to-life imprisonment, later being moved to a mental institution, where he remains today. Clapton's career successes in the 1970s were in stark contrast with the struggles he coped with in his personal life, which was troubled by romantic longings and drug and alcohol addiction. Still infatuated with Boyd and torn by his friendship with Harrison, he withdrew from recording and touring to isolation in his Surrey residence as the Dominos broke up. He nursed a heroin addiction, which resulted in a lengthy career hiatus interrupted only by performing at Harrison's Concert for Bangladesh benefit shows in New York in August 1971; there, he passed out on stage, was revived, and managed to finish his performance. In January 1973, the Who's Pete Townshend organised a comeback concert for Clapton at London's Rainbow Theatre, titled the "Rainbow Concert", to help Clapton kick his addiction. Clapton returned the favour by playing "The Preacher" in Ken Russell's film version of the Who's "Tommy" in 1975. His appearance in the film (performing "Eyesight to the Blind") is notable as he is clearly wearing a fake beard in some shots, the result of deciding to shave off his real beard after the initial takes in an attempt to force the director to remove his earlier scene from the film and leave the set. In 1974, Clapton started living with Boyd (they would not marry until 1979) and was no longer using heroin (although he gradually began to drink heavily). He assembled a low-key touring band that included Radle, Miami guitarist George Terry, keyboardist Dick Sims (who died in 2011), drummer Jamie Oldaker, and vocalists Yvonne Elliman and Marcy Levy (also known as Marcella Detroit). With this band Clapton recorded "461 Ocean Boulevard" (1974), an album with an emphasis on more compact songs and fewer guitar solos; the cover version of "I Shot the Sheriff" was Clapton's first number one hit and was important in bringing reggae and the music of Bob Marley to a wider audience. The 1975 album "There's One in Every Crowd" continued this trend. The album's original title, "The World's Greatest Guitar Player (There's One in Every Crowd)", was changed before pressing, as it was felt its ironic intention would be misunderstood. The band toured the world and subsequently released the 1975 live LP "E. C. Was Here". Clapton continued to release albums and toured regularly. Highlights of the period include "No Reason to Cry" (a collaboration with Bob Dylan and The Band); "Slowhand", which contained "Wonderful Tonight" and a second J. J. Cale cover, "Cocaine". In 1976 he performed as one of a string of notable guests at the farewell performance of The Band, filmed in a Martin Scorsese documentary titled "The Last Waltz". In 1981, Clapton was invited by producer Martin Lewis to appear at the Amnesty International benefit The Secret Policeman's Other Ball in London. Clapton accepted the invitation and teamed up with Jeff Beck to perform a series of duets – reportedly their first ever billed stage collaboration. Three of the performances were released on the album of the show, and one of the songs appeared in the film. The performances at London's Drury Lane theatre heralded a return to form and prominence for Clapton in the new decade. Many factors had influenced Clapton's comeback, including his "deepening commitment to Christianity", to which he had converted prior to his heroin addiction. After calling his manager and admitting he was an alcoholic, Clapton flew to Minneapolis–Saint Paul in January 1982 and checked in at Hazelden Treatment Center, located in Center City, Minnesota. On the flight over, Clapton indulged in a large number of drinks, for fear he would never be able to drink again. Clapton wrote in his autobiography: After being discharged, it was recommended by doctors of Hazelden that Clapton not partake in any activities that would act as triggers for his alcoholism or stress. A few months after his discharge, Clapton began working on his next album, against doctors' orders. Working with Tom Dowd, he produced what he thought as his "most forced" album to date, "Money and Cigarettes". Clapton chose the name of the album "because that's all I saw myself having left" after his first rehabilitation from alcoholism. In 1984 he performed on former Pink Floyd member Roger Waters' solo album "The Pros and Cons of Hitch Hiking", and participated in the supporting tour. Since then Waters and Clapton have had a close relationship. In 2005 they performed together for the Tsunami Relief Fund. In 2006 they performed at the Highclere Castle, in aid of the Countryside Alliance, playing two set pieces of "Wish You Were Here" and "Comfortably Numb". Clapton, now a regular charity performer, played at the Live Aid concert at John F. Kennedy Stadium in Philadelphia on 13 July 1985, playing with Phil Collins, Tim Renwick, Chris Stainton, Jamie Oldaker, Marcy Levy, Shaun Murphy and Donald 'Duck' Dunn. When offered a slot close to peak viewing hours, he was apparently flattered. As Clapton recovered from his addictions, his album output continued in the 1980s, including two produced with Phil Collins, 1985's "Behind the Sun", which produced the hits "Forever Man" and "She's Waiting", and 1986's "August". "August" was suffused with Collins's trademark drum and horn sound, and became Clapton's biggest seller in the UK to date, matching his highest chart position, number 3. The album's first track, the hit "It's in the Way That You Use It", appeared in the Tom Cruise–Paul Newman film "The Color of Money". The horn-driven "Run" echoed Collins' "Sussudio" and other work, while "Tearing Us Apart" (with Tina Turner) and "Miss You" continued Clapton's more angry sound. This rebound kicked off Clapton's two-year period of touring with Collins and their "August" collaborators, bassist Nathan East and keyboard player/songwriter Greg Phillinganes. While on tour for "August", two concert videos were recorded of the four-man band: "Eric Clapton Live from Montreux" and "Eric Clapton and Friends". Clapton later remade "After Midnight" as a single and a promotional track for the Michelob beer brand, which had also used earlier songs by Collins and Steve Winwood. Clapton won a British Academy Television Award for his collaboration with Michael Kamen on the score for the 1985 BBC Television thriller series "Edge of Darkness". At the 1987 Brit Awards in London, Clapton was awarded the prize for Outstanding Contribution to Music. Also in 1987, he played on George Harrison's album "Cloud Nine", contributing guitar to "Cloud 9", "That's What It Takes", "Devil's Radio" and "Wreck of the Hesperus." Clapton also got together with the Bee Gees for charity. The supergroup called itself the Bunburys, and recorded a charity album with the proceeds going to the Bunbury Cricket Club in Cheshire, which plays exhibition cricket matches to raise money for nonprofit organisations in England. The Bunburys recorded three songs for "The Bunbury Tails": "We're the Bunburys", "Bunbury Afternoon" and "Fight (No Matter How Long)". The last song also appeared on "The 1988 Summer Olympics Album" and went to No. 8 on the rock music chart. Clapton played at the cricket club's 25th anniversary celebrations in 2011, which were held at London's Grosvenor House Hotel. In 1988 he played with Dire Straits and Elton John at the Nelson Mandela 70th Birthday Tribute at Wembley Stadium and the Prince's Trust rock gala at the Royal Albert Hall. In 1989, Clapton released "Journeyman", an album that covered a wide range of styles, including blues, jazz, soul and pop. Collaborators included George Harrison, Phil Collins, Daryl Hall, Chaka Khan, Mick Jones, David Sanborn and Robert Cray. The song “Bad Love” was released as a single and later won the Grammy Award for Best Male Rock Vocal Performance. The 1990s brought a series of 32 concerts to the Royal Albert Hall, such as the 24 Nights series of concerts that took place around January through February 1990, and February to March 1991. On 1990, Dire Straits, Clapton and Elton John made a guest appearance in the Nordoff-Robbins charity show held at Knebworth in England. On 1990, fellow blues guitarist Stevie Ray Vaughan, who was touring with Clapton, and three members of their road crew were killed in a helicopter crash between concerts. Then, on 1991, Clapton's four-year-old son, Conor, died after falling from the 53rd-floor window of his mother's friend's New York City apartment at 117 East 57th Street. Conor's funeral took place on 28 March at St Mary Magdalene's Church in Clapton's home village in Ripley, Surrey. In 1991 Clapton appeared on Richie Sambora's album, "Stranger in This Town", in a song dedicated to him, called "Mr. Bluesman". He contributed guitar and vocals to "Runaway Train", a duet with Elton John on the latter's "The One" album the following year. Clapton's grief was expressed in the song "Tears in Heaven", which was co-written by Will Jennings. At the 35th Annual Grammy Awards, Clapton received six Grammys for the single "Tears in Heaven" and his "Unplugged" album, for which Clapton performed live in front of a small audience on 16 January 1992 at Bray Film Studios in Windsor, Berkshire, England. The album reached number one on the "Billboard" 200, and is certified Diamond by the RIAA for selling over 10 million copies in the US. It reached number two in the UK Albums Chart and is certified four times platinum in the UK. On 9 September 1992, Clapton performed "Tears in Heaven" at the 1992 MTV Video Music Awards, and won the award for Best Male Video. In 1992 Clapton received the Ivor Novello Award for Lifetime Achievement from the British Academy of Songwriters, Composers and Authors. In October 1992 Clapton was among the dozens of artists performing at Bob Dylan's 30th Anniversary Concert Celebration. Recorded at Madison Square Garden in New York City, the live two-disk CD/DVD captured a show full of celebrities performing classic Dylan songs, with Clapton playing the lead on a nearly 7-minute version of Dylan's "Knockin' on Heaven's Door" as part of the finale. While Clapton played acoustic guitar on "Unplugged", his 1994 album "From the Cradle" contained new versions of old blues standards, highlighted by his electric guitar playing. In 1995, Clapton for the first and only time appeared on a UK No. 1 single, collaborating with Chrissie Hynde, Cher and Neneh Cherry on a solo to a cover of "Love Can Build a Bridge" released in aid of the British charity telethon Comic Relief. On 12 September 1996 Clapton played a party for Armani at New York City's Lexington Armory with Greg Phillinganes, Nathan East and Steve Gadd. Sheryl Crow appeared on one number, performing "Tearing Us Apart", a track from "August", which was first performed by Tina Turner during the Prince's Trust All-Star Rock show in 1986. It was Clapton's sole US appearance that year, following the open-air concert held at Hyde Park. The concert was taped and the footage was released both on VHS video cassette and later, on DVD. Clapton's 1996 recording of the Wayne Kirkpatrick/Gordon Kennedy/Tommy Sims tune "Change the World" (on the soundtrack of the film "Phenomenon") won the Grammy Award for Song of the Year in 1997, the same year he recorded "Retail Therapy" (an album of electronic music with Simon Climie under the pseudonym TDF). On 15 September 1997, Clapton appeared at the "Music for Montserrat" concert at the Royal Albert Hall, London, performing "Layla" and "Same Old Blues" before finishing with "Hey Jude" alongside fellow English artists Paul McCartney, Elton John, Phil Collins, Mark Knopfler and Sting. That autumn, Clapton released the album "Pilgrim", the first record containing new material for almost a decade. In 1996 Clapton had a relationship with singer-songwriter Sheryl Crow. They remain friends, and Clapton appeared as a guest on Crow's Central Park Concert. The duo performed a Cream hit single, "White Room". Later, Clapton and Crow performed an alternate version of "Tulsa Time" with other guitar legends at the Crossroads Guitar Festival in June 2007 as well as Robert Johnson's blues classic "Crossroads" at London's Hyde Park in August 2008 with John Mayer and Robert Randolph. At the 41st Annual Grammy Awards on 24 February 1999, Clapton received his third Grammy Award for Best Male Pop Vocal Performance, for his song "My Father's Eyes". In October 1999, the compilation album, "", was released, which contained a new song, "Blue Eyes Blue", that also appears in soundtrack for the film, "Runaway Bride". Clapton finished the twentieth century with collaborations with Carlos Santana and B.B. King. Clapton looked up to King and had always wanted to make an album with him, while King said of Clapton, "I admire the man. I think he's No. 1 in rock 'n' roll as a guitarist and No. 1 as a great person." Clapton released the album "Reptile" in March 2001. One month after the 11 September attacks, Clapton appeared at the Concert for New York City, performing alongside Buddy Guy. An event marking the Golden Jubilee of Queen Elizabeth II in June 2002, Clapton performed "Layla" and "While My Guitar Gently Weeps" at the Party at the Palace concert in the grounds of Buckingham Palace. On 29 November 2002, the Concert for George was held at the Royal Albert Hall, a tribute to George Harrison, who had died a year earlier of lung cancer. Clapton was a performer and the musical director. The concert included Paul McCartney, Ringo Starr, Jeff Lynne, Tom Petty and the Heartbreakers, Ravi Shankar, Gary Brooker, Billy Preston, Joe Brown and Dhani Harrison. In 2004, Clapton released two albums of covers of songs by bluesman Robert Johnson, "Me and Mr. Johnson" and "Sessions for Robert J". Guitarist Doyle Bramhall II worked on the album with Clapton (after opening Clapton's 2001 tour with his band Smokestack) and would join him on his 2004 tour. In 2004, "Rolling Stone" ranked Clapton No. 53 on their list of the "100 Greatest Artists of All Time". Other media appearances include the Toots & the Maytals Grammy award-winning album "True Love", where he played guitar on the track "Pressure Drop". On 22 January 2005, Clapton performed in the Tsunami Relief Concert held at the Millennium Stadium in Cardiff, in aid of the victims of the 2004 Indian Ocean earthquake. In May 2005, Clapton, Jack Bruce, and Ginger Baker reunited as Cream for a series of concerts at the Royal Albert Hall in London. Concert recordings were released on CD and DVD. Later, Cream performed in New York at Madison Square Garden. Clapton's first album of new original material in nearly five years, "Back Home", was released on Reprise Records on . A collaboration with guitarist J. J. Cale, "The Road to Escondido", was released on 2006, featuring Derek Trucks and Billy Preston (Preston had also been a part of Clapton's 2004 touring band). He invited Trucks to join his band for his 2006–2007 world tour. Bramhall remained, giving Clapton three elite guitarists in his band, allowing him to revisit many Derek and the Dominos songs that he hadn't played in decades. Trucks became the third member of the Allman Brothers Band to tour supporting Clapton, the second being pianist/keyboardist Chuck Leavell, who appeared on the "MTV Unplugged" album and the "24 Nights" performances at the Royal Albert Hall, London in 1990 and 1991, as well as Clapton's 1992 US tour. On 20 May 2006, Clapton performed with Queen drummer Roger Taylor and former Pink Floyd bassist/songwriter Roger Waters at Highclere Castle, Hampshire, in support of the Countryside Alliance, which promotes issues relating to the British countryside. On 2006, Clapton made a guest appearance at the Bob Dylan concert in Columbus, Ohio, playing guitar on three songs in Jimmie Vaughan's opening act. The chemistry between Trucks and Clapton convinced him to invite the Derek Trucks Band to open for Clapton's set at his 2007 Crossroads Guitar Festival. Trucks remained on set afterward and performed with Clapton's band throughout his performances. The rights to Clapton's official memoirs, written by Christopher Simon Sykes and published in 2007, were sold at the 2005 Frankfurt Book Fair for . In 2007 Clapton learned more about his father, a Canadian soldier who left the UK after the war. Although Clapton's grandparents eventually told him the truth about his parentage, he only knew that his father's name was Edward Fryer. This was a source of disquiet for Clapton, as witnessed by his 1998 song "My Father's Eyes". A Montreal journalist named Michael Woloschuk researched Canadian Armed Forces service records and tracked down members of Fryer's family, and finally pieced together the story. He learned that Clapton's father was Edward Walter Fryer, born 1920, in Montreal and died in Newmarket, Ontario. Fryer was a musician (piano and saxophone) and a lifelong drifter who was married several times, had several children, and apparently never knew that he was the father of Eric Clapton. Clapton thanked Woloschuk in an encounter at Macdonald–Cartier Airport, in Ottawa, Ontario, Canada. On 26 February 2008, it was reported that North Korean officials had invited Clapton to play a concert in the communist state. Clapton's management received the invitation and passed it on to the singer, who agreed in principle and suggested it take place sometime in 2009. Kristen Foster, a spokesperson, said, "Eric Clapton receives numerous offers to play in countries around the world", and "[t]here is no agreement whatsoever for him to play in North Korea". In February 2008 Clapton performed with his long-time friend Steve Winwood at Madison Square Garden and guested on his recorded single, "Dirty City", on Winwood's album "Nine Lives". The two former Blind Faith bandmates met again for a series of 14 concerts throughout the United States in June 2009. Clapton's 2008 Summer Tour began on at the Ford Amphitheatre, Tampa, Florida, and then moved to Canada, Ireland, England, Norway, Iceland, Denmark, Poland, Germany, and Monaco. On 2008, he headlined Saturday night for Hard Rock Calling 2008 in London's Hyde Park (previously Hyde Park Calling) with support from Sheryl Crow and John Mayer. In September 2008 Clapton performed at a private charity fundraiser for The Countryside Alliance at Floridita in Soho, London, that included such guests as the London Mayor Boris Johnson. In March 2009, the Allman Brothers Band (amongst many notable guests) celebrated their 40th year, dedicating their string of concerts to the late Duane Allman on their annual run at the Beacon Theatre. Eric Clapton was one of the performers, with drummer Butch Trucks remarking that the performance was not the typical Allman Brothers experience, given the number and musical styles of the guests who were invited to perform. Songs like "In Memory of Elizabeth Reed" were punctuated with others, including "The Weight", with Levon Helm; Johnny Winter sitting in on Hendrix's "Red House"; and "Layla". On 2009 Clapton appeared at the Royal Albert Hall, playing "Further on Up the Road" with Joe Bonamassa. Clapton was scheduled to perform at the Rock and Roll Hall of Fame's 25th anniversary concert in Madison Square Garden on 2009, but cancelled due to gallstone surgery. Van Morrison (who also cancelled) said in an interview that he and Clapton were to do a "couple of songs", but that they would do something else together at "some other stage of the game". Clapton performed a two-night show with Jeff Beck at the O2 Arena in London on 2010. The two former Yardbirds extended their 2010 tour with stops at Madison Square Garden, the Air Canada Centre in Toronto, and the Bell Centre in Montreal. Clapton performed a series of concerts in 11 cities throughout the United States from to 2010, including Roger Daltrey as opening act. His third European tour with Steve Winwood began on and ended , including Tom Norris as opening act. He then began a short North American tour lasting from to , starting with his third Crossroads Guitar Festival on at Toyota Park in Bridgeview, Illinois. Clapton released a new studio album, "Clapton", on 2010 in the United Kingdom and 28 September 2010 in the United States. On 2010, Clapton performed as guest on the Prince's Trust rock gala held at the Royal Albert Hall, supported by the house band for the evening, which included Jools Holland, Midge Ure and Mark King. On 24 June 2011, Clapton was in concert with Pino Daniele in Cava de' Tirreni stadium before performing a series of concerts in South America from 6 to 16 October 2011. He spent November and December 2011 touring Japan with Steve Winwood, playing 13 shows in various cities throughout the country. On 24 February 2012 Clapton, Keith Richards, Gary Clark Jr., Derek Trucks, Doyle Bramhall II, Kim Wilson and other artists performed together in the Howlin' For Hubert Tribute concert held at the Apollo Theater of New York honouring blues guitarist Hubert Sumlin who died at age 80 on 4 December 2011. On 29 November 2012, Clapton joined The Rolling Stones at London's O2 Arena during the band's second of five arena dates celebrating their 50th anniversary. On 12 December, Clapton performed at Madison Square Garden, broadcast live via television, radio, cinemas and the Internet across six continents. In January 2013, Surfdog Records announced a signed deal with Clapton for the release of his forthcoming album "Old Sock" on 12 March. On 8 April 2013, Eric and Hard Rock International launched the limited-edition Eric Clapton Artist Spotlight merchandise programme benefiting Crossroads Centre Antigua. Clapton toured the US and Europe from 14 March to 19 June 2013 to celebrate 50 years as a professional musician. On 28 February 2013, Clapton announced his intention to stop touring in 2015 due to hassles with travel. On 15 October 2013, Clapton's popular "Unplugged" album and concert DVD were re-released, titled "Unplugged: Expanded & Remastered." The album includes the original 14 tracks, remastered, as well as 6 additional tracks, including 2 versions of "My Father's Eyes". The DVD includes a restored version of the concert, as well as over 60 minutes of unseen footage from the rehearsal. On 13 and 14 November 2013, Clapton headlined the final two evenings of the "Baloise Session", an annual indoor music festival in Basel, Switzerland. On 20 November 2013, Warner Bros released Crossroads Guitar Festival 2013 in CD/DVD/Blu-ray. On 30 April 2014, Clapton announced the release of "" as an homage to J. J. Cale who died on 26 July 2013. This tribute album is named after the 1972 single "Call Me the Breeze" and comprises 16 Cale songs performed by Clapton, Mark Knopfler, John Mayer, Willie Nelson, Tom Petty and others. On 21 June 2014, Clapton abruptly walked off stage during a concert at the Glasgow Hydro. Although he did return to perform one final song, thousands of fans were upset by the lack of explanation from Clapton or the venue and booed after the concert ended around 40 minutes before advertised to finish. Both Clapton and the venue apologised the next day, blaming 'technical difficulties' for making sound conditions 'unbearable' for Clapton on stage. A week later he confirmed his retirement plans attributing his decision to the road being "unbearable" in addition to "odd ailments" that may force him to put down his guitar permanently. In a 2016 interview with "Classic Rock" magazine, Clapton revealed that he had been diagnosed with peripheral neuropathy in 2013, a condition involving damage to peripheral nerves that typically causes stabbing, burning, or tingling pain in the arms and legs. Clapton performed two shows at Madison Square Garden in New York on 1 and 3 May 2015 followed by a 7-night residency at London's Royal Albert Hall from 14 to 23 May 2015 to celebrate his 70th birthday on 30 March. The shows also mark 50 years since Clapton first played at the Royal Albert Hall – his debut was on 7 December 1964 when he performed as part of The Yardbirds for the BBC's "Top Beat Show". The concert film, "Slowhand at 70 – Live at the Royal Albert Hall", was released by Eagle Rock Entertainment on 13 November 2015 on DVD, CD, Blu-ray and LP. The 2-night concerts in the US marked the 46th anniversary since Clapton, with Cream, opened the "new" Madison Square Garden on 2 November 1968. Clapton has performed more times at Madison Square Garden than any other US venue, a total of 45 times. On 20 May 2016, Clapton released his twenty-third studio album "I Still Do". On 30 September 2016 the live-album "Live in San Diego" was released. In August 2018, Clapton announced that he had recorded his twenty-fourth studio album, "Happy Xmas", which consists of blues-tinged interpretations of Christmas songs, with the album released on 12 October. On 25 June 2019, "The New York Times Magazine" listed Eric Clapton among hundreds of artists whose material was reportedly destroyed in the 2008 Universal fire. Clapton cites Muddy Waters, Freddie King, B.B. King, Albert King, Buddy Guy, and Hubert Sumlin as guitar playing influences. In his 2007 autobiography, Clapton referred to Muddy Waters as "the father figure I never really had." Until his death in 1983, Waters was a part of Clapton's life. "When I got to know Muddy, unfortunately, my drinking career was in full sway." In 2000 Clapton collaborated with B.B. King in their album "Riding with the King", with the music video for the title track seeing Clapton as the chauffeur with one of his idols in the back seat. Clapton has said that blues musician Robert Johnson is his single most important influence. In 2004 Clapton released CDs and DVDs entitled "Sessions for Robert Johnson", comprising covers of Robert Johnson songs using electric and acoustic guitars. Clapton co-authored with others the book "Discovering Robert Johnson", in which Clapton said Johnson was: Clapton also singled out Buddy Holly as an influence. "The Chirping Crickets" was the first album Clapton ever bought; he later saw Holly on "Sunday Night at the London Palladium". In his autobiography, Clapton recounted the first time he saw Holly and his Fender, saying, "I thought I'd died and gone to heaven ... it was like seeing an instrument from outer space and I said to myself: 'That's the future – that's what I want. In the 2017 documentary film, "Eric Clapton: Life in 12 Bars", Clapton cites Bismillah Khan as an influence: "I wanted my guitar to sound like his reed instrument". In the same documentary he also cited harmonica player Little Walter as an influence: "The sound he made with the harmonica playing through an amplifier. It was thick and fat and very melodic". Clapton has been referred to as one of the most important and influential guitarists of all time. Clapton is the only three-time inductee to the Rock and Roll Hall of Fame: once as a solo artist, and separately as a member of the Yardbirds and Cream. He ranked second in "Rolling Stone" magazine's list of the "100 Greatest Guitarists of All Time" and fourth in "Gibson's" Top 50 Guitarists of All Time. In 2011, "The Guardian" attributed the creation of the cult of the guitar hero to Clapton, ranking it number seven on their list of the 50 key events in rock music history; Elias Leight of "Rolling Stone" writes that Clapton "influenced recording techniques as well as guitar-playing technique". During recording sessions with John Mayall's group, Clapton was frustrated by technicians "that just came up to your amp with the microphone and just stuck it two inches away from the front of the amplifier. It seemed to me that if you wanted to get the atmosphere we were getting in the clubs, you needed it to sound like you were in the audience 10 feet away, not three inches". Clapton then moved the microphones, with Pink Floyd's Roger Waters stating, "That changed everything. Before Eric, guitar playing in England had been Hank Marvin of the Shadows, very simple, not much technique. Suddenly we heard something completely different. The records sounded unlike anything we had heard before". In 2012, Clapton was among the British cultural icons selected by artist Sir Peter Blake to appear in a new version of his most famous artwork – the Beatles' "Sgt. Pepper's Lonely Hearts Club Band" album cover – to celebrate the British cultural figures of his life that he most admires to mark his 80th birthday. Robert Christgau, in a dissenting appraisal of Clapton's legacy, writes: "A promiscuous sideman whose monklike aura has never diminished his extravagant appetites, Clapton likes to get paid, and he's amassed a discography that for an artist of his caliber is remarkably undistinguished. In his self-protective self-deprecation he often attributes this to his own laziness or his need for a catalyst, but it's also guitar hero's disease: like many other guys whose hand-ear coordination is off the curve, he's a casual tunesmith and a corny lyricist, and his band concepts are chronically hit-or-miss." Clapton's choice of electric guitars has been as notable as the man himself; like Hank Marvin, the Beatles and Jimi Hendrix, Clapton exerted a crucial and widespread influence in popularising particular models of electric guitar. With the Yardbirds, Clapton played a Fender Telecaster, a Fender Jazzmaster, a double-cutaway Gretsch 6120, and a 1964 Cherry-Red Gibson ES-335. He became exclusively a Gibson player for a period beginning in mid-1965, when he purchased a used sunburst Gibson Les Paul guitar from a guitar store in London. Clapton commented on the slim profile of the neck, which would indicate it was a 1960 model. Early during his stint in Cream, Clapton's first Les Paul Standard was stolen. He continued to play Les Pauls exclusively with Cream (one bought from Andy Summers was almost identical to the stolen guitar) until 1967, when he acquired his most famous guitar in this period, a 1964 Gibson SG, dubbed "the Fool". Clapton used both the Les Paul and the SG to create his self-described "woman tone". He explained in a 1967 interview, "I am playing more smoothly now. I’m developing what I call my 'woman tone.' It's a sweet sound, something like the solo on 'I Feel Free'." Writer Michael Dregni describes it as "thick yet piercing, overdriven yet smooth, distorted yet creamy". The tone is achieved by a combination of tone control settings on the guitars and Clapton's Marshall JTM45 amplifier. "Vintage Guitar" magazine identifies "the opening riff and solo of 'Sunshine of Your Love' are arguably the best illustrations of full-blown woman tone". Clapton's "Fool" acquired its name from its distinctive psychedelic paint job, created by the visual art collective also known as the Fool (just before Cream's first US appearance in 1967, Clapton's SG, Bruce's Fender VI, and Baker's drum head were all repainted in psychedelic designs). In 1968 Clapton bought a Gibson Firebird and started using the 1964 Cherry-Red Gibson ES-335 again. The aforementioned 1964 ES-335 had a storied career. Clapton used it at the last Cream show in November 1968 as well as with Blind Faith, played it sparingly for slide pieces in the 1970s, used it on "Hard Times" from "Journeyman", the Hyde Park live concert of 1996, and the "From the Cradle" sessions and tour of 1994–95. It was sold for US$847,500 at a 2004 auction. Gibson produced a limited run of 250 "Crossroads 335" replicas. The 335 was only the second electric guitar Clapton bought. In July 1968 Clapton gave George Harrison a 1957 'goldtop' Gibson Les Paul that been refinished with a red colour, nicknamed Lucy. The following September, Clapton played the guitar on the Beatles' recording of "While My Guitar Gently Weeps". Lucy was stolen from Harrison, though later tracked down and returned to him – he lent it to Clapton for his 1973 comeback concert at the Rainbow. His SG "The Fool" found its way into the hands of George Harrison's friend Jackie Lomax, who subsequently sold it to musician Todd Rundgren for US$500 in 1972. Rundgren restored the guitar and nicknamed it "Sunny", after "Sunshine of Your Love". He retained it until 2000, when he sold it at an auction for US$150,000. At the 1969 Blind Faith concert in Hyde Park, London Clapton played a Fender Custom Telecaster, which was fitted with "Brownie"s neck. In late 1969 Clapton made the switch to the Fender Stratocaster. "I had a lot of influences when I took up the Strat. First there was Buddy Holly, and Buddy Guy. Hank Marvin was the first well known person over here in England who was using one, but that wasn't really my kind of music. Steve Winwood had so much credibility, and when he started playing one, I thought, oh, if he can do it, I can do it". The first—used during the recording of "Eric Clapton"—was "Brownie", which in 1973 became the backup to the most famous of all Clapton's guitars, "Blackie". In November 1970 Eric bought six Fender Stratocasters from the Sho-bud guitar shop in Nashville, Tennessee while on tour with the Dominos. He gave one each to George Harrison, Steve Winwood, and Pete Townshend. Clapton assembled the best components of the remaining three to create "Blackie", which was his favourite stage guitar until its retirement in 1985. It was first played live 1973 at the Rainbow Concert. Clapton called the 1956/57 Strat a "mongrel". On 2004, Clapton sold "Blackie" at Christie's Auction House, New York, for US$959,500 to raise funds for his Crossroads Centre for drug and alcohol addictions. "Brownie" is now on display at the Experience Music Project. The Fender Custom Shop has since produced a limited run of 275 'Blackie' replicas, correct in every detail right down to the 'Duck Brothers' flight case, and artificially aged using Fender's 'Relic' process to simulate years of hard wear. One was presented to Clapton upon the model's release and was used for three numbers during a concert at the Royal Albert Hall on 2006. In 1979 Clapton gave his signed Fender Lead II guitar to the Hard Rock Cafe in London to designate his favourite bar stool. Pete Townshend also donated his own Gibson Les Paul guitar, with a note attached: "Mine's as good as his! Love, Pete". Signature guitars in Clapton's honour are made by Fender and C.F. Martin & Company. In 1988 Fender introduction his signature Eric Clapton Stratocaster. Several signature-model 000-sized acoustic guitars made by Martin. The first, of these, introduced in 1995, was a limited edition 000-42EC Eric Clapton signature model with a production run of 461. For the single "Change the World" (1996) and the album "Pilgrim" (1998) he used a Martin 000-28 EC Eric Clapton signature model, which he subsequently gave to guitarist Paul Wassif. His 1939 000-42 Martin that he played on the "Unplugged" album sold for US$791,500 at auction. Clapton uses Ernie Ball Slinky and Super Slinky strings, gauge .10 to.46. His guitar technician for over thirty years was Lee Dickson. Clapton appeared in the movie version of "Tommy", the first full length rock opera, written by the Who. In the movie version, Clapton appeared as the Preacher, performing Sonny Boy Williamson's song, "Eyesight to the Blind". He appeared in "Blues Brothers 2000" as one of the Louisiana Gator Boys. In addition to being in the band, he had a small speaking role. Clapton has appeared in an advertisement for the Mercedes-Benz G-Wagen. In March 2007 Clapton appeared in an advertisement for RealNetwork's Rhapsody online music service. In 2010 Clapton started appearing as a spokesman for T-Mobile, advertising their MyTouch Fender cell phone. Clapton also appeared in the 2011 BBC documentary "Reggae Got Soul: The Story of Toots and the Maytals", which was described as "The untold story of one of the most influential artists ever to come out of Jamaica." When asked to describe God by their minister, the characters Eric Forman and Steven Hyde both drew an image of Clapton in the episode "Holy Crap!" of season two of "That '70s Show". Clapton appeared on the BBC's "Top Gear" in 2013, during Series 19 Episode 4 and was involved in testing the new Kia Cee'd. He was called upon to test the Cee'd's auxiliary input, which he tested by plugging in one of his guitars and playing several bars of his most famous hits. He was introduced by "Top Gear" host Jeremy Clarkson as a "local guitarist". In 2017, a documentary film titled "Eric Clapton: Life in 12 Bars" was directed by Lili Fini Zanuck. Clapton wrote the film score for Zanuck's 1991 film "Rush" and the two remained friends. In an interview for BBC News, Zanuck said that Clapton only agreed to participate if she directed it: Clapton briefly dated funk singer Betty Davis. He married Pattie Boyd in 1979, but their marriage was marred by his infidelities and domestic violence. In an interview with "The Sunday Times", Clapton admitted he was abusive to Boyd when he was a "full-blown" alcoholic. In 1984, while recording "Behind the Sun", Clapton began a relationship with Yvonne Kelly, the manager of AIR Studios Montserrat. Although both were married to other partners at the time, they had a daughter in January 1985. She was named Ruth Kelly Clapton, but her existence was kept from the public until the media realized she was his child in 1991. Clapton and Boyd tried unsuccessfully to have children, even trying in vitro fertilisation in 1984, but were faced instead with miscarriages. They divorced in 1989 following his affair with Italian model Lory Del Santo, who gave birth to their son, Conor, on 1986. Conor died on 20 March 1991 at the age of four after falling out of an open bedroom window on the 53rd floor of a Manhattan apartment building. In 1998, Clapton, then 53, met 22-year-old administrative assistant Melia McEnery in Columbus, Ohio, at a party given for him after a performance. He quietly dated her for a year, and went public with the relationship in 1999. They married on 2002 at St Mary Magdalene church in Clapton's birthplace, Ripley. They have three daughters, Julie Rose (born 2001), Ella May (born 2003), and Sophie Belle (born 2005) His grandson Isaac Eric Owen Bartlett was born in June 2013 to his oldest daughter Ruth and her husband Dean Bartlett. Clapton is a supporter of the Countryside Alliance, which promotes field sports and issues relating to the British countryside. He has played in concerts to raise funds for the organisation and publicly opposed the Labour Party's ban on fox hunting with the 2004 Hunting Act. A spokesperson for Clapton said, "Eric supports the Countryside Alliance. He does not hunt himself, but does enjoy rural pursuits such as fishing and shooting. He supports the Alliance's pursuit to scrap the ban on the basis that he disagrees with the state's interference with people's private pursuits". On 5 August 1976, Clapton provoked an uproar and lingering controversy when he spoke out against increasing immigration during a concert in Birmingham. Visibly intoxicated, Clapton voiced his support of controversial political candidate Enoch Powell, and announced on stage that "I think Enoch's right, I think we should send them all back. Stop Britain from becoming a black colony. Get the foreigners out. Get the wogs out. Get the coons out. Keep Britain white." "Keep Britain white" was, at the time, a slogan of the far-right National Front (NF). This incident, along with some controversial remarks made around the same time by David Bowie, as well as uses of Nazi-related imagery by punk artists Sid Vicious and Siouxsie Sioux, were the main catalysts for the creation of Rock Against Racism, with a concert on 30 April 1978. In an interview from October 1976 with "Sounds" magazine, Clapton said that he was not a political person and that his rambling remarks that night were not appropriate. In a 2004 interview with "Uncut", Clapton referred to Powell as "outrageously brave". He complained that the UK was "... inviting people in as cheap labour and then putting them in ghettos." In 2004, Clapton told an interviewer for "Scotland on Sunday", "There's no way I could be a racist. It would make no sense." In his 2007 autobiography, Clapton claimed to be "oblivious to it all." In a December 2007 interview with Melvyn Bragg on "The South Bank Show", Clapton told Bragg that he was not a racist but still believed Powell's comments were relevant. In 2009, "Surrey Life Magazine" ranked Clapton as number 17 in their list of richest Surrey residents, estimating his fortune at £120 million in assets. This was a combination of income, property, a £9 million yacht, "Va Bene" (previously owned by Bernie Ecclestone), his back music catalogue, his touring income, and his holding company Marshbrook Ltd, which had earned him £110 million since 1989. In 2003, he purchased a 50% share of gentleman's outfitters Cordings Piccadilly. At the time, owner Noll Uloth was trying to save the shop from closure and is reported to have contacted Clapton, his "best client"; within five minutes, Clapton replied with "I can't let this happen". Since the 1970s, Clapton considered himself a "car enthusiast" and often stated his passion for the Ferrari brand. Clapton currently owns or has owned a range of Ferraris, and when asked about his Ferrari collection in 1989, he said he liked the touring cars the company produces for road use and commented "if I had more space and if I had been wise I would have a huge collection by now and I would be a multi-multi-millionaire". In 2010, he explained that for him "Ferrari has always been the number one car" to own and drive, and that he always supported Ferrari on the road and in Formula One motor racing. In 2012, Ferrari honoured Clapton with the one-off special project car, the Ferrari SP12 EC. In July 2013 Clapton displayed it at the Goodwood Festival of Speed in England in the Michelin Supercar Run. In 2014, Clapton explained that Ferrari is still his favourite car brand. Among the other vehicles Clapton owns or has owned are a vintage Mini Cooper Radford that was a gift from George Harrison. In 1993, Clapton was appointed a director of Clouds House, a UK treatment centre for drug and alcohol dependence, and served on their board until 1997. Clapton also served on the board of directors for The Chemical Dependency Centre from 1994 until 1999. Both charities subsequently merged to become Action on Addiction in 2007. In 1998, he established the Crossroads Centre in Antigua to help others overcome their addictions to drugs and alcohol and is active in its management oversight and fundraising to the present day. Clapton has organised the Crossroads Guitar Festival in 1999, 2004, 2007, 2010 and 2013 to raise funds for this centre. In 1999, Clapton auctioned off some of his guitar collection to raise more than US$5 million for continuing support of the Crossroads Centre. A second guitar auction, including the "Cream" of Clapton's collection – as well as guitars donated by famous friends – was held on 2004. His Lowden acoustic guitar sold for US$41,825. The revenue garnered by this auction at Christie's was US$7,438,624. In 2011 he auctioned off over 150 items at a New York auction with proceeds going to his Crossroads Centre. Items included his guitar from the Cream reunion tour in 2005, speaker cabinets used in the early 1970s from his days with Derek and the Dominos, and some guitars from Jeff Beck, J. J. Cale, and Joe Bonamassa. In March 2011 he raised more than £1.3 million when he auctioned off 138 lots, made up of 75 guitars and 55 amps from his personal collection, including a 1984 Gibson hollow body guitar, a Gianni Versace suit from his 1990 concert at the Royal Albert Hall, and a replica of the famous Fender Stratocaster known as "Blackie", which fetched more than $30,000. All proceeds again went to Crossroads. Clapton has performed at the "Secret Policeman's Ball", a benefit show co-founded by Monty Python member John Cleese on behalf of Amnesty International. He made his first appearance at the 1981 show held in London's Theatre Royal, Drury Lane, and he subsequently became an activist. Clapton has collaborated with The Prince's Trust, the leading UK youth charity, which provides training, personal development, business start up support, mentoring, and advice. He has performed at the charity's rock concert numerous times since the 1980s, most recently in 2010. In 2008, he donated a song to Aid Still Required's CD to assist with the restoration of the devastation done to Southeast Asia from the 2004 tsunami. Clapton is a fan of English football club West Bromwich Albion. In 1982 he performed a concert before West Brom player John Wile's testimonial game at The Hawthorns. It has been reported that the club rejected his offer to invest cash in the club around this time. In the late 1970s Clapton positioned a West Brom scarf on the back cover of his album, "Backless". In the 1978–79 season Clapton sponsored West Brom's UEFA Cup home game against Turkish club Galatasaray. Clapton's music has appeared in dozens of movies and television shows as far back as 1973's "Mean Streets", which included the song "I Looked Away". Other appearances in media include in the "Miami Vice" series ("Wonderful Tonight", "Knock on Wood", "She's Waiting", and "Layla"), "Back to the Future" ("Heaven Is One Step Away"), "The Color of Money" ("It's In The Way That You Use It"), "Lethal Weapon 2" ("Knockin' On Heaven's Door"), "Goodfellas" ("Layla" and "Sunshine of Your Love"), "Freaks and Geeks" episode "I'm With the Band" ("Sunshine of Your Love", "White Room" and "Crossroads (song)"), "Friends" episodes "The One with the Proposal, Part 2" ("Wonderful Tonight") and "The One Where Rachel Has A Baby" ("River of Tears"), "School Of Rock" ("Sunshine Of Your Love)", "Men in Black III" ("Strange Brew"), "Captain Phillips (film)" ("Wonderful Tonight"), "" ("Lay Down Sally"), for which it was a significant part of the soundtrack, being played in the intro and twice more later on, "Good Girls Revolt" episode "The Year-Ender" ("White Room)" and "Rick and Morty" episode "The Vat of Acid Episode" ("It's In The Way That You Use It"). Both Opel and Vauxhall used the guitar riff from "Layla" in their advertising campaigns throughout 1987–95. In addition to his music appearing in media, Clapton has contributed to several movies by writing or co-writing the musical scores or contributing original songs. These movies include "Lethal Weapon" (co-written with Michael Kamen), "Communion", "Rush", "Phenomenon" ("Change the World"), and "Lethal Weapon 3" (co-wrote and co-performed "It's Probably Me" with Sting and "Runaway Train" with Elton John).
https://en.wikipedia.org/wiki?curid=10049
Etiology Etiology (pronounced ; alternatively: aetiology or ætiology) is the study of causation or origination. The word is derived from the Greek ("aitiología") "giving a reason for" (, "aitía", "cause"; and , "-logía"). More completely, etiology is the study of the causes, origins, or reasons behind the way that things are, or the way they function, or it can refer to the causes themselves. The word is commonly used in medicine (pertaining to causes of disease) and in philosophy, but also in physics, psychology, government, geography, spatial analysis, theology, and biology, in reference to the causes or origins of various phenomena. In the past, when many physical phenomena were not well understood or when histories were not recorded, myths often arose to provide etiologies. Thus, an etiological myth, or origin myth, is a myth that has arisen, been told over time or written to explain the origins of various social or natural phenomena. For example, Virgil's "Aeneid" is a national myth written to explain and glorify the origins of the Roman Empire. In theology, many religions have creation myths explaining the origins of the world or its relationship to believers. In medicine, the etiology of an illness or condition refers to the frequent studies to determine one or more factors that come together to cause the illness. Relatedly, when disease is widespread, epidemiological studies investigate what associated factors, such as location, sex, exposure to chemicals, and many others, make a population more or less likely to have an illness, condition, or disease, thus helping determine its etiology. Sometimes determining etiology is an imprecise process. In the past, the etiology of a common sailor's disease, scurvy, was long unknown. When large, ocean-going ships were built, sailors began to put to sea for long periods of time, and often lacked fresh fruit and vegetables. Without knowing the precise cause, Captain James Cook suspected scurvy was caused by the lack of vegetables in the diet. Based on his suspicion, he forced his crew to eat sauerkraut, a cabbage preparation, every day, and based upon the positive outcomes, he inferred that it prevented scurvy, even though he did not know precisely why. It took about another two hundred years to discover the precise etiology: the lack of vitamin C in a sailor's diet. The following are examples of intrinsic factors: An etiological myth, or origin myth, is a myth intended to explain the origins of cult practices, natural phenomena, proper names and the like. For example, the name Delphi and its associated deity, "Apollon Delphinios", are explained in the Homeric Hymn which tells of how Apollo, in the shape of a dolphin ('), propelled Cretans over the seas to make them his priests. While Delphi is actually related to the word ' ("womb"), many etiological myths are similarly based on folk etymology (the term "Amazon", for example). In the "Aeneid" (published circa 17 BC), Virgil claims the descent of Augustus Caesar's Julian clan from the hero Aeneas through his son Ascanius, also called Iulus. The story of Prometheus' sacrifice trick at Mecone in Hesiod's "Theogony" relates how Prometheus tricked Zeus into choosing the bones and fat of the first sacrificial animal rather than the meat to justify why, after a sacrifice, the Greeks offered the bones wrapped in fat to the gods while keeping the meat for themselves. In Ovid's "Pyramus and Thisbe", the origin of the color of mulberries is explained, as the white berries become stained red from the blood gushing forth from their double suicide.
https://en.wikipedia.org/wiki?curid=10055
Empirical formula In chemistry, the empirical formula of a chemical compound is the simplest positive integer ratio of atoms present in a compound. A simple example of this concept is that the empirical formula of sulfur monoxide, or SO, would simply be SO, as is the empirical formula of disulfur dioxide, S2O2. Thus, sulfur monoxide and disulfur dioxide, both compounds of sulphur and oxygen, have the same empirical formula. However, their molecular formulas, which express the number of atoms in each molecule of a chemical compound, are not the same. An empirical formula makes no mention of the arrangement or number of atoms. It is standard for many ionic compounds, like calcium chloride (CaCl2), and for macromolecules, such as silicon dioxide (SiO2). The molecular formula, on the other hand, shows the number of each type of atom in a molecule. The structural formula shows the arrangement of the molecule. It is also possible for different types of compounds to have equal empirical formulas. Samples are analyzed in specific elemental analysis tests to determine what percent of a particular element the sample is composed of. A chemical analysis of a sample of methyl acetate provides the following elemental data: 48.64% carbon (C), 8.16% hydrogen (H), and 43.20% oxygen (O). For the purposes of determining empirical formulas assume that we have 100 grams of the compound. If this is the case, the percentages will be equal to the mass of each element in grams. Thus, the empirical formula of methyl acetate is . This formula also happens to be methyl acetate's molecular formula.
https://en.wikipedia.org/wiki?curid=10065
Episcopal polity An episcopal polity is a hierarchical form of church governance ("ecclesiastical polity") in which the chief local authorities are called bishops. (The word "bishop" derives, via the British Latin and Vulgar Latin term "*ebiscopus"/"*biscopus", from the Ancient Greek "epískopos" meaning "overseer".) It is the structure used by many of the major Christian Churches and denominations, such as the Catholic, Eastern Orthodox, Oriental Orthodox, Church of the East, Anglican, and Lutheran churches or denominations, and other churches founded independently from these lineages. Churches with an episcopal polity are governed by bishops, practicing their authorities in the dioceses and conferences or synods. Their leadership is both sacramental and constitutional; as well as performing ordinations, confirmations, and consecrations, the bishop supervises the clergy within a local jurisdiction and is the representative both to secular structures and within the hierarchy of the church. Bishops are considered to derive their authority from an unbroken, personal apostolic succession from the Twelve Apostles of Jesus. Bishops with such authority are said to represent the historical episcopate or historic episcopate. Churches with this type of government usually believe that the Church requires episcopal government as described in the New Testament (see 1 Timothy 3 and 2 Timothy 1). In some systems, bishops may be subject to bishops holding a higher office (variously called archbishops, metropolitans, or patriarchs, depending upon the tradition). They also meet in councils or synods. These gatherings, subject to presidency by higher ranking bishops, usually make important decisions, though the synod or council may also be purely advisory. For much of the written history of institutional Christianity, episcopal government was the only known form of church organization. This changed at the Reformation. Many Protestant churches are now organized by either congregational or presbyterian church polities, both descended from the writings of John Calvin, a Protestant reformer working and writing independently following the break with the Catholic Church precipitated by The Ninety-Five Theses of Martin Luther. The definition of the word "episcopal" has variation among Christian traditions. There are subtle differences in governmental principles among episcopal churches at the present time. To some extent the separation of episcopal churches can be traced to these differences in ecclesiology, that is, their theological understanding of church and church governance. For some, "episcopal churches" are churches that use a hierarchy of bishops that regard themselves as being in an unbroken, personal apostolic succession. "Episcopal" is also commonly used to distinguish between the various organizational structures of denominations. For instance, "Presbyterian" (, presbútēs) is used to describe a church governed by a hierarchy of assemblies of elected elders, referred to as Presbyterian polity. Similarly, "episcopal" is used to describe a church governed by bishops. Self-governed local congregations, governed neither by elders nor bishops, are usually described as "congregational". More specifically, the capitalized appellation "Episcopal" is applied to several churches historically based within Anglicanism ("Episcopalianism"), including those still in communion with the Church of England. Using these definitions, examples of specific episcopal churches include: Some Lutheran churches practice congregational polity or a form of presbyterian polity. Others, including the Church of Sweden, practice episcopal polity; the Church of Sweden also counts its bishops among the historic episcopate as do some American Lutheran churches like the Anglo-Lutheran Catholic Church, Lutheran Orthodox Church, Lutheran Church - International, and the Lutheran Episcopal Communion. Many Methodist churches (see The United Methodist Church, among others) retain the form and function of episcopal polity, although in a modified form, called connexionalism. Since all trace their ordinations to an Anglican priest, John Wesley, it is generally considered that their bishops do not share in apostolic succession, though United Methodists still affirm that their bishops share in the historic episcopate. All orthodox Christians were in churches with an episcopal government, that is, one Church under local bishops and regional Patriarchs. Writing between ca. 85 and 110, St. Ignatius of Antioch, Patriarch of Antioch, was the earliest of the Church fathers to define the importance of episcopal government. Assuming Ignatius' view was the Apostolic teaching and practice, the line of succession was unbroken and passed through the four ancient Patriarchal sees (those local churches known to be founded by apostles), Rome, Jerusalem, Antioch and Alexandria. Rome was the leading Patriarchate of the ancient four by virtue of its founding by Saints Peter and Paul and their martyrdom there, not to mention being the political center of the Roman empire at the time. Some organizations (e.g. the Assyrian Church of the East), though aloof from the political wranglings of imperial Christianity, nevertheless also practiced episcopal polity. Shortly after the Roman Emperor Constantine I legalized Christianity in 321, he also constructed an elaborate second capital of the Roman Empire located at Byzantium and renamed it Constantinople, in 324. The single Roman Empire was divided between these two autonomous administrative centers, Roman and Constantinopolitan, West and East, Latin speaking and Greek speaking. This remained the status quo through the fourth century. A deep chasm developed between the East and West, becoming critical around 350, known as the Arian, or Nicene controversy. The Eastern Christian Churches were thought by Constantine to believe against the Trinity; that Christ was lesser than God. Hilary, Bishop of Poitiers, France, believed that the Eastern Church should be given the opportunity to, at least, be educated on the subject. Constantine, in his wisdom, and upset by disagreement, banished Hilary to the East. Hilary perfected his Greek language skills while in exile, and determined the great divide between Rome and the East was actually not a disagreement at all, and was merely a linguistic ignorance on the part of his Latin speaking contemporaries. This truth became known in the West, though some differences lingered. Hilary of Poitiers later became St. Hilary, Doctor of the Church, for exposing the true Christian beliefs of the Eastern Church. Many of Hilary's writings were lost to time. In the fifth century, Pope Dioscorus, the Patriarch of Alexandria, rejected certain Christological dogmas promulgated by the Council of Chalcedon, and as a result, the Oriental Orthodox churches split from the rest; however they continued the episcopal tradition, and today in fact there is dialog between the various orthodox churches over whether the schism was due to real differences or simply translation failures. Also during the fifth century, the Western Roman Empire declined and was overrun by German and Frankish peoples. Although the city of Rome was in ruins, distant from the seat of secular power, and constantly harassed by invaders, the Roman Patriarchate remained the center of the Western or Latin Church. Claiming the ancient primacy of Peter and the title of "Apostolic See", it remained the last court of episcopal appeal in serious matters for the whole Church, East and West. However, the center of the civilized Roman world had shifted definitively to Constantinople, or New Rome, the capital of the Greek speaking Empire. Along with this shift, the effective administration of the Church in the Eastern Roman Empire also shifted. This practical eminence of Constantinople in the East is evident, first at the First Council of Constantinople 381, and then ecumenically at the Council of Chalcedon in 451. Beginning with John the Faster (John IV, 582-595), the Bishop of Constantinople adopted as a formal title for himself the by then customary honorific Ecumenical Patriarch ("pre-eminent father for the civilized world") over the strong objections of Rome, a title based on the political prestige of Constantinople and its economic and cultural centrality in the Empire. In the following years, Rome's appeals to the East were based on the unique authority of the Apostolic See and the primacy of Peter, over the powers of councils as defended by the East (councils, for example, had endorsed that lofty title which Rome contested). The sometimes subtle differences between Eastern and Western conceptions of authority and its exercise produced a gradually widening rift between the Churches which continued with some occasional relief throughout the following centuries until the final rupture of the Great Schism (marked by two dates: 16 July 1054 and the Council of Florence in 1439). The Catholic Church has an episcopate, with the Pope, who is the Bishop of Rome, at the top. The Catholic Church considers that juridical oversight over the Church is not a power that derives from human beings, but strictly from the authority of Christ, which was given to his twelve apostles. The See of Rome, as the unbroken line of apostolic authority descending from St. Peter (the "prince and head of the apostles"), is a visible sign and instrument of communion among the college of bishops and therefore also of the local churches around the world. In communion with the worldwide college of bishops, the Pope has all legitimate juridical and teaching authority over the whole Church. This authority given by Christ to St. Peter and the apostles is transmitted from one generation to the next by the power of the Holy Spirit, through the laying on of hands from the Apostles to the bishops, in unbroken succession. The conciliar idea of episcopal government continues in the Eastern Orthodox Church. In Eastern Orthodoxy, all autocephalous primates are seen as collectively gathering around Christ, with other archbishops and bishops gathering around them, and so forth, in a model called "conciliar hierarchy". This is based in part on the vision in the book of Revelation of the 24 elders gathered around the throne of Christ, who are believed to represent the 12 patriarchs of Israel and the 12 apostles of Jesus Christ. There is no single Patriarch with exclusive authority comparable to the Pope in Rome. However, the Patriarch of Constantinople (now Istanbul) is seen as the "primus inter pares", the "first among equals" of the autocephalous churches of Eastern Orthodoxy. The Oriental Orthodox Churches affirm the ideas of apostolic succession and episcopal government. Within each national Church, the bishops form a holy synod to which even the Patriarch is subject. The Syriac Orthodox Church traces its apostolic succession to St. Peter and recognises Antioch as the original See of St. Peter. The Armenian Apostolic Church traces its lineage to the Apostle Bartholomew. The Indian Orthodox Church traces its lineage to the Apostle Thomas. The Ethiopian Orthodox Church received its lines of succession through the Coptic Orthodox Church in the fifth century. Both the Greek and Coptic Orthodox Churches each recognise their own Pope of Alexandria (Pope and Patriarch of Alexandria and All Africa, and Pope of the Coptic Orthodox Church of Alexandria respectively), both of whom trace their apostolic succession back to the Mark the Evangelist. There are official ongoing efforts in recent times to heal this ancient breach. Already, the two recognize each other's baptisms, chrismations, and marriages, making intermarriage much easier. Historically, the Church of the East has traced its episcopal succession to St. Thomas the Apostle. Currently the bishops of the Assyrian Church of the East continue to maintain its apostolic succession. Anglicanism is the most prominent of the Reformation traditions to lay claim to the historic episcopate through apostolic succession in terms comparable to the various Roman Catholic and Orthodox Communions. Anglicans assert unbroken episcopal succession in and through the Church of England back to St. Augustine of Canterbury and to the first century Roman province of Britannia. While some Celtic Christian practices were changed at the Synod of Whitby, the church in the British Isles was under papal authority from earliest times. The legislation of Henry VIII effectively establishing the independence from Rome of the Church of England, did not alter its constitutional or pastoral structures. Royal supremacy was exercised through the extant legal structures of the church, whose leaders were bishops. Episcopacy was thus seen as a given of the Reformed "Ecclesia Anglicana", and a foundation in the institution's appeal to ancient and apostolic legitimacy. What did change was that bishops were now seen to be ministers of the Crown for the spiritual government of its subjects. The influence of Richard Hooker was crucial to an evolution in this understanding in which bishops came to be seen in their more traditional role as ones who delegate to the presbyterate inherited powers, act as pastors to presbyters, and holding a particular teaching office with respect to the wider church. Anglican opinion has differed as to the way in which episcopal government is "de jure divino" (by the Divine Right of Kings). On the one hand, the seventeenth century divine, John Cosin, held that episcopal authority is "jure divino", but that it stemmed from "apostolic practice and the customs of the Church...[not] absolute precept that either Christ or His Apostles gave about it" (a view maintained also by Hooker). In contrast, Lancelot Andrewes and others held that episcopal government is derived from Christ via the apostles. Regardless, both parties viewed the episcopacy as bearing the apostolic function of oversight, which both includes, and derives from the power of ordination, and is normative for the governance of the church. The practice of apostolic succession both ensures the legitimacy of the church's mission and establishes the unity, communion, and continuity of the local church with the universal church. This formulation, in turn, laid the groundwork for an independent view of the church as a "sacred society" distinct from civil society, which was so crucial for the development of local churches as non-established entities outside England, and gave direct rise to the Catholic Revival and disestablishmentarianism within England. Functionally, Anglican episcopal authority is expressed synodically, although individual provinces may accord their primate with more or less authority to act independently. Called variously "synods," "councils," or "conventions," they meet under episcopal chairmanship. In many jurisdictions, conciliar resolutions that have been passed require episcopal assent or consent to take force. Seen in this way, Anglicans often speak of "the bishop-in-synod" as the force and authority of episcopal governance. Such conciliar authority extends to the standard areas of doctrine, discipline, and worship, but in these regards is limited by Anglicanism's tradition of the limits of authority. Those limits are expressed in Article XXI of the Thirty-Nine Articles of Religion, ratified in 1571 (significantly, just as the Council of Trent was drawing to a close), which held that "General Councils...may err, and sometimes have erred...wherefore things ordained by them as necessary to salvation have neither strength nor authority, unless it may be declared that they be taken out of holy Scripture." Hence, Anglican jurisdictions have traditionally been conservative in their approach to either innovative doctrinal development or in encompassing actions of the church as doctrinal (see lex orandi, lex credendi). Anglican synodical government, though varied in expression, is characteristically representative. Provinces of the Anglican Communion, their ecclesiastical provinces and dioceses are governed by councils consisting not only of bishops, but also representatives of the presbyterate and laity. There is no international juridical authority in Anglicanism, although the tradition's common experience of episcopacy, symbolised by the historical link with the See of Canterbury, along with a common and complex liturgical tradition, has provided a measure of unity. This has been reinforced by the Lambeth Conferences of Anglican Communion bishops, which first met in 1867. These conferences, though they propose and pass resolutions, are strictly consultative, and the intent of the resolutions are to provide guideposts for Anglican jurisdictions—not direction. The Conferences also express the function of the episcopate to demonstrate the ecumenical and Catholic nature of the church. The Scottish Episcopal Church traces its history back to the origins of Christianity in Scotland. Following the 1560 Scottish Reformation the Church of Scotland was initially run by Superintendents, episcopal governance was restored in 1572, but episcopalianism alternated with periods when the Kirk was under presbyterian control until the 1711 Act allowed formation of the independent non-established Scottish Episcopal Church. The Nonjuring schism led to the British Government imposing penal laws against the church. In 1784 the Scottish church appointed Samuel Seabury as first bishop of the American Episcopal Church, beginning the worldwide Anglican Communion of churches, and in 1792 the penal laws were abolished. The church accepted the articles of the Church of England in 1804. The spread of increasingly democratic forms of representative governance has its origin in the formation of the first General Conventions of the American Episcopal Church in the 1780s, which established a "House of Bishops" and a "House of Deputies". In many jurisdictions, there is also a third, clerical House. Resolutions may be voted on jointly or by each House, in the latter case requiring passage in all Houses to be adopted by the particular council. Churches that are members of the Anglican Communion are episcopal churches in polity, and some are named "Episcopal". However, some churches that self-identify as Anglican do not belong to the Anglican Communion, and not all episcopally-governed churches are Anglican. The Roman Catholic Church, the Old Catholic Churches (in full communion with, but not members of, the Anglican Communion), and the Eastern Orthodox churches are recognized, and also their bishops, by Anglicans. As an offshoot of Anglicanism, Methodist churches often use episcopal polity for historical as well as practical reasons, albeit to limited use. Methodists often use the term "connexionalism" or "connexional polity" in addition to "episcopal". Nevertheless, the powers of the Methodist episcopacy can be relatively strong and wide-reaching compared to traditional conceptions of episcopal polity. For example, in the United Methodist Church, bishops are elected for life, can serve up to two terms in a specific conference (three if special permission is given), are responsible for ordaining and appointing clergy to pastor churches, perform many administrative duties, preside at the annual sessions of the regional Conferences and at the quadrennial meeting of the worldwide General Conference, have authority for teaching and leading the church on matters of social and doctrinal import, and serve to represent the denomination in ecumenical gatherings. United Methodist bishops in the United States serve in their appointed conferences, being moved to a new "Episcopal Area" after 8 (or 12) years, until their mandated retirement at the end of the quadrenium following their sixty-sixth birthday. British Methodism holds that all ordained ministers are equal in terms of spirituality. However, for practical management lines are drawn into President of Conference, Chair of District, Superintendent Minister, Minister. However, all are ministers. The Reformed Church of Hungary, and the Lutheran churches in continental Europe may sometimes be called "episcopal". In these latter cases, the form of government is not radically different from the presbyterian form, except that their councils of bishops have hierarchical jurisdiction over the local ruling bodies to a greater extent than in most Presbyterian and other Reformed churches. As mentioned, the Lutheran Church in Sweden and Finland are exceptions, claiming apostolic succession in a pattern somewhat like the Anglican churches. Otherwise, forms of polity are not mandated in the Lutheran churches, as it is not regarded as having doctrinal significance. Old World Lutheranism, for historical reasons, has tended to adopt Erastian theories of episcopal authority (by which church authority is to a limited extent sanctioned by secular government). In the United States, the Lutheran churches tend to adopt a form of government more comparable to congregationalism. A small minority of Episcopal Baptists exists. Most Anabaptist churches of the plain dress tradition follow an episcopal system, at least in name. Congregational governance is strongly emphasized, and each congregation elects its pastor. Bishops enforce inter-congregational unity and may discipline pastors for breaking from traditional norms. Although it never uses the term, The Church of Jesus Christ of Latter-day Saints (LDS Church) is episcopal, rather than presbyterian or congregational, in the sense that it has a strict hierarchy of leadership from the local bishop/branch president up to a single prophet/president, believed to be personally authorized and guided by Jesus Christ. Local congregations (branches, wards, and stakes) have "de jure" boundaries by which members are allocated, and membership records are centralized. This system developed gradually from a more presbyterian polity (Joseph Smith's original title in 1830 was "First Elder") for pragmatic and doctrinal reasons, reaching a full episcopacy during the Nauvoo period (1839–1846).
https://en.wikipedia.org/wiki?curid=10067
East Slavic languages The East Slavic languages constitute one of the three regional subgroups of Slavic languages, currently spoken throughout Eastern Europe, Northern Asia, and the Caucasus. It is the group with the largest numbers of speakers, far out-numbering the Western and Southern Slavic groups. The existing East Slavic languages are Belarusian, Russian and Ukrainian; Rusyn is considered to be either a separate language or a dialect of Ukrainian. The East Slavic languages descend from a common predecessor, the language of the medieval Kievan Rus' (9th to 13th centuries). All these languages use the Cyrillic script, but with particular modifications. The East Slavic territory shows a definite linguistic continuum with many transitional dialects. Between Belarusian and Ukrainian there is the Polesian dialect, which shares features from both languages. East Polesian is a transitional variety between Belarusian and Ukrainian on the one hand, and between South Russian and Ukrainian on the other hand. At the same time, Belarusian and Southern Russian form a continuous area, making it virtually impossible to draw a line between the two languages. Central or Middle Russian (with its Moscow sub-dialect), the transitional step between the North and the South, became a base for the Russian literary standard. Northern Russian with its predecessor, the Old Novgorod dialect, has many original and archaic features. Due to being under the influence of the Polish–Lithuanian Commonwealth for many centuries, Belarusian and Ukrainian have adopted several influences from Polish, a West Slavic language as a result. Ruthenian, the mixed Belarusian-Ukrainian literary language with Church Slavonic substratum and Polish adstratum, was together, with Middle Polish an official language in Belarus and Ukraine until the end of the 18th century. When the common Old East Slavic language became separated from the ancient Slavic tongue common to all Slavs is difficult to ascertain, though in the 12th century the common language of Rus' is still referred to in contemporary writing as Slavic. Therefore, a crucial differentiation has to be made between the history of the East Slavic "dialects" and that of the "literary languages" employed by the Eastern Slavs. Although most ancient texts betray the dialect their author or scribe spoke, it is also clearly visible that they tried to write in a language different from their dialects and to avoid those mistakes that enable us nowadays to locate them. In both cases one has to keep in mind that the history of the East Slavic languages is of course a history of written texts. We do not know how the writers of the preserved texts would have spoken in everyday life. After the conversion of the East Slavic region to Christianity the people used service books borrowed from Bulgaria, which were written in Old Church Slavonic. The Church Slavonic language was strictly used only in text, while the colloquial language of the Bulgarians was communicated in its spoken form. Throughout the Middle Ages (and in some way up to the present day) there existed a duality between the Church Slavonic language used as some kind of 'higher' register (not only) in religious texts and the popular tongue used as a 'lower' register for secular texts. It has been suggested to describe this situation as diglossia, although there do exist mixed texts where it is sometimes very hard to determine why a given author used a popular or a Church Slavonic form in a given context. Church Slavonic was a major factor in the evolution of modern Russian, where there still exists a "high stratum" of words that were imported from this language. All of these languages are today separate in their own right. In the Russian Empire the official view was that the Belarusian ("White Russian"), Ukrainian ("Little Russian"), and Russian ("Great Russian") languages were dialects of one common "Russian" language (the common languages of Eastern Slavic countries). Over the course of the 20th century, "Great Russian" came to be known as Russian proper, "Little Russian" as Ukrainian and "White Russian" as Belarusian.
https://en.wikipedia.org/wiki?curid=10070
Elizabeth Gracen Elizabeth Ward Gracen (born Elizabeth Grace Ward, April 3, 1961) is an American actress and beauty pageant contestant who won the title of Miss America in 1982. Elizabeth Grace Ward was born on April 3, 1961, in Ozark, Arkansas, the daughter of Patricia Hampe, a nurse, and Jimmy Young Ward, a poultry worker. She was raised in Booneville, Arkansas. The family later moved to Russellville, Arkansas, where Ward dated University of Arkansas trainer Mike Walker and graduated from Russellville High School in 1979. She was a junior accounting major at Arkansas Tech University at the time she entered the Miss America contest. Instead of returning to Arkansas Tech, she used her Miss America scholarship money to study acting at HB Studios in New York City. Gracen won the titles of Miss Arkansas in 1981 and Miss America in 1982. After her yearlong work as Miss America, she enrolled in acting classes then relocated to California to pursue a film and television career. Gracen posed nude for "Playboy" magazine's May 1992 issue. Gracen made her professional feature film debut in "Three For The Road" with Charlie Sheen. Her film credits also include a featured role in "Marked for Death", opposite Steven Seagal, "Pass The Ammo" with Tim Curry, and the CBS feature "83 Hours Till Dawn" with Peter Strauss and Robert Urich. Gracen starred in "Lower Level" and "Discretion Assured" with Michael York. On television, Gracen has appeared in Shelley Duvall's "Strange Case of Dr. Jekyll and Mr. Hyde", Sidney Sheldon's "The Sands of Time" and "Death of the Incredible Hulk". She also appeared with a starring role in the series "Extreme" for NBC and the syndicated series "Renegade" and "Queen of Swords". Gracen's best-known acting role has been as the recurring character, the Immortal Amanda, in the series "" and its spin-off series "". In December 1999, Gracen filed for bankruptcy protection. Afterward, Gracen was given a few television guest roles, and a supporting role in the made-for-television movie "Interceptor Force 2", before taking a long leave of absence from her acting career in 2002. Gracen began doing voiceover work for Blue Hours Productions, which has revived the classic radio anthology "Suspense", which airs on Sirius XM. In 2012, Gracen did a character voice-over in the Malaysian animated science fiction film "". In 2012, Gracen formed Flapper Films. In 2014, she starred in "Coherence", a sci-fi indie thriller. In January 2016, Gracen established Flapper Press and self-published "Shalilly", a young adult fantasy novel. Gracen made her directorial debut with a documentary short, "The Damn Deal" about three young drag queens from Arkansas who compete in female impersonator beauty pageants. Gracen married Jon Birmingham in 1982, but they divorced in 1984. In 1989, while filming "", she met actor Brendan Hughes, and they married soon after. They divorced in 1994. Gracen married Adam Murphy, and they have a daughter. According to Gracen, some time in 1983, she had a one-night stand with future President Bill Clinton when he was Governor of Arkansas. She was married at the time as was he. In 1992, rumors swirled that Gracen had conducted an affair with Bill Clinton. At first, Gracen dismissed this claim (as requested by Clinton's campaign manager Mickey Kantor); however, in spring 1998 Gracen recanted her six-year-old denial and stated she had a one-night stand with Clinton in 1982. After claiming this, Independent Counsel Kenneth Starr, who was investigating Clinton in the Paula Jones lawsuit, issued a subpoena to have her testify her claim in court. However, Gracen eluded the subpoena and was at one point able to avoid it because "" was being filmed out of the country. Paula Jones' legal team was also unable to track down Gracen because she had made unscheduled trips to Las Vegas and the Caribbean.
https://en.wikipedia.org/wiki?curid=10071
Epicurus Epicurus (; 341–270 BC) was an ancient Greek philosopher and sage who founded Epicureanism, a highly influential school of philosophy. He was born on the Greek island of Samos to Athenian parents. Influenced by Democritus, Aristippus, Pyrrho, and possibly the Cynics, he turned against the Platonism of his day and established his own school, known as "the Garden", in Athens. Epicurus and his followers were known for eating simple meals and discussing a wide range of philosophical subjects. He openly allowed women to join the school as a matter of policy. Epicurus is said to have originally written over 300 works on various subjects, but the vast majority of these writings have been lost. Only three letters written by him—the letters to "Menoeceus", "Pythocles", and "Herodotus"—and two collections of quotes—the "Principal Doctrines" and the "Vatican Sayings"—have survived intact, along with a few fragments of his other writings. Most knowledge of his teachings comes from later authors, particularly the biographer Diogenes Laërtius, the Epicurean Roman poet Lucretius and the Epicurean philosopher Philodemus, and with hostile but largely accurate accounts by the Pyrrhonist philosopher Sextus Empiricus, and the Academic Skeptic and statesman Cicero. For Epicurus, the purpose of philosophy was to help people attain a happy, tranquil life characterized by "ataraxia" (peace and freedom from fear) and "aponia" (the absence of pain). He advocated that people were best able to pursue philosophy by living a self-sufficient life surrounded by friends. He taught that the root of all human neurosis is death denial and the tendency for human beings to assume that death will be horrific and painful, which he claimed causes unnecessary anxiety, selfish self-protective behaviors, and hypocrisy. According to Epicurus, death is the end of both the body and the soul and therefore should not be feared. Epicurus taught that although the gods exist, they have no involvement in human affairs. He taught that people should behave ethically not because the gods punish or reward people for their actions, but because amoral behavior will burden them with guilt and prevent them from attaining "ataraxia". Like Aristotle, Epicurus was an empiricist, meaning he believed that the senses are the only reliable source of knowledge about the world. He derived much of his physics and cosmology from the earlier philosopher Democritus ( 460– 370 BC). Like Democritus, Epicurus taught that the universe is infinite and eternal and that all matter is made up of extremely tiny, invisible particles known as "atoms". All occurrences in the natural world are ultimately the result of atoms moving and interacting in empty space. Epicurus deviated from Democritus by proposing the idea of atomic "swerve", which holds that atoms may deviate from their expected course, thus permitting humans to possess free will in an otherwise deterministic universe. Though popular, Epicurean teachings were controversial from the beginning. Epicureanism reached the height of its popularity during the late years of the Roman Republic. It died out in late antiquity, subject to hostility from early Christianity. Throughout the Middle Ages Epicurus was popularly, though inaccurately, remembered as a patron of drunkards, whoremongers, and gluttons. His teachings gradually became more widely known in the fifteenth century with the rediscovery of important texts, but his ideas did not become acceptable until the seventeenth century, when the French Catholic priest Pierre Gassendi revived a modified version of them, which was promoted by other writers, including Walter Charleton and Robert Boyle. His influence grew considerably during and after the Enlightenment, profoundly impacting the ideas of major thinkers, including John Locke, Thomas Jefferson, Jeremy Bentham, and Karl Marx. Epicurus was born in the Athenian settlement on the Aegean island of Samos in February 341 BC. His parents, Neocles and Chaerestrate, were both Athenian-born, and his father was an Athenian citizen. Epicurus grew up during the final years of the Greek Classical Period. Plato had died seven years before Epicurus was born and Epicurus was seven years old when Alexander the Great crossed the Hellespont into Persia. As a child, Epicurus would have received a typical ancient Greek education. As such, according to Norman Wentworth DeWitt, "it is inconceivable that he would have escaped the Platonic training in geometry, dialectic, and rhetoric." Epicurus is known to have studied under the instruction of a Samian Platonist named Pamphilus, probably for about four years. His "Letter of Menoeceus" and surviving fragments of his other writings strongly suggest that he had extensive training in rhetoric. After the death of Alexander the Great, Perdiccas expelled the Athenian settlers on Samos to Colophon, on the coast of what is now Turkey. After the completion of his military service, Epicurus joined his family there. He studied under Nausiphanes, who followed the teachings of Democritus, and later those of Pyrrho, whose way of life Epicurus greatly admired. Epicurus's teachings were heavily influenced by those of earlier philosophers, particularly Democritus. Nonetheless, Epicurus differed from his predecessors on several key points of determinism and vehemently denied having been influenced by any previous philosophers, whom he denounced as "confused". Instead, he insisted that he had been "self-taught". According to DeWitt, Epicurus's teachings also show influences from the contemporary philosophical school of Cynicism. The Cynic philosopher Diogenes of Sinope was still alive when Epicurus would have been in Athens for his required military training and it is possible they may have met. Diogenes's pupil Crates of Thebes ( 365 – 285 BC) was a close contemporary of Epicurus. Epicurus agreed with the Cynics' quest for honesty, but rejected their "insolence and vulgarity", instead teaching that honesty must be coupled with courtesy and kindness. Epicurus shared this view with his contemporary, the comic playwright Menander. Epicurus's "Letter to Menoeceus", possibly an early work of his, is written in an eloquent style similar to that of the Athenian rhetorician Isocrates (436–338 BC), but, for his later works, he seems to have adopted the bald, intellectual style of the mathematician Euclid. Epicurus's epistemology also bears an unacknowledged debt to the later writings of Aristotle (384–322 BC), who rejected the Platonic idea of hypostatic Reason and instead relied on nature and empirical evidence for knowledge about the universe. During Epicurus's formative years, Greek knowledge about the rest of the world was rapidly expanding due to the Hellenization of the Near East and the rise of Hellenistic kingdoms. Epicurus's philosophy was consequently more universal in its outlook than those of his predecessors, since it took cognizance of non-Greek peoples as well as Greeks. He may have had access to the now-lost writings of the historian and ethnographer Megasthenes, who wrote during the reign of Seleucus I Nicator (ruled 305–281 BC). During Epicurus's lifetime, Platonism was the dominant philosophy in higher education. Epicurus's opposition to Platonism formed a large part of his thought. Over half of the forty Principal Doctrines of Epicureanism are flat contradictions of Platonism. In around 311 BC, Epicurus, when he was around thirty years old, began teaching in Mytilene. Around this time, Zeno of Citium, the founder of Stoicism, arrived in Athens, at the age of about twenty-one, but Zeno did not begin teaching what would become Stoicism for another twenty years. Although later texts, such as the writings of the first-century BC Roman orator Cicero, portray Epicureanism and Stoicism as rivals, this rivalry seems to have only emerged after Epicurus's death. Epicurus's teachings caused strife in Mytilene and he was forced to leave. He then founded a school in Lampsacus before returning to Athens in 306 BC, where he remained until his death. There he founded The Garden (κῆπος), a school named for the garden he owned that served as the school's meeting place, about halfway between the locations of two other schools of philosophy, the Stoa and the Academy. The Garden was more than just a school; it was "a community of like-minded and aspiring practitioners of a particular way of life." The primary members were Hermarchus, the financier Idomeneus, Leonteus and his wife Themista, the satirist Colotes, the mathematician Polyaenus of Lampsacus, and Metrodorus of Lampsacus, the most famous popularizer of Epicureanism. His school was the first of the ancient Greek philosophical schools to admit women as a rule rather than an exception, and the biography of Epicurus by Diogenes Laërtius lists female students such as Leontion and Nikidion. An inscription on the gate to The Garden is recorded by Seneca the Younger in epistle XXI of "Epistulae morales ad Lucilium": "Stranger, here you will do well to tarry; here our highest good is pleasure." According to Diskin Clay, Epicurus himself established a custom of celebrating his birthday annually with common meals, befitting his stature as "heros ktistes" ("founding hero") of the Garden. He ordained in his will annual memorial feasts for himself on the same date (10th of Gamelion month). Epicurean communities continued this tradition, referring to Epicurus as their "saviour" (soter) and celebrating him as hero. The hero cult of Epicurus may have operated as a Garden variety civic religion. However, clear evidence of an Epicurean hero cult, as well as the cult itself, seems buried by the weight of posthumous philosophical interpretation. Epicurus never married and had no known children. He was most likely a vegetarian. Diogenes Laërtius records that, according to Epicurus's successor Hermarchus, Epicurus died a slow and painful death in 270 BC at the age of seventy-two from a stone blockage of his urinary tract. Despite being in immense pain, Epicurus is said to have remained cheerful and to have continued to teach until the very end. Possible insights into Epicurus's death may be offered by the extremely brief "Epistle to Idomeneus", included by Diogenes Laërtius in Book X of his "Lives and Opinions of Eminent Philosophers". The authenticity of this letter is uncertain and it may be a later pro-Epicurean forgery intended to paint an admirable portrait of the philosopher to counter the large number of forged epistles in Epicurus's name portraying him unfavorably. I have written this letter to you on a happy day to me, which is also the last day of my life. For I have been attacked by a painful inability to urinate, and also dysentery, so violent that nothing can be added to the violence of my sufferings. But the cheerfulness of my mind, which comes from the recollection of all my philosophical contemplation, counterbalances all these afflictions. And I beg you to take care of the children of Metrodorus, in a manner worthy of the devotion shown by the young man to me, and to philosophy. If authentic, this letter would support the tradition that Epicurus was able to remain joyful to the end, even in the midst of his suffering. It would also indicate that he maintained an especial concern for the wellbeing of children. Epicurus and his followers had a well-developed epistemology, which developed as a result of their rivalry with other philosophical schools. Epicurus wrote a treatise entitled , or "Rule", in which he explained his methods of investigation and theory of knowledge. This book, however, has not survived, nor does any other text that fully and clearly explains Epicurean epistemology, leaving only mentions of this epistemology by several authors to reconstruct it. Epicurus was an ardent Empiricist; believing that the senses are the only reliable sources of information about the world. He rejected the Platonic idea of "Reason" as a reliable source of knowledge about the world apart from the senses and was bitterly opposed to the Pyrrhonists and Academic Skeptics, who not only questioned the ability of the senses to provide accurate knowledge about the world, but also whether it is even possible to know anything about the world at all. Epicurus maintained that the senses never deceive humans, but that the senses can be misinterpreted. Epicurus held that the purpose of all knowledge is to aid humans in attaining "ataraxia". He taught that knowledge is learned through experiences rather than innate and that the acceptance of the fundamental truth of the things a person perceives is essential to a person's moral and spiritual health. In the "Letter to Pythocles", he states, "If a person fights the clear evidence of his senses he will never be able to share in genuine tranquility." Epicurus regarded gut feelings as the ultimate authority on matters of morality and held that whether a person feels an action is right or wrong is a far more cogent guide to whether that act really is right or wrong than abstracts maxims, strict codified rules of ethics, or even reason itself. Epicurus permitted that any and every statement that is not directly contrary to human perception has the possibility to be true. Nonetheless, anything contrary to a person's experience can be ruled out as false. Epicureans often used analogies to everyday experience to support their argument of so-called "imperceptibles", which included anything that a human being cannot perceive, such as the motion of atoms. In line with this principle of non-contradiction, the Epicureans believed that events in the natural world may have multiple causes that are all equally possible and probable. Lucretius writes in "On the Nature of Things", as translated by William Ellery Leonard: There be, besides, some thing Of which 'tis not enough one only cause To state—but rather several, whereof one Will be the true: lo, if thou shouldst espy Lying afar some fellow's lifeless corse, 'Twere meet to name all causes of a death, That cause of his death might thereby be named: For prove thou mayst he perished not by steel, By cold, nor even by poison nor disease, Yet somewhat of this sort hath come to him We know—And thus we have to say the same In divers cases. Epicurus strongly favored naturalistic explanations over theological ones. In his "Letter to Pythocles", he offers four different possible natural explanations for thunder, six different possible natural explanations for lightning, three for snow, three for comets, two for rainbows, two for earthquakes, and so on. Although all of these explanations are now known to be false, they were an important step in the history of science, because Epicurus was trying to explain natural phenomena using natural explanations, rather than resorting to inventing elaborate stories about gods and mythic heroes. Epicurus was a hedonist, meaning he taught that what is pleasurable is morally good and what is painful is morally evil. He idiosyncratically defined "pleasure" as the absence of suffering and taught that all humans should seek to attain the state of "ataraxia", meaning "untroubledness", a state in which the person is completely free from all pain or suffering. He argued that most of the suffering which human beings experience is caused by the irrational fears of death, divine retribution, and punishment in the afterlife. In his "Letter to Menoeceus", Epicurus explains that people seek wealth and power on account of these fears, believing that having more money, prestige, or political clout will save them from death. He, however, maintains that death is the end of existence, that the terrifying stories of punishment in the afterlife are ridiculous superstitions, and that death is therefore nothing to be feared. He writes in his "Letter to Menoeceus": "Accustom thyself to believe that death is nothing to us, for good and evil imply sentience, and death is the privation of all sentience;... Death, therefore, the most awful of evils, is nothing to us, seeing that, when we are, death is not come, and, when death is come, we are not." From this doctrine arose the Epicurean epitaph: "Non fui, fui, non-sum, non-curo" ("I was not; I was; I am not; I do not care"), which is inscribed on the gravestones of his followers and seen on many ancient gravestones of the Roman Empire. This quotation is often used today at humanist funerals. The Tetrapharmakos presents a summary of the key points of Epicurean ethics: Don't fear god Don't worry about death What is good is easy to get What is terrible is easy to endure Although Epicurus has been commonly misunderstood as an advocate of the rampant pursuit of pleasure, he, in fact, maintained that a person can only be happy and free from suffering by living wisely, soberly, and morally. He strongly disapproved of raw, excessive sensuality and warned that a person must take into account whether the consequences of his actions will result in suffering, writing, "the pleasant life is produced not by a string of drinking bouts and revelries, nor by the enjoyment of boys and women, nor by fish and the other items on an expensive menu, but by sober reasoning." He also wrote that a single good piece of cheese could be equally pleasing as an entire feast. Furthermore, Epicurus taught that "it is not possible to live pleasurably without living sensibly and nobly and justly", because a person who engages in acts of dishonesty or injustice will be "loaded with troubles" on account of his own guilty conscience and will live in constant fear that his wrongdoings will be discovered by others. A person who is kind and just to others, however, will have no fear and will be more likely to attain "ataraxia". Epicurus distinguished between two different types of pleasure: "moving" pleasures (κατὰ κίνησιν ἡδοναί) and "static" pleasures (καταστηματικαὶ ἡδοναί). "Moving" pleasures occur when one is in the process of satisfying a desire and involve an active titillation of the senses. After one's desires have been satisfied (e.g. when one is full after eating), the pleasure quickly goes away and the suffering of wanting to fulfill the desire again returns. For Epicurus, static pleasures are the best pleasures because moving pleasures are always bound up with pain. Epicurus had a low opinion of sex and marriage, regarding both as having dubious value. Instead, he maintained that platonic friendships are essential to living a happy life. One of the "Principle Doctrines" states, "Of the things wisdom acquires for the blessedness of life as a whole, far the greatest is the possession of friendship." He also taught that philosophy is itself a pleasure to engage in. One of the quotes from Epicurus recorded in the "Vatican Sayings" declares, "In other pursuits, the hard-won fruit comes at the end. But in philosophy, delight keeps pace with knowledge. It is not after the lesson that enjoyment comes: learning and enjoyment happen at the same time." Epicurus distinguishes between three types of desires: natural and necessary, natural but unnecessary, and vain and empty. Natural and necessary desires include the desires for food and shelter. These are easy to satisfy, difficult to eliminate, bring pleasure when satisfied, and are naturally limited. Going beyond these limits produces unnecessary desires, such as the desire for luxury foods. Although food is necessary, luxury food is not necessary. Correspondingly, Epicurus advocates a life of hedonistic moderation by reducing desire, thus eliminating the unhappiness caused by unfulfilled desires. Vain desires include desires for power, wealth, and fame. These are difficult to satisfy because no matter how much one gets, one can always want more. These desires are inculcated by society and by false beliefs about what we need. They are not natural and are to be shunned. Epicurus' teachings were introduced into medical philosophy and practice by the Epicurean doctor Asclepiades of Bithynia, who was the first physician who introduced Greek medicine in Rome. Asclepiades introduced the friendly, sympathetic, pleasing and painless treatment of patients. He advocated humane treatment of mental disorders, had insane persons freed from confinement and treated them with natural therapy, such as diet and massages. His teachings are surprisingly modern; therefore Asclepiades is considered to be a pioneer physician in psychotherapy, physical therapy and molecular medicine. Epicurus writes in his "Letter to Herodotus" that "nothing ever arises from the nonexistent", indicating that all events therefore have causes, regardless of whether those causes are known or unknown. Similarly, he also writes that nothing ever passes away into nothingness, because, "if an object that passes from our view were completely annihilated, everything in the world would have perished, since that into which things were dissipated would be nonexistent." He therefore states: "The totality of things was always just as it is at present and will always remain the same because there is nothing into which it can change, inasmuch as there is nothing outside the totality that could intrude and effect change." Like Democritus before him, Epicurus taught that all matter is entirely made of extremely tiny particles called "atoms" (; "", meaning "indivisible"). For Epicurus and his followers, the existence of atoms was a matter of empirical observation; Epicurus's devoted follower, the Roman poet Lucretius, cites the gradual wearing down of rings from being worn, statues from being kissed, stones from being dripped on by water, and roads from being walked on in "On the Nature of Things" as evidence for the existence of atoms as tiny, imperceptible particles. Also like Democritus, Epicurus was a materialist who taught that the only things that exist are atoms and void. Void occurs in any place where there are no atoms. Epicurus and his followers believed that atoms and void are both infinite and that the universe is therefore boundless. In "On the Nature of Things", Lucretius argues this point using the example of a man throwing a javelin at the theoretical boundary of a finite universe. He states that the javelin must either go past the edge of the universe, in which case it is not really a boundary, or it must be blocked by something and prevented from continuing its path, but, if that happens, then the object blocking it must be outside the confines of the universe. As a result of this belief that the universe and the number of atoms in it are infinite, Epicurus and the Epicureans believed that there must also be infinitely many worlds within the universe. Epicurus taught that the motion of atoms is constant, eternal, and without beginning or end. He held that there are two kinds of motion: the motion of atoms and the motion of visible objects. Both kinds of motion are real and not illusory. Democritus had described atoms as not only eternally moving, but also eternally flying through space, colliding, coalescing, and separating from each other as necessary. In a rare departure from Democritus's physics, Epicurus posited the idea of atomic "swerve" ( ""; ), one of his best-known original ideas. According to this idea, atoms, as they are travelling through space, may deviate slightly from the course they would ordinarily be expected to follow. Epicurus's reason for introducing this doctrine was because he wanted to preserve the concepts of free will and ethical responsibility while still maintaining the deterministic physical model of atomism. Lucretius describes it, saying, "It is this slight deviation of primal bodies, at indeterminate times and places, which keeps the mind as such from experiencing an inner compulsion in doing everything it does and from being forced to endure and suffer like a captive in chains." Epicurus was first to assert human freedom as a result of the fundamental indeterminism in the motion of atoms. This has led some philosophers to think that, for Epicurus, free will was "caused directly by chance". In his "On the Nature of Things", Lucretius appears to suggest this in the best-known passage on Epicurus' position. In his "Letter to Menoeceus", however, Epicurus follows Aristotle and clearly identifies "three" possible causes: "some things happen of necessity, others by chance, others through our own agency." Aristotle said some things "depend on us" ("eph'hemin"). Epicurus agreed, and said it is to these last things that praise and blame naturally attach. For Epicurus, the "swerve" of the atoms simply defeated determinism to leave room for autonomous agency. In his "Letter to Menoeceus", a summary of his own moral and theological teachings, the first piece of advice Epicurus himself gives to his student is: "First, believe that a god is an indestructible and blessed animal, in accordance with the general conception of god commonly held, and do not ascribe to god anything foreign to his indestructibility or repugnant to his blessedness." Epicurus maintained that he and his followers knew that the gods exist because "our knowledge of them is a matter of clear and distinct perception", meaning that people can empirically sense their presences. He did not mean that people can see the gods as physical objects, but rather that they can see visions of the gods sent from the remote regions of interstellar space in which they actually reside. According to George K. Strodach, Epicurus could have easily dispensed of the gods entirely without greatly altering his materialist worldview, but the gods still play one important function in Epicurus's theology as the paragons of moral virtue to be emulated and admired. Epicurus rejected the conventional Greek view of the gods as anthropomorphic beings who walked the earth like ordinary people, fathered illegitimate offspring with mortals, and pursued personal feuds. Instead, he taught that the gods are morally perfect, but detached and immobile beings who live in the remote regions of interstellar space. In line with these teachings, Epicurus adamantly rejected the idea that deities were involved in human affairs in any way. Epicurus maintained that the gods are so utterly perfect and removed from the world that they are incapable of listening to prayers or supplications or doing virtually anything aside from contemplating their own perfections. In his "Letter to Herodotus", he specifically denies that the gods have any control over natural phenomena, arguing that this would contradict their fundamental nature, which is perfect, because any kind of worldly involvement would tarnish their perfection. He further warned that believing that the gods control natural phenomena would only mislead people into believing the superstitious view that the gods punish humans for wrongdoing, which only instills fear and prevents people from attaining "ataraxia". Epicurus himself criticizes popular religion in both his "Letter to Menoeceus" and his "Letter to Herodotus", but in a restrained and moderate tone. Later Epicureans mainly followed the same ideas as Epicurus, believing in the existence of the gods, but emphatically rejecting the idea of divine providence. Their criticisms of popular religion, however, are often less gentle than those of Epicurus himself. The "Letter to Pythocles", written by a later Epicurean, is dismissive and contemptuous towards popular religion and Epicurus's devoted follower, the Roman poet Lucretius ( 99 BC – 55 BC), passionately assailed popular religion in his philosophical poem "On the Nature of Things". In this poem, Lucretius declares that popular religious practices not only do not instill virtue, but rather result in "misdeeds both wicked and ungodly", citing the mythical sacrifice of Iphigenia as an example. Lucretius argues that divine creation and providence are illogical, not because the gods do not exist, but rather because these notions are incompatible with the Epicurean principles of the gods' indestructibility and blessedness. The later Pyrrhonist philosopher Sextus Empiricus ( 160 – 210 AD) rejected the teachings of the Epicureans specifically because he regarded them as theological "Dogmaticists". The Epicurean paradox or riddle of Epicurus or Epicurus' trilemma is a version of the problem of evil. Lactantius attributes this trilemma to Epicurus in "De Ira Dei", 13, 20-21: God, he says, either wishes to take away evils, and is unable; or He is able, and is unwilling; or He is neither willing nor able, or He is both willing and able. If He is willing and is unable, He is feeble, which is not in accordance with the character of God; if He is able and unwilling, He is envious, which is equally at variance with God; if He is neither willing nor able, He is both envious and feeble, and therefore not God; if He is both willing and able, which alone is suitable to God, from what source then are evils? Or why does He not remove them? In "Dialogues concerning Natural Religion" (1779), David Hume also attributes the argument to Epicurus: Epicurus’s old questions are yet unanswered. Is he willing to prevent evil, but not able? then is he impotent. Is he able, but not willing? then is he malevolent. Is he both able and willing? whence then is evil? No extant writings of Epicurus contain this argument. However, the vast majority of Epicurus's writings have been lost and it is possible that some form of this argument may have been found in his lost treatise "On the Gods", which Diogenes Laërtius describes as one of his greatest works. If Epicurus really did make some form of this argument, it would not have been an argument against the existence of deities, but rather an argument against divine providence. Epicurus's extant writings demonstrate that he did believe in the existence of deities. Furthermore, religion was such an integral part of daily life in Greece during the early Hellenistic Period that it is doubtful anyone during that period could have been an atheist in the modern sense of the word. Instead, the Greek word ("átheos"), meaning "without a god", was used as a term of abuse, not as an attempt to describe a person's beliefs. Epicurus promoted an innovative theory of justice as a social contract. Justice, Epicurus said, is an agreement neither to harm nor be harmed, and we need to have such a contract in order to enjoy fully the benefits of living together in a well-ordered society. Laws and punishments are needed to keep misguided fools in line who would otherwise break the contract. But the wise person sees the usefulness of justice, and because of his limited desires, he has no need to engage in the conduct prohibited by the laws in any case. Laws that are useful for promoting happiness are just, but those that are not useful are not just. ("Principal Doctrines" 31–40) Epicurus discouraged participation in politics, as doing so leads to perturbation and status seeking. He instead advocated not drawing attention to oneself. This principle is epitomised by the phrase "lathe biōsas" (), meaning "live in obscurity", "get through life without drawing attention to yourself", i.e., live without pursuing glory or wealth or power, but anonymously, enjoying little things like food, the company of friends, etc. Plutarch elaborated on this theme in his essay "Is the Saying "Live in Obscurity" Right?" (, "An recte dictum sit latenter esse vivendum") 1128c; cf. Flavius Philostratus, "Vita Apollonii" 8.28.12. Epicurus was an extremely prolific writer. According to Diogenes Laërtius, he wrote around 300 treatises on a variety of subjects. More original writings of Epicurus have survived to the present day than of any other Hellenistic Greek philosopher. Nonetheless, the vast majority of everything he wrote has now been lost and most of what is known about Epicurus's teachings come from the writings of his later followers, particularly the Roman poet Lucretius. The only surviving complete works by Epicurus are three relatively lengthy letters, which are quoted in their entirety in Book X of Diogenes Laërtius's "Lives and Opinions of Eminent Philosophers", and two groups of quotes: the "Principal Doctrines" (Κύριαι Δόξαι), which are likewise preserved through quotation by Diogenes Laërtius, and the "Vatican Sayings", preserved in a manuscript from the Vatican Library that was first discovered in 1888. In the "Letter to Herodotus" and the "Letter to Pythocles", Epicurus summarizes his philosophy on nature and, in the "Letter to Menoeceus", he summarizes his moral teachings. Numerous fragments of Epicurus's lost thirty-seven volume treatise "On Nature" have been found among the charred papyrus fragments at the Villa of the Papyri at Herculaneum. Scholars first began attempting to unravel and decipher these scrolls in 1800, but the efforts are painstaking and are still ongoing. According to Diogenes Laertius (10.27-9), the major works of Epicurus include: Epicureanism was extremely popular from the very beginning. Diogenes Laërtius records that the number of Epicureans throughout the world exceeded the populations of entire cities. Nonetheless, Epicurus was not universally admired and, within his own lifetime, he was vilified as an ignorant buffoon and egoistic sybarite. He remained the most simultaneously admired and despised philosopher in the Mediterranean for the next nearly five centuries. Epicureanism rapidly spread beyond the Greek mainland all across the Mediterranean world. By the first century BC, it had established a strong foothold in Italy. The Roman orator Cicero (106 – 43 BC), who deplored Epicurean ethics, lamented, "the Epicureans have taken Italy by storm." The overwhelming majority of surviving Greek and Roman sources are vehemently negative towards Epicureanism and, according to Pamela Gordon, they routinely depict Epicurus himself as "monstrous or laughable". Many Romans in particular took a negative view of Epicureanism, seeing its advocacy of the pursuit of "voluptas" ("pleasure") as contrary to the Roman ideal of "virtus" ("manly virtue"). The Romans therefore often stereotyped Epicurus and his followers as weak and effeminate. Prominent critics of his philosophy include prominent authors such as the Roman Stoic Seneca the Younger ( 4 BC – AD 65) and the Greek Middle Platonist Plutarch ( 46 – 120), who both derided it as immoral and disreputable. Gordon characterizes anti-Epicurean rhetoric as so "heavy-handed" and misrepresentative of Epicurus's actual teachings that they sometimes come across as "comical". In his "De vita beata", Seneca states that the "sect of Epicurus... has a bad reputation" and compares it to "a man in a dress: your chastity remains, your virility is unimpaired, your body has not submitted sexually, but in your hand is a tympanum." Epicureanism was a notoriously conservative philosophical school; although Epicurus's later followers did expand on his philosophy, they dogmatically retained what he himself had originally taught without modifying it. Epicureans and admirers of Epicureanism revered Epicurus himself as a great teacher of ethics, a savior, and even a god. His image was worn on finger rings, portraits of him were displayed in living rooms, and wealthy followers venerated likenesses of him in marble sculpture. His admirers revered his sayings as divine oracles, carried around copies of his writings, and cherished copies of his letters like the letters of an apostle. On the twentieth day of every month, admirers of his teachings would perform a solemn ritual to honor his memory. At the same time, opponents of his teachings denounced him with vehemence and persistence. However, in the first and second centuries AD, Epicureanism gradually began to decline as it failed to compete with Stoicism, which had an ethical system more in line with traditional Roman values. Epicureanism also suffered decay in the wake of Christianity, which was also rapidly expanding throughout the Roman Empire. Of all the Greek philosophical schools, Epicureanism was the one most at odds with the new Christian teachings, since Epicureans believed that the soul was mortal, denied the existence of an afterlife, denied that the divine had any active role in human life, and advocated pleasure as the foremost goal of human existence. As such, Christian writers such as Justin Martyr ( 100– 165 AD), Athenagoras of Athens ( 133– 190), Tertullian ( 155– 240), and Clement of Alexandria ( 150– 215), Arnobius (died 330), and Lactantius all singled it out for the most vitriolic criticism. In spite of this, DeWitt argues that Epicureanism and Christianity share much common language, calling Epicureanism "the first missionary philosophy" and "the first world philosophy". Both Epicureanism and Christianity placed strong emphasis on the importance of love and forgiveness and early Christian portrayals of Jesus are often similar to Epicurean portrayals of Epicurus. DeWitt argues that Epicureanism, in many ways, helped pave the way for the spread of Christianity by "helping to bridge the gap between Greek intellectualism and a religious way of life" and "shunt[ing] the emphasis from the political to the social virtues and offer[ing] what may be called a religion of humanity." By the early fifth century AD, Epicureanism was virtually extinct. The Christian Church Father Augustine of Hippo (354–430 AD) declared, "its ashes are so cold that not a single spark can be struck from them." While the ideas of Plato and Aristotle could easily be adapted to suit a Christian worldview, the ideas of Epicurus were not nearly as easily amenable. As such, while Plato and Aristotle enjoyed a privileged place in Christian philosophy throughout the Middle Ages, Epicurus was not held in such esteem. Information about Epicurus's teachings was available, through Lucretius's "On the Nature of Things", quotations of it found in medieval Latin grammars and "florilegia", and encyclopedias, such as Isidore of Seville's "Etymologiae" (seventh century) and Hrabanus Maurus's "De universo" (ninth century), but there is little evidence that these teachings were systematically studied or comprehended. During the Middle Ages, Epicurus was remembered by the educated as a philosopher, but he frequently appeared in popular culture as a gatekeeper to the Garden of Delights, the "proprietor of the kitchen, the tavern, and the brothel." He appears in this guise in Martianus Capella's "Marriage of Mercury and Philology" (fifth century), John of Salisbury's "Policraticus" (1159), John Gower's "Mirour de l'Omme", and Geoffrey Chaucer's "Canterbury Tales". Epicurus and his followers appear in Dante Alighieri's "Inferno" in the Sixth Circle of Hell, where they are imprisoned in flaming coffins for having believed that the soul dies with the body. In 1417, a manuscript-hunter named Poggio Bracciolini discovered a copy of Lucretius's "On the Nature of Things" in a monastery near Lake Constance. The discovery of this manuscript was met with immense excitement, because scholars were eager to analyze and study the teachings of classical philosophers and this previously-forgotten text contained the most comprehensive account of Epicurus's teachings known in Latin. The first scholarly dissertation on Epicurus, "De voluptate" ("On Pleasure") by the Italian Humanist and Catholic priest Lorenzo Valla was published in 1431. Valla made no mention of Lucretius or his poem. Instead, he presented the treatise as a discussion on the nature of the highest good between an Epicurean, a Stoic, and a Christian. Valla's dialogue ultimately rejects Epicureanism, but, by presenting an Epicurean as a member of the dispute, Valla lent Epicureanism credibility as a philosophy that deserved to be taken seriously. None of the Quattrocento Humanists ever clearly endorsed Epicureanism, but scholars such as Francesco Zabarella (1360–1417), Francesco Filelfo (1398–1481), Cristoforo Landino (1424–1498), and Leonardo Bruni ( 1370–1444) did give Epicureanism a fairer analysis than it had traditionally received and provided a less overtly hostile assessment of Epicurus himself. Nonetheless, "Epicureanism" remained a pejorative, synonymous with extreme egoistic pleasure-seeking, rather than a name of a philosophical school. This reputation discouraged orthodox Christian scholars from taking what others might regard as an inappropriately keen interest in Epicurean teachings. Epicureanism did not take hold in Italy, France, or England until the seventeenth century. Even the liberal religious skeptics who might have been expected to take an interest in Epicureanism evidently did not; Étienne Dolet (1509–1546) only mentions Epicurus once in all his writings and François Rabelais (between 1483 and 1494–1553) never mentions him at all. Michel de Montaigne (1533–1592) is the exception to this trend, quoting a full 450 lines of Lucretius's "On the Nature of Things" in his "Essays". His interest in Lucretius, however, seems to have been primarily literary and he is ambiguous about his feelings on Lucretius's Epicurean worldview. During the Protestant Reformation, the label "Epicurean" was bandied back and forth as an insult between Protestants and Catholics. In the seventeenth century, the French Catholic priest and scholar Pierre Gassendi (1592–1655) sought to dislodge Aristotelianism from its position of the highest dogma by presenting Epicureanism as a better and more rational alternative. In 1647, Gassendi published his book "De vita et moribus Epicuri" ("The Life and Morals of Epicurus"), a passionate defense of Epicureanism. In 1649, he published a commentary on Diogenes Laërtius's "Life of Epicurus". He left "Syntagma philosophicum" ("Philosophical Compendium"), a synthesis of Epicurean doctrines, unfinished at the time of his death in 1655. It was finally published in 1658, after undergoing revision by his editors. Gassendi modified Epicurus's teachings to make them palatable for a Christian audience. For instance, he argued that atoms were not eternal, uncreated, and infinite in number, instead contending that an extremely large but finite number of atoms were created by God at creation. As a result of Gassendi's modifications, his books were never censored by the Catholic Church. They came to exert profound influence on later writings about Epicurus. Gassendi's version of Epicurus's teachings became popular among some members of English scientific circles. For these scholars, however, Epicurean atomism was merely a starting point for their own idiosyncratic adaptations of it. To orthodox thinkers, Epicureanism was still regarded as immoral and heretical. For instance, Lucy Hutchinson (1620–1681), the first translator of Lucretius's "On the Nature of Things" into English, railed against Epicurus as "a lunatic dog" who formulated "ridiculous, impious, execrable doctrines". Epicurus's teachings were made respectable in England by the natural philosopher Walter Charleton (1619–1707), whose first Epicurean work, "The Darkness of Atheism Dispelled by the Light of Nature" (1652), advanced Epicureanism as a "new" atomism. His next work "Physiologia Epicuro-Gassendo-Charletoniana, or a Fabrick of Science Natural, upon a Hypothesis of Atoms, Founded by Epicurus, Repaired by Petrus Gassendus, and Augmented by Walter Charleton" (1654) emphasized this idea. These works, together with Charleton's "Epicurus's Morals" (1658), provided the English public with readily available descriptions of Epicurus's philosophy and assured orthodox Christians that Epicureanism was no threat to their beliefs. The Royal Society, chartered in 1662, advanced Epicurean atomism. One of the most prolific defenders of atomism was the chemist Robert Boyle (1627–1691), who argued for it in publications such as "The Origins of Forms and Qualities" (1666), "Experiments, Notes, etc. about the Mechanical Origin and Production of Divers Particular Qualities" (1675), and "Of the Excellency and Grounds of the Mechanical Hypothesis" (1674). By the end of the seventeenth century, Epicurean atomism was widely accepted by members of the English scientific community as the best model for explaining the physical world, but it had been modified so greatly that Epicurus was no longer seen as its original parent. The Anglican bishop Joseph Butler's anti-Epicurean polemics in his "Fifteen Sermons Preached at the Rolls Chapel" (1726) and "Analogy of Religion" (1736) set the tune for what most orthodox Christians believed about Epicureanism for the remainder of the eighteenth and nineteenth centuries. Nonetheless, there are a few indications from this time period of Epicurus's improving reputation. Epicureanism was beginning to lose its associations with indiscriminate and insatiable gluttony, which had been characteristic of its reputation ever since antiquity. Instead, the word "epicure" began to refer to a person with extremely refined taste in food. Examples of this usage include "Epicurean cooks / sharpen with cloyless sauce his appetite" from William Shakespeare's "Antony and Cleopatra" (Act II. scene i; 1607) and "such an epicure was Potiphar—to please his tooth and pamper his flesh with delicacies" from William Whately's "Prototypes" (1646). Around the same time, the Epicurean injunction to "live in obscurity" was beginning to gain popularity as well. In 1685, Sir William Temple (1628–1699) abandoned a promising career as a diplomat and instead retired to his garden, devoting himself to writing essays on Epicurus's moral teachings. That same year, John Dryden translated the celebrated lines from Book II of Lucretius's "On the Nature of Things": "'Tis pleasant, safely to behold from shore / The rowling ship, and hear the Tempest roar." Meanwhile, John Locke (1632–1704) adapted Gassendi's modified version of Epicurus's epistemology, which became highly influential on English empiricism. Many thinkers with sympathies towards the Enlightenment endorsed Epicureanism as an admirable moral philosophy. Thomas Jefferson (1743–1826), one of the Founding Fathers of the United States, declared in 1819, "I too am an Epicurean. I consider the genuine (not imputed) doctrines of Epicurus as containing everything rational in moral philosophy which Greece and Rome have left us." The German philosopher Karl Marx (1818–1883), whose ideas are basis of Marxism, was profoundly influenced as a young man by the teachings of Epicurus and his doctoral thesis was a Hegelian dialectical analysis of the differences between the natural philosophies of Democritus and Epicurus. Marx viewed Democritus as a rationalist skeptic, whose epistemology was inherently contradictory, but saw Epicurus as a dogmatic empiricist, whose worldview is internally consistent and practically applicable. The British poet Alfred, Lord Tennyson (1809–1892) praised "the sober majesties / of settled, sweet, Epicurean life" in his 1868 poem "Lucretius". Epicurus's ethical teachings also had an indirect impact on the philosophy of Utilitarianism in England during the nineteenth century. Academic interest in Epicurus and other Hellenistic philosophers increased over the course of the late twentieth and early twenty-first centuries, with an unprecedented number of monographs, articles, abstracts, and conference papers being published on the subject. The texts from the library of Philodemus of Gadara in the Villa of the Papyri in Herculaneum, first discovered between 1750 and 1765, are being deciphered, translated, and published by scholars part of the Philodemus Translation Project, funded by the United States National Endowment for the Humanities, and part of the Centro per lo Studio dei Papiri Ercolanesi in Naples. Epicurus's popular appeal among non-scholars is difficult to gauge, but it seems to be relatively comparable to the appeal of more traditionally popular ancient Greek philosophical subjects such as Stoicism, Aristotle, and Plato.
https://en.wikipedia.org/wiki?curid=10073
Epigram An epigram is a brief, interesting, memorable, and sometimes surprising or satirical statement. The word is derived from the "epigramma" 'inscription' from ἐπιγράφειν "epigraphein" 'to write on, to inscribe', and the literary device has been employed for over two millennia. The presence of wit or sarcasm tends to distinguish non-poetic epigrams from aphorisms and adages, which tend to lack those qualities. The Greek tradition of epigrams began as poems inscribed on votive offerings at sanctuariesincluding statues of athletesand on funerary monuments, for example "Go tell it to the Spartans, passersby...". These original epigrams did the same job as a short prose text might have done, but in verse. Epigram became a literary genre in the Hellenistic period, probably developing out of scholarly collections of inscriptional epigrams. Though modern epigrams are usually thought of as very short, Greek literary epigram was not always as short as later examples, and the divide between "epigram" and "elegy" is sometimes indistinct (they share a characteristic metre, elegiac couplets). In the classical period, the clear distinction between them was that epigrams were inscribed and meant to be read, while elegies were recited and meant to be heard. Some elegies could be quite short, but only public epigrams were longer than ten lines. All the same, the origin of epigram in inscription exerted a residual pressure to keep things concise, even when they were recited in Hellenistic times. Many of the characteristic types of literary epigram look back to inscriptional contexts, particularly funerary epigram, which in the Hellenistic era becomes a literary exercise. Many "sympotic" epigrams combine sympotic and funerary elementsthey tell their readers (or listeners) to drink and live for today because life is short. Generally, any theme found in classical elegies could be and were adapted for later literary epigrams. Hellenistic epigrams are also thought of as having a "point"that is, the poem ends in a punchline or satirical twist. By no means do all Greek epigrams behave this way; many are simply descriptive, but Meleager of Gadara and Philippus of Thessalonica, the first comprehensive anthologists, preferred the short and witty epigram. Since their collections helped form knowledge of the genre in Rome and then later throughout Europe, Epigram came to be associated with 'point,' especially because the European epigram tradition takes the Latin poet Martial as its principal model; he copied and adapted Greek models (particularly the contemporary poets Lucillius and Nicarchus) selectively and in the process redefined the genre, aligning it with the indigenous Roman tradition of 'satura', hexameter satire, as practised by (among others) his contemporary Juvenal. Greek epigram was actually much more diverse, as the Milan Papyrus now indicates. A major source for Greek literary epigram is the "Greek Anthology", a compilation from the 10th century AD based on older collections, including those of Meleager and Philippus. It contains epigrams ranging from the Hellenistic period through the Imperial period and Late Antiquity into the compiler's own Byzantine eraa thousand years of short elegiac texts on every topic under the sun. The "Anthology" includes one book of Christian epigrams as well as one book of erotic and amorous homosexual epigrams called the Μοῦσα Παιδικἠ (Mousa Paidike, "The Boyish Muse"). Roman epigrams owe much to their Greek predecessors and contemporaries. Roman epigrams, however, were often more satirical than Greek ones, and at times used obscene language for effect. Latin epigrams could be composed as inscriptions or graffiti, such as this one from Pompeii, which exists in several versions and seems from its inexact meter to have been composed by a less educated person. Its content makes it clear how popular such poems were: However, in the literary world, epigrams were most often gifts to patrons or entertaining verse to be published, not inscriptions. Many Roman writers seem to have composed epigrams, including Domitius Marsus, whose collection "Cicuta" (now lost) was named after the poisonous plant "Cicuta" for its biting wit, and Lucan, more famous for his epic "Pharsalia". Authors whose epigrams survive include Catullus, who wrote both invectives and love epigrams – his poem 85 is one of the latter. Martial, however, is considered to be the master of the Latin epigram. His technique relies heavily on the satirical poem with a joke in the last line, thus drawing him closer to the modern idea of epigram as a genre. Here he defines his genre against a (probably fictional) critic (in the latter half of 2.77): Poets known for their epigrams whose work has been lost include Cornificia. In early English literature the short couplet poem was dominated by the poetic epigram and proverb, especially in the translations of the Bible and the Greek and Roman poets. Since 1600, two successive lines of verse that rhyme with each other, known as a couplet featured as a part of the longer sonnet form, most notably in William Shakespeare's sonnets. Sonnet 76 is an excellent example. The two line poetic form as a closed couplet was also used by William Blake in his poem Auguries of Innocence, and also by Byron in his poem Don Juan, by John Gay in his fables and by Alexander Pope in his An Essay on Man. The first work of English literature penned in North America was Robert Hayman's "Quodlibets, Lately Come Over from New Britaniola, Old Newfoundland," which is a collection of over 300 epigrams, many of which do not conform to the two-line rule or trend. While the collection was written between 1618 and 1628 in what is now Harbour Grace, Newfoundland, it was published shortly after his return to Britain. In Victorian times the epigram couplet was often used by the prolific American poet Emily Dickinson. Her poem No. 1534 is a typical example of her eleven poetic epigrams. The novelist George Eliot also included couplets throughout her writings. Her best example is in her sequenced sonnet poem entitled "Brother and Sister" in which each of the eleven sequenced sonnet ends with a couplet. In her sonnets, the preceding lead-in-line, to the couplet ending of each, could be thought of as a title for the couplet, as is shown in Sonnet VIII of the sequence. During the early 20th century, the rhymed epigram couplet form developed into a fixed verse image form, with an integral title as the third line. Adelaide Crapsey codified the couplet form into a two line rhymed verse of ten syllables per line with her image couplet poem "On Seeing Weather-Beaten Trees" first published in 1915. By the 1930s, the five-line cinquain verse form became widely known in the poetry of the Scottish poet William Soutar. These were originally labelled epigrams but later identified as image cinquains in the style of Adelaide Crapsey. J. V. Cunningham was also a noted writer of epigrams, (a medium suited to a 'short-breathed' person).
https://en.wikipedia.org/wiki?curid=10075
El Cid Rodrigo Díaz de Vivar ( 10 July 1099) was a Castilian knight and warlord in medieval Spain. The Moors called him El Cid (), which meant "the Lord" (probably from the original ), and the Christians, El Campeador, which means "the Champion" in modern Spanish, but is literally translated as "the Master of the Battlefield" in Old Spanish. He was born in Vivar del Cid, a village near the city of Burgos. Díaz de Vivar became well known for his service in the armies of both Christian and Muslim rulers, his exile, and his temporary conquest of Valencia, which became independent for a brief period in the Reconquista. After his death, El Cid became Spain's celebrated national hero and the protagonist of the most significant medieval Spanish epic poem, "El Cantar de Mio Cid." To this day, El Cid remains a Spanish popular folk-hero and national icon, with his life and deeds remembered in plays, films, folktales, songs, and video games. Born a member of the minor nobility, El Cid was brought up at the court of Ferdinand the Great and served Ferdinand's son, Sancho II of León and Castile. He rose to become the commander and royal standard-bearer ("armiger regis") of Castile upon Sancho's ascension in 1065. Rodrigo went on to lead the Castilian military campaigns against Sancho's brothers, Alfonso VI of León and García II of Galicia, as well as in the Muslim kingdoms in al-Andalus. He became renowned for his military prowess in these campaigns, which helped expand the territory of the Crown of Castile at the expense of the Muslims and Sancho's brothers' kingdoms. When conspirators murdered Sancho in 1072, Rodrigo found himself in a difficult situation. Since Sancho was childless, the throne passed to his brother Alfonso, whom El Cid had helped remove from power. Although Rodrigo continued to serve the sovereign, he lost his ranking in the new court, which treated him suspiciously and kept him at arm's length. Finally, in 1081, he was exiled. El Cid found work fighting for the Muslim rulers of Zaragoza, whom he defended from its traditional enemy, Aragon. While in exile, he regained his reputation as a strategist and formidable military leader. He was repeatedly victorious in battle against the Muslim rulers of Lérida and their Christian allies, as well as against a large Christian army under King Sancho Ramírez of Aragon. In 1086, an expeditionary army of North African Almoravids inflicted a severe defeat to Castile, compelling Alfonso to overcome the resentment he harboured against El Cid. The terms for El Cid's return to Christian service must have been attractive enough since Rodrigo soon found himself fighting for his former lord. Over the next several years, however, El Cid set his sights on the kingdom-city of Valencia, operating more or less independently of Alfonso, while politically supporting the Banu Hud and other Muslim dynasties opposed to the Almoravids. He gradually increased his control over Valencia; the Islamic ruler, , became his tributary in 1092. When the Almoravids instigated an uprising that resulted in the death of Al-Cádir, El Cid responded by laying siege to the city. Valencia finally fell in 1094, and El Cid established an independent principality on the Mediterranean coast of Spain. He ruled over a pluralistic society with the popular support of Christians and Muslims alike. El Cid's final years were spent fighting the Almoravid Berbers. He inflicted upon them their first major defeat in 1094, on the plains of Caurte, outside Valencia, and continued opposing them until his death. Although Rodrigo remained undefeated in Valencia, Diego Rodríguez, his only son and heir, died fighting against the Almoravids in the service of Alfonso in 1097. After El Cid's death in 1099, his wife, Jimena Díaz, succeeded him as ruler of Valencia, but she was eventually forced to surrender the principality to the Almoravids in 1102. The name "El Cid" () is a modern Spanish denomination composed of the article "el" meaning "the" and "Cid", which derives from the Old Castilian loan word "Çid" borrowed from the dialectal Arabic word سيد "sîdi" or sayyid, which means "Lord" or "Master". The Mozarabs or the Arabs that served in his ranks may have addressed him in this way, which the Christians may have transliterated and adopted. Historians, however, have not yet found contemporary records referring to Rodrigo as "Cid." Arab sources use instead "Rudriq", "Ludriq al-Kanbiyatur" or "al-Qanbiyatur" ("Rodrigo el Campeador"). The cognomen "Campeador" derives from Latin "campi doctor," which means "battlefield master". He probably gained it during the campaigns of King Sancho II of Castile against his brothers King Alfonso VI of León and King García II of Galicia. While his contemporaries left no historical sources that would have addressed him as "Cid", they left plenty of Christian and Arab records, some even signed documents with his autograph, addressing him as "Campeador", which prove that he used the Christian cognomen himself. The whole combination "Cid Campeador" is first documented ca. 1195 in the Navarro-Aragonese "Linage de Rodric Díaz" included in the "Liber Regum" under the formula "mio Cid el Campeador". El Cid was born Rodrigo Díaz circa 1043 in Vivar, also known as Castillona de Bivar, a small town about six miles north of Burgos, the capital of Castile. His father, Diego Laínez, was a courtier, bureaucrat, and cavalryman who had fought in several battles. Despite the fact that El Cid's mother's family was aristocratic, in later years the peasants would consider him one of their own. However, his relatives were not major court officials; documents show that El Cid's paternal grandfather, Laín, confirmed only five documents of Ferdinand I's; his maternal grandfather, Rodrigo Álvarez, certified only two of Sancho II's; and El Cid's father confirmed only one. As a young man in 1057, Rodrigo fought against the Moorish stronghold of Zaragoza, making its emir al-Muqtadir a vassal of Sancho. In the spring of 1063, Rodrigo fought in the Battle of Graus, where Ferdinand's half-brother, Ramiro I of Aragon, was laying siege to the Moorish town of Graus, which was fought on Zaragozan lands in the valley of the river Cinca. Al-Muqtadir, accompanied by Castilian troops including El Cid, fought against the Aragonese. The party slew Ramiro I, setting the Aragonese army on the run, and emerged victorious. One legend has said that during the conflict, El Cid killed an Aragonese knight in single combat, thereby receiving the honorific title ""Campeador"". When Ferdinand died, Sancho continued to enlarge his territory, conquering both Christian strongholds and the Moorish cities of Zamora and Badajoz. When Sancho learned that Alfonso was planning on overthrowing him in order to gain his territory, Sancho sent Cid to bring Alfonso back so that Sancho could speak to him. Sancho was assassinated in 1072, possibly as the result of a pact between his brother Alfonso and his sister Urraca. Since Sancho died unmarried and childless, all of his power passed to his brother Alfonso who, almost immediately, returned from exile in Toledo and took his seat as king of Castile and León. He was, however, deeply suspected of having been involved in Sancho's murder. According to the epic of El Cid, the Castilian nobility led by El Cid and a dozen "oath-helpers" forced Alfonso to swear publicly on holy relics multiple times in front of Santa Gadea (Saint Agatha) Church in Burgos that he did not participate in the plot to kill his brother. This is widely reported as truth, but contemporary documents on the lives of both Rodrigo Diaz and Alfonso VI of Castile and León do not mention any such event. Rodrigo's position as "armiger regis" was taken away and given to Rodrigo's enemy, Count García Ordóñez. In 1079, Rodrigo was sent by Alfonso VI to Seville to the court of al-Mutamid to collect the "parias" owed by that "taifa" to León–Castile. While he was there Granada, assisted by other Castilian knights, attacked Seville, and Rodrigo and his forces repulsed the Christian and Grenadine attackers at the Battle of Cabra, in the (probably mistaken) belief that he was defending the king's tributary. Count García Ordóñez and the other Castilian leaders were taken captive and held for three days before being released. In the Battle of Cabra (1079), El Cid rallied his troops and turned the battle into a rout of Emir Abdullah of Granada and his ally García Ordóñez. This unauthorized expedition into Granada however, greatly angered Alfonso and May 8, 1080 was the last time El Cid confirmed a document in King Alfonso's court. This is the generally accepted reason for the exile of El Cid, although several others are plausible and indeed may have been contributing factors to the exile: jealous nobles turning Alfonso against El Cid through court intrigue, Alfonso's own personal animosity towards El Cid, as well as a possible misappropriation of some of the tribute from Seville by El Cid. At first he went to Barcelona, where Ramon Berenguer II refused his offer of service. The exile was not the end of El Cid, either physically or as an important figure. After being rejected by Ramon Berenguer II, El Cid journeyed to the Taifa of Zaragoza, where he received a warmer welcome. In 1081, El Cid went on to offer his services to the Moorish king of the northeast al-Andalus city of Zaragoza, Yusuf al-Mu'taman ibn Hud, and served both him and his successor, al-Musta'in II. He was given the title "El Cid" ("The Master") and served as a leading figure in a diverse Moorish force consisting of Muladis, Berbers, Arabs and Malians within the respective Taifa. According to Moorish accounts: Andalusi Knights found El Cid their foe ill, thirsty and exiled from the court of Alfonso, he was presented before the elderly Yusuf al-Mu'taman ibn Hud and accepted command of the forces of the Taifa of Zaragoza as their Master. In his "History of Medieval Spain" (Cornell University Press, 1975), Joseph F. O'Callaghan writes: That kingdom was divided between al-Mutamin (1081–1085) who ruled Zaragoza proper, and his brother al-Mundhir, who ruled Lérida and Tortosa. El Cid entered al-Mutamin's service and successfully defended Zaragoza against the assaults of al-Mundhir, Sancho I of Aragón, and Ramon Berenguer II, whom he held captive briefly in 1082. In 1084, the army of the Taifa of Zaragoza under El Cid defeated the Aragonese at the Battle of Morella near Tortosa, but in autumn the Castilians started a loose siege of Toledo and later the next year the Christians captured Salamanca, a stronghold of the Taifa of Toledo. In 1086, the Almoravid invasion of the Iberian Peninsula, through and around Gibraltar, began. The Almoravids, a Berber dynasty from North Africa, led by Yusuf ibn Tashfin, were asked to help defend the divided Moors from Alfonso. The Almoravid army, joined by that of several and Taifas, including Badajoz, Málaga, Granada, Tortosa and Seville, defeated a combined army of León, Aragón and Castile at the Battle of Sagrajas. In 1087, Raymond of Burgundy and his Christian allies attempted to weaken the Taifa of Zaragoza's northernmost stronghold by initiating the Siege of Tudela and Alfonso captured Aledo, Murcia, blocking the route between the Taifas in the eastern and western Iberian Peninsula. Terrified after his crushing defeat, Alfonso recalled El Cid. It has been shown that El Cid was at court on July 1087; however, what happened after that is unclear. El Cid returned to Alfonso, but now he had his own plans. He only stayed a short while and then returned to Zaragoza. El Cid was content to let the Almoravid armies and the armies of Alfonso fight without his help, even when there was a chance that the Almoravids might defeat Alfonso and take over all of Alfonso's lands. El Cid chose not to fight because he was hoping that both armies would weaken themselves. That would make it easier for him to carry out his own plan to become ruler of the Kingdom of Valencia. Around this time, El Cid, with a combined Christian and Moorish army, began maneuvering in order to create his own fiefdom in the Moorish Mediterranean coastal city of Valencia. Several obstacles lay in his way. First was Berenguer Ramon II, who ruled nearby Barcelona. In May 1090, El Cid defeated and captured Berenguer in the Battle of Tébar (nowadays Pinar de Tévar, near Monroyo, Teruel). Berenguer was later released and his nephew Ramon Berenguer III married El Cid's youngest daughter Maria to ward against future conflicts. Along the way to Valencia, El Cid also conquered other towns, many of which were near Valencia, such as El Puig and Quart de Poblet. El Cid gradually came to have more influence in Valencia, then ruled by , of the Hawwara Berber Dhunnunid dynasty. In October 1092 an uprising occurred in Valencia, inspired by the city's chief judge Ibn Jahhaf and the Almoravids. El Cid began a siege of Valencia. A December 1093 attempt to break the siege failed. By the time the siege ended in May 1094, El Cid had carved out his own principality on the coast of the Mediterranean. Officially El Cid ruled in the name of Alfonso; in reality, El Cid was fully independent. The city was both Christian and Muslim, and both Moors and Christians served in the army and as administrators. El Cid and his wife Jimena Díaz lived peacefully in Valencia for five years until the Almoravids besieged the city. El Cid died on July 10, 1099. His death was likely a result of the famine and deprivations caused by the siege. Valencia was captured by Masdali on May 5, 1102 and it did not become a Christian city again for over 125 years. Jimena fled to Burgos, Castile, in 1101. She rode into the town with her retinue and the body of El Cid. Originally buried in Castile in the monastery of , his body now lies at the center of Burgos Cathedral. After his demise, but still during the siege of Valencia, legend holds that Jimena ordered that the corpse of El Cid be fitted with his armour and set on his horse Babieca, to bolster the morale of his troops. In several variations of the story, the dead Rodrigo and his knights win a thundering charge against Valencia's besiegers, resulting in a war-is-lost-but-battle-is-won catharsis for generations of Christian Spaniards to follow. It is believed that the legend originated shortly after Jimena entered Burgos, and that it is derived from the manner in which Jimena's procession rode into Burgos, i.e., alongside her deceased husband. During his campaigns, El Cid often ordered that books by classic Roman and Greek authors on military themes be read aloud to him and his troops, for both entertainment and inspiration before battle. El Cid's army had a novel approach to planning strategy as well, holding what might be called "brainstorming" sessions before each battle to discuss tactics. They frequently used unexpected strategies, engaging in what modern generals would call psychological warfare — waiting for the enemy to be paralyzed with terror and then attacking them suddenly; distracting the enemy with a small group of soldiers, etc. (El Cid used this distraction in capturing the town of Castejón as depicted in "Cantar de Mio Cid" ("The Song of my Cid"). El Cid accepted or included suggestions from his troops. In "The Song" the man who served him as his closest adviser was his vassal and kinsman Álvar Fáñez ""Minaya"" (meaning ""My brother"", a compound word of Spanish possessive "Mi" (My) and "Anaia", the basque word for "brother"), although the historical Álvar Fáñez remained in Castile with Alfonso VI. Babieca or Bavieca was El Cid's warhorse. Several stories exist about El Cid and Babieca. One well-known legend about El Cid describes how he acquired the stallion. According to this story, Rodrigo's godfather, Pedro El Grande, was a monk at a Carthusian monastery. Pedro's coming-of-age gift to El Cid was his pick of a horse from an Andalusian herd. El Cid picked a horse that his godfather thought was a weak, poor choice, causing the monk to exclaim ""Babieca"!" (stupid!) Hence, it became the name of El Cid's horse. Another legend states that in a competition of battle to become King Sancho's "Campeador", or champion, a knight on horseback wished to challenge El Cid. The King wished a fair fight and gave El Cid his finest horse, Babieca, or Bavieca. This version says Babieca was raised in the royal stables of Seville and was a highly trained and loyal war horse, not a foolish stallion. The name in this instance could suggest that the horse came from the Babia region in León, Spain. In the poem "Carmen Campidoctoris", Babieca appears as a gift from "a barbarian" to El Cid, so its name could also be derived from "Barbieca", or "horse of the barbarian". Regardless, Babieca became a great warhorse, famous to the Christians, feared by El Cid's enemies, and loved by El Cid, who allegedly requested that Babieca be buried with him in the monastery of San Pedro de Cardeña. His name is mentioned in several tales and historical documents about El Cid, including "The Lay of El Cid". A weapon traditionally identified as El Cid's sword, Tizona, used to be displayed in the Army Museum (Museo del Ejército) in Toledo. In 1999, a small sample of the blade underwent metallurgical analysis which confirmed that the blade was made in Moorish Córdoba in the eleventh century and contained amounts of Damascus steel. In 2007, the Autonomous Community of Castile and León bought the sword for 1.6 million Euros, and it is currently on display at the Museum of Burgos. El Cid also had a sword called Colada. El Cid married Jimena Díaz, who was said to be part of an aristocratic family from Asturias, in the mid-1070s. The "Historia Roderici" calls her a daughter of a Count Diego Fernández de Oviedo. Tradition states that when El Cid first laid eyes on her, he was enamoured of her great beauty. El Cid and Jimena had two daughters and a son. The latter, Diego Rodríguez, was killed while fighting against the invading Muslim Almoravids from North Africa at the Battle of Consuegra in 1097. As both the Poem and Chronicle state, Cristina Rodríguez and María were originally married to the Infantes de Carrión along with the generous wedding gifts of his two famous swords, Tizona and Colada. However, in revenge for an incident that the Infantes blamed El Cid for, they tied their wives to trees and left them to die. El Cid saved their lives and avenged himself upon the Infantes, and found them more favorable husbands. Cristina Rodríguez, (known as Elvira by the Poem and the Chronicle) married Ramiro, Lord of Monzón and grandson of García Sánchez III of Navarre. Her own son, El Cid's grandson, would be elevated to the throne of Navarre as King García Ramírez. The other daughter, María (also known as Sol), is said first to have married a prince of Aragon, presumably the son of Peter I, and she later wed Ramon Berenguer III, count of Barcelona. The figure of El Cid has been the source for many literary works, beginning with the "Cantar de Mio Cid", an epic poem from the 12th century which gives a partly-fictionalized account of his life, and was one of the early chivalric romances. This poem, along with similar later works such as the "Mocedades de Rodrigo", contributed to portray El Cid as a chivalric hero of the Reconquista, making him a legendary figure in Spain. In the early 17th century the Spanish writer Guillén de Castro wrote a play called "Las Mocedades del Cid", on which French playwright Pierre Corneille based one of his most famous tragicomedies, "Le Cid". He was also a popular source of inspiration for Spanish writers of the Romantic period, such as Juan Eugenio Hartzenbusch, who wrote "La Jura de Santa Gadea", or José Zorrilla, who wrote a long poem called "La Leyenda del Cid". In 2019, Arturo Pérez-Reverte published the novel entitled "Sidi: Un relato de frontera". Georges Bizet worked on "Don Rodrigue" in 1873 that was set aside and never completed. Jules Massenet wrote an opera, "Le Cid", in 1885, based on Corneille's play of the same name. Claude Debussy began work in 1890 on an opera, "Rodrigue et Chimène", which he abandoned as unsuitable for his temperament; it was orchestrated for performance by Edison Denisov circa 1993. El Cid is portrayed by American actor Charlton Heston in a 1961 epic film of the same name directed by Anthony Mann, where the character of Doña Ximena is portrayed by Italian actress Sophia Loren. In 2019, Amazon Prime Video announced a new Spanish TV series with Jaime Lorente starring as "El Cid". In 1980, "Ruy, the Little Cid" was an animated series based on El Cid's childhood made by Nippon Animation. In the second "Age of Empires" video game installment, the "" expansion pack, there is a campaign starring El Cid Campeador. In both the first and second "" games, El Cid appears as a powerful independent general in the castle of Valencia. In 2003, the Spanish animated film "" was released. "The Ministry of Time", a Spanish science fiction television series, portrayed El Cid in Season 2, Episode 1 (air date: February 15, 2016)
https://en.wikipedia.org/wiki?curid=10076
European Convention on Nationality The European Convention on Nationality (E.T.S. No. 166) was signed in Strasbourg on 6 November 1997. It is a comprehensive convention of the Council of Europe dealing with the law of nationality. The Convention is open for signature by the member States of the Council of Europe and the non-member States which have participated in its elaboration and for accession by other non-member States. The Convention came into force on 1 March 2000 after ratification by 3 countries. As at 6 March 2014, the Convention has been signed by 29 countries, but has been ratified by only 20 of those countries. Article 4(d) provides that neither marriage nor dissolution of marriage shall automatically affect the nationality of either spouse, nor shall a change of nationality by one spouse during marriage automatically affect the nationality of their spouse. Common practice among states at the beginning of the 20th century was that a woman was to have the nationality of her husband; i.e., upon marrying a foreigner the wife would automatically acquire the nationality of her husband, and lose her previous nationality. Even after the nationality of a married woman was no longer dependent on the nationality of her husband, legal provisions were still retained which automatically naturalised married women, and sometimes married men as well. This led to a number of problems, such as loss of the spouses' original nationality, the spouse losing the right to consular assistance (since consular assistance cannot be provided to nationals under the jurisdiction of a foreign state of which they are also nationals), and becoming subject to military service obligations. Article 4d addresses this situation. Article 5 provides that no discrimination shall exist in a state's internal nationality law on the grounds of "sex, religion, race, colour or national or ethnic origin". It also provides that a state shall not discriminate amongst its nationals on the basis of whether they hold their nationality by birth or acquired it subsequently. Article 6 relates to the acquisition of nationality. It provides for nationality to be acquired at birth by descent from either parent to those born within the territory of the state. (States may exclude partially or fully children born abroad). It also provides for nationality by virtue of birth in the territory of state; however, states may limit this to only children who would be otherwise stateless. It requires the possibility of naturalisation, and provides that the period of residence required for eligibility cannot be more than ten years lawful and habitual residence. It also requires to "facilitate" the acquisition of nationality by certain persons, including spouses of nationals, children of its nationals born abroad, children one of whose parents has acquired the nationality, children adopted by a national, persons lawfully and habitually resident for a period before the age of eighteen, and stateless persons and refugees lawfully and habitually resident on its territory. Article 7 regulates the involuntary loss of nationality. It provides that states may deprive their nationals of their nationality in only the cases of voluntary acquisition of another nationality, fraud or failure to provide relevant information when acquiring nationality, voluntary military service in a foreign military force, or adoption as a child by foreign nationals. It also provides for the possibility of loss of nationality for nationals habitually residing abroad. Finally it provides loss of nationality for "conduct seriously prejudicial to the vital interests of the State Party". Article 8 provides nationals with the right to renounce their nationality, providing they do not thereby become stateless. States may however restrict this right with respect to nationals residing abroad. As at 6 March 2014, the following countries have signed or ratified the Convention:
https://en.wikipedia.org/wiki?curid=10080
English orthography English orthography is the system of writing conventions used to represent spoken English in written form that allows readers to connect spelling to sound to meaning. Like the orthography of most world languages, English orthography has a broad degree of standardisation. However, unlike with most languages, there are multiple ways to spell nearly every phoneme (sound), and most letters also have multiple pronunciations depending on their position in a word and the context. Several orthographic mistakes are common even among native speakers. This is mainly due to the large number of words that have been borrowed from a large number of other languages throughout the history of the English language without successful attempts at complete spelling reforms. Most of the spelling conventions in Modern English were derived from the phonetic spelling of a variety of Middle English, and generally do not reflect the sound changes that have occurred since the late 15th century (such as the Great Vowel Shift). Despite the various English dialects spoken from country to country and within different regions of the same country, there are only slight regional variations in English orthography, the two most recognised variations being British and American spelling, and its overall uniformity helps facilitate international communication. On the other hand, it also adds to the discrepancy between the way English is written and spoken in any given location. Letters in English orthography usually represent a particular sound (phoneme). For example, the word "cat" consists of three letters , , and , in which represents the sound , the sound , and the sound . Sequences of letters may perform this role as well as single letters. Thus, in the word "ship" (pronounced ), the digraph (two letters) represents the sound . In the word "ditch", the trigraph represents the sound . Less commonly, a single letter can represent multiple successive sounds. The most common example is the letter , which normally represents the consonant cluster (for example, in the word "six", pronounced ). The same letter (or sequence of letters) may be pronounced in different ways when it occurs in different positions within a word. For instance, the digraph represents the sound at the end of some words, such as "rough" , though not in others ("though" ). At the beginning of syllables (i.e. the syllable onset), the digraph is pronounced , as in the word "ghost" (pronounced ). Conversely, the digraph is never pronounced in syllable onsets other than in inflected forms, and is almost never pronounced in syllable codas (the proper name "Pittsburgh" is an exception). Some words contain silent letters, which do not represent any sound in modern English pronunciation. Examples include the in "doubt", "debt", "dumb", etc., the in "psychology" and "pneumatic", as mentioned above in numerous words such as "though", "daughter", "night", "brought", and the commonly encountered silent (discussed further below). Another type of spelling characteristic is related to word origin. For example, when representing a vowel, the letter represents the sound in some words borrowed from Greek (reflecting an original upsilon), whereas the letter usually representing this sound in non-Greek words is the letter . Thus, the word "myth" is of Greek origin, while "pith" is a Germanic word. Other examples include pronounced (which is usually spelled ), and pronounced (which is usually spelled or ) – the use of these spellings for these sounds often mark words that have been borrowed from Greek. Some researchers, such as Brengelman (1970), have suggested that, in addition to this marking of word origin, these spellings indicate a more formal level of style or register in a given text, although Rollings (2004) finds this point to be exaggerated as there would be many exceptions where a word with one of these spellings, such as for (like "telephone"), could occur in an informal text. Spelling may also be useful to distinguish between homophones (words with the same pronunciation but different meanings), although in most cases the reason for the difference is historical and was not introduced for the purpose of making a distinction. For example, the words "heir" and "air" are pronounced identically in most dialects, but in writing they are distinguished from each other by their different spellings. Another example is the pair of homophones "pain" and "pane", where both are pronounced but have two different spellings of the vowel . Often this is because of the historical pronunciation of each word where, over time, two separate sounds become the same but the different spellings remain: "pain" used to be pronounced as , with a diphthong, and "pane" as , but the diphthong merged with the long vowel in "pane", making "pain" and "pane" homophones ("pane"–"pain" merger). Later became a diphthong . In written language, this may help to resolve potential ambiguities that would arise otherwise (cf. "He's breaking the car" vs. "He's braking the car"). Nevertheless, many homophones remain that are unresolved by spelling (for example, the word "bay" has at least five fundamentally different meanings). Some letters in English provide information about the pronunciation of "other" letters in the word. Rollings (2004) uses the term "markers" for such letters. Letters may mark different types of information. For instance, the letter in the word "cottage" indicates that the preceding is pronounced , rather than the more common value of in word-final position as the sound , such as in "tag" . The letter also often marks an altered pronunciation of a preceding vowel. In the pair "ban" and "bane", the of "ban" has the value , whereas the of "bane" is marked by the as having the value . In this context, the is not pronounced, and is referred to as "silent e". A single letter may even fill multiple pronunciation-marking roles simultaneously. For example, in the word "wage", the marks not only the change of the from to , but also of the from to . In the word "vague", the marks the long a sound, but the keeps the g hard rather than soft. Doubled consonants usually indicate that the preceding vowel is pronounced short. For example, the doubled in "latter" indicates that the is pronounced , while the single of "later" gives . Doubled consonants only indicate any lengthening or gemination of the consonant sound itself when they come from different morphemes, as with the in "unnatural" = "un+natural". A given letter (or letters) may have dual functions. For example, the letter in the word "cinema" has a sound-representing function (representing the sound ) and a pronunciation-marking function (marking the as having the value opposed to the value ). Like many other alphabetic orthographies, English spelling does not represent non-contrastive phonetic sounds (that is, minor differences in pronunciation which are not used to distinguish between different words). Although the letter is pronounced by some speakers with aspiration at the beginning of words, this is never indicated in the spelling, and, indeed, this phonetic detail is probably not noticeable to the average native speaker not trained in phonetics. However, unlike some orthographies, English orthography often represents a very abstract underlying representation (or morphophonemic form) of English words. In these cases, a given morpheme (i.e. a component of a word) has a fixed spelling even though it is pronounced differently in different words. An example is the past tense suffix -, which may be pronounced variously as , , or (for example, "dip" , "dipped" , "boom" , "boomed" , "loot" , "looted" ). As it happens, these different pronunciations of - can be predicted by a few phonological rules, but that is not the reason why its spelling is fixed. Another example involves the vowel differences (with accompanying stress pattern changes) in several related words. For instance, the word "photographer" is derived from the word "photograph" by adding the derivational suffix -. When this suffix is added, the vowel pronunciations change largely owing to the moveable stress: Other examples of this type are the - suffix (as in "agile" vs "agility", "acid" vs "acidity", "divine" vs "divinity", "sane" vs "sanity"). See also: Trisyllabic laxing. Another such class of words includes "sign" and "bomb" with "silent" letters and , respectively. However, in the related words "signature" and "bombard" these letters are pronounced and , respectively. Here it could be argued that the underlying representation of "sign" and "bomb" is || and ||, in which the underlying || and || are only pronounced in the surface forms when followed by certain suffixes (-, -). Otherwise, the || and || are not realised in the surface pronunciation (e.g. when standing alone, or when followed by suffixes like - or -). In these cases, the orthography indicates the underlying consonants that are present in certain words but are absent in other related words. Other examples include the in "fast" and "fasten" , and the in "heir" and "inherit" . Another example includes words like "mean" and "meant" . Here the vowel spelling is pronounced differently in the two related words. Thus, again the orthography uses only a single spelling that corresponds to the single morphemic form rather than to the surface phonological form. English orthography does not always provide an underlying representation; sometimes it provides an intermediate representation between the underlying form and the surface pronunciation. This is the case with the spelling of the regular plural morpheme, which is written as either - (as in "tick, ticks" and "mite, mites") or - (as in "box, boxes"). Here the spelling - is pronounced either or (depending on the environment, e.g. "ticks" and "pigs" ) while - is usually pronounced (e.g. "boxes" ). Thus, there are two different spellings that correspond to the single underlying representation || of the plural suffix and the three surface forms. The spelling indicates the insertion of before the in the spelling -, but does not indicate the devoiced distinctly from the unaffected in the spelling -. The abstract representation of words as indicated by the orthography can be considered advantageous since it makes etymological relationships more apparent to English readers. This makes writing English more complex, but arguably makes reading English more efficient. However, very abstract underlying representations, such as that of Chomsky & Halle (1968) or of underspecification theories, are sometimes considered too abstract to accurately reflect the communicative competence of native speakers. Followers of these arguments believe the less abstract surface forms are more "psychologically real" and thus more useful in terms of pedagogy. English has some words that can be written with accent marks. These words have mostly been imported from other languages, usually French. As imported words become increasingly naturalised, there is an increasing tendency to omit the accent marks, even in formal writing. For example, words such as "rôle" and "hôtel" were first seen with accents when they were borrowed into English, but now the accent is almost never used. The words were originally considered foreign – and some people considered that English alternatives were preferable – but today their foreign origin is largely forgotten. Words most likely to retain the accent are those atypical of English morphology and therefore still perceived as slightly foreign. For example, "café" and "pâté" both have a pronounced final "e", which would otherwise be silent under the normal English pronunciation rules. However "café" is now sometimes facetiously pronounced "caff", while in "pâté", the acute accent is helpful to distinguish it from "pate". Further examples of words sometimes retaining diacritics when used in English are: Ångström (partly because the scientific symbol for this unit of measurement is "Å"), "appliqué", "attaché", "blasé", "bric-à-brac", "Brötchen", "cliché", "crème", "crêpe", "façade", "fiancé(e)", "flambé", "naïve", "naïveté", "né(e)", "papier-mâché", "passé", "piñata", "protégé", "résumé", "risqué", "über-", "voilà". Italics, with appropriate accents, are generally applied to foreign terms that are uncommonly used in or have not been assimilated into English: for example, "adiós, crème brûlée, pièce de résistance, raison d'être, über, vis-à-vis and belles-lettres." It was formerly common in American English to use a diaeresis mark to indicate a hiatus: for example, "coöperate", "daïs", "reëlect". "The New Yorker" and "Technology Review" magazines still use it for this purpose, even though it is increasingly rare in modern English. Nowadays the diaeresis is normally left out ("cooperate"), or a hyphen is used ("co-operate") if the hiatus is between two morphemes in a compound word. It is, however, still common in monomorphemic loanwords such as "naïve" and "Noël". Written accents are also used occasionally in poetry and scripts for dramatic performances to indicate that a certain normally unstressed syllable in a word should be stressed for dramatic effect, or to keep with the metre of the poetry. This use is frequently seen in archaic and pseudoarchaic writings with the "-ed" suffix, to indicate that the "e" should be fully pronounced, as with "cursèd". The acute and grave accents are occasionally used in poetry and lyrics: the acute to indicate stress overtly where it might be ambiguous (rébel vs. rebél) or nonstandard for metrical reasons (caléndar); the grave to indicate that an ordinarily silent or elided syllable is pronounced (warnèd, parlìament). In certain older texts (typically British), the use of the ligatures æ and œ is common in words such as "archæology", "diarrhœa", and "encyclopædia". Such words have Latin or Greek origin. Nowadays, the ligatures have been generally replaced in British English by the separated digraph "ae" and "oe" ("encyclopaedia", "diarrhoea"); but usually "economy", "ecology, " and in American English by "e" ("encyclopedia", "diarrhea"; but usually "paean", "amoeba", "oedipal", "Caesar"). In some cases, usage may vary; for instance, both "encyclopedia" and "encyclopaedia" are current in the UK. Partly because English has never had any official regulating authority for spelling, such as the Spanish "Real Academia Española", the French "Académie française", and the German "Rat für deutsche Rechtschreibung", English spelling, compared to many other languages, is quite irregular and complex. Although French, among other languages, presents a similar degree of difficulty when "encoding" (writing), English is more difficult when "decoding" (reading), as there are clearly many more possible pronunciations of a group of letters. For example, in French, the sound (as in "food", but short), can be spelled "ou", "ous", "out", or "oux" ("ou", "nous", "tout", "choux"), but the pronunciation of each of those sequences is always the same. In English, the sound can be spelled in up to 18 different ways (see the Sound-to-spelling correspondences section below), including "oo", "u", "ui", "ue", "o", "oe", "ou", "ough", and "ew" ("food", "truth", "fruit", "blues", "to", "shoe", "group", "through", "grew"), but all of these have other pronunciations as well (e.g. as in "flood", "trust", "build", "bluest", "go", "hoe", "grout", "rough", "sew"). The Spelling-to-sound correspondences section below presents a summary of pronunciation variations. Thus, in unfamiliar words and proper nouns the pronunciation of some sequences, "ough" being the prime example, is unpredictable to even educated native English speakers. Attempts to regularise or reform the spelling of English have usually failed. However, Noah Webster popularised more phonetic spellings in the United States, such as "flavor" for British "flavour", "fiber" for "fibre", "defense" for "defence", "analyze" for "analyse", "catalog" for "catalogue" and so forth. These spellings already existed as alternatives, but Webster's dictionaries helped make them standard in the US. See American and British English spelling differences for details. Besides the quirks the English spelling system has inherited from its past, there are other idiosyncrasies in spelling that make it tricky to learn. English contains, depending on dialect, 24–27 separate consonant phonemes and 13–20 vowels. However, there are only 26 letters in the modern English alphabet, so there is not a one-to-one correspondence between letters and sounds. Many sounds are spelled using different letters or multiple letters, and for those words whose pronunciation is predictable from the spelling, the sounds denoted by the letters depend on the surrounding letters. For example, the digraph "th" represents two different sounds (the voiced dental fricative and the voiceless dental fricative) (see Pronunciation of English "th"), and the voiceless alveolar sibilant can be represented by the letters "s" and "c". It is, however, not the shortage of letters which makes English spelling irregular. Its irregularities are caused mainly by the use of many different spellings for some of its sounds, such as the sounds /uː/, /iː/ and /oʊ/ ("too, "true, "shoe, "flew, "through; "sleeve", "leave", even", "seize", "siege"; "stole, "coal", "bowl", "roll", old", "mould"), and the use of identical sequences for spelling different sounds (over", oven", "move"). Furthermore, English no longer makes any attempt to anglicise the spellings of loanwords, but preserves the foreign spellings, even when they employ exotic conventions like the Polish "cz" in "Czech" (rather than "*Check") or the Norwegian "fj" in "fjord" (although "fiord" was formerly the most common spelling). In early Middle English, until roughly 1400, most imports from French were respelled according to English rules (e.g. "bataille"–"battle", "bouton"–"button", but not "double", or "trouble"). Instead of loans being respelled to conform to English spelling standards, sometimes the pronunciation changes as a result of pressure from the spelling. One example of this is the word "ski", which was adopted from Norwegian in the mid-18th century, although it did not become common until 1900. It used to be pronounced , which is similar to the Norwegian pronunciation, but the increasing popularity of the sport after the middle of the 20th century helped the pronunciation replace it. There was also a period when the spelling of a small number of words was altered in what is now regarded as a misguided attempt to make them conform to what were perceived to be the etymological origins of the words. For example, the letter "b" was added to "debt" (originally "dette") in an attempt to link it to the Latin "debitum", and the letter "s" in "island" is a misplaced attempt to link it to Latin "insula" instead of the Old English word "īġland", which is the true origin of the English word. The letter "p" in "ptarmigan" has no etymological justification whatsoever, only seeking to invoke Greek despite being a Gaelic word. The spelling of English continues to evolve. Many loanwords come from languages where the pronunciation of vowels corresponds to the way they were pronounced in Old English, which is similar to the Italian or Spanish pronunciation of the vowels, and is the value the vowel symbols and have in the International Phonetic Alphabet. As a result, there is a somewhat regular system of pronouncing "foreign" words in English, and some borrowed words have had their spelling changed to conform to this system. For example, "Hindu" used to be spelled "Hindoo", and the name "Maria" used to be pronounced like the name "Mariah", but was changed to conform to this system. Commercial advertisers have also had an effect on English spelling. They introduced new or simplified spellings like "lite" instead of "light", "thru" instead of "through", "smokey" instead of "smoky" (for "smokey bacon" flavour crisps), and "rucsac" instead of "rucksack". The spellings of personal names have also been a source of spelling innovations: diminutive versions of women's names that sound the same as men's names have been spelled differently: "Nikki" and "Nicky", "Toni" and "Tony", "Jo" and "Joe". The differentiation in between names that are spelled differently but have the same phonetic sound may come from modernization or different countries of origin. For example, "Isabelle" and "Isabel" sound the same but are spelled differently; these versions are from France and Spain respectively. As examples of the idiosyncratic nature of English spelling, the combination "ou" can be pronounced in at least nine different ways: in "out", in "soul", in "soup", in "touch", in "could", in "four", in "journal", in "cough", and in "famous". See the section Spelling-to-sound correspondences for a comprehensive treatment. In the other direction, the vowel sound in "me" can be spelled in at least 18 or 21 different ways: "be ("cede), "ski ("machine), "bologna (GA), "algae, "quay, "beach", "bee, "deceit", "people", "key, "volleyed", "field" ("hygiene), "amoeba", "chamois", "dengue, "beguine", "guyot", and "city. See the section Sound-to-spelling correspondences below. (These examples assume a more-or-less standard non-regional British English accent. Other accents will vary.) Sometimes everyday speakers of English change a counterintuitive pronunciation simply because it is counterintuitive. Changes like this are not usually seen as "standard", but can become standard if used enough. An example is the word "miniscule", which still competes with its original spelling of "minuscule", though this might also be because of analogy with the word "mini". Inconsistencies and irregularities in English pronunciation and spelling have gradually increased in number throughout the history of the English language. There are a number of contributing factors. First, gradual changes in pronunciation, such as the Great Vowel Shift, account for a tremendous number of irregularities. Second, relatively recent loan words from other languages generally carry their original spellings, which are often not phonetic in English. The Romanization of languages (e.g., Chinese) using alphabets derived from the Latin alphabet has further complicated this problem, for example when pronouncing Chinese proper names (of people or places). The regular spelling system of Old English was swept away by the Norman Conquest, and English itself was supplanted in some spheres by Norman French for three centuries, eventually emerging with its spelling much influenced by French. English had also borrowed large numbers of words from French, which naturally kept their French spellings as there was no reason or mechanism to change them. The spelling of Middle English, such as in the writings of Geoffrey Chaucer, is very irregular and inconsistent, with the same word being spelled in different ways, sometimes even in the same sentence. However, these were generally much better guides to the then pronunciation than modern English spelling is. For example, the sound , normally written "u", is spelled with an "o" in "son", "love", "come", etc., due to Norman spelling conventions which prohibited writing "u" before "v", "m", "n" due to the graphical confusion that would result. ("v", "u", "n" were identically written with two minims in Norman handwriting; "w" was written as two "u" letters; "m" was written with three minims, hence "mm" looked like "vun", "nvu", "uvu", etc.). Similarly, spelling conventions also prohibited final "v". Hence the identical spellings of the three different vowel sounds in "love", "grove" and "prove" are due to ambiguity in the Middle English spelling system, not sound change. In 1417 Henry V began using English for official correspondence, which had no standardised spelling, instead of Latin or French which had standardised spelling. For example, for the word "right", Latin had one spelling", rectus;" Old French as used in English law had six spellings; Middle English had 77 spellings. English, now used as the official replacement language for Latin and French, motivated writers to standardise spellings, an effort which lasted about 500 years. There was also a series of linguistic sound changes towards the end of this period, including the Great Vowel Shift, which resulted in the "i" in "mine", for example, changing from a pure vowel to a diphthong. These changes for the most part did not detract from the rule-governed nature of the spelling system; but in some cases they introduced confusing inconsistencies, like the well-known example of the many pronunciations of "ough" ("rough", "through", "though", "trough", "plough", etc.). Most of these changes happened before the arrival of printing in England. However, the arrival of the printing press froze the current system, rather than providing the impetus for a realignment of spelling with pronunciation. Furthermore, it introduced further inconsistencies, partly because of the use of typesetters trained abroad, particularly in the Low Countries. For example, the "h" in "ghost" was influenced by Dutch. The addition and deletion of a silent "e" at the ends of words was also sometimes used to make the right-hand margin line up more neatly. By the time dictionaries were introduced in the mid 17th century, the spelling system of English had started to stabilise. By the 19th century, most words had set spellings, though it took some time before they diffused throughout the English-speaking world. In "The Mill on the Floss" (1860), English novelist George Eliot satirised the attitude of the English rural gentry of the 1820s towards orthography: The modern English spelling system, with its national variants, spread together with the expansion of public education later in the 19th century. The most notorious group of letters in the English language, the "ough" tetragraph, can be pronounced in at least ten different ways, six of which are illustrated in the construct, "Though the tough cough and hiccough plough him through", which is quoted by Robert A. Heinlein in "The Door into Summer" to illustrate the difficulties facing automated speech transcription and reading. The "ough" tetragraph, usually representing a pronunciation of roughly , is in fact a word in its own right, though rarely known or used: an exclamation of disgust similar to "ugh". The following are recorded throughout English languages of the world: The following pronunciations are found in uncommon single words: The place name Loughborough uses two different pronunciations of "ough": the first "ough" has the sound as in "cuff" and the second rhymes with "thorough". In a generative approach to English spelling, Rollings (2004) identifies twenty main orthographic vowels of stressed syllables that are grouped into four main categories: "Lax", "Tense", "Heavy", "Tense-R". (As this classification is based on orthography, not all orthographic "lax" vowels are necessarily phonologically lax.) For instance, the letter "a" can represent the lax vowel , tense , heavy , or (often allophonically) before |r|. Heavy and tense-r vowels are the respective lax and tense counterparts followed by the letter "r". Tense vowels are distinguished from lax vowels with a "silent" "e" letter that is added at the end of words. Thus, the letter "a" in "hat" is lax , but when the letter "e" is added in the word "hate" the letter "a" is tense . Similarly, heavy and tense-r vowels pattern together: the letters "ar" in "car" are heavy , the letters "ar" followed by silent "e" in the word "care" are . The letter "u" represents two different vowel patterns, one being , the other . There is no distinction between heavy and tense-r vowels with the letter "o", and the letter "u" in the pattern does not have a heavy vowel member. Besides silent "e", another strategy for indicating tense and tense-r vowels, is the addition of another orthographic vowel forming a digraph. In this case, the first vowel is usually the main vowel while the second vowel is the "marking" vowel. For example, the word "man" has a lax "a" pronounced , but with the addition of "i" (as the digraph "ai") in the word "main" the "a" is marked as tense and pronounced . These two strategies produce words that are spelled differently but pronounced identically, as in "mane" (silent "e" strategy), "main" (digraph strategy) and "Maine" (both strategies). The use of two different strategies relates to the function of distinguishing between words that would otherwise be homonyms. Besides the 20 basic vowel spellings, Rollings (2004) has a reduced vowel category (representing the sounds ) and a miscellaneous category (representing the sounds and +V, +V, V+V). To reduce dialectal difficulties, the sound values given here correspond to the conventions at . This table includes H, W and Y when they represent vowel sounds. If no information is given, it is assumed that the vowel is in a stressed syllable. Deriving the pronunciation of an English word from its spelling requires not only a careful knowledge of the rules given below (many of which are not explicitly known even by native speakers: speakers merely learn the spelling of a word along with its pronunciation) and their many exceptions, but also: Notes: † Nearly 80% of Americans pronounce "luxurious" with , while two thirds of British people use . Half the American speakers pronounce "luxury" as , the rest says †† About half of both British and American speakers say , the other half says . * According to the Longman Pronunciation Dictionary, 75% of Americans pronounce "almond" as . † Where GA distinguishes between and in the letter combination ong, RP only has the vowel The following table shows for each sound the various spelling patterns used to denote it, starting with the prototypical pattern(s) followed by others in alphabetical order. Some of these patterns are very rare or unique (such as "gh" for , "ph" for , "i" for ). An ellipsis ("...") stands for an intervening consonant. Arranged in the order of the IPA consonant tables. * In 2008, 61% of British people pronounced "diphthong" as , though phoneticians prefer . ** In 2008, 20% of Americans pronounced "thespian" as . *** The majority of British people, and the great majority of younger ones, pronounce "crescent" as . † In 2008, 64% of Americans and 39% of British people pronounce "February" as . †† The majority of Americans, and the great majority of younger ones, pronounce "congratulate" as . Sorted more or less from close to open sounds in the vowel diagram. † Identical to previous vowel in non-rhotic dialects like RP.
https://en.wikipedia.org/wiki?curid=10081
Æthelred the Unready Æthelred II (Old English: "Æþelræd", ;  966 – 23 April 1016), known as the Unready, was King of the English from 978 to 1013 and again from 1014 until his death. His epithet does not derive from the modern word "unready", but rather from the Old English "unræd" meaning "poorly advised"; it is a pun on his name, which means "well advised". Æthelred was the son of King Edgar and Queen Ælfthryth. He came to the throne at about the age of 12, following the assassination of his older half-brother, Edward the Martyr. His brother's murder was carried out by supporters of his own claim to the throne, although he was too young to have any personal involvement. The chief problem of Æthelred's reign was conflict with the Danes. After several decades of relative peace, Danish raids on English territory began again in earnest in the 980s. Following the Battle of Maldon in 991, Æthelred paid tribute, or Danegeld, to the Danish king. In 1002, Æthelred ordered what became known as the St. Brice's Day massacre of Danish settlers. In 1013, King Sweyn Forkbeard of Denmark invaded England, as a result of which Æthelred fled to Normandy in 1013 and was replaced by Sweyn. However, he returned as king for two years after Sweyn's death in 1014. Æthelred's 37-year combined reign was the longest of any Anglo-Saxon king of England, and was only surpassed in the 13th century, by Henry III. Æthelred was briefly succeeded by his son, Edmund Ironside, but he died after a few months and was replaced by Sweyn's son, Cnut. Another of Æthelred's sons, Edward the Confessor, became king in 1042. Æthelred's first name, composed of the elements "æðele", "noble", and "ræd", "counsel, advice", is typical of the compound names of those who belonged to the royal House of Wessex, and it characteristically alliterates with the names of his ancestors, like Æthelwulf ("noble-wolf"), Ælfred ("elf-counsel"), Eadweard ("rich-protection"), and Eadgar ("rich-spear"). Æthelred's notorious nickname, Old English "Unræd", is commonly translated into present-day English as "The Unready" (less often, though less inaccurately, as "The Redeless"). The Anglo-Saxon noun "unræd" means "evil counsel", "bad plan", or "folly". It was most often used in reference to decisions and deeds, but once in reference to the ill-advised disobedience of Adam and Eve. The element "ræd" in "unræd" is the same element in Æthelred's name that means "counsel". Thus "Æþelræd Unræd" is an oxymoron: "Noble counsel, No counsel". The nickname has also been translated as "ill-advised", "ill-prepared", thus "Æthelred the ill-advised". Because the nickname was first recorded in the 1180s, more than 150 years after Æthelred's death, it is doubtful that it carries any implications as to the reputation of the king in the eyes of his contemporaries or near contemporaries. Sir Frank Stenton remarked that "much that has brought condemnation of historians on King Æthelred may well be due in the last resort to the circumstances under which he became king." Æthelred's father, King Edgar, had died suddenly in July 975, leaving two young sons behind. The elder, Edward (later Edward the Martyr), was probably illegitimate, and was "still a youth on the verge of manhood" in 975. The younger son was Æthelred, whose mother, Ælfthryth, Edgar had married in 964. Ælfthryth was the daughter of Ordgar, ealdorman of Devon, and widow of Æthelwold, Ealdorman of East Anglia. At the time of his father's death, Æthelred could have been no more than 10 years old. As the elder of Edgar's sons, Edward – reportedly a young man given to frequent violent outbursts – probably would have naturally succeeded to the throne of England despite his young age, had not he "offended many important persons by his intolerable violence of speech and behaviour." In any case, a number of English nobles took to opposing Edward's succession and to defending Æthelred's claim to the throne; Æthelred was, after all, the son of Edgar's last, living wife, and no rumour of illegitimacy is known to have plagued Æthelred's birth, as it might have his elder brother's. Both boys, Æthelred certainly, were too young to have played any significant part in the political manoeuvring which followed Edgar's death. It was the brothers' supporters, and not the brothers themselves, who were responsible for the turmoil which accompanied the choice of a successor to the throne. Æthelred's cause was led by his mother and included Ælfhere, Ealdorman of Mercia and Bishop Æthelwold of Winchester, while Edward's claim was supported by Dunstan, the Archbishop of Canterbury and Oswald, the Archbishop of York among other noblemen, notably Æthelwine, Ealdorman of East Anglia, and Byrhtnoth, ealdorman of Essex. In the end, Edward's supporters proved the more powerful and persuasive, and he was crowned king at Kingston upon Thames before the year was out. Edward reigned for only three years before he was murdered by members of his brother's household. Though little is known about Edward's short reign, it is known that it was marked by political turmoil. Edgar had made extensive grants of land to monasteries which pursued the new monastic ideals of ecclesiastical reform, but these disrupted aristocratic families' traditional patronage. The end of his firm rule saw a reversal of this policy, with aristocrats recovering their lost properties or seizing new ones. This was opposed by Dunstan, but according to Cyril Hart, "The presence of supporters of church reform on both sides indicates that the conflict between them depended as much on issues of land ownership and local power as on ecclesiastical legitimacy. Adherents of both Edward and Æthelred can be seen appropriating, or recovering, monastic lands." Nevertheless, favour for Edward must have been strong among the monastic communities. When Edward was killed at Æthelred's estate at Corfe Castle in Dorset in March 978, the job of recording the event, as well as reactions to it, fell to monastic writers. Stenton offers a summary of the earliest account of Edward's murder, which comes from a work praising the life of St Oswald: Nevertheless, at first, the outlook of the new king's officers and counsellors seems in no way to have been bleak. According to one chronicler, the coronation of Æthelred took place with much rejoicing by the councillors of the English people. Simon Keynes notes that "Byrhtferth of Ramsey states similarly that when Æthelred was consecrated king, by Archbishop Dunstan and Archbishop Oswald, 'there was great joy at his consecration', and describes the king in this connection as 'a young man in respect of years, elegant in his manners, with an attractive face and handsome appearance'." Æthelred could not have been older than 13 years of age in this year. Contemporary documents, including a number of charters, strongly suggest that Queen Ælfthryth acted as regent – along with advisors – before her son reached majority (which in Anglo-Saxon England was aged fifteen). Æthelred and Ælfthryth had already developed a close relationship to Æthelwold, bishop of Winchester, one who had supported his unsuccessful claim to the throne. When Æthelwold died, on 1 August 984, Æthelred deeply lamented the loss, and he wrote later in a charter from 993 that the event had deprived the country of one "whose industry and pastoral care administered not only to my interest but also to that of all inhabitants of the country." After the death of the bishop, Æthelred, now well into adulthood, took control of his kingdom from his mother for the first time. In the same year that the bishop died, she left the king's court and took up residence at a far off estate. Afterward, at least through the next decade, she rarely visited. Later she returned to raise his eldest child, but she never had the involvement in the internal affairs of the kingdom that she once did. Æthelred's changes in policy reflected a severance, at least politically, of his relationship to his mother. England had experienced a period of peace after the reconquest of the Danelaw in the mid-10th century by King Edgar, Æthelred's father. However, beginning in 980, when Æthelred could not have been more than 14 years old, small companies of Danish adventurers carried out a series of coastline raids against England. Hampshire, Thanet and Cheshire were attacked in 980, Devon and Cornwall in 981, and Dorset in 982. A period of six years then passed before, in 988, another coastal attack is recorded as having taken place to the south-west, though here a famous battle was fought between the invaders and the thegns of Devon. Stenton notes that, though this series of isolated raids had no lasting effect on England itself, "their chief historical importance is that they brought England for the first time into diplomatic contact with Normandy." During this period, the Normans offered shelter to Danes returning from raids on England. This led to tension between the English and Norman courts, and word of their enmity eventually reached Pope John XV. The pope was disposed to dissolve their hostility towards each other, and took steps to engineer a peace between England and Normandy, which was ratified in Rouen in 991. In August 991, a sizeable Danish fleet began a sustained campaign in the south-east of England. It arrived off Folkestone, in Kent, and made its way around the south-east coast and up the River Blackwater, coming eventually to its estuary and occupying Northey Island. About west of Northey lies the coastal town of Maldon, where Byrhtnoth, ealdorman of Essex, was stationed with a company of thegns. The battle that followed between English and Danes is immortalised by the Old English poem "The Battle of Maldon", which describes the doomed but heroic attempt of Byrhtnoth to defend the coast of Essex against overwhelming odds. This was the first of a series of crushing defeats felt by the English: beaten first by Danish raiders, and later by organised Danish armies. Stenton summarises the events of the poem: In the aftermath of Maldon, it was decided that the English should grant the tribute to the Danes that they desired, and so a "gafol" of £10,000 was paid them for their peace. Yet it was presumably the Danish fleet that had beaten Byrhtnoth at Maldon that continued to ravage the English coast from 991 to 993. In 994, the Danish fleet, which had swollen in ranks since 991, turned up the Thames estuary and headed toward London. The battle fought there was inconclusive. It was about this time that Æthelred met with the leaders of the Danish fleet and arranged an uneasy accord. A treaty was signed that provided for seemingly civilised arrangements between the then-settled Danish companies and the English government, such as regulation settlement disputes and trade. But the treaty also stipulated that the ravaging and slaughter of the previous year would be forgotten, and ended abruptly by stating that £22,000 of gold and silver had been paid to the raiders as the price of peace. In 994, Olaf Tryggvason, a Danish prince and already a baptised Christian, was confirmed as Christian in a ceremony at Andover; King Æthelred stood as his sponsor. After receiving gifts, Olaf promised "that he would never come back to England in hostility." Olaf then left England for Norway and never returned, though "other component parts of the Viking force appear to have decided to stay in England, for it is apparent from the treaty that some had chosen to enter into King Æthelred's service as mercenaries, based presumably on the Isle of Wight." In 997, Danish raids began again. According to Keynes, "there is no suggestion that this was a new fleet or army, and presumably the mercenary force created in 994 from the residue of the raiding army of 991 had turned on those whom it had been hired to protect." It harried Cornwall, Devon, western Somerset and south Wales in 997, Dorset, Hampshire and Sussex in 998. In 999, it raided Kent, and, in 1000, it left England for Normandy, perhaps because the English had refused in this latest wave of attacks to acquiesce to the Danish demands for "gafol" or tribute, which would come to be known as Danegeld, 'Dane-payment'. This sudden relief from attack Æthelred used to gather his thoughts, resources, and armies: the fleet's departure in 1000 "allowed Æthelred to carry out a devastation of Strathclyde, the motive for which is part of the lost history of the north." In 1001, a Danish fleet – perhaps the same fleet from 1000 – returned and ravaged west Sussex. During its movements, the fleet regularly returned to its base in the Isle of Wight. There was later an attempted attack in the south of Devon, though the English mounted a successful defence at Exeter. Nevertheless, Æthelred must have felt at a loss, and, in the Spring of 1002, the English bought a truce for £24,000. Æthelred's frequent payments of immense Danegelds are often held up as exemplary of the incompetency of his government and his own short-sightedness. However, Keynes points out that such payments had been practice for at least a century, and had been adopted by Alfred the Great, Charles the Bald and many others. Indeed, in some cases it "may have seemed the best available way of protecting the people against loss of life, shelter, livestock and crops. Though undeniably burdensome, it constituted a measure for which the king could rely on widespread support." Æthelred ordered the massacre of all Danish men in England to take place on 13 November 1002, St Brice's Day. No order of this kind could be carried out in more than a third of England, where the Danes were too strong, but Gunhilde, sister of Sweyn Forkbeard, King of Denmark, was said to have been among the victims. It is likely that a wish to avenge her was a principal motive for Sweyn's invasion of western England the following year. By 1004 Sweyn was in East Anglia, where he sacked Norwich. In this year, a nobleman of East Anglia, Ulfcytel Snillingr met Sweyn in force, and made an impression on the until-then rampant Danish expedition. Though Ulfcytel was eventually defeated, outside Thetford, he caused the Danes heavy losses and was nearly able to destroy their ships. The Danish army left England for Denmark in 1005, perhaps because of the losses they sustained in East Anglia, perhaps from the very severe famine which afflicted the continent and the British Isles in that year. An expedition the following year was bought off in early 1007 by tribute money of £36,000, and for the next two years England was free from attack. In 1008, the government created a new fleet of warships, organised on a national scale, but this was weakened when one of its commanders took to piracy, and the king and his council decided not to risk it in a general action. In Stenton's view: "The history of England in the next generation was really determined between 1009 and 1012...the ignominious collapse of the English defence caused a loss of morale which was irreparable." The Danish army of 1009, led by Thorkell the Tall and his brother Hemming, was the most formidable force to invade England since Æthelred became king. It harried England until it was bought off by £48,000 in April 1012. Sweyn then launched an invasion in 1013 intending to crown himself king of England, during which he proved himself to be a general greater than any other Viking leader of his generation. By the end of 1013 English resistance had collapsed and Sweyn had conquered the country, forcing Æthelred into exile in Normandy. But the situation changed suddenly when Sweyn died on 3 February 1014. The crews of the Danish ships in the Trent that had supported Sweyn immediately swore their allegiance to Sweyn's son Cnut the Great, but leading English noblemen sent a deputation to Æthelred to negotiate his restoration to the throne. He was required to declare his loyalty to them, to bring in reforms regarding everything that they disliked and to forgive all that had been said and done against him in his previous reign. The terms of this agreement are of great constitutional interest in early English History as they are the first recorded pact between a King and his subjects and are also widely regarded as showing that many English noblemen had submitted to Sweyn simply because of their distrust of Æthelred. According to the "Anglo-Saxon Chronicle": Æthelred then launched an expedition against Cnut and his allies. It was only the people of the Kingdom of Lindsey (modern North Lincolnshire) who supported Cnut. Æthelred first set out to recapture London apparently with the help of the Norwegian Olaf Haraldsson. According to the Icelandic historian, Snorri Sturluson, Ólaf led a successful attack on London bridge with a fleet of ships. He then went on to help Æthelred retake London and other parts of the country. Cnut and his army decided to withdraw from England, in April 1014, leaving his Lindsey allies to suffer Æthelred's revenge. In about 1016 it is thought that Ólaf left to concentrate on raiding western Europe. In the same year, Cnut returned to find a complex and volatile situation unfolding in England. Æthelred's son, Edmund Ironside, had revolted against his father and established himself in the Danelaw, which was angry at Cnut and Æthelred for the ravaging of Lindsey and was prepared to support Edmund in any uprising against both of them. Over the next few months Cnut conquered most of England, while Edmund rejoined Æthelred to defend London when Æthelred died on 23 April 1016. The subsequent war between Edmund and Cnut ended in a decisive victory for Cnut at the Battle of Ashingdon on 18 October 1016. Edmund's reputation as a warrior was such that Cnut nevertheless agreed to divide England, Edmund taking Wessex and Cnut the whole of the country beyond the Thames. However, Edmund died on 30 November and Cnut became king of the whole country. Æthelred was buried in Old St Paul's Cathedral, London. The tomb and his monument were destroyed along with the cathedral in the Great Fire of London in 1666. A modern monument in the crypt lists his among the important graves lost. Æthelred's government produced extensive legislation, which he "ruthlessly enforced". Records of at least six legal codes survive from his reign, covering a range of topics. Notably, one of the members of his council (known as the "Witan") was Wulfstan II, Archbishop of York, a well-known homilist. The three latest codes from Æthelred's reign seemed to have been drafted by Wulfstan. These codes are extensively concerned with ecclesiastical affairs. They also exhibit the characteristics of Wulfstan's highly rhetorical style. Wulfstan went on to draft codes for King Cnut, and recycled there many of the laws which were used in Æthelred's codes. Despite the failure of his government in the face of the Danish threat, Æthelred's reign was not without some important institutional achievements. The quality of the coinage, a good indicator of the prevailing economic conditions, significantly improved during his reign due to his numerous coinage reform laws. Later perspectives of Æthelred have been less than flattering. Numerous legends and anecdotes have sprung up to explain his shortcomings, often elaborating abusively on his character and failures. One such anecdote is given by William of Malmesbury (lived  1080 –  1143), who reports that Æthelred had defecated in the baptismal font as a child, which led St Dunstan to prophesy that the English monarchy would be overthrown during his reign. This story is, however, a fabrication, and a similar story is told of the Byzantine Emperor Constantine Copronymus, another mediaeval monarch who was unpopular among certain of his subjects. Efforts to rehabilitate Æthelred's reputation have gained momentum since about 1980. Chief among the rehabilitators has been Simon Keynes, who has often argued that our poor impression of Æthelred is almost entirely based upon after-the-fact accounts of, and later accretions to, the narrative of events during Æthelred's long and complex reign. Chief among the culprits is in fact one of the most important sources for the history of the period, the "Anglo-Saxon Chronicle", which, as it reports events with a retrospect of 15 years, cannot help but interpret events with the eventual English defeat a foregone conclusion. Yet, as virtually no strictly contemporary narrative account of the events of Æthelred's reign exists, historians are forced to rely on what evidence there is. Keynes and others thus draw attention to some of the inevitable snares of investigating the history of a man whom later popular opinion has utterly damned. Recent cautious assessments of Æthelred's reign have more often uncovered reasons to doubt, rather than uphold, Æthelred's later infamy. Though the failures of his government will always put Æthelred's reign in the shadow of the reigns of kings Edgar, Æthelstan, and Alfred, historians' current impression of Æthelred's personal character is certainly not as unflattering as it once was: "Æthelred's misfortune as a ruler was owed not so much to any supposed defects of his imagined character, as to a combination of circumstances which anyone would have found difficult to control." Æthelred has been credited with the formation of a local investigative body made up of twelve thegns who were charged with publishing the names of any notorious or wicked men in their respective districts. Because the members of these bodies were under solemn oath to act in accordance with the law and their own good consciences, they have been seen by some legal historians as the prototype for the English grand jury. Æthelred makes provision for such a body in a law code he enacted at Wantage in 997, which states: But the wording here suggests that Æthelred was perhaps revamping or re-confirming a custom which had already existed. He may actually have been expanding an established English custom for use among the Danish citizens in the North (the Danelaw). Previously, King Edgar had legislated along similar lines in his Whitbordesstan code: The 'legend' of an Anglo-Saxon origin to the jury was first challenged seriously by Heinrich Brunner in 1872, who claimed that evidence of the jury was only seen for the first time during the reign of Henry II, some 200 years after the end of the Anglo-Saxon period, and that the practice had originated with the Franks, who in turn had influenced the Normans, who thence introduced it to England. Since Brunner's thesis, the origin of the English jury has been much disputed. Throughout the 20th century, legal historians disagreed about whether the practice was English in origin, or was introduced, directly or indirectly, from either Scandinavia or Francia. Recently, the legal historians Patrick Wormald and Michael Macnair have reasserted arguments in favour of finding in practices current during the Anglo-Saxon period traces of the Angevin practice of conducting inquests using bodies of sworn, private witnesses. Wormald has gone as far as to present evidence suggesting that the English practice outlined in Æthelred's Wantage code is at least as old as, if not older than, 975, and ultimately traces it back to a Carolingian model (something Brunner had done). However, no scholarly consensus has yet been reached. "[A] youth of graceful manners, handsome countenance and fine person..." as well as "[A] tall, handsome man, elegant in manners, beautiful in countenance and interesting in his deportment." Æthelred married first Ælfgifu, daughter of Thored, earl of Northumbria, in about 985. Their known children are: In 1002 Æthelred married Emma of Normandy, sister of Richard II, Duke of Normandy. Their children were: All of Æthelred's sons were named after predecessors of Æthelred on the throne.
https://en.wikipedia.org/wiki?curid=10083
Edward Elgar Sir Edward William Elgar, 1st Baronet, (; 2 June 1857 – 23 February 1934) was an English composer, many of whose works have entered the British and international classical concert repertoire. Among his best-known compositions are orchestral works including the "Enigma Variations", the "Pomp and Circumstance Marches", concertos for violin and cello, and two symphonies. He also composed choral works, including "The Dream of Gerontius", chamber music and songs. He was appointed Master of the King's Musick in 1924. Although Elgar is often regarded as a typically English composer, most of his musical influences were not from England but from continental Europe. He felt himself to be an outsider, not only musically, but socially. In musical circles dominated by academics, he was a self-taught composer; in Protestant Britain, his Roman Catholicism was regarded with suspicion in some quarters; and in the class-conscious society of Victorian and Edwardian Britain, he was acutely sensitive about his humble origins even after he achieved recognition. He nevertheless married the daughter of a senior British army officer. She inspired him both musically and socially, but he struggled to achieve success until his forties, when after a series of moderately successful works his "Enigma Variations" (1899) became immediately popular in Britain and overseas. He followed the Variations with a choral work, "The Dream of Gerontius" (1900), based on a Roman Catholic text that caused some disquiet in the Anglican establishment in Britain, but it became, and has remained, a core repertory work in Britain and elsewhere. His later full-length religious choral works were well received but have not entered the regular repertory. In his fifties, Elgar composed a symphony and a violin concerto that were immensely successful. His second symphony and his cello concerto did not gain immediate public popularity and took many years to achieve a regular place in the concert repertory of British orchestras. Elgar's music came, in his later years, to be seen as appealing chiefly to British audiences. His stock remained low for a generation after his death. It began to revive significantly in the 1960s, helped by new recordings of his works. Some of his works have, in recent years, been taken up again internationally, but the music continues to be played more in Britain than elsewhere. Elgar has been described as the first composer to take the gramophone seriously. Between 1914 and 1925, he conducted a series of acoustic recordings of his works. The introduction of the moving-coil microphone in 1923 made far more accurate sound reproduction possible, and Elgar made new recordings of most of his major orchestral works and excerpts from "The Dream of Gerontius". Edward Elgar was born in the small village of Lower Broadheath, outside Worcester, England. His father, William Henry Elgar (1821–1906), was raised in Dover and had been apprenticed to a London music publisher. In 1841 William moved to Worcester, where he worked as a piano tuner and set up a shop selling sheet music and musical instruments. In 1848 he married Ann Greening (1822–1902), daughter of a farm worker. Edward was the fourth of their seven children. Ann Elgar had converted to Roman Catholicism shortly before Edward's birth, and he was baptised and brought up as a Roman Catholic, to the disapproval of his father. William Elgar was a violinist of professional standard and held the post of organist of St George's Roman Catholic Church, Worcester, from 1846 to 1885. At his instigation, masses by Cherubini and Hummel were first heard at the Three Choirs Festival by the orchestra in which he played the violin. All the Elgar children received a musical upbringing. By the age of eight, Elgar was taking piano and violin lessons, and his father, who tuned the pianos at many grand houses in Worcestershire, would sometimes take him along, giving him the chance to display his skill to important local figures. Elgar's mother was interested in the arts and encouraged his musical development. He inherited from her a discerning taste for literature and a passionate love of the countryside. His friend and biographer W. H. "Billy" Reed wrote that Elgar's early surroundings had an influence that "permeated all his work and gave to his whole life that subtle but none the less true and sturdy English quality". He began composing at an early age; for a play written and acted by the Elgar children when he was about ten, he wrote music that forty years later he rearranged with only minor changes and orchestrated as the suites titled "The Wand of Youth". Until he was fifteen, Elgar received a general education at Littleton (now Lyttleton) House school, near Worcester. However, his only formal musical training beyond piano and violin lessons from local teachers consisted of more advanced violin studies with Adolf Pollitzer, during brief visits to London in 1877–78. Elgar said, "my first music was learnt in the Cathedral ... from books borrowed from the music library, when I was eight, nine or ten." He worked through manuals of instruction on organ playing and read every book he could find on the theory of music. He later said that he had been most helped by Hubert Parry's articles in the "Grove Dictionary of Music and Musicians". Elgar began to learn German, in the hope of going to the Leipzig Conservatory for further musical studies, but his father could not afford to send him. Years later, a profile in "The Musical Times" considered that his failure to get to Leipzig was fortunate for Elgar's musical development: "Thus the budding composer escaped the dogmatism of the schools." However, it was a disappointment to Elgar that on leaving school in 1872 he went not to Leipzig but to the office of a local solicitor as a clerk. He did not find an office career congenial, and for fulfilment he turned not only to music but to literature, becoming a voracious reader. Around this time, he made his first public appearances as a violinist and organist. After a few months, Elgar left the solicitor to embark on a musical career, giving piano and violin lessons and working occasionally in his father's shop. He was an active member of the Worcester Glee club, along with his father, and he accompanied singers, played the violin, composed and arranged works, and conducted for the first time. Pollitzer believed that, as a violinist, Elgar had the potential to be one of the leading soloists in the country, but Elgar himself, having heard leading virtuosi at London concerts, felt his own violin playing lacked a full enough tone, and he abandoned his ambitions to be a soloist. At twenty-two he took up the post of conductor of the attendants' band at the Worcester and County Lunatic Asylum in Powick, from Worcester. The band consisted of: piccolo, flute, clarinet, two cornets, euphonium, three or four first and a similar number of second violins, occasional viola, cello, double bass and piano. Elgar coached the players and wrote and arranged their music, including quadrilles and polkas, for the unusual combination of instruments. "The Musical Times" wrote, "This practical experience proved to be of the greatest value to the young musician. ... He acquired a practical knowledge of the capabilities of these different instruments. ... He thereby got to know intimately the tone colour, the ins and outs of these and many other instruments." He held the post for five years, from 1879, travelling to Powick once a week. Another post he held in his early days was professor of the violin at the Worcester College for the Blind Sons of Gentlemen. Although rather solitary and introspective by nature, Elgar thrived in Worcester's musical circles. He played in the violins at the Worcester and Birmingham Festivals, and one great experience was to play Dvořák's Symphony No. 6 and "Stabat Mater" under the composer's baton. Elgar regularly played the bassoon in a wind quintet, alongside his brother Frank, an oboist (and conductor who ran his own wind band). Elgar arranged numerous pieces by Mozart, Beethoven, Haydn, and others for the quintet, honing his arranging and compositional skills. In his first trips abroad, Elgar visited Paris in 1880 and Leipzig in 1882. He heard Saint-Saëns play the organ at the Madeleine and attended concerts by first-rate orchestras. In 1882 he wrote, "I got pretty well dosed with Schumann (my ideal!), Brahms, Rubinstein & Wagner, so had no cause to complain." In Leipzig he visited a friend, Helen Weaver, who was a student at the Conservatoire. They became engaged in the summer of 1883, but for unknown reasons the engagement was broken off the next year. Elgar was greatly distressed, and some of his later cryptic dedications of romantic music may have alluded to Helen and his feelings for her. Throughout his life, Elgar was often inspired by close women friends; Helen Weaver was succeeded by Mary Lygon, Dora Penny, Julia Worthington, Alice Stuart Wortley and finally Vera Hockman, who enlivened his old age. In 1882, seeking more professional orchestral experience, Elgar was employed to play violin in Birmingham with William Stockley's Orchestra, for whom he would play every concert for the next seven years and where he later claimed he "learned all the music I know". On 13 December 1883 he took part with Stockley in a performance at Birmingham Town Hall of one of his first works for full orchestra, the "Sérénade mauresque" – the first time one of his compositions had been performed by a professional orchestra. Stockley had invited him to conduct the piece but later recalled "he declined, and, further, insisted upon playing in his place in the orchestra. The consequence was that he had to appear, fiddle in hand, to acknowledge the genuine and hearty applause of the audience." Elgar often went to London in an attempt to get his works published, but this period in his life found him frequently despondent and low on money. He wrote to a friend in April 1884, "My prospects are about as hopeless as ever ... I am not wanting in energy I think, so sometimes I conclude that 'tis want of ability. ... I have no money – not a cent." When Elgar was 29, he took on a new pupil, Caroline Alice Roberts, daughter of the late Major-General Sir Henry Roberts, and published author of verse and prose fiction. Eight years older than Elgar, Alice became his wife three years later. Elgar's biographer Michael Kennedy writes, "Alice's family was horrified by her intention to marry an unknown musician who worked in a shop and was a Roman Catholic. She was disinherited." They were married on 8 May 1889, at Brompton Oratory. From then until her death, she acted as his business manager and social secretary, dealt with his mood swings, and was a perceptive musical critic. She did her best to gain him the attention of influential society, though with limited success. In time, he would learn to accept the honours given him, realising that they mattered more to her and her social class and recognising what she had given up to further his career. In her diary, she wrote, "The care of a genius is enough of a life work for any woman." As an engagement present, Elgar dedicated his short violin-and-piano piece "Salut d'Amour" to her. With Alice's encouragement, the Elgars moved to London to be closer to the centre of British musical life, and Elgar started devoting his time to composition. Their only child, Carice Irene, was born at their home in West Kensington on 14 August 1890. Her name, revealed in Elgar's dedication of "Salut d'Amour", was a contraction of her mother's names Caroline and Alice. Elgar took full advantage of the opportunity to hear unfamiliar music. In the days before miniature scores and recordings were available, it was not easy for young composers to get to know new music. Elgar took every chance to do so at the Crystal Palace concerts. He and Alice attended day after day, hearing music by a wide range of composers. Among these were masters of orchestration from whom he learned much, such as Berlioz and Richard Wagner. His own compositions, however, made little impact on London's musical scene. August Manns conducted Elgar's orchestral version of "Salut d'amour" and the Suite in D at the Crystal Palace, and two publishers accepted some of Elgar's violin pieces, organ voluntaries, and part songs. Some tantalising opportunities seemed to be within reach but vanished unexpectedly. For example, an offer from the Royal Opera House, Covent Garden, to run through some of his works was withdrawn at the last second when Sir Arthur Sullivan arrived unannounced to rehearse some of his own music. Sullivan was horrified when Elgar later told him what had happened. Elgar's only important commission while in London came from his home city: the Worcester Festival Committee invited him to compose a short orchestral work for the 1890 Three Choirs Festival. The result is described by Diana McVeagh in the "Grove Dictionary of Music and Musicians", as "his first major work, the assured and uninhibited "Froissart"." Elgar conducted the first performance in Worcester in September 1890. For lack of other work, he was obliged to leave London in 1891 and return with his wife and child to Worcestershire, where he could earn a living conducting local musical ensembles and teaching. They settled in Alice's former home town, Great Malvern. During the 1890s, Elgar gradually built up a reputation as a composer, chiefly of works for the great choral festivals of the English Midlands. "The Black Knight" (1892) and "King Olaf" (1896), both inspired by Longfellow, "The Light of Life" (1896) and "Caractacus" (1898) were all modestly successful, and he obtained a long-standing publisher in Novello and Co. Other works of this decade included the "Serenade for Strings" (1892) and "Three Bavarian Dances" (1897). Elgar was of enough consequence locally to recommend the young composer Samuel Coleridge-Taylor to the Three Choirs Festival for a concert piece, which helped establish the younger man's career. Elgar was catching the attention of prominent critics, but their reviews were polite rather than enthusiastic. Although he was in demand as a festival composer, he was only just getting by financially and felt unappreciated. In 1898, he said he was "very sick at heart over music" and hoped to find a way to succeed with a larger work. His friend August Jaeger tried to lift his spirits: "A day's attack of the blues ... will not drive away your desire, your necessity, which is to exercise those creative faculties which a kind providence has given you. Your time of universal recognition will come." In 1899, that prediction suddenly came true. At the age of forty-two, Elgar produced the "Enigma Variations", which were premiered in London under the baton of the eminent German conductor Hans Richter. In Elgar's own words, "I have sketched a set of Variations on an original theme. The Variations have amused me because I've labelled them with the nicknames of my particular friends ... that is to say I've written the variations each one to represent the mood of the 'party' (the person) ... and have written what I think they would have written – if they were asses enough to compose". He dedicated the work "To my friends pictured within". Probably the best known variation is "Nimrod", depicting Jaeger. Purely musical considerations led Elgar to omit variations depicting Arthur Sullivan and Hubert Parry, whose styles he tried but failed to incorporate in the variations. The large-scale work was received with general acclaim for its originality, charm and craftsmanship, and it established Elgar as the pre-eminent British composer of his generation. The work is formally titled "Variations on an Original Theme"; the word "Enigma" appears over the first six bars of music, which led to the familiar version of the title. The enigma is that, although there are fourteen variations on the "original theme", there is another overarching theme, never identified by Elgar, which he said "runs through and over the whole set" but is never heard. Later commentators have observed that although Elgar is today regarded as a characteristically English composer, his orchestral music and this work in particular share much with the Central European tradition typified at the time by the work of Richard Strauss. The "Enigma Variations" were well received in Germany and Italy, and remain to the present day a worldwide concert staple. Elgar's biographer Basil Maine commented, "When Sir Arthur Sullivan died in 1900 it became apparent to many that Elgar, although a composer of another build, was his true successor as first musician of the land." Elgar's next major work was eagerly awaited. For the Birmingham Triennial Music Festival of 1900, he set Cardinal John Henry Newman's poem "The Dream of Gerontius" for soloists, chorus and orchestra. Richter conducted the premiere, which was marred by a poorly prepared chorus, which sang badly. Critics recognised the mastery of the piece despite the defects in performance. It was performed in Düsseldorf, Germany, in 1901 and again in 1902, conducted by Julius Buths, who also conducted the European premiere of the "Enigma Variations" in 1901. The German press was enthusiastic. "The Cologne Gazette" said, "In both parts we meet with beauties of imperishable value. ... Elgar stands on the shoulders of Berlioz, Wagner, and Liszt, from whose influences he has freed himself until he has become an important individuality. He is one of the leaders of musical art of modern times." "The Düsseldorfer Volksblatt" wrote, "A memorable and epoch-making first performance! Since the days of Liszt nothing has been produced in the way of oratorio ... which reaches the greatness and importance of this sacred cantata." Richard Strauss, then widely viewed as the leading composer of his day, was so impressed that in Elgar's presence he proposed a toast to the success of "the first English progressive musician, Meister Elgar." Performances in Vienna, Paris and New York followed, and "The Dream of Gerontius" soon became equally admired in Britain. According to Kennedy, "It is unquestionably the greatest British work in the oratorio form ... [it] opened a new chapter in the English choral tradition and liberated it from its Handelian preoccupation." Elgar, as a Roman Catholic, was much moved by Newman's poem about the death and redemption of a sinner, but some influential members of the Anglican establishment disagreed. His colleague, Charles Villiers Stanford complained that the work "stinks of incense". The Dean of Gloucester banned "Gerontius" from his cathedral in 1901, and at Worcester the following year, the Dean insisted on expurgations before allowing a performance. Elgar is probably best known for the first of the five "Pomp and Circumstance Marches", which were composed between 1901 and 1930. It is familiar to millions of television viewers all over the world every year who watch the Last Night of the Proms, where it is traditionally performed. When the theme of the slower middle section (technically called the "trio") of the first march came into his head, he told his friend Dora Penny, "I've got a tune that will knock 'em – will knock 'em flat". When the first march was played in 1901 at a London Promenade Concert, it was conducted by Henry J. Wood, who later wrote that the audience "rose and yelled ... the one and only time in the history of the Promenade concerts that an orchestral item was accorded a double encore." To mark the coronation of Edward VII, Elgar was commissioned to set A. C. Benson's "Coronation Ode" for a gala concert at the Royal Opera House in June 1901. The approval of the king was confirmed, and Elgar began work. The contralto Clara Butt had persuaded him that the trio of the first "Pomp and Circumstance" march could have words fitted to it, and Elgar invited Benson to do so. Elgar incorporated the new vocal version into the Ode. The publishers of the score recognised the potential of the vocal piece, "Land of Hope and Glory", and asked Benson and Elgar to make a further revision for publication as a separate song. It was immensely popular and is now considered an unofficial British national anthem. In the United States, the trio, known simply as "Pomp and Circumstance" or "The Graduation March", has been adopted since 1905 for virtually all high school and university graduations. In March 1904 a three-day festival of Elgar's works was presented at Covent Garden, an honour never before given to any English composer. "The Times" commented, "Four or five years ago if any one had predicted that the Opera-house would be full from floor to ceiling for the performance of an oratorio by an English composer he would probably have been supposed to be out of his mind." The king and queen attended the first concert, at which Richter conducted "The Dream of Gerontius", and returned the next evening for the second, the London premiere of "The Apostles" (first heard the previous year at the Birmingham Festival). The final concert of the festival, conducted by Elgar, was primarily orchestral, apart for an excerpt from "Caractacus" and the complete "Sea Pictures" (sung by Clara Butt). The orchestral items were "Froissart", the "Enigma Variations", "Cockaigne", the first two (at that time the only two) "Pomp and Circumstance" marches, and the premiere of a new orchestral work, "In the South", inspired by a holiday in Italy. Elgar was knighted at Buckingham Palace on 5 July 1904. The following month, he and his family moved to Plâs Gwyn, a large house on the outskirts of Hereford, overlooking the River Wye, where they lived until 1911. Between 1902 and 1914, Elgar was, in Kennedy's words, at the zenith of popularity. He made four visits to the US, including one conducting tour, and earned considerable fees from the performance of his music. Between 1905 and 1908, he held the post of Peyton Professor of Music at the University of Birmingham. He had accepted the post reluctantly, feeling that a composer should not head a school of music. He was not at ease in the role, and his lectures caused controversy, with his attacks on the critics and on English music in general: "Vulgarity in the course of time may be refined. Vulgarity often goes with inventiveness ... but the commonplace mind can never be anything but commonplace. An Englishman will take you into a large room, beautifully proportioned, and will point out to you that it is white – all over white – and somebody will say, 'What exquisite taste'. You know in your own mind, in your own soul, that it is not taste at all, that it is the want of taste, that is mere evasion. English music is white, and evades everything." He regretted the controversy and was glad to hand on the post to his friend Granville Bantock in 1908. His new life as a celebrity was a mixed blessing to the highly strung Elgar, as it interrupted his privacy, and he often was in ill-health. He complained to Jaeger in 1903, "My life is one continual giving up of little things which I love." Both W. S. Gilbert and Thomas Hardy sought to collaborate with Elgar in this decade. Elgar refused, but would have collaborated with George Bernard Shaw had Shaw been willing. Elgar paid three visits to the USA between 1905 and 1911. His first was to conduct his music and to accept a doctorate from Yale University. His principal composition in 1905 was the "Introduction and Allegro for Strings", dedicated to Samuel Sanford, p It was well received but did not catch the public imagination as "The Dream of Gerontius" had done and continued to do. Among keen Elgarians, however, "The Kingdom" was sometimes preferred to the earlier work: Elgar's friend Frank Schuster told the young Adrian Boult: "compared with "The Kingdom", "Gerontius" is the work of a raw amateur." As Elgar approached his fiftieth birthday, he began work on his first symphony, a project that had been in his mind in various forms for nearly ten years. His First Symphony (1908) was a national and international triumph. Within weeks of the premiere it was performed in New York under Walter Damrosch, Vienna under Ferdinand Löwe, St Petersburg under Alexander Siloti, and Leipzig under Arthur Nikisch. There were performances in Rome, Chicago, Boston, Toronto and fifteen British towns and cities. In just over a year, it received a hundred performances in Britain, America and continental Europe. The Violin Concerto (1910) was commissioned by Fritz Kreisler, one of the leading international violinists of the time. Elgar wrote it during the summer of 1910, with occasional help from W. H. Reed, the leader of the London Symphony Orchestra, who helped the composer with advice on technical points. Elgar and Reed formed a firm friendship, which lasted for the rest of Elgar's life. Reed's biography, "Elgar As I Knew Him" (1936), records many details of Elgar's methods of composition. The work was presented by the Royal Philharmonic Society, with Kreisler and the London Symphony Orchestra, conducted by the composer. Reed recalled, "the Concerto proved to be a complete triumph, the concert a brilliant and unforgettable occasion." So great was the impact of the concerto that Kreisler's rival Eugène Ysaÿe spent much time with Elgar going through the work. There was great disappointment when contractual difficulties prevented Ysaÿe from playing it in London. The Violin Concerto was Elgar's last popular triumph. The following year he presented his Second Symphony in London, but was disappointed at its reception. Unlike the First Symphony, it ends not in a blaze of orchestral splendour but quietly and contemplatively. Reed, who played at the premiere, later wrote that Elgar was recalled to the platform several times to acknowledge the applause, "but missed that unmistakable note perceived when an audience, even an English audience, is thoroughly roused or worked up, as it was after the Violin Concerto or the First Symphony." Elgar asked Reed, "What is the matter with them, Billy? They sit there like a lot of stuffed pigs." The work was, by normal standards, a success, with twenty-seven performances within three years of its premiere, but it did not achieve the international "furore" of the First Symphony. In June 1911, as part of the celebrations surrounding the coronation of King George V, Elgar was appointed to the Order of Merit, an honour limited to twenty-four holders at any time. The following year, the Elgars moved back to London, to a large house in Netherhall Gardens, Hampstead, designed by Norman Shaw. There Elgar composed his last two large-scale works of the pre-war era, the choral ode, "The Music Makers" (for the Birmingham Festival, 1912) and the symphonic study "Falstaff" (for the Leeds Festival, 1913). Both were received politely but without enthusiasm. Even the dedicatee of "Falstaff", the conductor Landon Ronald, confessed privately that he could not "make head or tail of the piece," while the musical scholar Percy Scholes wrote of "Falstaff" that it was a "great work" but, "so far as public appreciation goes, a comparative failure." When World War I broke out, Elgar was horrified at the prospect of the carnage, but his patriotic feelings were nonetheless aroused. He composed "A Song for Soldiers", which he later withdrew. He signed up as a special constable in the local police and later joined the Hampstead Volunteer Reserve of the army. He composed patriotic works, "Carillon", a recitation for speaker and orchestra in honour of Belgium, and "Polonia", an orchestral piece in honour of Poland. "Land of Hope and Glory", already popular, became still more so, and Elgar wished in vain to have new, less nationalistic, words sung to the tune. Elgar's other compositions during the war included incidental music for a children's play, "The Starlight Express" (1915); a ballet, "The Sanguine Fan" (1917); and "The Spirit of England" (1915–17, to poems by Laurence Binyon), three choral settings very different in character from the romantic patriotism of his earlier years. His last large-scale composition of the war years was "The Fringes of the Fleet", settings of verses by Rudyard Kipling, performed with great popular success around the country, until Kipling for unexplained reasons objected to their performance in theatres. Elgar conducted a recording of the work for the Gramophone Company. Towards the end of the war, Elgar was in poor health. His wife thought it best for him to move to the countryside, and she rented 'Brinkwells', a house near Fittleworth in Sussex, from the painter Rex Vicat Cole. There Elgar recovered his strength and, in 1918 and 1919, he produced four large-scale works. The first three of these were chamber pieces: the Violin Sonata in E minor, the Piano Quintet in A minor, and the String Quartet in E minor. On hearing the work in progress, Alice Elgar wrote in her diary, "E. writing wonderful new music". All three works were well received. "The Times" wrote, "Elgar's sonata contains much that we have heard before in other forms, but as we do not at all want him to change and be somebody else, that is as it should be." The quartet and quintet were premiered at the Wigmore Hall on 21 May 1919. "The Manchester Guardian" wrote, "This quartet, with its tremendous climaxes, curious refinements of dance-rhythms, and its perfect symmetry, and the quintet, more lyrical and passionate, are as perfect examples of chamber music as the great oratorios were of their type." By contrast, the remaining work, the Cello Concerto in E minor, had a disastrous premiere, at the opening concert of the London Symphony Orchestra's 1919–20 season in October 1919. Apart from the Elgar work, which the composer conducted, the rest of the programme was conducted by Albert Coates, who overran his rehearsal time at the expense of Elgar's. Lady Elgar wrote, "that brutal selfish ill-mannered bounder ... that brute Coates went on rehearsing." The critic of "The Observer", Ernest Newman, wrote, "There have been rumours about during the week of inadequate rehearsal. Whatever the explanation, the sad fact remains that never, in all probability, has so great an orchestra made so lamentable an exhibition of itself. ... The work itself is lovely stuff, very simple – that pregnant simplicity that has come upon Elgar's music in the last couple of years – but with a profound wisdom and beauty underlying its simplicity." Elgar attached no blame to his soloist, Felix Salmond, who played the work for him again later. In contrast with the First Symphony and its hundred performances in just over a year, the Cello Concerto did not have a second performance in London for more than a year. Although in the 1920s Elgar's music was no longer in fashion, his admirers continued to present his works when possible. Reed singles out a performance of the Second Symphony in March 1920 conducted by "a young man almost unknown to the public", Adrian Boult, for bringing "the grandeur and nobility of the work" to a wider public. Also in 1920, Landon Ronald presented an all-Elgar concert at the Queen's Hall. Alice Elgar wrote with enthusiasm about the reception of the symphony, but this was one of the last times she heard Elgar's music played in public. After a short illness, she died of lung cancer on 7 April 1920, at the age of seventy-two. Elgar was devastated by the loss of his wife. With no public demand for new works, and deprived of Alice's constant support and inspiration, he allowed himself to be deflected from composition. His daughter later wrote that Elgar inherited from his father a reluctance to "settle down to work on hand but could cheerfully spend hours over some perfectly unnecessary and entirely unremunerative undertaking", a trait that became stronger after Alice's death. For much of the rest of his life, Elgar indulged himself in his several hobbies. Throughout his life he was a keen amateur chemist, sometimes using a laboratory in his back garden. He even patented the "Elgar Sulphuretted Hydrogen Apparatus" in 1908. He enjoyed football, supporting Wolverhampton Wanderers F.C., for whom he composed an anthem, ""He Banged the Leather for Goal"", and in his later years he frequently attended horseraces. His protégés, the conductor Malcolm Sargent and violinist Yehudi Menuhin, both recalled rehearsals with Elgar at which he swiftly satisfied himself that all was well and then went off to the races. In his younger days, Elgar had been an enthusiastic cyclist, buying Royal Sunbeam bicycles for himself and his wife in 1903 (he named his "Mr. Phoebus"). As an elderly widower, he enjoyed being driven about the countryside by his chauffeur. In November and December 1923, he took a voyage to Brazil, journeying up the Amazon to Manaus, where he was impressed by its opera house, the Teatro Amazonas. Almost nothing is recorded about Elgar's activities or the events that he encountered during the trip, which gave the novelist James Hamilton-Paterson considerable latitude when writing "Gerontius", a fictional account of the journey. After Alice's death, Elgar sold the Hampstead house, and after living for a short time in a flat in St James's in the heart of London, he moved back to Worcestershire, to the village of Kempsey, where he lived from 1923 to 1927. He did not wholly abandon composition in these years. He made large-scale symphonic arrangements of works by Bach and Handel and wrote his "Empire March" and eight songs "Pageant of Empire" for the 1924 British Empire Exhibition. Shortly after these were published, he was appointed Master of the King's Musick on 13 May 1924, following the death of Sir Walter Parratt. From 1926 onwards, Elgar made a series of recordings of his own works. Described by the music writer Robert Philip as "the first composer to take the gramophone seriously", he had already recorded much of his music by the early acoustic-recording process for His Master's Voice (HMV) from 1914 onwards, but the introduction of electrical microphones in 1925 transformed the gramophone from a novelty into a realistic medium for reproducing orchestral and choral music. Elgar was the first composer to take full advantage of this technological advance. Fred Gaisberg of HMV, who produced Elgar's recordings, set up a series of sessions to capture on disc the composer's interpretations of his major orchestral works, including the "Enigma Variations", "Falstaff", the first and second symphonies, and the cello and violin concertos. For most of these, the orchestra was the LSO, but the "Variations" were played by the Royal Albert Hall Orchestra. Later in the series of recordings, Elgar also conducted two newly founded orchestras, Boult's BBC Symphony Orchestra and Sir Thomas Beecham's London Philharmonic Orchestra. Elgar's recordings were released on 78-rpm discs by both HMV and RCA Victor. After World War II, the 1932 recording of the Violin Concerto with the teenage Menuhin as soloist remained available on 78 and later on LP, but the other recordings were out of the catalogues for some years. When they were reissued by EMI on LP in the 1970s, they caused surprise to many by their fast tempi, in contrast to the slower speeds adopted by many conductors in the years since Elgar's death. The recordings were reissued on CD in the 1990s. In November 1931, Elgar was filmed by Pathé for a newsreel depicting a recording session of "Pomp and Circumstance March No. 1" at the opening of EMI's Abbey Road Studios in London. It is believed to be the only surviving sound film of Elgar, who makes a brief remark before conducting the London Symphony Orchestra, asking the musicians to "play this tune as though you've never heard it before." A memorial plaque to Elgar at Abbey Road was unveiled on 24 June 1993. A late piece of Elgar's, the "Nursery Suite", was an early example of a studio premiere: its first performance was in the Abbey Road studios. For this work, dedicated to the wife and daughters of the Duke of York, Elgar once again drew on his youthful sketch-books. In his final years, Elgar experienced a musical revival. The BBC organised a festival of his works to celebrate his seventy-fifth birthday, in 1932. He flew to Paris in 1933 to conduct the Violin Concerto for Menuhin. While in France, he visited his fellow composer Frederick Delius at his house at Grez-sur-Loing. He was sought out by younger musicians such as Adrian Boult, Malcolm Sargent and John Barbirolli, who championed his music when it was out of fashion. He began work on an opera, "The Spanish Lady", and accepted a commission from the BBC to compose a Third Symphony. His final illness, however, prevented their completion. He fretted about the unfinished works. He asked Reed to ensure that nobody would "tinker" with the sketches and attempt a completion of the symphony, but at other times he said, "If I can't complete the Third Symphony, somebody will complete it – or write a better one." After Elgar's death, Percy M. Young, in co-operation with the BBC and Elgar's daughter Carice, produced a version of "The Spanish Lady", which was issued on CD. The Third Symphony sketches were elaborated by the composer Anthony Payne into a complete score in 1998. Inoperable colorectal cancer was discovered during an operation on 8 October 1933. He told his consulting doctor, Arthur Thomson, that he had no faith in an afterlife: "I believe there is nothing but complete oblivion." Elgar died on 23 February 1934 at the age of seventy-six and was buried next to his wife at St Wulstan's Roman Catholic Church in Little Malvern. Elgar was contemptuous of folk music and had little interest in or respect for the early English composers, calling William Byrd and his contemporaries "museum pieces". Of later English composers, he regarded Purcell as the greatest, and he said that he had learned much of his own technique from studying Hubert Parry's writings. The continental composers who most influenced Elgar were Handel, Dvořák and, to some degree, Brahms. In Elgar's chromaticism, the influence of Wagner is apparent, but Elgar's individual style of orchestration owes much to the clarity of nineteenth-century French composers, Berlioz, Massenet, Saint-Saëns and, particularly, Delibes, whose music Elgar played and conducted at Worcester and greatly admired. Elgar began composing when still a child, and all his life he drew on his early sketchbooks for themes and inspiration. The habit of assembling his compositions, even large-scale ones, from scraps of themes jotted down randomly remained throughout his life. His early adult works included violin and piano pieces, music for the wind quintet in which he and his brother played between 1878 and 1881, and music of many types for the Powick Asylum band. Diana McVeagh in "Grove's Dictionary" finds many embryonic Elgarian touches in these pieces, but few of them are regularly played, except "Salut d'Amour" and (as arranged decades later into "The Wand of Youth" Suites) some of the childhood sketches. Elgar's sole work of note during his first spell in London in 1889–91, the overture "Froissart", was a romantic-bravura piece, influenced by Mendelssohn and Wagner, but also showing further Elgarian characteristics. Orchestral works composed during the subsequent years in Worcestershire include the "Serenade for Strings" and "Three Bavarian Dances". In this period and later, Elgar wrote songs and partsongs. W. H. Reed expressed reservations about these pieces, but praised the partsong "The Snow", for female voices, and "Sea Pictures", a cycle of five songs for contralto and orchestra which remains in the repertory. Elgar's principal large-scale early works were for chorus and orchestra for the Three Choirs and other festivals. These were "The Black Knight", "King Olaf", "The Light of Life", "The Banner of St George" and "Caractacus". He also wrote a "Te Deum" and "Benedictus" for the Hereford Festival. Of these, McVeagh comments favourably on his lavish orchestration and innovative use of leitmotifs, but less favourably on the qualities of his chosen texts and the patchiness of his inspiration. McVeagh makes the point that, because these works of the 1890s were for many years little known (and performances remain rare), the mastery of his first great success, the "Enigma Variations", appeared to be a sudden transformation from mediocrity to genius, but in fact his orchestral skills had been building up throughout the decade. Elgar's best-known works were composed within the twenty-one years between 1899 and 1920. Most of them are orchestral. Reed wrote, "Elgar's genius rose to its greatest height in his orchestral works" and quoted the composer as saying that, even in his oratorios, the orchestral part is the most important. The "Enigma Variations" made Elgar's name nationally. The variation form was ideal for him at this stage of his career, when his comprehensive mastery of orchestration was still in contrast to his tendency to write his melodies in short, sometimes rigid, phrases. His next orchestral works, "Cockaigne", a concert-overture (1900–1901), the first two "Pomp and Circumstance" marches (1901), and the gentle "Dream Children" (1902), are all short: the longest of them, "Cockaigne", lasting less than fifteen minutes. "In the South" (1903–1904), although designated by Elgar as a concert-overture, is, according to Kennedy, really a tone poem and the longest continuous piece of purely orchestral writing Elgar had essayed. He wrote it after setting aside an early attempt to compose a symphony. The work reveals his continuing progress in writing sustained themes and orchestral lines, although some critics, including Kennedy, find that in the middle part "Elgar's inspiration burns at less than its brightest." In 1905 Elgar completed the "Introduction and Allegro for Strings". This work is based, unlike much of Elgar's earlier writing, not on a profusion of themes but on only three. Kennedy called it a "masterly composition, equalled among English works for strings only by Vaughan Williams's "Tallis Fantasia"." Nevertheless, at less than a quarter of an hour, it was not by contemporary standards a lengthy composition. Gustav Mahler's Seventh Symphony, composed at the same time, runs for well over an hour. During the next four years, however, Elgar composed three major concert pieces, which, though shorter than comparable works by some of his European contemporaries, are among the most substantial such works by an English composer. These were his First Symphony, Violin Concerto, and Second Symphony, which all play for between forty-five minutes and an hour. McVeagh says of the symphonies that they "rank high not only in Elgar's output but in English musical history. Both are long and powerful, without published programmes, only hints and quotations to indicate some inward drama from which they derive their vitality and eloquence. Both are based on classical form but differ from it to the extent that ... they were considered prolix and slackly constructed by some critics. Certainly the invention in them is copious; each symphony would need several dozen music examples to chart its progress." Elgar's Violin Concerto and Cello Concerto, in the view of Kennedy, "rank not only among his finest works, but among the greatest of their kind". They are, however, very different from each other. The Violin Concerto, composed in 1909 as Elgar reached the height of his popularity, and written for the instrument dearest to his heart, is lyrical throughout and rhapsodical and brilliant by turns. The Cello Concerto, composed a decade later, immediately after World War I, seems, in Kennedy's words, "to belong to another age, another world ... the simplest of all Elgar's major works ... also the least grandiloquent." Between the two concertos came Elgar's symphonic study "Falstaff", which has divided opinion even among Elgar's strongest admirers. Donald Tovey viewed it as "one of the immeasurably great things in music", with power "identical with Shakespeare's", while Kennedy criticises the work for "too frequent reliance on sequences" and an over-idealised depiction of the female characters. Reed thought that the principal themes show less distinction than some of Elgar's earlier works. Elgar himself thought "Falstaff" the highest point of his purely orchestral work. The major works for voices and orchestra of the twenty-one years of Elgar's middle period are three large-scale works for soloists, chorus and orchestra: "The Dream of Gerontius" (1900), and the oratorios "The Apostles" (1903) and "The Kingdom" (1906); and two shorter odes, the "Coronation Ode" (1902) and "The Music Makers" (1912). The first of the odes, as a "pièce d'occasion", has rarely been revived after its initial success, with the culminating "Land of Hope and Glory". The second is, for Elgar, unusual in that it contains several quotations from his earlier works, as Richard Strauss quoted himself in "Ein Heldenleben". The choral works were all successful, although the first, "Gerontius", was and remains the best-loved and most performed. On the manuscript Elgar wrote, quoting John Ruskin, "This is the best of me; for the rest, I ate, and drank, and slept, loved and hated, like another. My life was as the vapour, and is not; but this I saw, and knew; this, if anything of mine, is worth your memory." All three of the large-scale works follow the traditional model with sections for soloists, chorus and both together. Elgar's distinctive orchestration, as well as his melodic inspiration, lifts them to a higher level than most of their British predecessors. Elgar's other works of his middle period include incidental music for "Grania and Diarmid", a play by George Moore and W. B. Yeats (1901), and for "The Starlight Express", a play based on a story by Algernon Blackwood (1916). Of the former, Yeats called Elgar's music "wonderful in its heroic melancholy". Elgar also wrote a number of songs during his peak period, of which Reed observes, "it cannot be said that he enriched the vocal repertory to the same extent as he did that of the orchestra." After the Cello Concerto, Elgar completed no more large-scale works. He made arrangements of works by Bach, Handel and Chopin, in distinctively Elgarian orchestration, and once again turned his youthful notebooks to use for the "Nursery Suite" (1931). His other compositions of this period have not held a place in the regular repertory. For most of the rest of the twentieth century, it was generally agreed that Elgar's creative impulse ceased after his wife's death. Anthony Payne's elaboration of the sketches for Elgar's Third Symphony led to a reconsideration of this supposition. Elgar left the opening of the symphony complete in full score, and those pages, along with others, show Elgar's orchestration changed markedly from the richness of his pre-war work. "The Gramophone" described the opening of the new work as something "thrilling ... unforgettably gaunt". Payne also subsequently produced a performing version of the sketches for a sixth "Pomp and Circumstance March", premiered at the Proms in August 2006. Elgar's sketches for a piano concerto dating from 1913 were elaborated by the composer Robert Walker and first performed in August 1997 by the pianist David Owen Norris. The realisation has since been extensively revised. Views of Elgar's stature have varied in the decades since his music came to prominence at the beginning of the twentieth century. Richard Strauss, as noted, hailed Elgar as a progressive composer; even the hostile reviewer in "The Observer", unimpressed by the thematic material of the First Symphony in 1908, called the orchestration "magnificently modern". Hans Richter rated Elgar as "the greatest modern composer" in any country, and Richter's colleague Arthur Nikisch considered the First Symphony "a masterpiece of the first order" to be "justly ranked with the great symphonic models – Beethoven and Brahms." By contrast, the critic W. J. Turner, in the mid-twentieth century, wrote of Elgar's "Salvation Army symphonies," and Herbert von Karajan called the "Enigma Variations" "second-hand Brahms". Elgar's immense popularity was not long-lived. After the success of his First Symphony and Violin Concerto, his Second Symphony and Cello Concerto were politely received but without the earlier wild enthusiasm. His music was identified in the public mind with the Edwardian era, and after the First World War he no longer seemed a progressive or modern composer. In the early 1920s, even the First Symphony had only one London performance in more than three years. Henry Wood and younger conductors such as Boult, Sargent and Barbirolli championed Elgar's music, but in the recording catalogues and the concert programmes of the middle of the century his works were not well represented. In 1924, the music scholar Edward J. Dent wrote an article for a German music journal in which he identified four features of Elgar's style that gave offence to a section of English opinion (namely, Dent indicated, the academic and snobbish section): "too emotional", "not quite free from vulgarity", "pompous", and "too deliberately noble in expression". This article was reprinted in 1930 and caused controversy. In the later years of the century there was, in Britain at least, a revival of interest in Elgar's music. The features that had offended austere taste in the inter-war years were seen from a different perspective. In 1955, the reference book "The Record Guide" wrote of the Edwardian background during the height of Elgar's career: By the 1960s, a less severe view was being taken of the Edwardian era. In 1966 the critic Frank Howes wrote that Elgar reflected the last blaze of opulence, expansiveness and full-blooded life, before World War I swept so much away. In Howes's view, there was a touch of vulgarity in both the era and Elgar's music, but "a composer is entitled to be judged by posterity for his best work. ... Elgar is historically important for giving to English music a sense of the orchestra, for expressing what it felt like to be alive in the Edwardian age, for conferring on the world at least four unqualified masterpieces, and for thereby restoring England to the comity of musical nations." In 1967 the critic and analyst David Cox considered the question of the supposed Englishness of Elgar's music. Cox noted that Elgar disliked folk-songs and never used them in his works, opting for an idiom that was essentially German, leavened by a lightness derived from French composers including Berlioz and Gounod. How then, asked Cox, could Elgar be "the most English of composers"? Cox found the answer in Elgar's own personality, which "could use the alien idioms in such a way as to make of them a vital form of expression that was his and his alone. And the personality that comes through in the music is English." This point about Elgar's transmuting his influences had been touched on before. In 1930 "The Times" wrote, "When Elgar's first symphony came out, someone attempted to prove that its main tune on which all depends was like the Grail theme in Parsifal. ... but the attempt fell flat because everyone else, including those who disliked the tune, had instantly recognized it as typically 'Elgarian', while the Grail theme is as typically Wagnerian." As for Elgar's "Englishness", his fellow-composers recognised it: Richard Strauss and Stravinsky made particular reference to it, and Sibelius called him, "the personification of the true English character in music ... a noble personality and a born aristocrat". Among Elgar's admirers there is disagreement about which of his works are to be regarded as masterpieces. The "Enigma Variations" are generally counted among them. "The Dream of Gerontius" has also been given high praise by Elgarians, and the Cello Concerto is similarly rated. Many rate the Violin Concerto equally highly, but some do not. Sackville-West omitted it from the list of Elgar masterpieces in "The Record Guide", and in a long analytical article in "The Musical Quarterly", Daniel Gregory Mason criticised the first movement of the concerto for a "kind of sing-songiness ... as fatal to noble rhythm in music as it is in poetry." "Falstaff" also divides opinion. It has never been a great popular favourite, and Kennedy and Reed identify shortcomings in it. In a "Musical Times" 1957 centenary symposium on Elgar led by Vaughan Williams, by contrast, several contributors share Eric Blom's view that "Falstaff" is the greatest of all Elgar's works. The two symphonies divide opinion even more sharply. Mason rates the Second poorly for its "over-obvious rhythmic scheme", but calls the First "Elgar's masterpiece. ... It is hard to see how any candid student can deny the greatness of this symphony." However, in the 1957 centenary symposium, several leading admirers of Elgar express reservations about one or both symphonies. In the same year, Roger Fiske wrote in "The Gramophone", "For some reason few people seem to like the two Elgar symphonies equally; each has its champions and often they are more than a little bored by the rival work." The critic John Warrack wrote, "There are no sadder pages in symphonic literature than the close of the First Symphony's Adagio, as horn and trombones twice softly intone a phrase of utter grief", whereas to Michael Kennedy, the movement is notable for its lack of anguished yearning and "angst" and is marked instead by a "benevolent tranquillity." Despite the fluctuating critical assessment of the various works over the years, Elgar's major works taken as a whole have in the twenty-first century recovered strongly from their neglect in the 1950s. "The Record Guide" in 1955 could list only one currently available recording of the First Symphony, none of the Second, one of the Violin Concerto, two of the Cello Concerto, two of the "Enigma Variations", one of "Falstaff", and none of "The Dream of Gerontius". Since then there have been multiple recordings of all the major works. More than thirty recordings have been made of the First Symphony since 1955, for example, and more than a dozen of "The Dream of Gerontius". Similarly, in the concert hall, Elgar's works, after a period of neglect, are once again frequently programmed. The Elgar Society's website, in its diary of forthcoming performances, lists performances of Elgar's works by orchestras, soloists and conductors across Europe, North America and Australia. Elgar was knighted in 1904, and in 1911 he was appointed a member of the Order of Merit. In 1920 he received the Cross of Commander of the Belgian Order of the Crown; in 1924 he was made Master of the King's Musick; the following year he received the Gold Medal of the Royal Philharmonic Society; and in 1928 he was appointed a Knight Commander of the Royal Victorian Order (KCVO). Between 1900 and 1931, Elgar received honorary degrees from the Universities of Cambridge, Durham, Leeds, Oxford, Yale (USA), Aberdeen, Western Pennsylvania (USA), Birmingham and London. Foreign academies of which he was made a member were Regia Accademia di Santa Cecilia, Rome; Accademia del Reale Istituto Musicale, Florence; Académie des Beaux Arts, Paris; Institut de France; and the American Academy. In 1931 he was created a Baronet, of Broadheath in the County of Worcester. In 1933 he was promoted within the Royal Victorian Order to Knight Grand Cross (GCVO). In Kennedy's words, he "shamelessly touted" for a peerage, but in vain. In "Who's Who", post World War I, he claimed to have been awarded "several Imperial Russian and German decorations (lapsed)". Elgar was offered, but declined, the office of Mayor of Hereford (despite not being a member of its city council) when he lived in the city in 1905. The same year he was made an honorary Freeman of the city of Worcester. The house in Lower Broadheath where Elgar was born is now the Elgar Birthplace Museum, devoted to his life and work. Elgar's daughter, Carice, helped to found the museum in 1936 and bequeathed to it much of her collection of Elgar's letters and documents on her death in 1970. Carice left Elgar manuscripts to musical colleges: "The Black Knight" to Trinity College of Music; "King Olaf" to the Royal Academy of Music; "The Music Makers" to Birmingham University; the Cello Concerto to the Royal College of Music; "The Kingdom" to the Bodleian Library; and other manuscripts to the British Museum. The Elgar Society dedicated to the composer and his works was formed in 1951. The University of Birmingham's Special Collections contain an archive of letters written by Elgar. Elgar's statue at the end of Worcester High Street stands facing the cathedral, only yards from where his father's shop once stood. Another statue of the composer by Rose Garrard is at the top of Church Street in Malvern, overlooking the town and giving visitors an opportunity to stand next to the composer in the shadow of the Hills that he so often regarded. In September 2005, a third statue sculpted by Jemma Pearson was unveiled near Hereford Cathedral in honour of his many musical and other associations with the city. It depicts Elgar with his bicycle. From 1999 until early 2007, new Bank of England twenty pound notes featured a portrait of Elgar. The change to remove his image generated controversy, particularly because 2007 was the 150th anniversary of Elgar's birth. From 2007 the Elgar notes were phased out, ceasing to be legal tender on 30 June 2010. There are around 65 roads in the UK named after Elgar, including six in the counties of Herefordshire and Worcestershire. Elgar had three locomotives named in his honour. Elgar's life and music have inspired works of literature including the novel "Gerontius" and several plays. "Elgar's Rondo", a 1993 stage play by David Pownall depicts the dead Jaeger offering ghostly advice on Elgar's musical development. Pownall also wrote a radio play, "Elgar's Third" (1994); another Elgar-themed radio play is Alick Rowe's "The Dorabella Variation" (2003). David Rudkin's BBC television "Play for Today" "Penda's Fen" (1974) deals with themes including sex and adolescence, spying, and snobbery, with Elgar's music, chiefly "The Dream of Gerontius", as its background. In one scene, a ghostly Elgar whispers the secret of the "Enigma" tune to the youthful central character, with an injunction not to reveal it. "Elgar on the Journey to Hanley", a novel by Keith Alldritt (1979), tells of the composer's attachment to Dora Penny, later Mrs Powell, (depicted as "Dorabella" in the "Enigma Variations"), and covers the fifteen years from their first meeting in the mid-1890s to the genesis of the Violin Concerto when, in the novel, Dora has been supplanted in Elgar's affections by Alice Stuart-Wortley. Perhaps the best-known work depicting Elgar is Ken Russell's 1962 BBC television film "Elgar", made when the composer was still largely out of fashion. This hour-long film contradicted the view of Elgar as a jingoistic and bombastic composer, and evoked the more pastoral and melancholy side of his character and music. Notes References
https://en.wikipedia.org/wiki?curid=10085
European Currency Unit The European Currency Unit ( or ECU) was a basket of the currencies of the European Community member states, used as the unit of account of the European Community before being replaced by the euro on 1 January 1999, at parity. The ECU itself replaced the European Unit of Account, also at parity, on 13 March 1979. The European Exchange Rate Mechanism attempted to minimize fluctuations between member state currencies and the ECU. The ECU was also used in some international financial transactions, where its advantage was that securities denominated in ECUs provided investors with the opportunity for foreign diversification without reliance on the currency of a single country. The ECU was conceived on 13 March 1979 as an internal accounting unit. It had the ISO 4217 currency code XEU. In 1990 the British Chancellor of the Exchequer proposed the creation of a 'hard' ECU, which different national currencies could compete against and, if the ECU was successful, could lead to a single currency. The move was seen by France and Germany as a wrecking tactic, especially when the increasingly Euro-sceptic Thatcher announced her outright opposition to European Economic and Monetary Union (EMU), and the idea was abandoned. On 1 January 1999, the euro (with the code EUR and symbol €) replaced the ECU, at the value €1 = 1 ECU. Unlike the ECU, the euro is a real currency, although not all member states participate (for details on euro membership see Eurozone). Two of the countries in the ECU basket of currencies, UK and Denmark, did not join the eurozone, and a third, Greece, joined late. On the other hand, Finland and Austria joined the eurozone from the beginning although their currencies were not part of the ECU basket (since they had joined the EU in 1995, two years after the ECU composition was "frozen") Due to the ECU being used in some international financial transactions, there was a concern that foreign courts might not recognize the euro as the legal successor to the ECU. This was unlikely to be a problem, since it is a generally accepted principle of private international law that states determine their currencies, and that therefore states would accept the European Union legislation to that effect. However, for abundant caution, several foreign jurisdictions adopted legislation to ensure a smooth transition. Of particular importance, the U.S. states of Illinois and New York adopted legislation to ensure a large proportion of international financial contracts recognized the euro as the successor of the ECU. Although the acronym ECU is formed from English words, "écu" is also the name of an ancient French coin. That was one reason that a new name was devised for its successor currency, "euro", which was felt not to favour any single language. The currency's symbol, (U+20A0), comprises an interlaced C and E, which are the initial letters of the phrase 'European Community' in many European languages. However, this symbol was not widely used: few systems at the time could render it and in any case banks preferred (as with all currencies) to use the ISO code XEU. As the ECU was only an electronic unit of account and not a full currency, it did not have any official coins or notes that could be used for everyday transactions. However, various European countries and organisations like the European Parliament made commemorative and mock-up coins and notes. A common theme on the coins was usually celebrating European unity, such as celebrating membership of the European Union. In 1989, the government of the Netherlands issued a series of ECU coins from ₠ to ₠200, which could be spent in shops in The Hague, during the European Capital of Culture festival. Gibraltar issued commemorative coins from 1993 through 1996, though these were (nominally) legal tender only in Gibraltar, which uses the pound sterling.
https://en.wikipedia.org/wiki?curid=10087
Eastern Caribbean dollar The Eastern Caribbean dollar (symbol: $; code: XCD) is the currency of all seven full members and one associate member of the Organisation of Eastern Caribbean States (OECS). The successor to the British West Indies dollar, it has existed since 1965, and it is normally abbreviated with the dollar sign "$" or, alternatively, "EC$" to distinguish it from other dollar-denominated currencies. The EC$ is subdivided into 100 cents. It has been pegged to the United States dollar since 7 July 1976, at the exchange rate of US$1 = EC$2.70. Six of the states using the EC$ are independent states: Antigua and Barbuda, Dominica, Grenada, Saint Kitts and Nevis, Saint Lucia, and Saint Vincent and the Grenadines. The other two, Anguilla and Montserrat, are British overseas territories. These states are all members of the Eastern Caribbean Currency Union. The other two associate members of the OECS do not use the Eastern Caribbean dollar as their official currency: the British Virgin Islands and Martinique. The British Virgin Islands were always problematic for currency purposes due to their proximity to the Danish West Indies, which became the United States Virgin Islands in 1917. Officially, the British Virgin Islands used to use sterling, but in practice the situation was more complicated and involved the circulation of French francs and U.S. dollars. In 1951, the British Virgin Islands adopted the British West Indies dollar which at that time operated in conjunction with the sterling coinage, and in 1959 they changed over officially to the U.S. dollar. Martinique, as part of France, uses the euro as its currency. British Guiana and Barbados had previously been members of the Eastern Caribbean Currency Union but withdrew in 1966 and 1972, respectively. Trinidad and Tobago had been a member of the earlier British West Indies currency union, but withdrew in 1964. The combined population of the EC$ area is about 613,000 (2014 census and estimates), which is comparable to Montenegro or the American capital city of Washington, D.C. The combined GDP is 5.46 billion US dollars, which is comparable to Bermuda. Queen Elizabeth II appears on the banknotes and also on the obverse of the coins. She is the head of state of all the states and territories using the EC$, except for Dominica. Dominica is nevertheless a member of the Commonwealth of Nations which recognises Queen Elizabeth II as Head of the Commonwealth. Queen Anne's proclamation of 1704 introduced the gold standard to the British West Indies, putting the West Indies about two hundred years ahead of the East Indies in this respect. Nevertheless, silver pieces of eight continued to form an important portion of the circulating coinage right up until the late 1870s. In 1822, the British government coined , , and fractional 'Anchor dollars' for use in Mauritius and the British West Indies (but not Jamaica). A few years later copper fractional dollars were coined for Mauritius, Sierra Leone, and the British West Indies. The first move to introduce British sterling silver coinage to the colonies came with an imperial order-in-council dated 1825. This move was inspired by a number of factors. The United Kingdom was now operating a very successful gold standard in relation to the gold sovereign that was introduced in 1816, and there was a desire to extend this system to the colonies. In addition to this, there was the fact that the supply of Spanish dollars (pieces of eight) had been cut off as a result of the revolutions in Latin America where most of the Spanish dollars were minted. The last Spanish Dollar was in fact minted at Potosi in 1825. There was now a growing desire to have a stable and steady supply of British shillings everywhere the British drum was beating. The 1825 order-in-council was largely a failure because it made sterling silver coinage legal tender at the unrealistic rating in relation to the Spanish dollar of $1 = 4 shillings 4 pence. It succeeded in Jamaica, Bermuda, and British Honduras because the authorities in those territories set aside the official ratings and used the more realistic rating of $1 = 4 shillings. The reality of the rating between the dollar and the pound was based on the silver content of the Spanish pieces of eight as compared to the gold content of the British gold sovereign. A second imperial order-in-council was passed in 1838 with the correct rating of $1 = 4 shillings 2 pence. In the years following the 1838 order-in-council, the British West Indies territories began to enact local legislation for the purposes of assimilating their monies of account with the British pound sterling. Gold discoveries in Australia in 1851 drove the silver dollar out of the West Indies, but it returned again with the great depreciation in the value of silver that followed with Germany's transition to the gold standard between 1871 and 1873. In the years immediately following 1873, there was a fear that the British West Indies might return to a silver standard. As such, legislation was passed in the individual territories to demonetize the silver dollars. Even though the British coinage was also silver, it represented fractions of the gold sovereign and so its value was based on a gold standard. During this period, and into the nineteenth century, accounts could be kept in either dollars or sterling. Jamaica, Bermuda, and the Bahamas preferred to use sterling accounts whereas British Guiana used dollar accounts. British Guiana used dollar accounts for the purpose of assisting in the transition from the Dutch guilder system of currency to the British pound sterling system. In the Eastern Caribbean territories the private sector preferred to use dollar accounts whereas the government preferred to use sterling accounts. In some of the Eastern Caribbean territories, notes were issued by various private banks, denominated in dollars equivalent to 4 shillings 2 pence. See Antigua dollar, Barbadian dollar, Dominican dollar, Grenadian dollar, Guyanese dollar, Saint Kitts dollar, Saint Lucia dollar, Saint Vincent dollar and Trinidad and Tobago dollar. In 1946, a West Indian Currency Conference saw Barbados, British Guiana, the Leeward Islands, Trinidad and Tobago and the Windward Islands agree to establish a unified decimal currency system based on a West Indian dollar to replace the current arrangement of having three different Boards of Commissioners of Currency (for Barbados (which also served the Leeward and Windward Islands), British Guiana and Trinidad & Tobago). In 1949, the British government formalized the dollar system of accounts in British Guiana and the Eastern Caribbean territories by introducing the British West Indies dollar (BWI$) at the already existing conversion rate of $4.80 per pound sterling (or $1 = 4 shillings 2 pence). It was one of the many experimental political and economic ventures tested by the British government to form a uniform system within the British West Indies territories. The symbol "BWI$" for frequently used and the currency was known verbally as the "Beewee" (slang for British West Indies) dollar. Shortly thereafter in the 1950, the British Caribbean Currency Board (BCCB) was set up in Trinidad with the sole right to issue notes and coins of the new unified currency and given the mandate of keeping full foreign exchange cover to ensure convertibility at $4.80 per pound sterling. In 1951, the British Virgin Islands joined the arrangement, but this led to discontent because that territory was more naturally drawn to the currency of the neighbouring U.S. Virgin Islands. In 1961, the British Virgin Islands withdrew from the arrangement and adopted the U.S. dollar. Until 1955, the BWI$ existed only as banknotes in conjunction with sterling fractional coinage. Decimal coins replaced the sterling coins in 1955. These decimal coins were denominated in cents, with each cent worth one halfpenny in sterling. In 1958, the West Indies Federation was established and the BWI$ was its currency. However, although Jamaica (including the Cayman Islands and the Turks and Caicos Islands) was part of the West Indies Federation, it retained the Jamaican pound, despite adopting the BWI$ as legal tender from 1954. Jamaica, the Cayman Islands, and the Turks and Caicos Islands were already long established users of the sterling accounts system of pounds, shillings, and pence. In 1964 Jamaica ended the legal tender status of the BWI$ and Trinidad and Tobago withdrew from the currency union (adopting the Trinidad and Tobago dollar) forcing the movement of the headquarters of the BCCB to Barbados and soon the "BWI$" dollar lost its regional support. In 1965, the British West Indies dollar of the now defunct West Indies Federation was replaced at par by the Eastern Caribbean dollar and the BCCB was replaced by the Eastern Caribbean Currency Authority or ECCA (established by the Eastern Caribbean Currency Agreement 1965). British Guiana withdrew from the currency union the following year. Grenada, which had used the Trinidad and Tobago dollar from 1964, rejoined the common currency arrangement in 1968. Barbados withdrew from the currency union in 1972, following which the ECCA headquarters were moved to St. Kitts. Between 1965 and 1983, the Eastern Caribbean Currency Authority issued the EC$, with banknotes from 1965 and coins from 1981. The EC$ is now issued by the Eastern Caribbean Central Bank, based in the city of Basseterre, in Saint Kitts and Nevis. The bank was established by an agreement (the Eastern Caribbean Central Bank Agreement) signed at Port of Spain on 5 July 1983. The exchange rate of $4.80 = £1 sterling (equivalent to the old $1 = 4s 2d) continued until 1976 for the new Eastern Caribbean dollar. For a wider outline of the history of currency in the region see Currencies of the British West Indies. Until 1981, the coins of the BWI$ circulated. In 1982, a new series of coins was introduced in denominations of 1, 2, 5, 10 and 25 cents and 1 dollar. The 1- and 5-cent coins were scalloped in shape while the 2-cent coin was square. These three were struck in aluminum. The 10- and 25-cent coins were round and cupro-nickel. The dollar was aluminum bronze and also round. The round, aluminum bronze dollar coin was replaced in 1989 with a decagonal, cupro-nickel type. In 2002 new and larger round-shaped 1-, 2-, and 5-cent pieces were introduced, along with a new 1-dollar coin which was also round. The effigy of Queen Elizabeth II was also changed that same year on all coin denominations to the Ian Rank-Broadley design, making it the last commonwealth currency up to that date to discontinue the Arnold Machin portrait. Their compositions remained aluminum and cupro-nickel, respectively. Higher denominations exist, but these were issued only as medal-coins. 1- and 2-cent coins were withdrawn from circulation in July 2015, and will remain legal tender until 30 June 2020. In 1965, the Eastern Caribbean Currency Authority issued banknotes in denominations of 1, 5, 20 and 100 dollars, all featuring Pietro Annigoni's 1956 portrait of Queen Elizabeth II in regalia of the Order of the Garter. The first issues in the name of the Eastern Caribbean Central Bank in 1985 were of the same denominations, with the addition of 10-dollar notes. The last 1-dollar notes were issued in 1989 and 50-dollar notes were introduced in 1993. On 1 April 2008, the Eastern Caribbean Central Bank issued a new series of banknotes which are like the preceding issues, except for omitting both the barcode and the country code letters which form part of the serial number on current notes. In 2012, the Eastern Caribbean Central Bank issued a series of banknotes with Braille features in an effort to provide notes which are easier for blind and visually impaired persons to use. The raised Braille characters on the upgraded notes feature a cricket theme in the form of balls and stumps. These characters have been added to the 10-, 20-, 50-, and 100-dollar notes. In 2019, the Eastern Caribbean Central Bank is set to introduce a new family of notes produced in polymer substrate and are to be presented in a vertical format. The 2012 issue included raised braille elements for the visually-impaired in the form of a cricket ball and stumps. These were added to the EC$10, $20, $50, and $100 banknotes.
https://en.wikipedia.org/wiki?curid=10088
Erythromycin Erythromycin is an antibiotic used for the treatment of a number of bacterial infections. This includes respiratory tract infections, skin infections, chlamydia infections, pelvic inflammatory disease, and syphilis. It may also be used during pregnancy to prevent Group B streptococcal infection in the newborn, as well as to improve delayed stomach emptying. It can be given intravenously and by mouth. An eye ointment is routinely recommended after delivery to prevent eye infections in the newborn. Common side effects include abdominal cramps, vomiting, and diarrhea. More serious side effects may include "Clostridium difficile" colitis, liver problems, prolonged QT, and allergic reactions. It is generally safe in those who are allergic to penicillin. Erythromycin also appears to be safe to use during pregnancy. While generally regarded as safe during breastfeeding, its use by the mother during the first two weeks of life may increase the risk of pyloric stenosis in the baby. This risk also applies if taken directly by the baby during this age. It is in the macrolide family of antibiotics and works by decreasing bacterial protein production. Erythromycin was first isolated in 1952 from the bacteria "Saccharopolyspora erythraea". It is on the World Health Organization's List of Essential Medicines, the safest and most effective medicines needed in a health system. The World Health Organization classifies it as critically important for human medicine. It is available as a generic medication and is not very expensive. In 2017, it was the 215th most commonly prescribed medication in the United States, with more than two million prescriptions. Erythromycin can be used to treat bacteria responsible for causing infections of the skin and upper respiratory tract, including "Streptococcus", "Staphylococcus", "Haemophilus" and "Corynebacterium" genera. The following represents MIC susceptibility data for a few medically significant bacteria: It may be useful in treating gastroparesis due to this promotility effect. It has been shown to improve feeding intolerances in those who are critically ill. Intravenous erythromycin may also be used in endoscopy to help clear stomach contents. Erythromycin is available in enteric-coated tablets, slow-release capsules, oral suspensions, ophthalmic solutions, ointments, gels, enteric-coated capsules, non enteric-coated tablets, non enteric-coated capsules, and injections. The following erythromycin combinations are available for oral dosage: For injection, the available combinations are: For ophthalmic use: Gastrointestinal disturbances, such as diarrhea, nausea, abdominal pain, and vomiting, are very common because erythromycin is a motilin agonist. Because of this, erythromycin tends not to be prescribed as a first-line drug. More serious side effects include arrhythmia with prolonged QT intervals, including "torsades de pointes", and reversible deafness. Allergic reactions range from urticaria to anaphylaxis. Cholestasis, Stevens–Johnson syndrome, and toxic epidermal necrolysis are some other rare side effects that may occur. Studies have shown evidence both for and against the association of pyloric stenosis and exposure to erythromycin prenatally and postnatally. Exposure to erythromycin (especially long courses at antimicrobial doses, and also through breastfeeding) has been linked to an increased probability of pyloric stenosis in young infants. Erythromycin used for feeding intolerance in young infants has not been associated with hypertrophic pyloric stenosis. Erythromycin estolate has been associated with reversible hepatotoxicity in pregnant women in the form of elevated serum glutamic-oxaloacetic transaminase and is not recommended during pregnancy. Some evidence suggests similar hepatotoxicity in other populations. It can also affect the central nervous system, causing psychotic reactions, nightmares, and night sweats. Erythromycin is metabolized by enzymes of the cytochrome P450 system, in particular, by isozymes of the CYP3A superfamily. The activity of the CYP3A enzymes can be induced or inhibited by certain drugs (e.g., dexamethasone), which can cause it to affect the metabolism of many different drugs, including erythromycin. If other CYP3A substrates — drugs that are broken down by CYP3A — such as simvastatin (Zocor), lovastatin (Mevacor), or atorvastatin (Lipitor)—are taken concomitantly with erythromycin, levels of the substrates increase, often causing adverse effects. A noted drug interaction involves erythromycin and simvastatin, resulting in increased simvastatin levels and the potential for rhabdomyolysis. Another group of CYP3A4 substrates are drugs used for migraine such as ergotamine and dihydroergotamine; their adverse effects may be more pronounced if erythromycin is associated. Earlier case reports on sudden death prompted a study on a large cohort that confirmed a link between erythromycin, ventricular tachycardia, and sudden cardiac death in patients also taking drugs that prolong the metabolism of erythromycin (like verapamil or diltiazem) by interfering with CYP3A4. Hence, erythromycin should not be administered to people using these drugs, or drugs that also prolong the QT interval. Other examples include terfenadine (Seldane, Seldane-D), astemizole (Hismanal), cisapride (Propulsid, withdrawn in many countries for prolonging the QT time) and pimozide (Orap). Theophylline, which is used mostly in asthma, is also contraindicated. Erythromycin and doxycycline can have a synergistic effect when combined and kill bacteria ("E. coli)" with a higher potency than the sum of the two drugs together. This synergistic relationship is only temporary. After approximately 72 hours, the relationship shifts to become antagonistic, whereby a 50/50 combination of the two drugs kills less bacteria than if the two drugs were administered separately. It may alter the effectiveness of combined oral contraceptive pills because of its effect on the gut flora. A review found that when erythromycin was given with certain oral contraceptives, there was an increase in the maximum serum concentrations and AUC of estradiol and dienogest. Erythromycin is an inhibitor of the cytochrome P450 system, which means it can have a rapid effect on levels of other drugs metabolised by this system, e.g., warfarin. Erythromycin displays bacteriostatic activity or inhibits growth of bacteria, especially at higher concentrations. By binding to the 50s subunit of the bacterial rRNA complex, protein synthesis and subsequent structure and function processes critical for life or replication are inhibited. Erythromycin interferes with aminoacyl translocation, preventing the transfer of the tRNA bound at the A site of the rRNA complex to the P site of the rRNA complex. Without this translocation, the A site remains occupied, thus the addition of an incoming tRNA and its attached amino acid to the nascent polypeptide chain is inhibited. This interferes with the production of functionally useful proteins, which is the basis of this antimicrobial action. Erythromycin increases gut motility by binding to Motillin, thus it is a Motillin receptor agonist in addition to its antimicrobial properties. Erythromycin is easily inactivated by gastric acid; therefore, all orally administered formulations are given as either enteric-coated or more-stable salts or esters, such as erythromycin ethylsuccinate. Erythromycin is very rapidly absorbed, and diffuses into most tissues and phagocytes. Due to the high concentration in phagocytes, erythromycin is actively transported to the site of infection, where, during active phagocytosis, large concentrations of erythromycin are released. Most of erythromycin is metabolised by demethylation in the liver by the hepatic enzyme CYP3A4. Its main elimination route is in the bile with little renal excretion, 2%-15% unchanged drug. Erythromycin's elimination half-life ranges between 1.5 and 2.0 hours and is between 5 and 6 hours in patients with end-stage renal disease. Erythromycin levels peak in the serum 4 hours after dosing; ethylsuccinate peaks 0.5-2.5 hours after dosing, but can be delayed if digested with food. Erythromycin crosses the placenta and enters breast milk. The American Association of Pediatrics determined erythromycin is safe to take while breastfeeding. Absorption in pregnant patients has been shown to be variable, frequently resulting in levels lower than in nonpregnant patients. Standard-grade erythromycin is primarily composed of four related compounds known as erythromycins A, B, C, and D. Each of these compounds can be present in varying amounts and can differ by lot. Erythromycin A has been found to have the most antibacterial activity, followed by erythromycin B. Erythromycins C and D are about half as active as erythromycin A. Some of these related compounds have been purified and can be studied and researched individually. Over the three decades after the discovery of erythromycin A and its activity as an antimicrobial, many attempts were made to synthesize it in the laboratory. The presence of 10 stereogenic carbons and several points of distinct substitution has made the total synthesis of erythromycin A a formidable task. Complete syntheses of erythromycins’ related structures and precursors such as 6-deoxyerythronolide B have been accomplished, giving way to possible syntheses of different erythromycins and other macrolide antimicrobials. Woodward successfully completed the synthesis of erythromycin A. In 1949 Abelardo B. Aguilar, a Filipino scientist, sent some soil samples to his employer Eli Lilly. Eli Lilly’s research team, led by J. M. McGuire, managed to isolate erythromycin from the metabolic products of a strain of "Streptomyces erythreus" (designation changed to "Saccharopolyspora erythraea") found in the samples. Lilly filed for patent protection on the compound which was granted in 1953. The product was launched commercially in 1952 under the brand name Ilosone (after the Philippine region of Iloilo where it was originally collected). Erythromycin was formerly also called Ilotycin. In 1981, Nobel laureate (1965 in chemistry) and professor of chemistry at Harvard University Robert B. Woodward (posthumously), along with a large number of members from his research group, reported the first stereocontrolled asymmetric chemical synthesis of erythromycin A. The antibiotic clarithromycin was invented by scientists at the Japanese drug company Taisho Pharmaceutical in the 1970s as a result of their efforts to overcome the acid instability of erythromycin. Scientists at Chugai Pharmaceuticals discovered an erythromycin-derived motilin agonist called mitemcinal that is believed to have strong prokinetic properties (similar to erythromycin) but lacking antibiotic properties. Erythromycin is commonly used off-label for gastric motility indications such as gastroparesis. If mitemcinal can be shown to be an effective prokinetic agent, it would represent a significant advance in the gastrointestinal field, as treatment with this drug would not carry the risk of unintentional selection for antibiotic-resistant bacteria. It is available as a generic medication and was inexpensive for many years. The wholesale price was between 0.03 and 0.06 USD per pill. In the United States in 2014 the price increased to seven dollars per tablet. The price of Erythromycin rose three times between 2010 and 2015, from 24 cents per tablet in 2010 to $8.96 in 2015. In 2017, a Kaiser Health News study found that the per-unit cost of dozens of generics doubled or even tripled from 2015 to 2016, increasing spending by the Medicaid program. Due to price increases by drug manufacturers, Medicaid paid on average $2,685,330 more for Erythromycin in 2016 compared to 2015 (not including rebates). By 2018, generic drug prices had climbed another 5% on average. Brand names include Robimycin, E-Mycin, E.E.S. Granules, E.E.S.-200, E.E.S.-400, E.E.S.-400 Filmtab, Erymax, Ery-Tab, Eryc, Ranbaxy, Erypar, EryPed, Eryped 200, Eryped 400, Erythrocin Stearate Filmtab, Erythrocot, E-Base, Erythroped, Ilosone, MY-E, Pediamycin, Zineryt, Abboticin, Abboticin-ES, Erycin, PCE Dispertab, Stiemycine, Acnasol, and Tiloryth. Erythromycin/tretinoin, a combination of tretinoin and the antibiotic erythromycin
https://en.wikipedia.org/wiki?curid=10090
Environmental law Environmental law is a collective term encompassing aspects of the law that provide protection to the environment. A related but distinct set of regulatory regimes, now strongly influenced by environmental legal principles, focus on the management of specific natural resources, such as forests, minerals, or fisheries. Other areas, such as environmental impact assessment, may not fit neatly into either category, but are nonetheless important components of environmental law. Early examples of legal enactments designed to consciously preserve the environment, for its own sake or human enjoyment, are found throughout history. In the common law, the primary protection was found in the law of nuisance, but this only allowed for private actions for damages or injunctions if there was harm to land. Thus, smells emanating from pig sties, strict liability against dumping rubbish, or damage from exploding dams. Private enforcement, however, was limited and found to be woefully inadequate to deal with major environmental threats, particularly threats to common resources. During the "Great Stink" of 1858, the dumping of sewerage into the River Thames began to smell so ghastly in the summer heat that Parliament had to be evacuated. Ironically, the Metropolitan Commission of Sewers Act 1848 had allowed the Metropolitan Commission for Sewers to close cesspits around the city in an attempt to "clean up" but this simply led people to pollute the river. In 19 days, Parliament passed a further Act to build the London sewerage system. London also suffered from terrible air pollution, and this culminated in the "Great Smog" of 1952, which in turn triggered its own legislative response: the Clean Air Act 1956. The basic regulatory structure was to set limits on emissions for households and business (particularly burning coal) while an inspectorate would enforce compliance. Notwithstanding early analogues, the concept of "environmental law" as a separate and distinct body of law is a twentieth-century development. The recognition that the natural environment was fragile and in need of special legal protections, the translation of that recognition into legal structures, the development of those structures into a larger body of "environmental law," and the strong influence of environmental law on natural resource laws, did not occur until about the 1960s. At that time, numerous influences - including a growing awareness of the unity and fragility of the biosphere; increased public concern over the impact of industrial activity on natural resources and human health; the increasing strength of the regulatory state; and more broadly the advent and success of environmentalism as a political movement - coalesced to produce a huge new body of law in a relatively short period of time. While the modern history of environmental law is one of continuing controversy, by the end of the twentieth century environmental law had been established as a component of the legal landscape in all developed nations of the world, many developing ones, and the larger project of international law. Air quality laws Water quality laws Waste management laws Environmental cleanup laws Chemical safety laws govern the use of chemicals in human activities, particularly man-made chemicals in modern industrial applications. As contrasted with media-oriented environmental laws (e.g., air or water quality laws), chemical control laws seek to manage the (potential) pollutants themselves. Regulatory efforts include banning specific chemical constituents in consumer products (e.g., Bisphenol A in plastic bottles), and regulating pesticides. Environmental impact assessment Water resources laws govern the ownership and use of water resources, including surface water and ground water. Regulatory areas may include water conservation, use restrictions, and ownership regimes. Mineral resource laws cover Wildlife laws govern the potential impact of human activity on wild animals, whether directly on individuals or populations, or indirectly via habitat degradation. Similar laws may operate to protect plant species. Such laws may be enacted entirely to protect biodiversity, or as a means for protecting species deemed important for other reasons. Regulatory efforts may including the creation of special conservation statuses, prohibitions on killing, harming, or disturbing protected species, efforts to induce and support species recovery, establishment of wildlife refuges to support conservation, and prohibitions on trafficking in species or animal parts to combat poaching. Fish and game laws regulate the right to pursue and take or kill certain kinds of fish and wild animal (game). Such laws may restrict the days to harvest fish or game, the number of animals caught per person, the species harvested, or the weapons or fishing gear used. Such laws may seek to balance dueling needs for preservation and harvest and to manage both environment and populations of fish and game. Game laws can provide a legal structure to collect license fees and other money which is used to fund conservation efforts as well as to obtain harvest information used in wildlife management practice. Environmental law has developed in response to emerging awareness of and concern over issues impacting the entire world. While laws have developed piecemeal and for a variety of reasons, some effort has gone into identifying key concepts and guiding principles common to environmental law as a whole. The principles discussed below are not an exhaustive list and are not universally recognized or accepted. Nonetheless, they represent important principles for the understanding of environmental law around the world. Defined by the United Nations Environment Programme as "development that meets the needs of the present without compromising the ability of future generations to meet their own needs," sustainable development may be considered together with the concepts of "integration" (development cannot be considered in isolation from sustainability) and "interdependence" (social and economic development, and environmental protection, are interdependent). Laws mandating environmental impact assessment and requiring or encouraging development to minimize environmental impacts may be assessed against this principle. The modern concept of sustainable development was a topic of discussion at the 1972 United Nations Conference on the Human Environment (Stockholm Conference), and the driving force behind the 1983 World Commission on Environment and Development (WCED, or Bruntland Commission). In 1992, the first UN Earth Summit resulted in the Rio Declaration, Principle 3 of which reads: "The right to development must be fulfilled so as to equitably meet developmental and environmental needs of present and future generations." Sustainable development has been a core concept of international environmental discussion ever since, including at the World Summit on Sustainable Development (Earth Summit 2002), and the United Nations Conference on Sustainable Development (Earth Summit 2012, or Rio+20). Defined by UNEP to include intergenerational equity - "the right of future generations to enjoy a fair level of the common patrimony" - and intragenerational equity - "the right of all people within the current generation to fair access to the current generation's entitlement to the Earth's natural resources" - environmental equity considers the present generation under an obligation to account for long-term impacts of activities, and to act to sustain the global environment and resource base for future generations. Pollution control and resource management laws may be assessed against this principle. Defined in the international law context as an obligation to protect one's own environment, and to prevent damage to neighboring environments, UNEP considers transboundary responsibility at the international level as a potential limitation on the rights of the sovereign state. Laws that act to limit externalities imposed upon human health and the environment may be assessed against this principle. Identified as essential conditions for "accountable governments... industrial concerns," and organizations generally, public participation and transparency are presented by UNEP as requiring "effective protection of the human right to hold and express opinions and to seek, receive and impart ideas... a right of access to appropriate, comprehensible and timely information held by governments and industrial concerns on economic and social policies regarding the sustainable use of natural resources and the protection of the environment, without imposing undue financial burdens upon the applicants and with adequate protection of privacy and business confidentiality," and "effective judicial and administrative proceedings." These principles are present in environmental impact assessment, laws requiring publication and access to relevant environmental data, and administrative procedure. One of the most commonly encountered and controversial principles of environmental law, the Rio Declaration formulated the precautionary principle as follows: The principle may play a role in any debate over the need for environmental regulation. The polluter pays principle stands for the idea that "the environmental costs of economic activities, including the cost of preventing potential harm, should be internalized rather than imposed upon society at large." All issues related to responsibility for cost for environmental remediation and compliance with pollution control regulations involve this principle. Environmental law is a continuing source of controversy. Debates over the necessity, fairness, and cost of environmental regulation are ongoing, as well as regarding the appropriateness of regulations vs. market solutions to achieve even agreed-upon ends. Allegations of scientific uncertainty fuel the ongoing debate over greenhouse gas regulation, and are a major factor in debates over whether to ban particular pesticides. In cases where the science is well-settled, it is not unusual to find that corporations intentionally hide or distort the facts, or sow confusion. It is very common for regulated industry to argue against environmental regulation on the basis of cost. Difficulties arise in performing cost-benefit analysis of environmental issues. It is difficult to quantify the value of an environmental value such as a healthy ecosystem, clean air, or species diversity. Many environmentalists' response to pitting economy vs. ecology is summed up by former Senator and founder of Earth Day Gaylord Nelson, "The economy is a wholly owned subsidiary of the environment, not the other way around." Furthermore, environmental issues are seen by many as having an ethical or moral dimension, which would transcend financial cost. Even so, there are some efforts underway to systemically recognize environmental costs and assets, and account for them properly in economic terms. While affected industries spark controversy in fighting regulation, there are also many environmentalists and public interest groups who believe that current regulations are inadequate, and advocate for stronger protection. Environmental law conferences - such as the annual Public Interest Environmental Law Conference in Eugene, Oregon - typically have this focus, also connecting environmental law with class, race, and other issues. An additional debate is to what extent environmental laws are fair to all regulated parties. For instance, researchers Preston Teeter and Jorgen Sandberg highlight how smaller organizations can often incur disproportionately larger costs as a result of environmental regulations, which can ultimately create an additional barrier to entry for new firms, thus stifling competition and innovation. Global and regional environmental issues are increasingly the subject of international law. Debates over environmental concerns implicate core principles of international law and have been the subject of numerous international agreements and declarations. Customary international law is an important source of international environmental law. These are the norms and rules that countries follow as a matter of custom and they are so prevalent that they bind all states in the world. When a principle becomes customary law is not clear cut and many arguments are put forward by states not wishing to be bound. Examples of customary international law relevant to the environment include the duty to warn other states promptly about icons of an environmental nature and environmental damages to which another state or states may be exposed, and Principle 21 of the Stockholm Declaration ('good neighbourliness' or sic utere). Numerous legally binding international agreements encompass a wide variety of issue-areas, from terrestrial, marine and atmospheric pollution through to wildlife and biodiversity protection. International environmental agreements are generally multilateral (or sometimes bilateral) treaties (a.k.a. convention, agreement, protocol, etc.). Protocols are subsidiary agreements built from a primary treaty. They exist in many areas of international law but are especially useful in the environmental field, where they may be used to regularly incorporate recent scientific knowledge. They also permit countries to reach agreement on a framework that would be contentious if every detail were to be agreed upon in advance. The most widely known protocol in international environmental law is the Kyoto Protocol, which followed from the United Nations Framework Convention on Climate Change. While the bodies that proposed, argued, agreed upon and ultimately adopted existing international agreements vary according to each agreement, certain conferences, including 1972's United Nations Conference on the Human Environment, 1983's World Commission on Environment and Development, 1992's United Nations Conference on Environment and Development and 2002's World Summit on Sustainable Development have been particularly important. Multilateral environmental agreements sometimes create an International Organization, Institution or Body responsible for implementing the agreement. Major examples are the Convention on International Trade in Endangered Species of Wild Fauna and Flora (CITES) and the International Union for Conservation of Nature (IUCN). International environmental law also includes the opinions of international courts and tribunals. While there are few and they have limited authority, the decisions carry much weight with legal commentators and are quite influential on the development of international environmental law. One of the biggest challenges in international decisions is to determine an adequate compensation for environmental damages. The courts include the International Court of Justice (ICJ), the international Tribunal for the Law of the Sea (ITLOS), the European Court of Justice, European Court of Human Rights and other regional treaty tribunals. According to the International Network for Environmental Compliance and Enforcement (INECE), the major environmental issues in Africa are “drought and flooding, air pollution, deforestation, loss of biodiversity, freshwater availability, degradation of soil and vegetation, and widespread poverty.” The U.S. Environmental Protection Agency (EPA) is focused on the “growing urban and industrial pollution, water quality, electronic waste and indoor air from cookstoves.” They hope to provide enough aid on concerns regarding pollution before their impacts contaminate the African environment as well as the global environment. By doing so, they intend to “protect human health, particularly vulnerable populations such as children and the poor.” In order to accomplish these goals in Africa, EPA programs are focused on strengthening the ability to enforce environmental laws as well as public compliance to them. Other programs work on developing stronger environmental laws, regulations, and standards. The Asian Environmental Compliance and Enforcement Network (AECEN) is an agreement between 16 Asian countries dedicated to improving cooperation with environmental laws in Asia. These countries include Cambodia, China, Indonesia, India, Maldives, Japan, Korea, Malaysia, Nepal, Philippines, Pakistan, Singapore, Sri Lanka, Thailand, Vietnam, and Lao PDR. The European Union issues secondary legislation on environmental issues that are valid throughout the EU (so called regulations) and many directives that must be implemented into national legislation from the 28 member states (national states). Examples are the Regulation (EC) No. 338/97 on the implementation of CITES; or the Natura 2000 network the centerpiece for nature & biodiversity policy, encompassing the bird Directive (79/409/EEC/ changed to 2009/147/EC)and the habitats directive (92/43/EEC). Which are made up of multiple SACs (Special Areas of Conservation, linked to the habitats directive) & SPAs (Special Protected Areas, linked to the bird directive), throughout Europe. EU legislation is ruled in Article 249 Treaty for the Functioning of the European Union (TFEU). Topics for common EU legislation are: The U.S. Environmental Protection Agency is working with countries in the Middle East to improve “environmental governance, water pollution and water security, clean fuels and vehicles, public participation, and pollution prevention.” The main concerns on environmental issues in the Oceanic Region are “illegal releases of air and water pollutants, illegal logging/timber trade, illegal shipment of hazardous wastes, including e-waste and ships slated for destruction, and insufficient institutional structure/lack of enforcement capacity”. The Secretariat of the Pacific Regional Environmental Programme (SPREP) is an international organization between Australia, the Cook Islands, FMS, Fiji, France, Kiribati, Marshall Islands, Nauru, New Zealand, Niue, Palau, PNG, Samoa, Solomon Island, Tonga, Tuvalu, USA, and Vanuatu. The SPREP was established in order to provide assistance in improving and protecting the environment as well as assure sustainable development for future generations. The Environment Protection and Biodiversity Conservation Act 1999 is the center piece of environmental legislation in the Australian Government. It sets up the “legal framework to protect and manage nationally and internationally important flora, fauna, ecological communities and heritage places”. It also focuses on protecting world heritage properties, national heritage properties, wetlands of international importance, nationally threatened species and ecological communities, migratory species, Commonwealth marine areas, Great Barrier Reef Marine Park, and the environment surrounding nuclear activities. "Commonwealth v Tasmania" (1983), also known as the "Tasmanian Dam Case", is the most influential case for Australian environmental law. The Brazilian government created the Ministry of Environment in 1992 in order to develop better strategies of protecting the environment, use natural resources sustainably, and enforce public environmental policies. The Ministry of Environment has authority over policies involving environment, water resources, preservation, and environmental programs involving the Amazon. The Department of the Environment Act establishes the Department of the Environment in the Canadian government as well as the position Minister of the Environment. Their duties include “the preservation and enhancement of the quality of the natural environment, including water, air and soil quality; renewable resources, including migratory birds and other non-domestic flora and fauna; water; meteorology;" The Environmental Protection Act is the main piece of Canadian environmental legislation that was put into place March 31, 2000. The Act focuses on “respecting pollution prevention and the protection of the environment and human health in order to contribute to sustainable development." Other principle federal statutes include the Canadian Environmental Assessment Act, and the Species at Risk Act. When provincial and federal legislation are in conflict federal legislation takes precedence, that being said individual provinces can have their own legislation such as Ontario's Environmental Bill of Rights, and Clean Water Act. According to the U.S. Environmental Protection Agency, "China has been working with great determination in recent years to develop, implement, and enforce a solid environmental law framework. Chinese officials face critical challenges in effectively implementing the laws, clarifying the roles of their national and provincial governments, and strengthening the operation of their legal system." Explosive economic and industrial growth in China has led to significant environmental degradation, and China is currently in the process of developing more stringent legal controls. The harmonization of Chinese society and the natural environment is billed as a rising policy priority. With the enactment of the 2008 Constitution, Ecuador became the first country in the world to codify the Rights of Nature. The Constitution, specifically Articles 10 and 71–74, recognizes the inalienable rights of ecosystems to exist and flourish, gives people the authority to petition on the behalf of ecosystems, and requires the government to remedy violations of these rights. The rights approach is a break away from traditional environmental regulatory systems, which regard nature as property and legalize and manage degradation of the environment rather than prevent it. The Rights of Nature articles in Ecuador's constitution are part of a reaction to a combination of political, economic, and social phenomena. Ecuador's abusive past with the oil industry, most famously the class-action litigation against Chevron, and the failure of an extraction-based economy and neoliberal reforms to bring economic prosperity to the region has resulted in the election of a New Leftist regime, led by President Rafael Correa, and sparked a demand for new approaches to development. In conjunction with this need, the principle of "Buen Vivir," or good living—focused on social, environmental and spiritual wealth versus material wealth—gained popularity among citizens and was incorporated into the new constitution. The influence of indigenous groups, from whom the concept of "Buen Vivir" originates, in the forming of the constitutional ideals also facilitated the incorporation of the Rights of Nature as a basic tenet of their culture and conceptualization of "Buen Vivir." The Environmental Protection Law outlines the responsibilities of the Egyptian government to “preparation of draft legislation and decrees pertinent to environmental management, collection of data both nationally and internationally on the state of the environment, preparation of periodical reports and studies on the state of the environment, formulation of the national plan and its projects, preparation of environmental profiles for new and urban areas, and setting of standards to be used in planning for their development, and preparation of an annual report on the state of the environment to be prepared to the President." In India, Environmental law is governed by the Environment Protection Act, 1986. This act is enforced by the Central Pollution Control Board and the numerous State Pollution Control Boards. Apart from this, there are also individual legislations specifically enacted for the protection of Water, Air, Wildlife, etc. Such legislations include :- The Basic Environmental Law is the basic structure of Japan's environmental policies replacing the Basic Law for Environmental Pollution Control and the Nature Conservation Law. The updated law aims to address “global environmental problems, urban pollution by everyday life, loss of accessible natural environment in urban areas and degrading environmental protection capacity in forests and farmlands.” The three basic environmental principles that the Basic Environmental Law follows are “the blessings of the environment should be enjoyed by the present generation and succeeded to the future generations, a sustainable society should be created where environmental loads by human activities are minimized, and Japan should contribute actively to global environmental conservation through international cooperation.” From these principles, the Japanese government have established policies such as “environmental consideration in policy formulation, establishment of the Basic Environment Plan which describes the directions of long-term environmental policy, environmental impact assessment for development projects, economic measures to encourage activities for reducing environmental load, improvement of social infrastructure such as sewerage system, transport facilities etc., promotion of environmental activities by corporations, citizens and NGOs, environmental education, and provision of information, promotion of science and technology." The Ministry for the Environment and Office of the Parliamentary Commissioner for the Environment were established by the Environment Act 1986. These positions are responsible for advising the Minister on all areas of environmental legislation. A common theme of New Zealand's environmental legislation is sustainably managing natural and physical resources, fisheries, and forests. The Resource Management Act 1991 is the main piece of environmental legislation that outlines the government's strategy to managing the “environment, including air, water soil, biodiversity, the coastal environment, noise, subdivision, and land use planning in general.” The Ministry of Natural Resources and Environment of the Russian Federation makes regulation regarding “conservation of natural resources, including the subsoil, water bodies, forests located in designated conservation areas, fauna and their habitat, in the field of hunting, hydrometeorology and related areas, environmental monitoring and pollution control, including radiation monitoring and control, and functions of public environmental policy making and implementation and statutory regulation." Vietnam is currently working with the U.S. Environmental Protection Agency on dioxin remediation and technical assistance in order to lower methane emissions. In March 2002, the U.S and Vietnam signed the U.S.-Vietnam Memorandum of Understanding on Research on Human Health and the Environmental Effects of Agent Orange/Dioxin.
https://en.wikipedia.org/wiki?curid=10091
Eurostar Eurostar is an international high-speed railway service connecting London with Amsterdam, Brussels, Bourg-Saint-Maurice, Disneyland Paris, Lille, Fréthun, Paris, and Rotterdam. All its trains travel through the Channel Tunnel between the United Kingdom and France, owned and operated separately by Getlink. The London terminus is St Pancras International, the other British calling points are Ebbsfleet International and Ashford International in Kent. Intermediate calling points in France are Calais-Fréthun and Lille-Europe, with trains to Paris terminating at Gare du Nord. Trains to Belgium terminate at Midi/Zuid station in Brussels. The only intermediate calling station in the Netherlands is Rotterdam Centraal and trains terminate at Amsterdam Centraal. There are direct services from London to Disneyland Paris at Marne-la-Vallée – Chessy, direct services to southern France (Lyon, Avignon and Marseille), and seasonal direct services to the French Alps in winter (December to April). The service is operated by eighteen-coach Class 373/1 trains and sixteen-coach trains which run at up to on a network of high-speed lines. The LGV Nord line in France opened before Eurostar services began in 1994, and newer lines enabling faster journeys were added later—HSL 1 in Belgium and High Speed 1 in south-east England. The French and Belgian parts of the network are shared with Paris–Brussels Thalys services and TGV trains. In the United Kingdom the two-stage Channel Tunnel Rail Link project was completed on 14 November 2007 and renamed "High Speed 1", when the London terminus of Eurostar moved from London Waterloo International to St Pancras International. Until 2010, Eurostar was operated jointly by the national railway companies of France and Belgium, SNCF and SNCB/NMBS, and Eurostar (UK) Ltd (EUKL), a subsidiary of London and Continental Railways (LCR) that owned the high-speed infrastructure and stations on the British side. On 1 September 2010, Eurostar was incorporated as a single corporate entity, Eurostar International Limited (EIL). The history of Eurostar can be traced to the 1986 choice of a rail tunnel to provide a cross-channel link between Britain and France. A previous attempt to construct a tunnel between the two nations had begun in 1974, but was quickly aborted. Construction began on a new basis in 1988. Eurotunnel was created to manage and own the tunnel, which was finished in 1993, the official opening taking place on 6 May 1994. In addition to the tunnel's shuttle trains carrying cars and lorries between Folkestone and Calais, the decision to build a railway tunnel opened up the possibility of through passenger and freight train services between places further afield. British Rail and SNCF contracted with Eurotunnel to use half the tunnel's capacity for this purpose. In 1987, Britain, France and Belgium set up an International Project Group to specify a train to provide an international high-speed passenger service through the tunnel. France had been operating high-speed TGV services since 1981, and had begun construction of a new high-speed line between Paris and the Channel Tunnel, LGV Nord; French TGV technology was chosen as the basis for the new trains. An order for 30 trainsets, to be manufactured in France but with some British and Belgian components, was placed in December 1989. On 20 June 1993, the first Eurostar test train travelled through the tunnel to the UK. Various technical difficulties in running the new trains on British tracks were quickly overcome. On 14 November 1994, Eurostar services began running from Waterloo International station in London, to Gare du Nord in Paris, and Brussels-South railway station in Brussels. The train service started with a limited "Discovery" service; the full daily service started from 28 May 1995. In 1995, Eurostar was achieving an average end-to-end speed of from London to Paris. On 8 January 1996, Eurostar launched services from a second railway station in the UK when Ashford International was opened. On 23 September 2003, passenger services began running on the first completed section of High Speed 1. Following a high-profile glamorous opening ceremony and a large advertising campaign, on 14 November 2007, Eurostar services in London transferred from Waterloo to the extended and extensively refurbished St Pancras International. The Channel Tunnel used by Eurostar services holds the record for having the longest underwater section of any tunnel in the world, and it is the second-longest railway tunnel in the world. On 30 July 2003, a Eurostar train set a new British speed record of on the first section of the "High Speed 1" railway between the Channel Tunnel, and Fawkham Junction in north Kent, two months before official public services began running. On 16 May 2006, Eurostar set a new record for the longest non-stop high-speed journey, a distance of from London to Cannes taking 7 hours 25 minutes. On 4 September 2007, a record-breaking train left Paris Gare du Nord at 10:44 (09:44 BST) and reached London St Pancras in 2 hours 3 minutes 39 seconds; carrying journalists and railway workers. This record trip was also the first passenger-carrying arrival to the new St Pancras International station. On 20 September 2007, Eurostar broke another record when it completed the journey from Brussels to London in 1 hour, 43 minutes. The original proposals for Eurostar included direct services to Paris and Brussels from cities north of London (NoL): Manchester via Birmingham on the West Coast Main Line and on the East Coast Main Line Leeds and Glasgow via Edinburgh, Newcastle and York. Seven 14-coach "North of London" Eurostar trains for these Regional Eurostar services were built, but these services never came to fruition. Predicted journey times of almost nine hours for Glasgow to Paris at the time of growth of low-cost air travel during the 1990s made the plans commercially unviable against the cheaper and quicker airlines. Other reasons that have been suggested for these services having never been run were both government policies and the disruptive privatisation of British Rail. Three of the Regional Eurostar units were leased by Great North Eastern Railway (GNER) to increase domestic services from London King's Cross to York and later Leeds. The lease expired in December 2005, and most of the NoL sets have since been transferred to SNCF for TGV services in northern France. An international Nightstar sleeper train was also planned; this would have travelled the same routes as Regional Eurostar, plus the Great Western Main Line to Cardiff. These were also deemed commercially unviable, and the scheme was abandoned with no services ever operated. In 2000, the coaches were sold to Via Rail in Canada. Ashford International station was the original station for Eurostar services in Kent. Once Ebbsfleet International railway station, also designed to serve the Kent region, had opened, only three trains a day to Paris and one to Disneyland Paris called at Ashford for a considerable amount of time. There were fears that services at Ashford International might be further reduced or withdrawn altogether as Eurostar planned to make Ebbsfleet the new regional hub instead. However, after a period during which no Brussels trains served the station, to the dissatisfaction of the local communities, Eurostar re-introduced a single daily Ashford-Brussels service on 23 February 2009. On 1 September 2010, Eurostar was incorporated as a single corporate entity, Eurostar International Limited (EIL), replacing the joint operation between EUKL, SNCF and SNCB/NMBS. EIL is owned by SNCF (55%), Caisse de dépôt et placement du Québec (CDPQ) (30%), Hermes Infrastructure (10%) and SNCB (5%). In June 2014, the UK shareholding in Eurostar International Limited was transferred from London and Continental Railways / Department for Transport to HM Treasury. That October, it was announced that the UK government planned to raise £300 million by selling its stake. In March 2015, the UK government announced that it would sell its 40% share to an Anglo-Canadian consortium made up of the Caisse and Hermes Infrastructure. The sale was completed in May 2015. In 2015, Eurostar threatened to require that cyclists dismantle bicycles before they could be transported on the trains. Following criticism from Boris Johnson and cycling groups, Eurostar reversed the edict. By March 2016, onboard entertainment was provided by GoMedia, including Wi-Fi connectivity and up to 300 hours of movies and television kept on the train's servers and accessed using the passenger's own devices: mobile phones, tablets, laptops etc. A tracker app allows customers to see where they are. LGV Nord is a French -long high-speed rail line that connects Paris to the Belgian border and the Channel Tunnel via Lille. It opened in 1993. Its extensions to Belgium and towards Paris, as well as connecting to the Channel Tunnel, have made LGV Nord a part of every Eurostar journey undertaken. A Belgian high-speed line, HSL 1, was added to the end of LGV Nord, at the Belgian border, in 1997. Of all French high-speed lines, LGV Nord sees the widest variety of high-speed rolling stock and is quite busy; a proposed cut-off bypassing Lille, which would reduce Eurostar journey times to Paris, is called LGV Picardie. The Channel Tunnel is the most important part of the route as it is the only rail connection between Great Britain and the European mainland. It joins LGV Nord in France with High Speed One in Britain. Tunnelling began in 1988, and the tunnel was officially opened by British sovereign Queen Elizabeth II and the French President François Mitterrand at a ceremony in Calais on 6 May 1994. It is owned by Getlink, which charges a significant toll to Eurostar for its use. In 1996, the American Society of Civil Engineers named the tunnel as one of the Seven Wonders of the Modern World. Along the current route of the Eurostar service, line speeds are except within the Channel Tunnel, where a reduced speed of applies for safety reasons. Since the launch of Eurostar services, severe disruptions and cancellations have been caused by fires breaking out within the Channel Tunnel, such as in 1996, 2006 (minor), 2008 and 2015. Until the opening on 2 June 1996, of the first phase of the Belgian high speed line, Eurostar trains were routed via the Belgian railway line 94. The Eurostar routes still use the line as a diversion if engineering works are taking place on HSL1, depending where it is. The 06:13 from London St Pancras to Brussels still uses the line as a diversion to bypass the peak time disruptions on HSL1 due to the extra TGV services from Brussels for the commuters. After 2 June 1996, some Eurostars to Brussels were routed via the first phase of the Belgian High Speed line and the Belgian railway line 78 via Mons. Although this line is still as a diversion if HSL1 is doing engineering, also depending where the maintenance is taking place. Journey times between London and Brussels were improved when an Belgian high-speed line, HSL 1, opened on 14 December 1997. It links with LGV Nord on the border with France, allowing Eurostar trains heading to Brussels to make the transition between the two without having to reduce speed. A further four-minute improvement for London–Brussels trains was achieved in December 2006, with the opening of the Brussels South Viaduct. Linking the international platforms of Brussels-South railway station with the high-speed line, the viaduct separates Eurostar (and Thalys) from local services. High Speed 1 (HS1), formerly known as the Channel Tunnel Rail Link (CTRL), is a high-speed railway line running from London through Kent to the British end of the Channel Tunnel. It was built in two stages. The first section between the tunnel and Fawkham Junction in north Kent opened in September 2003, cutting London–Paris journey times by 21 minutes to 2 hours 35 minutes, and London–Brussels to 2 hours 20 minutes. On 14 November 2007, commercial services began over the whole of the new HS1 line. The redeveloped St Pancras International station became the new London terminus for all Eurostar services. The completion of High Speed 1 has brought the British part of Eurostar's route up to the same standards as the French and Belgian high-speed lines. Non-stop journey times were reduced by a further 20 minutes to 2 hours 15 minutes for London–Paris and 1 hour 51 minutes for London–Brussels. Eurostar offers up to fifteen weekday London – Paris services (nineteen on Fridays) including nine non-stop (thirteen on Fridays). There are also nine (ten on Friday) London–Brussels services, with two running non-stop (continuing to Amsterdam) and a further two calling at Lille only. Three services daily (two on weekends) operate to Amsterdam via Brussels and Rotterdam with one also calling at Lille. These return to Brussels only. In addition, there is a return trip from London to Marne-la-Vallée - Chessy for Disneyland Paris, which runs 4 times a week with increased frequency during school holidays. There are also seasonal services: an up-to-4-times-a-week service to Marseille via Lyon and Avignon in the summer; and, in the winter, "Snow trains", aimed at skiers, to Bourg-Saint-Maurice, Aime-la-Plagne and Moûtiers in the Alps; these run twice-weekly, one overnight and one during the daytime. Intermediate stations are Ebbsfleet International in northwest Kent, Ashford International in southeast Kent, and Calais-Fréthun and Lille-Europe in Nord-Pas-de-Calais. In February 2018, Eurostar announced the start of its long planned service from London to Amsterdam, with an initial two trains per day from April of that year running between St Pancras and Amsterdam Centraal. This launched as a one-way service, with return trains carrying passengers to Rotterdam and Brussels Midi/Zuid, making a 28-minute stop (which has not been deemed long enough to process UK-bound passengers) and then carrying different passengers from Brussels to London. Passengers travelling back must take Thalys services to Brussels Midi/Zuid where they can join the Eurostar. This is owing to the lack of facilities for juxtaposed controls by the UK Border Force at Amsterdam Centraal and Rotterdam Centraal. On 4 February 2020, the Dutch Minister of Infrastructure and Water Management, Cora van Nieuwenhuizen, and the UK Transport Secretary, Grant Shapps, announced that juxtaposed controls would be established at Amsterdam Centraal and Rotterdam Centraal. The direct train from Amsterdam was due to launch on 30 April 2020, and from Rotterdam on 18 May 2020. However, the launch has since been postponed due to the COVID-19 pandemic. Since 14 November 2007, all Eurostar trains have been routed via High Speed 1 to or from the redeveloped London terminus at St Pancras International, which at a cost of £800 million was extensively rebuilt and extended to cope with long Eurostar trains. It had been intended to retain some Eurostar services at Waterloo International, but this was ruled out on cost grounds. Completion of High Speed 1 has increased the potential number of trains serving London. Separation of Eurostar from British domestic services through Kent meant that timetabling was no longer affected by peak-hour restrictions. Eurostar's fares were significantly higher in its early years; the cheapest fare in 1994 was £99 return. In 2002, Eurostar was planning cheaper fares, an example of which was an offer of £50-day returns from London to Paris or Brussels. By March 2003, the cheapest fare from the UK was £59 return, available all year around. In June 2009 it was announced that one-way single fares would be available at £31 at the cheapest. Competition between Eurostar and airline services was a large factor in ticket prices being reduced from the initial levels. Business Premier fares also slightly undercut air fares on similar routes, targeted at regular business travellers. In 2009, Eurostar greatly increased its budget ticket availability to help maintain and grow its dominant market share. The Eurostar ticketing system is very complex, being distributed through no fewer than 48 individual sales systems. Eurostar is a member of the Amadeus CRS distribution system, making its tickets available alongside those of airlines worldwide. First class on Eurostar is called Business Premier; benefits include guaranteed faster checking-in and meals served at-seat, as well as the improved furnishings and interior of Business Premier carriages. The rebranding is part of Eurostar's marketing drive to attract more business professionals. Increasingly, business people in a group have been chartering private carriages as opposed to individual seats on the train. Without the operation of Regional Eurostar services using the North of London trainsets across the rest of Britain, Eurostar has developed its connections with other transport services instead, such as integrating effectively with traditional UK rail operators' schedules and routes, making it possible for passengers to use Eurostar as a quick connection to further destinations on the continent. All three main terminals used by the Eurostar service – St Pancras International, Paris Gare du Nord, and Brussels Midi/Zuid – are served by domestic trains and by local urban transport networks such as the London Underground and the Paris Metro. Standard Eurostar tickets no longer include free onward connections to or from any other station in Belgium: this is now available for a flat-rate supplement, currently £5.50. Eurostar has announced several partnerships with other rail services, most notably Thalys connections at Lille and Brussels for passengers to go beyond current Eurostar routes towards the Netherlands and Germany. In 2002, Eurostar initiated the Eurostar-Plus program, offering connecting tickets for onward journeys from Lille and Paris to dozens of destinations in France. Through fares are also available from 68 British towns and cities to destinations in France and Belgium. In May 2009 Eurostar announced that a formal connection to Switzerland had been established in a partnership between Eurostar and Lyria, which will operate TGV services from Lille to the Swiss Alps for Eurostar connection. In May 2019, Eurostar ended its agreement with Deutsche Bahn that allowed passengers to travel by train from the UK to Germany, Austria and Switzerland. Under the agreement passengers could travel on a single booking which made rescheduling easier. However, the direct tickets will no longer be sold from 9 November 2019. Because the UK is not part of the Schengen Area, and because the Netherlands, Belgium and France are not part of the Common Travel Area, all Eurostar passengers must go through border controls. Both the British Government and the Schengen governments concerned (Belgium, Netherlands and France) have legal obligations to check the travel documents of those entering their respective countries (as well as those leaving, in the case of Belgium and France). To allow passengers to walk off the train without arrival checks in most cases, juxtaposed controls ordinarily take place at the embarkation station. In order to comply with special UK legislation, there are full security checks similar to those at airports, scanning both bags and people's pockets. Security checks at Eurostar are comparable to those at a small airport and generally much quicker than at London Heathrow. The recommended check-in time is 30 minutes except for business class where it is 10 minutes. Eurostar passengers travelling inside the Schengen Area (Brussels–Lille and Brussels–Calais; Lille–Calais trips are not allowed to be performed by Eurostar) pass through a separate corridor in Brussels bypassing border checks, and enter the preallocated cars of the train, which is reserved for these passengers. This arrangement was set up after numerous illegal immigrants entered the UK without the right to do so, by buying a ticket from Brussels to Lille or Calais but remaining on the train until London – an issue exacerbated by Belgian police threatening to arrest UK Border Agency staff at Brussels-Midi if they tried to prevent passengers whom they suspected of attempting to exploit this loophole from boarding Eurostar trains. Travel from Calais or Lille towards Brussels and Amsterdam, from Brussels to Rotterdam and Amsterdam, and from Amsterdam and Rotterdam to Brussels has no border or security control; trains from Amsterdam are emptied at Brussels and passengers seeking to continue onwards towards London must alight and pass through security and border control. When the tripartite agreements were signed, the Belgian Government said that it had serious questions about the compatibility of this agreement with the Schengen Convention and the principle of free movement of people enshrined in various European Treaties. On 30 June 2009 Eurostar raised concerns at the UK House of Commons Home Affairs Select Committee that it was illegal under French law for the collection of information desired by the UK government under the e-Borders scheme, and they would be unable to cooperate. On the northbound Marne la Vallée-Chessy - London train, the security check and French passport check take place at Marne la Vallée-Chessy, while the UK passport check takes place at the UK arrival stations - this is the only route where passengers are not cleared by UK border officials before crossing the channel. On the northbound Marseille-London train, there is no facility for security or passport checks at the southern French stations, so passengers must leave the train at Lille-Europe, taking all their belongings with them, and undergo security and border checks there before rejoining the train which waits at the station for just over an hour. On several occasions, people have illegally tried to stow away on board the train, sometimes in large groups, trying to enter the UK; border monitoring and security is therefore extremely tight. Eurostar claims to have good and well-funded security measures. Eurostar's punctuality has fluctuated from year to year, but usually remains over 90%; in the first quarter of 1999, 89% of services operated were on time, and in the second quarter it reached 92%. Eurostar's best punctuality record was 97.35%, between 16 and 22 August 2004. In 2006, it was 92.7%, and in 2007, 91.5% were on time. In the first quarter of 2009, 96% of Eurostar services were punctual, compared with rival air routes' 76%. An advantage held by Eurostar is the convenience and speed of the service: with shorter check-in times than at most airports and hence quicker boarding and less queueing and high punctuality, it takes less time to travel between central London and central Paris by high-speed rail than by air. Eurostar now has a dominant share of the combined rail–air market on its routes to Paris and Brussels. In 2004, it had a 66% share of the London–Paris market, and a 59% share of the London–Brussels market. In 2007, it achieved record market shares of 71% for London–Paris and 65% for London–Brussels routes. Eurostar's passenger numbers initially failed to meet predictions. In 1996, London and Continental Railways forecast that passenger numbers would reach 21.4 million annually by 2004, but only 7.3 million was achieved. 82 million passengers used Waterloo International Station from its opening in 1994 to its closure in 2007. 2008 was a record year for Eurostar, with a 10.3% rise in passenger use, which was attributed to the use of High Speed 1 and the move to St Pancras. The following year, Eurostar saw an 11.5% fall in passenger numbers during the first three months of 2009, attributed to the 2008 Channel Tunnel fire and the 2009 recession. As a result of the poor economic conditions, Eurostar received state aid in May 2009 to cancel out some of the accumulated debt from the High Speed 1 construction programme. Later that year, during snowy conditions in the run-up to Christmas, thousands of passengers were left stranded as several trains broke down and many more were cancelled. In an independent review commissioned by Eurostar, the company came in for serious criticism about its handling of the incident and lack of plans for such a scenario. In 2006, the Department for Transport predicted that, by 2037, annual cross-channel passenger numbers would probably reach 16 million, considerably less optimistic than London and Continental Railways's original 1996 forecast. In 2007 Eurostar set a target of carrying 10 million passengers by 2010. The company cited several factors to support this objective, such as improved journey times, punctuality and station facilities. Passengers in general, it stated, are becoming increasingly aware of the environmental effects of air travel, and Eurostar services emit much less carbon dioxide. and that its remaining carbon emissions are now offset, making its services carbon neutral. Further expansion of the high-speed rail network in Europe, such as the HSL-Zuid line between Belgium and the Netherlands, continues to bring more destinations within rail-competitive range, giving Eurostar the possibility of opening up new services in future. The following chart presents the estimated number of passengers annually transported by the Eurostar service since 1995: Eurostar has been hailed as having set new standards in international rail travel and has won praise several times over, recognising its high standards. Eurostar won the Train Operator of the Year award in the HSBC Rail Awards for 2005. It was declared the Best Train Company in the joint Guardian/Observer Travel Awards 2008. Eurostar had previously struggled with its reputation and brand image. One commentator had defined the situation at the time as: By 2008, Eurostar's environmental credentials had become highly developed and promoted. In 2006 Eurostar's Environment Group was set up, with the aim of making changes in the Eurostar services' daily running to decrease the environmental impact, the organisation setting itself a target of reducing carbon emissions per passenger journey by 25% by 2012. Drivers are trained in techniques to achieve maximum energy efficiency, and lighting has been minimised; the provider of the bulk of the energy for the Channel Tunnel has been switched to nuclear power stations in France. Eurostar's current target is to reduce emissions by 35 percent per passenger journey by 2012, putting itself beyond the efforts of other railway companies in this field and thereby winning the 2007 Network Rail Efficiency Award. In the grand opening ceremony of St Pancras International, one of the Eurostar trains was given the name 'Tread Lightly', said to symbolise their smaller impact on the environment compared to planes. Since 2010, Eurostar has been owned by Eurostar International Limited (EIL), a company jointly owned by SNCF (55%), (CDPQ) (30%), Hermes Infrastructure (10%) and SNCB (5%). Eurostar is a member of Railteam, a marketing alliance formed in July 2007 of seven European high-speed rail operators. The alliance plans to allow tickets to be booked from one end of Europe to the other on a single website. In June 2009 London and Continental Railways, and the Eurostar UK operations they held ownership of, became fully nationalised by the UK government. In addition to its multiple unit fleet units, Eurostar operates a single Class 08 diesel shunter as the pilot at Temple Mills depot. Built between 1992 and 1996, Eurostar's fleet consists of 38 EMU trains, designated Class 373 in the United Kingdom and TGV TMST in France. The units have also been branded as the Eurostar e300 by Eurostar since 2015. There are two variants: The trains are essentially modified TGV sets, and can operate at up to on high-speed lines, and in the Channel Tunnel. It is possible to exceed the 300 km/h speed limit, but only with special permission from the safety authorities in the respective country. Speed limits in the Channel Tunnel are dictated by air-resistance, energy (heat) dissipation and the need to be used with other, slower trains. The trains were designed with Channel Tunnel safety in mind, and consist of two independent "half-sets" each with its own power car. In the event of a serious fire on board while travelling through the tunnel, passengers would be transferred into the undamaged half of the train, which would then be detached and driven out of the tunnel to safety. If the undamaged part were the rear half of the train, this would be driven by the Chef du Train, who is a fully authorised driver and occupies the rear driving cab while the train travels through the tunnel for this purpose. As 27 of the 31 Inter-Capital sets are sufficient to operate the service, four are currently used by SNCF for domestic TGV services; one of these regularly operates a Paris–Lille shuttle. The Eurostar logos have been removed from these sets, but the base colours of white, black, and yellow remain. SNCF's lease of the sets is scheduled to last until 2011, with an option for a further two years. Each train has a unique four-digit number starting with "3" (3xxx). This designates the train as a Mark 3 TGV (Mark 1 being SNCF TGV Sud-Est; Mark 2 being SNCF TGV Atlantique). The second digit denotes the country of ownership: In 2004–2005 the "Inter-Capital" sets still in daily use for international services were refurbished with a new interior designed by Philippe Starck. The original grey-yellow scheme in Standard class and grey-red of First/Premium First were replaced with a grey-brown look in Standard and grey-burnt-orange in First class. Power points were added to seats in First class and coaches 5 and 14 in Standard class. Premium First class was renamed BusinessPremier. In 2008, Eurostar announced that it would be carrying out a mid-life refurbishment of its Class 373 trains to allow the fleet to remain in service beyond 2020. This will include the 28 units making up the Eurostar fleet, but not the three Class 373/1 units used by SNCF or the seven Class 373/2 "North of London" sets. As part of the refurbishment, the Italian company Pininfarina was contracted to redesign the interiors, and The Yard Creative was selected to design the new buffet cars. On 11 May 2009 Eurostar revealed the new look for its first-class compartments. The first refurbished train was due in service in 2012, and Eurostar plans to have completed the entire process by 2014. On 13 November 2014 Eurostar announced the first refurbished trains would not re-enter the fleet until the 3rd or 4th quarter of 2015 due to delays at the completion centre. The last refurbished "e300" re-entered service in April 2019. In addition to the announced mid-life update of the existing Class 373 fleet, Eurostar in 2009 reportedly entered prequalification bids for eight new trainsets to be purchased. Any new trains would need to meet the same safety rules governing passage through the Channel Tunnel as the existing Class 373 fleet. The replacement to the Class 373 trains has been decided jointly between the French Transport Ministry and the UK Department for Transport. The new trains will be equipped to use the new ERTMS in-cab signalling system, due to be fitted to High Speed 1 around 2040. On 7 October 2010, it was reported that Eurostar had selected Siemens as preferred bidder to supply 10 Siemens Velaro e320 trainsets at a cost of €600 million (and a total investment of more than £700 million with the refurbishment of the existing fleet included) to operate an expanded route network, including services from London to Cologne and Amsterdam. These would be sixteen-car, long trainsets built to meet current Channel Tunnel regulations. The top speed will be and they will have 894–950 seats, unlike the current fleet built by the French company Alstom, which has a top speed of and a seating capacity of 750. Total traction power will be rated at 16 MW. The nomination of Siemens would see it break into the French high-speed market for the first time, as all French and French subsidiary high-speed operators use TGV derivatives produced by Alstom. Alstom attempted legal action to prevent Eurostar from acquiring German-built trains, claiming that the Siemens sets ordered would breach Channel Tunnel safety rules, but this was thrown out of court. Alstom said, after its High Court defeat, that it would "pursue alternative legal options to uphold its position". On 4 November 2010, the company lodged a complaint with the European Commission over the tendering process, which then asked the British government for "clarification". Alstom then announced it had started legal action against Eurostar, again in the High Court in London. In July 2011, the High Court rejected Alstom's claim that the tender process was "ineffective", and in April 2012 Alstom said it would call off pending court actions against Siemens. This effectively freed the way for Siemens to build the new Eurostar trains, the first of which were expected to enter service in late 2015. On 13 November 2014 Eurostar announced the purchase of an additional seven e320s for delivery in the second half of 2016. At the same time, Eurostar announced the first five e320s from the original order of ten would be available by December 2015, with the remaining five entering service by May 2016. Of the five sets ready by December 2015, three of them were planned to be used on London-Paris and London-Brussels routes. In 2005, the chief executive of Eurostar, Richard Brown, suggested that existing Eurostar trains could be replaced by double-deck trains similar to the TGV Duplex units when they are withdrawn from service. According to Brown, a double-deck fleet could carry 40 million passengers per year from Britain to Continental Europe, equivalent to adding an extra runway at a London airport. A number of technical incidents have affected Eurostar services over the years, but there has only been one major accident involving a service operated by Eurostar, a derailment in June 2000. Other incidents in the Channel Tunnel – such as the 1996 and 2008 Channel Tunnel fires – have affected Eurostar services but were not directly related to Eurostar's operations. However, the breakdowns in the tunnel, which resulted in cessation of service and inconvenience to thousands of passengers, in the run-up to Christmas 2009, proved a public-relations disaster. There have been several minor incidents with a few Eurostar services. In October 1994 there were teething problems relating to the start of operations. The first preview train, carrying 400 members of the press and media, was delayed for two hours by technical problems. On 29 May 2002 a Eurostar train was initially sent down a wrong line – towards London Victoria railway station instead of London Waterloo – causing the service to arrive 25 minutes late. A signalling error that led to the incorrect routeing was stated to have caused "no risk" as a result. On 11 April 2006, a house collapsed next to a railway line near London which caused Eurostar services to have to terminate and start from Ashford International instead of London Waterloo. Passengers waiting at Waterloo International were initially directed on to local trains towards Ashford leaving from the adjacent London Waterloo East railway station, until overcrowding occurred at Ashford. Approximately 1,000 passengers were trapped in darkness for several hours inside two Eurostar trains on the night of 19/20 February 1996. The trains stopped inside the tunnels due to electronic failures caused by snow and ice. Questions were raised at the time about the ability of the train and tunnel electronics to withstand the mix of snow, salt and ice which collect in the tunnels during periods of extreme cold. On 5 June 2000 a Eurostar train travelling from Paris to London derailed on the LGV Nord high-speed line while traveling at . Fourteen people were treated for light injuries or shock, with no fatalities or major injuries. The articulated nature of the trainset was credited with maintaining stability during the incident and all of the train stayed upright. The incident was caused by a traction link on the second bogie of the front power car coming loose, leading to components of the transmission system on that bogie impacting the track. The first departures from St Pancras on 14 November 2007 coincided with an open-ended strike by French rail unions as part of general strike actions over proposed public-sector pension reforms. The trains were operated by uninvolved British employees and service was not interrupted. On 23 September 2009 an overhead power line dropped on to a Class 373 train arriving at St Pancras station, activating a circuit breaker and delaying eleven other trains. Two days later, on 25 September 2009, electrical power via the overhead lines was lost on a section of high-speed line outside Lille, delaying passengers on two evening Eurostar-operated services. During the December 2009 European snowfall, four Eurostar trains broke down inside the Channel Tunnel, after leaving France, and one in Kent on 18 December. Although the trains had been winterised, the systems had not coped with the conditions. Over 2,000 passengers were stuck inside failed trains inside the tunnel, and over 75,000 had their services disrupted. All Eurostar services were cancelled from Saturday 19 December to Monday 21 December 2009. An independent review, published on 12 February 2010, was critical of the contingency plans in place for assisting passengers stranded by the delays, calling them "insufficient". On 7 January 2010 a Brussels-London train broke down in the Channel Tunnel, resulting in three other trains failing to complete their journeys. The cause of the failure was the onboard signalling system. Due to the severe weather, a limited service was operated in the next few days. On 15 February 2010, services between Brussels and London were interrupted following the Halle train collision, this time after the dedicated HSL 1 lines in the suburbs of the Belgian capital were blocked by debris from a serious train crash on the suburban commuter lines alongside. No efforts were made to reroute trains around the blockage; Eurostar instead terminated services to Brussels at Lille, directing passengers to continue their journey on local trains. Brussels services resumed on a limited scale on 22 February. On 21 February 2010 the 21:43 service from Paris to London broke down just outside Ashford International, stranding 740 passengers for several hours while a rescue train was called in. On 15 April 2010 air traffic in Western Europe closed because of the eruption of the Eyjafjallajökull volcano. Many travellers between the UK and the European mainland instead took the Eurostar train, all tickets between Brussels and London on 15 and 16 April being sold out within 3½ hours after the closure of British airspace. Between 15 and 20 April, Eurostar put on 33 additional trains and carried 165,000 passengers – 50,000 more than had been scheduled to travel during this period. On 20 December 2010, the Channel Tunnel was closed off for a day due to snowy weather. Eurostar Trains were suspended that day with thousands of passengers stranded in the run up to Christmas. On 17 October 2011 a man fell from the 17:04 service from London to Brussels as it passed through Westenhanger and Cheriton in Folkestone, near the entry to the Channel Tunnel. The individual was an Albanian who had been refused entry to the United Kingdom and was voluntarily returning to Brussels. He was pronounced dead at the scene. The line was closed for several hours following the incident. The train itself returned north to Ashford International, where passengers were transferred to a Eurostar service operating from Marne-la-Vallée to London. Eurostar trains do not currently call at , originally intended to be the London stop for the regional Eurostars. This was to be reviewed following the 2012 Olympics. However, in 2013, Eurostar claimed that its 'business would be hit' by stopping trains there. Although the original plan for Regional Eurostar services to destinations north of London was abandoned, the significantly improved journey times available since the opening of High Speed 1 — which is physically connected to both the East Coast Main Line and the North London Line (for the West Coast Main Line) at St Pancras – and the recently increased maximum speeds on the West Coast Main Line may make potential Regional Eurostar services more commercially viable. This would be even more likely if proposals are adopted for a new high-speed line from London to the north of Britain. Simon Montague, Eurostar's Director of Communications, commented that: "...International services to the regions are only likely once High Speed 2 is built." However, as of 2014 the current plans for High Speed 2 do not allow for a direct rail link between that new line, and High Speed 1, meaning passengers would still be required to change at Euston and take some form of transportation to St Pancras International. Key pieces of infrastructure still belong to LCR via its subsidiary London & Continental Stations and Property, such as the Manchester International Depot, and Eurostar (UK) still owns several track access rights and the rights to paths on both the East Coast and West Coast Main Lines. While no announcement has been made of plans to start Regional Eurostar services, it remains a possibility for the future. In the meantime, the closest equivalent to Regional Eurostar services are same-station connections with East Midlands Railway and Thameslink, changing at St Pancras. The construction of a new concourse at adjacent King's Cross Station improved interchange with St Pancras and provided London North Eastern Railway, Great Northern, Hull Trains and Grand Central services with easier connections to Eurostar. Eurostar has already been involved in reviewing and publishing reports into High Speed 2 for the British Government and looks favourably upon such an undertaking. The operation of Regional Eurostar services will not be considered until such time as High Speed 2 has been completed. Alternatively, future loans of the North of London sets to other operators would enable the trains to operate at their full speed, unlike GNER's previous loan between 2000 and 2005, where the trains were limited to on regular track. A separate company called High Speed Two (HS2) Ltd has been set up to investigate the feasibility and viability of a new line likely serving a similar route to the West Coast Main Line. LGV Picardie is a proposed high-speed line between Paris and Calais via Amiens. By cutting off the corner of the LGV Nord at Lille, it would enable Eurostar trains to save 20 minutes on the journey between Paris and Calais, bringing the London–Paris journey time under 2 hours. In 2008 the French Government announced its future investment plans for new LGVs to be built up to 2020; LGV Picardie was not included but was listed as planned in the longer term. The reduced journey times offered by the opening of High Speed 1 and the opening of the LGV Est and HSL-Zuid bring more continental destinations within a range from London where rail is competitive with air travel. By Eurostar's estimates a train would then take 3 hours 30 minutes from London to Amsterdam. At present Eurostar is concentrating on developing its connections with other services, but direct services to other destinations would be possible. With the new e320 rolling stock allowed Eurostar to enter the Netherlands and possibly Germany in future. In additional with every new country Eurostar enters there are security issues, due to the UK's not having signed up to the Schengen Agreement, which allows unrestricted movement across borders of member countries. For example, on the current Amsterdam to London route it is only direct one way, with people needing to get a train to Brussels to go through the juxtaposed controls; the direct connection is subject to talks between the UK and Dutch governments, set to be completed in 2020 for services to start. The difficulties that Eurostar faces in expanding its services would also be faced by any potential competitors to Eurostar. As the UK is outside the Schengen Agreement, London-bound trains must use platforms that are physically isolated, a constraint which other international operators such as Thalys do not face. In addition, the British authorities are required to make passenger security and passport checks before they board the train, which might deter domestic passengers. Compounding the difficulties in providing a similar service are the Channel Tunnel safety rules, the major ones being the "half-train rule" and the "length rule". The "half-train rule" stipulated that passenger trains had to be able to split in the case of emergency. Class 373 trains were designed as two half-sets, which when coupled form a complete train, enabling them to be split easily in the event of an emergency while in the tunnel, with the unaffected set able to be driven out. The half-train rule was finally abolished in May 2010. However, the "length rule", which states that passenger trains must be at least 375 metres long with a through corridor (to match the distance between the safety doors in the tunnel), was retained, preventing any potential operators from applying to run services with existing fleets (the majority of both TGV and ICE trains are only 200m long). At the same time as Pepy's announcement, Richard Brown announced that Eurostar's plans for expanding its network potentially included Amsterdam and Rotterdam as destinations, using the HSL Zuid line. This would require either equipment upgrades of the existing fleet, or a new fleet equipped for both ERTMS and the domestic signalling systems used by Nederlandse Spoorwegen. Following the December 2009 opening of HSL Zuid, a London–Amsterdam journey is estimated to take 4 hr 16 min. In an interview with Eurostar's Chief Executive Nicolas Petrovic in the "Financial Times" in May 2012, an intention for Eurostar to serve ten new destinations was expressed, including Amsterdam, Frankfurt, Cologne, Lyon, Marseille and Geneva, along with a likely second hub to be created in Brussels. In March 2016 in an interview with Bloomberg, Eurostar's Chief Executive expressed interest in operating a direct train service between London and Bordeaux, but not before 2019. Journey time was said to be around four and a half hours using the new LGV Sud Europe Atlantique. In December 2012 Eurostar announced that on Saturdays during May 2013–June 2013 a new seasonal service would be introduced to Aix-en-Provence, also serving Lyon Part-Dieu and Avignon TGV on the way (the latter being from central Avignon). This is in addition to the long-standing seasonal summer service on Saturdays during July and August and the first week of September travelling to Avignon Centre. The Aix-en-Provence services did not run in 2014 but was replaced along with the seasonal Avignon Centre services with the new year-round service to Lyon and Marseille as of 1 May 2015. In 2018, at least, direct services to Lyon, Avignon and Marseille ran only from May to September, with connections during the rest of the year being offered via Eurostar but requiring a change to SNCF trains in Paris or Lille. Travel time from London to Marseille was roughly 6.5 hours in 2018. In September 2013, Eurostar announced an agreement with the Government of Netherlands and NS, the Dutch railway company, to start twice daily services between London and Amsterdam Centraal; the launch was initially planned for December 2016. The service will use the newly bought Siemens Velaro trainsets and will also call at Brussels, and Rotterdam. The journey time will be around four hours. Initially, trains stop in Brussels for about half an hour to allow domestic passengers from Amsterdam and Rotterdam to leave and allow London bound passengers to board. Eventually, Passengers for London from Amsterdam and Rotterdam will undertake all security checks before boarding and will be able to travel direct both ways. The trains will also convey passengers from the Netherlands on journeys to Brussels who will not need to pass through security and they will be allocated half the train which will be kept separate from the London-bound passengers by locking the intermediate door. The Brussels-bound half of the train will be security swept on arrival at Brussels before Brussels-to-London passengers can board. The journey from London to Amsterdam Centraal will take 3 hr 41 min and trains will call at Brussels Zuid/Midi and Rotterdam Centraal Station. From Amsterdam Centraal to London St Pancras, trains will take 4 hr 9 min to include the 28-minute stop at Brussels. Eurostar trains from London will also call at Antwerp Centraal and Schiphol Airport, although trains from Amsterdam will miss out Antwerp on the journey back to London. In November 2014, Eurostar announced the service to Amsterdam would start in "2016-2017", and would include a stop at Schiphol Airport in addition to the previously announced destinations. Eurostar have indicated that the calling pattern 'is not set in stone' and if a business case supports it the service might be extended to additional cities such as Utrecht. The service was finally planned to start running on 4 April 2018, with fare prices starting at £35 for a single ticket. An "inaugural train" from St Pancras International to Amsterdam via Rotterdam broke a speed record for the journey to Brussels (1hr 46mins) on 20 February 2018. The first regular service to Amsterdam left St Pancras at 08:31 on 4 April 2018. On 27 September 2019, the heads of two of Eurostar's major shareholders, Guillaume Pepy of SNCF, and the chair of SNCB publicized that Eurostar is planning to come together with its sister company the Franco-Belgian transnational rail service Thalys. The arrangement is to merge their operations under the working title of "Green Speed" is to expand services outside the core London-Paris-Brussels-Amsterdam service, to create a grand Western European high-speed rail service covering the UK, France, Belgium, the Netherlands and Germany to serve up to 30 million customers by 2030. , Thalys assisted Eurostar with onwards connections between the two cites of Amsterdam and Brussels, and to provide the Amsterdam to London service, in lieu of passport and customs checks at Amsterdam Centraal station. Thalys currently uses two ex-Eurostar class 373 trains (half-sets 373213 + 373224) to provide Thalys' no frills service IZY, between Paris and Brussels. In 2010, international rail travel was liberalised by new European Union directives, designed to break up monopolies in order to encourage competition for services between countries. This sparked interest among other companies in providing services in competition to Eurostar and new services to destinations beyond Paris and Brussels. The only rail carrier to formally propose and secure permission for such a service up to now is Deutsche Bahn, which intends to run services between London and Germany and the Netherlands. The sale of High Speed One by the British Government having effectively nationalised LCR in June 2009 is also likely to stimulate competition on the line. In March 2010, it was announced that Eurotunnel was in discussions with the Intergovernment Commission, which oversees the tunnel, with the aim of amending elements of the safety code governing the tunnel's usage. Most saliently, the requirement that trains be able to split within the tunnel and each part of the train be driven out to opposite ends has been removed. However, the proposal to allow shorter trains was not passed. Eurotunnel Chairman & Chief Executive Jacques Gounon said that he hoped the liberalisation of rules would allow entry into the market of competitors such as Deutsche Bahn. Sources at Eurotunnel suggested that Deutsche Bahn could have entered the market at the timetable change in December 2012. This, however, did not happen. In July 2010 Deutsche Bahn (DB) announced that it intended to make a test run with a high-speed ICE-3MF train through the Channel Tunnel in October 2010 in preparation for possible future operations. The trial ran on 19 October 2010 with a Class 406 ICE train specially liveried with a British "Union flag" decal. The train was then put on display for the press at St Pancras International. However, this is not the class of train that would be used for the proposed service. At the St Pancras ceremony, DB revealed that it planned to operate from London to Frankfurt and Amsterdam (two of the biggest air travel markets in Europe), with trains 'splitting & joining' in Brussels. It hoped to begin these services in 2013 using Class 407 ICE units, with three trains per day each way—morning, midday and afternoon. Initially the only calling points would be Rotterdam on the way to Amsterdam, and Cologne on the way to Frankfurt. Amsterdam and Cologne would be under four hours from London, Frankfurt around five hours. DB decided to put this on hold mainly due to advance passport check requirements. DB had hoped that immigration checks could be done on board, but British authorities required immigration and security checks to be done at Lille Europe station, taking at least 30 minutes. In August 2010, Trenitalia announced its desire to eventually run high-speed trains from Italy to the United Kingdom, using its newly ordered high-speed trains. The trains will be delivered from 2013. Cumulative ridership since 1994 has reached 190 million, 11 million passengers had used its international services during 2018, the highest ever, a 7% increase on the 10·3 million carried in 2017.
https://en.wikipedia.org/wiki?curid=10093
Equinox An equinox is commonly regarded as the instant of time when the plane (extended indefinitely in all directions) of Earth's equator passes through the center of the Sun. This occurs twice each year, around 20 March and 23 September. In other words, it is the moment at which the center of the visible Sun is directly above the equator. The word is derived from the Latin ', from ' (equal) and ' (genitive ') (night). On the day of an equinox, daytime and nighttime are of approximately equal duration all over the planet. They are not exactly equal, however, due to the angular size of the Sun, atmospheric refraction, and the rapidly changing duration of the length of day that occurs at most latitudes around the equinoxes. Long before conceiving this equality, primitive equatorial cultures noted the day when the Sun rises due east and sets due west, and indeed this happens on the day closest to the astronomically defined event. In the Northern Hemisphere, the March equinox is called the vernal or spring equinox while the September equinox is called the autumnal or fall equinox. In the Southern Hemisphere, the reverse is true. The dates slightly vary due to leap years and other factors. Since the Moon (and to a lesser extent the planets) causes Earth's orbit to slightly vary from a perfect ellipse, the equinox is officially defined by the Sun's more regular ecliptic longitude rather than by its declination. The instants of the equinoxes are currently defined to be when the apparent geocentric longitude of the Sun is 0° and 180°. Systematically observing the sunrise, people discovered that it occurs between two extreme locations at the horizon and eventually noted the midpoint between the two. Later it was realized that this happens on a day when the durations of the day and the night are practically equal and the word "equinox" comes from Latin "aequus", meaning "equal", and "nox", meaning "night". In the northern hemisphere, the vernal equinox (March) conventionally marks the beginning of spring in most cultures and is considered the start of the New Year in the Assyrian calendar, Hindu, and the Persian or Iranian calendars, while the autumnal equinox (September) marks the beginning of autumn. The equinoxes are the only times when the solar terminator (the "edge" between night and day) is perpendicular to the equator. As a result, the northern and southern hemispheres are equally illuminated. For the same reason, this is also the time when the Sun rises for an observer at one of Earth's rotational poles and sets at the other; for a brief period, both North and South Poles are in daylight. In other words, the equinoxes are the only times when the subsolar point is on the equator, meaning that the Sun is exactly overhead at a point on the equatorial line. The subsolar point crosses the equator moving northward at the March equinox and southward at the September equinox. When Julius Caesar established the Julian calendar in 45 BC, he set 25 March as the date of the spring equinox; this was already the starting day of the year in the Persian and Indian calendars. Because the Julian year is longer than the tropical year by about 11.3 minutes on average (or 1 day in 128 years), the calendar "drifted" with respect to the two equinoxes – so that in 300 AD the spring equinox occurred on about 21 March, and by 1500 AD it had drifted backwards to 11 March. This drift induced Pope Gregory XIII to establish the modern Gregorian calendar. The Pope wanted to continue to conform with the edicts of the Council of Nicaea in 325 AD concerning the date of Easter, which means he wanted to move the vernal equinox to the date on which it fell at that time (21 March is the day allocated to it in the Easter table of the Julian calendar), and to maintain it at around that date in the future, which he achieved by reducing the number of leap years from 100 to 97 every 400 years. However, there remained a small residual variation in the date and time of the vernal equinox of about ±27 hours from its mean position, virtually all because the distribution of 24 hour centurial leap-days causes large jumps (see ). This in turn raised the possibility that it could fall on 22 March, and thus Easter Day might theoretically commence before the equinox. The consulting astronomers chose the appropriate number of days to omit so that the equinox would swing from 19 to 21 March but never fall on 22 March (within Europe). The dates of the equinoxes change progressively during the leap-year cycle, because the Gregorian calendar year is not commensurate with the period of the Earth's revolution about the Sun. It is only after a complete Gregorian leap-year cycle of 400 years that the seasons commence at approximately the same time. In the 21st century the earliest March equinox will be 19 March 2096, while the latest was 21 March 2003. The earliest September equinox will be 21 September 2096 while the latest was 23 September 2003 (Universal Time). Day is usually defined as the period when sunlight reaches the ground in the absence of local obstacles. On the date of the equinox, the center of the Sun spends a roughly equal amount of time above and below the horizon at every location on the Earth, so night and day are about the same length. Sunrise and sunset can be defined in several ways, but a widespread definition is the time that the top limb of the Sun is level with the horizon. With this definition, the day is longer than the night at the equinoxes: In sunrise/sunset tables, the atmospheric refraction is assumed to be 34 arcminutes, and the assumed semidiameter (apparent radius) of the Sun is 16 arcminutes. (The apparent radius varies slightly depending on time of year, slightly larger at perihelion in January than aphelion in July, but the difference is comparatively small.) Their combination means that when the upper limb of the Sun is on the visible horizon, its centre is 50 arcminutes below the geometric horizon, which is the intersection with the celestial sphere of a horizontal plane through the eye of the observer. These effects make the day about 14 minutes longer than the night at the equator and longer still towards the poles. The real equality of day and night only happens in places far enough from the equator to have a seasonal difference in day length of at least 7 minutes, actually occurring a few days towards the winter side of each equinox. The times of sunset and sunrise vary with the observer's location (longitude and latitude), so the dates when day and night are equal also depend upon the observer's location. A third correction for the visual observation of a sunrise (or sunset) is the angle between the apparent horizon as seen by an observer and the geometric (or sensible) horizon. This is known as the dip of the horizon and varies from 3 arcminutes for a viewer standing on the sea shore to 160 arcminutes for a mountaineer on Everest. The effect of a larger dip on taller objects (reaching over 2½° of arc on Everest) accounts for the phenomenon of snow on a mountain peak turning gold in the sunlight long before the lower slopes are illuminated. The date on which the day and night are exactly the same is known as an "equilux"; the neologism, believed to have been coined in the 1980s, achieved more widespread recognition in the 21st century. At the most precise measurements, a true equilux is rare, because the lengths of day and night change more rapidly than any other time of the year around the equinoxes. In the mid-latitudes, daylight increases or decreases by about three minutes per day at the equinoxes, and thus adjacent days and nights only reach within one minute of each other. The date of the closest approximation of the equilux varies slightly by latitude; in the mid-latitudes, it occurs a few days before the spring equinox and after the fall equinox in each respective hemisphere. In the half-year centered on the June solstice, the Sun rises north of east and sets north of west, which means longer days with shorter nights for the northern hemisphere and shorter days with longer nights for the southern hemisphere. In the half-year centered on the December solstice, the Sun rises south of east and sets south of west and the durations of day and night are reversed. Also on the day of an equinox, the Sun rises everywhere on Earth (except at the poles) at about 06:00 and sets at about 18:00 (local solar time). These times are not exact for several reasons: Some of the statements above can be made clearer by picturing the day arc (i.e., the path along which the Sun appears to move across the sky). The pictures show this for every hour on equinox day. In addition, some 'ghost' suns are also indicated below the horizon, up to 18° below it; the Sun in such areas still causes twilight. The depictions presented below can be used for both the northern and the southern hemispheres. The observer is understood to be sitting near the tree on the island depicted in the middle of the ocean; the green arrows give cardinal directions. The following special cases are depicted: The March equinox occurs about when the Sun appears to cross the celestial equator northward. In the Northern Hemisphere, the term "vernal point" is used for the time of this occurrence and for the precise direction in space where the Sun exists at that time. This point is the origin of some celestial coordinate systems, which are usually rooted to an astronomical epoch since it gradually varies (precesses) over time: Strictly speaking, at the equinox, the Sun's ecliptic longitude is zero. Its latitude will not be exactly zero, since Earth is not exactly in the plane of the ecliptic. Its declination will not be exactly zero either. The mean ecliptic is defined by the barycenter of Earth and the Moon combined, so the Earth wanders slightly above and below the ecliptic due to the orbital tilt of the Moon. The modern definition of equinox is the instants when the Sun's apparent geocentric longitude is 0° (northward equinox) or 180° (southward equinox). See the adjacent diagram. Because of the precession of the Earth's axis, the position of the vernal point on the celestial sphere changes over time, and the equatorial and the ecliptic coordinate systems change accordingly. Thus when specifying celestial coordinates for an object, one has to specify at what time the vernal point and the celestial equator are taken. That reference time is called the equinox of date. The upper culmination of the vernal point is considered the start of the sidereal day for the observer. The hour angle of the vernal point is, by definition, the observer's sidereal time. Using the current official IAU constellation boundaries – and taking into account the variable precession speed and the rotation of the celestial equator – the equinoxes shift through the constellations as follows (expressed in astronomical year numbering when the etc.): The equinoxes are sometimes regarded as the start of spring and autumn. A number of traditional harvest festivals are celebrated on the date of the equinoxes. One effect of equinoctial periods is the temporary disruption of communications satellites. For all geostationary satellites, there are a few days around the equinox when the Sun goes directly behind the satellite relative to Earth (i.e. within the beam-width of the ground-station antenna) for a short period each day. The Sun's immense power and broad radiation spectrum overload the Earth station's reception circuits with noise and, depending on antenna size and other factors, temporarily disrupt or degrade the circuit. The duration of those effects varies but can range from a few minutes to an hour. (For a given frequency band, a larger antenna has a narrower beam-width and hence experiences shorter duration "Sun outage" windows.) Satellites in geostationary orbit also experience difficulties maintaining power during the equinox, due to the fact that they now have to travel through Earth's shadow and rely only on battery power. Usually, a satellite will travel either north or south of the Earth's shadow due to its shifted axis throughout the year. During the equinox, since geostationary satellites are situated above the Equator, they will be put into Earth's shadow for the longest duration all year. Equinoxes occur on any planet with a tilted rotational axis. A dramatic example is Saturn, where the equinox places its ring system edge-on facing the Sun. As a result, they are visible only as a thin line when seen from Earth. When seen from above – a view seen during an equinox for the first time from the "Cassini" space probe in 2009 – they receive very little sunshine, indeed more planetshine than light from the Sun. This phenomenon occurs once every 14.7 years on average, and can last a few weeks before and after the exact equinox. Saturn's most recent equinox was on 11 August 2009, and its next will take place on 6 May 2025. Mars's most recent equinox was on 8 April 2020 (northern autumn), and the next will be on 7 February 2021 (northern spring).
https://en.wikipedia.org/wiki?curid=10100
Eugene Wigner Eugene Paul "E. P." Wigner (, ; November 17, 1902 – January 1, 1995) was a Hungarian-American theoretical physicist and mathematician. He received the Nobel Prize in Physics in 1963 "for his contributions to the theory of the atomic nucleus and the elementary particles, particularly through the discovery and application of fundamental symmetry principles". A graduate of the Technical University of Berlin, Wigner worked as an assistant to Karl Weissenberg and Richard Becker at the Kaiser Wilhelm Institute in Berlin, and David Hilbert at the University of Göttingen. Wigner and Hermann Weyl were responsible for introducing group theory into physics, particularly the theory of symmetry in physics. Along the way he performed ground-breaking work in pure mathematics, in which he authored a number of mathematical theorems. In particular, Wigner's theorem is a cornerstone in the mathematical formulation of quantum mechanics. He is also known for his research into the structure of the atomic nucleus. In 1930, Princeton University recruited Wigner, along with John von Neumann, and he moved to the United States. Wigner participated in a meeting with Leo Szilard and Albert Einstein that resulted in the Einstein-Szilard letter, which prompted President Franklin D. Roosevelt to initiate the Manhattan Project to develop atomic bombs. Wigner was afraid that the German nuclear weapon project would develop an atomic bomb first. During the Manhattan Project, he led a team whose task was to design nuclear reactors to convert uranium into weapons grade plutonium. At the time, reactors existed only on paper, and no reactor had yet gone critical. Wigner was disappointed that DuPont was given responsibility for the detailed design of the reactors, not just their construction. He became Director of Research and Development at the Clinton Laboratory (now the Oak Ridge National Laboratory) in early 1946, but became frustrated with bureaucratic interference by the Atomic Energy Commission, and returned to Princeton. In the postwar period he served on a number of government bodies, including the National Bureau of Standards from 1947 to 1951, the mathematics panel of the National Research Council from 1951 to 1954, the physics panel of the National Science Foundation, and the influential General Advisory Committee of the Atomic Energy Commission from 1952 to 1957 and again from 1959 to 1964. In later life, he became more philosophical, and published "The Unreasonable Effectiveness of Mathematics in the Natural Sciences", his best-known work outside technical mathematics and physics. Wigner Jenő Pál was born in Budapest, Austria-Hungary on November 17, 1902, to middle class Jewish parents, Elisabeth (Einhorn) and Anthony Wigner, a leather tanner. He had an older sister, Bertha, known as Biri, and a younger sister Margit, known as Manci, who later married British theoretical physicist Paul Dirac. He was home schooled by a professional teacher until the age of 9, when he started school at the third grade. During this period, Wigner developed an interest in mathematical problems. At the age of 11, Wigner contracted what his doctors believed to be tuberculosis. His parents sent him to live for six weeks in a sanatorium in the Austrian mountains, before the doctors concluded that the diagnosis was mistaken. Wigner's family was Jewish, but not religiously observant, and his Bar Mitzvah was a secular one. From 1915 through 1919, he studied at the secondary grammar school called Fasori Evangélikus Gimnázium, the school his father had attended. Religious education was compulsory, and he attended classes in Judaism taught by a rabbi. A fellow student was János von Neumann, who was a year behind Wigner. They both benefited from the instruction of the noted mathematics teacher László Rátz. In 1919, to escape the Béla Kun communist regime, the Wigner family briefly fled to Austria, returning to Hungary after Kun's downfall. Partly as a reaction to the prominence of Jews in the Kun regime, the family converted to Lutheranism. Wigner explained later in his life that his family decision to convert to Lutheranism "was not at heart a religious decision but an anti-communist one". On religious views, Wigner was an atheist. After graduating from the secondary school in 1920, Wigner enrolled at the Budapest University of Technical Sciences, known as the "Műegyetem". He was not happy with the courses on offer, and in 1921 enrolled at the "Technische Hochschule Berlin" (now Technical University of Berlin), where he studied chemical engineering. He also attended the Wednesday afternoon colloquia of the German Physical Society. These colloquia featured such luminaries as Max Planck, Max von Laue, Rudolf Ladenburg, Werner Heisenberg, Walther Nernst, Wolfgang Pauli, and Albert Einstein. Wigner also met the physicist Leó Szilárd, who at once became Wigner's closest friend. A third experience in Berlin was formative. Wigner worked at the Kaiser Wilhelm Institute for Physical Chemistry and Electrochemistry (now the Fritz Haber Institute), and there he met Michael Polanyi, who became, after László Rátz, Wigner's greatest teacher. Polanyi supervised Wigner's DSc thesis, "Bildung und Zerfall von Molekülen" ("Formation and Decay of Molecules"). Wigner returned to Budapest, where he went to work at his father's tannery, but in 1926, he accepted an offer from Karl Weissenberg at the Kaiser Wilhelm Institute in Berlin. Weissenberg wanted someone to assist him with his work on x-ray crystallography, and Polanyi had recommended Wigner. After six months as Weissenberg's assistant, Wigner went to work for Richard Becker for two semesters. Wigner explored quantum mechanics, studying the work of Erwin Schrödinger. He also delved into the group theory of Ferdinand Frobenius and Eduard Ritter von Weber. Wigner received a request from Arnold Sommerfeld to work at the University of Göttingen as an assistant to the great mathematician David Hilbert. This proved a disappointment, as the aged Hilbert's abilities were failing, and his interests had shifted to logic. Wigner nonetheless studied independently. He laid the foundation for the theory of symmetries in quantum mechanics and in 1927 introduced what is now known as the Wigner D-matrix. Wigner and Hermann Weyl were responsible for introducing group theory into quantum mechanics. The latter had written a standard text, "Group Theory and Quantum Mechanics" (1928), but it was not easy to understand, especially for younger physicists. Wigner's "Group Theory and Its Application to the Quantum Mechanics of Atomic Spectra" (1931) made group theory accessible to a wider audience. In these works, Wigner laid the foundation for the theory of symmetries in quantum mechanics. Wigner's theorem proved by Wigner in 1931, is a cornerstone of the mathematical formulation of quantum mechanics. The theorem specifies how physical symmetries such as rotations, translations, and CPT symmetry are represented on the Hilbert space of states. According to the theorem, any symmetry transformation is represented by a linear and unitary or antilinear and antiunitary transformation of Hilbert space. The representation of a symmetry group on a Hilbert space is either an ordinary representation or a projective representation. In the late 1930s, Wigner extended his research into atomic nuclei. By 1929, his papers were drawing notice in the world of physics. In 1930, Princeton University recruited Wigner for a one-year lectureship, at 7 times the salary that he had been drawing in Europe. Princeton recruited von Neumann at the same time. Jenő Pál Wigner and János von Neumann had collaborated on three papers together in 1928 and two in 1929. They anglicized their first names to "Eugene" and "John", respectively. When their year was up, Princeton offered a five-year contract as visiting professors for half the year. The Technische Hochschule responded with a teaching assignment for the other half of the year. This was very timely, since the Nazis soon rose to power in Germany. At Princeton in 1934, Wigner introduced his sister Manci to the physicist Paul Dirac, whom she married. Princeton did not rehire Wigner when his contract ran out in 1936. Through Gregory Breit, Wigner found new employment at the University of Wisconsin. There he met his first wife, Amelia Frank, who was a physics student there. However she died unexpectedly in 1937, leaving Wigner distraught. He therefore accepted a 1938 offer from Princeton to return there. Wigner became a naturalized citizen of the United States on January 8, 1937, and he brought his parents to the United States. Although he was a professed political amateur, on August 2, 1939, he participated in a meeting with Leó Szilárd and Albert Einstein that resulted in the Einstein–Szilárd letter, which prompted President Franklin D. Roosevelt to initiate the Manhattan Project to develop atomic bombs. Wigner was afraid that the German nuclear weapon project would develop an atomic bomb first, and even refused to have his fingerprints taken because they could be used to track him down if Germany won. "Thoughts of being murdered," he later recalled, "focus your mind wonderfully." On June 4, 1941, Wigner married his second wife, Mary Annette Wheeler, a professor of physics at Vassar College, who had completed her Ph.D. at Yale University in 1932. After the war she taught physics on the faculty of Rutgers University's Douglass College in New Jersey until her retirement in 1964. They remained married until her death in November 1977. They had two children, David Wigner and Martha Wigner Upton. During the Manhattan Project, Wigner led a team that included Alvin M. Weinberg, Katharine Way, Gale Young and Edward Creutz. The group's task was to design the production nuclear reactors that would convert uranium into weapons grade plutonium. At the time, reactors existed only on paper, and no reactor had yet gone critical. In July 1942, Wigner chose a conservative 100 MW design, with a graphite neutron moderator and water cooling. Wigner was present at a converted rackets court under the stands at the University of Chicago's abandoned Stagg Field on December 2, 1942, when the world's first atomic reactor, Chicago Pile One (CP-1) achieved a controlled nuclear chain reaction. Wigner was disappointed that DuPont was given responsibility for the detailed design of the reactors, not just their construction. He threatened to resign in February 1943, but was talked out of it by the head of the Metallurgical Laboratory, Arthur Compton, who sent him on vacation instead. As it turned out, a design decision by DuPont to give the reactor additional load tubes for more uranium saved the project when neutron poisoning became a problem. Without the additional tubes, the reactor could have been run at 35% power until the boron impurities in the graphite were burned up and enough plutonium produced to run the reactor at full power; but this would have set the project back a year. During the 1950s, he would even work for DuPont on the Savannah River Site. Wigner did not regret working on the Manhattan Project, and sometimes wished the atomic bomb had been ready a year earlier. An important discovery Wigner made during the project was the Wigner effect. This is a swelling of the graphite moderator caused by the displacement of atoms by neutron radiation. The Wigner effect was a serious problem for the reactors at the Hanford Site in the immediate post-war period, and resulted in production cutbacks and a reactor being shut down entirely. It was eventually discovered that it could be overcome by controlled heating and annealing. Through Manhattan project funding, Wigner and Leonard Eisenbud also developed an important general approach to nuclear reactions, the Wigner–Eisenbud R-matrix theory, which was published in 1947.
https://en.wikipedia.org/wiki?curid=10101
Electroweak interaction In particle physics, the electroweak interaction or electroweak force is the unified description of two of the four known fundamental interactions of nature: electromagnetism and the weak interaction. Although these two forces appear very different at everyday low energies, the theory models them as two different aspects of the same force. Above the unification energy, on the order of 246 GeV, they would merge into a single force. Thus, if the universe is hot enough (approximately 1015 K, a temperature not exceeded since shortly after the Big Bang), then the electromagnetic force and weak force merge into a combined electroweak force. During the quark epoch, the electroweak force split into the electromagnetic and weak force. Sheldon Glashow, Abdus Salam, and Steven Weinberg were awarded the 1979 Nobel Prize in Physics for their contributions to the unification of the weak and electromagnetic interaction between elementary particles, known as the Weinberg–Salam theory. The existence of the electroweak interactions was experimentally established in two stages, the first being the discovery of neutral currents in neutrino scattering by the Gargamelle collaboration in 1973, and the second in 1983 by the UA1 and the UA2 collaborations that involved the discovery of the W and Z gauge bosons in proton–antiproton collisions at the converted Super Proton Synchrotron. In 1999, Gerardus 't Hooft and Martinus Veltman were awarded the Nobel prize for showing that the electroweak theory is renormalizable. After the Wu experiment discovered parity violation in the weak interaction, a search began for a way to relate the weak and electromagnetic interactions. Extending his doctoral advisor Julian Schwinger's work, Sheldon Glashow first experimented with introducing two different symmetries, one chiral and one achiral, and combined them such that their overall symmetry was unbroken. This didn't yield a renormalizable theory, and had to be broken by hand as no spontaneous mechanism was known, but it predicted a new particle, the Z boson. This received little notice, as it matched no experimental finding. In 1964, Salam and Weinberg had the same idea, but predicted a massless photon and three massive gauge bosons with a manually broken symmetry. Later around 1967, while investigating spontaneous symmetry breaking, Weinberg found a set of symmetries predicting a massless, neutral gauge boson. Initially rejecting such a particle as useless, he later realized his symmetries produced the electroweak force, and he proceeded to predict rough masses for the W and Z bosons. Significantly, he suggested this new theory was renormalizable. In 1971, Gerard 't Hooft proved that spontaneously broken gauge symmetries are renormalizable even with massive gauge bosons. Mathematically, electromagnetism is unified with the weak interactions as a Yang–Mills field with an SU(2) × U(1) gauge group, which describes the formal operations that can be applied to the electroweak gauge fields without changing the dynamics of the system. These fields are the weak isospin fields "W"1, "W"2, and "W"3, and the weak hypercharge field "B". This invariance is known as electroweak symmetry. The generators of SU(2) and U(1) are given the name weak isospin (labeled "T") and weak hypercharge (labeled "Y") respectively. These then give rise to the gauge bosons which mediate the electroweak interactions – the three W bosons of weak isospin ("W"1, "W"2, and "W"3), and the "B" boson of weak hypercharge, respectively, all of which are "initially" massless. These are not physical fields yet, before spontaneous symmetry breaking and the associated Higgs mechanism. In the Standard Model, the and bosons, and the photon, are produced through the spontaneous symmetry breaking of the electroweak symmetry SU(2) × U(1)"Y" to U(1)em, effected by the Higgs mechanism (see also Higgs boson), an elaborate quantum field theoretic phenomenon that "spontaneously" alters the realization of the symmetry and rearranges degrees of freedom. The electric charge arises as a (nontrivial) linear combination of "Y" (weak hypercharge) and the "T"3 component of weak isospin that does not couple to the Higgs boson – that is to say, the Higgs and the electromagnetic field have no effect on each other at the level of the fundamental forces ("tree level"), while any other linear combination of the hypercharge and the weak isospin will interact with the Higgs. This causes an apparent separation between the weak force, which interacts with the Higgs, and electromagnetism, which does not. Mathematically, the electric charge is a specific combination of the hypercharge and "T"3 outlined in the figure. U(1)em (the symmetry group of electromagnetism) is defined to be the group generated by this special linear combination, and the symmetry described by this group is unbroken as it does not interact with the Higgs directly (but does through quantum fluctuations). The above spontaneous symmetry breaking makes the 3 and bosons coalesce into two different physical bosons with different masses – the boson, and the photon (), where is the "weak mixing angle". The axes representing the particles have essentially just been rotated, in the ("W"3, "B") plane, by the angle . This also introduces a mismatch between the mass of the and the mass of the particles (denoted as and , respectively), The and bosons, in turn, combine to yield massive charged bosons The Lagrangian for the electroweak interactions is divided into four parts before electroweak symmetry breaking becomes manifest, The formula_7 term describes the interaction between the three vector bosons and the vector boson, where formula_9 (formula_10) and formula_11 are the field strength tensors for the weak isospin and weak hypercharge gauge fields. formula_12 is the kinetic term for the Standard Model fermions. The interaction of the gauge bosons and the fermions are through the gauge covariant derivative, where the subscript runs over the three generations of fermions; , , and are the left-handed doublet, right-handed singlet up, and right handed singlet down quark fields; and and are the left-handed doublet and right-handed singlet electron fields. The Feynman slash formula_14 means the contraction of the 4-gradient with the Dirac matrices and the covariant derivative is (excluding the gluon gauge field for the strong interaction) Here formula_17 is the weak hypercharge and the formula_18 are the components of the weak isospin. The formula_19 term describes the Higgs field and its interactions with itself and the gauge bosons, The formula_21 term describes the Yukawa interaction with the fermions, and generates their masses, manifest when the Higgs field acquires a nonzero vacuum expectation value, discussed next. The Lagrangian reorganizes itself as the Higgs boson acquires a non-vanishing vacuum expectation value dictated by the potential of the previous section. As a result of this rewriting, the symmetry breaking becomes manifest. In the history of the universe, this is believed to have happened shortly after the hot big bang, when the universe was at a temperature 159.5±1.5 GeV (assuming the Standard Model of particle physics). Due to its complexity, this Lagrangian is best described by breaking it up into several parts as follows. The kinetic term formula_24 contains all the quadratic terms of the Lagrangian, which include the dynamic terms (the partial derivatives) and the mass terms (conspicuously absent from the Lagrangian before symmetry breaking) where the sum runs over all the fermions of the theory (quarks and leptons), and the fields formula_26, formula_27, formula_28, and formula_29 are given as with ‘formula_31’ to be replaced by the relevant field (formula_32, formula_33, formula_34), and by the structure constants of the appropriate gauge group. The neutral current formula_35 and charged current formula_36 components of the Lagrangian contain the interactions between the fermions and gauge bosons, where formula_38 The electromagnetic current formula_39 is where formula_41 is the fermions' electric charges. The neutral weak current formula_42 is where formula_44 is the fermions' weak isospin. The charged current part of the Lagrangian is given by where formula_46 contains the Higgs three-point and four-point self interaction terms, formula_48 contains the Higgs interactions with gauge vector bosons, formula_50 contains the gauge three-point self interactions, formula_52 contains the gauge four-point self interactions, formula_54 contains the Yukawa interactions between the fermions and the Higgs field, Note the formula_56 factors in the weak couplings: these factors project out the left handed components of the spinor fields. This is why electroweak theory is said to be a chiral theory.
https://en.wikipedia.org/wiki?curid=10103
Erasmus Reinhold Erasmus Reinhold (October 22, 1511 – February 19, 1553) was a German astronomer and mathematician, considered to be the most influential astronomical pedagogue of his generation. He was born and died in Saalfeld, Saxony. He was educated, under Jacob Milich, at the University of Wittenberg, where he was first elected dean and later became rector. In 1536 he was appointed professor of higher mathematics by Philipp Melanchthon. In contrast to the limited modern definition, "mathematics" at the time also included applied mathematics, especially astronomy. His colleague, Georg Joachim Rheticus, also studied at Wittenberg and was appointed professor of lower mathematics in 1536. Reinhold catalogued a large number of stars. His publications on astronomy include a commentary (1542, 1553) on Georg Purbach's "Theoricae novae planetarum". Reinhold knew about Copernicus and his heliocentric ideas prior to the publication of his "De revolutionibus", and made a favourable reference to him in his commentary on Purbach. However, Reinhold (like other astronomers before Kepler and Galileo) translated Copernicus' mathematical methods back into a geocentric system, rejecting heliocentric cosmology on physical and theological grounds. Duke Albert of Brandenburg Prussia supported Reinhold and financed the printing of Reinhold's "Prutenicae Tabulae" (1551, 1562, 1571 & 1585) or "Prussian Tables". These astronomical tables helped to disseminate calculation methods of Copernicus throughout the Empire, however, Gingerich notes that they showed a "notable lack of commitment" to heliocentricity and were "carefully framed" to be independent of the movement of the Earth. Both Reinhold's "Prutenic Tables" and Copernicus' studies were the foundation for the Calendar Reform by Pope Gregory XIII in 1582. It was Reinhold's heavily annotated copy of "De revolutionibus" in the Royal Observatory, Edinburgh, that started Owen Gingerich on his search for copies of the first and second editions which he describes in "The Book Nobody Read". In Reinhold's unpublished commentary on "De revolutionibus", he calculated the distance from the Earth to the sun. He "massaged" his calculation method in order to arrive at an answer close to that of Ptolemy. His name has been given to a prominent lunar impact crater that lies to the south-southwest of the crater Copernicus, on the Mare Insularum.
https://en.wikipedia.org/wiki?curid=10105
Earthquake An earthquake (also known as a quake, tremor or temblor) is the shaking of the surface of the Earth resulting from a sudden release of energy in the Earth's lithosphere that creates seismic waves. Earthquakes can range in size from those that are so weak that they cannot be felt to those violent enough to propel objects and people into the air, and wreak destruction across entire cities. The seismicity, or seismic activity, of an area is the frequency, type, and size of earthquakes experienced over a period of time. The word "tremor" is also used for non-earthquake seismic rumbling. At the Earth's surface, earthquakes manifest themselves by shaking and displacing or disrupting the ground. When the epicenter of a large earthquake is located offshore, the seabed may be displaced sufficiently to cause a tsunami. Earthquakes can also trigger landslides and occasionally, volcanic activity. In its most general sense, the word "earthquake" is used to describe any seismic event—whether natural or caused by humans—that generates seismic waves. Earthquakes are caused mostly by rupture of geological faults but also by other events such as volcanic activity, landslides, mine blasts, and nuclear tests. An earthquake's point of initial rupture is called its hypocenter or focus. The epicenter is the point at ground level directly above the hypocenter. Tectonic earthquakes occur anywhere in the earth where there is sufficient stored elastic strain energy to drive fracture propagation along a fault plane. The sides of a fault move past each other smoothly and aseismically only if there are no irregularities or asperities along the fault surface that increase the frictional resistance. Most fault surfaces do have such asperities, which leads to a form of stick-slip behavior. Once the fault has locked, continued relative motion between the plates leads to increasing stress and therefore, stored strain energy in the volume around the fault surface. This continues until the stress has risen sufficiently to break through the asperity, suddenly allowing sliding over the locked portion of the fault, releasing the stored energy. This energy is released as a combination of radiated elastic strain seismic waves, frictional heating of the fault surface, and cracking of the rock, thus causing an earthquake. This process of gradual build-up of strain and stress punctuated by occasional sudden earthquake failure is referred to as the elastic-rebound theory. It is estimated that only 10 percent or less of an earthquake's total energy is radiated as seismic energy. Most of the earthquake's energy is used to power the earthquake fracture growth or is converted into heat generated by friction. Therefore, earthquakes lower the Earth's available elastic potential energy and raise its temperature, though these changes are negligible compared to the conductive and convective flow of heat out from the Earth's deep interior. There are three main types of fault, all of which may cause an interplate earthquake: normal, reverse (thrust), and strike-slip. Normal and reverse faulting are examples of dip-slip, where the displacement along the fault is in the direction of dip and where movement on them involves a vertical component. Normal faults occur mainly in areas where the crust is being extended such as a divergent boundary. Reverse faults occur in areas where the crust is being shortened such as at a convergent boundary. Strike-slip faults are steep structures where the two sides of the fault slip horizontally past each other; transform boundaries are a particular type of strike-slip fault. Many earthquakes are caused by movement on faults that have components of both dip-slip and strike-slip; this is known as oblique slip. Reverse faults, particularly those along convergent plate boundaries, are associated with the most powerful earthquakes, megathrust earthquakes, including almost all of those of magnitude 8 or more. Strike-slip faults, particularly continental transforms, can produce major earthquakes up to about magnitude 8. Earthquakes associated with normal faults are generally less than magnitude 7. For every unit increase in magnitude, there is a roughly thirtyfold increase in the energy released. For instance, an earthquake of magnitude 6.0 releases approximately 32 times more energy than a 5.0 magnitude earthquake and a 7.0 magnitude earthquake releases 1,000 times more energy than a 5.0 magnitude of earthquake. An 8.6 magnitude earthquake releases the same amount of energy as 10,000 atomic bombs like those used in World War II. This is so because the energy released in an earthquake, and thus its magnitude, is proportional to the area of the fault that ruptures and the stress drop. Therefore, the longer the length and the wider the width of the faulted area, the larger the resulting magnitude. The topmost, brittle part of the Earth's crust, and the cool slabs of the tectonic plates that are descending down into the hot mantle, are the only parts of our planet that can store elastic energy and release it in fault ruptures. Rocks hotter than about flow in response to stress; they do not rupture in earthquakes. The maximum observed lengths of ruptures and mapped faults (which may break in a single rupture) are approximately . Examples are the earthquakes in Alaska (1957), Chile (1960), and Sumatra (2004), all in subduction zones. The longest earthquake ruptures on strike-slip faults, like the San Andreas Fault (1857, 1906), the North Anatolian Fault in Turkey (1939), and the Denali Fault in Alaska (2002), are about half to one third as long as the lengths along subducting plate margins, and those along normal faults are even shorter. The most important parameter controlling the maximum earthquake magnitude on a fault, however, is not the maximum available length, but the available width because the latter varies by a factor of 20. Along converging plate margins, the dip angle of the rupture plane is very shallow, typically about 10 degrees. Thus, the width of the plane within the top brittle crust of the Earth can become (Japan, 2011; Alaska, 1964), making the most powerful earthquakes possible. Strike-slip faults tend to be oriented near vertically, resulting in an approximate width of within the brittle crust. Thus, earthquakes with magnitudes much larger than 8 are not possible. Maximum magnitudes along many normal faults are even more limited because many of them are located along spreading centers, as in Iceland, where the thickness of the brittle layer is only about . In addition, there exists a hierarchy of stress level in the three fault types. Thrust faults are generated by the highest, strike-slip by intermediate, and normal faults by the lowest stress levels. This can easily be understood by considering the direction of the greatest principal stress, the direction of the force that "pushes" the rock mass during the faulting. In the case of normal faults, the rock mass is pushed down in a vertical direction, thus the pushing force ("greatest" principal stress) equals the weight of the rock mass itself. In the case of thrusting, the rock mass "escapes" in the direction of the least principal stress, namely upward, lifting the rock mass up, and thus, the overburden equals the "least" principal stress. Strike-slip faulting is intermediate between the other two types described above. This difference in stress regime in the three faulting environments can contribute to differences in stress drop during faulting, which contributes to differences in the radiated energy, regardless of fault dimensions. Where plate boundaries occur within the continental lithosphere, deformation is spread out over a much larger area than the plate boundary itself. In the case of the San Andreas fault continental transform, many earthquakes occur away from the plate boundary and are related to strains developed within the broader zone of deformation caused by major irregularities in the fault trace (e.g., the "Big bend" region). The Northridge earthquake was associated with movement on a blind thrust within such a zone. Another example is the strongly oblique convergent plate boundary between the Arabian and Eurasian plates where it runs through the northwestern part of the Zagros Mountains. The deformation associated with this plate boundary is partitioned into nearly pure thrust sense movements perpendicular to the boundary over a wide zone to the southwest and nearly pure strike-slip motion along the Main Recent Fault close to the actual plate boundary itself. This is demonstrated by earthquake focal mechanisms. All tectonic plates have internal stress fields caused by their interactions with neighboring plates and sedimentary loading or unloading (e.g., deglaciation). These stresses may be sufficient to cause failure along existing fault planes, giving rise to intraplate earthquakes. The majority of tectonic earthquakes originate at the ring of fire in depths not exceeding tens of kilometers. Earthquakes occurring at a depth of less than are classified as "shallow-focus" earthquakes, while those with a focal-depth between are commonly termed "mid-focus" or "intermediate-depth" earthquakes. In subduction zones, where older and colder oceanic crust descends beneath another tectonic plate, deep-focus earthquakes may occur at much greater depths (ranging from ). These seismically active areas of subduction are known as Wadati–Benioff zones. Deep-focus earthquakes occur at a depth where the subducted lithosphere should no longer be brittle, due to the high temperature and pressure. A possible mechanism for the generation of deep-focus earthquakes is faulting caused by olivine undergoing a phase transition into a spinel structure. Earthquakes often occur in volcanic regions and are caused there, both by tectonic faults and the movement of magma in volcanoes. Such earthquakes can serve as an early warning of volcanic eruptions, as during the 1980 eruption of Mount St. Helens. Earthquake swarms can serve as markers for the location of the flowing magma throughout the volcanoes. These swarms can be recorded by seismometers and tiltmeters (a device that measures ground slope) and used as sensors to predict imminent or upcoming eruptions. A tectonic earthquake begins by an initial rupture at a point on the fault surface, a process known as nucleation. The scale of the nucleation zone is uncertain, with some evidence, such as the rupture dimensions of the smallest earthquakes, suggesting that it is smaller than while other evidence, such as a slow component revealed by low-frequency spectra of some earthquakes, suggest that it is larger. The possibility that the nucleation involves some sort of preparation process is supported by the observation that about 40% of earthquakes are preceded by foreshocks. Once the rupture has initiated, it begins to propagate along the fault surface. The mechanics of this process are poorly understood, partly because it is difficult to recreate the high sliding velocities in a laboratory. Also the effects of strong ground motion make it very difficult to record information close to a nucleation zone. Rupture propagation is generally modeled using a fracture mechanics approach, likening the rupture to a propagating mixed mode shear crack. The rupture velocity is a function of the fracture energy in the volume around the crack tip, increasing with decreasing fracture energy. The velocity of rupture propagation is orders of magnitude faster than the displacement velocity across the fault. Earthquake ruptures typically propagate at velocities that are in the range 70–90% of the S-wave velocity, which is independent of earthquake size. A small subset of earthquake ruptures appear to have propagated at speeds greater than the S-wave velocity. These supershear earthquakes have all been observed during large strike-slip events. The unusually wide zone of coseismic damage caused by the 2001 Kunlun earthquake has been attributed to the effects of the sonic boom developed in such earthquakes. Some earthquake ruptures travel at unusually low velocities and are referred to as slow earthquakes. A particularly dangerous form of slow earthquake is the tsunami earthquake, observed where the relatively low felt intensities, caused by the slow propagation speed of some great earthquakes, fail to alert the population of the neighboring coast, as in the 1896 Sanriku earthquake. Tides may induce some seismicity. See tidal triggering of earthquakes for details. Most earthquakes form part of a sequence, related to each other in terms of location and time. Most earthquake clusters consist of small tremors that cause little to no damage, but there is a theory that earthquakes can recur in a regular pattern. An aftershock is an earthquake that occurs after a previous earthquake, the mainshock. An aftershock is in the same region of the main shock but always of a smaller magnitude. If an aftershock is larger than the main shock, the aftershock is redesignated as the main shock and the original main shock is redesignated as a foreshock. Aftershocks are formed as the crust around the displaced fault plane adjusts to the effects of the main shock. Earthquake swarms are sequences of earthquakes striking in a specific area within a short period of time. They are different from earthquakes followed by a series of aftershocks by the fact that no single earthquake in the sequence is obviously the main shock, so none has a notable higher magnitude than another. An example of an earthquake swarm is the 2004 activity at Yellowstone National Park. In August 2012, a swarm of earthquakes shook Southern California's Imperial Valley, showing the most recorded activity in the area since the 1970s. Sometimes a series of earthquakes occur in what has been called an "earthquake storm", where the earthquakes strike a fault in clusters, each triggered by the shaking or stress redistribution of the previous earthquakes. Similar to aftershocks but on adjacent segments of fault, these storms occur over the course of years, and with some of the later earthquakes as damaging as the early ones. Such a pattern was observed in the sequence of about a dozen earthquakes that struck the North Anatolian Fault in Turkey in the 20th century and has been inferred for older anomalous clusters of large earthquakes in the Middle East. Quaking or shaking of the earth is a common phenomenon undoubtedly known to humans from earliest times. Prior to the development of strong-motion accelerometers that can measure peak ground speed and acceleration directly, the intensity of the earth-shaking was estimated on the basis of the observed effects, as categorized on various seismic intensity scales. Only in the last century has the source of such shaking been identified as ruptures in the Earth's crust, with the intensity of shaking at any locality dependent not only on the local ground conditions but also on the strength or "magnitude" of the rupture, and on its distance. The first scale for measuring earthquake magnitudes was developed by Charles F. Richter in 1935. Subsequent scales (see seismic magnitude scales) have retained a key feature, where each unit represents a ten-fold difference in the amplitude of the ground shaking and a 32-fold difference in energy. Subsequent scales are also adjusted to have approximately the same numeric value within the limits of the scale. Although the mass media commonly reports earthquake magnitudes as "Richter magnitude" or "Richter scale", standard practice by most seismological authorities is to express an earthquake's strength on the moment magnitude scale, which is based on the actual energy released by an earthquake. It is estimated that around 500,000 earthquakes occur each year, detectable with current instrumentation. About 100,000 of these can be felt. Minor earthquakes occur nearly constantly around the world in places like California and Alaska in the U.S., as well as in El Salvador, Mexico, Guatemala, Chile, Peru, Indonesia, Philippines, Iran, Pakistan, the Azores in Portugal, Turkey, New Zealand, Greece, Italy, India, Nepal and Japan. Larger earthquakes occur less frequently, the relationship being exponential; for example, roughly ten times as many earthquakes larger than magnitude 4 occur in a particular time period than earthquakes larger than magnitude 5. In the (low seismicity) United Kingdom, for example, it has been calculated that the average recurrences are: an earthquake of 3.7–4.6 every year, an earthquake of 4.7–5.5 every 10 years, and an earthquake of 5.6 or larger every 100 years. This is an example of the Gutenberg–Richter law. The number of seismic stations has increased from about 350 in 1931 to many thousands today. As a result, many more earthquakes are reported than in the past, but this is because of the vast improvement in instrumentation, rather than an increase in the number of earthquakes. The United States Geological Survey estimates that, since 1900, there have been an average of 18 major earthquakes (magnitude 7.0–7.9) and one great earthquake (magnitude 8.0 or greater) per year, and that this average has been relatively stable. In recent years, the number of major earthquakes per year has decreased, though this is probably a statistical fluctuation rather than a systematic trend. More detailed statistics on the size and frequency of earthquakes is available from the United States Geological Survey (USGS). A recent increase in the number of major earthquakes has been noted, which could be explained by a cyclical pattern of periods of intense tectonic activity, interspersed with longer periods of low intensity. However, accurate recordings of earthquakes only began in the early 1900s, so it is too early to categorically state that this is the case. Most of the world's earthquakes (90%, and 81% of the largest) take place in the , horseshoe-shaped zone called the circum-Pacific seismic belt, known as the Pacific Ring of Fire, which for the most part bounds the Pacific Plate. Massive earthquakes tend to occur along other plate boundaries too, such as along the Himalayan Mountains. With the rapid growth of mega-cities such as Mexico City, Tokyo and Tehran in areas of high seismic risk, some seismologists are warning that a single quake may claim the lives of up to three million people. While most earthquakes are caused by movement of the Earth's tectonic plates, human activity can also produce earthquakes. Four main activities contribute to this phenomenon: storing large amounts of water behind a dam (and possibly building an extremely heavy building), drilling and injecting liquid into wells, coal mining, and oil drilling. Perhaps the best known example is the 2008 Sichuan earthquake in China's Sichuan Province in May; this tremor resulted in 69,227 fatalities and is the 19th deadliest earthquake of all time. The Zipingpu Dam is believed to have fluctuated the pressure of the fault away; this pressure probably increased the power of the earthquake and accelerated the rate of movement for the fault. The instrumental scales used to describe the size of an earthquake began with the Richter magnitude scale in the 1930s. It is a relatively simple measurement of an event's amplitude, and its use has become minimal in the 21st century. Seismic waves travel through the Earth's interior and can be recorded by seismometers at great distances. The surface wave magnitude was developed in the 1950s as a means to measure remote earthquakes and to improve the accuracy for larger events. The moment magnitude scale not only measures the amplitude of the shock but also takes into account the seismic moment (total rupture area, average slip of the fault, and rigidity of the rock). The Japan Meteorological Agency seismic intensity scale, the Medvedev–Sponheuer–Karnik scale, and the Mercalli intensity scale are based on the observed effects and are related to the intensity of shaking. Every tremor produces different types of seismic waves, which travel through rock with different velocities: Propagation velocity of the seismic waves through solid rock ranges from approx. up to , depending on the density and elasticity of the medium. In the Earth's interior, the shock- or P-waves travel much faster than the S-waves (approx. relation 1.7:1). The differences in travel time from the epicenter to the observatory are a measure of the distance and can be used to image both sources of quakes and structures within the Earth. Also, the depth of the hypocenter can be computed roughly. In the upper crust, P-waves travel in the range per second (or lower) in soils and unconsolidated sediments, increasing to per second in solid rock. In the lower crust, they travel at about per second; the velocity increases within the deep mantle to about per second. The velocity of S-waves ranges from per second in light sediments and per second in the Earth's crust up to per second in the deep mantle. As a consequence, the first waves of a distant earthquake arrive at an observatory via the Earth's mantle. On average, the kilometer distance to the earthquake is the number of seconds between the P- and S-wave times 8. Slight deviations are caused by inhomogeneities of subsurface structure. By such analyses of seismograms the Earth's core was located in 1913 by Beno Gutenberg. S-waves and later arriving surface waves do most of the damage compared to P-waves. P-waves squeeze and expand material in the same direction they are traveling, whereas S-waves shake the ground up and down and back and forth. Earthquakes are not only categorized by their magnitude but also by the place where they occur. The world is divided into 754 Flinn–Engdahl regions (F-E regions), which are based on political and geographical boundaries as well as seismic activity. More active zones are divided into smaller F-E regions whereas less active zones belong to larger F-E regions. Standard reporting of earthquakes includes its magnitude, date and time of occurrence, geographic coordinates of its epicenter, depth of the epicenter, geographical region, distances to population centers, location uncertainty, a number of parameters that are included in USGS earthquake reports (number of stations reporting, number of observations, etc.), and a unique event ID. Although relatively slow seismic waves have traditionally been used to detect earthquakes, scientists realized in 2016 that gravitational measurements could provide instantaneous detection of earthquakes, and confirmed this by analyzing gravitational records associated with the 2011 Tohoku-Oki ("Fukushima") earthquake. The effects of earthquakes include, but are not limited to, the following: Shaking and ground rupture are the main effects created by earthquakes, principally resulting in more or less severe damage to buildings and other rigid structures. The severity of the local effects depends on the complex combination of the earthquake magnitude, the distance from the epicenter, and the local geological and geomorphological conditions, which may amplify or reduce wave propagation. The ground-shaking is measured by ground acceleration. Specific local geological, geomorphological, and geostructural features can induce high levels of shaking on the ground surface even from low-intensity earthquakes. This effect is called site or local amplification. It is principally due to the transfer of the seismic motion from hard deep soils to soft superficial soils and to effects of seismic energy focalization owing to typical geometrical setting of the deposits. Ground rupture is a visible breaking and displacement of the Earth's surface along the trace of the fault, which may be of the order of several meters in the case of major earthquakes. Ground rupture is a major risk for large engineering structures such as dams, bridges, and nuclear power stations and requires careful mapping of existing faults to identify any that are likely to break the ground surface within the life of the structure. Soil liquefaction occurs when, because of the shaking, water-saturated granular material (such as sand) temporarily loses its strength and transforms from a solid to a liquid. Soil liquefaction may cause rigid structures, like buildings and bridges, to tilt or sink into the liquefied deposits. For example, in the 1964 Alaska earthquake, soil liquefaction caused many buildings to sink into the ground, eventually collapsing upon themselves. An earthquake may cause injury and loss of life, road and bridge damage, general property damage, and collapse or destabilization (potentially leading to future collapse) of buildings. The aftermath may bring disease, lack of basic necessities, mental consequences such as panic attacks, depression to survivors, and higher insurance premiums. Earthquakes can produce slope instability leading to landslides, a major geological hazard. Landslide danger may persist while emergency personnel are attempting rescue. Earthquakes can cause fires by damaging electrical power or gas lines. In the event of water mains rupturing and a loss of pressure, it may also become difficult to stop the spread of a fire once it has started. For example, more deaths in the 1906 San Francisco earthquake were caused by fire than by the earthquake itself. Tsunamis are long-wavelength, long-period sea waves produced by the sudden or abrupt movement of large volumes of water—including when an earthquake occurs at sea. In the open ocean the distance between wave crests can surpass , and the wave periods can vary from five minutes to one hour. Such tsunamis travel 600–800 kilometers per hour (373–497 miles per hour), depending on water depth. Large waves produced by an earthquake or a submarine landslide can overrun nearby coastal areas in a matter of minutes. Tsunamis can also travel thousands of kilometers across open ocean and wreak destruction on far shores hours after the earthquake that generated them. Ordinarily, subduction earthquakes under magnitude 7.5 do not cause tsunamis, although some instances of this have been recorded. Most destructive tsunamis are caused by earthquakes of magnitude 7.5 or more. Floods may be secondary effects of earthquakes, if dams are damaged. Earthquakes may cause landslips to dam rivers, which collapse and cause floods. The terrain below the Sarez Lake in Tajikistan is in danger of catastrophic flooding if the landslide dam formed by the earthquake, known as the Usoi Dam, were to fail during a future earthquake. Impact projections suggest the flood could affect roughly 5 million people. One of the most devastating earthquakes in recorded history was the 1556 Shaanxi earthquake, which occurred on 23 January 1556 in Shaanxi province, China. More than 830,000 people died. Most houses in the area were yaodongs—dwellings carved out of loess hillsides—and many victims were killed when these structures collapsed. The 1976 Tangshan earthquake, which killed between 240,000 and 655,000 people, was the deadliest of the 20th century. The 1960 Chilean earthquake is the largest earthquake that has been measured on a seismograph, reaching 9.5 magnitude on 22 May 1960. Its epicenter was near Cañete, Chile. The energy released was approximately twice that of the next most powerful earthquake, the Good Friday earthquake (March 27, 1964), which was centered in Prince William Sound, Alaska. The ten largest recorded earthquakes have all been megathrust earthquakes; however, of these ten, only the 2004 Indian Ocean earthquake is simultaneously one of the deadliest earthquakes in history. Earthquakes that caused the greatest loss of life, while powerful, were deadly because of their proximity to either heavily populated areas or the ocean, where earthquakes often create tsunamis that can devastate communities thousands of kilometers away. Regions most at risk for great loss of life include those where earthquakes are relatively rare but powerful, and poor regions with lax, unenforced, or nonexistent seismic building codes. Earthquake prediction is a branch of the science of seismology concerned with the specification of the time, location, and magnitude of future earthquakes within stated limits. Many methods have been developed for predicting the time and place in which earthquakes will occur. Despite considerable research efforts by seismologists, scientifically reproducible predictions cannot yet be made to a specific day or month. While forecasting is usually considered to be a type of prediction, earthquake forecasting is often differentiated from earthquake prediction. Earthquake forecasting is concerned with the probabilistic assessment of general earthquake hazard, including the frequency and magnitude of damaging earthquakes in a given area over years or decades. For well-understood faults the probability that a segment may rupture during the next few decades can be estimated. Earthquake warning systems have been developed that can provide regional notification of an earthquake in progress, but before the ground surface has begun to move, potentially allowing people within the system's range to seek shelter before the earthquake's impact is felt. The objective of earthquake engineering is to foresee the impact of earthquakes on buildings and other structures and to design such structures to minimize the risk of damage. Existing structures can be modified by seismic retrofitting to improve their resistance to earthquakes. Earthquake insurance can provide building owners with financial protection against losses resulting from earthquakes Emergency management strategies can be employed by a government or organization to mitigate risks and prepare for consequences. Individuals can also take preparedness steps like securing water heaters and heavy items that could injure someone, locating shutoffs for utilities, and being educated about what to do when shaking starts. For areas near large bodies of water, earthquake preparedness encompasses the possibility of a tsunami caused by a large quake. From the lifetime of the Greek philosopher Anaxagoras in the 5th century BCE to the 14th century CE, earthquakes were usually attributed to "air (vapors) in the cavities of the Earth." Thales of Miletus (625–547 BCE) was the only documented person who believed that earthquakes were caused by tension between the earth and water. Other theories existed, including the Greek philosopher Anaxamines' (585–526 BCE) beliefs that short incline episodes of dryness and wetness caused seismic activity. The Greek philosopher Democritus (460–371 BCE) blamed water in general for earthquakes. Pliny the Elder called earthquakes "underground thunderstorms". In recent studies, geologists claim that global warming is one of the reasons for increased seismic activity. According to these studies, melting glaciers and rising sea levels disturb the balance of pressure on Earth's tectonic plates, thus causing an increase in the frequency and intensity of earthquakes. In Norse mythology, earthquakes were explained as the violent struggling of the god Loki. When Loki, god of mischief and strife, murdered Baldr, god of beauty and light, he was punished by being bound in a cave with a poisonous serpent placed above his head dripping venom. Loki's wife Sigyn stood by him with a bowl to catch the poison, but whenever she had to empty the bowl the poison dripped on Loki's face, forcing him to jerk his head away and thrash against his bonds, which caused the earth to tremble. In Greek mythology, Poseidon was the cause and god of earthquakes. When he was in a bad mood, he struck the ground with a trident, causing earthquakes and other calamities. He also used earthquakes to punish and inflict fear upon people as revenge. In Japanese mythology, Namazu (鯰) is a giant catfish who causes earthquakes. Namazu lives in the mud beneath the earth, and is guarded by the god Kashima who restrains the fish with a stone. When Kashima lets his guard fall, Namazu thrashes about, causing violent earthquakes. In modern popular culture, the portrayal of earthquakes is shaped by the memory of great cities laid waste, such as Kobe in 1995 or San Francisco in 1906. Fictional earthquakes tend to strike suddenly and without warning. For this reason, stories about earthquakes generally begin with the disaster and focus on its immediate aftermath, as in "Short Walk to Daylight" (1972), "The Ragged Edge" (1968) or "" (1999). A notable example is Heinrich von Kleist's classic novella, "The Earthquake in Chile", which describes the destruction of Santiago in 1647. Haruki Murakami's short fiction collection "After the Quake" depicts the consequences of the Kobe earthquake of 1995. The most popular single earthquake in fiction is the hypothetical "Big One" expected of California's San Andreas Fault someday, as depicted in the novels "Richter 10" (1996), "Goodbye California" (1977), "2012" (2009) and "San Andreas" (2015) among other works. Jacob M. Appel's widely anthologized short story, "A Comparative Seismology", features a con artist who convinces an elderly woman that an apocalyptic earthquake is imminent. Contemporary depictions of earthquakes in film are variable in the manner in which they reflect human psychological reactions to the actual trauma that can be caused to directly afflicted families and their loved ones. Disaster mental health response research emphasizes the need to be aware of the different roles of loss of family and key community members, loss of home and familiar surroundings, loss of essential supplies and services to maintain survival. Particularly for children, the clear availability of caregiving adults who are able to protect, nourish, and clothe them in the aftermath of the earthquake, and to help them make sense of what has befallen them has been shown even more important to their emotional and physical health than the simple giving of provisions. As was observed after other disasters involving destruction and loss of life and their media depictions, recently observed in the 2010 Haiti earthquake, it is also important not to pathologize the reactions to loss and displacement or disruption of governmental administration and services, but rather to validate these reactions, to support constructive problem-solving and reflection as to how one might improve the conditions of those affected.
https://en.wikipedia.org/wiki?curid=10106
Emperor of Japan The Emperor of Japan is the head of the Imperial Family and the head of state of Japan. Under the 1947 constitution, he is defined as "the symbol of the State and of the unity of the people." Historically, he is also the highest authority of the Shinto religion. In Japanese, the Emperor is called , literally "heavenly sovereign". In English, the use of the term for the Emperor was once common, but is now considered obsolete. Currently, the Emperor of Japan is the only head of state in the world with the English title of "emperor". The Imperial House of Japan is the oldest continuing monarchical house in the world. The historical origins of the emperors lie in the late Kofun period of the 3rd–6th centuries AD, but according to the traditional account of the "Kojiki" (finished 712) and "Nihon Shoki" (finished 720), Japan was founded in 660 BC by Emperor Jimmu, who was said to be a direct descendant of the sun-goddess Amaterasu. Naruhito is the current Emperor of Japan. He acceded to the Chrysanthemum Throne upon the abdication of his father, the now-Emperor Emeritus Akihito on 1 May 2019 at midnight JST. The role of the Emperor of Japan has historically alternated between a largely ceremonial symbolic role and that of an actual imperial ruler. Since the establishment of the first shogunate in 1199, the Emperors of Japan have rarely taken on a role as supreme battlefield commander, unlike many Western monarchs. Japanese emperors have nearly always been controlled by external political forces, to varying degrees. For example, between 1192 and 1867, the "shōguns", or their "shikken" regents in Kamakura (1203–1333), were the "de facto" rulers of Japan, although they were nominally appointed by the Emperor. After the Meiji Restoration in 1867, the Emperor was the embodiment of all sovereign power in the realm, as enshrined in the Meiji Constitution of 1889. Since the enactment of the 1947 constitution, the role of the Emperor has been relegated to a ceremonial head of state without even nominal political powers. Since the mid-nineteenth century, the Imperial Palace has been called , later , and is on the former site of Edo Castle in the heart of Tokyo (the current capital of Japan). Earlier, emperors resided in Kyoto (the ancient capital) for nearly eleven centuries. The Emperor's Birthday (currently 23 February) is a national holiday. Unlike most constitutional monarchs, the Emperor is not the "nominal" chief executive. explicitly vests executive power in the Cabinet, of which the Prime Minister is the leader. The Emperor is also not the commander-in-chief of the Japan Self-Defense Forces. The Japan Self-Defense Forces Act of 1954 also explicitly vests this role with the Prime Minister. The Emperor's powers are limited only to important ceremonial functions. of the Constitution stipulates that the Emperor "shall perform only such acts in matters of state as are provided for in the Constitution and he shall not have powers related to government." It also stipulates that "the advice and approval of the Cabinet shall be required for all acts of the Emperor in matters of state" (Article 3). Article 4 also states that these duties can be delegated by the Emperor as provided for by law. While the Emperor formally appoints the Prime Minister to office, Article 6 of the Constitution requires him to appoint the candidate "as designated by the Diet", without giving the Emperor the right to decline appointment. The Emperor's other duties are laid down in article 7 of the Constitution, where it is stated that "the Emperor, with the advice and approval of the Cabinet, shall perform the following acts in matters of state on behalf of the people." In practice, all of these duties are exercised only in accordance with the binding instructions of the Cabinet: Regular ceremonies of the Emperor with a constitutional basis are the Imperial Investitures "(Shinninshiki)" in the Tokyo Imperial Palace and the Speech from the Throne ceremony in the House of Councillors in the National Diet Building. The latter ceremony opens ordinary and extra sessions of the Diet. Ordinary sessions are opened each January and also after new elections to the House of Representatives. Extra sessions usually convene in the autumn and are opened then. Although the Emperor has been a symbol of continuity with the past, the degree of power exercised by the Emperor has varied considerably throughout Japanese history. In the early 7th century, the Emperor had begun to be called the . The title of Emperor was borrowed from China, being derived from Chinese characters and was retroactively applied to the legendary Japanese rulers who reigned before the 7th–8th centuries AD. According to the traditional account of the Nihon Shoki, Japan was founded by Emperor Jimmu in 660 BC. Modern historians generally believe that the Emperors till Suinin are 'largely legendary' as there is insufficient material available for verification and study regarding their lives. Emperor Keiko till Emperor Ingyo (376-453 AD) are considered as might be real. Emperor Ankō (401-456) of the 5th century AD, traditionally the 20th emperor, is the earliest generally agreed upon historical ruler of all or a part of Japan. The reign of Emperor Kinmei ( – 571 AD), the 29th emperor, is the first for which contemporary historiography is able to assign verifiable dates; however, the conventionally accepted names and dates of the early emperors were not to be confirmed as "traditional" until the reign of Emperor Kanmu (737–806), the 50th sovereign of the Yamato dynasty. Archaeological information about the earliest historical rulers of Japan may lie within the ancient tombs known as kofun, constructed between the early 3rd century and the early 7th century AD. However, since the Meiji period, the Imperial Household Agency has refused to open the kofun to the public nor to archaeologists, citing their desire not to disturb the spirits of the past Emperors. In December 2006, the Imperial Household Agency reversed its position and decided to allow researchers to enter some of the kofun with no restrictions. There have been six non-imperial families who have controlled Japanese emperors: the Soga (530s–645), the Fujiwara (850s–1070), the Taira (1159-1180s), the Minamoto (and Kamakura Bakufu) (1192–1333), the Ashikaga (1336–1565), and the Tokugawa (1603–1867). However, every shogun from the Minamoto, Ashikaga, and Tokugawa families had to be officially recognized by the Emperors, who were still the source of sovereignty, although they could not exercise their powers independently from the Shogunate. The growth of the samurai class from the 10th century gradually weakened the power of the imperial family over the realm, leading to a time of instability. Emperors have been known to come into conflict with the reigning shogun from time to time. Some instances, such as Emperor Go-Toba's 1221 rebellion against the Kamakura shogunate and the 1336 Kenmu Restoration under Emperor Go-Daigo, show the power struggle between the imperial court and the military governments of Japan. Until recent centuries, Japan's territory did not include several remote regions of its modern-day territory. The name "Nippon" came into use only many centuries after the start of the current imperial line. Centralized government only began to appear shortly before and during the time of Prince Shōtoku (572–622). The Emperor was more like a revered embodiment of divine harmony than the head of an actual governing administration. In Japan, it has always been easy for ambitious lords to hold actual power, as such positions have not been inherently contradictory to the Emperor's position. The parliamentary government today continues a similar coexistence with the Emperor as have various shoguns, regents, warlords, guardians, etc. Historically the titles of "Tennō" in Japanese have never included territorial designations as is the case with many European monarchs. The position of Emperor is a territory-independent phenomenon—the Emperor is the Emperor, even if he has followers only in one province (as was the case sometimes with the southern and northern courts). From 1192 to 1867, sovereignty of the state was exercised by the "shōguns", or their "shikken" regents (1203–1333), whose authority was conferred by Imperial warrant. When Portuguese explorers first came into contact with the Japanese (see "Nanban period"), they described Japanese conditions in analogy, likening the Emperor with great symbolic authority but little political power, to the Pope, and the "shōgun" to secular European rulers (e.g., the Holy Roman Emperor). In keeping with the analogy, they even used the term "Emperor" in reference to the "shōguns" and their regents, e.g. in the case of Toyotomi Hideyoshi, whom missionaries called "Emperor Taico-sama" (from Taikō and the honorific "sama"). After the United States Navy Commodore Matthew C. Perry's Black Ships forcibly opened Japan to foreign trade, and the shogunate proved incapable of hindering the "barbarian" interlopers, the Emperor Kōmei began to assert himself politically. By the early 1860s, the relationship between the imperial court and the Shogunate was changing radically. Disaffected domains and "rōnin" began to rally to the call of "sonnō jōi" ("revere the Emperor, expel the barbarians"). The domains of Satsuma and Chōshū, historic enemies of the Tokugawa, used this turmoil to unite their forces and won an important military victory outside of Kyoto against Tokugawa forces. In 1868, Emperor Meiji was restored to nominal full power and the shogunate was dissolved. A new constitution described the Emperor as "the head of the Empire, combining in Himself the rights of sovereignty", whose rights included to sanction and promulgate laws, to execute them and to exercise "supreme command of the Army and the Navy". The liaison conference created in 1893 also made the Emperor the leader of the Imperial General Headquarters. The role of the Emperor as head of the State Shinto religion was exploited during the war, creating an Imperial cult that led to kamikaze bombers and other manifestations of fanaticism. This in turn led to the requirement in the Potsdam Declaration for the elimination "for all time [of] the authority and influence of those who have deceived and misled the people of Japan into embarking on world conquest". In State Shinto, the Emperor was believed to be an "arahitogami" (a living god). Following Japan's surrender, the Allies issued the Shinto Directive separating church and state within Japan. The constitution provides for a parliamentary system of government and guarantees certain fundamental rights. Under its terms, the Emperor of Japan is "the symbol of the State and of the unity of the people" and exercises a purely ceremonial role without the possession of sovereignty. The constitution, also known as the , the or the , was drawn up under the Allied occupation that followed World War II and was intended to replace Japan's previous militaristic and quasi-absolute monarchy system with a form of liberal democracy. Currently, it is a rigid document and no subsequent amendment has been made to it since its adoption. In Japanese mythology, according to "Kojiki" and "Nihon Shoki", the Emperor and his family are said to be direct descendants of the sun-goddess Amaterasu. During World War II, the role of the Emperor as head of the Shinto religion was exploited, which resulted in the creation of State Shinto and an Imperial cult. Following the end of the Second World War, the Allies issued the Shinto Directive which abolished the state support for the Shinto religion, leading to the Humanity Declaration of the incumbent Emperor which refuted the idea that the Emperor is a living divine being, and dismissed the importance of "myths and legends" for the Emperor's status. However, the Emperors have continued to perform many traditional ceremonies privately. The Emperors traditionally had an education officer. In recent times, Emperor Taishō had Count Nogi Maresuke, Emperor Shōwa had Marshal-Admiral Marquis Tōgō Heihachirō, and Emperor Akihito had Elizabeth Gray Vining as well as as their tutors. There are two Japanese words equivalent to the English word "Emperor": , which is used exclusively to refer to the Emperor of Japan, and , which is used primarily to describe non-Japanese Emperors. "Sumeramikoto" ("the Imperial person") was also used in Old Japanese. The term "tennō" was used by the Emperors up until the Middle Ages; then, following a period of disuse, it was used again from the 19th century. In English, the term "mikado" ( or ), literally meaning "the honorable gate" (i.e. the gate of the imperial palace, which indicates the person who lives in and possesses the palace; compare Sublime Porte, an old term for the Ottoman government), was once used (as in "The Mikado", a 19th-century operetta), but this term is now obsolete. Traditionally, the Japanese considered it disrespectful to call any person by his given name, and more so for a person of noble rank. This convention is only slightly relaxed in the modern age and it is still inadvisable among friends to use the given name, use of the family name being the common form of address. In the case of the imperial family, it is considered extremely inappropriate to use the given name. Since Emperor Meiji, it has been customary to have one era per Emperor and to rename each Emperor after his death using the name of the era over which he presided. Before Emperor Meiji, the names of the eras were changed more frequently, and the posthumous names of the Emperors were chosen differently. Hirohito, as usually called in English outside Japan, was never referred to by his name in Japan. He was given posthumous name "Shōwa Tennō" after his death, which is the only name that Japanese speakers currently use when referring to him. The current Emperor on the throne is typically referred to as "Tennō Heika" (, "His (Imperial) Majesty the Emperor"), "Kinjō Heika" (, "His Current Majesty") or simply "Tennō", when speaking Japanese. Emperor Akihito received the title "Daijō Tennō" (, Emperor Emeritus), often shortened to "Jōkō" (), upon his abdication on 30 April 2019, and is expected to be renamed "Heisei Tennō" () after his death and will then be referred to exclusively by that name in Japanese. In Japanese culture, it is considered a major faux pas to refer to a living Emperor by his era name, since it is only after his death when his era name becomes his posthumous name. Originally, the ruler of Japan was known as either / (Yamato-ōkimi, Grand King of Yamato), / ("Wa-ō"/"Wakoku-ō", King of Wa, used externally) or ("Ame-no-shita shiroshimesu ōkimi" or "Sumera no mikoto", Grand King who rules all under heaven, used internally) in Japanese and Chinese sources before the 7th century. The oldest documented use of the word "Tennō" is on a wooden slat, or "mokkan", that was unearthed in Asuka-mura, Nara Prefecture in 1998 and dated back to the reign of Emperor Tenmu and Empress Jitō. Throughout history, Japanese Emperors and noblemen appointed the position of chief wife, rather than just keeping a harem or an assortment of female attendants. The Japanese imperial dynasty consistently practiced official polygamy, a practice that only ended in the Taishō period (1912–1926). Besides the Empress, the Emperor could take, and nearly always took, several secondary consorts ("concubines") of various hierarchical degrees. Concubines were allowed also to other dynasts (Shinnōke, Ōke). After a decree by Emperor Ichijō, some Emperors even had two empresses simultaneously ("kōgō" and "chūgū" are the two separate titles for that situation). With the help of all this polygamy, the imperial clan thus was capable of producing more offspring. (Sons by secondary consorts were usually recognized as imperial princes, too, and could be recognized as heir to the throne if the empress did not give birth to an heir.) Of the eight female reigning empresses of Japan, none married or gave birth after ascending the throne. Some of them, being widows, had produced children before their reigns. In the succession, children of the empress were preferred over sons of secondary consorts. Thus it was significant which quarters had preferential opportunities in providing chief wives to imperial princes, i.e. supplying future empresses. Apparently, the oldest tradition of official marriages within the imperial dynasty were marriages between dynasty members, even half-siblings or uncle and niece. Such marriages were deemed to preserve better the imperial blood or were aimed at producing children symbolic of a reconciliation between two branches of the imperial dynasty. Daughters of others remained concubines, until Emperor Shōmu (701–706)in what was specifically reported as the first elevation of its kindelevated his Fujiwara consort Empress Kōmyō to chief wife. Japanese monarchs have been, as much as others elsewhere, dependent on making alliances with powerful chiefs and other monarchs. Many such alliances were sealed by marriages. The specific feature in Japan has been the fact that these marriages have been soon incorporated as elements of tradition which controlled the marriages of later generations, though the original practical alliance had lost its real meaning. A repeated pattern has been an imperial son-in-law under the influence of his powerful non-imperial father-in-law. Beginning from the 7th and 8th centuries, Emperors primarily took women of the Fujiwara clan as their highest wives – the most probable mothers of future monarchs. This was cloaked as a tradition of marriage between heirs of two "kami" (Shinto deities): descendants of Amaterasu with descendants of the family "kami" of the Fujiwara. (Originally, the Fujiwara were descended from relatively minor nobility, thus their "kami" is an unremarkable one in the Japanese myth world.) To produce imperial children, heirs of the nation, with two-side descent from the two kami, was regarded as desirable – or at least it suited powerful Fujiwara lords, who thus received preference in the imperial marriage market. The reality behind such marriages was an alliance between an imperial prince and a Fujiwara lord, his father-in-law or grandfather, the latter with his resources supporting the prince to the throne and most often controlling the government. These arrangements created the tradition of regents (Sesshō and Kampaku), with these positions held only by a Fujiwara sekke lord. Earlier, the Emperors had married women from families of the government-holding Soga lords, and women of the imperial clan itself, i.e. various-degree cousins and often even their own sisters (half-sisters). Several imperials of the 5th and 6th centuries such as Prince Shōtoku were children of half-sibling couples. These marriages often were alliance or succession devices: the Soga lord ensured his domination of a prince who would be put on the throne as a puppet; or a prince ensured the combination of two imperial descents, to strengthen his own and his children's claim to the throne. Marriages were also a means to seal a reconciliation between two imperial branches. After a couple of centuries, Emperors could no longer take anyone from outside such families as primary wife, no matter what the expediency of such a marriage and power or wealth brought by such might have been. Only very rarely did a prince ascend the throne whose mother was not descended from the approved families. The earlier necessity and expediency had mutated into a strict tradition that did not allow for current expediency or necessity, but only dictated that daughters of a restricted circle of families were eligible brides, because they had produced eligible brides for centuries. Tradition had become more forceful than law. Fujiwara women were often Empresses, and concubines came from less exalted noble families. In the last thousand years, sons of an imperial male and a Fujiwara woman have been preferred in the succession. The five Fujiwara families, Ichijō, Kujō, Nijō, Konoe, and Takatsukasa, were the primary source of imperial brides from the 8th century to the 19th century, even more often than daughters of the imperial clan itself. Fujiwara daughters were thus the usual empresses and mothers of Emperors. This restriction on brides for the Emperor and crown prince was made explicit in the Meiji-era Imperial House Law of 1889. A clause stipulated that daughters of Sekke (the five main branches of the higher Fujiwara) and daughters of the imperial clan itself were primarily acceptable brides. The law was repealed in the aftermath of World War II. The now-Emperor Emeritus Akihito became the first crown prince for over a thousand years to marry a consort from outside the previously eligible circle. In Japanese mythology, the sacred treasures were bestowed on Ninigi-no-Mikoto, the grandson of the goddess Amaterasu, at the advent of Tenson kōrin. Amaterasu sent him to pacify Japan by bringing the three celestial gifts that are used by the Emperor. The account of Ninigi being sent to earth appears in the "Nihon Shoki". The Three Sacred Treasures were inherited by successive Japanese emperors, which are the same as or similar to the sacred treasures in mythology. These three gifts signify that the Emperor is the descendant of Amaterasu herself. The three sacred treasures are: During the succession rite (senso, 践祚), possessing the jewel Yasakani no Magatama, the sword Kusanagi and the mirror Yata no Kagami are a testament of the legitimate serving emperor. The origins of the Japanese imperial dynasty are obscure, and it bases its position on the claim that it has "reigned since time immemorial". There are no records of any Emperor who was not said to have been a descendant of other, yet earlier Emperor ( "bansei ikkei"). There is suspicion that Emperor Keitai (c. AD 500) may have been an unrelated outsider, though the sources (Kojiki, Nihon-Shoki) state that he was a male-line descendant of Emperor Ōjin. However, his descendants, including his successors, were according to records descended from at least one and probably several imperial princesses of the older lineage. The tradition built by those legends has chosen to recognize just the putative male ancestry as valid for legitimizing his succession, not giving any weight to ties through the said princesses. Millennia ago, the Japanese imperial family developed its own peculiar system of hereditary succession. It has been non-primogenitural, more or less agnatic, based mostly on rotation. Today, Japan uses strict agnatic primogeniture, which was adopted from Prussia, by which Japan was greatly influenced in the 1870s. The controlling principles and their interaction were apparently very complex and sophisticated, leading to even idiosyncratic outcomes. Some chief principles apparent in the succession have been: Historically, the succession to the Chrysanthemum Throne has always passed to descendants in male line from the imperial lineage. Generally, they have been males, though over the reign of one hundred monarchs there have been nine women (one pre-historical and eight historical) as Emperor on eleven occasions. Over a thousand years ago, a tradition started that an Emperor should ascend relatively young. A dynast who had passed his toddler years was regarded suitable and old enough. Reaching the age of legal majority was not a requirement. Thus, a multitude of Japanese Emperors have ascended as children, as young as 6 or 8 years old. The high-priestly duties were deemed possible for a walking child. A reign of around ten years was regarded a sufficient service. Being a child was apparently a fine property, to better endure tedious duties and to tolerate subjugation to political power-brokers, as well as sometimes to cloak the truly powerful members of the imperial dynasty. Almost all Japanese empresses and dozens of Emperors abdicated, and lived the rest of their lives in pampered retirement, wielding influence behind the scenes. Several Emperors abdicated to their entitled retirement while still in their teens. These traditions show in Japanese folklore, theater, literature, and other forms of culture, where the Emperor is usually described or depicted as an adolescent. Before the Meiji Restoration, Japan had eleven reigns of reigning empresses, all of them daughters of the male line of the Imperial House. None ascended purely as a wife or as a widow of an Emperor. Imperial daughters and granddaughters, however, usually ascended the throne as a sort of a "stop gap" measure – if a suitable male was not available or some imperial branches were in rivalry so that a compromise was needed. Over half of Japanese empresses and many Emperors abdicated once a suitable male descendant was considered to be old enough to rule (just past toddlerhood, in some cases). Four empresses, Empress Suiko, Empress Kōgyoku (also Empress Saimei), and Empress Jitō, as well as the mythical Empress Jingū, were widows of deceased Emperors and princesses of the blood imperial in their own right. One, Empress Genmei, was the widow of a crown prince and a princess of the blood imperial. The other four, Empress Genshō, Empress Kōken (also Empress Shōtoku), Empress Meishō, and Empress Go-Sakuramachi, were unwed daughters of previous Emperors. None of these empresses married or gave birth after ascending the throne. Article 2 of the Meiji Constitution (the Constitution of the Empire of Japan) stated, "The Imperial Throne shall be succeeded to by imperial male descendants, according to the provisions of the Imperial House Law." The 1889 Imperial Household Law fixed the succession on male descendants of the imperial line, and specifically excluded female descendants from the succession. In the event of a complete failure of the main line, the throne would pass to the nearest collateral branch, again in the male line. If the Empress did not give birth to an heir, the Emperor could take a concubine, and the son he had by that concubine would be recognized as heir to the throne. This law, which was promulgated on the same day as the Meiji Constitution, enjoyed co-equal status with that constitution. Article 2 of the Constitution of Japan, promulgated in 1947 by influence of the U.S. occupation administration, provides that "The Imperial Throne shall be dynastic and succeeded to in accordance with the Imperial Household Law passed by the Diet." The Imperial Household Law of 1947, enacted by the ninety-second and last session of the Imperial Diet, retained the exclusion on female dynasts found in the 1889 law. The government of Prime Minister Yoshida Shigeru hastily cobbled together the legislation to bring the Imperial Household in compliance with the American-written Constitution of Japan that went into effect in May 1947. In an effort to control the size of the imperial family, the law stipulates that only legitimate male descendants in the male line can be dynasts, that imperial princesses lose their status as Imperial Family members if they marry outside the Imperial Family, and that the Emperor and other members of the Imperial Family may not adopt children. It also prevented branches, other than the branch descending from Taishō, from being imperial princes any longer. Succession is now regulated by laws passed by the National Diet. The current law excludes women from the succession. A change to this law had been considered until Princess Kiko gave birth to a son. Until the birth of Prince Hisahito, son of Prince Akishino, on September 6, 2006, there was a potential succession problem, since Prince Akishino was the only male child to be born into the imperial family since 1965. Following the birth of Princess Aiko, there was public debate about amending the current Imperial Household Law to allow women to succeed to the throne. In January 2005, Prime Minister Junichiro Koizumi appointed a special panel composed of judges, university professors, and civil servants to study changes to the Imperial Household Law and to make recommendations to the government. The panel dealing with the succession issue recommended on October 25, 2005, amending the law to allow females of the male line of imperial descent to ascend the Japanese throne. On January 20, 2006, Prime Minister Junichiro Koizumi devoted part of his annual keynote speech to the controversy, pledging to submit a bill allowing women to ascend the throne to ensure that the succession continues in the future in a stable manner. Shortly after the announcement that Princess Kiko was pregnant with her third child, Koizumi suspended such plans. Her son, Prince Hisahito, is the third in line to the throne under the current law of succession. On January 3, 2007, Prime Minister Shinzō Abe announced that he would drop the proposal to alter the Imperial Household Law. Another proposed plan is to allow unmarried men from the abolished collateral branches of the imperial family to rejoin through adoption or marriage. This would be an emergency measure to ensure stable succession. It doesn't revise the Imperial Household Law. This doesn't restore the royalty of the 11 collateral branches of the Imperial House that were abolished in October 1947. During the Kofun period, so-called "archaic funerals" were held for the dead Emperors, but only the funerary rites from the end of the period, which the chronicles describe in more detail, are known. They were centered around the rite of the "mogari" (), a provisional depository between death and permanent burial. Empress Jitō was the first Japanese imperial personage to be cremated (in 703). After that, with a few exceptions, all Emperors were cremated up to the Edo period. For the next 350 years, in-ground burial became the favoured funeral custom. Until 1912, the Emperors of Japan were usually buried in Kyoto. From Emperor Taishō onward, the Emperors have been buried at the Musashi Imperial Graveyard in Tokyo. In 2013, the Imperial Household Agency announced that Emperor Akihito and Empress Michiko would be cremated after they die. Until the end of World War II, the Japanese monarchy was thought to be among the wealthiest in the world. Before 1911, no distinction was made between the imperial crown estates and the Emperor's personal properties, which were considerable. The Imperial Property Law, which came into effect in January 1911, established two categories of imperial properties: the hereditary or crown estates and the personal ("ordinary") properties of the imperial family. The Imperial Household Minister was given the responsibility for observing any judicial proceedings concerning imperial holdings. Under the terms of the law, imperial properties were only taxable in cases where no conflict with the Imperial House Law existed; however, crown estates could only be used for public or imperially-sanctioned undertakings. Personal properties of certain members of the imperial family, in addition to properties held for imperial family members who were minors, were exempted from taxation. Those family members included the Empress Dowager, the Empress, the Crown Prince and Crown Princess, the Imperial Grandson and the consort of the Imperial Grandson. As a result of the poor economic conditions in Japan, 289,259.25 acres of crown lands (about 26% of the total landholdings) were either sold or transferred to government and private-sector interests in 1921. In 1930, the Nagoya Detached Palace (Nagoya Castle) was donated to the city of Nagoya, with six other imperial villas being either sold or donated at the same time. In 1939, Nijō Castle, the former Kyoto residence of the Tokugawa shoguns and an imperial palace since the Meiji Restoration, was likewise donated to the city of Kyoto. At the end of 1935, according to official government figures, the Imperial Court owned roughly 3,111,965 acres of landed estates, the bulk of which (2,599,548 acres) were the Emperor's private lands, with the total acreage of the crown estates amounting to some 512,161 acres; those landholdings comprised palace complexes, forest and farm lands and other residential and commercial properties. The total value of the imperial properties was then estimated at ¥650 million, or roughly US$195 million at prevailing exchange rates. This was in addition to the Emperor's personal fortune, which itself amounted to hundreds of millions of yen and included numerous family heirlooms and furnishings, purebred livestock and investments in major Japanese firms, such as the Bank of Japan, other major Japanese banks, the Imperial Hotel and Nippon Yusen. Following Japan's defeat in the Second World War, all of the collateral branches of the imperial family were abolished under the Allied occupation of the country and the subsequent constitutional reforms, forcing those families to sell their assets to private or government owners. Staff numbers in the imperial households were slashed from a peak of roughly 6000 to about 1000. The imperial estates and the Emperor's personal fortune (then estimated at US$17.15 million, or roughly US$625 million in 2017 terms) were transferred to either state or private ownership, excepting 6,810 acres of landholdings. Since the 1947 constitutional reforms, the imperial family has been supported by an official civil list sanctioned by the Japanese government. The largest imperial divestments were the former imperial Kiso and Amagi forest lands in Gifu and Shizuoka prefectures, grazing lands for livestock in Hokkaido and a stock farm in the Chiba region, all of which were transferred to the Ministry of Agriculture, Forestry and Fisheries. Imperial property holdings have been further reduced since 1947 after several handovers to the government. Today, the primary imperial properties include the two Imperial Palaces at Tokyo and Kyoto, several imperial villas and a number of imperial farms and game preserves. As of 2017, Akihito, the previous Emperor, has an estimated net worth of US$40 million. The exact wealth and expenditures of the Emperor and the imperial family have remained a subject of speculation, and were largely withheld from the public until 2003, when Mori Yohei, a former royal correspondent for the "Mainichi Shimbun", obtained access to 200 documents through a recently passed public information law. Mori's findings, which he published in a book, revealed details of the imperial family's US$240 million civil list (in 2003 values). Among other details, the book revealed the royal family employed a staff of over 1,000 people. Informational notes Citations Bibliography
https://en.wikipedia.org/wiki?curid=10110
Politics of Cape Verde Politics of Cape Verde takes place in a framework of a semi-presidential representative democratic republic, whereby the Prime Minister of Cape Verde is the head of government and the President of the Republic of Cape Verde is the head of state, and of a multi-party system. Executive power is exercised by the President and the Government. Legislative power is vested in both the Government and the National Assembly. The Judiciary is independent of the executive and the legislature. The constitution first approved in 1980 and substantially revised in 1992 forms the basis of government organization. It declares that the government is the "organ that defines, leads, and executes the general internal and external policy of the country" and is responsible to the National Assembly. Following independence in 1975, the African Party for the Independence of Guinea and Cape Verde (PAIGC) established a one-party political system. This became the African Party for the Independence of Cape Verde (PAICV) in 1980, as Cape Verde sought to distance itself from Guinea-Bissau, following unrest in that country. In 1991, following growing pressure for a more pluralistic society, multi-party elections were held for the first time. The opposition party, the Movement for Democracy ("Movimento para a Democracia", MpD), won the legislative elections, and formed the government. The MpD candidate also defeated the PAICV candidate in the presidential elections. In the 1996 elections, the MpD increased their majority, but in the 2001 the PAICV returned to power, winning both the Legislative and the Presidential elections. Generally, Cape Verde enjoys a stable democratic system. The elections have been considered free and fair, there is a free press, and the rule of law is respected by the State. In acknowledgment of this, Freedom House granted Cape Verde two 1st in its annual Freedom in the World report, a perfect score. It is the only African country to receive this score. The Prime Minister is the head of the government and as such proposes other ministers and secretaries of state. The Prime Minister is nominated by the National Assembly and appointed by the President. The President is the head of state and is elected by popular vote for a five-year term; the most recent elections were held in 2016. Also in the legislative branch the National Assembly ("Assembleia Nacional") has 72 members, elected for a five-year term by proportional representation. The judicial system is composed of the Supreme Court and the regional courts. Of the five Supreme Court judges, one is appointed by the President, one by the National Assembly, and three by the Superior Judiciary Council. This council consists of the President of the Supreme Court, the Attorney General, eight private citizens, two judges, two prosecutors, the senior legal inspector of the Attorney General's office, and a representative of the Ministry of Justice. Judges are independent and may not belong to a political party. In October 2000, a female judge who was known for taking strict legal measures in cases of domestic violence was transferred from the capital to the countryside. Separate courts hear civil, constitutional and criminal cases. Appeal is to the Supreme Court. Reforms to strengthen an overburdened judiciary were implemented in 1998. Free legal counsel is provided to indigents, defendants are presumed innocent until proven guilty, and trials are public. Judges must lay charges within 24 hours of arrests.. The Constitution provides for an independent judiciary, and the Government generally respects this provision in practice. The Constitution provides for the right to a fair trial and due process, and an independent judiciary usually enforces this right. Unlike in the previous year, there were no reports of politicization and biased judgement in the judiciary. Cases involving former public office holders still are under investigation. For example, the investigations continued in the case of the former prime minister accused of embezzlement in the privatization of ENACOL (a parastatal oil supply firm) in which he allegedly embezzled approximately $16,250 (2 million Cape Verdean escudos) from the buyers of the parastatal. The case of four persons accused of church desecration in 1996 also was under investigation. These individuals filed a complaint with the Attorney General against the judiciary police for alleged fabrication of evidence. The Constitution provides for the right to a fair trial. Defendants are presumed to be innocent; they have the right to a public, nonjury trial; to counsel; to present witnesses; and to appeal verdicts. Free counsel is provided for the indigent. Regional courts adjudicate minor disputes on the local level in rural areas. The Ministry of Justice does not have judicial powers; such powers lie with the courts. The judiciary generally provides due process rights; however, the right to an expeditious trial is constrained by a seriously overburdened and understaffed judicial system. A backlog of cases routinely leads to trial delays of 6 months or more; more than 10,780 cases were pending at year's end. In addition the right of victims to compensation and recovery for pain and mental suffering are overlooked, due both to the low damage assessments imposed and ineffective enforcement of court sentences. Cape Verde is divided into 22 municipalities (concelhos, singular - concelho): Boa Vista, Brava, Maio, Mosteiros, Paul, Porto Novo, Praia, Ribeira Grande, Ribeira Grande de Santiago, Sal, Santa Catarina, Santa Catarina do Fogo, Santa Cruz, São Domingos, São Filipe, São Lourenço dos Órgãos, São Miguel, São Nicolau, São Salvador do Mundo, São Vicente, Tarrafal, Tarrafal de São Nicolau. Article 24 of the Cape Verde Constitution states that The website of the governmental Institute of Cape Verde Communities states that such a measure was adopted ""to stimulate reciprocity from host countries of Cape Verdian migrants"". A law nr. 36/V/97 was promulgated on August 25, 1997 regulating the "Statute of Lusophone Citizen", concerning nationals from any country member of the Community of Portuguese Language Countries (article 2), stating in its article 3 that ""The lusophone citizen with residence in Cape Verde is recognized the active and passive electoral capacity for municipal elections, under conditions of the law. The lusophone citizen with residence in Cape Verde has the right to exercise political activity related to his electoral capacity."" ACCT, ACP, AfDB, AU, CCC, ECA, ECOWAS, FAO, G-77, IBRD, ICAO, ICRM, IDA, IFAD, IFC, IFRCS, ILO, IMF, IMO, Intelsat, Interpol, IOC, IOM (observer), ITU, ITUC, NAM, OAU, OPCW, UN, UNCTAD, UNESCO, UNIDO, UPU, WHO, WIPO, WMO, WTO (applicant)
https://en.wikipedia.org/wiki?curid=5462
Transport in Cape Verde Most transportation in Cape Verde is done by air. There are regular flights between the major islands (Santiago, Sal and São Vicente), with less frequent flights to the other islands. Boat transportation is available, though not widely used nor dependable. In the major cities, public bus transport runs periodically and taxis are common. In smaller towns, there are mostly hiaces and/or taxis. Railways: 0 km - There are no railways in Cape Verde. There was a short overhead conveyor system for salt from the open salt lake on Sal to the port at Pedra de Lume, and a short rail track to the pier at Santa Maria for similar purposes. Both are now disused. Roadways: "total:" 10,000 km including unpaved tracks accessible only to four wheel drive vehicles "asphalt:" 360 km "cobbled:" 5,000 km (2007 estimates) The majority of Cape Verdean roads are paved with cobblestones cut from local basalt. Recent international aid has allowed the asphalting of many roads including all of the highway between Praia and Tarrafal, all of the highway between Praia and Cidade Velha, and all of the highway between Praia, Pedra Badejo, and Calheta de São Miguel on Santiago, and the dual carriageway between Santa Maria and Espargos on Sal. A new ring road has been built from Praia International Airport around the city of Praia. The primary method of intercity and inter-village transport for Cape Verdeans is by aluguer shared taxis, commonly called "Yasi", which is a derived from the name HiAce, because the Toyota HiAce is the most common shared taxi model. Few Cape Verdeans own cars, but ownership is rising rapidly with increasing prosperity, particularly on Santiago Island. Ports and harbors: Mindelo on São Vicente is the main port for cruise liners and the terminus for the ferry service to Santo Antão. A marina for yachts is undergoing enlargement (2007). Praia on Santiago is a main hub for ferry service to other islands. Palmeira on Sal supplies fuel for the main airport on the island, Amílcar Cabral International Airport, and is important for hotel construction on the island. Porto Novo on Santo Antão is the only source for imports and exports of produce from the island as well as passenger traffic since the closure of the airstrip at Ponta do Sol. There are smaller harbors, essentially single jetties at Tarrafal on São Nicolau, Sal Rei on Boa Vista, Vila do Maio (Porto Inglês) on Maio, São Filipe on Fogo and Furna on Brava. These are terminals for inter island ferry service carrying freight and passengers. There are small harbors, with protective breakwaters, used by fishing boats at Tarrafal on Santiago, Pedra de Lume on Sal and Ponta do Sol on Santo Antão. Some offer suitable protection for small yachts. The pier at Santa Maria on Sal used by both fishing and dive boats has been rehabilitated. Merchant marine: "total:" 10 "ships by type:" chemical tanker 1, trawler/cargo ship 5, passenger/cargo 5 "foreign-owned:" 2 (Spain 1, UK 1) (2008) Airports International Airports: International passenger traffic is forecast to exceed 250,000 passengers for 2007. Annual growth, mostly of tourists from Europe is anticipated to continue at just under 20%. (Source ASA Cape Verde airport authority) Main Airlines serving the country: TACV flies daily international flights from Lisbon to Sal or Praia and once a week from Amsterdam, Munich, Paris, Las Palmas, Fortaleza and Boston to one or other of the international airports. It operates on a frequency varying from daily to thrice weekly on inter-island flights to each of the seven islands with operational airports and also to Dakar. It has a fleet of two Boeing 757s and three ATR42s have been replaced by ATR72s. It is currently (2010) undergoing privatization at the insistence of the World Bank. The road network of Cape Verde is managed by the national government ("Instituto de Estradas") and by the municipalities. The total length of the road network is 1,650 km, of which 1,113 km national roads and 537 km municipal roads. Of the national roads, 36% is asphalted. TACV Cabo Verde Airlines, the national airline, flies weekly from Boston Logan International Airport to Praia International Airport at Praia Santiago island. Currently (2007) these flights are on Wednesdays, but schedules vary and are subject to change. It also has flights four times weekly from Lisbon to Francisco Mendes (the recently-opened airport at Praia on Santiago island) and four times weekly from Lisbon to Amílcar Cabral International Airport on Sal island. There is a flight on Mondays from Paris-Charles de Gaulle Airport to Sal and on Thursdays from Amsterdam Schiphol Airport via Munich-Riem Airport to Sal. Return flights are just after midnight on the same day. From Las Palmas in the Canary Islands, Spain there are night flights on Mondays and Thursdays, with departures just after midnight. Return flights are the previous day. There is a service from Praia to Fortaleza, Brazil on Mondays and Thursdays departing early evening and returning at night. All international flights are operated by Boeing 757 aircraft. Most international flights are subject to delay and sometimes cancellation. TAP Air Portugal the Portuguese national carrier operates a daily service from Lisbon to Sal with late evening departures returning after midnight and reaching Lisbon in the early morning. Most flights are delayed and onward connections from Lisbon can be missed as a result. TAP and other European carriers provide connections with most European capitals, enabling same day through flights. From the UK, direct routes by Astraeus from London Gatwick and Manchester to Sal ceased in April 2008; their website has not taken reservations since May 2008. TACV Cabo Verde Airlines opened a route from London Stansted in October 2008 though it was rumoured that flights were being cancelled due to minimum take up though with effect from May 2008, TACV have ceased flights from London Gatwick. There is a Fly TACV website, and you can book from their website. Reservations should be made via the UK TACV office on 0870 774 7338. Thomson Airways have opened additional routes from London Gatwick and Manchester on Mondays and Fridays. Various options and bookings can be made via Thomsonfly to both Sal and Boa Vista. Hamburg International provides a charter service from Hamburg via Düsseldorf on Thursdays and Condor operates from Frankfurt Rhein Main on Tuesdays returning on Wednesday. Neos operates charter flights from Milan Malpensa, Rome Fiumcino and Bologna on Wednesdays. TACV Cabo Verde Airlines, the national airline has been a monopoly carrier within the island archipelago (2007). It operates services from the main hub airports at Sal and Santiago to Boa Vista, Fogo, Maio, São Nicolau and São Vicente at frequencies ranging from thrice weekly to thrice daily. Air strips on the other islands of Brava and Santo Antão are closed (2007) and can only be reached by ferry services from other islands. TACV does not publish timetables; flight times are listed on departure boards. Tickets can be bought at the TACV shop at each airport by queuing and paying in cash (euros or escudos). Flights are often delayed and sometimes cancelled due to weather or operational conditions. Services are operated by ATR 42 turboprop aircraft, which are being replaced (2007) by the enlarged ATR 72 aircraft. Inter island tariffs vary depending on the distance but are generally around €180 return. Air passes are obtainable for multiple flights, when buying an international ticket on TACV. Halcyonair a private carrier with Portuguese and Cape Verdean shareholders is commenced operations on inter-island flights during 2007. It has obtained the necessary licensing from the Cape Verde Government. The frequency and regularity of publicly-accessible ground transportation services vary between the islands and municipalities. There are some common features that can be found throughout Cape Verde. The primary mode of transportation between municipalities is via shared minibuses commonly referred to as a "yasi", because of the Toyota HiAce which makeup the majority of the minibuses in service. While 12-14 passenger "yasi" class minibuses connect the major municipalities at their end points, modified pickup trucks with partially covered cabs and benches installed in the back transport passengers along shorter distances through minor municipalities and the rural areas in between. These modified pickup trucks are referred to as "hilux" after the Toyota Hilux, the common model adapted. Notably, both "yasi" and "hilux" transportation will stop and pickup any passenger that hails them, as well as drop off any passenger that requests to disembark at any point. intermuncipality transportation licenses are granted on an individual basis to each vehicle in the name of the owner by the Direcção Geral dos Transportes e Rodoviários (General Directorate of Transport and Roads). With the exception of the Praia ⇄ Assomada route on Santiago, all yasi and hilux class vehicles licensed to carry passengers act as individual freelancers, not collectively. As such, they do not adhere to scheduling, and have no obligation to provide service. This includes many vehicles running the same route, owned by the same person. Brava Hiluxes and yasis connect Furna and Nova Sintra mostly when boats arrive. Other parts of the island are connected by these vehicles. Fogo Fogo has many yasis running the routes between São Filipe and Mosteiros, São Filipe and Chã das Calderas. Unlike many other islands, these buses depart at roughly the same time every day, and despite the presence of multiple vehicles running each route, passengers can find themselves stranded if they do not board a vehicle during the limited departure window. Yasis tend to depart Mosteiros headed to São Filipe around 6am, and tend to Depart São Filipe headed to Chã around noon. São Vicente Mindelo has a municipal bus service run by the company Transcor. Yasi and hilux transportation connects Mindelo with other parts of the island. Other transportation companies especially minibuses include Transporte Morabeza, Transporte Alegría, Amizade, Sotral y Automindelo. Santiago Maura Company and Sol Atlántico are the only two companies that have been granted municipal bus service licenses. Over the past decade, Maura Company, which had previously been the dominant bus company, has retired the majority of its buses, while many that continue to run are in a state of disrepair due to financial difficulties. Sol Atlántico, in contrast, has greatly increased its fleet of buses, adding several new high capacity buses in 2015. Municipal bus prices are regulated at 44 escudos per ride. Transfers are not allowed. Bus schedules do not exist, but buses start running around 6am and stop around 9pm. Bus stops exist, and are frequently infiltrated by minibus vehicles (also called "yasis") and both (taxi) licensed and unlicensed "clan" taxis illegally running municipal bus routes without a municipal license. No other city on Santiago has a municipal bus service. The government of Assomada has solicited requests for a bus service but so far none has been approved, and there are no short term plans for any bus company to enter the municipal market. Transportation between the municipalities and rural areas is handled predominantly by yasi and hilux transportation. Rates are not fixed and range from 20 escudos for short trips between rural areas up to 500 escudos for Praia ⇄ Tarrafal. Some commonly accepted prices charged between municipalities are 100 escudos for Praia ⇄ São Domingos, 150 escudos for Praia ⇄ Orgãos, and 250 escudos for Praia ⇄ Assomada. Some of the yasis start collecting passengers before dawn to transport between Praia and Assomada and Praia and Pedra Badejo, and the last departures usually occur between 7 and 8pm. These vehicles do not maintain a schedule (with the exception of two early morning vehicles departing Assomada at 5:40 and 6:20 headed to Praia), instead choosing to drive around in circles within the urban centers of Praia, Assomada, and Pedra Badejo to pick up passengers until they are full, or over capacity (14 passengers is the legal limit for an actual Toyota HiAce), at which point they depart. Yasi drivers employ helpers to hawk out the window the destination of the yasi, as well as the obligatory "cheio", meaning full, with little regard for the number of people aboard. Helpers and drivers sometimes use shills (fake passengers) to overcome the common chicken and egg problem wherein passengers will not board an empty (or low passenger) minibus in an urban center because they know it will not depart until it is full. They will board a nearly-full (or over capacity) bus because they know it is likely to depart soon. In 2015 a project called EcobusCV started running a fleet of dual fuel waste vegetable oil / diesel modified Toyota HiACE minibuses using a scheduled service model between Praia and Assomada. Buses depart one per hour, on the hour, from designated bus stops in Praia, at Igreja Nova Apostólica in Fazenda, and Assomada, in front of the court house. The current departure schedule as of September 15 is one departure per hour, every hour starting at 7am, with the last departure at 6pm. EcobusCV plans to expand to departures in 30 minute intervals before the end of 2015. EcobusCV has instituted aggressive, transparent pricing undercutting the informal generally accepted prices between municipalities, which has started to cause freelance yasis to alter their pricing. Taxis are common in Praia and Assomada. Taxis with a base in Praia are painted beige, while taxis with a base in Assomada are painted white. They can carry passengers between municipalities, but they are prohibited from circulating and picking up passengers outside of their base city, though they will usually pickup passengers if they get hailed on their way back to their home city. Taximeters are installed in most legal taxis, but many are not functional and they are almost never used because the generally accepted rates are cheaper than what the taximeter would usually count. In Praia there is a large number of "clan" or clandestine taxis that operate without paying for a license. Most people identify Toyota Corolla hatchbacks as clans and they are frequently hailed. While the minimum taximeter price is officially 80, in practice 100 is the minimum a person pays if they board a taxi. Taxi rates in Praia generally go up to 250 escudos from the furthest points of the city to Plateau, and cross town taxis cap out at 400 during the day. Rates generally go up by 50 escudos after 10pm at night, though for longer distances some will try to charge an extra 100. An exception to this rule is the airport. Airport rates generally range from 500-1000 depending on the starting place or destination, and can go up by several hundred at night. Sal Sal has unscheduled yasi service between Espargos and Santa Maria, with frequent departures in the morning from Espargos, where most locals live, to Santa Maria, where most locals work, and vice versa in the afternoon. Several ferries operate between the islands with much lower fares than the airlines. These are provided by various independent shipping companies and their conditions and seaworthiness vary. Many services depart from Praia at about midnight, arriving in outlying islands at breakfast time. Return trips often depart around mid-day. Service schedules are approximate and delays or cancellations of service are common. Conditions can be very crowded it is advisable to pre-book a cabin for all but the shortest of trips. Passages can be very rough in winter. Departure days vary according to the season and are frequently altered. Enquire at the shipping offices in Praia and other Cape Verdean ports. In early 2011, the Kriola, the first of a proposed fleet of ferryboats belonging to the company Cabo Verde Fast Ferry (CVFF) arrived in Praia directly from Singapore. It was custom-built there by the Dutch shipbuilding company, Damen Group. The Kriola operates regular service among the Sotavento islands of Brava, Fogo, and Santiago. Lesser ferry routes:
https://en.wikipedia.org/wiki?curid=5465
Foreign relations of Cape Verde Cape Verde follows a policy of nonalignment and seeks cooperative relations with all friendly states. Angola, Brazil, the People's Republic of China, Cuba, France, Germany, Portugal, Senegal, Russia, South Korea and the United States maintain embassies in Praia. Cape Verde is actively interested in foreign affairs, especially in Africa. It has bilateral relations with some Lusophone nations and holds membership in a number of international organizations. It also participates in most international conferences on economic and political issues.
https://en.wikipedia.org/wiki?curid=5467
Cayman Islands The Cayman Islands ( or ) is an autonomous British Overseas Territory in the western Caribbean Sea. The territory comprises the three islands of Grand Cayman, Cayman Brac and Little Cayman, which are located to the south of Cuba and northeast of Honduras, between Jamaica and Mexico's Yucatán Peninsula. The capital city is George Town on Grand Cayman, which is the most populous of the three islands. The Cayman Islands is considered to be part of the geographic Western Caribbean Zone as well as the Greater Antilles. The territory is often considered a major world offshore financial haven for international businesses and many wealthy individuals. It has been placed on the EU's blacklist of foreign tax havens. No archaeological evidence for an indigenous presence has been found on the Cayman Islands. Therefore, it is believed that they were discovered by Christopher Columbus on 10 May 1503 during his final voyage to the Americas. He named them 'Las Tortugas' due to the large number of turtles found on the islands (which were soon hunted to near-extinction); however, in the succeeding decades the islands began to be referred to as the Caymans, after the caimans present there. No immediate colonisation occurred following Columbus's discovery; however, a variety of settlers from various backgrounds made their home on the islands, including pirates, shipwrecked sailors, and deserters from Oliver Cromwell's army in Jamaica. Sir Francis Drake briefly visited the islands in 1586. The first recorded permanent inhabitant of the Cayman Islands, Isaac Bodden, was born on Grand Cayman around 1661. He was the grandson of the original settler named Bodden who was probably one of Oliver Cromwell's soldiers at the taking of Jamaica in 1655. England took formal control of the Cayman Islands, along with Jamaica, as a result of the Treaty of Madrid of 1670. That same year saw an attack on a turtle fishing settlement on Little Cayman by the Spanish under Manuel Ribeiro Pardal. Following several unsuccessful attempts at settlement in what had by now become a haven for pirates, a permanent English-speaking population in the islands dates from the 1730s. With settlement, after the first royal land grant by the Governor of Jamaica in 1734, came the perceived need for slaves. Many were brought to the islands from Africa; this is evident today with the majority of native Caymanians being of African and/or English descent. The results of the first census taken in the islands in 1802 showed the population on Grand Cayman to be 933, with 545 of those inhabitants being enslaved. Slavery was abolished in the Cayman Islands in 1833. At the time of abolition, there were over 950 blacks of African ancestry enslaved by 116 white families of English ancestry. On June 22 1863, the Cayman Islands became officially declared and administered as a dependency of the Crown Colony of Jamaica. The islands continued to be governed as part of the Colony of Jamaica until 1962, when they became a separate Crown colony while Jamaica became an independent Commonwealth realm. On February 8 1794, the Caymanians rescued the crews of a group of ten merchant ships, including HMS "Convert", an incident that has since become known as the Wreck of the Ten Sail. The ships had struck a reef and run aground during rough seas. Legend has it that King George III rewarded the island with a promise never to introduce taxes as compensation for their generosity, as one of the ships carried a member of the King's own family. While this remains a popular legend, the story is not true. In the 1950s, tourism began to take off with the opening of the airport, a bank and several hotels, plus a number of scheduled flights and cruise stop-overs. Politically the Cayman Islands were an internally self-governing territory of Jamaica from 1958 to 1962; however, they reverted to direct British rule following the independence of Jamaica in 1962. In 1972, a large degree of internal autonomy was granted by a new constitution, with further revisions being made in 1994. The Cayman Islands government focused on boosting the territory's economy via tourism and off-shore finance, both of which mushroomed from the 1970s onwards. The Caymans have historically been a tax-exempt destination, and the government has always relied on indirect and not direct taxes. The territory has never levied income tax, capital gains tax, or any wealth tax, making them a popular tax haven. The constitution was further modified in 2001 and 2009, codifying various aspects of human rights legislation. On 11 September 2004 the island of Grand Cayman, which lies largely unprotected at sea level, was hit by Hurricane Ivan, creating an 8-ft (2.4 m) storm surge which flooded many areas of Grand Cayman. An estimated 83% of the dwellings on the island were damaged including 4% requiring complete reconstruction. A reported 70% of all dwellings suffered severe damage from flooding or wind. Another 26% sustained minor damage from partial roof removal, low levels of flooding, or impact with floating or wind driven hurricane debris. Power, water and communications were disrupted for months in some areas as Ivan was the worst hurricane to hit the islands in 86 years. Grand Cayman began a major rebuilding process and within two years its infrastructure was nearly returned to pre-hurricane status. Due to the tropical location of the islands, more hurricanes or tropical systems have affected the Cayman Islands than any other region in the Atlantic basin; it has been brushed or directly hit, on average, every 2.23 years. The islands are in the western Caribbean Sea and are the peaks of an undersea mountain range called the Cayman Ridge (or Cayman Rise). This ridge flanks the Cayman Trough, deep which lies to the south. The islands lie in the northwest of the Caribbean Sea, east of Quintana Roo, Mexico and Yucatán State, Mexico, northeast of Costa Rica, north of Panama, south of Cuba and west of Jamaica. They are situated about south of Miami, east of Mexico, south of Cuba, and about northwest of Jamaica. Grand Cayman is by far the largest, with an area of . Grand Cayman's two "sister islands", Cayman Brac and Little Cayman, are about east north-east of Grand Cayman and have areas of respectively. The nearest land mass from Grand Cayman is the Canarreos Archipelago (about 150 miles away), whereas the nearest from the easternmost island Cayman Brac is the Jardines de la Reina archipelago (about 100 miles away) - both of which are part of Cuba. All three islands were formed by large coral heads covering submerged ice age peaks of western extensions of the Cuban Sierra Maestra range and are mostly flat. One notable exception to this is The Bluff on Cayman Brac's eastern part, which rises to above sea level, the highest point on the islands. Terrain is mostly a low-lying limestone base surrounded by coral reefs. The mammalian species in the Cayman Islands include the introduced Central American agouti and eight species of bats. At least three now extinct native rodent species were present until the discovery of the islands by Europeans. Marine life around the island of the Grand Cayman includes tarpon, silversides ("Atheriniformes"), French angelfish ("Pomacanthus paru"), and giant barrel sponges. A number of cetaceans are found in offshore waters. These species include the goose-beaked whale ("Ziphius cavirostris"), Blainville's beaked whale ("Mesoplodon densirostris") and sperm whale ("Physeter macrocephalus"). Cayman avian fauna includes two endemic subspecies of "Amazona" parrots: "Amazona leucocephala hesterna" or Cuban amazon, presently restricted to the island of Cayman Brac, but formerly also on Little Cayman, and "Amazona leucocephala caymanensis" or Grand Cayman parrot, which is native to the Cayman Islands, forested areas of Cuba, and the Isla de la Juventud. Little Cayman and Cayman Brac are also home to red-footed and brown boobies. Although the barn owl ("Tyto alba") occurs in all three of the islands they are not commonplace. The Cayman Islands also possess five endemic subspecies of butterflies. These butterfly breeds can be viewed at the Queen Elizabeth II Botanic Park on the Grand Cayman. Among other notable fauna at the Queen Elizabeth II Botanic Park is the critically threatened blue iguana which is also known as the Grand Cayman iguana ("Cyclura lewisi"). The blue iguana is endemic to the Grand Cayman particularly because of rocky, sunlit, open areas near the island's shores that are advantageous for the laying of eggs. Nevertheless, habitat destruction and invasive mammalian predators remain primary reasons that blue iguana hatchlings do not survive naturally. The Cuban crocodile ("Crocodylus rhombifer") once inhabited the islands. The name "Cayman" is derived from a Carib word for various crocodilians. The Cayman Islands has a tropical wet and dry climate, with a wet season from May to October, and a dry season that runs from November to April. Seasonally, there is little temperature change. A major natural hazard is the tropical cyclones that form during the Atlantic hurricane season from June to November. On 11 and 12 September 2004, Hurricane Ivan struck the Cayman Islands. The storm resulted in two deaths and caused great damage to the infrastructure on the islands. The total economic impact of the storms was estimated to be $3.4 billion. According to the Economics and Statistics Office of the Government of the Cayman Islands, the Cayman Islands had a population of about 68,076 . The Cayman Islands has more registered businesses than people. In , the Cayman Islands had an estimated population of about according to the United Nations, representing a mix of more than 100 nationalities. The majority of the population resides on Grand Cayman, followed by Cayman Brac and Little Cayman. According to the Cayman Islands 2018 Compendium of Statistics released by the Economics and Statistics Office (ESO) the estimated resident population is 65,813 people, broken down as follows: The main work on the island is financial services: there are 100,000 registered companies there, more than the population itself. Referring to one building on Cayman, Ugland House, former US president Barack Obama said it housed "12,000 corporations. Now, that's either the biggest building or the biggest tax scam on record." With an average income of around KYD$47,000, Caymanians have the highest standard of living in the Caribbean. According to the CIA World Factbook, the Cayman Islands GDP per capita is the 38th highest in the world, but the CIA's data for Cayman dates to 2004 and is likely to be lower than present-day values. The territory prints its own currency, the Cayman Islands dollar (KYD), which is pegged to the US dollar US$1.227 to 1 KYD. However, in many retail stores throughout the islands, the KYD is typically traded at US$1.25. The government has established a Needs Assessment Unit to relieve poverty in the islands. Despite this, the Cayman Islands holds a relatively low unemployment rate of about 4.24% as of 2015, lower than the value of 4.7% that was recorded in 2014. The government's primary source of income is indirect taxation: there is no income tax, capital gains tax, or corporation tax. An import duty of 5% to 22% (automobiles 29.5% to 100%) is levied against goods imported into the islands. Few goods are exempt; notable exemptions include books, cameras, gold, and perfume. One of Grand Cayman's main attractions is Seven Mile Beach, site of a number of the island's hotels and resorts. Named one of the Ultimate Beaches by Caribbean Travel and Life, Seven Mile Beach is a public beach on the western shore of Grand Cayman Island. Historical sites in Grand Cayman, such as Pedro St. James Castle in Savannah, also attract visitors. All three islands offer scuba diving, and the Cayman Islands is home to several snorkelling locations where tourists can swim with stingrays. The most popular area to do this is Stingray City, Grand Cayman. Stingray City is a top attraction in Grand Cayman and originally started in the 1980s, when divers started feeding squid to stingrays. The stingrays started to associate the sound of the boat motors with food, and thus visit this area year round. There are two shipwrecks off the shores of Cayman Brac, including the MV Captain Keith Tibbetts; Grand Cayman also has several shipwrecks off its shores, including one deliberate one. On 30 September 1994 the was decommissioned and struck from the Naval Vessel Register. In November 2008 her ownership was transferred for an undisclosed amount to the government of the Cayman Islands, which had decided to sink the "Kittiwake" in June 2009 to form a new artificial reef off Seven Mile Beach, Grand Cayman. Following several delays, the ship was finally scuttled according to plan on 5 January 2011. The "Kittiwake" has become a dynamic environment for marine life. While visitors are not allowed to take anything, there are endless sights. Each of the five decks of the ship offers squirrelfish, rare sponges, Goliath groupers, urchins, and more. Experienced and beginner divers are invited to swim around the "Kittiwake". Pirates Week is an annual 11-day November festival started in 1977 by then-Minister of Tourism Jim Bodden to boost tourism during the country's tourism slow season. Other Grand Cayman tourist attractions include: the ironshore landscape of Hell; the marine theme park "Cayman Turtle Centre: Island Wildlife Encounter", previously known as "Boatswain's Beach"; the production of gourmet sea salt; and the Mastic Trail, a hiking trail through the forests in the centre of the island. The National Trust for the Cayman Islands provides guided tours weekly on the Mastic Trail and other locations. Another attraction to visit on Grand Cayman is the Observation Tower, located in Camana Bay. The Observation Tower is 75 feet tall and provides 360-degree views across Seven Mile Beach, George Town, the North Sound, and beyond. It is free to the public and climbing the tower has become a popular thing to do in the Cayman Islands. Points of interest include the East End Light (sometimes called Gorling Bluff Light), a lighthouse at the east end of Grand Cayman island. The lighthouse is the centrepiece of East End Lighthouse Park, managed by the National Trust for the Cayman Islands; the first navigational aid on the site was the first lighthouse in the Cayman Islands. 360 commercial vessels and 1,674 pleasure craft were registered in the Cayman Islands totaling 4.3 million GT. The Cayman Islands has a population of 68,076 () and therefore a limited workforce. Work permits may, therefore, be granted to foreigners. On average, there have been more than 21,000 foreigners holding valid work permits. To work in the Cayman Islands as a non-citizen, a work permit is required. This involves passing a police background check and a health check. A prospective immigrant worker will not be granted a permit unless certain medical conditions are met, including testing negative for syphilis and HIV. A permit may be granted to individuals on special work. A foreigner must first have a job to move to the Cayman Islands. The employer applies and pays for the work permit. Work permits are not granted to foreigners who are in the Cayman Islands (unless it is a renewal). The Cayman Islands Immigration Department requires foreigners to remain out of the country until their work permit has been approved. The Cayman Islands presently imposes a controversial "rollover" in relation to expatriate workers who require a work permit. Non-Caymanians are only permitted to reside and work within the territory for a maximum of nine years unless they satisfy the criteria of key employees. Non-Caymanians who are "rolled over" may return to work additional nine-year periods, subject to a one-year gap between their periods of work. The policy has been the subject of some controversy within the press. Law firms have been particularly upset by the recruitment difficulties that it has caused. Other less well-remunerated employment sectors have been affected as well. Concerns about safety have been expressed by diving instructors, and realtors have also expressed concerns. Others support the rollover as necessary to protect Caymanian identity in the face of immigration of large numbers of expatriate workers. Concerns have been expressed that in the long term, the policy may damage the preeminence of the Cayman Islands as an offshore financial centre by making it difficult to recruit and retain experienced staff from onshore financial centres. Government employees are no longer exempt from this "rollover" policy, according to this report in a local newspaper. The governor has used his constitutional powers, which give him absolute control over the disposition of civil service employees, to determine which expatriate civil servants are dismissed after seven years service and which are not. This policy is incorporated in the Immigration Law (2003 revision), written by the United Democratic Party government, and subsequently enforced by the People's Progressive Movement Party government. Both governments agree to the term limits on foreign workers, and the majority of Caymanians also agree it is necessary to protect local culture and heritage from being eroded by a large number of foreigners gaining residency and citizenship. In recognition of the CARICOM (Free Movement) Skilled Persons Act which came into effect in July 1997 in some of the CARICOM countries such as Jamaica and which has been adopted in other CARICOM countries, such as Trinidad and Tobago it is possible that CARICOM nationals who hold the "A Certificate of Recognition of Caribbean Community Skilled Person" may be allowed to work in the Cayman Islands under normal working conditions. The Cayman Islands is a British overseas territory, listed by the UN Special Committee of 24 as one of the 16 non-self-governing territories. The current Constitution, incorporating a Bill of Rights, was ordained by a statutory instrument of the United Kingdom in 2009. A 19-seat (not including two non-voting members appointed by the Governor which brings the total to 21 members) Legislative Assembly is elected by the people every four years to handle domestic affairs. Of the elected Members of the Legislative Assembly (MLAs), seven are chosen to serve as government Ministers in a Cabinet headed by the Governor. The Premier is appointed by the Governor. A Governor is appointed by the Queen of the United Kingdom on the advice of the British Government to represent the monarch. Governors can exercise complete legislative and executive authority if they wish through blanket powers reserved to them in the constitution. Bills which have passed the Legislative Assembly require royal assent before becoming effective. The Constitution empowers the Governor to withhold royal assent in cases where the legislation appears to him or her to be repugnant to or inconsistent with the Constitution or affects the rights and privileges of the Legislative Assembly or the Royal Prerogative, or matters reserved to the Governor by article 55. The executive authority of the Cayman Islands is vested in the Queen and is exercised by the Government, consisting of the Governor and the Cabinet. There is an office of the Deputy Governor, who must be a Caymanian and have served in a senior public office. The Deputy Governor is the acting Governor when the office of Governor is vacant, or the Governor is not able to discharge his or her duties or is absent from the Cayman Islands. The current Governor of the Cayman Islands is Martyn Roper. The Cabinet is composed of two official members and seven elected members, called Ministers; one of whom is designated Premier. The Premier can serve for two consecutive terms after which he or she is barred from attaining the office again. Although an MLA can only be Premier twice any person who meets the qualifications and requirements for a seat in the Legislative Assembly can be elected to the Legislative Assembly indefinitely. There are two official members of the Legislative Assembly, the Deputy Governor and the Attorney General. They are appointed by the Governor in accordance with Her Majesty's instructions, and although they have seats in the Legislative Assembly, under the 2009 Constitution, they do not vote. They serve in a professional and advisory role to the MLAs, the Deputy Governor represents the Governor who is a representative of the Queen and the British Government. While the Attorney General serves to advise on legal matters and has special responsibilities in the LA, he is generally responsible for changes to the Penal code among other things. The seven Ministers are voted into office by the 19 elected members of the Legislative Assembly of the Cayman Islands. One of the Ministers, the leader of the majority political party, is appointed Premier by the Governor. After consulting the Premier, the Governor allocates a portfolio of responsibilities to each Cabinet Minister. Under the principle of collective responsibility, all Ministers are obliged to support in the Assembly any measures approved by Cabinet. Almost 80 departments, sections and units carry out the business of government, joined by a number of statutory boards and authorities set up for specific purposes, such as the Port Authority, the Civil Aviation Authority, the Immigration Board, the Water Authority, the University College Board of Governors, the National Pensions Board and the Health Insurance Commission. Since 2000, there have been two official major political parties: The Cayman Democratic Party (CDP) and the People's Progressive Movement (PPM). While there has been a shift to political parties, many contending for office still run as independents. The two parties are notably similar, though they consider each other rivals in most cases, their differences are generally in personality and implementation rather than actual policy. The Cayman Islands currently lacks any real liberal or progressive representation in the Legislative Assembly or in the form of organized political parties. As of the May 2017 General Election, members of the PPM and CDP have joined together with 3 independent members to form a government coalition despite many years of enmity. The law enforcement in the country is provided chiefly by the RCIPS or Royal Cayman Islands Police Service and the CICBC or Cayman Islands Customs & Border Control. These two agencies co-operate in aspects of law enforcement, including their joint marine unit. The defence of the Cayman Islands is the responsibility of the United Kingdom. On 12 October 2019, the government announced the formation of the Cayman Regiment, a new British Armed Forces unit. The Cayman Regiment was planned to become operational in 2020, with an initial 35-50 personnel of mostly reservists that are locally recruited and likely initially trained as engineers and expected growth of an increase of personnel to several hundred. In mid-December 2019, recruitment for commanding officers and junior officers began, with the commanding officers expected to begin work in Jan 2020 and the junior officers expected to begin in February 2020. In January 2020, the first officers were chosen for the Cayman Regiment. The Cayman Islands Cadet Corps was formed in March 2001 and carries out military-type training with teenage citizens of the country. As of 2017 the PPM led Coalition government have pledged to form a Coast Guard to protect the interests of the Islands, especially in terms of illegal immigration and illegal drug importation as well as Search and Rescue. in Mid 2018 the Commander and 2nd in Command of the Cayman Islands Coast Guard were appointed. The First Commanding officer Appointed is Commander Robert Scotland and Second In Command Lieutenant Commander Leo Anglin In Mid 2019 the Commander and 2nd in Command took part in international joint operations with the United States Coast Guard and the Jamaica Defense Force Coast Guard called Operation Riptide. This making it first deployment for the Cayman Islands Coast Guard and first in 10 years any Cayman Representative has been on a foreign military ship for a counternarcotic operation. In late November 2019 it was announced that the Cayman Islands Coast Guard would become operational in January 2020, with initial total of 21 Coast Guardsman half of which would come from the joint marine unit, with further recruitment in the new year. One of the many taskings of the Coast Guard will be to push enforcement of all laws that apply to the designated Wildlife Interaction Zone No direct taxation is imposed on residents and Cayman Islands companies. The government receives the majority of its income from indirect taxation. Duty is levied against most imported goods, which is typically in the range of 22% to 25%. Some items are exempted, such as baby formula, books, cameras and certain items are taxed at 5%. Duty on automobiles depends on their value. The duty can amount to 29.5% up to $20,000.00 KYD CIF (cost, insurance and freight) and up to 42% over $30,000.00 KYD CIF for expensive models. The government charges flat licensing fees on financial institutions that operate in the islands and there are work permit fees on foreign labour. A 13% government tax is placed on all tourist accommodations in addition to US$37.50 airport departure tax which is built into the cost of an airline ticket. There are no taxes on corporate profits, capital gains, or personal income. There are no estate or death inheritance taxes payable on Cayman Islands real estate or other assets held in the Cayman Islands. The legend behind the lack of taxation comes from the Wreck of the Ten Sail, when a British prince was supposedly involved in a shipwreck of the north coast of Grand Cayman. Luckily, a local spotted the shipwreck and swam to his help, rescuing the Prince and his crew. Since that day the Prince declared that in gratitude for his help - the citizens of the country of the Cayman Islands would never pay tax. There is, however, no documented evidence for this story, nor even that a British prince was involved in the shipwreck. Foreign policy is controlled by the United Kingdom, as the islands remain an overseas territory of the United Kingdom. Although in its early days, the Cayman Islands' most important relationships were with Britain and Jamaica, in recent years, as a result of economic dependence, a relationship with the United States has developed. Though the Cayman Islands is involved in no major international disputes, they have come under some criticism due to the use of their territory for narcotics trafficking and money laundering. In an attempt to address this, the government entered into the Narcotics Agreement of 1984 and the Mutual Legal Assistance Treaty of 1986 with the United States, to reduce the use of their facilities associated with these activities. In more recent years, they have stepped up the fight against money laundering, by limiting banking secrecy, introducing requirements for customer identification and record keeping, and requiring banks to co-operate with foreign investigators. Due to their status as an overseas territory of the UK, the Cayman Islands has no representation either in the United Nations or in most other international organisations. However, the Cayman Islands still participates in some international organisations, being an associate member of Caricom and UNESCO, and a member of a sub-bureau of Interpol. George Town is the port capital of Grand Cayman. There are no berthing facilities for cruise ships, but up to four cruise ships can anchor in designated anchorages. There are three cruise terminals in George Town, the North, South, and Royal Watler Terminals. The ride from the ship to the terminal is about 5 minutes. The Cayman Islands Education Department operates state schools. Caymanian children are entitled to free primary and secondary education. There are two public high schools on Grand Cayman, John Gray High School and Clifton Hunter High School, and one on Cayman Brac, Layman E. Scott High School. Various churches and private foundations operate several private schools. The University College of the Cayman Islands has campuses on Grand Cayman and Cayman Brac and is the only government-run university on the Cayman Islands. The International College of the Cayman Islands is a private college in Grand Cayman. The college was established in 1970 and offers associate's, bachelor's and master's degree programmes. Grand Cayman is also home to St. Matthew's University, which includes a medical school and a school of veterinary medicine. The Cayman Islands Law School, a branch of the University of Liverpool, is based on Grand Cayman. The Cayman Islands Civil Service College, a unit of Cayman Islands government organised under the Portfolio of the Civil Service, is in Grand Cayman. Co-situated with University College of the Cayman Islands, it offers both degree programs and continuing education units of various sorts. The college opened in 2007 and is also used as a government research centre. There is a University of the West Indies Open campus in the territory. "See Health in the Cayman Islands." The Royal Cayman Islands Police Service (RCIPS) provides law enforcement for the three islands. Regular off-shore marine and air patrols are conducted by the RCIP using a small fleet of vessels and a helicopter. Grand Cayman is a port of call for Britain's Royal Navy and the United States Coast Guard who often assist with sea rescues when their resources are in the Cayman Islands area. The Cayman Islands Fire Service provides fire prevention, fire fighting and rescue. Its headquarters are in George Town and has substations in Frank Sound, West Bay, Cayman Brac and Little Cayman. Access to Emergency Services is available using 9-1-1, the Emergency telephone number, the same number as is used in Canada and the United States. The Cayman Islands Department of Public Safety's Communications Centre processes 9-1-1 and non-emergency Police assistance, Ambulance Service, Fire Service and search and rescue calls for all three islands. The Communications Centre dispatches RCIP and EMS units directly; however, the Cayman Islands Fire Service maintains their own dispatch room at the airport fire station. Truman Bodden Sports Complex is a multi-use complex in George Town. The complex is separated into an outdoor, six-lane swimming pool, full purpose track and field and basketball/netball courts. The field surrounded by the track is used for association football matches as well as other field sports. The track stadium holds 3,000 people. Association football is the national and most popular sport, with the Cayman Islands national football team representing the Cayman Islands in FIFA. The Cayman Islands Basketball Federation joined the international basketball governing body FIBA in 1976. The country's national team attended the Caribbean Basketball Championship for the first time in 2011. Rugby union is a developing sport, and has its own national men's team, women's team, and Sevens team. The Cayman Men's Rugby 7s team is second in the region after the 2011 NACRA 7s Championship. The Cayman Islands is a member of FIFA, the International Olympic Committee and the Pan American Sports Organisation, and also competes in the biennial Island Games. The Cayman Islands is a member of the International Cricket Council which they joined in 1997 as an Affiliate, before coming an Associate member in 2002. The Cayman Islands national cricket team represents the islands in international cricket. The team has previously played the sport at first-class, List A and Twenty20 level. It competes in Division Five of the World Cricket League. Squash is popular in the Cayman Islands with a vibrant community of mostly ex-pats playing out of the 7 court South Sound Squash Club. In addition, the women's professional squash association hosts one of their major events each year in an all glass court being set up in Camana Bay. In December 2012, the former Cayman Open will be replaced by the Women's World Championships, the largest tournament in the world. The top Cayman men's player, Cameron Stafford is No. 2 in the Caribbean and ranked top 200 on the men's professional circuit. Flag football (CIFFA) has men's, women's and co-ed leagues. Other organised sports leagues include softball, beach volleyball, Gaelic football and ultimate frisbee. The Cayman Islands Olympic Committee was founded in 1973 and was recognised by the IOC (International Olympic Committee) in 1976. In the 21st century, skateboarding has become popular among the youth. In February 2010, the first purpose-built track for kart racing in the Cayman Islands was opened. Corporate karting Leagues at the track have involved widespread participation with 20 local companies and 227 drivers taking part in the 2010 Summer Corporate Karting League. The Cayman National Cultural Foundation manages the F.J. Harquail Cultural Centre and the US$4 million Harquail Theatre. The Cayman National Cultural Foundation, established in 1984, helps to preserve and promote Cayman folk music, including the organisation of festivals such as Cayman Islands International Storytelling Festival, the Cayman JazzFest, Seafarers Festival and Cayfest. The jazz, calypso and reggae genres of music styles feature prominently in Cayman music as celebrated cultural influences. The National Gallery of the Cayman Islands is an art museum in George Town. Founded in 1996, NGCI is an arts organisation that seeks to fulfil its mission through exhibitions, artist residencies, education/outreach programmes and research projects in the Cayman Islands. The NGCI is a non profit institution, part of the Ministry of Health and Culture. There is one print newspaper currently in circulation throughout the islands: the "Cayman Compass". Online news services include the Cayman News Service and the Cayman Compass online edition. Local radio stations are broadcast throughout the islands. Feature films that have been filmed in the Cayman Islands include: "The Firm", "Haven", "Cayman Went" and "Zombie Driftwood".
https://en.wikipedia.org/wiki?curid=5468
Geography of the Cayman Islands The Cayman Islands are a British dependency and island country. It is a three-island archipelago in the Caribbean Sea, consisting of Grand Cayman, Cayman Brac, and Little Cayman. Georgetown, the capital of the Cayman Islands is south of Havana, Cuba, and northwest of Kingston, Jamaica, northeast of Costa Rica, north of Panama and are between Cuba and Central America. Georgetown's geographic coordinates are 19.300° north, 81.383° west. The Cayman Islands have a land area of approximately 1.5 times the size of Washington, D.C. and just larger than Saint Kitts and Nevis. The Cayman Islands have a coastline of . The Cayman Islands make a maritime claim of a exclusive fishing zone and a territorial sea of . The islands are located on the Cayman Rise which forms the northern margin of the Cayman Trough. The trough is the deepest point in the Caribbean Sea and forms part of the tectonic boundary between the North American Plate and the Caribbean Plate. The Cayman Rise extends from southeastern Cuba along the northern margin of the Cayman Trough toward Costa Rica and resulted from Paleocene to Eocene island arc formation with associated volcanism along an extinct subduction zone. The islands are formed of marine limestone and dolomite that was uplifted during the late Miocene epoch. Due to the Islands' location, the Cayman Islands do get earthquakes. The Cayman Islands have a tropical wet and dry climate, with a wet season from May to December, and a dry season that runs from January to April. Terrain is mostly a low-lying limestone base surrounded by coral reefs. Besides earthquakes another major natural hazard is the tropical cyclones that form during the Atlantic hurricane season from July to November. An important environmental issue is the lack of fresh water resources. Drinking water supplies must be met by rainwater catchment and desalination. There is also a problem with trash washing up on the beaches or being deposited by there by residents. The Cayman Islands have no recycling or waste treatment facilities. Natural resources include fish and a climate and beaches that foster tourism, which is the islands' major industry. A 2012 estimate of land use determined that the Cayman Islands' had 0.83 percent arable land and 2.08 percent permanent crops. The territory is subdivided into six districts that are governed by district managers and that have a role as electoral districts and as regional units for statistics. Five of the districts are located on the main island, Grand Cayman. The sixth, Sister Islands, comprises the islands of Little Cayman and Cayman Brac. The population is concentrated in the three (south-)western districts George Town (capital), West Bay, and Bodden Town. Those have a population density many times higher than all remaining districts.
https://en.wikipedia.org/wiki?curid=5470
Demographics of the Cayman Islands This article is about the demographic features of the population of the Cayman Islands, including population density, ethnicity, religious affiliations and other aspects of the population. With its success in the tourism and financial service industries, the Cayman Islands have attracted many international businesses and citizens to relocate. The largest numbers of expatriates living in the Cayman Islands (as of the government's 1999 Census Report) hail from Jamaica (8,320), the United Kingdom (2,392), the United States (2,040), Canada (1,562), Honduras (873) and 800 Indian citizens. Approximately 3,300 more residents are citizens of various other countries. While the government does not restrict foreign land ownership, it does strongly enforce its immigration laws. Businesses are required to grant access to job openings to Caymanian citizens first; if none of them are suitable, the business may then seek employees from other countries. In order to work in the Cayman Islands, foreigners must have a job offer before immigrating. The estimated population of is (). The vast majority of its residents live on the island of Grand Cayman. According to the 2010 census only 2,277 people lived on Cayman Brac or Little Cayman. According to the Cayman Islands 2010 census the estimated resident population is 54,878 people, broken down as follows: Structure of the population (31.12.2013) (estimates): Although many Caribbean islands were initially populated by Amerindian groups such as the Taíno and Caribs, no evidence of this has been found in the Cayman Islands. Therefore, native Caymanians do not have any Amerindian heritage from their own islands; however, a significant number of Jamaicans have settled in the Cayman Islands over the years, so they and their descendants may have some Amerindian blood via Jamaica. Slavery was less common on the Cayman Islands than in many other parts of the Caribbean, resulting in a more even division of African and European ancestry. Those of mixed race make up 40% of the population, with blacks and whites following at 20% each. The remaining 20% belong to various immigrant ethnic groups. According to CIA factbook of 2013, Caymanian people of mixed-race of mixed black African and white European ancestry are the plurality ethnic group in the Cayman Islands, accounting for 40% of the country's population, with white 20%, black 20%, and expatriates of various ethnic groups 20%. The official language of the Cayman Islands is English. Islanders' accents retain elements passed down from English, Scottish, and Welsh settlers (among others) in a language variety known as Cayman Creole. Caymanians of Jamaican origin speak in their own vernacular (see Jamaican Creole and Jamaican English). It is also quite commonplace to hear some residents converse in Spanish as many citizens have relocated from Latin America to work and live on Grand Cayman. The Latin American nations with greatest representation are Honduras, Cuba, Colombia, Nicaragua, and the Dominican Republic. Spanish speakers comprise approximately between 10-12% of the population and is predominantly of the Caribbean dialect. Tagalog is spoken by about 5% of inhabitants most of whom are Filipino residents on work permits. The predominant religion on the Cayman Islands is Christianity. Collectively, Protestants account for 67.8%, while the Roman Catholic Church accounts for 14.1%. Denominations practiced include United Church, Church of God, Anglican Church, Baptist Church, Roman Catholic Church, Seventh-day Adventist Church, and Pentecostal Church. Roman Catholic churches are St. Ignatius Church, George Town and Stella Maris Church, Cayman Brac. Many citizens are deeply religious, regularly going to church. Ports are closed on Sundays and Christian holidays. There are places of worship in George Town for Jehovah's Witnesses and followers of the Bahá'í faith. The Cayman Islands also hosts a growing Jewish community. Hinduism is a minority religion in the Cayman Islands and is one the smallest religions. Although it is unknown as to when Hinduism was introduced to the Cayman Islands. There is no Hindu temple located in the Cayman Islands, but there is at least one home which is set aside for the purpose of worship. There were only 98 Hindus in the Caymans according to the 2000 census (about 0.25% of the population). In the 2008 census, the number of Hindus increased to 510 (1% of the total population). The 2010 Census showed the number of Hindus decreasing to 454 (0.8% of the total Cayman Islands population).
https://en.wikipedia.org/wiki?curid=5471
Economy of the Cayman Islands The economy of the Cayman Islands, a British overseas territory located in the western Caribbean Sea, is mainly fueled by the tourism sector and by the financial services sector, together representing 50–60 percent of the country's gross domestic product (GDP). The Cayman Islands Investment Bureau, a government agency, has been established with the mandate of promoting investment and economic development in the territory. The emergence of what are now considered the Cayman Islands' "twin pillars of economic development" (tourism and international finance) started in the 1950s with the introduction of modern transportation and telecommunications. From the earliest settlement of the Cayman Islands, economic activity was hindered by isolation and a limited natural resource base. The harvesting of sea turtles to resupply passing sailing ships was the first major economic activity on the islands, but local stocks were depleted by the 1790s. Agriculture, while sufficient to support the small early settler population, has always been limited by the scarcity of arable land. Fishing, shipbuilding, and cotton production boosted the economy during the early days of settlement. In addition, settlers scavenged shipwreck remains from the surrounding coral reefs. The boom in the Cayman Islands' international finance industry can also be at least partly attributed to the British overseas territory having no direct taxation. A popular legend attributes the tax-free status to the heroic acts of the inhabitants during a maritime tragedy in 1794, often referred to as "Wreck of the Ten Sails". The wreck involved nine British merchant vessels and their naval escort, the frigate HMS Convert, that ran aground on the reefs off Grand Cayman. Due to the rescue efforts by the Caymanians using canoes, the loss of life was limited to eight. However, records from the colonial era indicate that Cayman Islands, then a dependency of Jamaica, was not tax-exempt during the period that followed. In 1803, the inhabitants signed a petition addressed to the Jamaican governor asking him to grant them a tax exemption from the "Transient Tax on Wreck Goods". Sir Vassel Johnson, who became the only Caymanian ever knighted, was a pioneer of Cayman’s financial services industry. Cayman Islands Past Governor Stuart Jack said ‘As one of the architects of modern Cayman, especially the financial industry, Sir Vassel guided the steady growth of these Islands as the first financial secretary. His remarkable vision set the foundation for the prosperity and economic stability of these islands. Without his input, Cayman might well have remained the islands that time forgot.’ The Cayman Islands' tax-free status has attracted numerous banks and other companies to its shores. More than 92,000 companies were registered in the Cayman Islands as of 2014, including almost 600 banks and trust companies, with banking assets exceeding $500 billion. Numerous large corporations are based in the Cayman Islands, including, for example, Semiconductor Manufacturing International Corporation (SMIC). The Cayman Islands Stock Exchange was opened in 1997. The Cayman Islands is a major international financial centre. The largest sectors are "banking, hedge fund formation and investment, structured finance and securitisation, captive insurance, and general corporate activities". Regulation and supervision of the financial services industry is the responsibility of the Cayman Islands Monetary Authority (CIMA). Sir Vassel Johnson was a pioneer of Cayman’s financial services industry. Sir Vassel, who became the only Caymanian ever knighted in 1994, served as the Cayman Islands financial secretary from 1965 through 1982 and then as an Executive Council member from 1984 through 1988. In his government roles, Sir Vassel was a driving force in shaping the Cayman Islands financial services industry. The Cayman Islands is the fifth-largest banking centre in the world, with $1.5 trillion in banking liabilities . In March 2017 there were 158 banks, 11 of which were licensed to conduct banking activities with domestic (Cayman-based) and international clients, and the remaining 147 were licensed to operate on an international basis with only limited domestic activity. Financial services generated KYD$1.2 billion of GDP in 2007 (55% of the total economy), 36% of all employment and 40% of all government revenue. In 2010, the country ranked fifth internationally in terms of value of liabilities booked and sixth in terms of assets booked. It has branches of 40 of the world's 50 largest banks. The Cayman Islands is the second largest captive domicile (Bermuda is largest) in the world with more than 700 captives, writing more than US$7.7 billion of premiums and with US$36.8 billion of assets under management. There are a number of service providers. These include global financial institutions including HSBC, Deutsche Bank, UBS, and Goldman Sachs; over 80 administrators, leading accountancy practices (incl. the Big Four auditors), and offshore law practices including Maples & Calder. They also include wealth management such as Rothschilds private banking and financial advice. Since the introduction of the Mutual Funds Law in 1993, which has been copied by jurisdictions around the world, the Cayman Islands has grown to be the world's leading offshore hedge fund jurisdiction. In June 2008, it passed 10,000 hedge fund registrations, and over the year ending June 2008 CIMA reported a net growth rate of 12% for hedge funds. Starting in the mid-late 1990s, offshore financial centres, such as the Cayman Islands, came under increasing pressure from the OECD for their allegedly harmful tax regimes, where the OECD wished to prevent low-tax regimes from having an advantage in the global marketplace. The OECD threatened to place the Cayman Islands and other financial centres on a "black list" and impose sanctions against them. However, the Cayman Islands successfully avoided being placed on the OECD black list in 2000 by committing to regulatory reform to improve transparency and begin information exchange with OECD member countries about their citizens. In 2004, under pressure from the UK, the Cayman Islands agreed in principle to implement the European Union Savings Directive (EUSD), but only after securing some important benefits for the financial services industry in the Cayman Islands. As the Cayman Islands is not subject to EU laws, the implementation of the EUSD is by way of bilateral agreements between each EU member state and the Cayman Islands. The government of the Cayman Islands agreed on a model agreement, which set out how the EUSD would be implemented with the Cayman Islands. A report published by the International Monetary Fund (IMF), in March 2005, assessing supervision and regulation in the Cayman Islands' banking, insurance and securities industries, as well as its money laundering regime, recognised the jurisdiction's comprehensive regulatory and compliance frameworks. "An extensive program of legislative, rule and guideline development has introduced an increasingly effective system of regulation, both formalizing earlier practices and introducing enhanced procedures", noted IMF assessors. The report further stated that "the supervisory system benefits from a well-developed banking infrastructure with an internationally experienced and qualified workforce as well as experienced lawyers, accountants and auditors", adding that, "the overall compliance culture within Cayman is very strong, including the compliance culture related to AML (anti-money laundering) obligations". On 4 May 2009, the United States President, Barack Obama, declared his intentions to curb the use of financial centres by multinational corporations. In his speech, he singled out the Cayman Islands as a tax shelter. The next day, the Cayman Island Financial Services Association submitted an open letter to the president detailing the Cayman Islands' role in international finance and its value to the US financial system. The Cayman Islands was ranked as the world's second most significant tax haven on the Tax Justice Network's "Financial Secrecy Index" from 2011, scoring slightly higher than Luxembourg and falling behind only Switzerland. In 2013, the Cayman Islands was ranked by the Financial Secrecy Index as the fourth safest tax haven in the world, behind Hong Kong but ahead of Singapore. In the first conviction of a non-Swiss financial institution for US tax evasion conspiracy, two Cayman Islands financial institutions pleaded guilty in Manhattan Federal Court in 2016 to conspiring to hide more than $130 million in Cayman Islands bank accounts. The companies admitted to helping US clients hide assets in offshore accounts, and agreed to produce account files of non-compliant US taxpayers. On 30 June 2014, the tax jurisdiction of the Cayman Islands was deemed to have an inter-governmental agreement (IGA) with the United States of America with respect to the "Foreign Account Tax Compliance Act" of the United States of America. The Model 1 Agreement recognizes: On 26 March 2017, the US Treasury site disclosed that the Model 1 agreement and related agreement were "In Force" on 1 July 2014. Under the UK Sanctions and Anti-Money Laundering Act of 2018, beneficial ownership of companies in British overseas territories such as the Cayman Islands must be publicly registered for disclosure by 31 December 2020. The Government of the Cayman Islands plans to challenge this law, arguing that it violates the Constitutional sovereignty granted to the islands. The British National Crime Agency said in September 2018 that the authorities in the Cayman Islands were not supplying information about the beneficial ownership of firms registered in the Cayman Islands. Tourism is also a mainstay, accounting for about 70% of GDP and 75% of foreign currency earnings. The tourist industry is aimed at the luxury market and caters mainly to visitors from North America. Unspoiled beaches, duty-free shopping, scuba diving, and deep-sea fishing draw almost a million visitors to the islands each year. Due to the well-developed tourist industry, many citizens work in service jobs in that sector. The Cayman Islands is seeking to diversify beyond its two traditional industries, and invest in health care and technology. Health City Cayman Islands, opened in 2014, is a medical tourism hospital in East End, led by surgeon Devi Shetty. Cayman Enterprise City is a special economic zone that was opened in 2011 for technology, finance, and education investment. Cayman Sea Salt (producing gourmet sea salt) and Cayman Logwood products are now made in the Cayman Islands. Because the islands cannot produce enough goods to support the population, about 90% of their food and consumer goods must be imported. In addition, the islands have few natural fresh water resources. Desalination of sea water is used to solve this. Despite those challenges, the Caymanians enjoy one of the highest outputs per capita and one of the highest standards of living in the world. Education is compulsory to the age of 16 and is free to all Caymanian children. Most schools follow the British educational system. Ten primary, one special education and two highs schools ('junior high and senior high') are operated by the government, along with eight private high schools. In addition, there is a law school, a university-college and a medical school. Poverty relief is provided by the Needs Assessment Unit, a government agency established by the Poor Persons (Relief) Law in January 1964. GDP: purchasing power parity - $1.939 billion (2004 est.) GDP - real growth rate: 1.7% (2005 est.) GDP - per capita: purchasing power parity - $43,800 (2004 est.) GDP - composition by sector: "agriculture:" 1.4% "industry:" 3.2% "services:" 95.4% (1994 est.) Population below poverty line: NA% Household income or consumption by percentage share: "lowest 10%:" NA% "highest 10%:" NA% Inflation rate (consumer prices): 4.4% (2004) Labour force: 23,450 (2004) Labour force - by occupation: agriculture 1.4%, industry 12.6%, services 86% (1995) Unemployment rate: 1.7% (2010) Budget: "revenues:" $423.8 million "expenditures:" $392.6 million, including capital expenditures of $NA (1997) Industries: tourism, banking, insurance and finance, construction, construction materials, furniture Industrial production growth rate: NA% Electricity - production: 441.9 million kWh (2003) Electricity - production by source: "fossil fuel:" 100% "hydro:" 0% "nuclear:" 0% "other:" 0% (1998) Electricity - consumption: 411 million kWh (2003) Electricity - exports: 0 kWh (1998) Electricity - imports: 0 kWh (1998) Agriculture - products: vegetables, fruit; livestock, turtle, sea salt farming Exports: $2.52 million (2004) Exports - commodities: turtle products, sea salt, manufactured consumer goods Exports - partners: mostly US (2004) Imports: $866.9 million (2004) Imports - commodities: foodstuffs, manufactured goods Imports - partners: UK, US, Netherlands Antilles, Japan (2004) Debt - external: $70 million (1996) Economic aid - recipient: $NA Currency: 1 Cayman Islands dollar (CI$) = 100 cents Exchange rates: Caymanian dollars per US dollar - 0.80 Fiscal year: 1 April–31 March
https://en.wikipedia.org/wiki?curid=5473
Telecommunications in the Cayman Islands Communications in the Cayman Islands Telephones - main lines in use: 37,400 (2009) Telephones - mobile cellular: 99,900 (2004) Telephone system: Radio broadcast stations: AM 0, FM 17, shortwave 0 (2006) Radios: 36,000 (1997) Television broadcast stations: Cayman27 Internet Service Providers (ISPs): 3 (2009) Cable & Wireless, Logic, Tele Cayman Country code (Top-level domain): .ky
https://en.wikipedia.org/wiki?curid=5474
Transport in the Cayman Islands The transport infrastructure of the Cayman Islands consists of a public road network, two seaports, and three airports. As of 2000, the Cayman Islands had a total of 488 miles (785 km) of paved highway. Driving is on the left, and speed is reckoned in miles per hour, same as UK. Two ports, Cayman Brac and George Town, serve the islands. One hundred and twenty-three ships (of 1,000 GT or more) are registered in the Cayman Islands, with a total capacity of 2,402,058 GT/. Some foreign ships (including vessels from Cyprus, Denmark, Greece, Norway, the UK, and US) are registered in the Cayman Islands under a flag of convenience. (All figures are 2002 estimates.) There are three airports on the Islands. The main airport Owen Roberts International Airport serving Grand Cayman, Charles Kirkconnell International Airport serving Cayman Brac and Edward Bodden Airfield serving Little Cayman. A fleet of Share taxi minibuses serves Grand Cayman. A daily service starts at 6.00 from the depot and runs as follows from George Town to: Colour-coded logos on the front and rear of the buses (white mini-vans) identify the routes:
https://en.wikipedia.org/wiki?curid=5475
Central African Republic The Central African Republic (CAR; Sango: "Ködörösêse tî Bêafrîka";   , or "" ) is a landlocked country in Central Africa. It is bordered by Chad to the north, Sudan to the northeast, South Sudan to the southeast, the Democratic Republic of the Congo to the south, the Republic of the Congo to the southwest and Cameroon to the west. The CAR covers a land area of about and had an estimated population of around /1e6 round 1 million . , the CAR is the scene of a civil war, ongoing since 2012. Most of the CAR consists of Sudano-Guinean savannas, but the country also includes a Sahelo-Sudanian zone in the north and an equatorial forest zone in the south. Two thirds of the country is within the Ubangi River basin (which flows into the Congo), while the remaining third lies in the basin of the Chari, which flows into Lake Chad. What is today the Central African Republic has been inhabited for millennia; however, the country's current borders were established by France, which ruled the country as a colony starting in the late 19th century. After gaining independence from France in 1960, the Central African Republic was ruled by a series of autocratic leaders, including an abortive attempt at a monarchy; by the 1990s, calls for democracy led to the first multi-party democratic elections in 1993. Ange-Félix Patassé became president, but was later removed by General François Bozizé in the 2003 coup. The Central African Republic Bush War began in 2004 and, despite a peace treaty in 2007 and another in 2011, civil war resumed in 2012, perpetuating the country's poor human rights record, characterized by widespread and increasing abuses by various participating armed groups, such as arbitrary imprisonment, torture and restrictions on freedom of the press and freedom of movement. Despite its significant mineral deposits and other resources, such as uranium reserves, crude oil, gold, diamonds, cobalt, lumber, and hydropower, as well as significant quantities of arable land, the Central African Republic is among the ten poorest countries in the world, with the lowest GDP per capita at purchasing power parity in the world as of 2017. , according to the Human Development Index (HDI), the country had the second lowest level of human development, ranking 188th out of 189 countries. It is also estimated to be the unhealthiest country as well as the worst country in which to be young. The Central African Republic is a member of the United Nations, the African Union, the Economic Community of Central African States, the Organisation internationale de la Francophonie and the Non-Aligned Movement. Approximately 10,000 years ago, desertification forced hunter-gatherer societies south into the Sahel regions of northern Central Africa, where some groups settled. Farming began as part of the Neolithic Revolution. Initial farming of white yam progressed into millet and sorghum, and before 3000 BC the domestication of African oil palm improved the groups' nutrition and allowed for expansion of the local populations. This Agricultural Revolution, combined with a "Fish-stew Revolution", in which fishing began to take place, and the use of boats, allowed for the transportation of goods. Products were often moved in ceramic pots, which are the first known examples of artistic expression from the region's inhabitants. The Bouar Megaliths in the western region of the country indicate an advanced level of habitation dating back to the very late Neolithic Era (c. 3500–2700 BC). Ironworking arrived in the region around 1000 BC. Central Sudanic-speaking people settled along the Ubangi River in what is today Central and East CAR. while Bantoid peoples migrated from the southwest from Cameroon. Bananas arrived in the region during the first millennium BCE and added an important source of carbohydrates to the diet; they were also used in the production of alcoholic beverages. Production of copper, salt, dried fish, and textiles dominated the economic trade in the Central African region. During the 16th and 17th centuries slave traders began to raid the region as part of the expansion of the Saharan and Nile River slave routes. Their captives were enslaved and shipped to the Mediterranean coast, Europe, Arabia, the Western Hemisphere, or to the slave ports and factories along the West and North Africa or South along the Ubanqui and Congo rivers. In the mid 19th century, the Bobangi people became major slave traders and sold their captives to the Americas using the Ubangi river to reach the coast. During the 18th century Bandia-Nzakara peoples established the Bangassou Kingdom along the Ubangi River. In 1875, the Sudanese sultan Rabih az-Zubayr governed Upper-Oubangui, which included present-day CAR. The European invasion of Central African territory began in the late 19th century during the Scramble for Africa. Europeans, primarily the French, Germans, and Belgians, arrived in the area in 1885. France seized and colonized Ubangi-Shari territory in 1894. In 1911 at the Treaty of Fez, France ceded a nearly 300,000 km² portion of the Sangha and Lobaye basins to the German Empire which ceded a smaller area (in present-day Chad) to France. After World War I France again annexed the territory. Modeled on King Leopold's Congo Free State, concessions were doled out to private companies that endeavored to strip the region's assets as quickly and cheaply as possible before depositing a percentage of their profits into the French treasury. The concessionary companies forced local people to harvest rubber, coffee, and other commodities without pay and held their families hostage until they met their quotas. Between 1890, a year after the French first arrived, and 1940, the population declined by half due to diseases, famine and exploitation by private companies. In 1920 French Equatorial Africa was established and Ubangi-Shari was administered from Brazzaville. During the 1920s and 1930s the French introduced a policy of mandatory cotton cultivation, a network of roads was built, attempts were made to combat sleeping sickness, and Protestant missions were established to spread Christianity. New forms of forced labor were also introduced and a large number of Ubangians were sent to work on the Congo-Ocean Railway. Through the period of construction until 1934 there was a continual heavy cost in human lives, with total deaths among all workers along the railway estimated in excess of 17,000 of the construction workers, from a combination of both industrial accidents and diseases including malaria. In 1928, a major insurrection, the Kongo-Wara rebellion or 'war of the hoe handle', broke out in Western Ubangi-Shari and continued for several years. The extent of this insurrection, which was perhaps the largest anti-colonial rebellion in Africa during the interwar years, was carefully hidden from the French public because it provided evidence of strong opposition to French colonial rule and forced labor. In September 1940, during the Second World War, pro-Gaullist French officers took control of Ubangi-Shari and General Leclerc established his headquarters for the Free French Forces in Bangui. In 1946 Barthélémy Boganda was elected with 9,000 votes to the French National Assembly, becoming the first representative of the CAR in the French government. Boganda maintained a political stance against racism and the colonial regime but gradually became disheartened with the French political system and returned to CAR to establish the Movement for the Social Evolution of Black Africa ("Mouvement pour l'évolution sociale de l'Afrique noire", MESAN) in 1950. In the Ubangi-Shari Territorial Assembly election in 1957, MESAN captured 347,000 out of the total 356,000 votes, and won every legislative seat, which led to Boganda being elected president of the Grand Council of French Equatorial Africa and vice-president of the Ubangi-Shari Government Council. Within a year, he declared the establishment of the Central African Republic and served as the country's first prime minister. MESAN continued to exist, but its role was limited. After Boganda's death in a plane crash on 29 March 1959, his cousin, David Dacko, took control of MESAN and became the country's first president after the CAR had formally received independence from France. Dacko threw out his political rivals, including former Prime Minister and Mouvement d'évolution démocratique de l'Afrique centrale (MEDAC), leader Abel Goumba, whom he forced into exile in France. With all opposition parties suppressed by November 1962, Dacko declared MESAN as the official party of the state. On 31 December 1965, Dacko was overthrown in the Saint-Sylvestre coup d'état by Colonel Jean-Bédel Bokassa, who suspended the constitution and dissolved the National Assembly. President Bokassa declared himself President for Life in 1972, and named himself Emperor Bokassa I of the Central African Empire (as the country was renamed) on 4 December 1976. A year later, Emperor Bokassa crowned himself in a lavish and expensive ceremony that was ridiculed by much of the world. In April 1979, young students protested against Bokassa's decree that all school attendees would need to buy uniforms from a company owned by one of his wives. The government violently suppressed the protests, killing 100 children and teenagers. Bokassa himself may have been personally involved in some of the killings. In September 1979, France overthrew Bokassa and restored Dacko to power (subsequently restoring the name of the country and the original government to the Central African Republic). Dacko, in turn, was again overthrown in a coup by General André Kolingba on 1 September 1981. Kolingba suspended the constitution and ruled with a military junta until 1985. He introduced a new constitution in 1986 which was adopted by a nationwide referendum. Membership in his new party, the Rassemblement Démocratique Centrafricain (RDC), was voluntary. In 1987 and 1988, semi-free elections to parliament were held, but Kolingba's two major political opponents, Abel Goumba and Ange-Félix Patassé, were not allowed to participate. By 1990, inspired by the fall of the Berlin Wall, a pro-democracy movement arose. Pressure from the United States, France, and from a group of locally represented countries and agencies called GIBAFOR (France, the US, Germany, Japan, the EU, the World Bank, and the UN) finally led Kolingba to agree, in principle, to hold free elections in October 1992 with help from the UN Office of Electoral Affairs. After using the excuse of alleged irregularities to suspend the results of the elections as a pretext for holding on to power, President Kolingba came under intense pressure from GIBAFOR to establish a "Conseil National Politique Provisoire de la République" (Provisional National Political Council, CNPPR) and to set up a "Mixed Electoral Commission", which included representatives from all political parties. When a second round of elections were finally held in 1993, again with the help of the international community coordinated by GIBAFOR, Ange-Félix Patassé won in the second round of voting with 53% of the vote while Goumba won 45.6%. Patassé's party, the "Mouvement pour la Libération du Peuple Centrafricain" (MLPC) or Movement for the Liberation of the Central African People, gained a simple but not an absolute majority of seats in parliament, which meant Patassé's party required coalition partners. Patassé purged many of the Kolingba elements from the government and Kolingba supporters accused Patassé's government of conducting a "witch hunt" against the Yakoma. A new constitution was approved on 28 December 1994 but had little impact on the country's politics. In 1996–1997, reflecting steadily decreasing public confidence in the government's erratic behaviour, three mutinies against Patassé's administration were accompanied by widespread destruction of property and heightened ethnic tension. During this time (1996) the Peace Corps evacuated all its volunteers to neighboring Cameroon. To date, the Peace Corps has not returned to the Central African Republic. The Bangui Agreements, signed in January 1997, provided for the deployment of an inter-African military mission, to Central African Republic and re-entry of ex-mutineers into the government on 7 April 1997. The inter-African military mission was later replaced by a U.N. peacekeeping force (MINURCA). Since 1997, the country has hosted almost a dozen peacekeeping interventions, earning it the title of "world champion of peacekeeping". In 1998, parliamentary elections resulted in Kolingba's RDC winning 20 out of 109 seats. The next year, however, in spite of widespread public anger in urban centers over his corrupt rule, Patassé won a second term in the presidential election. On 28 May 2001, rebels stormed strategic buildings in Bangui in an unsuccessful coup attempt. The army chief of staff, Abel Abrou, and General François N'Djadder Bedaya were killed, but Patassé regained the upper hand by bringing in at least 300 troops of the Congolese rebel leader Jean-Pierre Bemba and Libyan soldiers. In the aftermath of the failed coup, militias loyal to Patassé sought revenge against rebels in many neighborhoods of Bangui and incited unrest including the murder of many political opponents. Eventually, Patassé came to suspect that General François Bozizé was involved in another coup attempt against him, which led Bozizé to flee with loyal troops to Chad. In March 2003, Bozizé launched a surprise attack against Patassé, who was out of the country. Libyan troops and some 1,000 soldiers of Bemba's Congolese rebel organization failed to stop the rebels and Bozizé's forces succeeded in overthrowing Patassé. François Bozizé suspended the constitution and named a new cabinet, which included most opposition parties. Abel Goumba was named vice-president, which gave Bozizé's new government a positive image. Bozizé established a broad-based National Transition Council to draft a new constitution, and announced that he would step down and run for office once the new constitution was approved. In 2004, the Central African Republic Bush War began, as forces opposed to Bozizé took up arms against his government. In May 2005, Bozizé won the presidential election, which excluded Patassé, and in 2006 fighting continued between the government and the rebels. In November 2006, Bozizé's government requested French military support to help them repel rebels who had taken control of towns in the country's northern regions. Though the initially public details of the agreement pertained to logistics and intelligence, by December the French assistance included airstrikes by Dassault Mirage 2000 fighters against rebel positions. The Syrte Agreement in February and the Birao Peace Agreement in April 2007 called for a cessation of hostilities, the billeting of FDPC fighters and their integration with FACA, the liberation of political prisoners, integration of FDPC into government, an amnesty for the UFDR, its recognition as a political party, and the integration of its fighters into the national army. Several groups continued to fight but other groups signed on to the agreement, or similar agreements with the government (e.g. UFR on 15 December 2008). The only major group not to sign an agreement at the time was the CPJP, which continued its activities and signed a peace agreement with the government on 25 August 2012. In 2011, Bozizé was reelected in an election which was widely considered fraudulent. In November 2012, Séléka, a coalition of rebel groups, took over towns in the northern and central regions of the country. These groups eventually reached a peace deal with the Bozizé's government in January 2013 involving a power sharing government but this deal broke down and the rebels seized the capital in March 2013 and Bozizé fled the country. Michel Djotodia took over as president. Prime Minister Nicolas Tiangaye requested a UN peacekeeping force from the UN Security Council and on 31 May former President Bozizé was indicted for crimes against humanity and incitement of genocide. By the end of the year there were international warnings of a "genocide" and fighting was largely from reprisal attacks on civilians from Seleka's predominantly Muslim fighters and Christian militias called "anti-balaka." By August 2013, there were reports of over 200,000 internally displaced persons (IDPs) French President François Hollande called on the UN Security Council and African Union to increase their efforts to stabilize the country. On 18 February 2014, United Nations Secretary-General Ban Ki-moon called on the UN Security Council to immediately deploy 3,000 troops to the country, bolstering the 6,000 African Union soldiers and 2,000 French troops already in the country, to combat civilians being murdered in large numbers. The "Séléka" government was said to be divided, and in September 2013, Djotodia officially disbanded Seleka, but many rebels refused to disarm, becoming known as ex-Seleka, and veered further out of government control. It is argued that the focus of the initial disarmament efforts exclusively on the Seleka inadvertently handed the anti-Balaka the upper hand, leading to the forced displacement of Muslim civilians by anti-Balaka in Bangui and western CAR. On 11 January 2014, Michael Djotodia and Nicolas Tiengaye resigned as part of a deal negotiated at a regional summit in neighboring Chad. Catherine Samba-Panza was elected as interim president by the National Transitional Council, becoming the first ever female Central African president. On 23 July 2014, following Congolese mediation efforts, Séléka and anti-balaka representatives signed a ceasefire agreement in Brazzaville. By the end of 2014, the country was de facto partitioned with the anti-Balaka in the southwest and ex-Seleka in the northeast. In March 2015, Samantha Power, the U.S. ambassador to the United Nations, said 417 of the country's 436 mosques had been destroyed, and Muslim women were so scared of going out in public they were giving birth in their homes instead of going to the hospital. On 14 December 2015, Séléka rebel leaders declared an independent Republic of Logone. The Central African Republic is a landlocked nation within the interior of the African continent. It is bordered by Cameroon, Chad, Sudan, South Sudan, the Democratic Republic of the Congo, and the Republic of the Congo. The country lies between latitudes 2° and 11°N, and longitudes 14° and 28°E. Much of the country consists of flat or rolling plateau savanna approximately above sea level. Most of the northern half lies within the World Wildlife Fund's East Sudanian savanna ecoregion. In addition to the Fertit Hills in the northeast of the CAR, there are scattered hills in the southwest regions. In the northwest is the Yade Massif, a granite plateau with an altitude of . At , the Central African Republic is the world's 44th-largest country. It is comparable in size to Ukraine, as the Ukraine is in area, according to List of countries and dependencies by area. Much of the southern border is formed by tributaries of the Congo River; the Mbomou River in the east merges with the Uele River to form the Ubangi River, which also comprises portions of the southern border. The Sangha River flows through some of the western regions of the country, while the eastern border lies along the edge of the Nile River watershed. It has been estimated that up to 8% of the country is covered by forest, with the densest parts generally located in the southern regions. The forests are highly diverse and include commercially important species of Ayous, Sapelli and Sipo. The deforestation rate is about 0.4% per annum, and lumber poaching is commonplace. In 2008, Central African Republic was the world's least light pollution affected country. The Central African Republic is the focal point of the Bangui Magnetic Anomaly, one of the largest magnetic anomalies on Earth. In the southwest, the Dzanga-Sangha National Park is located in a rain forest area. The country is noted for its population of forest elephants and western lowland gorillas. In the north, the Manovo-Gounda St Floris National Park is well-populated with wildlife, including leopards, lions, cheetahs and rhinos, and the Bamingui-Bangoran National Park is located in the northeast of CAR. The parks have been seriously affected by the activities of poachers, particularly those from Sudan, over the past two decades. The climate of the Central African Republic is generally tropical, with a wet season that lasts from June to September in the northern regions of the country, and from May to October in the south. During the wet season, rainstorms are an almost daily occurrence, and early morning fog is commonplace. Maximum annual precipitation is approximately in the upper Ubangi region. The northern areas are hot and humid from February to May, but can be subject to the hot, dry, and dusty trade wind known as the Harmattan. The southern regions have a more equatorial climate, but they are subject to desertification, while the extreme northeast regions of the country are a steppe. The Central African Republic is divided into 16 administrative prefectures ("préfectures"), two of which are economic prefectures ("préfectures economiques"), and one an autonomous commune; the prefectures are further divided into 71 sub-prefectures ("sous-préfectures"). The prefectures are Bamingui-Bangoran, Basse-Kotto, Haute-Kotto, Haut-Mbomou, Kémo, Lobaye, Mambéré-Kadéï, Mbomou, Nana-Mambéré, Ombella-M'Poko, Ouaka, Ouham, Ouham-Pendé and Vakaga. The economic prefectures are Nana-Grébizi and Sangha-Mbaéré, while the commune is the capital city of Bangui. The population of the Central African Republic has almost quadrupled since independence. In 1960, the population was 1,232,000; as of a UN estimate, it is approximately . The United Nations estimates that approximately 4% of the population aged between 15 and 49 is HIV positive. Only 3% of the country has antiretroviral therapy available, compared to a 17% coverage in the neighbouring countries of Chad and the Republic of the Congo. The nation is divided into over 80 ethnic groups, each having its own language. The largest ethnic groups are the Baggara Arabs, Baka, Banda, Bayaka, Fula, Gbaya, Kara, Kresh, Mbaka, Mandja, Ngbandi, Sara, Vidiri, Wodaabe, Yakoma, Yulu, Zande, with others including Europeans of mostly French descent. According to the 2003 national census, 80.3% of the population was Christian—51.4% Protestant and 28.9% Roman Catholic—, 10% is Muslim and 4.5 percent other religious groups, with 5.5 percent having no religious beliefs. More recent work from the Pew Research Center estimated that, as of 2010, Christians constituted 89.8% of the population (with Protestantism at 60.7% and Catholicism 28.5%) while Muslims make up 8.9%. The Catholic Church claims over 1.5 million adherents, approximately one-third of the population. Indigenous belief (animism) is also practiced, and many indigenous beliefs are incorporated into Christian and Islamic practice. A UN director described religious tensions between Muslims and Christians as being high. There are many missionary groups operating in the country, including Lutherans, Baptists, Catholics, Grace Brethren, and Jehovah's Witnesses. While these missionaries are predominantly from the United States, France, Italy, and Spain, many are also from Nigeria, the Democratic Republic of the Congo, and other African countries. Large numbers of missionaries left the country when fighting broke out between rebel and government forces in 2002–3, but many of them have now returned to continue their work. According to Overseas Development Institute research, during the crisis ongoing since 2012, religious leaders have mediated between communities and armed groups; they also provided refuge for people seeking shelter. The Central African Republic's two official languages are French and Sango (also spelled Sangho), a creole developed as an inter-ethnic lingua franca based on the local Ngbandi language. CAR is one of the few African countries to have an African language as their official language. Basketball is the country's most popular sport and a good way to connect with its people. Its national team won the African Championship twice and was the first Sub-Saharan African team to qualify for the Basketball World Cup, in 1974. The country also has a national football team, which is governed by the Central African Football Federation, and stages matches at the Barthélemy Boganda Stadium. Politics in the Central African Republic formally take place in a framework of a semi-presidential republic. In this system, the President is the head of state, with a Prime Minister as head of government. Executive power is exercised by the government. Legislative power is vested in both the government and parliament. Changes in government have occurred in recent years by three methods: violence, negotiations, and elections. A new constitution was approved by voters in a referendum held on 5 December 2004. The government was rated 'Partly Free' from 1991 to 2001 and from 2004 to 2013. The president is elected by popular vote for a six-year term, and the prime minister is appointed by the president. The president also appoints and presides over the Council of Ministers, which initiates laws and oversees government operations. However, as of 2018 the official government is not in control of large parts of the country, which are governed by rebel groups. Acting president since April 2016 is Faustin Archange Touadera who followed the interim government under Catherine Samba-Panza, interim prime minister André Nzapayeké. The National Assembly ("Assemblée Nationale") has 140 members, elected for a five-year term using the two-round (or Run-off) system. As in many other former French colonies, the Central African Republic's legal system is based on French law. The Supreme Court, or "Cour Supreme", is made up of judges appointed by the president. There is also a Constitutional Court, and its judges are also appointed by the president. The Central African Republic relies heavily on Russian mercenaries for the protection of its diamond mines. The Central African Republic is heavily dependent upon foreign aid and numerous NGOs provide services that the government does not provide. In 2019, over $100 million USD in foreign aid was spent in the country, mostly on humanitarian assistance. In 2006, due to ongoing violence, over 50,000 people in the country's northwest were at risk of starvation but this was averted due to assistance from the United Nations. On 8 January 2008, the UN Secretary-General Ban Ki-Moon declared that the Central African Republic was eligible to receive assistance from the Peacebuilding Fund. Three priority areas were identified: first, the reform of the security sector; second, the promotion of good governance and the rule of law; and third, the revitalization of communities affected by conflicts. On 12 June 2008, the Central African Republic requested assistance from the UN Peacebuilding Commission, which was set up in 2005 to help countries emerging from conflict avoid devolving back into war or chaos. In response to concerns of a potential genocide, a peacekeeping force – the International Support Mission to the Central African Republic (MISCA) – was authorized in December 2013. This African Union force of 6,000 personnel was accompanied by the French Operation Sangaris. In 2017, Central African Republic signed the UN treaty on the Prohibition of Nuclear Weapons. The per capita income of the Republic is often listed as being approximately $400 a year, one of the lowest in the world, but this figure is based mostly on reported sales of exports and largely ignores the unregistered sale of foods, locally produced alcoholic beverages, diamonds, ivory, bushmeat, and traditional medicine. The currency of Central African Republic is the CFA franc, which is accepted across the former countries of French West Africa and trades at a fixed rate to the euro. Diamonds constitute the country's most important export, accounting for 40–55% of export revenues, but it is estimated that between 30% and 50% of those produced each year leave the country clandestinely. Agriculture is dominated by the cultivation and sale of food crops such as cassava, peanuts, maize, sorghum, millet, sesame, and plantain. The annual real GDP growth rate is just above 3%. The importance of food crops over exported cash crops is indicated by the fact that the total production of cassava, the staple food of most Central Africans, ranges between 200,000 and 300,000 tonnes a year, while the production of cotton, the principal exported cash crop, ranges from 25,000 to 45,000 tonnes a year. Food crops are not exported in large quantities, but still constitute the principal cash crops of the country, because Central Africans derive far more income from the periodic sale of surplus food crops than from exported cash crops such as cotton or coffee. Much of the country is self-sufficient in food crops; however, livestock development is hindered by the presence of the tsetse fly. The Republic's primary import partner is the Netherlands (19.5%). Other imports come from Cameroon (9.7%), France (9.3%), and South Korea (8.7%). Its largest export partner is Belgium (31.5%), followed by China (27.7%), the Democratic Republic of Congo (8.6%), Indonesia (5.2%), and France (4.5%). The CAR is a member of the Organization for the Harmonization of Business Law in Africa (OHADA). In the 2009 World Bank Group's report "Doing Business", it was ranked 183rd of 183 as regards 'ease of doing business', a composite index which takes into account regulations that enhance business activity and those that restrict it. Bangui is the transport hub of the Central African Republic. As of 1999, eight roads connected the city to other main towns in the country, Cameroon, Chad and South Sudan; of these, only the toll roads are paved. During the rainy season from July to October, some roads are impassable. River ferries sail from the river port at Bangui to Brazzaville and Zongo. The river can be navigated most of the year between Bangui and Brazzaville. From Brazzaville, goods are transported by rail to Pointe-Noire, Congo's Atlantic port. The river port handles the overwhelming majority of the country's international trade and has a cargo handling capacity of 350,000 tons; it has length of wharfs and of warehousing space. Bangui M'Poko International Airport is Central African Republic's only international airport. As of June 2014 it had regularly scheduled direct flights to Brazzaville, Casablanca, Cotonou, Douala, Kinshasha, Lomé, Luanda, Malabo, N'Djamena, Paris, Pointe-Noire, and Yaoundé. Since at least 2002 there have been plans to connect Bangui by rail to the Transcameroon Railway. The Central African Republic primarily uses hydroelectricity as there are few other resources for energy and power. The Central African Republic will likely be among the main winners after the global energy transition is completed as the country is ranked no. 7 out of 156 nations in the index of geopolitical gains and losses after energy transition (GeGaLo Index). Presently, the Central African Republic has active television services, radio stations, internet service providers, and mobile phone carriers; Socatel is the leading provider for both internet and mobile phone access throughout the country. The primary governmental regulating bodies of telecommunications are the Ministère des Postes and Télécommunications et des Nouvelles Technologies. In addition, the Central African Republic receives international support on telecommunication related operations from ITU Telecommunication Development Sector (ITU-D) within the International Telecommunication Union to improve infrastructure. Public education in the Central African Republic is free and is compulsory from ages 6 to 14. However, approximately half of the adult population of the country is illiterate. The two institutions of higher education in the Central African Republic are the University of Bangui, a public university located in Bangui, which includes a medical school; and Euclid University, an international university. The largest hospitals in the country are located in the Bangui district. As a member of the World Health Organization, the Central African Republic receives vaccination assistance, such as a 2014 intervention for the prevention of a measles epidemic. In 2007, female life expectancy at birth was 48.2 years and male life expectancy at birth was 45.1 years. Women's health is poor in the Central African Republic. , the country had the 4th highest maternal mortality rate in the world. The total fertility rate in 2014 was estimated at 4.46 children born/woman. Approximately 25% of women had undergone female genital mutilation. Many births in the country are guided by traditional birth attendants, who often have little or no formal training. Malaria is endemic in the Central African Republic, and one of the leading causes of death. According to 2009 estimates, the HIV/AIDS prevalence rate is about 4.7% of the adult population (ages 15–49). This is in general agreement with the 2016 United Nations estimate of approximately 4%. Government expenditure on health was US$20 (PPP) per person in 2006 and 10.9% of total government expenditure in 2006. There was only around 1 physician for every 20,000 persons in 2009. The 2009 Human Rights Report by the United States Department of State noted that human rights in CAR were poor and expressed concerns over numerous government abuses. The U.S. State Department alleged that major human rights abuses such as extrajudicial executions by security forces, torture, beatings and rape of suspects and prisoners occurred with impunity. It also alleged harsh and life-threatening conditions in prisons and detention centers, arbitrary arrest, prolonged pretrial detention and denial of a fair trial, restrictions on freedom of movement, official corruption, and restrictions on workers' rights. The State Department report also cites widespread mob violence, the prevalence of female genital mutilation, discrimination against women and Pygmies, human trafficking, forced labor, and child labor. Freedom of movement is limited in the northern part of the country "because of actions by state security forces, armed bandits, and other nonstate armed entities", and due to fighting between government and anti-government forces, many persons have been internally displaced. Violence against children and women in relation to accusations of witchcraft has also been cited as a serious problem in the country. Witchcraft is a criminal offense under the penal code. Freedom of speech is addressed in the country's constitution, but there have been incidents of government intimidation of the media. A report by the International Research & Exchanges Board's media sustainability index noted that "the country minimally met objectives, with segments of the legal system and government opposed to a free media system". Approximately 68% of girls are married before they turn 18, and the United Nations' Human Development Index ranked the country 188 out of 188 countries surveyed. The Bureau of International Labor Affairs has also mentioned it in its last edition of the "List of Goods Produced by Child Labor or Forced Labor".
https://en.wikipedia.org/wiki?curid=5478
History of the Central African Republic The history of the Central African Republic is roughly composed of four distinct periods. The earliest period of settlement began around 10,000 years ago when nomadic people first began to settle, farm and fish in the region. The next period began around 1,000 to 3,000 years ago when several non-indigenous groups began to migrate into the region from other parts of the continent. The third period involved the colonial conquest and rule of the country by France and Germany which spanned from the late 1800s until 1960 when the Central African Republic became an independent state. The final period has been the era during which the Central African Republic has been an independent state. Approximately 10,000 years ago, desertification forced hunter-gatherer societies south into the Sahel regions of northern Central Africa, where some groups settled and began farming as part of the Neolithic Revolution. Initial farming of white yam progressed into millet and sorghum, and then later the domestication of African oil palm improved the groups' nutrition and allowed for expansion of the local populations. Bananas arrived in the region and added an important source of carbohydrates to the diet; they were also used in the production of alcohol. This Agricultural Revolution, combined with a "Fish-stew Revolution", in which fishing began to take place, and the use of boats, allowed for the transportation of goods. Products were often moved in ceramic pots, which are the first known examples of artistic expression from the region's inhabitants. The Bouar Megaliths in the western region of the country indicate an advanced level of habitation dating back to the very late Neolithic Era (c. 3500-2700 BC). Ironworking arrived in the region by around 1000 BC, likely from early Bantu cultures in what is today southeast Nigeria and/or Cameroon. The site of Gbabiri (in the Central African Republic) has yielded evidence of iron metallurgy, from a reduction furnace and blacksmith workshop; with earliest dates of 896-773 BC and 907-796 BC respectively. Some earlier iron metallurgy dates of 2,000 BC from the site of Oboui (also in the Central Africa Republic) have also been proposed, but these are disputed by some archaeologists. During the Bantu Migrations from about 1000 BC to AD 1000, Ubangian-speaking people spread eastward from Cameroon to Sudan, Bantu-speaking people settled in the southwestern regions of the CAR, and Central Sudanic-speaking people settled along the Ubangi River in what is today Central and East CAR. Production of copper, salt, dried fish, and textiles dominated the economic trade in the Central African region. The territory of modern Central African Republic is known to have been settled from at least the 7th century on by overlapping empires, including the Kanem-Bornu, Ouaddai, Baguirmi, and Dafour groups based on the Lake Chad region and along the Upper Nile. During the 16th and 17th centuries Muslim slave traders began to raid the region and their captives were shipped to the Mediterranean coast, Europe, Arabia, the Western Hemisphere, or to the slave ports and factories along the West African coast. The Bobangi people became major slave traders and sold their captives to the Americas using the Ubangi river to reach the coast. During the 18th century Bandia-Nzakara peoples established the Bangassou Kingdom along the Ubangi river. Population migration in the 18th and 19th centuries brought new migrants into the area, including the Zande, Banda, and Baya-Mandjia. In 1875, the Sudanese sultan Rabih az-Zubayr governed Upper-Oubangui, which included present-day CAR. Europeans, primarily the French, German, and Belgians, arrived in the area in 1885. The French consolidated their legal claim to the area through an 1887 convention with Congo Free State, which granted France possession of the right bank of the Oubangui River. Two years later, the French established an outpost at Bangui, and in 1894, Oubangui-Chari became a French territory. However, the French did not consolidate their control over the area until 1903, after having defeated the forces of Rabih in the battle of Kousséri, and established colonial administration throughout the territory. In 1906, the Oubangui-Chari territory was united with the Chad colony; in 1910, it became one of the four territories of the Federation of French Equatorial Africa (AEF), along with Chad, Middle Congo, and Gabon. The next thirty years were marked by mostly small scale revolts against French rule and the development of a plantation-style economy. The largest of these revolts was the Kongo-Wara rebellion, when over 350,000 natives rebelled against the colonial administration. The European penetration of Central African territory began in the late 19th century during the Scramble for Africa. Count Savorgnan de Brazza established the French Congo and sent expeditions up the Ubangi River from Brazzaville in an effort to expand France's claims to territory in Central Africa. Belgium, Germany, and the United Kingdom also competed to establish their claims to territory in the region. In 1889, the French established a post on the Ubangi River at Bangui. In 1890–91, de Brazza sent expeditions up the Sangha River, in what is now south-western CAR, up the center of the Ubangi basin toward Lake Chad, and eastward along the Ubangi River toward the Nile, with the intention of expanding the borders of the French Congo to link up the other French territories in Africa. In 1894, the French Congo's borders with Leopold II of Belgium's Congo Free State and German Cameroon were fixed by diplomatic agreements. In 1899, the French Congo's border with Sudan was fixed along the Congo-Nile divide. This situation left France without her much coveted outlet on the Nile. Once European negotiators had agreed upon the borders of the French Congo, France had to decide how to pay for the costly occupation, administration, and development of the territory it had acquired. The reported financial successes of Leopold II's concessionary companies in the Congo Free State convinced the French government to grant 17 private companies large concessions in the Ubangi-Shari region in 1899. In return for the right to exploit these lands by buying local products and selling European goods, the companies promised to pay rent to France and to promote the development of their concessions. The companies employed European and African agents who frequently used brutal methods to force the natives to labor. At the same time, the French colonial administration began to force the local population to pay taxes and to provide the state with free labor. The companies and the French administration at times collaborated in forcing the Central Africans to work for them. Some French officials reported abuses committed by private company militias, and their own colonial colleagues and troops, but efforts to hold these people accountable almost invariably failed. When any news of atrocities committed against Central Africans reached France and caused an outcry, investigations were undertaken and some feeble attempts at reform were made, but the situation on the ground in Ubangi-Shari remained virtually unchanged. During the first decade of French colonial rule, from about 1900 to 1910, the rulers of the Ubangi-Shari region increased both their slave-raiding activities and the selling of local produce to Europe. They took advantage of their treaties with the French to procure more weapons, which were used to capture more slaves: much of the eastern half of Ubangi-Shari was depopulated as a result of slave-trading by local rulers during the first decade of colonial rule. After the power of local African rulers was destroyed by the French, slave raiding greatly diminished. In 1911, the Sangha and Lobaye basins were ceded to Germany as part of an agreement which gave France a free hand in Morocco. Western Ubangi-Shari remained under German rule until World War I, after which France again annexed the territory using Central African troops. From 1920 to 1930, a network of roads was built, cash crops were promoted and mobile health services were formed to combat sleeping sickness. Protestant missions were established in different parts of the country. New forms of forced labor were also introduced, however, as the French conscripted large numbers of Ubangians to work on the Congo-Ocean Railway; many of these recruits died of exhaustion and illness as a result of the poor conditions. In 1925, the French writer André Gide published "Voyage au Congo", in which he described the alarming consequences of conscription for the Congo-Ocean railroad. He exposed the continuing atrocities committed against Central Africans in Western Ubangi-Shari by such employers as the Forestry Company of Sangha-Ubangi. In 1928, a major insurrection, the Kongo-Wara rebellion or 'war of the hoe handle', broke out in Western Ubangi-Shari and continued for several years. The extent of this insurrection, which was perhaps the largest anti-colonial rebellion in Africa during the interwar years, was carefully hidden from the French public because it provided evidence of strong opposition to French colonial rule and forced labor. During the 1930s, cotton, tea, and coffee emerged as important cash crops in Ubangi-Shari and the mining of diamonds and gold began in earnest. Several cotton companies were granted purchasing monopolies over large areas of cotton production and were able to fix the prices paid to cultivators, which assured profits for their shareholders. In September 1940, during the Second World War, pro-Gaullist French officers took control of Ubangi-Shari. In August 1940, the territory responded, with the rest of the AEF, to the call from General Charles de Gaulle to fight for Free France. After World War II, the French Constitution of 1946 inaugurated the first of a series of reforms that led eventually to complete independence for all French territories in western and equatorial Africa. In 1946, all AEF inhabitants were granted French citizenship and allowed to establish local assemblies. The assembly in CAR was led by Barthélemy Boganda, a Catholic priest who also was known for his forthright statements in the French Assembly on the need for African emancipation. In 1956 French legislation eliminated certain voting inequalities and provided for the creation of some organs of self-government in each territory. The French constitutional referendum of September 1958 dissolved the AEF, and on 1 December of the same year the Assembly declared the birth of the autonomous Central African Republic with Boganda as head of government. Boganda ruled until his death in a plane crash on 29 March 1959. His cousin, David Dacko, replaced him as head of Government. On 12 July 1960 France agreed to the Central African Republic becoming fully independent. On 13 August 1960 the Central African Republic became an independent country and David Dacko became its first President. On 1 January 1966, following a swift and almost bloodless overnight coup, Colonel Jean-Bédel Bokassa assumed power as president of the Republic. Bokassa abolished the constitution of 1959, dissolved the National Assembly, and issued a decree that placed all legislative and executive powers in the hands of the president. On the 4 March 1972, Bokassa's presidency was extended to a life term. On 4 December 1976, the republic became a monarchy – the Central African Empire – with the promulgation of the imperial constitution and the coronation of the president as Emperor Bokassa I. His authoritarian regime was characterized by numerous human rights violations. On 20 September 1979, Dacko, with French support, led a bloodless coup that overthrew Bokassa while he was out of the country. The republic was restored, and Bokassa, who took refuge in Ivory Coast and France, was sentenced to death in absentia for various crimes, including cannibalism. Moreover, an African judicial commission reported that he had "almost certainly" taken part in the massacre of some 100 children for refusing to wear the compulsory school uniforms. In January 1981, six of his supporters, including two sons-in-law, were executed. Bokassa returned unexpectedly in October 1986 and on 12 June 1987, he was convicted of having ordered the murders of at least 20 prisoners and the arrest of the schoolchildren who were murdered. He was sentenced to death, but this was commuted to a life term in February 1988. He was released from prison on 1 September 1993, as a result of an amnesty and died of a heart attack in Bangui on 3 November 1996 at age 75. Dacko's efforts to promote economic and political reforms proved ineffectual, and on 20 September 1981, he in turn was overthrown in a bloodless coup by General André Kolingba. For four years, Kolingba led the country as head of the Military Committee for National Recovery (CRMN). Kolingba suspended the constitution and ruled with a military junta until 1985. In 1985, the CRMN was dissolved, and Kolingba named a new cabinet with increased civilian participation, signaling the start of a return to civilian rule. The process of democratization quickened in 1986 with the creation of a new political party, the Rassemblement Démocratique Centrafricain (RDC), and the drafting of a new constitution that subsequently was ratified in a national referendum. General Kolingba was sworn in as constitutional President on 29 November 1986. The constitution established a National Assembly made up of 52 elected deputies, elected in July 1987 and municipal elections were held in 1988. Kolingba's two major political opponents, Abel Goumba and Ange-Félix Patassé, boycotted these elections because their parties were not allowed to participate. By 1990, inspired by the fall of the Berlin Wall, a pro-democracy movement became very active. In May 1990, a letter signed by 253 prominent citizens asked for the convocation of a National Conference, but Kolingba refused this request and instead detained several opponents. Pressure from the United States, France, and from a group of locally represented countries and agencies called GIBAFOR (France, the US, Germany, Japan, the EU, the World Bank, and the UN) finally led Kolingba to agree, in principle, to hold free elections in October 1992 with help from the UN Office of Electoral Affairs. After using the excuse of alleged irregularities to suspend the results of the elections as a pretext for holding on to power, President Kolingba came under intense pressure from GIBAFOR to establish a "Conseil National Politique Provisoire de la République" (Provisional National Political Council, CNPPR) and to set up a "Mixed Electoral Commission", which included representatives from all political parties. When elections were finally held in 1993, again with the help of the international community, Ange-Félix Patassé led in the first round and Kolingba came in fourth behind Abel Goumba and David Dacko. In the second round, Patassé won 53% of the vote while Goumba won 45.6%. Most of Patassé's support came from Gbaya, Kare, and Kaba voters in seven heavily populated prefectures in the northwest while Goumba's support came largely from ten less-populated prefectures in the south and east. Furthermore, Patassé's party, the "Mouvement pour la Libération du Peuple Centrafricain" (MLPC) or Movement for the Liberation of the Central African People, gained a simple but not an absolute majority of seats in parliament, which meant Patassé's party required coalition partners. Patassé relieved former President Kolingba of his military rank of general in March 1994 and then charged several former ministers with various crimes. Patassé also removed many Yakoma from important, lucrative posts in the government. Two hundred predominantly Yakoma members of the presidential guard were also dismissed or reassigned to the army. Kolingba's RDC loudly proclaimed that Patassé's government was conducting a "witch hunt" against the Yakoma. A new constitution was approved on 28 December 1994 and promulgated on 14 January 1995, but this constitution, like those before it, did not have much impact on the country's politics. In 1996–1997, reflecting steadily decreasing public confidence in the government's erratic behaviour, three mutinies against Patassé's administration were accompanied by widespread destruction of property and heightened ethnic tension. On 25 January 1997, the Bangui Agreements, which provided for the deployment of an inter-African military mission, the "Mission Interafricaine de Surveillance des Accords de Bangui" (MISAB), were signed. Mali's former president, Amadou Touré, served as chief mediator and brokered the entry of ex-mutineers into the government on 7 April 1997. The MISAB mission was later replaced by a U.N. peacekeeping force, the "Mission des Nations Unies en RCA" (MINURCA). In 1998, parliamentary elections resulted in Kolingba's RDC winning 20 out of 109 seats, constituting a comeback. However, in 1999, in spite of widespread public anger in urban centers over his corrupt rule, Patassé won free elections to become president for a second term. On 28 May 2001, rebels stormed strategic buildings in Bangui in an unsuccessful coup attempt. The army chief of staff, Abel Abrou, and General François N'Djadder Bedaya were killed, but Patassé regained the upper hand by bringing in at least 300 troops of the rebel leader Jean-Pierre Bemba (from across the river in the Democratic Republic of the Congo) and by Libyan soldiers. In the aftermath of the failed coup, militias loyal to Patassé sought revenge against rebels in many neighborhoods of the capital, Bangui, and incited unrest which resulted in the destruction of many homes as well as the torture and murder of many opponents. Eventually, Patassé came to suspect that General François Bozizé was involved in another coup attempt against him, which led Bozizé to flee with loyal troops to Chad. In March 2003, Bozizé launched a surprise attack against Patassé, who was out of the country. Libyan troops and some 1,000 soldiers of Bemba's Congolese rebel organization failed to stop the rebels, who took control of the country and thus succeeded in overthrowing Patassé. On 15 March 2003, while Patassé was abroad, rebels moved into Bangui and installed their commander, General François Bozizé, as president. Bozizé has since been elected President in an election considered by observers to be fair and free. Patassé has been found guilty of major crimes in Bangui and CAR has brought a case to the International Criminal Court against him and Jean Pierre Bemba from the neighbouring Democratic Republic of Congo accusing them both of multiple crimes in suppressing one of the mutinies against Patasse. Civil tranquillity has yet to be established and parts of the country remain out of government control. The UN continues to maintain a peace building mission in Bangui. Despite these shortcomings, and his promise to step down at the end of the transition, Bozizé contested the 13 March 2005 presidential elections in which all of the leading opposition candidates were allowed to run except for Patassé. Bozizé won on the second run-off round on 8 May 2005, defeating Martin Ziguélé, who ran on the ticket of the MLPC, the former ruling party. The National Elections Commission declared Bozizé the winner with 64.6 percent of the vote to 35.4 percent for Ziguélé. The election was generally considered to be fair, although the absence of Patassé cast a shadow over the legitimacy of the process. After Bozizé seized power in 2003, the Central African Republic Bush War began with the rebellion by the Union of Democratic Forces for Unity (UFDR), led by Michel Djotodia. This quickly escalated into major fighting during 2004. The UFDR rebel forces consisted of three allies, the "Groupe d'action patriotique pour la liberation de Centrafrique" (GAPLC), the Convention of Patriots for Justice and Peace (CPJP), the People's Army for the Restoration of Democracy (APRD), the Movement of Central African Liberators for Justice (MLCJ), and the "Front démocratique Centrafricain" (FDC). On 8 May 2005, Bozizé gained yet a further victory when his coalition, Convergence Kwa Na Kwa, won 42 parliamentary seats in the legislative run-off vote. The MLPC came in second with 11 seats while the RDC won only eight seats. The remaining seats were won by independents or by smaller parties. In June later that year, the African Union (AU) lifted sanctions against the country, which had been applied after the 2003 coup. In early 2006, Bozizé's government appeared stable. However, Patassé, who was living in exile in Togo, could not be ruled out as a leader of a future uprising. His supporters reportedly were joining or were prepared to join rebel movements in belief that their leader was still the rightful head of state of the country. Further, members of Kolingba's Yakoma tribe in the south posed a potential threat to Bozizé's government because of their widespread boycott of the second round of the legislative elections. Members of the Yakoma dominate the army. On 13 April 2007, a peace agreement between the government and the UFDR was signed in Birao. The agreement provided for an amnesty for the UFDR, its recognition as a political party, and the integration of its fighters into the army. Further negotiations resulted in an agreement in 2008 for reconciliation, a unity government, and local elections in 2009 and parliamentary and presidential elections in 2010. The new unity government that resulted was formed in January 2009. In late 2012, a coalition of old rebel groups under new name of Séléka renewed fighting. Two other, previously unknown groups, the Alliance for Revival and Rebuilding (A2R) and the Patriotic Convention for Saving the Country (CPSK) also joined the coalition, as well as the Chadian group FPR. On 27 December 2012, CAR President Francois Bozizé requested international assistance to help with the rebellion, in particular from France and the United States. French President François Hollande rejected the plea, saying that the 250 French troops stationed at Bangui M'Poko International Airport are there "in no way to intervene in the internal affairs". On 11 January 2013, a ceasefire agreement was signed Libreville, Gabon. The rebels dropped their demand for President François Bozizé to resign, but he had to appoint a new prime minister from the opposition party by 18 January 2013. On 13 January, Bozizé signed a decree that removed Prime Minister Faustin-Archange Touadéra from power, as part of the agreement with the rebel coalition. On 17 January, Nicolas Tiangaye was appointed Prime Minister. On 24 March 2013, rebel forces heavily attacked the capital Bangui and took control of major structures, including the presidential palace. Bozizé's family fled across the river to the Democratic Republic of the Congo and then to Yaounde, the capital of Cameroon where he has been granted temporary refuge. Séléka leader Michel Djotodia declared himself President. Djotodia said that there would be a three-year transitional period and that Tiangaye would continue to serve as Prime Minister. Djotodia promptly suspended the constitution and dissolved the government, as well as the National Assembly. He then reappointed Tiangaye as Prime Minister on 27 March 2013. Top military and police officers met with Djotodia and recognized him as President on 28 March 2013. Catherine Samba-Panza assumed the office of interim president on 23 January 2014. By 2015, there was little government control outside of the capital, Bangui. The dissolution of Seleka led to ex-Seleka fighters forming new militia that often fight each other. The rebel leader Noureddine Adam declared the autonomous Republic of Logone on 14 December 2015. Peacekeeping largely transitioned from the ECCAS led MICOPAX to the AU led MISCA to the UN led MINUSCA while the French peacekeeping mission was known as Operation Sangaris. Armed entrepreneurs had carved out personal fiefdoms in which they set up checkpoints, collect illegal taxes, and take in millions of dollars from the illicit coffee, mineral, and timber trades. By 2017, more than 14 armed groups vied for territory, notably four factions formed by ex-Séléka leaders who control about 60% of the country's territory. With the de facto partition of the country between ex-Séléka militias in the north and east and Antibalaka militias in the south and west, hostilities between both sides decreased but sporadic fighting continued. In February 2016, after a peaceful election, the former Prime Minister Faustin-Archange Touadéra was elected president. In October 2016, France announced that it was ending its peacekeeping mission in the country, Operation Sangaris and largely withdrew its troops, saying that the operation was a success. Tensions erupted in competition between Ex-Seleka militias arising over control of a goldmine in November 2016, where MPC and the FPRC coalition which incorporated elements of their former enemy, the Anti-balaka, attacked UPC. The violence is often ethnic in nature with the FPRC associated with the Gula and Runga people and the UPC associated with the Fulani. Most of the fighting was in the centrally located Ouaka prefecture, which has the country's second largest city Bambari, because of its strategic location between the Muslim and Christian regions of the country and its wealth. The fight for Bambari in early 2017 displaced 20,000. MINUSCA made a robust deployment to prevent FPRC taking the city and in February 2017, Joseph Zoundeiko, the chief of staff of FPRC who previously led the military wing of Seleka, was killed by MINUSCA after crossing one of the red lines. At the same time, MINUSCA negotiated the removal of Darassa from the city. This led to UPC to find new territory, spreading the fighting from urban to rural areas previously spared. Additionally, the thinly spread MINUSCA relied on Ugandan as well as American special forces to keep the peace in the southeast as they were part of a campaign to eliminate the Lord's Resistance Army but the mission ended in April 2017. By the latter half of 2017, the fighting largely shifted to the Southeast where the UPC reorganized and were pursued by the FPRC and antibalaka with the level of violence only matched by the early stage of the war. About 15,000 people fled from their homes in an attack in May and six U.N. peacekeepers were killed - the deadliest month for the mission yet. In June 2017, another ceasefire was signed in Rome by the government and 14 armed groups including FPRC but the next day fighting between an FPRC faction and antibalaka militias killed more than 100 people. In October 2017, another ceasefire was signed between the UPC, the FPRC, and anti-balaka groups and FPRC announced Ali Darassa as coalition vice-president but fighting continued afterward. By July 2018, FPRC, now headed by Abdoulaye Hissène and based in the northeastern town of Ndélé, had troops threatening to move onto Bangui. In Western CAR, another rebel group, with no known links to Seleka or Antibalaka, called "Return, Reclamation, Rehabilitation" (3R) formed in 2015 reportedly by self-proclaimed general Sidiki Abass, claiming to be protecting Muslim Fulani people from an Antibalaka militia led by Abbas Rafal. They are accused of displacing 17,000 people in November 2016 and at least 30,000 people in the Ouham-Pendé prefecture in December 2016. In Northwestern CAR around Paoua, fighting since December 2017 between Revolution and Justice (RJ) and Movement for the Liberation of the Central African Republic People (MNLC) displaced around 60,000 people. MNLC, founded in October 2017, was led by Ahamat Bahar, a former member and co-founder of FPRC and MRC, and is allegedly backed by Fulani fighters from Chad. The Christian militant group RJ was formed in 2013, mostly by members of the presidential guard of former President Ange Felix Patassé, and were composed mainly of ethnic Sara-Kaba. While both groups had previously divided the territory in the Northwest, tensions erupted after the killing of RJ leader, Clément Bélanga, in November 2017.
https://en.wikipedia.org/wiki?curid=5479
Geography of the Central African Republic The Central African Republic is a landlocked nation within the interior of the African continent. It is bordered by the countries of Cameroon, Chad, Sudan, South Sudan, the Democratic Republic of the Congo and the Republic of the Congo. Much of the country consists of flat, or rolling plateau savanna, about above sea level. In the northeast are the Fertit Hills, and there are scattered hills in the southwestern part of the country. To the northwest are the Karre Mountains (also known as Yade Massif), a granite plateau with an altitude of . At , the Central African Republic is the world's 45th-largest country (after Somalia). It is comparable in size to Ukraine or slightly smaller than the US state of Texas. Much of the southern border is formed by tributaries of the Congo River, with the Mbomou River in the east merging with the Uele River to form the Ubangi River. In the west, the Sangha River flows through part of the country. The eastern border lies along the edge of the Congo-Nile watershed. Estimates of the amount of the country covered by forest ranges up to 8%, with the densest parts in the south. The forest is highly diverse, and includes commercially important species of Ayous, Sapelli and Sipo. The deforestation rate is 0.4% per annum, and lumber poaching is commonplace. The climate of the C.A.R. is generally tropical. The northern areas are subject to harmattan winds, which are hot, dry, and carry dust. The tip of the northern regions have been subject to desertification. The remainder of the country is prone to flooding from nearby rivers. About one third of the Central African Republic's population do not have access to clean water. Location: Central Africa, north of Democratic Republic of the Congo Area - comparative: Land boundaries: "total:" 5,920 km "border countries:" Cameroon 901 km, Chad 1,556 km, Democratic Republic of the Congo 1,747 km, Republic of the Congo 487 km, Sudan 174 km and South Sudan 1,055 km Coastline: 0 km (landlocked) Terrain: vast, flat to rolling, monotonous plateau; scattered hills in northeast and southwest Elevation extremes: "lowest point:" Oubangui River 335 m "highest point:" Mont Ngaoui 1,420 m Natural resources: diamonds, uranium, timber, gold, petroleum, hydropower Land use: "arable land:" 2.89% "permanent crops:" 0.13% "other:" 96.98% (2012 est.) Irrigated land: 1.35 km² (2003) Total renewable water resources: 144.4 km3 (2011) Freshwater withdrawal (domestic/industrial/agricultural): "total:" 0.07 km3/yr (83%/17%/1%) "per capita:" 17.42 m3/yr (2005) Natural hazards: hot, dry, dusty harmattan winds affect northern areas; floods are common Environment - current issues: tap water is not potable; poaching has diminished its reputation as one of the last great wildlife refuges; desertification Environment - international agreements: "party to:" Biodiversity, Climate Change, Desertification, Endangered Species, Hazardous Wastes, Nuclear Test Ban, Ozone Layer Protection, Tropical Timber 94, Wetlands "signed, but not ratified:" Law of the Sea Geography - note: landlocked; almost the precise center of Africa This is a list of the extreme points of the Central African Republic, the points that are farther north, south, east or west than any other location.
https://en.wikipedia.org/wiki?curid=5480
Demographics of the Central African Republic This article is about the demographic features of the population of the Central African Republic, including population density, ethnicity, education level, health of the populace, economic status, religious affiliations and other aspects of the population. There are more than 80 ethnic groups in the Central African Republic (CAR), each with its own language. About 50% are Baya-Mandjia, 40% Banda (largely located in the northern and central parts of the country), and 7% are M'Baka (southwestern corner of the CAR). Sangho, the language of a small group along the Oubangui River, is the national language spoken by the majority of Central Africans. Only a small part of the population has more than an elemental knowledge of French, the official language. More than 55% of the population of the CAR lives in rural areas. The chief agricultural areas are around the Bossangoa and Bambari. Bangui, Berberati, Bangassou, and Bossangoa are the most densely populated urban centers. According to the total population was in , compared to only 1 327 000 in 1950. The proportion of children below the age of 15 in 2010 was 40.4%, 55.6% was between 15 and 65 years of age, while 4% was 65 years or older . Registration of vital events is in the Central African Republic not complete. The Population Departement of the United Nations prepared the following estimates. Total Fertility Rate (TFR) (Wanted Fertility Rate) and Crude Birth Rate (CBR): Demographic statistics according to the World Population Review in 2019. The following demographic statistics are from the CIA World Factbook. An approximate distribution of the ethnic groups is shown in the chart below:
https://en.wikipedia.org/wiki?curid=5481
Economy of the Central African Republic The economy of the Central African Republic is one of the world's least developed, with an estimated annual per capita income of just $805 as measured by purchasing power parity in 2019. Sparsely populated and landlocked, the nation is overwhelmingly agrarian. The vast bulk of the population engages in subsistence farming and 55% of the country's GDP arises from agriculture. Subsistence agriculture, together with forestry, remains the backbone of the economy of the Central African Republic (CAR), with more than 70% of the population living in outlying areas. Principal food crops include cassava, peanuts, sorghum, millet, maize, sesame, and plantains. Principal cash crops for export include cotton, coffee, and tobacco. Timber has accounted for about 16% of export earnings and the diamond industry for nearly 54%. Much of the country's limited electrical supply is provided by hydroelectric plants located in Boali. Fuel supplies must be barged in via the Oubangui River or trucked overland through Cameroon, resulting in frequent shortages of gasoline, diesel, and jet fuel. The C.A.R.'s transportation and communication network is limited. The country has only 429 kilometers of paved road, limited international, and no domestic air service, and does not possess a railroad. River traffic on the Oubangui River is impossible from April to July, and conflict in the region has sometimes prevented shipments from moving between Kinshasa and Bangui. The telephone system functions, albeit imperfectly. Four radio stations currently operate in the C.A.R., as well as one television station. Numerous newspapers and pamphlets are published on a regular basis, and one company has begun providing Internet service. There are 22.2 million hectares (56.5 million acres) of forest (36 percent of the total land area), but only 3.4 million hectares (8.4 million acres) of dense forest, all in the south in the regions bordering the DRC, Republic of Congo and Cameroon. The CAR’s exploitable forests cover 27 million hectares (68 million acres), or 43% of the total land area. Transportation bottlenecks on rivers and lack of rail connections are serious hindrances to commercial exploitation. Most timber is shipped down the Ubangi and Zaire rivers and then on the Congo railway to the Atlantic. About 34 species of trees are felled, but 85 percent of the total is ayous, Aniegré, iroko, sapele, and sipo. A dozen sawmills produced of sawn logs and veneer logs in 2014. The government is encouraging production of plywood and veneer due to the high exporting costs due to poor transportation infrastructure. Competition from lower-cost Asian and Latin American loggers has hurt the local industry, which is encumbered with high transportation and labor costs. In 2014, the country exported 59.3 million US dollars of forest products. This accounts for 40% of the countries total export earnings. The country has rich natural resources in the form of diamonds, gold, uranium, and other minerals. Diamonds constitute one of the most important exports of the CAR, frequently accounting for 20-30% of export revenues, but an estimated 30-50% of the diamonds produced each year leave the country clandestinely. There may be petroleum deposits along the country's northern border with Chad (Two billion barrels of oil are present in private estimates). Diamonds are the only of these mineral resources currently being developed; reported sales of largely uncut diamonds range between 20-30% of the CAR's export earnings. Industry contributes less than 20% of the country's GDP, with artesian diamond mining, breweries, and sawmills making up the bulk of the sector. Services currently account for 25% of GDP, largely because of the oversized government bureaucracy and high transportation costs arising from the country's landlocked position. 74 percent (2013) of the population in the Central African Republic works in the agriculture industry, so Central African Republic's economy is dominated by the cultivation and sale of foodcrops such as yams, cassava, peanuts, maize, sorghum, millet, sesame, and plantains. The importance of foodcrops over exported cash crops is illustrated by the fact that the total production of cassava, the staple food of most Central Africans, ranges between c. 200,000 and 300,000 tons a year, while the production of cotton, the principal exported cash crop, ranges from c. 25,000 to 45,000 tons a year. Foodcrops are not exported in large quantities but they still constitute the principal cash crops of the country because Central Africans derive far more income from the periodic sale of surplus foodcrops than from exported cash crops such as cotton or coffee. Many rural and urban women also transform some foodcrops into alcoholic drinks such as sorghum beer or hard liquor and derive considerable income from the sale of these drinks. Much of the income derived from the sale of foods and alcohol is not "on the books" and thus is not considered in calculating per capita income, which is one reason why official figures for per capita income are not accurate in the case of the CAR. The per capita income of the CAR is often listed as being around $400 a year, said to be one of the lowest in the world, but this figure is based mostly on reported sales of exports and largely ignores the more important but unregistered sale of foods, locally produced alcohol, diamonds, ivory, bushmeat, and traditional medicines, for example. The informal economy of the CAR is more important than the formal economy for most Central Africans. The financial sector of the CAR, the smallest in the CEMAC, plays a limited role in supporting economic growth. Suffering from weak market infrastructure and legal and judicial frameworks, the financial system remains small, undeveloped, and dominated by commercial banks. Because of economic and security concerns, financial institutions, and particularly microfinance institutions (MFIs), have consolidated their business in the capital, Bangui, over the past few years. With less than 1 percent of the total population holding a bank account, access to financial services is extremely limited in the CAR. Microfinance accounts only for 1 percent of the total credit facilities, serving 0.5 percent of the population. Low levels of mobile penetration – which stand at 30 percent, a significantly lower percentage than in the rest of the continent – dampen the potential expansion of access to financial services through mobile technology. Fishing is carried on extensively along the rivers, but most of the catch is sold or bartered on the Democratic Republic of the Congo (DRC) side of the Ubangi River. In 1950, the government began a fish-farming program, and by the end of 1968 there were almost 12,000 ponds. The 2003 fish catch was about 15,000 tons. The CAR is heavily dependent upon multilateral foreign aid and the presence of numerous NGO's which provide numerous services which the government fails to provide. As one UNDP official put it, the CAR is a country "sous serum," or a country hooked up to an IV. (Mehler 2005:150) The very presence of numerous foreign personnel and organizations in the country, including peacekeepers and even refugee camps, provides an important source of revenue for many Central Africans. In the 40 years since independence, the CAR has made slow progress toward economic development. Economic mismanagement, poor infrastructure, a limited tax base, scarce private investment, and adverse external conditions have led to deficits in both its budget and external trade. Its debt burden is considerable, and the country has seen a decline in per capita gross national product over the last 40 years. Important constraints to economic development include the CAR's landlocked position, a poor transportation system, a largely unskilled work force, and a legacy of misdirected macroeconomic policies. The 50% devaluation of the currencies of 14 Francophone African nations on 12 January 1994 had mixed effects on the CAR's economy. Diamond, timber, coffee, and cotton exports increased, leading an estimated rise of GDP of 70% in 1994 and nearly 50% in 1995. Military rebellions and social unrest in 1996 were accompanied by widespread destruction of property and a drop in GDP of 2%. Ongoing violence between the government and rebel military groups over pay issues, living conditions, and political representation has destroyed many businesses in the capital and reduced tax revenues for the government. The International Monetary Fund (IMF) approved an Extended Structure Adjustment Facility in 1998. The government has set targets of annual 5% growth and 25% inflation for 2000-2001. Structural adjustment programs with the World Bank and IMF and interest-free credits to support investments in the agriculture, livestock, and transportation sectors have had limited impact. The World Bank and IMF are now encouraging the government to concentrate exclusively on implementing much-needed economic reforms to jump-start the economy and defining its fundamental priorities with the aim of alleviating poverty. As a result, many of the state-owned business entities have been privatized and limited efforts have been made to standardize and simplify labor and investment codes and to address problems of corruption. The Central African Government is currently in the process of adopting new labor and investment codes. The following table shows the main economic indicators in 1980–2017.
https://en.wikipedia.org/wiki?curid=5483
Telecommunications in the Central African Republic Telecommunications in the Central African Republic includes radio, television, fixed and mobile telephones, and the Internet as well as the postal system. Persistent conflict has hampered telecommunication and media development in the Central African Republic. There are active television services, radio stations, internet service providers, and mobile phone carriers. Radio is the most-popular communications medium. Socatel is the leading provider for both Internet and mobile phone access throughout the country. The primary governmental regulating bodies of telecommunications are the "Ministère des Postes" (Ministry of Posts), and "Télécommunications et des Nouvelles Technologies" (Telecommunications and New Technologies). Support is received from the ITU Telecommunication Development Sector (ITU-D) within the International Telecommunication Union to improve telecommunications infrastructure. In Bangui, UN-sponsored Radio "Ndeke Luka" ("bird of luck") provides balanced output, and rebroadcasts international news programming. Other radio and TV stations are run by the state-run "Radiodiffusion-Television Centrafricaine" and provide little coverage of the political opposition. There are about two dozen privately owned radio stations. Many, such as Radio Notre Dame, run by religious organisations. Starting 24 November 2011 "La Radio et la Télévision nationale centrafricaine (TVCA)" (Central African Radio (CAR) and National Television) are available via satellite. This move brought state-run national radio and television coverage to the entire country. GSM coverage was for a long time limited to the capital area. But in June 2006 coverage was expanded at least to Berberati in the western part of the country. There are currently two GSM-900 mobile operators, Telecel CAR and NationLink Telecom RCA. A third network, Centrafrique Telecom Plus closed down in late 2005 after failing to attract more than 5,000 subscribers and failing to renew its license, in accordance with the then new Bozize government policies. There are no government restrictions on access to the Internet or credible reports that the government monitors e-mail or Internet chat rooms without judicial oversight. Although the constitution and law provide for freedom of speech and press, authorities occasionally arrest journalists critical of the government and in some cases the government impedes individuals’ right to free speech. Imprisonment for defamation and censorship were abolished in 2005; however, journalists found guilty of libel or slander face fines of 100,000 to eight million CFA francs ($200 to US$16,000). The law provides for imprisonment and fines of as much as one million CFA francs (US$2,000) for journalists who use the media to incite disobedience among security forces or incite persons to violence, hatred, or discrimination. Similar fines and imprisonment of six months to two years may be imposed for the publication or broadcast of false or fabricated information that "would disturb the peace." L'Office National de la Poste et de l'Épargne (ONPE) is the government organization responsible for the postal service.
https://en.wikipedia.org/wiki?curid=5484
Transport in the Central African Republic Modes of transport in the Central African Republic include road, water, and air. Most but not all of the country is connected to the road network, much of which is unpaved, and which centres on the "routes nationales" identified as RN1 to RN11. Bangui serves as a seaport, and 900 km of inland waterways are navigable, the main route being the Oubangui river. There is one international airport at Bangui-Mpoko, two other paved airports, and over 40 with unpaved runways. There are presently no railways in the Central African Republic. A line from Cameroon port of Kribi to Bangui was proposed in 2002. In 2002, the country had 23,810 km (14,796 mi) of roads, of which only 429 km (267 mi) were paved. A rehabilitation project, begun in 1974 and completed ten years later, centered on three highways running north, west, and south from Nola. In 2003, there were about 1,850 passenger cars and 1,650 commercial vehicles in use. Major roads include: The roads east to Sudan and north to rishon kabur 900 km; traditional trade carried on by means of shallow-draft dugouts; Oubangui is the most important river, navigable all year to craft drawing 0.6 m or less; 282 km navigable to craft drawing as much as 1.8 m. There is only one river port. It is at the city of Bangui. In 2004, there were an estimated 50 airports. However, only three had paved runways, as of 2005. There is an international airport at Bangui-Mpoko. Five airlines provide international transport. The Republic was also a partner in Air Afrique, before it ceased operations in 2002. Inter-RCA provides domestic service. In 2003, about 46,000 passengers were carried on domestic and international flights. The most important airport in the Central African Republic is Bangui M'Poko International Airport (ICAO: FEFF)
https://en.wikipedia.org/wiki?curid=5485
Central African Armed Forces The Central African Armed Forces () are the armed forces of the Central African Republic and have been barely functional since the outbreak of the civil war in 2012. Today they are among the world's weakest armed forces, dependent on international support to provide security in the country. In recent years the government has struggled to form a unified national army. It consists of the Ground Force (which includes the air service), the gendarmerie, and the National Police. Its disloyalty to the president came to the fore during the mutinies in 1996–1997, and since then has faced internal problems. It has been strongly criticised by human rights organisations due to terrorism, including killings, torture and sexual violence. In 2013 when militants of the Séléka rebel coalition seized power and overthrew President Bozizé they executed many FACA troops. The military has played an important role in the history of Central African Republic. The immediate former president, General François Bozizé was a former army chief-of-staff and his government included several high-level military officers. Among the country's five presidents since independence in 1960, three have been former army chiefs-of-staff, who have taken power through coups d'état. No president with a military background has, however, ever been succeeded by a new military president. The country's first president, David Dacko was overthrown by his army chief-of-staff, Jean-Bédel Bokassa in 1966. Following Bokassa, David Dacko was restored in 1981, only to be overthrown once again by his new army chief of staff, General André Kolingba, after only a few months in power. In 1993, Ange-Félix Patassé became the Central African Republic's first elected president. He soon became unpopular within the army, resulting in violent mutinies in 1996–1997. In May 2001, there was an unsuccessful coup attempt by Kolingba and once again Patassé had to turn to friends abroad for support, this time Libya and DR Congo. Some months later, at the end of October, Patassé sacked his army chief-of-staff, François Bozizé, and attempted to arrest him. Bozizé then fled to Chad and gathered a group of rebels. In 2002, he seized Bangui for a short period, and in March 2003 took power in a coup d'état . When General Kolingba became president in 1981, he implemented an ethnicity-based recruitment policy for the administration. Kolingba was a member of the Yakoma people from the south of the country, which made up approximately 5% of the total population. During his rule, members of Yakoma were granted all key positions in the administration and made up a majority of the military. This later had disastrous consequences when Kolingba was replaced by a member of a northerner tribe, Ange-Félix Patassé. Soon after the election 1993, Patassé became unpopular within the army, not least because of his inability to pay their wages (partly due to economic mismanagement and partly because France suddenly ended its economic support for the soldiers' wages). Another reason for the irritation was that most of FACA consisted of soldiers from Kolingba's ethnic group, the Yakoma. During Patassé's rule they had become increasingly marginalised, while he created militias favouring his own Gbaya tribe, as well as neighbouring Sara and Kaba. This resulted in army mutinies in 1996–1997, where fractions of the military clashed with the presidential guard, the "Unité de sécurité présidentielle (USP)" and militias loyal to Patassé. On 6 December, a negotiation process started, facilitated by Gabon, Burkina-Faso, Chad and Mali. The military — supported by the opposition parties — insisted that Patassé had to resign. In January, 1997, however, the Bangui Agreements were signed and the French EFAO troop were replaced by the 1,350 soldiers of the "Mission interafricaine de surveillance des Accords de Bangui (MISAB)". In March, all mutineers were granted amnesty. The fighting between MISAB and the mutineers continued with a large offensive in June, resulting in up to 200 casualties. After this final clash, the mutineers calmed. After the mutinies, President Patassé suffered from a typical "dictator's paranoia", resulting in a period of cruel terror executed by the presidential guard and various militias within the FACA loyal to the president, such as the Karako. The violence was directed against the Yakoma tribe, of which it is estimated that 20,000 persons fled during this period. The oppression also targeted other parts of the society. The president accused his former ally France of supporting his enemies and sought new international ties. When he strengthened his presidential guard (creating the FORSIDIR, see below), Libya sent him 300 additional soldiers for his own personal safety. When former President Kolingba attempted a coup d'état in 2001 (which was, according to Patassé, supported by France), the Movement for the Liberation of Congo (MLC) of Jean-Pierre Bemba in DR Congo came to his rescue. Crimes conducted by Patassé's militias and Congolese soldiers during this period are now being investigated by the International Criminal Court, who wrote that "sexual violence appears to have been a central feature of the conflict", having identified more than 600 rape victims. The FACA has been dominated by soldiers from the Yakoma ethnic group since the time of Kolingba. It has hence been considered disloyal by the two northerner presidents Patassé and Bozizé, both of whom have equipped and run their own militias outside FACA. The military also proved its disloyalty during the mutinies in 1996–1997. Although Francois Bozizé had a background in FACA himself (being its chief-of-staff from 1997 to 2001), he was cautious by retaining the defence portfolio, as well as by appointing his son Jean-Francis Bozizé cabinet director in charge of running the Ministry of Defence. He kept his old friend General Antoine Gambi as Chief of Staff. Due to failure to curb deepening unrest in the northern part of the country, Gambi was in July 2006 replaced with Bozizé's old friend from the military academy, Jules Bernard Ouandé. The forces assisting Bozizé in seizing the power in 2003 were not paid what they were promised and started looting, terrorising and killing ordinary citizens. Summary executions took place with the implicit approval of the government. The situation has deteriorated since early 2006, and the regular army and the presidential guard regularly execute extortion, torture, killings and other human rights violations. There is no possibility for the national judicial system to investigate these cases. At the end of 2006, there were an estimated 150,000 internally displaced people in CAR. During a UN mission in the northern part of the country in November 2006, the mission had a meeting with a prefect who said that he could not maintain law and order over the military and the presidential guards. The FACA currently conducts summary executions and burns houses. On the route between Kaga-Bandoro and Ouandago some 2,000 houses have been burnt, leaving an estimated 10,000 persons homeless. Both the "Multinational Force in the Central African Republic" (FOMUC) and France are assisting in the current reform of the army. One of the key priorities of the reform of the military is make it more ethnically diversified. It should also integrate Bozizé's own rebel group (mainly consisting of members of his own Gbaya tribe). Many of the Yakoma soldiers who left the country after the mutinies in 1996–1997 have now returned and must also be reintegrated into the army. At the same time, BONUCA holds seminars in topics such as the relationship between military and civil parts of society. Most of the army's heavy weapons and equipment were destroyed or captured by Séléka militants during the 2012–2014 civil war. In the immediate aftermath of the war, the army was only in possession of 70 rifles. The majority of its arsenals were plundered during the fighting by the Séléka coalition and other armed groups. Thousands of the army's small arms were also distributed to civilian supporters of former President Bozizé in 2013. Prior to 2014, the army's stocks of arms and ammunition were primarily of French, Soviet, and Chinese origin. In 2018, the army's equipment stockpiles were partly revitalized by a donation of 900 pistols, 5,200 rifles, and 270 unspecified rocket launchers from Russia. Since the mutinies, a number of peacekeeping and peace enforcing international missions have been present in Central African Republic. There has been discussion of the deployment of a regional United Nations (UN) peacekeeping force in both Chad and Central African Republic, in order to potentially shore up the ineffectual Darfur Peace Agreement. The missions deployed in the country during the last 10 years are the following: In addition to the multilateral forces, CAR has received bilateral support from other African countries, such as the Libyan and Congolese assistance to Patassé mentioned above. Bozizé is in many ways dependent on Chad's support. Chad has an interest in CAR, since it needs to ensure calmness close to its oil fields and the pipeline leading to the Cameroonian coast, close to CAR's troubled northwest. Before seizing power, Bozizé built up his rebel force in Chad, trained and augmented by the Chadian military. Chadian President Déby assisted him actively in taking the power in March 2003 (his rebel forces included 100 Chadian soldiers). After the coup, Chad provided another 400 soldiers. Current direct support includes 150 non-FOMUC Chadian troops that patrol the border area near Goré, a contingent of soldiers in Bangui, and troops within the presidential lifeguard. The CEMAC Force includes 121 Chadian soldiers. There has been an almost uninterrupted French military presence in Central African Republic since independence, regulated through agreements between the two Governments. French troops were allowed to be based in the country and to intervene in cases of destabilisation. This was particularly important during the cold war era, when Francophone Africa was regarded as a natural French sphere of influence. Additionally, the strategic location of the country made it a more interesting location for military bases than its neighbours, and Bouar and Bangui were hence two of the most important French bases abroad. However, in 1997, following Lionel Jospin's expression "Neither interference nor indifference", France came to adopt new strategic principles for its presence in Africa. This included a reduced permanent presence on the continent and increased support for multilateral interventions. In Central African Republic, the Bouar base and the Béal Camp (at that time home to 1,400 French soldiers) in Bangui were shut down, as the French concentrated its African presence to Abidjan, Dakar, Djibouti, Libreville and N'Djamena and the deployment of a "Force d'action rapide", based in France. However, due to the situation in the country, France has retained a military presence. During the mutinies, 2,400 French soldiers patrolled the streets of Bangui. Their official task was to evacuate foreign citizens, but this did not prevent direct confrontations with the mutineers (resulting in French and mutineer casualties). The level of French involvement resulted in protests among the Central African population, since many sided with the mutineers and accused France of defending a dictator against the people's will. Criticism was also heard in France, where some blamed their country for its protection of a discredited ruler, totally incapable of exerting power and managing the country. After the mutinies in 1997, the MISAB became a multilateral force, but it was armed, equipped, trained and managed by France. The Chadian, Gabonese and Congolese troops of the current "Force multinationale en Centrafrique (FOMUC)" mission in the country also enjoy logistical support from French soldiers. A study carried out by the US Congressional Research Service revealed that France has again increased its arms sales to Africa, and that during the 1998–2005 period it was the leading supplier of arms to the continent. The Air Force is almost inoperable. Lack of funding has almost grounded the air force apart from an AS 350 Ecureuil delivered in 1987. Mirage F1 planes from the French Air Force regularly patrolled troubled regions of the country and also participated in direct confrontations until they were withdrawn and retired in 2014. According to some sources, Bozizé used the money he got from the mining concession in Bakouma to buy two old MI 8 helicopters from Ukraine and one Lockheed C-130 Hercules, built in the 1950s, from the USA. The air force otherwise operates 7 light aircraft, including a single helicopter: The Presidential Guard ("garde présidentielle") or Republican Guard is officially part of FACA but it is often regarded as a separate entity under the direct command of the President. Since 2010 the Guard has received training from South Africa and Sudan, with Belgium and Germany providing support. GR consists of so-called patriots that fought for Bozizé when he seized power in 2003 (mainly from the Gbaya tribe), together with soldiers from Chad. They are guilty of numerous assaults on the civil population, such as terror, aggression, sexual violence. Only a couple of months after Bozizé's seizure of power, in May 2003, taxi and truck drivers conducted a strike against these outrages. However, post-civil leaders have been cautious in attempting to significantly reform the Republican Guard. Bozizé has created an amphibious force. It is called the Second Battalion of the Ground Forces and it patrols the Ubangi river. The staff of the sixth region in Bouali (mainly made up of members of the former president's lifeguard) was transferred to the city of Mongoumba, located on the river. This city had previously been plundered by forces from the MLC, that had crossed the CAR/Congo border. The riverine patrol force has approximately one hundred personnel and operates seven patrol boats. A program for disarmament and reintegration of veteran soldiers is currently taking place. A national commission for the disarmament, demobilisation and reintegration was put in place in September 2004. The commission is in charge of implementing a program wherein approximately 7,500 veteran soldiers will be reintegrated in civil life and obtain education.
https://en.wikipedia.org/wiki?curid=5486
Foreign relations of the Central African Republic President François Bozizé has said that one of his priorities is to get the support of the international community. This has indeed been visible in his relations to donor countries and international organisations. At the same time it is difficult to have an open policy towards neighbouring countries when they are used as safe haven by rebels regularly attacking Central African Republic (C.A.R.), or when one allied country is in war with another (as is Chad–Sudan). The Military of the Central African Republic cannot–even with the support of France and the Multinational Force of the Economic and Monetary Community of Central Africa (FOMUC)–exert control over its own borders. Hence, armed groups are regularly entering the country from Chad and Sudan. The President said in an interview that he has a good relation with neighbours and fellow CEMAC countries, "put aside the incident with Sudan when the border had to be closed since militia entered C.A.R. territory". The Central African Republic is an active member in several Central African organizations, including the Economic and Monetary Union (CEMAC), the Economic Community of Central African States (CEEAC), the Central African Peace and Security Council (COPAX- still under formation), and the Central Bank of Central African States (BEAC). Standardization of tax, customs, and security arrangements between the Central African states is a major foreign policy objective of the C.A.R. Government. The C.A.R. is a participant in the Community of Sahel–Saharan States (CEN-SAD), and the African Union (AU). Other multilateral organizations—including the World Bank, International Monetary Fund, United Nations agencies, European Union, and the African Development Bank—and bilateral donors—including Germany, Japan, the European Union, and the United States—are significant development partners for the C.A.R. Nineteen countries have resident diplomatic representatives in Bangui, and the C.A.R. maintains approximately the same number of missions abroad. Since early 1989 the government recognizes both Israel and the Palestinian state. The C.A.R. also maintains diplomatic relations with the People's Republic of China. The C.A.R. generally joins other African and developing country states in consensus positions on major policy issues. The most important countries the C.A.R. maintains bilateral relations include the following.
https://en.wikipedia.org/wiki?curid=5487
Chad Chad (; , ; , ), officially known as the Republic of Chad ( ""; ), is a landlocked country in north-central Africa. It is bordered by Libya to the north, Sudan to the east, the Central African Republic to the south, Cameroon to the south-west, Nigeria to the southwest (at Lake Chad), and Niger to the west. Chad has several regions: a desert zone in the north, an arid Sahelian belt in the centre and a more fertile Sudanian Savanna zone in the south. Lake Chad, after which the country is named, is the largest wetland in Chad and the second-largest in Africa. The capital N'Djamena is the largest city. Chad's official languages are Arabic and French. Chad is home to over 200 different ethnic and linguistic groups. The most popular religion of Chad is Islam (at 55%), followed by Christianity (at 40%). Beginning in the 7th millennium BC, human populations moved into the Chadian basin in great numbers. By the end of the 1st millennium AD, a series of states and empires had risen and fallen in Chad's Sahelian strip, each focused on controlling the trans-Saharan trade routes that passed through the region. France conquered the territory by 1920 and incorporated it as part of French Equatorial Africa. In 1960, Chad obtained independence under the leadership of François Tombalbaye. Resentment towards his policies in the Muslim north culminated in the eruption of a long-lasting civil war in 1965. In 1979 the rebels conquered the capital and put an end to the South's hegemony. But, the rebel commanders fought amongst themselves until Hissène Habré defeated his rivals. Chadian–Libyan conflict erupted in 1978 by the Libyan invasion which stopped in 1987 with a French military intervention (Operation Épervier). Hissène Habré was overthrown in turn in 1990 by his general Idriss Déby. With French support, a modernization of the Chadian armed forces was initiated in 1991. Since 2003 the Darfur crisis in Sudan has spilt over the border and destabilised the nation. Poor already, the nation and people struggled to accommodate the hundreds of thousands of Sudanese refugees who live in and around camps in eastern Chad. While many political parties are active, power lies firmly in the hands of President Déby and his political party, the Patriotic Salvation Movement. Chad remains plagued by political violence and recurrent attempted coups d'état. Chad is one of the poorest and most corrupt countries in the world; most inhabitants live in poverty as subsistence herders and farmers. Since 2003 crude oil has become the country's primary source of export earnings, superseding the traditional cotton industry. Chad has a poor human rights record, with frequent abuses such as arbitrary imprisonment, extrajudicial killings, and limits on civil liberties by both security forces and armed militias. In the 7th millennium BCE, ecological conditions in the northern half of Chadian territory favored human settlement, and its population increased considerably. Some of the most important African archaeological sites are found in Chad, mainly in the Borkou-Ennedi-Tibesti Region; some date to earlier than 2000 BCE. For more than 2,000 years, the Chadian Basin has been inhabited by agricultural and sedentary people. The region became a crossroads of civilizations. The earliest of these were the legendary Sao, known from artifacts and oral histories. The Sao fell to the Kanem Empire, the first and longest-lasting of the empires that developed in Chad's Sahelian strip by the end of the 1st millennium AD. Two other states in the region, Sultanate of Bagirmi and Wadai Empire emerged in the 16th and 17th centuries. The power of Kanem and its successors was based on control of the trans-Saharan trade routes that passed through the region. These states, at least tacitly Muslim, never extended their control to the southern grasslands except to raid for slaves. In Kanem, about a third of the population were slaves. French colonial expansion led to the creation of the in 1900. By 1920, France had secured full control of the colony and incorporated it as part of French Equatorial Africa. French rule in Chad was characterised by an absence of policies to unify the territory and sluggish modernisation compared to other French colonies. The French primarily viewed the colony as an unimportant source of untrained labour and raw cotton; France introduced large-scale cotton production in 1929. The colonial administration in Chad was critically understaffed and had to rely on the dregs of the French civil service. Only the Sara of the south was governed effectively; French presence in the Islamic north and east was nominal. The educational system was affected by this neglect. After World War II, France granted Chad the status of overseas territory and its inhabitants the right to elect representatives to the National Assembly and a Chadian assembly. The largest political party was the Chadian Progressive Party (, PPT), based in the southern half of the colony. Chad was granted independence on 11 August 1960 with the PPT's leader, François Tombalbaye, an ethnic Sara, as its first president. Two years later, Tombalbaye banned opposition parties and established a one-party system. Tombalbaye's autocratic rule and insensitive mismanagement exacerbated inter-ethnic tensions. In 1965, Muslims in the north, led by the National Liberation Front of Chad (, FRONILAT), began a civil war. Tombalbaye was overthrown and killed in 1975, but the insurgency continued. In 1979 the rebel factions led by Hissène Habré took the capital, and all central authority in the country collapsed. Armed factions, many from the north's rebellion, contended for power. The disintegration of Chad caused the collapse of France's position in the country. Libya moved to fill the power vacuum and became involved in Chad's civil war. Libya's adventure ended in disaster in 1987; the French-supported president, Hissène Habré, evoked a united response from Chadians of a kind never seen before and forced the Libyan army off Chadian soil. Habré consolidated his dictatorship through a power system that relied on corruption and violence with thousands of people estimated to have been killed under his rule. The president favoured his own Toubou ethnic group and discriminated against his former allies, the Zaghawa. His general, Idriss Déby, overthrew him in 1990. Attempts to prosecute Habré led to his placement under house arrest in Senegal in 2005; in 2013, Habré was formally charged with war crimes committed during his rule. In May 2016, he was found guilty of human-rights abuses, including rape, sexual slavery, and ordering the killing of 40,000 people, and sentenced to life in prison. Déby attempted to reconcile the rebel groups and reintroduced multiparty politics. Chadians approved a new constitution by referendum, and in 1996, Déby easily won a competitive presidential election. He won a second term five years later. Oil exploitation began in Chad in 2003, bringing with it hopes that Chad would at last have some chances of peace and prosperity. Instead, internal dissent worsened, and a new civil war broke out. Déby unilaterally modified the constitution to remove the two-term limit on the presidency; this caused an uproar among the civil society and opposition parties. In 2006 Déby won a third mandate in elections that the opposition boycotted. Ethnic violence in eastern Chad has increased; the United Nations High Commissioner for Refugees has warned that a genocide like that in Darfur may yet occur in Chad. In 2006 and in 2008 rebel forces attempted to take the capital by force, but failed on both occasions. An agreement for the restoration of harmony between Chad and Sudan, signed 15 January 2010, marked the end of a five-year war. The fix in relations led to the Chadian rebels from Sudan returning home, the opening of the border between the two countries after seven years of closure, and the deployment of a joint force to secure the border. In May 2013, security forces in Chad foiled a coup against President Idriss Déby that had been in preparation for several months. Chad is currently one of the leading partners in a West African coalition in the fight against Boko Haram. Chad has also been included on Presidential Proclamation 9645, the expanded version of United States president Donald Trump's Executive Order 13780, which restricts entry by nationals from 8 countries, including Chad, into the US. This move has angered the Chadian government. At , Chad is the world's 20th-largest country. It is slightly smaller than Peru and slightly larger than South Africa. Chad is in north central Africa, lying between latitudes 7° and 24°N, and 13° and 24°E. Chad is bounded to the north by Libya, to the east by Sudan, to the west by Niger, Nigeria and Cameroon, and to the south by the Central African Republic. The country's capital is from the nearest seaport, Douala, Cameroon. Because of this distance from the sea and the country's largely desert climate, Chad is sometimes referred to as the "Dead Heart of Africa". The dominant physical structure is a wide basin bounded to the north and east by the Ennedi Plateau and Tibesti Mountains, which include Emi Koussi, a dormant volcano that reaches above sea level. Lake Chad, after which the country is named (and which in turn takes its name from the Kanuri word for "lake"), is the remains of an immense lake that occupied of the Chad Basin 7,000 years ago. Although in the 21st century it covers only , and its surface area is subject to heavy seasonal fluctuations, the lake is Africa's second largest wetland. The region's tall grasses and extensive marshes make it favourable for birds, reptiles, and large mammals. Chad's major rivers—the Chari, Logone and their tributaries—flow through the southern savannas from the southeast into Lake Chad. Each year a tropical weather system known as the intertropical front crosses Chad from south to north, bringing a wet season that lasts from May to October in the south, and from June to September in the Sahel. Variations in local rainfall create three major geographical zones. The Sahara lies in the country's northern third. Yearly precipitations throughout this belt are under ; only the occasional spontaneous palm grove survives, and the only ones to do so are south of the Tropic of Cancer. The Sahara gives way to a Sahelian belt in Chad's centre; precipitation there varies from per year. In the Sahel, a steppe of thorny bushes (mostly acacias) gradually gives way to the south to East Sudanian savanna in Chad's Sudanese zone. Yearly rainfall in this belt is over . Chad's animal and plant life correspond to the three climatic zones. In the Saharan region, the only flora is the date-palm groves of the oasis. Palms and acacia trees grow in the Sahelian region. The southern, or Sudanic, zone consists of broad grasslands or prairies suitable for grazing. As of 2002, there were at least 134 species of mammals, 509 species of birds (354 species of residents and 155 migrants), and over 1,600 species of plants throughout the country. Elephants, lions, buffalo, hippopotamuses, rhinoceroses, giraffes, antelopes, leopards, cheetahs, hyenas, and many species of snakes are found here, although most large carnivore populations have been drastically reduced since the early 20th century. Elephant poaching, particularly in the south of the country in areas such as Zakouma National Park, is a severe problem. The small group of surviving West African crocodiles in the Ennedi Plateau represents one of the last colonies known in the Sahara today. Extensive deforestation has resulted in loss of trees such as acacias, baobab, dates and palm trees. This has also caused loss of natural habitat for wild animals; one of the main reasons for this is also hunting and livestock farming by increasing human settlements. Populations of animals like lions, leopards and rhino have fallen significantly. Efforts have been made by the Food and Agriculture Organization to improve relations between farmers, agro-pastoralists and pastoralists in the Zakouma National Park (ZNP), Siniaka-Minia, and Aouk reserve in southeastern Chad to promote sustainable development. As part of the national conservation effort, more than 1.2 million trees have been replanted to check the advancement of the desert, which incidentally also helps the local economy by way of financial return from acacia trees, which produce gum arabic, and also from fruit trees. Poaching is a serious problem in the country, particularly of elephants for the profitable ivory industry and a threat to lives of rangers even in the national parks such as Zakouma. Elephants are often massacred in herds in and around the parks by organized poaching. The problem is worsened by the fact that the parks are understaffed and that a number of wardens have been murdered by poachers. Chad's national statistical agency projected the country's 2015 population between 13,630,252 and 13,679,203, with 13,670,084 as its medium projection; based on the medium projection, 3,212,470 people lived in urban areas and 10,457,614 people lived in rural areas. The country's population is young: an estimated 47.3% is under 15. The birth rate is estimated at 42.35 births per 1,000 people, the mortality rate at 16.69. The life expectancy is 52 years. Chad's population is unevenly distributed. Density is in the Saharan Borkou-Ennedi-Tibesti Region but in the Logone Occidental Region. In the capital, it is even higher. About half of the nation's population lives in the southern fifth of its territory, making this the most densely populated region. Urban life is concentrated in the capital, whose population is mostly engaged in commerce. The other major towns are Sarh, Moundou, Abéché and Doba, which are considerably smaller but growing rapidly in population and economic activity. Since 2003, 230,000 Sudanese refugees have fled to eastern Chad from war-ridden Darfur. With the 172,600 Chadians displaced by the civil war in the east, this has generated increased tensions among the region's communities. Polygamy is common, with 39% of women living in such unions. This is sanctioned by law, which automatically permits polygamy unless spouses specify that this is unacceptable upon marriage. Although violence against women is prohibited, domestic violence is common. Female genital mutilation is also prohibited, but the practice is widespread and deeply rooted in tradition; 45% of Chadian women undergo the procedure, with the highest rates among Arabs, Hadjarai, and Ouaddaians (90% or more). Lower percentages were reported among the Sara (38%) and the Toubou (2%). Women lack equal opportunities in education and training, making it difficult for them to compete for the relatively few formal-sector jobs. Although property and inheritance laws based on the French code do not discriminate against women, local leaders adjudicate most inheritance cases in favour of men, according to traditional practice. The peoples of Chad carry significant ancestry from Eastern, Central, Western, and Northern Africa. Chad has more than 200 distinct ethnic groups, which create diverse social structures. The colonial administration and independent governments have attempted to impose a national society, but for most Chadians the local or regional society remains the most important influence outside the immediate family. Nevertheless, Chad's people may be classified according to the geographical region in which they live. In the south live sedentary people such as the Sara, the nation's main ethnic group, whose essential social unit is the lineage. In the Sahel sedentary peoples live side-by-side with nomadic ones, such as the Arabs, the country's second major ethnic group. The north is inhabited by nomads, mostly Toubous. Chad's official languages are Arabic and French, but over 100 languages and dialects are spoken. Due to the important role played by itinerant Arab traders and settled merchants in local communities, Chadian Arabic has become a lingua franca. Chad is a religiously diverse country. Estimates from Pew Research Center in 2010 found that 57.7% of the population was Muslim, while 21.5% was Catholic and a further 16.6% was Protestant. Among Muslims, 48% professed to be Sunni, 21% Shia, 4% Ahmadi and 23% just Muslim. A small proportion of the population continues to practice indigenous religions. Animism includes a variety of ancestor and place-oriented religions whose expression is highly specific. Islam is expressed in diverse ways; for example, 55% of Muslim Chadians belong to Sufi orders. Christianity arrived in Chad with the French and American missionaries; as with Chadian Islam, it syncretises aspects of pre-Christian religious beliefs. Muslims are largely concentrated in northern and eastern Chad, and animists and Christians live primarily in southern Chad and Guéra. The constitution provides for a secular state and guarantees religious freedom; different religious communities generally co-exist without problems. The majority of Muslims in the country are adherents of a moderate branch of mystical Islam (Sufism). Its most common expression is the Tijaniyah, an order followed by the 35% of Chadian Muslims which incorporates some local African religious elements. A small minority of the country's Muslims hold more fundamentalist practices, which, in some cases, may be associated with Saudi-oriented Salafi movements. Roman Catholics represent the largest Christian denomination in the country. Most Protestants, including the Nigeria-based "Winners' Chapel", are affiliated with various evangelical Christian groups. Members of the Bahá'í and Jehovah's Witnesses religious communities also are present in the country. Both faiths were introduced after independence in 1960 and therefore are considered to be "new" religions in the country. Chad is home to foreign missionaries representing both Christian and Islamic groups. Itinerant Muslim preachers, primarily from Sudan, Saudi Arabia, and Pakistan, also visit. Saudi Arabian funding generally supports social and educational projects and extensive mosque construction. Chad's constitution provides for a strong executive branch headed by a president who dominates the political system. The president has the power to appoint the prime minister and the cabinet, and exercises considerable influence over appointments of judges, generals, provincial officials and heads of Chad's para-statal firms. In cases of grave and immediate threat, the president, in consultation with the National Assembly, may declare a state of emergency. The president is directly elected by popular vote for a five-year term; in 2005 constitutional term limits were removed, allowing a president to remain in power beyond the previous two-term limit. Most of Déby's key advisers are members of the Zaghawa ethnic group, although southern and opposition personalities are represented in government. Chad's legal system is based on French civil law and Chadian customary law where the latter does not interfere with public order or constitutional guarantees of equality. Despite the constitution's guarantee of judicial independence, the president names most key judicial officials. The legal system's highest jurisdictions, the Supreme Court and the Constitutional Council, have become fully operational since 2000. The Supreme Court is made up of a chief justice, named by the president, and 15 councillors, appointed for life by the president and the National Assembly. The Constitutional Court is headed by nine judges elected to nine-year terms. It has the power to review legislation, treaties and international agreements prior to their adoption. The National Assembly makes legislation. The body consists of 155 members elected for four-year terms who meet three times per year. The Assembly holds regular sessions twice a year, starting in March and October, and can hold special sessions when called by the prime minister. Deputies elect a National Assembly president every two years. The president must sign or reject newly passed laws within 15 days. The National Assembly must approve the prime minister's plan of government and may force the prime minister to resign through a majority vote of no confidence. However, if the National Assembly rejects the executive branch's programme twice in one year, the president may disband the Assembly and call for new legislative elections. In practice, the president exercises considerable influence over the National Assembly through his party, the Patriotic Salvation Movement (MPS), which holds a large majority. Until the legalisation of opposition parties in 1992, Déby's MPS was the sole legal party in Chad. Since then, 78 registered political parties have become active. In 2005, opposition parties and human rights organisations supported the boycott of the constitutional referendum that allowed Déby to stand for re-election for a third term amid reports of widespread irregularities in voter registration and government censorship of independent media outlets during the campaign. Correspondents judged the 2006 presidential elections a mere formality, as the opposition deemed the polls a farce and boycotted them. Chad is listed as a failed state by the Fund for Peace (FFP). In 2007 Chad had the seventh highest score on the failed state index. Since then the trend has been upwards each year. Chad had the fourth highest score (behind Sudan) on the Failed State Index of 2012 and , is ranked fifth. Corruption is rife at all levels; Transparency International's Corruption Perceptions Index for 2005 named Chad (tied with Bangladesh) as the most corrupt country in the world. Chad's ranking on the index has improved only marginally in recent years. Since its first inclusion on the index in 2004, Chad's best score has been 2/10 for 2011. Critics of President Déby have accused him of cronyism and tribalism. In southern Chad, bitter conflicts over land are becoming more and more common. They frequently turn violent. Long-standing community culture is being eroded – and so are the livelihoods of many farmers. Homosexual acts are illegal and can be punished by 15 to 20 years in prison. In December 2016, Chad passed a law criminalising both male and female same-sex sexual activity by a vote of 111 to 1. Déby faces armed opposition from groups who are deeply divided by leadership clashes but united in their intention to overthrow him. These forces stormed the capital on 13 April 2006, but were ultimately repelled. Chad's greatest foreign influence is France, which maintains 1,000 soldiers in the country. Déby relies on the French to help repel the rebels, and France gives the Chadian army logistical and intelligence support for fear of a complete collapse of regional stability. Nevertheless, Franco-Chadian relations were soured by the granting of oil drilling rights to the American Exxon company in 1999. There have been numerous rebel groups in Chad throughout the last few decades. In 2007, a peace treaty was signed that integrated United Front for Democratic Change or FUC soldiers into the Chadian Army. The Movement for Justice and Democracy in Chad or MDJT also clashed with government forces in 2003 in an attempt to overthrow President Idriss Déby. In addition, there have been various conflicts with Khartoum's Janjaweed rebels in eastern Chad, who killed civilians by use of helicopter gunships. Presently, the Union of Resistance Forces or UFR are a rebel group that continues to battle with the government of Chad. In 2010, the UFR reportedly had a force estimating 6,000 men and 300 vehicles. Since 2012 Chad has been divided into 23 regions. The subdivision of Chad in regions came about in 2003 as part of the decentralisation process, when the government abolished the previous 14 prefectures. Each region is headed by a presidentially appointed governor. Prefects administer the 61 departments within the regions. The departments are divided into 200 sub-prefectures, which are in turn composed of 446 cantons. The cantons are scheduled to be replaced by "communautés rurales", but the legal and regulatory framework has not yet been completed. The constitution provides for decentralised government to compel local populations to play an active role in their own development. To this end, the constitution declares that each administrative subdivisions be governed by elected local assemblies, but no local elections have taken place, and communal elections scheduled for 2005 have been repeatedly postponed. The CIA World Factbook estimates the military budget of Chad to be 4.2% of GDP as of 2006. Given the then GDP ($7.095 bln) of the country, military spending was estimated to be about $300 million. This estimate however dropped after the end of the Civil war in Chad (2005–2010) to 2.0% as estimated by the World Bank for the year 2011. The United Nations' Human Development Index ranks Chad as the seventh poorest country in the world, with 80% of the population living below the poverty line. The GDP (Purchasing power parity) per capita was estimated as US$1,651 in 2009. Chad is part of the Bank of Central African States, the Customs and Economic Union of Central Africa (UDEAC) and the Organization for the Harmonization of Business Law in Africa (OHADA). Chad's currency is the CFA franc. In the 1960s, the Mining industry of Chad produced sodium carbonate, or natron. There have also been reports of gold-bearing quartz in the Biltine Prefecture. However, years of civil war have scared away foreign investors; those who left Chad between 1979 and 1982 have only recently begun to regain confidence in the country's future. In 2000 major direct foreign investment in the oil sector began, boosting the country's economic prospects. An uneven inclusion into the global political economy as a site for colonial resource extraction (primarily cotton and crude oil), a global economic system that does not promote nor encourage the development of Chadian industrialization, and the failure to support local agricultural production has meant that the majority of Chadians live in daily uncertainty and hunger. Over 80% of Chad's population relies on subsistence farming and livestock raising for its livelihood. The crops grown and the locations of herds are determined by the local climate. In the southernmost 10% of the territory lies the nation's most fertile cropland, with rich yields of sorghum and millet. In the Sahel only the hardier varieties of millet grow, and these with much lower yields than in the south. On the other hand, the Sahel is ideal pastureland for large herds of commercial cattle and for goats, sheep, donkeys and horses. The Sahara's scattered oases support only some dates and legumes. Chad's cities face serious difficulties of municipal infrastructure; only 48% of urban residents have access to potable water and only 2% to basic sanitation. Before the development of oil industry, cotton dominated industry and the labour market accounted for approximately 80% of export earnings. Cotton remains a primary export, although exact figures are not available. Rehabilitation of Cotontchad, a major cotton company weakened by a decline in world cotton prices, has been financed by France, the Netherlands, the European Union, and the International Bank for Reconstruction and Development (IBRD). The parastatal is now expected to be privatised. Other than Cotton, Cattle and Gum Arabic are dominant. According to the United Nations, Chad has been affected by a humanitarian crisis since at least 2001. , the country of Chad hosts over 280,000 refugees from the Sudan's Darfur region, over 55,000 from the Central African Republic, as well as over 170,000 internally displaced persons. In February 2008 in the aftermath of the battle of N'Djamena, UN Under-Secretary-General for Humanitarian Affairs John Holmes expressed "extreme concern" that the crisis would have a negative effect on the ability of humanitarians to deliver life-saving assistance to half a million beneficiaries, most of whom – according to him – heavily rely on humanitarian aid for their survival. UN spokesperson Maurizio Giuliano stated to "The Washington Post": "If we do not manage to provide aid at sufficient levels, the humanitarian crisis might become a humanitarian catastrophe". In addition, organizations such as Save the Children have suspended activities due to killings of aid workers. Civil war crippled the development of transport infrastructure; in 1987, Chad had only of paved roads. Successive road rehabilitation projects improved the network to by 2004. Nevertheless, the road network is limited; roads are often unusable for several months of the year. With no railways of its own, Chad depends heavily on Cameroon's rail system for the transport of Chadian exports and imports to and from the seaport of Douala. Chad had an estimated 59 airports, only 9 of which had paved runways. An international airport serves the capital and provides regular nonstop flights to Paris and several African cities. Chad's energy sector has had years of mismanagement by the parastatal Chad Water and Electric Society (STEE), which provides power for 15% of the capital's citizens and covers only 1.5% of the national population. Most Chadians burn biomass fuels such as wood and animal manure for power. ExxonMobil leads a consortium of Chevron and Petronas that has invested $3.7 billion to develop oil reserves estimated at one billion barrels in southern Chad. Oil production began in 2003 with the completion of a pipeline (financed in part by the World Bank) that links the southern oilfields to terminals on the Atlantic coast of Cameroon. As a condition of its assistance, the World Bank insisted that 80% of oil revenues be spent on development projects. In January 2006 the World Bank suspended its loan programme when the Chadian government passed laws reducing this amount. On 14 July 2006, the World Bank and Chad signed a memorandum of understanding under which the Government of Chad commits 70% of its spending to priority poverty reduction programmes. The telecommunication system is basic and expensive, with fixed telephone services provided by the state telephone company SotelTchad. In 2000, there were only 14 fixed telephone lines per 10,000 inhabitants in the country, one of the lowest telephone densities in the world. Gateway Communications, a pan-African wholesale connectivity and telecommunications provider also has a presence in Chad. In September 2013, Chad's Ministry for Posts and Information & Communication Technologies (PNTIC) announced that the country will be seeking a partner for fiber optic technology. Chad is ranked last in the World Economic Forum's Network Readiness Index (NRI) – an indicator for determining the development level of a country's information and communication technologies. Chad ranked number 148 out of 148 overall in the 2014 NRI ranking, down from 142 in 2013. In September 2010 the mobile phone penetration rate was estimated at 24.3% over a population estimate of 10.7 million. Chad's television audience is limited to N'Djamena. The only television station is the state-owned Télé Tchad. Radio has a far greater reach, with 13 private radio stations. Newspapers are limited in quantity and distribution, and circulation figures are small due to transportation costs, low literacy rates, and poverty. While the constitution defends liberty of expression, the government has regularly restricted this right, and at the end of 2006 began to enact a system of prior censorship on the media. Educators face considerable challenges due to the nation's dispersed population and a certain degree of reluctance on the part of parents to send their children to school. Although attendance is compulsory, only 68 percent of boys attend primary school, and more than half of the population is illiterate. Higher education is provided at the University of N'Djamena. At 33 percent, Chad has one of the lowest literacy rates of Sub-Saharan Africa. In 2013, the U.S. Department of Labor's Findings on the Worst Forms of Child Labor in Chad reported that school attendance of children aged 5 to 14 was as low as 39%. This can also be related to the issue of child labor as the report also stated that 53% of children aged 5 to 14 were working children, and that 30% of children aged 7 to 14 combined work and school. A more recent DOL report listed cattle herding as a major agricultural activity that employed underage children. Because of its great variety of peoples and languages, Chad possesses a rich cultural heritage. The Chadian government has actively promoted Chadian culture and national traditions by opening the Chad National Museum and the Chad Cultural Centre. Six national holidays are observed throughout the year, and movable holidays include the Christian holiday of Easter Monday and the Muslim holidays of Eid ul-Fitr, Eid ul-Adha, and Eid Milad Nnabi. The music of Chad includes a number of instruments such as the "kinde", a type of bow harp; the "kakaki", a long tin horn; and the "hu hu", a stringed instrument that uses calabashes as loudspeakers. Other instruments and their combinations are more linked to specific ethnic groups: the Sara prefer whistles, balafones, harps and "kodjo" drums; and the Kanembu combine the sounds of drums with those of flute-like instruments. Millet is the staple food of Chadian cuisine. It is used to make balls of paste that are dipped in sauces. In the north this dish is known as "alysh"; in the south, as "biya". Fish is popular, which is generally prepared and sold either as "salanga" (sun-dried and lightly smoked "Alestes" and "Hydrocynus") or as "banda" (smoked large fish). "Carcaje" is a popular sweet red tea extracted from hibiscus leaves. Alcoholic beverages, though absent in the north, are popular in the south, where people drink millet beer, known as "billi-billi" when brewed from red millet, and as "coshate" when from white millet. The music group Chari Jazz formed in 1964 and initiated Chad's modern music scene. Later, more renowned groups such as African Melody and International Challal attempted to mix modernity and tradition. Popular groups such as Tibesti have clung faster to their heritage by drawing on "sai", a traditional style of music from southern Chad. The people of Chad have customarily disdained modern music. However, in 1995 greater interest has developed and fostered the distribution of CDs and audio cassettes featuring Chadian artists. Piracy and a lack of legal protections for artists' rights remain problems to further development of the Chadian music industry. As in other Sahelian countries, literature in Chad has seen an economic, political and spiritual drought that has affected its best known writers. Chadian authors have been forced to write from exile or expatriate status and have generated literature dominated by themes of political oppression and historical discourse. Since 1962, 20 Chadian authors have written some 60 works of fiction. Among the most internationally renowned writers are Joseph Brahim Seïd, Baba Moustapha, Antoine Bangui and Koulsy Lamko. In 2003 Chad's sole literary critic, Ahmat Taboye, published his to further knowledge of Chad's literature internationally and among youth and to make up for Chad's lack of publishing houses and promotional structure. The development of a Chadian film industry, which began with the short films of Edouard Sailly in the 1960s, was hampered by the devastations of civil wars and from the lack of cinemas, of which there is currently only one in the whole country (the "Normandie" in N'Djamena). The Chadian feature film industry began growing again in the 1990s, with the work of directors Mahamat-Saleh Haroun, Issa Serge Coelo and Abakar Chene Massar. Haroun's film "Abouna" was critically acclaimed, and his "Daratt" won the Grand Special Jury Prize at the 63rd Venice International Film Festival. The 2010 feature film "A Screaming Man" won the Jury Prize at the 2010 Cannes Film Festival, making Haroun the first Chadian director to enter, as well as win, an award in the main Cannes competition. Issa Serge Coelo directed the films, "Daresalam" and "". Football is Chad's most popular sport. The country's national team is closely followed during international competitions and Chadian footballers have played for French teams. Basketball and freestyle wrestling are widely practiced, the latter in a form in which the wrestlers put on traditional animal hides and cover themselves with dust.
https://en.wikipedia.org/wiki?curid=5488
Chile Chile (, ; ), officially the Republic of Chile (), is a country in western South America. It occupies a long, narrow strip of land between the Andes to the east and the Pacific Ocean to the west. Chile covers an area of and has a population of 17.5 million as of 2017. The capital and largest city is Santiago and the national language is Spanish. Chile borders Peru to the north, Bolivia to the northeast, Argentina to the east, and the Drake Passage in the far south. Chilean territory includes the Pacific islands of Juan Fernández, Salas y Gómez, Desventuradas, and Easter Island in Oceania. Chile also claims about of Antarctica under the Chilean Antarctic Territory. Spain conquered and colonized the region in the mid-16th century, replacing Inca rule, but failing to conquer the independent Mapuche who inhabited what is now south-central Chile. After declaring independence from Spain in 1818, Chile emerged in the 1830s as a relatively stable authoritarian republic. In the 19th century, Chile saw significant economic and territorial growth, ending Mapuche resistance in the 1880s and gaining its current northern territory in the War of the Pacific (1879–83) after defeating Peru and Bolivia. In the 1960s and 1970s, the country experienced severe left-right political polarization and turmoil. This development culminated with the 1973 Chilean coup d'état that overthrew Salvador Allende's democratically elected left-wing government and instituted a 16-year right-wing military dictatorship of Augusto Pinochet that left more than 3,000 people dead or missing. The regime ended in 1990 after a referendum in 1988 and was succeeded by a center-left coalition which ruled until 2010. Chile is a high-income economy with high living standards. It is among South America's most economically and socially stable and prosperous nations and it leads Latin American nations in rankings of competitiveness, income per capita, globalization, state of peace, economic freedom, and low perception of corruption. It also ranks high regionally in sustainability of the state, and democratic development. Currently it also has the lowest homicide rate in the Americas after Canada. Chile is a founding member of the United Nations, the Community of Latin American and Caribbean States (CELAC) and the Pacific Alliance, and joined the OECD in 2010. There are various theories about the origin of the word "Chile". According to 17th-century Spanish chronicler Diego de Rosales, the Incas called the valley of the Aconcagua "Chili" by corruption of the name of a Picunche tribal chief ("cacique") called "Tili", who ruled the area at the time of the Incan conquest in the 15th century. Another theory points to the similarity of the valley of the Aconcagua with that of the Casma Valley in Peru, where there was a town and valley named "Chili". Other theories say Chile may derive its name from a Native American word meaning either "ends of the earth" or "sea gulls"; from the Mapuche word "chilli," which may mean "where the land ends;" or from the Quechua "chiri", "cold", or "tchili", meaning either "snow" or "the deepest point of the Earth". Another origin attributed to "chilli" is the onomatopoeic "cheele-cheele"—the Mapuche imitation of the warble of a bird locally known as trile. The Spanish conquistadors heard about this name from the Incas, and the few survivors of Diego de Almagro's first Spanish expedition south from Peru in 1535–36 called themselves the "men of Chilli". Ultimately, Almagro is credited with the universalization of the name "Chile", after naming the Mapocho valley as such. The older spelling "Chili" was in use in English until at least 1900 before switching to "Chile". Stone tool evidence indicates humans sporadically frequented the Monte Verde valley area as long as 18,500 years ago. About 10,000 years ago, migrating Indigenous Peoples settled in fertile valleys and coastal areas of what is present-day Chile. Settlement sites from very early human habitation include Monte Verde, Cueva del Milodón and the Pali-Aike Crater's lava tube. The Incas briefly extended their empire into what is now northern Chile, but the Mapuche (or Araucanians as they were known by the Spaniards) successfully resisted many attempts by the Inca Empire to subjugate them, despite their lack of state organization. They fought against the Sapa Inca Tupac Yupanqui and his army. The result of the bloody three-day confrontation known as the Battle of the Maule was that the Inca conquest of the territories of Chile ended at the Maule river. In 1520, while attempting to circumnavigate the globe, Ferdinand Magellan discovered the southern passage now named after him (the Strait of Magellan) thus becoming the first European to set foot on what is now Chile. The next Europeans to reach Chile were Diego de Almagro and his band of Spanish conquistadors, who came from Peru in 1535 seeking gold. The Spanish encountered various cultures that supported themselves principally through slash-and-burn agriculture and hunting. The conquest of Chile began in earnest in 1540 and was carried out by Pedro de Valdivia, one of Francisco Pizarro's lieutenants, who founded the city of Santiago on 12 February 1541. Although the Spanish did not find the extensive gold and silver they sought, they recognized the agricultural potential of Chile's central valley, and Chile became part of the Spanish Empire. Conquest took place gradually, and the Europeans suffered repeated setbacks. A massive Mapuche insurrection that began in 1553 resulted in Valdivia's death and the destruction of many of the colony's principal settlements. Subsequent major insurrections took place in 1598 and in 1655. Each time the Mapuche and other native groups revolted, the southern border of the colony was driven northward. The abolition of slavery by the Spanish crown in 1683 was done in recognition that enslaving the Mapuche intensified resistance rather than cowing them into submission. Despite royal prohibitions, relations remained strained from continual colonialist interference. Cut off to the north by desert, to the south by the Mapuche, to the east by the Andes Mountains, and to the west by the ocean, Chile became one of the most centralized, homogeneous colonies in Spanish America. Serving as a sort of frontier garrison, the colony found itself with the mission of forestalling encroachment by both the Mapuche and Spain's European enemies, especially the British and the Dutch. Buccaneers and English adventurers menaced the colony in addition to the Mapuche, as was shown by Sir Francis Drake's 1578 raid on Valparaíso, the colony's principal port. Chile hosted one of the largest standing armies in the Americas, making it one of the most militarized of the Spanish possessions, as well as a drain on the treasury of the Viceroyalty of Peru. The first general census was conducted by the government of Agustín de Jáuregui between 1777 and 1778; it indicated that the population consisted of 259,646 inhabitants: 73.5 percent of European descent, 7.9 percent mestizos, 8.6 percent indigenous peoples and 9.8 percent blacks. Francisco Hurtado, Governor of the province of Chiloé, conducted a census in 1784 and found the population consisted of 26,703 inhabitants, 64.4 percent of whom were whites and 33.5 percent of whom were natives. The Diocese of Concepción conducted a census in areas south of the Maule river in 1812, but did not include the indigenous population or the inhabitants of the province of Chiloé. The population is estimated at 210,567, 86.1 percent of whom were Spanish or of European descent, 10 percent of whom were indigenous and 3.7 percent of whom were mestizos, blacks and mulattos. In 1808, Napoleon's enthronement of his brother Joseph as the Spanish King precipitated the drive by the colony for independence from Spain. A national junta in the name of Ferdinand – heir to the deposed king – was formed on 18 September 1810. The Government Junta of Chile proclaimed Chile an autonomous republic within the Spanish monarchy (in memory of this day, Chile celebrates its National Day on 18 September each year). After these events, a movement for total independence, under the command of José Miguel Carrera (one of the most renowned patriots) and his two brothers Juan José and Luis Carrera, soon gained a wider following. Spanish attempts to re-impose arbitrary rule during what was called the Reconquista led to a prolonged struggle, including infighting from Bernardo O'Higgins, who challenged Carrera's leadership. Intermittent warfare continued until 1817. With Carrera in prison in Argentina, O'Higgins and anti-Carrera cohort José de San Martín, hero of the Argentine War of Independence, led an army that crossed the Andes into Chile and defeated the royalists. On 12 February 1818, Chile was proclaimed an independent republic. The political revolt brought little social change, however, and 19th-century Chilean society preserved the essence of the stratified colonial social structure, which was greatly influenced by family politics and the Roman Catholic Church. A strong presidency eventually emerged, but wealthy landowners remained powerful. Chile slowly started to expand its influence and to establish its borders. By the Tantauco Treaty, the archipelago of Chiloé was incorporated in 1826. The economy began to boom due to the discovery of silver ore in Chañarcillo, and the growing trade of the port of Valparaíso, which led to conflict over maritime supremacy in the Pacific with Peru. At the same time, attempts were made to strengthen sovereignty in southern Chile intensifying penetration into Araucanía and colonizing Llanquihue with German immigrants in 1848. Through the founding of Fort Bulnes by the Schooner Ancud under the command of John Williams Wilson, the Magallanes region joined the country in 1843, while the Antofagasta region, at the time part of Bolivia, began to fill with people. Toward the end of the 19th century, the government in Santiago consolidated its position in the south by the Occupation of Araucanía. The Boundary treaty of 1881 between Chile and Argentina confirmed Chilean sovereignty over the Strait of Magellan. As a result of the War of the Pacific with Peru and Bolivia (1879–83), Chile expanded its territory northward by almost one-third, eliminating Bolivia's access to the Pacific, and acquired valuable nitrate deposits, the exploitation of which led to an era of national affluence. Chile had joined the stand as one of the high-income countries in South America by 1870. The 1891 Chilean Civil War brought about a redistribution of power between the President and Congress, and Chile established a parliamentary style democracy. However, the Civil War had also been a contest between those who favored the development of local industries and powerful Chilean banking interests, particularly the House of Edwards who had strong ties to foreign investors. Soon after, the country engaged in a vastly expensive naval arms race with Argentina that nearly led to war. The Chilean economy partially degenerated into a system protecting the interests of a ruling oligarchy. By the 1920s, the emerging middle and working classes were powerful enough to elect a reformist president, Arturo Alessandri, whose program was frustrated by a conservative congress. In the 1920s, Marxist groups with strong popular support arose. A military coup led by General Luis Altamirano in 1924 set off a period of political instability that lasted until 1932. Of the ten governments that held power in that period, the longest lasting was that of General Carlos Ibáñez del Campo, who briefly held power in 1925 and then again between 1927 and 1931 in what was a de facto dictatorship (although not really comparable in harshness or corruption to the type of military dictatorship that have often bedeviled the rest of Latin America). By relinquishing power to a democratically elected successor, Ibáñez del Campo retained the respect of a large enough segment of the population to remain a viable politician for more than thirty years, in spite of the vague and shifting nature of his ideology. When constitutional rule was restored in 1932, a strong middle-class party, the Radicals, emerged. It became the key force in coalition governments for the next 20 years. During the period of Radical Party dominance (1932–52), the state increased its role in the economy. In 1952, voters returned Ibáñez del Campo to office for another six years. Jorge Alessandri succeeded Ibáñez del Campo in 1958, bringing Chilean conservatism back into power democratically for another term. The 1964 presidential election of Christian Democrat Eduardo Frei Montalva by an absolute majority initiated a period of major reform. Under the slogan "Revolution in Liberty", the Frei administration embarked on far-reaching social and economic programs, particularly in education, housing, and agrarian reform, including rural unionization of agricultural workers. By 1967, however, Frei encountered increasing opposition from leftists, who charged that his reforms were inadequate, and from conservatives, who found them excessive. At the end of his term, Frei had not fully achieved his party's ambitious goals. In the 1970 election, Senator Salvador Allende of the Socialist Party of Chile (then part of the "Popular Unity" coalition which included the Communists, Radicals, Social-Democrats, dissident Christian Democrats, the Popular Unitary Action Movement, and the Independent Popular Action), achieved a partial majority in a plurality of votes in a three-way contest, followed by candidates Radomiro Tomic for the Christian Democrat Party and Jorge Alessandri for the Conservative Party. Allende was not elected with an absolute majority, receiving fewer than 35 percent of votes. The Chilean Congress conducted a runoff vote between the leading candidates, Allende and former president Jorge Alessandri, and, keeping with tradition, chose Allende by a vote of 153 to 35. Frei refused to form an alliance with Alessandri to oppose Allende, on the grounds that the Christian Democrats were a workers' party and could not make common cause with the right wing. An economic depression that began in 1972 was exacerbated by capital flight, plummeting private investment, and withdrawal of bank deposits in response to Allende's socialist program. Production fell and unemployment rose. Allende adopted measures including price freezes, wage increases, and tax reforms, to increase consumer spending and redistribute income downward. Joint public-private public works projects helped reduce unemployment. Much of the banking sector was nationalized. Many enterprises within the copper, coal, iron, nitrate, and steel industries were expropriated, nationalized, or subjected to state intervention. Industrial output increased sharply and unemployment fell during the Allende administration's first year. Allende's program included advancement of workers' interests, replacing the judicial system with "socialist legality", nationalization of banks and forcing others to bankruptcy, and strengthening "popular militias" known as MIR. Started under former President Frei, the Popular Unity platform also called for nationalization of Chile's major copper mines in the form of a constitutional amendment. The measure was passed unanimously by Congress. As a result, the Richard Nixon administration organized and inserted secret operatives in Chile, in order to swiftly destabilize Allende's government. In addition, US financial pressure restricted international economic credit to Chile. The economic problems were also exacerbated by Allende's public spending which was financed mostly by printing money and poor credit ratings given by commercial banks. Simultaneously, opposition media, politicians, business guilds and other organizations helped to accelerate a campaign of domestic political and economical destabilization, some of which was backed by the United States. By early 1973, inflation was out of control. The crippled economy was further battered by prolonged and sometimes simultaneous strikes by physicians, teachers, students, truck owners, copper workers, and the small business class. On 26 May 1973, Chile's Supreme Court, which was opposed to Allende's government, unanimously denounced the Allende "disruption of the legality of the nation". Although illegal under the Chilean constitution, the court supported and strengthened Pinochet's soon-to-be seizure of power. A military coup overthrew Allende on 11 September 1973. As the armed forces bombarded the presidential palace, Allende apparently committed suicide. After the coup, Henry Kissinger told U.S. president Richard Nixon that the United States had "helped" the coup. A military junta, led by General Augusto Pinochet, took control of the country. The first years of the regime were marked by human rights violations. Chile actively participated in Operation Condor. On October 1973, at least 72 people were murdered by the Caravan of Death. According to the Rettig Report and Valech Commission, at least 2,115 were killed, and at least 27,265 were tortured (including 88 children younger than 12 years old). In 2011, Chile recognized an additional 9,800 victims, bringing the total number of killed, tortured or imprisoned for political reasons to 40,018. At the national stadium, filled with detainees, one of those tortured and killed was internationally known poet-singer Víctor Jara (see "Music and Dance", below). The stadium was renamed for Jara in 2003. A new Constitution was approved by a controversial plebiscite on 11 September 1980, and General Pinochet became president of the republic for an eight-year term. After Pinochet obtained rule of the country, several hundred committed Chilean revolutionaries joined the Sandinista army in Nicaragua, guerrilla forces in Argentina or training camps in Cuba, Eastern Europe and Northern Africa. In the late 1980s, largely as a result of events such as the 1982 economic collapse and mass civil resistance in 1983–88, the government gradually permitted greater freedom of assembly, speech, and association, to include trade union and political activity. The government launched market-oriented reforms with Hernán Büchi as Minister of Finance. Chile moved toward a free market economy that saw an increase in domestic and foreign private investment, although the copper industry and other important mineral resources were not opened for competition. In a plebiscite on 5 October 1988, Pinochet was denied a second eight-year term as president (56% against 44%). Chileans elected a new president and the majority of members of a bicameral congress on 14 December 1989. Christian Democrat Patricio Aylwin, the candidate of a coalition of 17 political parties called the "Concertación", received an absolute majority of votes (55%). President Aylwin served from 1990 to 1994, in what was considered a transition period. In December 1993, Christian Democrat Eduardo Frei Ruiz-Tagle, the son of previous president Eduardo Frei Montalva, led the Concertación coalition to victory with an absolute majority of votes (58%). Frei Ruiz-Tagle was succeeded in 2000 by Socialist Ricardo Lagos, who won the presidency in an unprecedented runoff election against Joaquín Lavín of the rightist Alliance for Chile. In January 2006, Chileans elected their first female president, Michelle Bachelet Jeria, of the Socialist Party, defeating Sebastián Piñera, of the National Renewal party, extending the "Concertación" governance for another four years. In January 2010, Chileans elected Sebastián Piñera as the first rightist President in 20 years, defeating former President Eduardo Frei Ruiz-Tagle of the "Concertación", for a four-year term succeeding Bachelet. Due to term limits, Sebastián Piñera did not stand for re-election in 2013, and his term expired in March 2014 resulting in Michelle Bachelet returning to office. On 27 February 2010, Chile was struck by an 8.8 earthquake, the fifth largest ever recorded at the time. More than 500 people died (most from the ensuing tsunami) and over a million people lost their homes. The earthquake was also followed by multiple aftershocks. Initial damage estimates were in the range of US$15–30 billion, around 10 to 15 percent of Chile's real gross domestic product. Chile achieved global recognition for the successful rescue of 33 trapped miners in 2010. On 5 August 2010 the access tunnel collapsed at the San José copper and gold mine in the Atacama Desert near Copiapó in northern Chile, trapping 33 men below ground. A rescue effort organized by the Chilean government located the miners 17 days later. All 33 men were brought to the surface two months later on 13 October 2010 over a period of almost 24 hours, an effort that was carried on live television around the world. 2019 Chilean protests are a series of country-wide protests in response to a raise in the Santiago Metro's subway fare, the increased cost of living, privatization and inequality prevalent in the country. On 15 November, most of the political parties represented in the National Congress signed an agreement to call a national referendum on April 2020 regarding the creation of a new Constitution, later postponed to October due to the COVID-19 pandemic. The current Constitution of Chile was approved in a national plebiscite—regarded as "highly irregular" by some observers—in September 1980, under the military dictatorship of Augusto Pinochet. It entered into force in March 1981. After Pinochet's defeat in the 1988 plebiscite, the constitution was amended to ease provisions for future amendments to the Constitution. In September 2005, President Ricardo Lagos signed into law several constitutional amendments passed by Congress. These include eliminating the positions of appointed senators and senators for life, granting the President authority to remove the commanders-in-chief of the armed forces, and reducing the presidential term from six to four years. The Congress of Chile has a 38-seat Senate and a 120-member Chamber of Deputies. Senators serve for eight years with staggered terms, while deputies are elected every 4 years. The last congressional elections were held on 17 November 2013, concurrently with the presidential election. The current Senate has a 21–15 split in favor of the governing coalition and 2 independents. The current lower house, the Chamber of Deputies, contains 67 members of the governing center-left coalition, 48 from the center-right opposition and 5 from small parties or independents. The Congress is located in the port city of Valparaíso, about west of the capital, Santiago. Chile's congressional elections are governed by a binomial system that, for the most part, rewards the two largest representations equally, often regardless of their relative popular support. Parties are thus forced to form wide coalitions and, historically, the two largest coalitions (Concertación and Alianza) split most of the seats. Only if the leading coalition ticket out-polls the second place coalition by a margin of more than 2-to-1 does the winning coalition gain both seats, which tends to lock the legislature in a roughly 50–50 split. Chile's judiciary is independent and includes a court of appeal, a system of military courts, a constitutional tribunal, and the Supreme Court of Chile. In June 2005, Chile completed a nationwide overhaul of its criminal justice system. The reform has replaced inquisitorial proceedings with an adversarial system more similar to that of the United States. In the 2001 congressional elections, the conservative Independent Democratic Union (UDI) surpassed the Christian Democrats for the first time to become the largest party in the lower house. In the 2005 parliamentary election, both leading parties, the Christian Democrats and the UDI lost representation in favor of their respective allies Socialist Party (which became the biggest party in the Concertación block) and National Renewal in the right-wing alliance. In the 2009 legislative elections in Chile, the Communist Party won 3 out of 120 seats in the Chamber of Deputies for the first time in 30 years (the Communist Party was not allowed to exist as such during the dictatorship). Chileans voted in the first round of presidential elections on 17 November 2013. None of the nine presidential candidates got more than 50 percent of the vote. As a result, the top two candidates, center-left Nueva Mayoría coalition's Michelle Bachelet and center-right Alianza coalition's Evelyn Matthei, competed in a run-off election on 15 December 2013, which Bachelet won. This was Chile's sixth presidential election since the end of the Pinochet era. All six have been judged free and fair. The president is constitutionally barred from serving consecutive terms. Since the early decades after independence, Chile has always had an active involvement in foreign affairs. In 1837 the country aggressively challenged the dominance of Peru's port of Callao for preeminence in the Pacific trade routes, defeating the short-lived alliance between Peru and Bolivia, the Peru-Bolivian Confederation (1836–39) in the War of the Confederation. The war dissolved the confederation while distributing power in the Pacific. A second international war, the War of the Pacific (1879–83), further increased Chile's regional role, while adding considerably to its territory. During the 19th century, Chile's commercial ties were primarily with Britain, a nation that had a major influence on the formation of the Chilean navy. The French influenced Chile's legal and educational systems and had a decisive impact on Chile, through the architecture of the capital in the boom years at the turn of the 20th century. German influence came from the organization and training of the army by Prussians. On 26 June 1945, Chile participated as a founding member of the United Nations being among 50 countries that signed the United Nations Charter in San Francisco, California. With the military coup of 1973, Chile became isolated politically as a result of widespread human rights abuses. Since its return to democracy in 1990, Chile has been an active participant in the international political arena. Chile completed a 2-year non-permanent position on the UN Security Council in January 2005. Jose Miguel Insulza, a Chilean national, was elected Secretary General of the Organization of American States in May 2005 and confirmed in his position, being re-elected in 2009. Chile is currently serving on the International Atomic Energy Agency (IAEA) Board of Governors, and the 2007–2008 chair of the board is Chile's ambassador to the IAEA, Milenko E. Skoknic. The country is an active member of the UN family of agencies and participates in UN peacekeeping activities. It was re-elected as a member of the UN Human Rights Council in 2011 for a three-year term. It was also elected to one of five non-permanent seats on the UN Security Council in 2013. Chile hosted the Defense Ministerial of the Americas in 2002 and the APEC summit and related meetings in 2004. It also hosted the Community of Democracies ministerial in April 2005 and the Ibero-American Summit in November 2007. An associate member of Mercosur and a full member of APEC, Chile has been a major player in international economic issues and hemispheric free trade. There have been many arguments between Chileans and Peruvians since the 1800s because they both claim boundary coastal lines. Peruvians claim the Northern part of Chile, which is now South West of Peru. This is also known as the "triangulation" which was made to fix the boundary problem between Chile and Peru. The decisions were accomplished on 27 January 1839 by the International Court of Justice This became the War of the Pacific which was in the years 1879–1883. This conflict was caused by the mineral resources Chile had, Peruvians believed they belonged to them as well. Chile had to control sea shipping to Peru and sent out an army to invade Peru on 8 October 1879. The attempt was also made to The United States but failed badly on October 1880. There was a resistance between Peruvians and Chileans for a few years because they could not reach an agreement. The United States offered help with a treaty for both Peru and Chile, better known as the Treaty of Ancón. This was later signed by both of them on 20 October 1883 to keep peace between them. In 2008, Peru took Chile to court over their maritime disputes. Then later in 2014, the International Court of Justice's ruling resulted in Chile losing 80 miles from Northwestern ocean zone. This ruling severely impacted fishers in the country making Chile lose a valuable trade in Northwest Chile. The Chilean government has diplomatic relations with most countries. It settled all its territorial disputes with Argentina during the 1990s except for part of the border at Southern Patagonian Ice Field. Chile and Bolivia severed diplomatic ties in 1978 over Bolivia's desire to regain sovereign access to the Pacific Ocean after losing it to Chile in the 1879–83 War of the Pacific. The two countries maintain consular relations and are represented at the Consul General level. In 1978 Chile was administratively divided into regions, and in 1979 subdivided into provinces and these into communes. In total the country has 16 regions, 56 provinces and 348 communes. Each region is designated by a name and a Roman numeral assigned from north to south, except for the Santiago Metropolitan Region, which does not have a number. The creation of two new regions in 2007, Arica and Parinacota (XV) and Los Ríos (XIV), and a third region in 2018, Ñuble (XVI) made this numbering lose its original order meaning. The national flower is the copihue ("Lapageria rosea", Chilean bellflower), which grows in the woods of southern Chile. The coat of arms depicts the two national animals: the condor ("Vultur gryphus", a very large bird that lives in the mountains) and the huemul ("Hippocamelus bisulcus," an endangered white tail deer). It also has the legend "Por la razón o la fuerza" ("By reason or by force"). The flag of Chile consists of two equal horizontal bands of white (top) and red; there is a blue square the same height as the white band at the hoist-side end of the white band; the square bears a white five-pointed star in the center representing a guide to progress and honor; blue symbolizes the sky, white is for the snow-covered Andes, and red stands for the blood spilled to achieve independence. The flag of Chile is similar to the Flag of Texas, although the Chilean flag is 21 years older. However, like the Texan flag, the flag of Chile is modeled after the Flag of the United States. The Armed Forces of Chile are subject to civilian control exercised by the president through the Minister of Defense. The president has the authority to remove the commanders-in-chief of the armed forces. The commander in chief of the Chilean Army is General Humberto Oviedo Arriagada. The Chilean Army is 45,000 strong and is organized with an Army headquarters in Santiago, six divisions throughout its territory, an Air Brigade in Rancagua, and a Special Forces Command in Colina. The Chilean Army is one of the most professional and technologically advanced armies in Latin America. Admiral Julio Leiva Molina directs around 25,000-person Chilean Navy, including 2,500 Marines. Of the fleet of 29 surface vessels, only eight are operational major combatants (frigates). Those ships are based in Valparaíso. The Navy operates its own aircraft for transport and patrol; there are no Navy fighter or bomber aircraft. The Navy also operates four submarines based in Talcahuano. Air Force General (four star) Jorge Rojas Ávila heads the 12,500 strong Chilean Air Force. Air assets are distributed among five air brigades headquartered in Iquique, Antofagasta, Santiago, Puerto Montt, and Punta Arenas. The Air Force also operates an airbase on King George Island, Antarctica. The Air Force took delivery of the final two of ten F-16s, all purchased from the U.S., in March 2007 after several decades of U.S. debate and previous refusal to sell. Chile also took delivery in 2007 of a number of reconditioned Block 15 F-16s from the Netherlands, bringing to 18 the total of F-16s purchased from the Dutch. After the military coup in September 1973 the Chilean national police (Carabineros) were incorporated into the Defense Ministry. With the return of democratic government, the police were placed under the operational control of the Interior Ministry but remained under the nominal control of the Defense Ministry. Gen. Gustavo González Jure is the head of the national police force of 40,964 men and women who are responsible for law enforcement, traffic management, narcotics suppression, border control, and counter-terrorism throughout Chile. In 2017, Chile signed the UN treaty on the Prohibition of Nuclear Weapons. A long and narrow coastal Southern Cone country on the west side of the Andes Mountains, Chile stretches over north to south, but only at its widest point east to west. This encompasses a remarkable variety of climates and landscapes. It contains of land area. It is situated within the Pacific Ring of Fire. Excluding its Pacific islands and Antarctic claim, Chile lies between latitudes 17° and 56°S, and longitudes 66° and 75°W. Chile is among the longest north–south countries in the world. If one considers only mainland territory, Chile is unique within this group in its narrowness from east to west, with the other long north–south countries (including Brazil, Russia, Canada, and the United States, among others) all being wider from east to west by a factor of more than 10. Chile also claims of Antarctica as part of its territory (Chilean Antarctic Territory). However, this latter claim is suspended under the terms of the Antarctic Treaty, of which Chile is a signatory. It is the world's southernmost country that is geographically on the mainland. Chile controls Easter Island and Sala y Gómez Island, the easternmost islands of Polynesia, which it incorporated to its territory in 1888, and the Juan Fernández Islands, more than from the mainland. Also controlled but only temporarily inhabited (by some local fishermen) are the small islands of San Ambrosio and San Felix. These islands are notable because they extend Chile's claim to territorial waters out from its coast into the Pacific Ocean. The northern Atacama Desert contains great mineral wealth, primarily copper and nitrates. The relatively small Central Valley, which includes Santiago, dominates the country in terms of population and agricultural resources. This area is also the historical center from which Chile expanded in the late 19th century, when it integrated the northern and southern regions. Southern Chile is rich in forests, grazing lands, and features a string of volcanoes and lakes. The southern coast is a labyrinth of fjords, inlets, canals, twisting peninsulas, and islands. The Andes Mountains are located on the eastern border. The diverse climate of Chile ranges from the world's driest desert in the north—the Atacama Desert—through a Mediterranean climate in the center, humid subtropical in Easter Island, to an oceanic climate, including alpine tundra and glaciers in the east and south. According to the Köppen system, Chile within its borders hosts at least ten major climatic subtypes. There are four seasons in most of the country: summer (December to February), autumn (March to May), winter (June to August), and spring (September to November). The flora and fauna of Chile are characterized by a high degree of endemism, due to its particular geography. In continental Chile, the Atacama Desert in the north and the Andes mountains to the east are barriers that have led to the isolation of flora and fauna. Add to that the enormous length of Chile (over ) and this results in a wide range of climates and environments that can be divided into three general zones: the desert provinces of the north, central Chile, and the humid regions of the south. The native flora of Chile consists of relatively fewer species compared to the flora of other South American countries. The northernmost coastal and central region is largely barren of vegetation, approaching the most absolute desert in the world. On the slopes of the Andes, in addition to the scattered tola desert brush, grasses are found. The central valley is characterized by several species of cacti, the hardy espinos, the Chilean pine, the southern beeches and the copihue, a red bell-shaped flower that is Chile's national flower. In southern Chile, south of the Biobío River, heavy precipitation has produced dense forests of laurels, magnolias, and various species of conifers and beeches, which become smaller and more stunted to the south. The cold temperatures and winds of the extreme south preclude heavy forestation. Grassland is found in Atlantic Chile (in Patagonia). Much of the Chilean flora is distinct from that of neighboring Argentina, indicating that the Andean barrier existed during its formation. Some of Chile's flora has an Antarctic origin due to land bridges which formed during the Cretaceous ice ages, allowing plants to migrate from Antarctica to South America. Just over 3,000 species of fungi are recorded in Chile, but this number is far from complete. The true total number of fungal species occurring in Chile is likely to be far higher, given the generally accepted estimate that only about 7 percent of all fungi worldwide have so far been discovered. Although the amount of available information is still very small, a first effort has been made to estimate the number of fungal species endemic to Chile, and 1995 species have been tentatively identified as possible endemics of the country. Chile's geographical isolation has restricted the immigration of faunal life, so that only a few of the many distinctive South American animals are found. Among the larger mammals are the puma or cougar, the llama-like guanaco and the fox-like chilla. In the forest region, several types of marsupials and a small deer known as the pudu are found. There are many species of small birds, but most of the larger common Latin American types are absent. Few freshwater fish are native, but North American trout have been successfully introduced into the Andean lakes. Owing to the vicinity of the Humboldt Current, ocean waters abound with fish and other forms of marine life, which in turn support a rich variety of waterfowl, including several penguins. Whales are abundant, and some six species of seals are found in the area. Chile is located along a highly seismic and volcanic zone, part of the Pacific Ring of Fire, due to the subduction of the Nazca and Antarctic plates in the South American plate. Late Paleozoic, 251 million years ago, Chile belonged to the continental block called Gondwana. It was just a depression accumulated marine sediments began to rise at the end of the Mesozoic, 66 million years ago, due to the collision between the Nazca and South American plates, resulting in the Andes. The territory would be shaped by millions of years due to the folding of the rocks, forming the current relief. The Chilean relief consists of the central depression, which crosses the country longitudinally, flanked by two mountain ranges that make up about 80% of the territory: the Andes mountains to the east-natural border with Bolivia and Argentina in the region of Atacama and the Coastal Range west-minor height from the Andes. Chile's highest peak is the Nevado Ojos del Salado, at 6891.3 m, which is also the highest volcano in the world. The highest point of the Coastal Range is Vicuña Mackenna, at 3114 meters, located in the Sierra Vicuña Mackenna, the south of Antofagasta. Among the coastal mountains and the Pacific is a series of coastal plains, of variable length, which allow the settlement of coastal towns and big ports. Some areas of the plains territories encompass territory east of the Andes, and the Patagonian steppes and Magellan, or are high plateaus surrounded by high mountain ranges, such as the Altiplano or Puna de Atacama. The Far North is the area between the northern boundary of the country and the parallel 26° S, covering the first three regions. It is characterized by the presence of the Atacama desert, the most arid in the world. The desert is fragmented by streams that originate in the area known as the pampas Tamarugal. The Andes, split in two and whose eastern arm runs Bolivia, has a high altitude and volcanic activity, which has allowed the formation of the Andean altiplano and salt structures as the Salar de Atacama, due to the gradual accumulation of sediments over time. To the south is the Norte Chico, extending to the Aconcagua river. Los Andes begin to decrease its altitude to the south and closer to the coast, reaching 90 km away at the height of Illapel, the narrowest part of the Chilean territory. The two mountain ranges intersect, virtually eliminating the intermediate depression. The existence of rivers flowing through the territory allows the formation of transverse valleys, where agriculture has developed strongly in recent times, while the coastal plains begin to expand. The Central area is the most populated region of the country. The coastal plains are wide and allow the establishment of cities and ports along the Pacific. The Andes maintains altitudes above 6000m but descend slowly starts approaching the 4000 meters on average. The intermediate depression reappears becoming a fertile valley that allows agricultural development and human settlement, due to sediment accumulation. To the south, the Cordillera de la Costa reappears in the range of Nahuelbuta while glacial sediments originate a series of lakes in the area of La Frontera. Patagonia extends from within Reloncavi, at the height of parallel 41°S, to the south. During the last glaciation, this area was covered by ice that strongly eroded Chilean relief structures. As a result, the intermediate depression sinks in the sea, while the coastal mountains rise to a series of archipelagos, such as Chiloé and the Chonos, disappearing in Taitao peninsula, in the parallel 47°S. The Andes mountain range loses height and erosion caused by the action of glaciers has caused fjords. East of the Andes, on the continent, or north of it, on the island of Tierra del Fuego are located relatively flat plains, which in the Strait of Magellan cover large areas. The Andes, as he had done previously Cordillera de la Costa, begins to break in the ocean causing a myriad of islands and islets and disappear into it, sinking and reappearing in the Southern Antilles arc and then the Antarctic Peninsula, where it is called Antartandes, in the Chilean Antarctic Territory, lying between the meridians 53°W and 90°W. In the middle of the Pacific, the country has sovereignty over several islands of volcanic origin, collectively known as Insular Chile. Of these, we highlight the archipelago of Juan Fernandez and Easter Island, which is located in the fracture zone between the Nazca plate and the Pacific plate known as East Pacific Rise. Due to the characteristics of the territory, Chile is crossed by numerous rivers generally short in length and with low torrential flow. They commonly extend from the Andes to the Pacific Ocean, flowing in an East to West. Because of the Atacama desert, in the Norte Grande there are only short endorheic character streams, except for the Loa River, the longest in the country 440 km. In the high valleys, wetland areas generate Chungará Lake, located at 4500 meters above sea level. It and the Lauca River are shared with Bolivia, as well as the Lluta River. In the center-north of the country, the number of rivers that form valleys of agricultural importance increases. Noteworthy are the Elqui with 75 km long, 142 km Aconcagua, Maipo with 250 km and its tributary, the Mapocho with 110 km, and Maule with 240 km. Their waters mainly flow from Andean snowmelt in the summer and winter rains. The major lakes in this area are the artificial lake Rapel, the Colbun Maule lagoon and the lagoon of La Laja. Chile's 2017 census reported a population of 17,574,003. Its rate of population growth has been decreasing since 1990, due to a declining birth rate. By 2050 the population is expected to reach approximately 20.2 million people. About 85 percent of the country's population lives in urban areas, with 40 percent living in Greater Santiago. The largest agglomerations according to the 2002 census are Greater Santiago with 5.6 million people, Greater Concepción with 861,000 and Greater Valparaíso with 824,000. Mexican professor Francisco Lizcano, of the National Autonomous University of Mexico, estimated that 52.7% of Chileans were white, 39.3% were mestizo, and 8% were Amerindian. The most recent study in the Candela Project establishes that the genetic composition of Chile is 52% of European origin, with 44% of the genome coming from Native Americans (Amerindians), and 4% coming from Africa, making Chile a primarily mestizo country with traces of African descent present in half of the population. Another genetic study conducted by the University of Brasilia in several American countries shows a similar genetic composition for Chile, with a European contribution of 51.6%, an Amerindian (Native) contribution of 42.1%, and an African contribution of 6.3%. A public health booklet from the University of Chile states that 64% of the population is of Caucasian origin; "predominantly White" Mestizos are estimated to amount a total of 35%, while Native Americans (Amerindians) comprise the remaining 5%. Despite the genetic considerations, many Chileans, if asked, would self-identify as White. The 2011 Latinobarómetro survey asked respondents in Chile what race they considered themselves to belong to. Most answered "White" (59%), while 25% said "Mestizo" and 8% self-classified as "indigenous". A 2002 national poll revealed that a majority of Chileans believed they possessed some (43.4%) or much (8.3%) "indigenous blood", while 40.3% responded that they had none. The 1907 census reported 101,118 Natives, or 3.1% of the total population. Only those that practiced their native culture or spoke their native language were considered to be Natives, irrespective of their "racial purity". In 2002 a census took place, directly asking the public whether they considered themselves as part of any of the eight Chilean ethnic groups, regardless of whether or not they maintained their culture, traditions and language, and 4.6 percent of the population (692,192 people) fitted that description of indigenous peoples in Chile. Of that number, 87.3% declared themselves Mapuche. Most of the indigenous population shows varying degrees of mixed ancestry. Chile is one of 22 countries to have signed and ratified the only binding international law concerning indigenous peoples, the Indigenous and Tribal Peoples Convention, 1989. It was adopted in 1989 as the International Labour Organization (ILO) Convention 169. Chile ratified it in 2008. A Chilean court decision in November 2009 considered to be a landmark ruling on indigenous rights and made use of the convention. The Supreme Court decision on Aymara water rights upheld rulings by both the Pozo Almonte tribunal and the Iquique Court of Appeals, and marks the first judicial application of ILO Convention 169 in Chile. The earliest European immigrants were Spanish colonisers who arrived in the 16th century. The Amerindian population of central Chile was absorbed into the Spanish settler population in the beginning of the colonial period to form the large mestizo population that exists in Chile today; mestizos create modern middle and lower classes. In the 18th and 19th centuries, many Basques came to Chile where they integrated into the existing elites of Castilian origin. Postcolonial Chile was never a particularly attractive destination for migrants, owing to its remoteness and distance from Europe. Europeans preferred to stay in countries closer to their homelands instead of taking the long journey through the Straits of Magellan or crossing the Andes. European migration did not result in a significant change in the ethnic composition of Chile, except in the region of Magellan. Spaniards were the only major European migrant group to Chile, and there was never large-scale immigration such as that to Argentina or Uruguay. Between 1851 and 1924, Chile only received 0.5% of European immigration to Latin America, compared to 46% to Argentina, 33% to Brazil, 14% to Cuba, and 4% to Uruguay. However, it is undeniable that immigrants have played a significant role in Chilean society. Other groups of Europeans have followed but are found in smaller numbers, like the descendants of Austrians and Dutch people. Currently, these are estimated at about 50,000 people. After the failed liberal revolution of 1848 in the German states, a noticeable German immigration took place, laying the foundation for the German Chileans. Sponsored by the Chilean government to "unbarbarize" and colonize the southern region, these Germans (notably the Swiss, Silesians, Alsatians and Austrians) settled mainly in Valdivia, Osorno and Llanquihue. Descendants of different European ethnic groups often intermarried in Chile. This intermarriage and mixture of cultures and races have helped to shape the present society and culture of the Chilean middle and upper classes. Due in part to its economic fortunes, Chile has recently become a new magnet for immigrants, mostly from neighboring Venezuela, Peru, Haiti, Colombia, Bolivia and Argentina. According to the 2002 national census, Chile's foreign-born population has increased by 75% since 1992. According to an estimate by the Migration and Foreign Residency Department 1,200,000 foreigners were living in Chile in 2017. Roughly 500,000 of Chile's population is of full or partial Palestinian origin. , 66.6% of Chilean population over 15 years of age claimed to adhere to the Roman Catholic church, a decrease from the 70% reported in the 2002 census. In the same census of 2012, 17% of Chileans reported adherence to an Evangelical church ("Evangelical" in the census referred to all Christian denominations other than the Roman Catholic and Orthodox—Greek, Persian, Serbian, Ukrainian, and Armenian—churches, The Church of Jesus Christ of Latter-day Saints or Mormons, Seventh-day Adventists, and Jehovah's Witnesses: essentially, those denominations generally still termed "Protestant" in most English-speaking lands, although Adventism is often considered an Evangelical denomination as well). Approximately 90% of Evangelical Christians are Pentecostal. but Wesleyan, Lutheran, Anglican, Episcopalian, Presbyterian, other Reformed, Baptist, and Methodist churches also are present amongst Chilean Evangelical churches. Irreligious people, atheists, and agnostics account for around 12% of the population. By 2015, the major religion in Chile remained Christianity (68%), with an estimated 55% of Chileans belonging to the Roman Catholic church, 13% to various Evangelical churches, and just 7% adhering to any other religion. Agnostics and atheist were estimated at 25% of the population. Chile has a Bahá'í religious community, and is home to the mother temple of the Bahá'ís for Latin America, which was completed in 2016 and has been described as futuristic and translucent in its architectural style. The Constitution guarantees the right to freedom of religion, and other laws and policies contribute to generally free religious practice. The law at all levels fully protects this right against abuse by either governmental or private actors. Church and state are officially separate in Chile. A 1999 law on religion prohibits religious discrimination. However, the Roman Catholic church for mostly historical and social reasons enjoys a privileged status and occasionally receives preferential treatment. Government officials attend Roman Catholic events as well as major Evangelical and Jewish ceremonies. The Chilean government treats the religious holidays of Christmas, Good Friday, the Feast of the Virgin of Carmen, the Feast of Saints Peter and Paul, the Feast of the Assumption, All Saints' Day, and the Feast of the Immaculate Conception as national holidays. Recently, the government declared 31 October, Reformation Day, to be an additional national holiday, in honor of the Evangelical churches of the country. The patron saints of Chile are Our Lady of Mount Carmel and Saint James the Greater ("Santiago"). In 2005, Pope Benedict XVI canonized Alberto Hurtado, who became the country's second native Roman Catholic saint after Teresa de los Andes. The Spanish spoken in Chile is distinctively accented and quite unlike that of neighboring South American countries because final syllables are often dropped, and some consonants have a soft pronunciation. Accent varies only very slightly from north to south; more noticeable are the differences in accent based on social class or whether one lives in the city or the country. That the Chilean population was largely formed in a small section at the center of the country and then migrated in modest numbers to the north and south helps explain this relative lack of differentiation, which was maintained by the national reach of radio, and now television, which also helps to diffuse and homogenize colloquial expressions. There are several indigenous languages spoken in Chile: Mapudungun, Quechua, Aymara and Rapa Nui. After the Spanish invasion, Spanish took over as the lingua franca and the indigenous languages have become minority languages, with some now extinct or close to extinction. German is still spoken to some extent in southern Chile, either in small country side pockets or as a second language among the communities of larger cities. Through initiatives such as the English Opens Doors Program, the government made English mandatory for students in fifth-grade and above in public schools. Most private schools in Chile start teaching English from kindergarten. Common English words have been absorbed and appropriated into everyday Spanish speech. In Chile, education begins with preschool until the age of 5. Primary school is provided for children between ages 6 and 13. Students then attend secondary school until graduation at age 17. Secondary education is divided into two parts: During the first two years, students receive a general education. Then, they choose a branch: scientific humanistic education, artistic education, or technical and professional education. Secondary school ends two years later on the acquirement of a certificate (licencia de enseñanza media). Chilean education is segregated by wealth in a three-tiered system — the quality of the schools reflect socioeconomic backgrounds: Upon successful graduation of secondary school, students may continue into higher education. The higher education schools in Chile consist of Chilean Traditional Universities and are divided into public universities or private universities. There are medical schools and both the Universidad de Chile and Universidad Diego Portales offer law schools in a partnership with Yale University. The Ministry of Health ("Minsal") is the cabinet-level administrative office in charge of planning, directing, coordinating, executing, controlling and informing the public health policies formulated by the President of Chile. The National Health Fund ("Fonasa"), created in 1979, is the financial entity entrusted to collect, manage and distribute state funds for health in Chile. It is funded by the public. All employees pay 7 percent of their monthly income to the fund. Fonasa is part of the NHSS and has executive power through the Ministry of Health (Chile). Its headquarters are in Santiago and decentralized public service is conducted by various Regional Offices. More than 12 million beneficiaries benefit from Fonasa. Beneficiaries can also opt for more costly private insurance through Isapre. Hospitals in Chile are mainly located in the Santiago Metropolitan Region. The Central Bank of Chile in Santiago serves as the central bank for the country. The Chilean currency is the Chilean peso (CLP). Chile is one of South America's most stable and prosperous nations, leading Latin American nations in human development, competitiveness, income per capita, globalization, economic freedom, and low perception of corruption. Since July 2013, Chile is considered by the World Bank as a "high-income economy". Chile has the highest degree of economic freedom in South America (ranking 7th worldwide), owing to its independent and efficient judicial system and prudent public finance management. In May 2010 Chile became the first South American country to join the OECD. In 2006, Chile became the country with the highest nominal GDP per capita in Latin America. Copper mining makes up 20% of Chilean GDP and 60% of exports. Escondida is the largest copper mine in the world, producing over 5% of global supplies. Overall, Chile produces a third of the world's copper. Codelco, the state mining firm, competes with private copper mining companies. Sound economic policies, maintained consistently since the 1980s, have contributed to steady economic growth in Chile and have more than halved poverty rates. Chile began to experience a moderate economic downturn in 1999. The economy remained sluggish until 2003, when it began to show clear signs of recovery, achieving 4.0% GDP growth. The Chilean economy finished 2004 with growth of 6 percent. Real GDP growth reached 5.7 percent in 2005 before falling back to 4 percent in 2006. GDP expanded by 5 percent in 2007. Faced with an international economic downturn the government announced an economic stimulus plan to spur employment and growth, and despite the global financial crisis, aimed for an expansion of between 2 percent and 3 percent of GDP for 2009. Nonetheless, economic analysts disagreed with government estimates and predicted economic growth at a median of 1.5 percent. Real GDP growth in 2012 was 5.5%. Growth slowed to 4.1% in the first quarter of 2013. The unemployment rate was 6.4% in April 2013. There are reported labor shortages in agriculture, mining, and construction. The percentage of Chileans with per capita household incomes below the poverty line—defined as twice the cost of satisfying a person's minimal nutritional needs—fell from 45.1 percent in 1987 to 11.5 percent in 2009, according to government surveys. Critics in Chile, however, argue that true poverty figures are considerably higher than those officially published. Using the relative yardstick favoured in many European countries, 27% of Chileans would be poor, according to Juan Carlos Feres of the ECLAC. , about 11.1 million people (64% of the population) benefit from government welfare programs, via the "Social Protection Card", which includes the population living in poverty and those at a risk of falling into poverty. The privatized national pension system (AFP) has encouraged domestic investment and contributed to an estimated total domestic savings rate of approximately 21 percent of GDP. Under the compulsory private pension system, most formal sector employees pay 10 percent of their salaries into privately managed funds. Chile has signed free trade agreements (FTAs) with a whole network of countries, including an FTA with the United States that was signed in 2003 and implemented in January 2004. Internal Government of Chile figures show that even when factoring out inflation and the recent high price of copper, bilateral trade between the U.S. and Chile has grown over 60 percent since then. Chile's total trade with China reached US$8.8 billion in 2006, representing nearly 66 percent of the value of its trade relationship with Asia. Exports to Asia increased from US$15.2 billion in 2005 to US$19.7 billion in 2006, a 29.9 percent increase. Year-on-year growth in imports was especially strong from a number of countries: Ecuador (123.9%), Thailand (72.1%), South Korea (52.6%), and China (36.9%). Chile's approach to foreign direct investment is codified in the country's Foreign Investment Law. Registration is reported to be simple and transparent, and foreign investors are guaranteed access to the official foreign exchange market to repatriate their profits and capital. The Chilean Government has formed a Council on Innovation and Competition, hoping to bring in additional FDI to new parts of the economy. Standard & Poor's gives Chile a credit rating of AA-. The Government of Chile continues to pay down its foreign debt, with public debt only 3.9 percent of GDP at the end of 2006. The Chilean central government is a net creditor with a net asset position of 7% of GDP at end 2012. The current account deficit was 4% in the first quarter of 2013, financed mostly by foreign direct investment. 14% of central government revenue came directly from copper in 2012. Chile is rich in mineral resources, especially copper and lithium. It is thought that due to the importance of lithium for batteries for electric vehicles and stabilization of electric grids with large proportions of intermittent renewables in the electricity mix, Chile could be strengthened geopolitically. However, this perspective has also been criticised for underestimating the power of economic incentives for expanded production in other parts of the world. Agriculture in Chile encompasses a wide range of different activities due to its particular geography, climate and geology and human factors. Historically agriculture is one of the bases of Chile's economy. Now agriculture and allied sectors like forestry, logging and fishing account for only 4.9% of the GDP and employ 13.6% of the country's labor force. Some major agriculture products of Chile include grapes, apples, pears, onions, wheat, maize, oats, peaches, garlic, asparagus, beans, beef, poultry, wool, fish, timber and hemp. Due to its geographical isolation and strict customs policies Chile is free from diseases such as mad cow disease, fruit fly and Phylloxera. This, its location in the Southern Hemisphere, which has quite different harvesting times from the Northern Hemisphere, and its wide range of agriculture conditions are considered Chile's main comparative advantages. However, Chile's mountainous landscape limits the extent and intensity of agriculture so that arable land corresponds only to 2.62% of the total territory. Chile currently utilizes 14,015 Hectares of agricultural land. Tourism in Chile has experienced sustained growth over the last few decades. In 2005, tourism grew by 13.6 percent, generating more than 4.5 billion dollars of which 1.5 billion was attributed to foreign tourists. According to the National Service of Tourism (Sernatur), 2 million people a year visit the country. Most of these visitors come from other countries in the American continent, mainly Argentina; followed by a growing number from the United States, Europe, and Brazil with a growing number of Asians from South Korea and China. The main attractions for tourists are places of natural beauty situated in the extreme zones of the country: San Pedro de Atacama, in the north, is very popular with foreign tourists who arrive to admire the Incaic architecture, the altiplano lakes, and the Valley of the Moon. In Putre, also in the north, there is the Chungará Lake, as well as the Parinacota and the Pomerape volcanoes, with altitudes of 6,348 m and 6,282 m, respectively. Throughout the central Andes there are many ski resorts of international repute, including Portillo, Valle Nevado and Termas de Chillán. The main tourist sites in the south are national parks (the most popular is Conguillío National Park in the Araucanía) and the coastal area around Tirúa and Cañete with the Isla Mocha and the Nahuelbuta National Park, Chiloé Archipelago and Patagonia, which includes Laguna San Rafael National Park, with its many glaciers, and the Torres del Paine National Park. The central port city of Valparaíso, which is World Heritage with its unique architecture, is also popular. Finally, Easter Island in the Pacific Ocean is one of the main Chilean tourist destinations. For locals, tourism is concentrated mostly in the summer (December to March), and mainly in the coastal beach towns. Arica, Iquique, Antofagasta, La Serena and Coquimbo are the main summer centers in the north, and Pucón on the shores of Lake Villarrica is the main center in the south. Because of its proximity to Santiago, the coast of the Valparaíso Region, with its many beach resorts, receives the largest number of tourists. Viña del Mar, Valparaíso's northern affluent neighbor, is popular because of its beaches, casino, and its annual song festival, the most important musical event in Latin America. Pichilemu in the O'Higgins Region is widely known as South America's "best surfing spot" according to Fodor's. In November 2005 the government launched a campaign under the brand "Chile: All Ways Surprising" intended to promote the country internationally for both business and tourism. Museums in Chile such as the Chilean National Museum of Fine Arts built in 1880, feature works by Chilean artists. Chile is home to the world renowned Patagonian Trail that resides on the border between Argentina and Chile. Chile recently launched a massive scenic route for tourism in hopes of encouraging development based on conservation. The Route of Parks covers 1,740 miles and was designed by Tompkin Conservation (founders Douglas Tompkins and wife Kristine). Due to Chile's topography a functioning transport network is vital to its economy. Buses are now the main means of long-distance transportation in Chile, following the decline of its railway network. The bus system covers the entire country, from Arica to Santiago (a 30-hour journey) and from Santiago to Punta Arenas (about 40 hours, with a change at Osorno). Chile has a total of 372 runways (62 paved and 310 unpaved). Important airports in Chile include Chacalluta International Airport (Arica), Diego Aracena International Airport (Iquique), Andrés Sabella Gálvez International Airport (Antofagasta), Carriel Sur International Airport (Concepción), El Tepual International Airport (Puerto Montt), Presidente Carlos Ibáñez del Campo International Airport (Punta Arenas), La Araucanía International Airport (Temuco), Mataveri International Airport (Easter Island), the most remote airport in the world, and the Comodoro Arturo Merino Benítez International Airport (Santiago) with a traffic of 12,105,524 passengers in 2011. Santiago is headquarters of Latin America's largest airline holding company and Chilean flag carrier LATAM Airlines. Chile has a telecommunication system which covers much of the country, including Chilean insular and Antarctic bases. Privatization of the telephone system began in 1988; Chile has one of the most advanced telecommunications infrastructure in South America with a modern system based on extensive microwave radio relay facilities and domestic satellite system with 3 earth stations. In 2012, there were 3.276 million main lines in use and 24.13 million mobile cellular telephone subscribers. According to a 2012 database of the International Telecommunications Union (ITU), 61.42% of the Chilean population uses the internet, making Chile the country with the highest internet penetration in South America. The Chilean internet country code is ".cl". From the period between early agricultural settlements and up to the late pre-Hispanic period, northern Chile was a region of Andean culture that was influenced by altiplano traditions spreading to the coastal valleys of the north, while southern regions were areas of Mapuche cultural activities. Throughout the colonial period following the conquest, and during the early Republican period, the country's culture was dominated by the Spanish. Other European influences, primarily English, French, and German began in the 19th century and have continued to this day. German migrants influenced the Bavarian style rural architecture and cuisine in the south of Chile in cities such as Valdivia, Frutillar, Puerto Varas, Osorno, Temuco, Puerto Octay, Llanquihue, Faja Maisan, Pitrufquén, Victoria, Pucón and Puerto Montt. Music in Chile ranges from folkloric, popular and classical music. Its large geography generates different musical styles in the north, center and south of the country, including also Easter Island and Mapuche music. The national dance is the cueca. Another form of traditional Chilean song, though not a dance, is the tonada. Arising from music imported by the Spanish colonists, it is distinguished from the cueca by an intermediate melodic section and a more prominent melody. Between 1950 and 1970 appears a rebirth in folk music leading by groups such as Los de Ramón, Los Cuatro Huasos and Los Huasos Quincheros, among others with composers such as Raúl de Ramón, Violeta Parra and others. In the mid-1960s native musical forms were revitalized by the Parra family with the Nueva canción Chilena, which was associated with political activists and reformers such as Víctor Jara, Inti-Illimani, and Quilapayún. Other important folk singer and researcher on folklore and Chilean ethnography, is Margot Loyola. Also many Chilean rock bands like Los Jaivas, Los Prisioneros, La Ley, and Los Tres have reached international success. In February, annual music festivals are held in Viña del Mar. Chile is a "country of poets". Gabriela Mistral was the first Latin American to receive a Nobel Prize in Literature (1945). Chile's most famous poet is Pablo Neruda, who received the Nobel Prize for Literature (1971) and is world-renowned for his extensive library of works on romance, nature, and politics. His three highly personalized homes in Isla Negra, Santiago and Valparaíso are popular tourist destinations. Among the list of other Chilean poets are Carlos Pezoa Véliz, Vicente Huidobro, Gonzalo Rojas, Pablo de Rokha, Nicanor Parra and Raúl Zurita. Isabel Allende is the best-selling Chilean novelist, with 51 millions of her novels sold worldwide. Novelist José Donoso's novel "The Obscene Bird of Night" is considered by critic Harold Bloom to be one of the canonical works of 20th-century Western literature. Another internationally recognized Chilean novelist and poet is Roberto Bolaño whose translations into English have had an excellent reception from the critics. Chilean cuisine is a reflection of the country's topographical variety, featuring an assortment of seafood, beef, fruits, and vegetables. Traditional recipes include asado, cazuela, empanadas, humitas, pastel de choclo, pastel de papas, curanto and sopaipillas. Crudos is an example of the mixture of culinary contributions from the various ethnic influences in Chile. The raw minced llama, heavy use of shellfish and rice bread were taken from native Quechua Andean cuisine, (although now beef brought to Chile by Europeans is also used in place of the llama meat), lemon and onions were brought by the Spanish colonists, and the use of mayonnaise and yogurt was introduced by German immigrants, as was beer. The folklore of Chile, cultural and demographic characteristics of the country, is the result of mixture of Spanish and Amerindian elements that occurred during the colonial period. Due to cultural and historical reasons, they are classified and distinguished four major areas in the country: northern areas, central, southern and south. Most of the traditions of the culture of Chile have a festive purpose, but some, such as dances and ceremonies, have religious components. Chilean mythology is the mythology and beliefs of the Folklore of Chile. This includes Chilote mythology, Rapa Nui mythology and Mapuche mythology. Film production originated in Valparaíso on 26 May 1902 with the premiere of the documentary "Exercise General Fire Brigade", the first film completely filmed and processed in the country. In the following decades, marked milestones "The deck of Death" (or "The Enigma of Lord Street") (1916), considered the first film of a Chilean story, "The transmission of presidential" (1920), the first animated film in the country, and "North and South" (1934), the first sound film of Chile. Chile's most popular sport is association football. Chile has appeared in nine FIFA World Cups which includes hosting the 1962 FIFA World Cup where the national football team finished third. Other results achieved by the national football team include two Copa América titles (2015 and 2016), and two runners up positions, one silver and two bronze medals at the Pan American Games, a bronze medal at the 2000 Summer Olympics and two third places finishes in the FIFA under-17 and under-20 youth tournaments. The top league in the Chilean football league system is the Chilean Primera División, which is named by the IFFHS as the ninth strongest national football league in the world. The main football clubs are Colo-Colo, Universidad de Chile and Universidad Católica. Colo-Colo is the country's most successful football club, having both the most national and international championships, including the coveted Copa Libertadores South American club tournament. Universidad de Chile was the last international champion (Copa Sudamericana 2011). Tennis is Chile's most successful sport. Its national team won the World Team Cup clay tournament twice (2003 & 2004), and played the Davis Cup final against Italy in 1976. At the 2004 Summer Olympics the country captured gold and bronze in men's singles and gold in men's doubles. Marcelo Ríos became the first Latin American man to reach the number one spot in the ATP singles rankings in 1998. Anita Lizana won the US Open in 1937, becoming the first woman from Latin America to win a Grand Slam tournament. Luis Ayala was twice a runner-up at the French Open and both Ríos and Fernando González reached the Australian Open men's singles finals. González also won a silver medal in singles at the 2008 Summer Olympics in Beijing. At the Summer Olympic Games Chile boasts a total of two gold medals (tennis), seven silver medals (athletics, equestrian, boxing, shooting and tennis) and four bronze medals (tennis, boxing and football). In 2012, Chile won its first Paralympic Games medal (gold in Athletics). Rodeo is the country's national sport and is practiced in the more rural areas of the nation. A sport similar to hockey called "chueca" was played by the Mapuche people during the Spanish conquest. Skiing and snowboarding are practiced at ski centers located in the Central Andes, and in southern ski centers near to cities as Osorno, Puerto Varas, Temuco and Punta Arenas. surfing is popular at some coastal towns. Polo is professionally practiced within Chile, with the country achieving top prize in the 2008 and 2015 World Polo Championship. Basketball is a popular sport in which Chile has earned a bronze medal in the first men's FIBA World Championship held in 1950 and winning a second bronze medal when Chile hosted the 1959 FIBA World Championship. Chile hosted the first FIBA World Championship for Women in 1953 finishing the tournament with the silver medal. San Pedro de Atacama is host to the annual "Atacama Crossing", a six-stage, footrace which annually attracts about 150 competitors from 35 countries. The Dakar Rally off-road automobile race has been held in both Chile and Argentina since 2009. The cultural heritage of Chile consists, first, of their intangible heritage, composed of various cultural events, such as visual arts, crafts, dances, holidays, cuisine, games, music and traditions, and, secondly, by its tangible, consists of those buildings, objects and sites of archaeological, architectural, traditional, artistic, ethnographic, folkloric, historical, religious or technological scattered through Chilean territory, among them, those goods are declared World Heritage Site by UNESCO, in accordance with the provisions of the Convention concerning the Protection of World Cultural and Natural Heritage of 1972, ratified by Chile in 1980. These cultural sites are the Rapa Nui National Park (1995), the Churches of Chiloé (2000), the historical district of the port city of Valparaíso (2003), Humberstone and Santa Laura Saltpeter Works (2005) and the mining city Sewell (2006). In 1999 the "Cultural Heritage Day" was established as a way to honour and commemorate Chiles cultural heritage. It is an official national event celebrated in May every year.
https://en.wikipedia.org/wiki?curid=5489
History of Chile The territory of Chile has been populated since at least 3000 BC. By the 16th century, Spanish conquistadors began to subdue and colonize the region of present-day Chile, and the territory was a colony between 1540 and 1818, when it gained independence from Spain. The country's economic development was successively marked by the export of first agricultural produce, then saltpeter and later copper. The wealth of raw materials led to an economic upturn, but also led to dependency, and even wars with neighboring states. Chile was governed during most of its first 150 years of independence by different forms of restricted government, where the electorate was carefully vetted and controlled by an elite. Failure to address the economic and social increases and increasing political awareness of the less-affluent population, as well as indirect intervention and economic funding to the main political groups by the CIA, as part of the Cold War, led to a political polarization under Socialist President Salvador Allende. This in turn resulted in the 1973 coup d'état and the military dictatorship of General Augusto Pinochet, whose subsequent 17-year regime was responsible for both numerous human rights violations and deep market-oriented economic reforms. In 1990, Chile finally made a peaceful transition to democracy. About 10,000 years ago, migrating Native Americans settled in the fertile valleys and coastal areas of what is present day Chile. Pre-Hispanic Chile was home to over a dozen different Amerindian societies. The current prevalent theories are that the initial arrival of humans to the continent took place either along the Pacific coast southwards in a rather rapid expansion long preceding the Clovis culture, or even trans-Pacific migration. These theories are backed by findings in the Monte Verde archaeological site, which predates the Clovis site by thousands of years. Specific early human settlement sites from the very early human habitation in Chile include the Cueva del Milodon and the Pali Aike Crater's lava tube. Despite such diversity, it is possible to classify the indigenous people into three major cultural groups: the northern people, who developed rich handicrafts and were influenced by pre-Incan cultures; the Araucanian culture, who inhabited the area between the river Choapa and the island of Chiloé, and lived primarily off agriculture; and the Patagonian culture composed of various nomadic tribes, who supported themselves through fishing and hunting (and who in Pacific/Pacific Coast immigration scenario would be descended partly from the most ancient settlers). No elaborate, centralized, sedentary civilization reigned supreme. The Araucanians, a fragmented society of hunters, gatherers, and farmers, constituted the largest Native American group in Chile. A mobile people who engaged in trade and warfare with other indigenous groups, they lived in scattered family clusters and small villages. Although the Araucanians had no written language, they did use a common tongue. Those in what became central Chile were more settled and more likely to use irrigation. Those in the south combined slash-and-burn agriculture with hunting. Of the three Araucanian groups, the one that mounted the fiercest resistance to the attempts at seizure of their territory were the Mapuche, meaning "people of the land." The Inca Empire briefly extended their empire into what is now northern Chile, where they collected tribute from small groups of fishermen and oasis farmers but were not able to establish a strong cultural presence in the area. As the Spaniards would after them, the Incas encountered fierce resistance and so were unable to exert control in the south. During their attempts at conquest in 1460 and again in 1491, the Incas established forts in the Central Valley of Chile, but they could not colonize the region. The Mapuche fought against the Sapa Tupac Inca Yupanqui (c. 1471–1493) and his army. The result of the bloody three-day confrontation known as the Battle of the Maule was that the Inca conquest of the territories of Chile ended at the Maule river, which subsequently became the boundary between the Incan empire and the Mapuche lands until the arrival of the Spaniards. Scholars speculate that the total Araucanian population may have numbered 1.5 million at most when the Spaniards arrived in the 1530s; a century of European conquest and disease reduced that number by at least half. During the conquest, the Araucanians quickly added horses and European weaponry to their arsenal of clubs and bows and arrows. They became adept at raiding Spanish settlements and, albeit in declining numbers, managed to hold off the Spaniards and their descendants until the late 19th century. The Araucanians' valor inspired the Chileans to mythologize them as the nation's first national heroes, a status that did nothing, however, to elevate the wretched living standard of their descendants. The Chilean Patagonia located south of the Calle-Calle River in Valdivia was composed of many tribes, mainly Tehuelches, who were considered giants by Spaniards during Magellan's voyage of 1520. The name Patagonia comes from the word "patagón" used by Magellan to describe the native people whom his expedition thought to be giants. It is now believed the Patagons were actually Tehuelches with an average height of 1.80 m (~5′11″) compared to the 1.55 m (~5′1″) average for Spaniards of the time. The Argentine portion of Patagonia includes the provinces of Neuquén, Río Negro, Chubut and Santa Cruz, as well as the eastern portion of Tierra del Fuego archipelago. The Argentine politico-economic Patagonic Region includes the Province of La Pampa. The Chilean part of Patagonia embraces the southern part of Valdivia, Los Lagos in Lake Llanquihue, Chiloé, Puerto Montt and the Archaeological site of Monte Verde, also the fiords and islands south to the regions of Aisén and Magallanes, including the west side of Tierra del Fuego and Cape Horn. The first European to sight Chilean territory was Ferdinand Magellan, who crossed the Strait of Magellan on November 1, 1520. However, the title of discoverer of Chile is usually assigned to Diego de Almagro. Almagro was Francisco Pizarro's partner, and he received the Southern area ("Nueva Toledo"). He organized an expedition that brought him to central Chile in 1537, but he found little of value to compare with the gold and silver of the Incas in Peru. Left with the impression that the inhabitants of the area were poor, he returned to Peru, later to be garotted following defeat by Hernando Pizarro in a Civil War. After this initial excursion there was little interest from colonial authorities in further exploring modern-day Chile. However, Pedro de Valdivia, captain of the army, realizing the potential for expanding the Spanish empire southward, asked Pizarro's permission to invade and conquer the southern lands. With a couple of hundred men, he subdued the local inhabitants and founded the city of Santiago de Nueva Extremadura, now Santiago de Chile, on February 12, 1541. Although Valdivia found little gold in Chile he could see the agricultural richness of the land. He continued his explorations of the region west of the Andes and founded over a dozen towns and established the first "encomiendas". The greatest resistance to Spanish rule came from the Mapuche people, who opposed European conquest and colonization until the 1880s; this resistance is known as the Arauco War. Valdivia died at the Battle of Tucapel, defeated by Lautaro, a young Mapuche "toqui" (war chief), but the European conquest was well underway. The Spaniards never subjugated the Mapuche territories; various attempts at conquest, both by military and peaceful means, failed. The Great Uprising of 1598 swept all Spanish presence south of the Bío-Bío River except Chiloé (and Valdivia which was decades later reestablished as a fort), and the great river became the frontier line between Mapuche lands and the Spanish realm. North of that line cities grew up slowly, and Chilean lands eventually became an important source of food for the Viceroyalty of Peru. Valdivia became the first governor of the Captaincy General of Chile. In that post, he obeyed the viceroy of Peru and, through him, the King of Spain and his bureaucracy. Responsible to the governor, town councils known as "Cabildo" administered local municipalities, the most important of which was Santiago, which was the seat of a Royal Appeals Court ("Real Audiencia") from 1609 until the end of colonial rule. Chile was the least wealthy realm of the Spanish Crown for most of its colonial history. Only in the 18th century did a steady economic and demographic growth begin, an effect of the reforms by Spain's Bourbon dynasty and a more stable situation along the frontier. The drive for independence from Spain was precipitated by usurpation of the Spanish throne by Napoleon's brother Joseph Bonaparte. The Chilean War of Independence was part of the larger Spanish American independence movement, and it was far from having unanimous support among Chileans, who became divided between independentists and royalists. What started as an elitist political movement against their colonial master, finally ended as a full-fledged civil war between pro-Independence Criollos who sought political and economic independence from Spain and royalist Criollos, who supported the continued allegiance to and permanence within the Spanish Empire of the Captaincy General of Chile. The struggle for independence was a war within the upper class, although the majority of troops on both sides consisted of conscripted mestizos and Native Americans. The beginning of the Independence movement is traditionally dated as September 18, 1810 when a national junta was established to govern Chile in the name of the deposed king Ferdinand VII. Depending on what terms one uses to define the end, the movement extended until 1821 (when the Spanish were expelled from mainland Chile) or 1826 (when the last Spanish troops surrendered and Chiloé was incorporated to the Chilean republic). The independence process is normally divided into three stages: "Patria Vieja", "Reconquista", and "Patria Nueva". Chile's first experiment with self-government, the "Patria Vieja" ("old republic", 1810–1814), was led by José Miguel Carrera, an aristocrat then in his mid-twenties. The military-educated Carrera was a heavy-handed ruler who aroused widespread opposition. Another of the earliest advocates of full independence, Bernardo O'Higgins, captained a rival faction that plunged the Criollos into civil war. For him and for certain other members of the Chilean elite, the initiative for temporary self-rule quickly escalated into a campaign for permanent independence, although other Criollos remained loyal to Spain. Among those favoring independence, conservatives fought with liberals over the degree to which French revolutionary ideas would be incorporated into the movement. After several efforts, Spanish troops from Peru took advantage of the internecine strife to reconquer Chile in 1814, when they reasserted control by the Battle of Rancagua on October 12. O'Higgins, Carrera and many of the Chilean rebels escaped to Argentina. The second period was characterized by the Spanish attempts to reimpose arbitrary rule during the period known as the "Reconquista" of 1814–1817 ("Reconquest": the term echoes the "Reconquista" in which the Christian kingdoms retook Iberia from the Muslims). During this period, the harsh rule of the Spanish loyalists, who punished suspected rebels, drove more and more Chileans into the insurrectionary camp. More members of the Chilean elite were becoming convinced of the necessity of full independence, regardless of who sat on the throne of Spain. As the leader of guerrilla raids against the Spaniards, Manuel Rodríguez became a national symbol of resistance. In exile in Argentina, O'Higgins joined forces with José de San Martín. Their combined army freed Chile with a daring assault over the Andes in 1817, defeating the Spaniards at the Battle of Chacabuco on February 12 and marking the beginning of the "Patria Nueva". San Martín considered the liberation of Chile a strategic stepping-stone to the emancipation of Peru, which he saw as the key to hemispheric victory over the Spanish. Chile won its formal independence when San Martín defeated the last large Spanish force on Chilean soil at the Battle of Maipú on April 5, 1818. San Martín then led his Argentine and Chilean followers north to liberate Peru; and fighting continued in Chile's southern provinces, the bastion of the royalists, until 1826. A declaration of independence was officially issued by Chile on February 12, 1818 and formally recognized by Spain in 1840, when full diplomatic relations were established. From 1817 to 1823, Bernardo O'Higgins ruled Chile as supreme director. He won plaudits for defeating royalists and founding schools, but civil strife continued. O'Higgins alienated liberals and provincials with his authoritarianism, conservatives and the church with his anticlericalism, and landowners with his proposed reforms of the land tenure system. His attempt to devise a constitution in 1818 that would legitimize his government failed, as did his effort to generate stable funding for the new administration. O'Higgins's dictatorial behavior aroused resistance in the provinces. This growing discontent was reflected in the continuing opposition of partisans of Carrera, who was executed by the Argentine regime in Mendoza in 1821, like his two brothers were three years earlier. Although opposed by many liberals, O'Higgins angered the Roman Catholic Church with his liberal beliefs. He maintained Catholicism's status as the official state religion but tried to curb the church's political powers and to encourage religious tolerance as a means of attracting Protestant immigrants and traders. Like the church, the landed aristocracy felt threatened by O'Higgins, resenting his attempts to eliminate noble titles and, more important, to eliminate entailed estates. O'Higgins's opponents also disapproved of his diversion of Chilean resources to aid San Martín's liberation of Peru. O'Higgins insisted on supporting that campaign because he realized that Chilean independence would not be secure until the Spaniards were routed from the Andean core of the empire. However, amid mounting discontent, troops from the northern and southern provinces forced O'Higgins to resign. Embittered, O'Higgins departed for Peru, where he died in 1842. After O'Higgins went into exile in 1823, civil conflict continued, focusing mainly on the issues of anticlericalism and regionalism. Presidents and constitutions rose and fell quickly in the 1820s. The civil struggle's harmful effects on the economy, and particularly on exports, prompted conservatives to seize national control in 1830. In the minds of most members of the Chilean elite, the bloodshed and chaos of the late 1820s were attributable to the shortcomings of liberalism and federalism, which had been dominant over conservatism for most of the period. The political camp became divided by supporters of O'Higgins, Carrera, liberal Pipiolos and conservative Pelucones, being the two last the main movements that prevailed and absorbed the rest. The abolition of slavery in 1823—long before most other countries in the Americas—was considered one of the Pipiolos' few lasting achievements. One Pipiolo leader from the south, Ramón Freire, rode in and out of the presidency several times (1823–1827, 1828, 1829, 1830) but could not sustain his authority. From May 1827 to September 1831, with the exception of brief interventions by Freire, the presidency was occupied by Francisco Antonio Pinto, Freire's former vice president. In August 1828, Pinto's first year in office, Chile abandoned its short-lived federalist system for a unitary form of government, with separate legislative, executive, and judicial branches. By adopting a moderately liberal constitution in 1828, Pinto alienated both the federalists and the liberal factions. He also angered the old aristocracy by abolishing estates inherited by primogeniture (mayorazgo) and caused a public uproar with his anticlericalism. After the defeat of his liberal army at the Battle of Lircay on April 17, 1830, Freire, like O'Higgins, went into exile in Peru. Although never president, Diego Portales dominated Chilean politics from the cabinet and behind the scenes from 1830 to 1837. He installed the "autocratic republic", which centralized authority in the national government. His political program enjoyed support from merchants, large landowners, foreign capitalists, the church, and the military. Political and economic stability reinforced each other, as Portales encouraged economic growth through free trade and put government finances in order. Portales was an agnostic who said that he believed in the clergy but not in God. He realized the importance of the Roman Catholic Church as a bastion of loyalty, legitimacy, social control and stability, as had been the case in the colonial period. He repealed Liberal reforms that had threatened church privileges and properties. The "Portalian State" was institutionalized by the Chilean Constitution of 1833. One of the most durable charters ever devised in Latin America, the Portalian constitution lasted until 1925. The constitution concentrated authority in the national government, more precisely, in the hands of the president, who was elected by a tiny minority. The chief executive could serve two consecutive five-year terms and then pick a successor. Although the Congress had significant budgetary powers, it was overshadowed by the president, who appointed provincial officials. The constitution also created an independent judiciary, guaranteed inheritance of estates by primogeniture, and installed Catholicism as the state religion. In short, it established an autocratic system under a republican veneer. Portales also achieved his objectives by wielding dictatorial powers, censoring the press, and manipulating elections. For the next forty years, Chile's armed forces would be distracted from meddling in politics by skirmishes and defensive operations on the southern frontier, although some units got embroiled in domestic conflicts in 1851 and 1859. The Portalian president was General Joaquín Prieto, who served two terms (1831–1836, 1836–1841). President Prieto had four main accomplishments: implementation of the 1833 constitution, stabilization of government finances, defeat of provincial challenges to central authority, and victory over the Peru-Bolivia Confederation. During the presidencies of Prieto and his two successors, Chile modernized through the construction of ports, railroads, and telegraph lines, some built by United States entrepreneur William Wheelwright. These innovations facilitated the export-import trade as well as domestic commerce. Prieto and his adviser, Portales, feared the efforts of Bolivian general Andrés de Santa Cruz to unite with Peru against Chile. These qualms exacerbated animosities toward Peru dating from the colonial period, now intensified by disputes over customs duties and loans. Chile also wanted to become the dominant South American military and commercial power along the Pacific. Santa Cruz united Peru and Bolivia in the Peru–Bolivian Confederation in 1836 with a desire to expand control over Argentina and Chile. Portales got Congress to declare war on the Confederation. Portales was killed by traitors in 1837. The general Manuel Bulnes defeated the Confederation in the Battle of Yungay in 1839. After his success Bulnes was elected president in 1841. He served two terms (1841–1846, 1846–1851). His administration concentrated on the occupation of the territory, especially the Strait of Magellan and the Araucanía. The Venezuelan Andres Bello make in this period important intellectual advances, specially the creation of the University of Santiago. But political tensions, including a liberal rebellion, led to the Chilean Civil War of 1851. Finally the conservatives defeated the liberals. The last conservative president was Manuel Montt, who also served two terms (1851–1856, 1856–1861), but his bad administration led to the liberal rebellion in 1859. Liberals triumphed in 1861 with the election of Jose Joaquin Perez as president. The political revolt brought little social change, however, and 19th century Chilean society preserved the essence of the stratified colonial social structure, which was greatly influenced by family politics and the Roman Catholic Church. A strong presidency eventually emerged, but wealthy landowners remained powerful. Toward the end of the 19th century, the government in Santiago consolidated its position in the south by persistently suppressing the Mapuche during the Occupation of the Araucanía. In 1881, it signed the Boundary Treaty of 1881 between Chile and Argentina confirming Chilean sovereignty over the Strait of Magellan, but conceding all of oriental Patagonia, and a considerable fraction of the territory it had during colonial times. As a result of the War of the Pacific with Peru and Bolivia (1879–1883), Chile expanded its territory northward by almost one-third and acquired valuable nitrate deposits, the exploitation of which led to an era of national affluence. In the 1870s, the church influence started to diminish slightly with the passing of several laws that took some old roles of the church into the State's hands such as the registry of births and marriages. In 1886, José Manuel Balmaceda was elected president. His economic policies visibly changed the existing liberal policies. He began to violate the constitution and slowly began to establish a dictatorship. Congress decided to depose Balmaceda, who refused to step down. Jorge Montt, among others, directed an armed conflict against Balmaceda, which soon extended into the 1891 Chilean Civil War. Defeated, Balmaceda fled to Argentina's embassy, where he committed suicide. Jorge Montt became the new president. The so-called Parliamentary Republic was not a true parliamentary system, in which the chief executive is elected by the legislature. It was, however, an unusual regime in presidentialist Latin America, for Congress really did overshadow the rather ceremonial office of the president and exerted authority over the chief executive's cabinet appointees. In turn, Congress was dominated by the landed elites. This was the heyday of classic political and economic liberalism. For many decades thereafter, historians derided the Parliamentary Republic as a quarrel-prone system that merely distributed spoils and clung to its laissez-faire policy while national problems mounted. The characterization is epitomized by an observation made by President Ramón Barros Luco (1910–1915), reputedly made in reference to labor unrest: "There are only two kinds of problems: those that solve themselves and those that can't be solved." At the mercy of Congress, cabinets came and went frequently, although there was more stability and continuity in public administration than some historians have suggested. Chile also temporarily resolved its border disputes with Argentina with the Puna de Atacama Lawsuit of 1899, the Boundary treaty of 1881 between Chile and Argentina and the 1902 General Treaty of Arbitration, though not without engaging in an expensive naval arms race beforehand. Political authority ran from local electoral bosses in the provinces through the congressional and executive branches, which reciprocated with payoffs from taxes on nitrate sales. Congressmen often won election by bribing voters in this clientelistic and corrupt system. Many politicians relied on intimidated or loyal peasant voters in the countryside, even though the population was becoming increasingly urban. The lackluster presidents and ineffectual administrations of the period did little to respond to the country's dependence on volatile nitrate exports, spiraling inflation, and massive urbanization. In recent years, however, particularly when the authoritarian regime of Augusto Pinochet is taken into consideration, some scholars have reevaluated the Parliamentary Republic of 1891–1925. Without denying its shortcomings, they have lauded its democratic stability. They have also hailed its control of the armed forces, its respect for civil liberties, its expansion of suffrage and participation, and its gradual admission of new contenders, especially reformers, to the political arena. In particular, two young parties grew in importance – the Democrat Party, with roots among artisans and urban workers, and the Radical Party, representing urban middle sectors and provincial elites. By the early 20th century, both parties were winning increasing numbers of seats in Congress. The more leftist members of the Democrat Party became involved in the leadership of labor unions and broke off to launch the Socialist Workers' Party ( – POS) in 1912. The founder of the POS and its best-known leader, Luis Emilio Recabarren, also founded the Communist Party of Chile ( – PCCh) in 1922. By the 1920s, the emerging middle and working classes were powerful enough to elect a reformist president, Arturo Alessandri Palma. Alessandri appealed to those who believed the social question should be addressed, to those worried by the decline in nitrate exports during World War I, and to those weary of presidents dominated by Congress. Promising "evolution to avoid revolution", he pioneered a new campaign style of appealing directly to the masses with florid oratory and charisma. After winning a seat in the Senate representing the mining north in 1915, he earned the sobriquet "Lion of Tarapacá." As a dissident Liberal running for the presidency, Alessandri attracted support from the more reformist Radicals and Democrats and formed the so-called Liberal Alliance. He received strong backing from the middle and working classes as well as from the provincial elites. Students and intellectuals also rallied to his banner. At the same time, he reassured the landowners that social reforms would be limited to the cities. Alessandri soon discovered that his efforts to lead would be blocked by the conservative Congress. Like Balmaceda, he infuriated the legislators by going over their heads to appeal to the voters in the congressional elections of 1924. His reform legislation was finally rammed through Congress under pressure from younger military officers, who were sick of the neglect of the armed forces, political infighting, social unrest, and galloping inflation, whose program was frustrated by a conservative congress. A double military coup set off a period of great political instability that lasted until 1932. First military right-wingers opposing Alessandri seized power in September 1924, and then reformers in favor of the ousted president took charge in January 1925. The "Saber noise" (ruido de sables) incident of September 1924, provoked by discontent of young officers, mostly lieutenants from middle and working classes, lead to the establishment of the September Junta led by General Luis Altamirano and the exile of Alessandri. However, fears of a conservative restoration in progressive sectors of the army led to another coup in January, which ended with the establishment of the January Junta as interim government while waiting for Alessandri's return. The latter group was led by two colonels, Carlos Ibáñez del Campo and Marmaduke Grove. They returned Alessandri to the presidency that March and enacted his promised reforms by decree. The latter re-assumed power in March, and a new Constitution encapsulating his proposed reforms was ratified in a plebiscite in September 1925. The new constitution gave increased powers to the presidency. Alessandri broke with the classical liberalism's policies of "laissez-faire" by creating a Central Bank and imposing a revenue tax. However, social discontents were also crushed, leading to the Marusia massacre in March 1925 followed by the La Coruña massacre. The longest lasting of the ten governments between 1924 and 1932 was that of General Carlos Ibáñez, who briefly held power in 1925 and then again between 1927 and 1931 in what was a "de facto" dictatorship. When constitutional rule was restored in 1932, a strong middle-class party, the Radicals, emerged. It became the key force in coalition governments for the next 20 years. The Seguro Obrero Massacre took place on September 5, 1938, in the midst of a heated three-way election campaign between the ultraconservative Gustavo Ross Santa María, the radical Popular Front's Pedro Aguirre Cerda, and the newly formed Popular Alliance candidate, Carlos Ibáñez del Campo. The National Socialist Movement of Chile supported Ibáñez's candidacy, which had been announced on September 4. In order to preempt Ross's victory, the National Socialists mounted a coup d'état that was intended to take down the rightwing government of Arturo Alessandri Palma and place Ibáñez in power. During the period of Radical Party dominance (1932–1952), the state increased its role in the economy. In 1952, voters returned Ibáñez to office for another 6 years. Jorge Alessandri succeeded Ibáñez in 1958. The 1964 presidential election of Christian Democrat Eduardo Frei Montalva by an absolute majority initiated a period of major reform. Under the slogan "Revolution in Liberty", the Frei administration embarked on far-reaching social and economic programs, particularly in education, housing, and agrarian reform, including rural unionization of agricultural workers. By 1967, however, Frei encountered increasing opposition from leftists, who charged that his reforms were inadequate, and from conservatives, who found them excessive. At the end of his term, Frei had accomplished many noteworthy objectives, but he had not fully achieved his party's ambitious goals. In the 1970 presidential election, Senator Salvador Allende Gossens won a plurality of votes in a three-way contest. He was a Marxist physician and member of Chile's Socialist Party, who headed the "Popular Unity" (UP or "Unidad Popular") coalition of the Socialist, Communist, Radical, and Social-Democratic Parties, along with dissident Christian Democrats, the Popular Unitary Action Movement (MAPU), and the Independent Popular Action. Allende had two main competitors in the election — Radomiro Tomic, representing the incumbent Christian Democratic party, who ran a left-wing campaign with much the same theme as Allende's, and the right-wing former president Jorge Alessandri. In the end, Allende received a plurality of the votes cast, getting 36% of the vote against Alessandri's 35% and Tomic's 28%. Despite pressure from the government of the United States, the Chilean Congress, keeping with tradition, conducted a runoff vote between the leading candidates, Allende and former president Jorge Alessandri. This procedure had previously been a near-formality, yet became quite fraught in 1970. After assurances of legality on Allende's part, the murder of the Army Commander-in-Chief, General René Schneider and Frei's refusal to form an alliance with Alessandri to oppose Allende – on the grounds that the Christian Democrats were a workers' party and could not make common cause with the oligarchs – Allende was chosen by a vote of 153 to 35. The Popular Unity platform included the nationalization of U.S. interests in Chile's major copper mines, the advancement of workers' rights, deepening of the Chilean land reform, reorganization of the national economy into socialized, mixed, and private sectors, a foreign policy of "international solidarity" and national independence and a new institutional order (the "people's state" or "poder popular"), including the institution of a unicameral congress. Immediately after the election, the United States expressed its disapproval and raised a number of economic sanctions against Chile. In addition, the CIA's website reports that the agency aided three different Chilean opposition groups during that time period and "sought to instigate a coup to prevent Allende from taking office". The action plans to prevent Allende from coming to power were known as Track I and Track II. In the first year of Allende's term, the short-term economic results of Economics Minister Pedro Vuskovic's expansive monetary policy were unambiguously favorable: 12% industrial growth and an 8.6% increase in GDP, accompanied by major declines in inflation (down from 34.9% to 22.1%) and unemployment (down to 3.8%). Allende adopted measures including price freezes, wage increases, and tax reforms, which had the effect of increasing consumer spending and redistributing income downward. Joint public-private public works projects helped reduce unemployment. Much of the banking sector was nationalized. Many enterprises within the copper, coal, iron, nitrate, and steel industries were expropriated, nationalized, or subjected to state intervention. Industrial output increased sharply and unemployment fell during the administration's first year. However, these results were not sustainable and in 1972 the Chilean "escudo" had runaway inflation of 140%. An economic depression that had begun in 1967 peaked in 1972, exacerbated by capital flight, plummeting private investment, and withdrawal of bank deposits in response to Allende's socialist program. Production fell and unemployment rose. The combination of inflation and government-mandated price-fixing led to the rise of black markets in rice, beans, sugar, and flour, and a "disappearance" of such basic commodities from supermarket shelves. Recognizing that U.S. intelligence forces were trying to destabilize his presidency through a variety of methods, the KGB offered financial assistance to the first democratically elected Marxist president. However, the reason behind the U.S. covert actions against Allende concerned not the spread of Marxism but fear over losing control of its investments. "By 1968, 20 percent of total U.S. foreign investment was tied up in Latin America...Mining companies had invested $1 billion over the previous fifty years in Chile's copper mining industry – the largest in the world – but they had sent $7.2 billion home." Part of the CIA's program involved a propaganda campaign that portrayed Allende as a would-be Soviet dictator. In fact, however, "the U.S.'s own intelligence reports showed that Allende posed no threat to democracy." Nevertheless, the Richard Nixon administration organized and inserted secret operatives in Chile, in order to quickly destabilize Allende's government. In addition, Nixon gave instructions to make the Chilean economy scream, and international financial pressure restricted economic credit to Chile. Simultaneously, the CIA funded opposition media, politicians, and organizations, helping to accelerate a campaign of domestic destabilization. By 1972, the economic progress of Allende's first year had been reversed, and the economy was in crisis. Political polarization increased, and large mobilizations of both pro- and anti-government groups became frequent, often leading to clashes. By 1973, Chilean society had grown highly polarized, between strong opponents and equally strong supporters of Salvador Allende and his government. Military actions and movements, separate from the civilian authority, began to manifest in the countryside. The Tanquetazo was a failed military coup d'état attempted against Allende in June 1973. In its "", on August 22, 1973, the Chamber of Deputies of Chile asserted that Chilean democracy had broken down and called for "redirecting government activity", to restore constitutional rule. Less than a month later, on September 11, 1973, the Chilean military deposed Allende, who shot himself in the head to avoid capture as the Presidential Palace was surrounded and bombed. Subsequently, rather than restore governmental authority to the civilian legislature, Augusto Pinochet exploited his role as Commander of the Army to seize total power and to establish himself at the head of a junta. CIA involvement in the coup is documented. As early as the Church Committee Report (1975), publicly available documents have indicated that the CIA attempted to prevent Allende from taking office after he was elected in 1970; the CIA itself released documents in 2000 acknowledging this and that Pinochet was one of their favored alternatives to take power. According to the Vasili Mitrokhin and Christopher Andrew, the KGB and the Cuban Intelligence Directorate launched a campaign known as Operation TOUCAN. For instance, in 1976, the New York Times published 66 articles on alleged human rights abuses in Chile and only 4 on Cambodia, where the communist Khmer Rouge killed some 1.5 million people of 7.5 million people in the country. By early 1973, inflation had risen 800% under Allende's presidency."" The crippled economy was further battered by prolonged and sometimes simultaneous strikes by physicians, teachers, students, truck owners, copper workers, and the small business class. A military coup overthrew Allende on September 11, 1973. As the armed forces bombarded the presidential palace (Palacio de La Moneda), Allende committed suicide. A military government, led by General Augusto Pinochet Ugarte, took over control of the country. The first years of the regime were marked by human rights violations. The junta jailed, tortured, and executed thousands of Chileans. In October 1973, at least 72 people were murdered by the Caravan of Death. At least a thousand people were executed during the first six months of Pinochet in office, and at least two thousand more were killed during the next sixteen years, as reported by the Rettig Report. At least 29,000 were imprisoned and tortured. According to the Latin American Institute on Mental Health and Human Rights (ILAS), "situations of extreme trauma" affected about 200,000 persons.; this figure includes individuals killed, tortured or exiled, and their immediate families. About 30,000 left the country. The four-man junta headed by General Augusto Pinochet abolished civil liberties, dissolved the national congress, banned union activities, prohibited strikes and collective bargaining, and erased the Allende administration's agrarian and economic reforms. The junta embarked on a radical program of liberalization, deregulation and privatization, slashing tariffs as well as government welfare programs and deficits. Economic reforms were drafted by a group of technocrats who became known as the Chicago Boys because many of them had been trained or influenced by University of Chicago professors. Under these new policies, the rate of inflation dropped: A new constitution was approved by plebiscite characterized by the absence of registration lists, on September 11, 1980, and General Pinochet became president of the republic for an 8-year term. In 1982–1983 Chile witnessed a severe economic crisis with a surge in unemployment and a meltdown of the financial sector. 16 out of 50 financial institutions faced bankruptcy. In 1982 the two biggest banks were nationalized to prevent an even worse credit crunch. In 1983 another five banks were nationalized and two banks had to be put under government supervision. The central bank took over foreign debts. Critics ridiculed the economic policy of the Chicago Boys as "Chicago way to socialism“. After the economic crisis, Hernan Buchi became Minister of Finance from 1985 to 1989, introducing a more pragmatic economic policy. He allowed the peso to float and reinstated restrictions on the movement of capital in and out of the country. He introduced Bank regulations, simplified and reduced the corporate tax. Chile went ahead with privatizations, including public utilities plus the re-privatization of companies that had returned to the government during the 1982–1983 crisis. From 1984 to 1990, Chile's gross domestic product grew by an annual average of 5.9%, the fastest on the continent. Chile developed a good export economy, including the export of fruits and vegetables to the northern hemisphere when they were out of season, and commanded high prices. The military junta began to change during the late 1970s. Due to problems with Pinochet, Leigh was expelled from the junta in 1978 and replaced by General Fernando Matthei. In the late 1980s, the government gradually permitted greater freedom of assembly, speech, and association, to include trade union and political activity. Due to the Caso Degollados ("slit throats case"), in which three Communist party members were assassinated, César Mendoza, member of the junta since 1973 and representants of the carabineros, resigned in 1985 and was replaced by Rodolfo Stange. The next year, Carmen Gloria Quintana was burnt alive in what became known as the Caso Quemado ("Burnt Alive case"). Chile's constitution established that in 1988 there would be another plebiscite in which the voters would accept or reject a single candidate proposed by the Military Junta. Pinochet was, as expected, the candidate proposed, but was denied a second 8-year term by 54.5% of the vote. Chileans elected a new president and the majority of members of a two-chamber congress on December 14, 1989. Christian Democrat Patricio Aylwin, the candidate of a coalition of 17 political parties called the "Concertación", received an absolute majority of votes (55%). President Aylwin served from 1990 to 1994, in what was considered a transition period. In February 1991 Aylwin created the National Commission for Truth and Reconciliation, which released in February 1991 the Rettig Report on human rights violations committed during the military rule. This report counted 2,279 cases of "disappearances" which could be proved and registered. Of course, the very nature of "disappearances" made such investigations very difficult. The same problem arose, several years later, with the Valech Report, released in 2004 and which counted almost 30,000 victims of torture, among testimonies from 35,000 persons. In December 1993, Christian Democrat Eduardo Frei Ruiz-Tagle, the son of previous president Eduardo Frei Montalva, led the Concertación coalition to victory with an absolute majority of votes (58%). Frei Ruiz-Tagle was succeeded in 2000 by Socialist Ricardo Lagos, who won the presidency in an unprecedented runoff election against Joaquín Lavín of the rightist Alliance for Chile, by a very tight score of less than 200,000 votes (51,32%). In 1998, Pinochet traveled to London for back surgery. But under orders of Spanish judge Baltasar Garzón, he was arrested there, attracting worldwide attention, not only because of the history of Chile and South America, but also because this was one of the first arrests of a former president based on the universal jurisdiction principle. Pinochet tried to defend himself by referring to the State Immunity Act of 1978, an argument rejected by the British justice. However, UK Home Secretary Jack Straw took the responsibility to release him on medical grounds, and refused to extradite him to Spain. Thereafter, Pinochet returned to Chile in March 2000. Upon descending the plane on his wheelchair, he stood up and saluted the cheering crowd of supporters, including an army band playing his favorite military march tunes, which was awaiting him at the airport in Santiago. President Ricardo Lagos later commented that the retired general's televised arrival had damaged the image of Chile, while thousands demonstrated against him. The "Concertación" coalition has continued to dominate Chilean politics for last two decades. In January 2006 Chileans elected their first woman president, Michelle Bachelet, of the Socialist Party. She was sworn in on March 11, 2006, extending the "Concertación" coalition governance for another four years. In 2002 Chile signed an association agreement with the European Union (comprising a free trade agreement and political and cultural agreements), in 2003, an extensive free trade agreement with the United States, and in 2004 with South Korea, expecting a boom in import and export of local produce and becoming a regional trade-hub. Continuing the coalition's free trade strategy, in August 2006 President Bachelet promulgated a free trade agreement with the People's Republic of China (signed under the previous administration of Ricardo Lagos), the first Chinese free trade agreement with a Latin American nation; similar deals with Japan and India were promulgated in August 2007. In October 2006, Bachelet promulgated a multilateral trade deal with New Zealand, Singapore and Brunei, the Trans-Pacific Strategic Economic Partnership (P4), also signed under Lagos' presidency. Regionally, she has signed bilateral free trade agreements with Panama, Peru and Colombia. After 20 years, Chile went in a new direction with the win of center-right Sebastián Piñera, in the Chilean presidential election of 2009–2010. On 27 February 2010, Chile was struck by an 8.8 MW earthquake, the fifth largest ever recorded at the time. More than 500 people died (most from the ensuing tsunami) and over a million people lost their homes. The earthquake was also followed by multiple aftershocks. Initial damage estimates were in the range of US$15–30 billion, around 10 to 15 percent of Chile's real gross domestic product. Chile achieved global recognition for the successful rescue of 33 trapped miners in 2010. On 5 August 2010, the access tunnel collapsed at the San José copper and gold mine in the Atacama Desert near Copiapó in northern Chile, trapping 33 men below ground. A rescue effort organized by the Chilean government located the miners 17 days later. All 33 men were brought to the surface two months later on 13 October 2010 over a period of almost 24 hours, an effort that was carried on live television around the world. Despite good macroeconomic indicators, there was increased social dissatisfaction, focused on demands for better and fairer education, culminating in massive protests demanding more democratic and equitable institutions. Approval of Piñera's administration fell irrevocably. In 2013, Bachelet, a Social Democrat, was elected again as President, seeking to make the structural changes claimed in recent years by the society relative to education reform, tributary reform, same sex civil union, and definitely end the Binomial System, looking to further equality and the end of what remains of the dictatorship. In 2015 a series of corruption scandals (most notably Penta case and Caval case) became public, threatening the credibility of the political and business class. On 17 December 2017, Sebastián Piñera was elected president of Chile for a second term. He received 36% of the votes, the highest percentage among all 8 candidates. In the second round, Piñera faced Alejandro Guillier, a television news anchor who represented Bachelet's New Majority (Nueva Mayoría) coalition. Piñera won the elections with 54% of the votes. In October 2019 there were violent protests about costs of living and inequality, resulting in Piñera declaring a state of emergency. On 15 November, most of the political parties represented in the National Congress signed an agreement to call a national referendum on April 2020 regarding the creation of a new Constitution. But the COVID-19 pandemic postponed the date of the elections, while Chile was one of the hardest hit nations in the Americas as of May 2020. General:
https://en.wikipedia.org/wiki?curid=5490
Politics of Chile Chile's government is a representative democratic republic, whereby the President of Chile is both head of state and head of government, and of a formal multi-party system. Executive power is exercised by the president and his or her cabinet. Legislative power is vested in both the government and the two chambers of the National Congress. The judiciary is independent of the executive and the legislature of Chile. The Constitution of Chile was approved in a national plebiscite in September 1980, under the military dictatorship of Augusto Pinochet. It entered into force in March 1981. After Pinochet left power in 1988, saying this country was ready to keep going along with a plebiscite, the Constitution was amended to ease provisions for future amendments to the Constitution. In September 2005, President Ricardo Lagos signed into law several constitutional amendments passed by Congress. These include eliminating the positions of appointed senators and senators for life, granting the President authority to remove the commanders-in-chief of the armed forces, and reducing the presidential term from six to four years while also disabling immediate re-election. The autocratic and conservative republic (1831-1861) was replaced by the liberal republic (1861-1891) during which some political conquests were made, such as proportional representation (1871) and the abolition of the condition of ownership to have the right to vote (1885) When the era of the parliamentary republic began in 1891, the struggle between liberals (pipiolos) and conservatives (pelucones) had already evolved due to the emergence of a multi-party system. In the 1880s, the Liberals split into two factions: the moderates, who did not want to impose secularism too quickly and were willing to compromise with the Conservatives, and the radical Liberals, who joined the Radical Party founded in 1862 or the new Democratic Party with more progressive, if not socialist, ideas. European and particularly British companies having appropriated a large part of the country's economy (saltpeter, bank, railway, trade), President José Balmaceda (1886-1891), leader of moderate liberals, decided to react by directing his policy in two directions: the nationalisation of saltpeter mines and the intervention of the State in economic matters. Already facing the conservative aristocracy, he alienated the bankers. He was dismissed by a vote of Parliament and pressure from part of the army. He committed suicide by firearm at the end of the civil war that his supporters lost. A new parliamentary regime emerged from the civil war; it was the government of Fronda aristocrática. From 1906 onwards, the Radical Party demanded social reforms and the establishment of a democratic regime. That same year, the leader of the Federation of Workers, Luis Emilio Recabarren, was elected to the House but his election was canceled by the House. In 1912 he founded the Socialist Workers Party. Despite the country's good economic performance, life remains particularly hard for a large part of the population (12 or 14-hour working days for workers, very low wages, illiteracy of more than 50% in the years 1900–1910, etc.). Trade unionism was organized and fought; strikes and workers' demonstrations multiplied, sometimes very harshly repressed: general strike in Santiago (1905), railways and mines in Antofagasta (1906), a demonstration in Iquique (1907). From 1911 to 1920, there were 293 strikes. Some repressions kill hundreds of people. The workers' movement was organized in the 1910s with the creation of the Chilean Regional Workers' Federation in 1913 and the Chilean branch of the Industrial Workers of the World in 1919. In 1920, the economic crisis worsened the standard of living of the middle classes, which were politically closer to the working classes. This new situation led to the election of Arturo Alessandri Palma. During his first term in office, he pursued a progressive policy: labor law, the establishment of the tax on property income, the establishment of the Central Bank, creation of social security funds, etc. However, it must constantly deal with the Senate, always under Conservative control, which systematically tries to block its reforms. Shortly before his withdrawal from power, he drew up a new Constitution that was considered to be the advent of true democracy in Chile. This Constitution enshrines the separation of Church and State and religious freedom, declares compulsory primary education, restores presidentialism but by electing the president by universal suffrage, and above all proclaims that property must be regulated in such a way as to ensure its social function. The bicameral National Congress "(Congreso Nacional)" consists of the Senate "(Senado)" and the Chamber of Deputies "(Cámara de Diputados)." Chile's congressional elections replaces the binominal electoral system applicable to the parliamentary elections, by one of an inclusive proportional nature and strengthens the representativeness of the National Congress (D'Hondt System). Elections are very labor-intensive but efficient, and vote counting normally takes place the evening of the election day. One voting table, with a ballot-box each, is set up for at-most 200 names in the voting registry. Each table is manned by five people ("vocales de mesa") from the same registry. Vocales have the duty to work as such during a cycle of elections, and can be penalized legally if they do not show up. A registered citizen can only vote after his identity has been verified at the table corresponding to his registry. Ballots are manually counted by the five vocales, after the table has closed, at least eight hours after opening, and the counting witnessed by representatives of all the parties who choose to have observers. The Senate is made up of 50 members elected from regions or subregions. Senators serve approximately eight-year terms. The Chamber of Deputies has 155 members, who are elected by popular vote to serve four-year terms. The last congressional elections were held on November 19, 2017. The next congressional elections are scheduled for December 2021. The current Senate composition is as follows: 36 seats are held by the Progressive Convergence coalition: seven Socialists (PS), seven Party for Democracy (PPD) and one Social Democrat Radical Party (PRSD); 6 by the Democratic Convergence coalition: Six Christian Democrats (PDC); 17 by the Chile let's go coalition: nine Independent Democratic Union (UDI) and eight National Renewal (RN); and two Politic Evolution (EVOPOLI); 1 by Broad Front coalition: One Democratic Revolution (RD); 1 by Throughout Chile coalition: One Country (PAIS); and 1 independent. The current lower house—the Chamber of Deputies —contains: 65 members of Progressive Convergence coalition: nineteen Socialists (PS), seven Party for Democracy (PPD), eight Social Democrat Radical Party (PRSD) and three independent pro Progressive Convergence; 8 by the Chilean Communist Party; 14 by the Democratic Convergence coalition: fourteen Christian Democrats (PDC); 71 by the Chile let's go coalition: twenty seven Independent Democratic Union (UDI), thirty-three National Renewal (RN); four Politic Evolution (EVOPOLI), and seven independent pro Chile let's go; 4 from the Regional Green Socialist Federation (FRVS); 20 by Broad Front coalition: eight Democratic Revolution (RD), two Liberal Party (PL), 1 Humanist Party (PH), one Green Environmentalist Party (PEV), one Power (PODER), and five independent pro Broad Front; 1 by Throughout Chile coalition: Progressive Party (PRO); and 2 independent. Since 1987, the Congress operates in the port city of Valparaíso, about 110 kilometers (~70 mi.) northwest of the capital, Santiago. However some commissions are allowed to meet in other places, especially Santiago. Congressional members have tried repeatedly to relocate the Congress back to Santiago, where it operated until the Chilean Coup of 1973, but have not been successful. The last attempt was in 2000, when the project was rejected by the Constitutional Court, because it allocated funds from the national budget, which, under the Chilean Constitution, is a privilege of the President. Chile's legal system is civil law based. It is primarily based on the Civil code of 1855, derived from Spanish law and subsequent codes influenced by European law of the last half of the 19th Century. It does not accept compulsory ICJ jurisdiction. From the year 2000 onward, Chile completely overhauled its criminal justice system; a new, US-style adversarial system has been gradually implemented throughout the country with the final stage of implementation in the Santiago metropolitan region completed on June 9, 2001 Pressure groups according to the CIA World Factbook: Chile or Chilean organizations participate in the following international organizations:
https://en.wikipedia.org/wiki?curid=5493
Economy of Chile The economy of Chile is a high-income economy as ranked by the World Bank, and is considered one of South America's most prosperous nations, leading Latin American nations in competitiveness, income per capita, globalization, economic freedom, and low perception of corruption. Although Chile has high economic inequality, as measured by the Gini index, it is close to the regional mean. In 2006, Chile became the country with the highest nominal GDP per capita in Latin America. In May 2010 Chile became the first South American country to join the OECD. Tax revenues, all together 20.2% of GDP in 2013, were the second lowest among the 34 OECD countries, and the lowest in 2010. Chile has an inequality-adjusted human development index of 0.696, compared to 0.703, 0.714 and 0.575 for neighboring Uruguay, Argentina and Brazil, respectively. In 2017, only 0.7% of the population lived on less than US$1.90 a day. The Global Competitiveness Report for 2009–2010 ranked Chile as being the 30th most competitive country in the world and the first in Latin America, well above Brazil (56th), Mexico (60th), and Argentina which ranks 85th; it has since fallen out of the top 30. The ease of doing business index, created by the World Bank, listed Chile as 34th in the world as of 2014, 41st for 2015, and 48th as of 2016. The privatized national pension system (AFP) has an estimated total domestic savings rate of approximately 21% of GDP. After Spanish arrival in the 15th century Chilean economy came to revolve around autarchy estates called fundos and around the army that was engaged in the Arauco War. During early colonial times there were gold exports to Perú from placer deposits which soon depleted. Trade restrictions and monopolies established by the Spanish crown are credited for having held back economic development for much of the colonial times. As effect of these restrictions the country incorporated very few new crops and animal breeds after initial conquest. Other sectors that were held back by restrictions were the wine and mining industries. The Bourbon reforms in the 18th century eased many monopolies and trade restrictions. In the 1830s Chile consolidated under the ideas of Diego Portales as a stable state open to foreign trade. Foreign investment in Chile grew over the 19th century. After the War of the Pacific the Chilean treasury grew by 900%. The League of Nations labeled Chile the country hardest hit by the Great Depression because 80% of government revenue came from exports of copper and nitrates, which were in low demand. After the Great Depression Chilean economic policies changed toward import substitution industrialization and the Production Development Corporation was established. Under the influence of the Chicago Boys the Pinochet regime made of Chile a leading country in establishing neoliberal policies. These policies allowed large corporations to consolidate their power over the Chilean economy, leading to long-term economic growth. The crisis of 1982 caused the appointment of Hernán Büchi as minister of finance and a sharp revision of economic policy. Despite a general selling of state property and contrary to neoliberal prescriptions, the regime retained the lucrative state owned mining company Codelco which stands for about 30% of government income. According to the CIA World Factbook, during the early 1990s, Chile's reputation as a role model for economic reform was strengthened when the democratic government of Patricio Aylwin, who took over from the military in 1990, deepened the economic reform initiated by the military government. The Aylwin government departed significantly from the neoliberal doctrine of the Chicago boys, as evidenced by higher government expenditure on social programs to tackle poverty and poor quality housing. Growth in real GDP averaged 8% from 1991 to 1997, but fell to half that level in 1998 because of tight monetary policies (implemented to keep the current account deficit in check) and lower exports due to the Asian financial crisis. Chile's economy has since recovered and has seen growth rates of 5–7% over the past several years. After a decade of impressive growth rates, Chile began to experience a moderate economic downturn in 1999, brought on by unfavorable global economic conditions related to the Asian financial crisis, which began in 1997. The economy remained sluggish until 2003, when it began to show clear signs of recovery, achieving 4.0% real GDP growth. The Chilean economy finished 2004 with growth of 6.0%. Real GDP growth reached 5.7% in 2005 before falling back to 4.0% in 2006. GDP expanded by 5.1% in 2007. In 2019, fewer than half of Chilean women are in paid work, and 31% have no contract or social or health insurance. Chile has a 45-hour working week, 70% of employees earn less than $825 a month. During 2012, the largest sectors by GDP were mining (mainly copper), business services, personal services, manufacturing and wholesale and retail trade. Mining also represented 59.5% of exports in the period, while the manufacturing sector accounted for 34% of exports, concentrated mainly in food products, chemicals and pulp, paper and others. Agriculture and allied sectors like forestry, logging and fishing accounts only for 4.9% of the GDP as of 2007 and employed 13.6% of the country's labor force. Some major agriculture products of Chile includes grapes, apples, pears, onions, wheat, corn, oats, peaches, garlic, asparagus, beans, beef, poultry, wool, fish and timber. Chile's position in the Southern Hemisphere leads to an agricultural season cycle opposite to those of the principal consumer markets, primarily located in the Northern Hemisphere. Chile's extreme north–south orientation produces seven different macro-regions distinguished by climate and geographical features, which allows the country itself to stagger harvests and results in extended harvesting seasons. However, the mountainous landscape of Chile limits the extent and intensity of agriculture so that arable land corresponds only to 2.62% of the total territory. Through Chile's trade agreements, its agricultural products have gained access to a market controlling 77% of the world's GDP and by approximately 2012, 74% of Chilean agribusiness exports will be duty-free. Chile's principal growing region and agricultural heartland is the Central Valley delimited by the Chilean Coast Range in the west, the Andes in the east Aconcagua River by the north and Bío-Bío River by the south. In the northern half of Chile cultivation is highly dependent on irrigation. South of the Central Valley cultivation is gradually replaced by aquaculture, silviculture, sheep and cattle farming. Chile is the second largest producer of salmon in the world. As of August 2007, Chile's share of worldwide salmon industry sales was 38.2%, rising from just 10% in 1990. The average growth rate of the industry for the 20 years between 1984 and 2004 was 42% per year. The presence of large foreign firms in the salmon industry has brought what probably most contributes to Chile's burgeoning salmon production, technology. Technology transfer has allowed Chile to build its global competitiveness and innovation and has led to the expansion of production as well as to an increase in average firm size in the industry. In November 2018, the Chinese company Joyvio Group (Legend Holdings) bought the Chilean salmon producer Australis Seafoods for $880 million, thus gaining control over 30% of all Chilean salmon exports. The Chilean forestry industry grew to comprise 13% of the country's total exports in 2005, making it one of the largest export sectors for Chile. Radiata Pine and Eucalyptus comprise the vast majority of Chile's forestry exports. Within the forestry sector, the largest contributor to total production is pulp, followed by wood-based panels and lumber. Due to popular and increasing demands for Chile's forestry products, the government is currently focusing on increasing the already vast acreage of Chile's Pine and Eucalyptus plantations as well as opening new industrial plants. Chile's unique geography and climate make it ideal for winegrowing and the country has made the top ten list of wine producers many times in the last few decades. The popularity of Chilean wine has been attributed not just to the quantity produced but also to increasing levels of quality. The combination of quantity and quality allows Chile to export excellent wines at reasonable prices to the international market. The mining sector in Chile is one of the pillars of Chilean economy. The Chilean government strongly supports foreign investment in the sector and has modified its mining industry laws and regulations to create a favourable investing environment for foreigners. Thanks to a large amount of copper resources, progressive legislation and a healthy investment environment, Chile has become the copper mining capital of the world, producing over 1/3 of the global copper output. The service sector in Chile has grown fast and consistently in recent decades, reinforced by the rapid development of communication and information technology, access to education and an increase in specialist skills and knowledge among the workforce. Chilean foreign policy has recognized the importance of the tertiary sector or service sector to the economy, boosting its international liberalization and leading to the signing of several free trade area agreements. Chilean service exportation consists mainly of maritime and aeronautical services, tourism, retail (department stores, supermarkets, and shopping centres), engineering and construction services, informatics, health and education. Chile ranked first among Latin American countries (and No. 32 worldwide) in Adecco's 2019 Global Talent Competitiveness Index (GTCI). Chile's financial sector has grown quickly in recent years, with a banking reform law approved in 1997 that broadened the scope of permissible foreign activity for Chilean banks. The Chilean Government implemented a further liberalization of capital markets in 2001, and there is further pending legislation proposing further liberalization. Over the last ten years, people who live in Chile have enjoyed the introduction of new financial tools such as home equity loans, currency futures and options, factoring, leasing, and debit cards. The introduction of these new products has also been accompanied by an increased use of traditional instruments such as loans and credit cards. Chile's private pension system, with assets worth roughly $70 billion at the end of 2006, has been an important source of investment capital for the capital market. However, by 2009, it has been reported that $21 billion had been lost from the pension system to the global financial crisis. Tourism in Chile has experienced sustained growth over the last decades. Chile received about 2.25 million foreign visitors in 2006, up to 2.50 million in 2007 The percentages of foreign tourists arrivals by land, air and sea were, respectively, 55.3%, 40.5% and 4.2% for that year. The two main gateways for international tourists visiting Chile are Comodoro Arturo Merino Benítez International Airport and Paso Los Libertadores. Chile has a great diversity of natural landscapes, from the Mars-like landscapes of the hyperarid Atacama Desert to the glacier-fed fjords of the Chilean Patagonia, passing by the winelands backdropped by the Andes of the Central Valley and the old-growth forests of the Lakes District. Easter Island and Juan Fernández Archipelago, including Robinson Crusoe Island, are also major attractions. Many of the most visited attractions in Chile are protected areas. The extensive Chilean protected areas system includes 32 protected parks, 48 natural reserves and 15 natural monuments. According to the CIA World Factbook, Chile's "sound economic policies", maintained consistently since the 1980s, "have contributed to steady economic growth in Chile and have more than halved poverty rates." The 1973–90 military government sold many state-owned companies, and the three democratic governments since 1990 have implemented export promotion policies and continued privatization, though at a slower pace. The government's role in the economy is mostly limited to regulation, although the state continues to operate copper giant CODELCO and a few other enterprises (there is one state-run bank). Under the compulsory private pension system, most formal sector employees pay 10% of their salaries into privately managed funds. As of 2006, Chile invested 0.6% of its annual GDP in research and development (R&D). Even then, two-thirds of that was government spending. Beyond its general economic and political stability, the government has also encouraged the use of Chile as an "investment platform" for multinational corporations planning to operate in the region. Chile's approach to foreign direct investment is codified in the country's Foreign Investment Law, which gives foreign investors the same treatment as Chileans. Registration is reported to be simple and transparent, and foreign investors are guaranteed access to the official foreign exchange market to repatriate their profits and capital. Faced with the financial crisis of 2007–2008, the government announced a $4 billion economic stimulus plan to spur employment and growth, and despite the global financial crisis, aimed for an expansion of between 2 percent and 3 percent of GDP for 2009. Nonetheless, economic analysts disagreed with government estimates and predicted economic growth at a median of 1.5 percent. According to the CIA World FactBook, the GDP contracted an estimated −1.7% in 2009. The Chilean Government has formed a Council on Innovation and Competition, which is tasked with identifying new sectors and industries to promote. It is hoped that this, combined with some tax reforms to encourage domestic and foreign investment in research and development, will bring in additional FDI to new parts of the economy. According to The Heritage Foundation's "Index of Economic Freedom" in 2012, Chile has the strongest private property rights in Latin America, scoring 90 on a scale of 100. Chile's AA- S&P credit rating is the highest in Latin America, while Fitch Ratings places the country one step below, in A+. There are three main ways for Chilean firms to raise funds abroad: bank loans, issuance of bonds, and the selling of stocks on U.S. markets through American Depository Receipts (ADRs). Nearly all of the funds raised through these means go to finance domestic Chilean investment. In 2006, the Government of Chile ran a surplus of $11.3 billion, equal to almost 8% of GDP. The Government of Chile continues to pay down its foreign debt, with public debt only 3.9% of GDP at the end of 2006. One of Chile's fiscal policy central features has been its counter-cyclical nature. This has been facilitated by the voluntary application since 2001 of a structural balance policy based on the commitment to an announced goal of a medium-term structural balance as a percentage of GDP. The structural balance nets out the effect of the economic cycle (including copper price volatility) on fiscal revenues and constrains expenditures to a correspondingly consistent level. In practice, this means that expenditures rise when activity is low and decrease in booms The target was of 1% of GDP between 2001 and 2007, it was reduced to 0.5% in 2008 and then to 0% in 2009 in the wake of the global financial crisis. In 2005, key elements of this voluntary policy were incorporated into legislation through the Fiscal Responsibility Law (Law 20,128). The Fiscal Responsibility Law also allowed for the creation of two sovereign wealth funds: the Pension Reserve Fund (PRF), to face increased expected old-age benefits liabilities, and the Economic and Social Stabilization Fund (ESSF), to stabilize fiscal spending by providing funds to finance fiscal deficits and debt amortization. By the end of 2012, they had respective market values of US$5.883 million and US$14.998 million. The main taxes in Chile in terms of revenue collection are the value added tax (45.8% of total revenues in 2012) and the income tax (41.8% of total revenues in 2012). The value added tax is levied on sales of goods and services (including imports) at a rate of 19%, with a few exemptions. The income tax revenue comprises different taxes. While there is a corporate income tax of 20% over profits from companies (called First Category Tax), the system is ultimately designed to tax individuals. Therefore, corporate income taxes paid constitute a credit towards two personal income taxes: the Global Complementary Tax (in the case of residents) or the Additional Tax (in the case of non-residents). The Global Complementary Tax is payable by those that have different sources of income, while those receiving income solely from dependent work are subject to the Second Category Tax. Both taxes are equally progressive in statutory terms, with a top marginal rate of 40%. Income arising from corporate activity under the Global Complementary Tax only becomes payable when effectively distributed to the individual. There are also special sales taxes on alcohol and luxury goods, as well as specific taxes on tobacco and fuel. Other taxes include the inheritance tax and custom duties. In 2012, general government expenditure reached 21.5% of GDP, while revenues were equivalent to 22% of GDP. Gross financial debt amounted to 12.2% of GDP, while in net terms it was −6.9% of GDP, both well below OECD averages. Chile's monetary authority is the Central Bank of Chile (CBoC). The CBoC pursues an inflation target of 3%, with a tolerance range of 1% (below or above). Inflation has followed a relatively stable trajectory since the year 2000, remaining under 10%, despite the temporary surge of some inflationary pressures in the year 2008. The Chilean peso's rapid appreciation against the U.S. dollar in recent years has helped dampen inflation. Most wage settlements and loans are indexed, reducing inflation's volatility. The CBoC is granted autonomous status by Chile's National Constitution, providing credibility and stability beyond the political cycle. According to the Basic Constitutional Act of the Central Bank of Chile (Law 18,840), its main objectives are to safeguard "the stability of the currency and the normal functioning of internal and external payments". To meet these objectives, the CBoC is enabled to use monetary and foreign exchange policy instruments, along with some discretion on financial regulation. In practice, the CBoC monetary policy is guided by an inflation targeting regime, while the foreign exchange policy is led by a floating exchange rate and, although unusual, the bank reserves the right to intervene in the foreign exchange markets. Chile is strongly committed to free trade and has welcomed large amounts of foreign investment. Chile has signed free trade agreements (FTAs) with a whole network of countries, including an FTA with the United States that was signed in 2003 and implemented in January 2004. Chile unilaterally lowered its across-the-board import tariff for all countries with which it does not have a trade agreement to 6% in 2003. Higher effective tariffs are charged only on imports of wheat, wheat flour, and sugar as a result of a system of import price bands. The price bands were ruled inconsistent with Chile's World Trade Organization (WTO) obligations in 2002, and the government has introduced legislation to modify them. Under the terms of the U.S.–Chile FTA, the price bands will be completely phased out for U.S. imports of wheat, wheat flour, and sugar within 12 years. Chile is a strong proponent of pressing ahead on negotiations for a Free Trade Area of the Americas (FTAA) and is active in the WTO's Doha round of negotiations, principally through its membership in the G-20 and Cairns Group. Most imports are not subject to the full statutory tariff, due to the extensive preferences negotiated outside the multilateral system through Regional Trade Agreements (RTAs). By the last version of the World Trade Organization's Trade Policy Review (October 2009), Chile had signed 21 RTAs with 57 countries and the number has continued to rise in recent years More recently, Chile has also been an active participant of deeper plurilateral trade agreement negotiations. Notably, Chile is currently in talks with eleven other economies in the Trans-Pacific Partnership (TPP), a proposed agreement that would stem from the existing P-4 Agreement between Brunei, Chile, New Zealand and Singapore. Chile has signed some form of bilateral or plurilateral agreement with each of the parties at TPP, although with different degrees of integration. Chile is also a party in conversations to establish the Pacific Alliance along with Peru, Mexico and Colombia. 2006 was a record year for Chilean trade. Total trade registered a 31% increase over 2005. During 2006, exports of goods and services totaled US$58 billion, an increase of 41%. This figure was somewhat distorted by the skyrocketing price of copper. In 2006, copper exports reached a historical high of US$33.3 billion. Imports totaled US$35 billion, an increase of 17% compared to the previous year. Chile thus recorded a positive trade balance of US$2.3 billion in 2006. The main destinations for Chilean exports were the Americas (US$39 billion), Asia (US$27.8 billion) and Europe (US$22.2 billion). Seen as shares of Chile's export markets, 42% of exports went to the Americas, 30% to Asia and 24% to Europe. Within Chile's diversified network of trade relationships, its most important partner remained the United States. Total trade with the U.S. was US$14.8 billion in 2006. Since the U.S.–Chile Free Trade Agreement went into effect on 1 January 2004, U.S.–Chilean trade has increased by 154%. Internal Government of Chile figures show that even when factoring out inflation and the recent high price of copper, bilateral trade between the U.S. and Chile has grown over 60% since then. Total trade with Europe also grew in 2006, expanding by 42%. The Netherlands and Italy were Chile's main European trading partners. Total trade with Asia also grew significantly at nearly 31%. Trade with Korea and Japan grew significantly, but China remained Chile's most important trading partner in Asia. Chile's total trade with China reached U.S. $8.8 billion in 2006, representing nearly 66% of the value of its trade relationship with Asia.= The growth of exports in 2006 was mainly caused by a strong increase in sales to the United States, the Netherlands, and Japan. These three markets alone accounted for an additional US$5.5 billion worth of Chilean exports. Chilean exports to the United States totaled US$9.3 billion, representing a 37.7% increase compared to 2005 (US$6.7 billion). Exports to the European Union were US$15.4 billion, a 63.7% increase compared to 2005 (US$9.4 billion). Exports to Asia increased from US$15.2 billion in 2005 to US$19.7 billion in 2006, a 29.9% increase. During 2006, Chile imported US$26 billion from the Americas, representing 54% of total imports, followed by Asia at 22%, and Europe at 16%. Mercosur members were the main suppliers of imports to Chile at US$9.1 billion, followed by the United States with US$5.5 billion and the European Union with US$5.2 billion. From Asia, China was the most important exporter to Chile, with goods valued at US$3.6 billion. Year-on-year growth in imports was especially strong from a number of countries – Ecuador (123.9%), Thailand (72.1%), Korea (52.6%), and China (36.9%). Chile's overall trade profile has traditionally been dependent upon copper exports. The state-owned firm CODELCO is the world's largest copper-producing company, with recorded copper reserves of 200 years. Chile has made an effort to expand nontraditional exports. The most important non-mineral exports are forestry and wood products, fresh fruit and processed food, fishmeal and seafood, and wine. Over the last several years, Chile has signed FTAs with the European Union, South Korea, New Zealand, Singapore, Brunei, China, and Japan. It reached a partial trade agreement with India in 2005 and began negotiations for a full-fledged FTA with India in 2006. Chile conducted trade negotiations in 2007 with Australia, Malaysia, and Thailand, as well as with China to expand an existing agreement beyond just trade in goods. Chile concluded FTA negotiations with Australia and an expanded agreement with China in 2008. The members of the P4 (Chile, Singapore, New Zealand, and Brunei) also plan to conclude a chapter on finance and investment in 2008. Successive Chilean governments have actively pursued trade-liberalizing agreements. During the 1990s, Chile signed free trade agreements (FTA) with Canada, Mexico, and Central America. Chile also concluded preferential trade agreements with Venezuela, Colombia, and Ecuador. An association agreement with Mercosur-Argentina, Brazil, Paraguay, and Uruguay-went into effect in October 1996. Continuing its export-oriented development strategy, Chile completed landmark free trade agreements in 2002 with the European Union and South Korea. Chile, as a member of the Asia-Pacific Economic Cooperation (APEC) organization, is seeking to boost commercial ties to Asian markets. To that end, it has signed trade agreements in recent years with New Zealand, Singapore, Brunei, India, China, and most recently Japan. In 2007, Chile held trade negotiations with Australia, Thailand, Malaysia, and China. In 2008, Chile hopes to conclude an FTA with Australia, and finalize an expanded agreement (covering trade in services and investment) with China. The P4 (Chile, Singapore, New Zealand, and Brunei) also plan to expand ties through adding a finance and investment chapter to the existing P4 agreement. Chile's trade talks with Malaysia and Thailand are also scheduled to continue in 2008. After two years of negotiations, the United States and Chile signed an agreement in June 2003 that will lead to completely duty-free bilateral trade within 12 years. The U.S.-Chile FTA entered into force 1 January 2004, following approval by the U.S. and Chilean congresses. The FTA has greatly expanded U.S.-Chilean trade ties, with total bilateral trade jumping by 154% during the FTA's first three years. On 1 January 2014, Chile-Vietnam Free Trade Agreement officially took effect. Unemployment hovered at 8–10% after the start of the economic slowdown in 1999, above the 7% average for the 1990s. Unemployment finally dipped to 7.8% in 2006, and continued to fall in 2007, averaging 6.8% monthly (up to August). Wages have risen faster than inflation as a result of higher productivity, boosting national living standards. The percentage of Chileans with household incomes below the poverty line – defined as twice the cost of satisfying a person's minimal nutritional needs – fell from 45.1% in 1987 to 11.7% in 2015, according to government polls. Critics in Chile, however, argue that poverty figures are considerably higher than those officially published; until 2016, the government defined the poverty line based on an outdated 1987 household consumption poll, instead of more recent polls from 1997 or 2007. According to critics who use data from the 1997 poll, the poverty rate goes up to 29%; a study published in 2017 claims that it reaches 26%. Using the relative yardstick favoured in many European countries, 27% of Chileans would be poor, according to Juan Carlos Feres of the ECLAC. Starting in 2016, a new Multidimensional Poverty Index is also used, which reached 20.9% using 2015 data. The percent of total income earned by the richest 20% of the Chilean population in 2000 was 61.0% of GDP, while the percent of total income earned by the poorest 20% of the Chilean population was 3.3% of GDP. Chile's Gini Coefficient in 2003 (53.8) has slightly changed in comparison with the value in 1995 (56.4). In 2005 the 10% poorest among the Chileans received 1.2% of GNP (2000 = 1.4%), while the 10% richest received 47% of GNP (2000 = 46%). Regarding the census, assessments have exhibited mixed results. An initial evaluation by a domestic independent experts panel released in August 2013 placed the omission rate in 9.3%, three times as much as other census in the region, and recommended annulling the census to hold a new version in 2015. The government sought an assessment by international experts before making a final decision. The team, which included three experts that represented the World Bank and the E.U. Statistics Commission, found "no basis for doubting the usability of the census data for most, if perhaps not all, of the customary uses" and recommended its release subject to the elimination of the imputation of housing units not observed on the ground during the enumeration and the concurrent publication of a methodological and administrative report. The following table shows the main economic indicators in 1980–2025. Inflation below 5% is in green. Main macroeconomic aggregates of GDP. Note: Data are preliminary. Source: Cuentas Nacionales de Chile – Evolución de la actividad económica en el año 2015 (p. 29), Central Bank of Chile, accessed on 23 March 2016. Gross domestic product by sector of the economy. Note: 2011 data are preliminary. Source: Cuentas Nacionales – Evolución de la actividad económica en el año 2011 (p. 34). Central Bank of Chile. accessed on 22 March 2012. Chile's top exports in 2013. Source: Central Bank of Chile's statistics database.
https://en.wikipedia.org/wiki?curid=5494
Transport in Chile Transport in Chile is mostly by road. The far south of the country is not directly connected to central Chile by road, and water transport also plays a part there. The railways were historically important in Chile, but now play a relatively small part in the country's transport system. Because of the country's geography and long distances between major cities, aviation is also important. "Total:" 79,799 km "Paved:" 41,012 km "Unpaved:" 38,788 km (1996 est.) 2,653 km Buses are now the main means of long distance transportation in Chile, following the decline of the rail network. The bus system covers the whole country, from Arica to Santiago (a 30-hour journey) and from Santiago to Punta Arenas (about 40 hours, with a change at Osorno). There are also international services to most other countries in South America. Longer-distance services are mostly on "semi-cama" (reclining seat) or "cama" (sleeper) buses, often double deck. Santiago began its public bus system Transantiago in 2007. Concepción's "Bio Bus" integrates with the electric train, Biotren, and is based on a dedicated right of way for buses. Not all lines connect. Chile's railways (except for a few dedicated industrial lines ) are operated by the state owned company "Empresa de los Ferrocarriles del Estado" (EFE), which completed a major investment programme in 2005. The rail system once served the entire country, running rail lines from Arica in the north to Puerto Montt in the south. Due partly to the nature of the terrain and evolution in transportation systems, rail travel has suffered greatly at the hands of bus and air competition. The train usually takes longer to reach a destination than a bus, and the comfort is comparable. Prices also tend to be uncompetitive. Rail freight transport has also suffered at the hands of the trucking industry and will continue to do so due to the immense leverage the truck driver's union can bring to bear if they were to feel threatened. The Ferrocarril de Antofagasta a Bolivia is a metre gauge railway in the north of the country. It was originally constructed in gauge. The northern rail line out of Santiago is now disused past the intersection with the Valparaíso line. Until there, it is used nearly exclusively for freight. Although the rest of the northern line is still in place, it is in a state of serious disrepair. The southern line runs as far as Puerto Montt and is electrified as far as the city of Temuco, from where diesel locomotives are used. Due to lack of budget and care, the 389 km Temuco to Puerto Montt section was abandoned in 1992 but after a $44m upgrade it has been back in use since 6 December 2005 with daily service between Victoria (north of Temuco) and Puerto Montt; today, however only the service between Victoria and Temuco still operates. Work to build/restore(?) the South Trans-Andean Railway link between Zapala, Argentina and Lonquimay, Chile was underway in 2005. Possible break-of-gauge. Possible rack railway. Construction was undertaken by Patagonia Ferrocanal SA, formed and funded by the province. The first 7 km was completed by January 2006. Commuter rail lines in Santiago are planned to connect to Melipilla and Batuco. There have been repeated case studies regarding the installation of a high-speed line between the cities of Valparaíso and Santiago, some even considering maglev trains, but no serious action has ever been taken on the matter. "total:" 45 ships ( or over) totaling / "ships by type:" (1999 est.) Arturo Merino Benitez International Airport, located in Santiago, is Chile's largest aviation facility. Chacao Channel bridge is a planned suspension bridge that was to link the island of Chiloé with mainland Chile crossing the Chacao Channel. It was one of the several projects that were planned to commemorate the Chile's bicentennial in 2010. If completed, it would have been the largest suspension bridge in South America.
https://en.wikipedia.org/wiki?curid=5496
Christmas Island The Territory of Christmas Island is an Australian external territory comprising the island of the same name. Christmas Island is located in the Indian Ocean, around south of Java and Sumatra and around north-west of the closest point on the Australian mainland. It has an area of . Christmas Island had a population of 1,843 residents , the majority of whom live in settlements on the northern tip of the island. The main settlement is Flying Fish Cove. Historically, Asian Australians of Chinese, Malay, and Indian descent have collectively formed the majority of the population. Today, around two-thirds of the island's population is estimated to have Malaysian Chinese or Singaporean Chinese origin (though just 21.2% of the population declared a Chinese ancestry in 2016), with significant numbers of Malays and European Australians and smaller numbers of Malaysian Indians and Eurasians. Several languages are in use, including English, Malay, and various Chinese dialects. Islam and Buddhism are major religions on the island. The religion question in the Australian census is optional and 28% of the population do not declare their religious belief, if any. The first European to sight the island was Richard Rowe of the "Thomas" in 1615. The island was later named on Christmas Day (25 December) 1643 by Captain William Mynors but only settled in the late 19th century. Its geographic isolation and history of minimal human disturbance has led to a high level of endemism among its flora and fauna, which is of interest to scientists and naturalists. The majority (63 percent) of the island is included in the Christmas Island National Park, which features several areas of primary monsoonal forest. Phosphate, deposited originally as guano, has been mined on the island since 1899. The first European to sight the island was Richard Rowe of the "Thomas" in 1615. Captain William Mynors of the "Royal Mary", an English East India Company vessel, named the island when he sailed past it on Christmas Day, in 1643. The island was included on English and Dutch navigation charts as early as the beginning of the 17th century, but it was not until 1666 that a map published by Dutch cartographer Pieter Goos included the island. Goos labelled the island "Mony" or "Moni", the meaning of which is unclear. English navigator William Dampier, aboard the privateer Charles Swan's ship, "Cygnet", made the earliest recorded visit to the sea around the island in March 1688. He found it uninhabited. Dampier wrote an account of the visit. Dampier was trying to reach Cocos from New Holland. His ship was blown off course in an easterly direction, arriving at Christmas Island twenty-eight days later. Dampier landed on the west coast, at "the Dales". Two of his crewmen became the first Europeans to set foot on Christmas Island. Captain Daniel Beeckman of the "Eagle" passed the island on 5 April 1714, chronicled in his 1718 book, "A Voyage to and from the Island of Borneo, in the East-Indies". The first attempt at exploring the island was in 1857 by the crew of the "Amethyst". They tried to reach the summit of the island but found the cliffs impassable. During the 1872–1876 "Challenger" expedition to Indonesia, naturalist John Murray carried out extensive surveys. In 1886, Captain John Maclear of , having discovered an anchorage in a bay that he named "Flying Fish Cove", landed a party and made a small collection of the flora and fauna. In the next year, Pelham Aldrich, on board HMS "Egeria", visited the island for ten days, accompanied by J. J. Lister, who gathered a larger biological and mineralogical collection. Among the rocks then obtained and submitted to Murray for examination were many of nearly pure phosphate of lime. This discovery led to annexation of the island by the British Crown on 6 June 1888. Soon afterwards, a small settlement was established in Flying Fish Cove by G. Clunies Ross, the owner of the Cocos (Keeling) Islands some to the southwest, to collect timber and supplies for the growing industry on Cocos. In 1897 the island was visited by Charles W. Andrews, who did extensive research on the natural history of the island, on behalf of the British Museum. Phosphate mining began in 1899 using indentured workers from Singapore, Malaya, and China. John Davis Murray, a mechanical engineer and recent graduate of Purdue University, was sent to supervise the operation on behalf of the Phosphate Mining and Shipping Company. Murray was known as the "King of Christmas Island" until 1910, when he married and settled in London. The island was administered jointly by the British Phosphate commissioners and district officers from the United Kingdom Colonial Office through the Straits Settlements, and later the Crown Colony of Singapore. Hunt (2011) provides a detailed history of Chinese indentured labor on the island during those years. In 1922, scientists attempted unsuccessfully to view a solar eclipse from the island to test Albert Einstein's Theory of Relativity. From the outbreak of the South-East Asian theatre of World War II in December 1941, Christmas Island was a target for Japanese occupation because of its rich phosphate deposits. A naval gun was installed under a British officer and four NCOs and 27 Indian soldiers. The first attack was carried out on 20 January 1942, by , which torpedoed a Norwegian freighter, the "Eidsvold". The vessel drifted and eventually sank off West White Beach. Most of the European and Asian staff and their families were evacuated to Perth. In late February and early March 1942, there were two aerial bombing raids. Shelling from a Japanese naval group on 7 March led the District Officer to hoist the white flag. But after the Japanese naval group sailed away, the British officer raised the Union Flag once more. During the night of 10–11 March, a mutiny of the Indian troops, abetted by Sikh policemen, led to the killing of the five British officers in the garrison, and the imprisonment of the remaining 21 Europeans. At dawn on 31 March 1942, a dozen Japanese bombers launched the attack, destroying the radio station. The same day, a Japanese fleet of nine vessels arrived, and the island was surrounded. About 850 men of the Japanese 21st and 24th Special Base Forces and 102nd Construction Unit came ashore at Flying Fish Cove and occupied the island. They rounded up the workforce, most of whom had fled to the jungle. Sabotaged equipment was repaired and preparations were made to resume the mining and export of phosphate. Only twenty men from the 21st Special Base Force were left as a garrison. Isolated acts of sabotage and the torpedoing of the "Nissei Maru" at the wharf on 17 November 1942 meant that only small amounts of phosphate were exported to Japan during the occupation. In November 1943, over 60% of the island's population was evacuated to Surabaya prison camps, leaving a total population of just under 500 Chinese and Malays and 15 Japanese to survive as best they could. In October 1945, re-occupied Christmas Island. After the war, seven mutineers were traced and prosecuted by the Military Court in Singapore. In 1947, five of them were sentenced to death. However, following representations made by the newly independent government of India, their sentences were reduced to penal servitude for life. At Australia's request, the United Kingdom transferred sovereignty to Australia, with a $20 million payment from the Australian government to Singapore as compensation for the loss of earnings from the phosphate revenue. The United Kingdom's Christmas Island Act was given royal assent on 14 May 1958, enabling Britain to transfer authority over Christmas Island from Singapore to Australia by an order-in-council. Australia's Christmas Island Act was passed in September 1958 and the island was officially placed under the authority of the Commonwealth of Australia on 1 October 1958. Under Commonwealth Cabinet Decision 1573 of 9 September 1958, D. E. Nickels was appointed the first official representative of the new territory. In a media statement on 5 August 1960, the minister for territories, Paul Hasluck, said, among other things, that, "His extensive knowledge of the Malay language and the customs of the Asian people ... has proved invaluable in the inauguration of Australian administration ... During his two years on the island he had faced unavoidable difficulties ... and constantly sought to advance the island's interests." John William Stokes succeeded him and served from 1 October 1960, to 12 June 1966. On his departure, he was lauded by all sectors of the island community. In 1968, the official secretary was retitled an administrator and, since 1997, Christmas Island and the Cocos (Keeling) Islands together are called the Australian Indian Ocean Territories and share a single administrator resident on Christmas Island. Recollections of the island's history and lifestyle, and lists and timetables of the island's leaders and events since its settlement are at the World Statesmen site and in Neale (1988), Bosman (1993), Hunt (2011), and Stokes (2012). The settlement of Silver City was built in the 1970s, with aluminium-clad houses that were supposed to be cyclone-proof. The 2004 Indian Ocean earthquake and tsunami centred off the western shore of Sumatra in Indonesia, resulted in no reported casualties, but some swimmers were swept some out to sea for a time before being swept back in. From the late 1980s and early 1990s, boats carrying asylum seekers, mainly departing from Indonesia, began landing on the island. In 2001, Christmas Island was the site of the "Tampa" controversy, in which the Australian government stopped a Norwegian ship, MV "Tampa", from disembarking 438 rescued asylum-seekers. The ensuing standoff and the associated political reactions in Australia were a major issue in the 2001 Australian federal election. The Howard government operated the "Pacific Solution" from 2001 to 2007, excising Christmas Island from Australia's migration zone so that asylum seekers on the island could not apply for refugee status. Asylum seekers were relocated from Christmas Island to Manus Island and Nauru. In 2006, an immigration detention centre, containing approximately 800 beds, was constructed on the island for the Department of Immigration and Multicultural Affairs. Originally estimated to cost  million, the final cost was over $400 million. In 2007, the Rudd government decommissioned Manus Regional Processing Centre and Nauru detention centre; processing would then occur on Christmas Island itself. In December 2010, 48 asylum-seekers died just off the coast of the island in what became known as the Christmas Island boat disaster when their boat hit the rocks near Flying Fish Cove, and then smashed against nearby cliffs. In the case "Plaintiff M61/2010E v Commonwealth of Australia", the High Court of Australia ruled, in a 7–0 joint judgment, that asylum seekers detained on Christmas Island were entitled to the protections of the Migration Act. Accordingly, the Commonwealth was obliged to afford asylum seekers a minimum of procedural fairness when assessing their claims. , after the interception of four boats in six days, carrying 350 people, the Immigration Department stated that there were 2,960 "irregular maritime arrivals" being held in the island's five detention facilities, which exceeded not only the "regular operating capacity" of 1,094 people, but also the "contingency capacity" of 2,724. The Christmas Island Immigration Reception and Processing Centre closed on 30 September 2018. On 13 February 2019 the Morrison government announced it would re-open the centre, after Australia's parliament passed legislation giving sick asylum seekers easier access to mainland hospitals. During the 2020 COVID-19 outbreak the government opened parts of the Immigration Reception and Processing Centre to be used as a quarantine facility to accommodate Australian citizens who had been in Wuhan, the point of origin of the COVID-19 pandemic. The evacuees arrived on February 3. They left 14 days later to their homes on the mainland. The island is about in greatest length and in breadth. The total land area is , with of coastline. The island is the flat summit of an underwater mountain more than high, which rises from about below the sea and only about above it. The mountain was originally a volcano, and some basalt is exposed in places such as The Dales and Dolly Beach, but most of the surface rock is limestone accumulated from coral growth. The karst terrain supports numerous anchialine caves. The summit of this mountain peak is formed by a succession of Tertiary limestones ranging in age from the Eocene or Oligocene up to recent reef deposits, with intercalations of volcanic rock in the older beds. Steep cliffs along much of the coast rise abruptly to a central plateau. Elevation ranges from sea level to at Murray Hill. The island is mainly tropical rainforest, 63% of which is national parkland. The narrow fringing reef surrounding the island poses a maritime hazard. Christmas Island lies northwest of Perth, Western Australia, south of Indonesia, ENE of the Cocos (Keeling) Islands, and west of Darwin, Northern Territory. Its closest point to the Australian mainland is from the town of Exmouth, Western Australia. Christmas Island has of shoreline but only small parts of the shoreline are easily accessible. The island's perimeter is embodied by sharp cliff faces, making many of the island's beaches difficult to get to. Some of the easily accessible beaches include Flying Fish Cove (main beach), Lily Beach, Ethel Beach, and Isabel Beach, while the more difficult beaches to access include Greta Beach, Dolly Beach, Winifred Beach, Merrial Beach, and West White Beach, which all require a vehicle with four wheel drive and a difficult walk through dense rainforest. Christmas Island is located towards the southern edge of the equatorial region. Climate is tropical and temperatures vary little throughout the year. The highest temperature is usually around in March and April, while the lowest temperature is and occurs in August. There is a dry season from July to October with only occasional showers. The wet season is between November and June, and includes monsoons, with downpours of rain at random times of the day. Tropical cyclones also occur in the wet season, bringing very strong winds, rain and enormous seas. As of the 2016 Australian census, the population of Christmas Island is 1,843. 21.2% of the population had Chinese ancestry (up from 18.3% in 2001), 12.7% had generic Australian ancestry (11.7% in 2001), 12% had Malay ancestry (9.3% in 2001), 10% had English ancestry (8.9% in 2001), and 2.3% of the population was of Irish origin (the same share as in 2001). 48.1% of the population has undetermined ancestry. , most are people born in Christmas Island and many are of Chinese and Malay origin. 38.5% of people were born in Australia. The next most common country of birth was Malaysia at 20.1%. 27.8% of the population spoke English as their family language, while 17.2% spoke Mandarin Chinese, 17.2% spoke Malay, 3.7% Cantonese, 1.5% Southern Min, and 1% Tagalog. Additionally, there are small local populations of Malaysian Indians and Eurasians. The 2016 Australian census recorded that the population of Christmas Island was 38.7% female and 61.3% male, while in 2011 the figures had been 29.3% female and 70.7% male. In contrast, the 2016 figures for the whole of Australia were 50.7% female, 49.3% male. Since 1998 there has been no provision for childbirth on the island; expectant mothers travel to mainland Australia approximately one month before their expected date in order to give birth. Christmas Island is a non-self-governing external territory of Australia, administered by the Department of Infrastructure, Regional Development and Cities (from 29 November 2007 until 14 September 2010, administration was carried out by the Attorney-General's Department, and prior to this by the Department of Transport and Regional Services). The legal system is under the authority of the Governor-General of Australia and Australian law. An administrator appointed by the Governor-General represents the monarch and Australia and lives on the island. The territory falls under no formal state jurisdiction, but the Western Australian Government provides many services as established by the Christmas Island Act. The Australian government provides services through the Christmas Island Administration and the Department of Infrastructure and Regional Development. Under the federal government's Christmas Island Act 1958, Western Australian laws are applied to Christmas Island; non-application or partial application of such laws is at the discretion of the federal government. The act also gives Western Australian courts judicial power over Christmas Island. Christmas Island remains constitutionally distinct from Western Australia, however; the power of the state to legislate for the territory is delegated by the federal government. The kind of services typically provided by a state government elsewhere in Australia are provided by departments of the Western Australian government, and by contractors, with the costs met by the federal government. A unicameral Shire of Christmas Island with nine seats provides local government services and is elected by popular vote to serve four-year terms. Elections are held every two years, with four or five of the members standing for election. Christmas Island residents who are Australian citizens vote in Australian federal elections. Christmas Island residents are represented in the House of Representatives by the Division of Lingiari in the Northern Territory and in the Senate by Northern Territory senators. At the 2019 federal election, the Labor Party received majorities from Christmas Island electors in both the House of Representatives and the Senate. Residents find the system of administration frustrating, with the island run by bureaucrats in the federal government, but subject to the laws of Western Australia and enforced by federal police. There is a feeling of resignation that any progress on local issues is hampered by the confusing governance system. A number of islanders support self-governance, including shire president Gordon Thompson, who also believes that a lack of news media to cover local affairs had contributed to political apathy among residents. In early 1986, the Christmas Island Assembly held a design competition for an island flag; the winning design was adopted as the informal flag of the territory for over a decade, and in 2002 it was made the official flag of Christmas Island. Phosphate mining had been the only significant economic activity, but in December 1987 the Australian government closed the mine. In 1991, the mine was reopened by a consortium which included many of the former mine workers as shareholders. With the support of the government, the $34 million Christmas Island Casino and Resort opened in 1993 but was closed in 1998. , the resort has re-opened without the casino. The Australian government in 2001 agreed to support the creation of a commercial spaceport on the island; however, this has not yet been constructed and appears that it will not proceed. The Howard government built a temporary immigration detention centre on the island in 2001 and planned to replace it with a larger, modern facility at North West Point until Howard's defeat in the 2007 elections. Historically, the majority of Christmas Islanders were those of Chinese, Malay, and Indian origins, the initial permanent settlers. Today, the majority of residents are Chinese, with significant numbers of European Australians and Malays as well as smaller Indian and Eurasian communities too. The main languages of Christmas Island are English and Chinese. Since the turn of the 21st century and right up to the present, Europeans have mainly confined themselves to the Settlement, where there is a small supermarket and several restaurants; the Malays live in the Flying Fish Cove,also known as Kampong; and the Chinese reside in Poon San (Mandarin for “in the middle of the hill”). Religious beliefs are diverse and include Buddhism, Taoism, Christianity, Islam and Confucianism. There is a mosque, a Christian church, a Bahá'í centre and around twenty Chinese temples and shrines, which include seven Buddhist temples (like Guan Yin Monastery (观音寺) at Gaze Road), ten Taoist temples (like Soon Tian Kong (顺天宫) in South Point and Grants Well Guan Di Temple) and shrines dedicated to Na Tuk Kong or Datuk Keramat on the island. There are many religious festivals, such as Spring Festival, Chap Goh Meh, Qingming Festival, Zhong Yuan Festival, Hari Raya, Christmas and Easter. The main local organization that "promotes and supports" the "status and interests" of female Christmas Islanders is the "Christmas Island Women's Association" that was established in 1989 and is a member organisation of the Associated Country Women of the World. Christmas Island is well known for its biological diversity. There are many rare species of animals and plants on the island, making nature-walking a popular activity. Along with the diversity of species, many different types of caves exist, such as plateau caves, coastal caves, raised coastal caves and alcoves, sea caves, fissure caves, collapse caves, and basalt caves; most of these are near the sea and have been formed by the action of water. Altogether, there are approximately 30 caves on the island, with Lost Lake Cave, Daniel Roux Cave, and Full Frontal Cave being the most well-known. The many freshwater springs include Hosnies Spring Ramsar, which also has a mangrove stand. The Dales is a rainforest in the western part of the island and consists of seven deep valleys, all of which were formed by spring streams. Hugh's Dale waterfall is part of this area and is a popular attraction. The annual breeding migration of the Christmas Island red crabs is a popular event. Fishing is another common activity. There are many distinct species of fish in the oceans surrounding Christmas Island. Snorkeling and swimming in the ocean are two other activities that are extremely popular. Walking trails are also very popular, for there are many beautiful trails surrounded by extravagant flora and fauna. 63% of the island is national park making it one of the main attractions to experience when visiting. Christmas Island was uninhabited until the late 19th century, allowing many species to evolve without human interference. Two-thirds of the island has been declared a National Park, which is managed by the Australian Department of Environment and Heritage through Parks Australia. Christmas Island contains unique species, both of flora and fauna, some of which are threatened with, or have become, extinct. The dense rainforest has grown in the deep soils of the plateau and on the terraces. The forests are dominated by 25 tree species. Ferns, orchids and vines grow on the branches in the humid atmosphere beneath the canopy. The 135 plant species include at least 18 that are found nowhere else. The rainforest is in great condition despite the mining activities over the last 100 years. Areas that have been damaged by mining are now a part of an ongoing rehabilitation project. The island is small and covers 135 square kilometres of land which 63% of that land has been declared National park. Christmas Island's endemic plants include the trees "Arenga listeri", "Pandanus elatus" and "Dendrocnide peltata" var. "murrayana"; the shrubs "Abutilon listeri", "Colubrina pedunculata", "Grewia insularis" and "Pandanus christmatensis"; the vines "Hoya aldrichii" and "Zehneria alba"; the herbs "Asystasia alba", "Dicliptera maclearii" and "Peperomia rossii"; the grass "Ischaemum nativitatis"; the fern "Asplenium listeri"; and the orchids "Brachypeza archytas", "Flickingeria nativitatis", "Phreatia listeri" and "Zeuxine exilis". Two species of native rats, the Maclear's and bulldog rats, have become extinct since the island was settled, while the Javan rusa deer has been introduced. The endemic Christmas Island shrew has not been seen since the mid-1980s and may be already extinct, while the Christmas Island pipistrelle (a small bat) is presumed to be extinct. The fruit bat (flying fox) species "Pteropus natalis" is only found on Christmas Island, its epithet "natalis" is a reference to that name. The species is probably the last native mammal, and an important pollinator and rainforest seed-disperser; the population is also in decline and under increasing pressure from land clearing and introduced pest species. The flying fox's low rate of reproduction (one pup each year) and high infant mortality rate makes it especially vulnerable and the conservation status is as critically endangered. Flying foxes are an 'umbrella' species helping forests regenerate and other species survive in stressed environments. The land crabs and seabirds are the most noticeable fauna on the island. Christmas Island has been identified by BirdLife International as both an Endemic Bird Area and an Important Bird Area because it supports five endemic species and five subspecies as well as over one percent of the world populations of five other seabirds. Twenty terrestrial and intertidal species of crab have been described here, of which thirteen are regarded as true land crabs, being dependent on the ocean only for larval development. Robber crabs, known elsewhere as coconut crabs, also exist in large numbers on the island. The annual red crab mass migration (around 100 million animals) to the sea to spawn has been called one of the wonders of the natural world. This takes place each year around November – after the start of the wet season and in synchronisation with the cycle of the moon. Once at the ocean, the mothers release the embryos where they can survive and grow until they are able to live on land. The island is a focal point for seabirds of various species. Eight species or subspecies of seabirds nest on it. The most numerous is the red-footed booby, which nests in colonies, using trees on many parts of the shore terrace. The widespread brown booby nests on the ground near the edge of the seacliff and inland cliffs. Abbott's booby (listed as endangered) nests on tall emergent trees of the western, northern and southern plateau rainforest, the only remaining nesting habitat for this bird in the world. Another endangered and endemic bird, the Christmas frigatebird, has nesting areas on the northeastern shore terraces. The more widespread great frigatebirds nest in semi-deciduous trees on the shore terrace, with the greatest concentrations being in the North West and South Point areas. The common noddy and two species of bosun or tropicbirds, with their brilliant gold or silver plumage and distinctive streamer tail feathers, also nest on the island. Of the ten native land birds and shorebirds, seven are endemic species or subspecies. This includes the Christmas thrush and the Christmas imperial pigeon. Some 86 migrant bird species have been recorded as visitors to the island. Six species of butterfly are known to occur on Christmas Island. These are the Christmas swallowtail ("Papilio memnon"), striped albatross ("Appias olferna"), Christmas emperor ("Polyura andrewsi"), king cerulean ("Jamides bochus"), lesser grass-blue ("Zizina otis"), and Papuan grass-yellow ("Eurema blanda"). Insect species include the yellow crazy ant ("Anoplolepis gracilipes"), introduced to the island and since subjected to attempts to destroy the supercolonies that emerged with aerial spraying of the insecticide Fipronil. Christmas Island has access to a range of modern communication services. Radio broadcasts from Australia include ABC Radio National, ABC Kimberley, Triple J and Red FM. All services are provided by satellite links from the mainland. Broadband internet became available to subscribers in urban areas in mid-2005 through the local internet service provider, CIIA (formerly dotCX). Christmas Island, due to its close proximity to Australia's northern neighbors, falls within many of the satellite footprints throughout the region. This results in ideal conditions for receiving various Asian broadcasts, which locals sometimes prefer to those emanating from Western Australia. Additionally, ionospheric conditions are conducive to terrestrial radio transmissions, from HF through VHF and sometimes into UHF. The island plays home to a small array of radio equipment that spans a good chunk of the usable spectrum. A variety of government-owned and operated antenna systems are employed on the island to take advantage of this. Free-to-air digital television stations from Australia are broadcast in the same time zone as Perth, and are broadcast from three separate locations: Cable television from Australia, Malaysia, Singapore and the United States commenced in January 2013. Telephone services are provided by Telstra and are a part of the Australian network with the same prefix as Western Australia, South Australia and the Northern Territory (08). A GSM mobile telephone system on the 900 MHz band replaced the old analogue network in February 2005. A postal agency was opened on the island in 1901 and sold stamps of the Strait Settlements. After the Japanese occupation (1942–1945), postage stamps of the British Military Administration in Malaya were in use, then stamps of Singapore. In 1958, the island received its own postage stamps after being put under Australian custody. It had a large philatelic and postal independence, managed first by the Phosphate Commission (1958–1969) and then by the island's administration (1969–1993). This ended on 2 March 1993 when Australia Post became the island's postal operator; Christmas Island stamps may be used in Australia and Australian stamps may be used on the island. A container port exists at Flying Fish Cove with an uncompleted alternative container-unloading point to the east of the island at Norris Point, intended for use during the December-to-March "swell season" of rough seas. The standard gauge Christmas Island Phosphate Co.'s Railway from Flying Fish Cove to the phosphate mine was constructed in 1914. It was closed in December 1987, when the Australian government closed the mine, and since has been recovered as scrap, leaving only earthworks in places. Virgin Australia Regional Airlines provides two weekly flights to Christmas Island Airport from Perth, Western Australia, Garuda Indonesia conduct weekly open-charter flights from/to Jakarta with bookings done through Christmas Island Travel Exchange and Malindo Air operate fortnightly open-charter flights from/to Kuala Lumpur with bookings done through Evercrown Air Services. Hire cars are available from the airport however no franchised companies are represented. CI Taxi Service also operates most days. Due to the lack of 3G or 4G, the island's sole taxi operator couldn't meet the requirement issued by WA Department of Transport to install electronic meters, and the operator was forced to close at the end of June 2019. The road network covers most of the island and is of generally good quality, although four-wheel drive vehicles are needed to reach some of the more distant parts of the rainforest or the more isolated beaches on the rough dirt roads. The island-operated crèche is in the Recreation Centre. Christmas Island District High School, catering to students in grades P-12, is run by the Western Australian Education Department. There are no universities on Christmas Island. The island has one public library. Cricket and rugby league are the two main organised sports on the island. Christmas Island Cricket Club was founded in 1959 now known as Christmas Island cricket and sporting club. In 2019 the club celebrated its 60 years anniversary. The club entered its first representative team into the WACA country week this year 2020. Where they were runners up in F division. Rugby league is growing in the island: the first game was played in 2016, and a local committee, with the support of NRL Western Australia, is willing to organize matches with nearby Cocos Islands and to create a Rugby League Competition in the Indian Ocean area.
https://en.wikipedia.org/wiki?curid=5500
Clipperton Island Clipperton Island ( or ; ) is an uninhabited coral atoll in the eastern Pacific Ocean off the coast of Central America. It is from Paris, France, from Papeete, Tahiti, and from Mexico. It is an overseas state private property of France, under direct authority of the Minister of Overseas France. The atoll is south-west of Mexico, west of Nicaragua, west of Costa Rica and north-west of the Galápagos Islands, Ecuador, at . Clipperton is about south-east of Socorro Island in the Revillagigedo Archipelago, which is the nearest land while the nearest French-owned island is Hiva Oa. It is low-lying and largely barren, with some scattered grasses and a few clumps of coconut palms ("Cocos nucifera"). Land elevations average , though a small volcanic outcrop rising to on its south-east side is considerably higher and is referred to as "Clipperton Rock". The surrounding reef is exposed at low tide. The presence of this rock means that technically Clipperton is not an atoll but an island with a barrier reef. Clipperton has had no permanent inhabitants since 1945. It is visited on occasion by fishermen, French Navy patrols, scientific researchers, film crews, and shipwreck survivors. It has become a popular site for transmissions by ham radio operators. Clipperton has a ring-shaped atoll which completely encloses a stagnant fresh water lagoon, and is in circumference. The lagoon is devoid of fish, and contains some deep basins with depths of , including a spot known as Trou-Sans-Fond, or "the bottomless hole", with acidic water at its base. The water is described as being almost fresh at the surface, and highly eutrophic. Seaweed beds cover approximately 45 percent of the lagoon's surface. The rim averages in width, reaching in the west and narrows to in the north-east, where sea waves occasionally spill over into the lagoon. While some sources have rated the lagoon water as non-potable, testimony from the crew of the tuna clipper M/V "Monarch," stranded for 23 days in 1962 after their boat sank, indicates otherwise. Their report reveals that the lagoon water, while not tasting very good, was drinkable, though "muddy and dirty". Several of the castaways drank it, with no apparent ill effects. Survivors of an ill-fated Mexican military colony in 1917 (see below) indicated that they were dependent upon rain for their water supply, catching it in old boats they used for this purpose. Aside from the lagoon and water caught from rain, no other freshwater sources are known to exist. It has a tropical oceanic climate, with average temperatures of . The rainy season occurs from May to October, when it is subject to tropical storms and hurricanes. Surrounding ocean waters are warm, pushed by equatorial and counter-equatorial currents. It has no known natural resources (its guano having been depleted early in the 20th century). Although 115 species of fish have been identified in nearby waters the only economic activity in the area is tuna fishing. When Snodgrass and Heller visited in 1898, they reported that "no land plant is native to the island". Historical accounts from 1711, 1825, and 1839 show a low grassy or suffrutescent (partially woody) flora. During Sachet's visit in 1958, the vegetation was found to consist of a sparse cover of spiny grass and low thickets, a creeping plant ("Ipomoea" spp.), and stands of coconut palm. This low-lying herbaceous flora seems to be a pioneer in nature, and most of it is believed to be composed of recently introduced species. Sachet suspected that "Heliotropium curassavicum" and possibly "Portulaca oleracea" were native. Coconut palms and pigs were introduced in the 1890s by guano miners. The pigs reduced the crab population, which in turn allowed grassland to gradually cover about 80 percent of the land surface. The elimination of these pigs in 1958 — the result of a personal project by Kenneth E. Stager — has caused most of this vegetation to disappear, as the population of land crabs ("Johngarthia planata") has returned to millions. The result is virtually a sandy desert, with only 674 palms counted by Christian Jost during the "Passion 2001" French mission, and five islets in the lagoon with grass that the terrestrial crabs cannot reach. On the north-west side the most abundant plant species are "Cenchrus echinatus", "Sida rhombifolia", and "Corchorus aestuans". These plants compose a shrub cover up to 30 cm in height and are intermixed with "Eclipta", "Phyllanthus", and "Solanum", as well as a taller plant, "Brassica juncea". A unique feature is that the vegetation is arranged in parallel rows of species, with dense rows of taller species alternating with lower, more open vegetation. This was assumed to be a result of the phosphate mining method of trench-digging. The only land animals known to exist are two species of reptiles ("Gehyra insulensis", a gecko, and "Emoia cyanura," a skink), bright-orange land crabs ("Johngarthia planata", sometimes known as the 'Clipperton Crab', although it is also found on other islands in the eastern Pacific), birds, and rats. The rats probably arrived on large fishing boats that were wrecked on the island in 1999 and 2000. Bird species include white terns, masked boobies, sooty terns, brown boobies, brown noddies, black noddies, great frigatebirds, coots, martins (swallows), cuckoos and yellow warblers. Ducks have been reported in the lagoon. The island has been identified as an Important Bird Area by BirdLife International because of the large breeding colony of masked boobies, with 110,000 individual birds recorded. The lagoon harbors millions of isopods, which are said to deliver an especially painful sting. A 2005 report by the NOAA's Southwest Fisheries Science Center indicated that the increased rat presence had led to a decline in both crab and bird populations, causing a corresponding increase in both vegetation and coconut palms. This report urgently recommended eradication of rats so that vegetation might be reduced and the island might return to its "pre-human" state. The island was discovered by Alvaro Saavedra Cedrón on 15 November 1528. The expedition was commissioned by Hernán Cortés, the Spanish Conquistador in Mexico, to find a route to the Philippines. The island was rediscovered on Good Friday, 3 April 1711 by Frenchmen Martin de Chassiron and Michel Du Bocage, commanding the French ships "La Princesse" and "La Découverte". It was given the name "Île de la Passion" (). They drew up the first map and claimed the island for France. The first scientific expedition took place in 1725 under Frenchman M. Bocage, who lived on the island for several months. In 1858, France formally laid claim. The current name comes from John Clipperton, an English pirate and privateer who fought the Spanish during the early 18th century, and who is said to have passed by the island. Some sources claim that he used it as a base for his raids on shipping. Other claimants included the United States, whose American Guano Mining Company claimed it under the Guano Islands Act of 1856; Mexico also claimed it due to activities undertaken there as early as 1848–1849. On 17 November 1858 Emperor Napoleon III annexed it as part of the French colony of Tahiti. This did not settle the ownership question. On 24 November 1897, French naval authorities found three Americans working for the American Guano Company, who had raised the American flag. U.S. authorities denounced their act, assuring the French that they did not intend to assert American sovereignty. Mexico reasserted its claim late in the 19th century, and on 13 December 1897 sent the gunboat "La Demócrata" to occupy and annex it. A colony was established, and a series of military governors was posted, the last one being Ramón Arnaud (1906–1916). France insisted on its ownership, and a lengthy diplomatic correspondence between Mexico and France led to the conclusion of a treaty on March 2, 1909, to seek binding international arbitration by King Victor Emmanuel III of Italy, with each nation promising to abide by his determination. In 1931, King Victor Emmanuel III issued his arbitral decision in the "Clipperton Island Case", declaring Clipperton to be a French possession. The French rebuilt the lighthouse and settled a military outpost, which remained for seven years before being abandoned. The British Pacific Island Company acquired the rights to guano deposits in 1906 and built a mining settlement in conjunction with the Mexican government. That same year, a lighthouse was erected under the orders of President Porfirio Díaz. By 1914 around 100 people—men, women, and children—were living there, resupplied every two months by a ship from Acapulco. With the escalation of fighting in the Mexican Revolution, the regular resupply visits ceased and the inhabitants were left to their own devices. By 1917 all but one of the male inhabitants had died. Many had perished from scurvy, while others (including Captain Arnaud) died during an attempt to sail after a passing ship to fetch help. Lighthouse keeper Victoriano Álvarez was the last man on the island, together with 15 women and children. Álvarez proclaimed himself "king" and began a campaign of rape and murder, before being killed by Tirza Rendon, who was the recipient of his unwanted attention. Almost immediately after Álvarez's death four women and seven children, the last survivors, were picked up by the US Navy gunship on 18 July 1917. No more attempts were made to colonize it, though it was briefly occupied during the 1930s and 1940s. The story of the Mexican colony has been the subject of several novels, including Ivo Mansmann's "Clipperton, Schicksale auf einer vergessenen Insel" ("Clipperton, Destinies on a Forgotten Island") in German, Colombian writer Laura Restrepo's "La Isla de la Pasión" ("Passion Island") in Spanish, and Ana Garcia Bergua's "Isla de Bobos" ("Island of Fools"), also in Spanish. The island was abandoned by the end of World War II after being briefly occupied by the US from 1944–1945. Since then it has been visited by sports fishermen, patrols of the French Navy, and by Mexican tuna and shark fishermen. There have been infrequent scientific and amateur radio expeditions, and in 1978 Jacques-Yves Cousteau visited with his team of divers and a survivor from the 1917 evacuation to film a television special called "Clipperton: The Island that Time Forgot." It was visited by ornithologist Ken Stager of the Los Angeles County Museum in 1958. Appalled at the depredations visited by feral pigs upon the island's brown booby and masked booby colonies (reduced to 500 and 150 birds, respectively), Stager procured a shotgun and killed all 58 pigs. By 2003, the booby colonies had 25,000 brown boobies and 112,000 masked boobies, the world's second-largest brown booby colony and its largest masked booby colony. When the independence of Algeria in 1962 threatened French nuclear testing sites in Algeria, the French Ministry of Defense considered Clipperton Island as a possible replacement. This was eventually ruled out due to the island's hostile climate and remote location. The French explored reopening the lagoon and developing a harbor for trade and tourism during the 1970s, but this too was abandoned. An automatic weather installation was completed on 7 April 1980, with data collected by this station transmitted directly by satellite to Brittany. In 1981, the Academy of Sciences for Overseas Territories recommended that the island have its own economic infrastructure, with an airstrip and a fishing port in the lagoon. This would mean opening up the lagoon by creating a passage in the atoll rim. For this purpose, an agreement was signed with the French government, represented by the High Commissioner for French Polynesia, whereby the island became French state property. In 1986 a meeting took place regarding the establishment of a permanent base for fishing, between the High Commissioner and the survey firm for the development and exploitation of the island (SEDEIC). Taking into account the economic constraints, the distance from markets, and the small size of the atoll, nothing apart from preliminary studies was undertaken. All plans for development were abandoned. In early 1962 the island provided a home to nine crewmen of the sunken tuna clipper MV "Monarch", stranded for 23 days from 6 February to 1 March. They reported that the lagoon water was drinkable, though they preferred to drink water from the coconuts they found. Unable to use any of the dilapidated buildings, they constructed a crude shelter from cement bags and tin salvaged from Quonset huts built by the American military 20 years earlier. Wood from the huts was used for firewood, and fish caught off the fringing reef combined with some potatoes and onions they had saved from their sinking vessel augmented the island's meager supply of coconuts. The crewmen reported that they tried eating bird's eggs, but found them to be rancid, and they decided after trying to cook a "little black bird" that it did not have enough meat to make the effort worthwhile. Pigs had been eradicated, though the crewmen reported seeing their skeletons around the atoll. The crewmen were eventually discovered by another fishing boat and rescued by the United States Navy destroyer USS "Robison". In 1988, five Mexican fishermen became lost at sea after a storm during their trip along the coast of Costa Rica. They drifted within sight of the island but were unable to reach it. Steven Longbaugh and David Heritage, two American deckhands from a fishing boat based in California, were stranded for three weeks in 1998. They were rescued after rebuilding a survival radio and using distress flares to signal for help. The Mexican and French oceanographic expedition SURPACLIP (UNAM Mexico and UNC Nouméa) made extensive studies in 1997. In 2001, French geographer Christian Jost extended the 1997 studies through his French "Passion 2001" expedition, explaining the evolution of the ecosystem, and releasing several papers, a video film, and a website. In 2003 Lance Milbrand stayed for 41 days on a National Geographic Society expedition, recording his adventure in video, photos, and a written diary (see links below). In 2005, the ecosystem was extensively studied for four months by a scientific mission organized by Jean-Louis Étienne, which made a complete inventory of mineral, plant, and animal species, studied algae as deep as 100 m (330 ft) below sea level, and examined the effects of pollution. A 2008 expedition from the University of Washington's School of Oceanography collected sediment cores from the lagoon to study climate change over the last millennium. On 21 February 2007, administration was transferred from the High Commissioner of the Republic in French Polynesia to the Minister of Overseas France. In 2007 a recreational scuba diving expedition explored the reefs surrounding Clipperton and compared the marine life with the reports of the Connie Limbaugh (Scripps) expeditions in 1956 and 1958. Recreational scuba diving expeditions are now made every spring. During the night of 10 February 2010, the "Sichem Osprey", a Maltese chemical tanker, ran aground on its way from the Panama Canal to South Korea. The ship contained xylene, a clear, flammable volatile liquid. All 19 crew members were reported safe, and the vessel reported no leaks. The vessel was refloated on 6 March and returned to service. In mid-March 2012, the crew from The Clipperton Project noted the widespread presence of refuse, particularly on the northeast shore and around the Rock. Debris including plastic bottles and containers create a potentially harmful environment to its flora and fauna. This trash is common to only two beaches (North East and South West) and the rest of the island is fairly clean. Other refuse has been left over after the occupation by the Americans in 1944–1945, the French in 1966–1969 and the 2008 scientific expedition. The island has long been an attractive destination for amateur radio groups, due to its remoteness, the difficulty of landing, permit requirements, garish history, and interesting environment. While some radio operation was done ancillary to other expeditions, major DX-peditions include FO0XB (1978), FO0XX (1985), FO0CI (1992), FO0AAA (2000), and TX5C (2008). One DX-pedition was the Cordell Expedition in March 2013 using the callsign TX5K, organized and led by Robert Schmieder. The project combined radio operations with selected scientific investigations. The team of 24 radio operators made more than 114,000 contacts, breaking the previous record of 75,000. The activity included extensive operation at 6 meters, including EME (Earth–Moon–Earth communication or 'moonbounce') contacts. A notable accomplishment was the use of DXA, a real-time satellite-based online graphic radio log web page that allowed anyone anywhere with a browser to see the radio activity. Scientific work carried out during the expedition included the first collection and identification of foraminifera, and extensive aerial imaging of the island using kite-borne cameras. The team included two scientists from the French-Polynesian University of Tahiti and a TV crew from the French documentary television series Thalassa. An April 2015 DX-pedition using callsign TX5P was conducted by Alain Duchauchoy, F6BFH, concurrent with the Passion 2015 scientific expedition to Clipperton Island, and engaging in research of Mexican use of the island during the early 1900s.
https://en.wikipedia.org/wiki?curid=5510
Cocos (Keeling) Islands The Territory of Cocos (Keeling) Islands is an Australian external territory in the Indian Ocean, comprising a small archipelago approximately midway between Australia and Sri Lanka and closer to the Indonesian island of Sumatra. It is part of Southeast Asia and is in the Southern Hemisphere. The territory's dual name (official since the islands’ incorporation into Australia in 1955) reflects that the islands have historically been known as either the Cocos Islands or the Keeling Islands. The territory consists of two atolls made up of 27 coral islands, of which only two – West Island and Home Island – are inhabited. The population of around 600 people consists mainly of Cocos Malays, who mostly practise Sunni Islam and speak a dialect of Malay as their first language. The territory is administered by the Australian federal government's Department of Infrastructure, Regional Development and Cities, and together with Christmas Island (which is about to the east) forms the Australian Indian Ocean Territories administrative unit. However, the islanders do have a degree of self-government through the local shire council. Many public services – including health, education, and policing – are provided by the state of Western Australia, and Western Australian law applies except where the federal government has determined otherwise. The islands were discovered in 1609 by William Keeling, but no settlement occurred until the early 19th century. One of the first settlers was John Clunies-Ross, a Scottish merchant; much of the island's current population is descended from the Malay workers he brought in to work his copra plantation. The Clunies-Ross family ruled the islands as a private fiefdom for almost 150 years, with the head of the family usually recognised as resident magistrate. The British annexed the islands in 1857, and for the next century they were administered from either Ceylon or Singapore. The territory was transferred to Australia in 1955, although until 1979 virtually all of the island's real estate still belonged to the Clunies-Ross family. The islands have been called the "Cocos Islands" (from 1622), the "Keeling Islands" (from 1703), the "Cocos–Keeling Islands" (since James Horsburgh in 1805) and the "Keeling–Cocos Islands" (19th century). "Cocos" refers to the abundant coconut trees, while "Keeling" is William Keeling, who discovered the islands in 1609. John Clunies-Ross, who sailed there in the "Borneo" in 1825, called the group the "Borneo Coral Isles", restricting "Keeling" to North Keeling, and calling South Keeling "the Cocos properly so called". The form "Cocos (Keeling) Islands", attested from 1916, was made official by the Cocos (Keeling) Islands Act 1955. The territory's Malay name is "Pulu Kokos (Keeling)". Sign boards on the island also feature Malay translations. The Cocos (Keeling) Islands consist of two flat, low-lying coral atolls with an area of , of coastline, a highest elevation of and thickly covered with coconut palms and other vegetation. The climate is pleasant, moderated by the southeast trade winds for about nine months of the year and with moderate rainfall. Tropical cyclones may occur in the early months of the year. North Keeling Island is an atoll consisting of just one C-shaped island, a nearly closed atoll ring with a small opening into the lagoon, about wide, on the east side. The island measures in land area and is uninhabited. The lagoon is about . North Keeling Island and the surrounding sea to from shore form the Pulu Keeling National Park, established on 12 December 1995. It is home to the only surviving population of the endemic, and endangered, Cocos Buff-banded Rail. South Keeling Islands is an atoll consisting of 24 individual islets forming an incomplete atoll ring, with a total land area of . Only Home Island and West Island are populated. The Cocos Malays maintain weekend shacks, referred to as pondoks, on most of the larger islands. There are no rivers or lakes on either atoll. Fresh water resources are limited to water lenses on the larger islands, underground accumulations of rainwater lying above the seawater. These lenses are accessed through shallow bores or wells. Cocos (Keeling) Islands experiences tropical rainforest climate ("Af") according to the Köppen climate classification; the archipelago lies approximately midway between the Equator and the Tropic of Capricorn. The archipelago has two distinct seasons, the wet season and the dry season. The wettest month is April with precipitation totaling , and the driest month is October with precipitation totaling . The temperature varies a little as its location is some distance from the Equator. The hottest month is March with an average high temperature of , while the coolest month is August with an average low temperature of . In the , the population of the islands was 544 people of which 75% are Islamic and 69.6% speak Malay rather than English at home. The population on the two inhabited islands generally is split between the ethnic Europeans on West Island (estimated population 100) and the ethnic Malays on Home Island (estimated population 500). The archipelago was discovered in 1609 by Captain William Keeling of the East India Company, on a return voyage from the East Indies. North Keeling was sketched by Ekeberg, a Swedish captain, in 1749, showing the presence of coconut palms. It also appears on a 1789 chart produced by British hydrographer Alexander Dalrymple. In 1825, Scottish merchant seaman Captain John Clunies-Ross stopped briefly at the islands on a trip to India, nailing up a Union Jack and planning to return and settle on the islands with his family in the future. Wealthy Englishman Alexander Hare had similar plans, and hired a captain coincidentally, Clunies-Ross's brotherto bring him and a volunteer harem of 40 Malay women to the islands, where he hoped to establish his private residence. Hare had previously served as resident of Banjarmasin, a town in Borneo, and found that "he could not confine himself to the tame life that civilisation affords". Clunies-Ross returned two years later with his wife, children and mother-in-law, and found Hare already established on the island and living with the private harem. A feud grew between the two. Clunies-Ross's eight sailors "began at once the invasion of the new kingdom to take possession of it, women and all". After some time, Hare's women began deserting him, and instead finding themselves mates amongst Clunies-Ross's sailors. Disheartened, Hare left the island. He died in Bencoolen in 1834. Encouraged by members of the former harem, Clunies-Ross then recruited Malays to come to the island for work and wives. Clunies-Ross's workers were paid in a currency called the Cocos rupee, a currency John Clunies-Ross minted himself that could only be redeemed at the company store. On 1 April 1836, under Captain Robert FitzRoy arrived to take soundings to establish the profile of the atoll as part of the survey expedition of the "Beagle". To the naturalist Charles Darwin, aboard the ship, the results supported a theory he had developed of how atolls formed, which he later published as "The Structure and Distribution of Coral Reefs". He studied the natural history of the islands and collected specimens. Darwin's assistant Syms Covington noted that "an Englishman [he was in fact Scottish] and HIS family, with about sixty or seventy mulattos from the Cape of Good Hope, live on one of the islands. Captain Ross, the governor, is now absent at the Cape." The islands were annexed by the British Empire in 1857. This annexation was carried out by Captain Stephen Grenville Fremantle in command of . Fremantle claimed the islands for the British Empire and appointed Ross II as Superintendent. In 1878, by Letters Patent, the Governor of Ceylon was made Governor of the islands, and, by further Letters Patent in 1886, responsibility for the islands was transferred to the Governor of the Straits Settlement to exercise his functions as "Governor of Cocos Islands". The islands were made part of the Straits Settlement under an Order in Council of 20 May 1903. Meanwhile, in 1886 Queen Victoria had, by indenture, granted the islands in perpetuity to John Clunies-Ross. The head of the family enjoyed semi-official status as Resident Magistrate and Government representative. In 1901 a telegraph cable station was established on Direction Island. Undersea cables went to Rodrigues, Mauritius, Batavia, Java and Fremantle, Western Australia. In 1910 a wireless station was established to communicate with passing ships. The cable station ceased operation in 1966. On the morning of 9 November 1914, the islands became the site of the Battle of Cocos, one of the first naval battles of World War I. A landing party from the German cruiser captured and disabled the wireless and cable communications station on Direction Island, but not before the station was able to transmit a distress call. An Allied troop convoy was passing nearby, and the Australian cruiser was detached from the convoy escort to investigate. "Sydney" spotted the island and "Emden" at 09:15, with both ships preparing for combat. At 11:20, the heavily damaged "Emden" beached herself on North Keeling Island. The Australian warship broke to pursue "Emden"s supporting collier, which scuttled herself, then returned to North Keeling Island at 16:00. At this point, "Emden"s battle ensign was still flying: usually a sign that a ship intends to continue fighting. After no response to instructions to lower the ensign, two salvoes were shot into the beached cruiser, after which the Germans lowered the flag and raised a white sheet. "Sydney" had orders to ascertain the status of the transmission station, but returned the next day to provide medical assistance to the Germans. Casualties totaled 134 personnel aboard "Emden" killed, and 69 wounded, compared to four killed and 16 wounded aboard "Sydney". The German survivors were taken aboard the Australian cruiser, which caught up to the troop convoy in Colombo on 15 November, then transported to Malta and handed over the prisoners to the British Army. An additional 50 German personnel from the shore party, unable to be recovered before "Sydney" arrived, commandeered a schooner and escaped from Direction Island, eventually arriving in Constantinople. "Emden" was the last active Central Powers warship in the Indian or Pacific Ocean, which meant troopships from Australia and New Zealand could sail without naval escort, and Allied ships could be deployed elsewhere. During World War II, the cable station was once again a vital link. The Cocos were valuable for direction finding by the Y service, the worldwide intelligence system used during the war. Allied planners noted that the islands might be seized as an airfield for German planes and as a base for commerce raiders operating in the Indian Ocean. Following Japan's entry into the war, Japanese forces occupied neighbouring islands. To avoid drawing their attention to the Cocos cable station and its islands' garrison, the seaplane anchorage between Direction and Horsburgh islands was not used. Radio transmitters were also kept silent, except in emergencies. After the Fall of Singapore in 1942, the islands were administered from Ceylon (Sri Lanka), and West and Direction Islands were placed under Allied military administration. The islands' garrison initially consisted of a platoon from the British Army's King's African Rifles, located on Horsburgh Island, with two guns to cover the anchorage. The local inhabitants all lived on Home Island. Despite the importance of the islands as a communication centre, the Japanese made no attempt either to raid or to occupy them and contented themselves with sending over a reconnaissance aircraft about once a month. On the night of 8–9 May 1942, 15 members of the garrison, from the Ceylon Defence Force, mutinied under the leadership of Gratien Fernando. The mutineers were said to have been provoked by the attitude of their British officers and were also supposedly inspired by Japanese anti-British propaganda. They attempted to take control of the gun battery on the islands. The Cocos Islands Mutiny was crushed, but the mutineers murdered one non-mutinous soldier and wounded one officer. Seven of the mutineers were sentenced to death at a trial that was later alleged to have been improperly conducted, though the guilt of the accused was admitted. Four of the sentences were commuted, but three men were executed, including Fernando. These were to be the only British Commonwealth soldiers executed for mutiny during the Second World War. On 25 December 1942, the Japanese submarine "I-166" bombarded the islands but caused no damage. Later in the war, two airstrips were built, and three bomber squadrons were moved to the islands to conduct raids against Japanese targets in South East Asia and to provide support during the planned reinvasion of Malaya and reconquest of Singapore. The first aircraft to arrive were Supermarine Spitfire Mk VIIIs of No. 136 Squadron RAF. They included some Liberator bombers from No. 321 (Netherlands) Squadron RAF (members of exiled Dutch forces serving with the Royal Air Force), which were also stationed on the islands. When in July 1945 No. 99 and No. 356 RAF squadrons arrived on West Island, they brought with them a daily newspaper called "Atoll" which contained news of what was happening in the outside world. Run by airmen in their off-duty hours, it achieved fame when dropped by Liberator bombers on POW camps over the heads of the Japanese guards. In 1946, the administration of the islands reverted to Singapore and it became part of the Colony of Singapore. On 23 November 1955, the islands were transferred from the United Kingdom to the Commonwealth of Australia. Immediately before the transfer the islands were part of the United Kingdom's Colony of Singapore, in accordance with the Straits Settlements (Repeal) Act, 1946 of the United Kingdom and the British Settlements Acts, 1887 and 1945, as applied by the Act of 1946. The legal steps for effecting the transfer were as follows: The reason for this comparatively complex machinery was due to the terms of the Straits Settlement (Repeal) Act, 1946. According to Sir Kenneth Roberts-Wray "any other procedure would have been of doubtful validity". The separation involved three steps: separation from the Colony of Singapore; transfer by United Kingdom and acceptance by Australia. H. J. Hull was appointed the first official representative (now administrator) of the new territory. He had been a lieutenant-commander in the Royal Australian Navy and was released for the purpose. Under Commonwealth Cabinet Decision 1573 of 9 September 1958, Hull's appointment was terminated and John William Stokes was appointed on secondment from the Northern Territory police. A media release at the end of October 1958 by the Minister for Territories, Hasluck, commended Hull's three years of service on Cocos. Stokes served in the position from 31 October 1958 to 30 September 1960. His son's boyhood memories and photos of the Islands have been published. C. I. Buffett MBE from Norfolk Island succeeded him and served from 28 July 1960 to 30 June 1966, and later acted as Administrator back on Cocos and on Norfolk Island. In 1974, Ken Mullen wrote a small book about his time with wife and son from 1964 to 1966 working at the Cable Station on Direction Island. In the 1970s, the Australian government's dissatisfaction with the Clunies-Ross feudal style of rule of the island increased. In 1978, Australia forced the family to sell the islands for the sum of , using the threat of compulsory acquisition. By agreement, the family retained ownership of Oceania House, their home on the island. In 1983, the Australian government reneged on this agreement and told John Clunies-Ross that he should leave the Cocos. The following year the High Court of Australia ruled that resumption of Oceania House was unlawful, but the Australian government ordered that no government business was to be granted to Clunies-Ross's shipping company, an action that contributed to his bankruptcy. John Clunies-Ross now lives in Perth, Western Australia. However, some members of the Clunies-Ross family still live on the Cocos. Extensive preparations were undertaken by the government of Australia to prepare the Cocos Malays to vote in their referendum of self-determination. Discussions began in 1982, with an aim of holding the referendum, under United Nations supervision, in mid-1983. Under guidelines developed by the UN Decolonization Committee, residents were to be offered three choices: full independence, free association, or integration with Australia. The last option was preferred by both the islanders and the Australian government. A change in government in Canberra following the March 1983 Australian elections delayed the vote by one year. While the Home Island Council stated a preference for a traditional communal consensus "vote", the UN insisted on a secret ballot. The referendum was held on 6 April 1984, with all 261 eligible islanders participating, including the Clunies-Ross family: 229 voted for integration, 21 for Free Association, nine for independence, and two failed to indicate a preference. In recent years a series of disputes have occurred between the Muslim and the non-Muslim population of the islands. Descendants of the Cocos Malays who were brought to the islands from British Malaya and the Dutch East Indies as well as South Africa and New Guinea by Hare and by Clunies-Ross as indentured workers, slaves or convicts are seeking recognition from the Australian Federal Government to be acknowledged as Indigenous Australians. The capital of the Territory of Cocos (Keeling) Islands is West Island while the largest settlement is the village of Bantam (Home Island). Governance of the islands is based on the "Cocos (Keeling) Islands Act 1955" and depends heavily on the laws of Australia. The islands are administered from Canberra by the Department of Infrastructure, Regional Development and Cities through a non-resident Administrator appointed by the Governor-General. They were previously the responsibility of the Department of Transport and Regional Services (before 2007), the Attorney-General's Department (2007–2013), and Department of Infrastructure and Regional Development (2013–2017) The current Administrator is Natasha Griggs, who was appointed on 5 October 2017 and is also the Administrator of Christmas Island. These two Territories comprise the Australian Indian Ocean Territories. The Australian Government provides Commonwealth-level government services through the Christmas Island Administration and the Department of Infrastructure and Regional Development. As per the Federal Government's Territories Law Reform Act 1992, which came into force on 1 July 1992, Western Australian laws are applied to the Cocos Islands, "so far as they are capable of applying in the Territory."; non-application or partial application of such laws is at the discretion of the federal government. The Act also gives Western Australian courts judicial power over the islands. The Cocos Islands remain constitutionally distinct from Western Australia, however; the power of the state to legislate for the territory is power delegated by the federal government. The kind of services typically provided by a state government elsewhere in Australia are provided by departments of the Western Australian Government, and by contractors, with the costs met by the federal government. There also exists a unicameral Cocos (Keeling) Islands Shire Council with seven seats. A full term lasts four years, though elections are held every two years; approximately half the members retire each two years. Cocos (Keeling) Islands residents who are Australian citizens also vote in federal elections. Cocos (Keeling) Islanders are represented in the House of Representatives by the member for the Division of Lingiari (in the Northern Territory) and in the Senate by Northern Territory senators. At the 2016 federal election, the Labor Party received absolute majorities from Cocos electors in both the House of Representatives and the Senate. Defence is the responsibility of the Australian Defence Force. There are no active military installations or defence personnel on the islands. The Administrator may request the assistance of the Australian Defence Force if required. The 2016 Australian Defence White Paper stated that the airfield in the island would be upgraded to support the RAAF's P-8 Poseidon maritime patrol aircraft. Civilian law enforcement and community policing is provided by the Australian Federal Police. The normal deployment to the island is one sergeant and one constable. These are augmented by two locally engaged Special Members who have police powers. Since 1992, court services have been provided by the Western Australian Department of the Attorney-General under a service delivery arrangement with the Australian Government. Western Australian Court Services provide Magistrates Court, District Court, Supreme Court, Family Court, Children's Court, Coroner's Court and Registry for births, deaths and marriages and change of name services. Magistrates and judges from Western Australia convene a circuit court as required. Home Island and West Island have medical clinics providing basic health services, but serious medical conditions and injuries cannot be treated on the island and patients are sent to Perth for treatment, a distance of . The population of the islands is approximately 600. There is a small and growing tourist industry focused on water-based or nature activities. In 2016, a beach on Direction Island was named the best beach in Australia by Brad Farmer, an Aquatic and Coastal Ambassador for Tourism Australia and co-author of "101 Best Beaches 2017". Small local gardens and fishing contribute to the food supply, but most food and most other necessities must be imported from Australia or elsewhere. The Cocos Islands Cooperative Society Ltd. employs construction workers, stevedores, and lighterage worker operations. Tourism employs others. The unemployment rate was 6.7% in 2011. A 2019 study led by Jennifer Lavers from the University of Tasmania’s Institute of Marine and Antarctic Studies published in the journal "Scientific Reports" estimated the volume of plastic rubbish on the Islands as around 414 million pieces, weighing 238 tonnes, 93% of which lies buried under the sand. It said that previous surveys which only assessed surface garbage probably “drastically underestimated the scale of debris accumulation”. The plastic waste found in the study consisted mostly of single-use items such as bottles, plastic cutlery, bags and drinking straws. The Cocos Islands are strategically important because of their proximity to shipping lanes in the Indian and Pacific oceans. The United States and Australia have expressed interest in stationing surveillance drones on the Cocos Islands. Euronews described the plan as Australian support for an increased American presence in Southeast Asia, but expressed concern that it was likely to upset Chinese officials. James Cogan has written for the World Socialist Web Site that the plan to station surveillance drones at Cocos is one component of former US President Barack Obama's "pivot" towards Asia, facilitating control of the sea lanes and potentially allowing US forces to enforce a blockade against China. After plans to construct airbases were leaked to the "Washington Post", Australian defence minister Stephen Smith stated that the Australian government views the "Cocos as being potentially a long-term strategic location, but that is down the track." The Cocos (Keeling) Islands have of highway. There is one paved airport on the West Island. A tourist bus operates on Home Island. The only airport is Cocos (Keeling) Islands Airport with a single paved runway. Virgin Australia operates scheduled jet services from Perth Airport via Christmas Island. After 1952, the airport at Cocos Islands was a stop for airline flights between Australia and South Africa, and Qantas and South African Airways stopped there to refuel. The arrival of long-range jet aircraft ended this need in 1967. An interisland ferry, the "Cahaya Baru", connects West, Home and Direction Islands. There is a lagoon anchorage between Horsburgh and Direction islands for larger vessels, while yachts have a dedicated anchorage area in the southern lee of Direction Island. There are no major seaports on the islands. The islands are connected within Australia's telecommunication system (with number range +61 8 9162 xxxx). Public phones are located on both West Island and Home Island. A reasonably reliable GSM mobile phone network (number range +61 406 xxx), run by CiiA (Christmas Island Internet Association), operates on Cocos (Keeling) Islands. SIM cards (full size) and recharge cards can be purchased from the Telecentre on West Island to access this service. Australia Post provides mail services with the postcode 6799. There are post offices on West Island and Home Island. Standard letters and express post items are sent by air twice weekly, but all other mail is sent by sea and can take up to two months for delivery. .cc is the Internet country code top-level domain (ccTLD) for Cocos (Keeling) Islands. It is administered by VeriSign through a subsidiary company eNIC, which promotes it for international registration as "the next .com"; .cc was originally assigned in October 1997 to eNIC Corporation of Seattle WA by the IANA. The Turkish Republic of Northern Cyprus also uses the .cc domain, along with .nc.tr. Internet access on Cocos is provided by CiiA (Christmas Island Internet Association), and is supplied via satellite ground station on West Island, and distributed via a wireless PPPoE-based WAN on both inhabited islands. Casual internet access is available at the Telecentre on West Island and the Indian Ocean Group Training office on Home Island. The National Broadband Network announced in early 2012 that it would extend service to Cocos in 2015 via high-speed satellite link. The Cocos (Keeling) Islands have access to a range of modern communication services. Digital television stations are broadcast from Western Australia via satellite. A local radio station, 6CKI – Voice of the Cocos (Keeling) Islands, is staffed by community volunteers and provides some local content. The Cocos (Keeling) Islands receives a range of digital channels from Western Australia via satellite and is broadcast from the Airport Building on the West Island on the following VHF frequencies: ABC6, SBS7, WAW8, WOW10 and WDW11 From 2013 onwards, Cocos Island will receive four Malaysian channels via satellite: TV3, ntv7, 8TV and TV9. There is a school in the archipelago, Cocos Islands District High School, with campuses located on West Island (Kindergarten to Year 10), and the other on Home Island (Kindergarten to Year 6). CIDHS is part of the Western Australia Department of Education. School instruction is in English on both campuses, with Cocos Malay teacher aides assisting the younger children in Kindergarten, Pre-Preparatory and early Primary with the English curriculum on the Home Island Campus. The Home Language of Cocos Malay is valued whilst students engage in learning English. The culture of the Islands is essentially Western, with extensive influences from Malaysia and Indonesia. The West Island Mosque on Alexander Street is listed on the Australian Commonwealth Heritage List.
https://en.wikipedia.org/wiki?curid=5520